{ "cells": [ { "cell_type": "markdown", "metadata": {}, "source": [ "### Dataset Loading and testing" ] }, { "cell_type": "code", "execution_count": 1, "metadata": { "ExecuteTime": { "end_time": "2018-10-18T16:08:14.804305Z", "start_time": "2018-10-18T16:08:14.300247Z" } }, "outputs": [ { "name": "stdout", "output_type": "stream", "text": [ "Dataset has been loaded\n", "x-train (60000, 8)\n", "x-test (40000, 8)\n", "y-train (60000, 256)\n", "y-test (40000, 256)\n" ] } ], "source": [ "%load_ext autoreload\n", "%autoreload 2\n", "\n", "import numpy as np\n", "import matplotlib.pyplot as plt\n", "import h5py\n", "from sklearn.model_selection import train_test_split\n", "#import jtplot submodule from jupyterthemes\n", "from jupyterthemes import jtplot\n", "#currently installed theme will be used to\n", "#set plot style if no arguments provided\n", "jtplot.style()\n", "\n", "#now load this dataset \n", "h5f = h5py.File('./datasets/s8_sio2tio2_v2.h5','r')\n", "X = h5f['sizes'][:]\n", "Y = h5f['spectrum'][:]\n", "\n", "\n", "\n", "\n", "#get the ranges of the loaded data\n", "num_layers = X.shape[1]\n", "num_lpoints = Y.shape[1]\n", "size_max = np.amax(X)\n", "size_min = np.amin(X)\n", "size_av = 0.5*(size_max + size_min)\n", "\n", "#this information is not given in the dataset\n", "lam_min = 300\n", "lam_max = 1200\n", "lams = np.linspace(lam_min, lam_max, num_lpoints)\n", "\n", "# X = np.expand_dims(X, 1)\n", "# #X = np.expand_dims(X, 3)\n", "# Y = np.expand_dims(Y, 1)\n", "# #Y = np.expand_dims(Y, 3)\n", "\n", "\n", "\n", "\n", "\n", "\n", "#create a train - test split of the dataset\n", "x_train, x_test, y_train, y_test = train_test_split(X, Y, test_size=0.4, random_state=42)\n", "\n", "# normalize inputs \n", "x_train = (x_train - 50)/20 \n", "x_test = (x_test - 50)/20 \n", "\n", "print(\"Dataset has been loaded\")\n", "print(\"x-train\", x_train.shape)\n", "print(\"x-test \", x_test.shape)\n", "print(\"y-train\", y_train.shape)\n", "print(\"y-test \", y_test.shape)\n" ] }, { "cell_type": "markdown", "metadata": {}, "source": [ "#### Model Development" ] }, { "cell_type": "code", "execution_count": null, "metadata": { "ExecuteTime": { "end_time": "2018-09-27T05:04:59.838628Z", "start_time": "2018-09-27T05:04:59.808085Z" } }, "outputs": [], "source": [ "from keras.utils import to_channels_first\n", "x_train = to_channels_first(x_train)\n" ] }, { "cell_type": "code", "execution_count": 2, "metadata": { "ExecuteTime": { "end_time": "2018-10-18T16:10:50.680462Z", "start_time": "2018-10-18T16:10:49.935191Z" }, "code_folding": [ 48, 65, 84, 123, 169, 234 ] }, "outputs": [ { "name": "stderr", "output_type": "stream", "text": [ "Using MXNet backend\n" ] } ], "source": [ "from keras import backend as K\n", "from keras.models import Sequential, Model\n", "from keras.layers import Dense, Dropout, Reshape, UpSampling1D, Conv1D, Flatten, Activation\n", "from keras.utils import np_utils, multi_gpu_model\n", "from keras.regularizers import l2\n", "from keras.wrappers.scikit_learn import KerasRegressor\n", "from keras.optimizers import Adam\n", "from keras.layers.normalization import BatchNormalization\n", "from keras.layers import PReLU\n", "\n", "\n", "from sklearn.model_selection import cross_val_score, KFold\n", "from sklearn.preprocessing import StandardScaler\n", "from sklearn.pipeline import Pipeline\n", "\n", "num_gpus = 2\n", "gpu_list = [\"gpu(%d)\" % i for i in range(num_gpus)]\n", "\n", "\n", "\n", "#define various models here\n", "#naive percentage loss\n", "def size_percent_loss(y_true, y_pred):\n", " y_true_a = 0.5*y_true*(size_max - size_min) + size_av\n", " y_pred_a = 0.5*y_pred*(size_max - size_min) + size_av\n", " y_err = np.abs(y_true_a - y_pred_a)/y_true_a\n", " y_err_f = K.flatten(y_err)\n", " return K.sum(y_err_f)\n", "\n", "#naive percentage loss\n", "def naive_percent_loss(y_true, y_pred):\n", " y_err = np.abs(y_true - y_pred)/y_true\n", " y_err_f = K.flatten(y_err)\n", " return K.sum(y_err_f)\n", "\n", "\n", "#function to test performance on testset \n", "def calc_mre(y_true, y_pred):\n", " y_err = 100*np.abs(y_true - y_pred)/y_true\n", " return np.mean(y_err)\n", "\n", "#function to test performance on testset \n", "def calc_mre_K(y_true, y_pred):\n", " y_err = 100*np.abs(y_true - y_pred)/y_true\n", " return K.mean(y_err)\n", "\n", "\n", "\n", "def naiveploss_mgpu_model():\n", " # create model\n", " model = Sequential()\n", " model = multi_gpu_model(model, gpus=num_gpus)\n", " model.add(Dense(250, input_dim=x_train.shape[1], kernel_initializer='normal', activation='relu', \n", " name='first' ))\n", " model.add(Dense(250, input_dim=x_train.shape[1], kernel_initializer='normal', activation='relu', \n", " name='second' ))\n", " model.add(Dense(250, input_dim=x_train.shape[1], kernel_initializer='normal', activation='relu', \n", " name='third' ))\n", " model.add(Dense(250, input_dim=x_train.shape[1], kernel_initializer='normal', activation='relu', \n", " name='fourth' ))\n", " model.add(Dense(250, kernel_initializer='normal', name='last'))\n", " # Compile model\n", " model.compile(loss=naive_percent_loss, optimizer='adam', context = gpu_list)\n", " return model\n", "\n", "def naiveploss_model():\n", " # create model\n", " model = Sequential()\n", " model.add(Dense(256, input_dim=x_train.shape[1], kernel_initializer='normal', activation='relu', \n", " name='first' ))\n", " model.add(Dense(256, input_dim=x_train.shape[1], kernel_initializer='normal', activation='relu', \n", " name='second' ))\n", " model.add(Dense(256, input_dim=x_train.shape[1], kernel_initializer='normal', activation='relu', \n", " name='third' ))\n", " model.add(Dense(256, input_dim=x_train.shape[1], kernel_initializer='normal', activation='relu', \n", " name='fourth' ))\n", " model.add(Dense(256, kernel_initializer='normal', name='last'))\n", " # Compile model\n", " model.compile(loss=naive_percent_loss, optimizer='adam', metrics=['accuracy'])\n", " return model\n", "\n", "import timeit\n", "#here we must have a function that calls the training routine n times and then gives avg and stddev \n", "# of the resulting figures\n", "def net_performance(modelfunc, num_trials=3, batch_size=32, num_epochs=200, num_gpus=2):\n", " models = []\n", " train_err = np.ones(num_trials)\n", " test_err = np.ones(num_trials)\n", " val_err = np.ones(num_trials)\n", " train_time = np.ones(num_trials)\n", " for tnum in np.arange(num_trials):\n", " print(\"iteration: \" + str(tnum + 1))\n", " model_curr = modelfunc()\n", " x_t, x_v, y_t, y_v = train_test_split(x_train, y_train, test_size=0.2, random_state=42)\n", " start_time = timeit.default_timer()\n", " history = model_curr.fit(x_t, y_t,\n", " batch_size=batch_size*num_gpus,\n", " epochs=num_epochs, \n", " verbose=1,\n", " validation_data=(x_v, y_v))\n", " train_time[tnum] = timeit.default_timer() - start_time\n", " models.append(model_curr)\n", " train_err[tnum] = (100.0/num_lpoints)*history.history['loss'][-1]/(batch_size*num_gpus)\n", " val_err[tnum] = (100.0/num_lpoints)*history.history['val_loss'][-1]/(batch_size*num_gpus)\n", " test_err[tnum] = calc_mre(y_test, models[tnum].predict(x_test))\n", " return train_err, val_err, test_err, train_time\n", "\n", "#staging area for new models \n", "def plot_training_history(history, factor):\n", " loss, val_loss = history.history['loss'], history.history['val_loss']\n", " loss = np.asarray(loss)/(factor)\n", " val_loss = np.asarray(val_loss)/(factor)\n", " epochs = len(loss)\n", " \n", " fig, axs = plt.subplots(1,1, figsize=(5,2.5))\n", " axs.semilogy(np.arange(1, epochs + 1), loss, label='Train error')\n", " axs.semilogy(np.arange(1, epochs + 1), val_loss, label='Test error')\n", " axs.set_xlabel('Epoch number')\n", " #axs.set_ylim((0.4, 3))\n", " axs.set_xlim(left=1)\n", "# plt.yticks(np.array([0.1, 0.2, 0.3, 0.4, 0.5, 0.75, 1.0, 1.5, 2]), \n", "# ('0.1', '0.2', '0.3', '0.4', '0.5', '0.75', '1.0', '1.5', '2'))\n", " plt.yticks(np.array([0.5, 0.75, 1.0, 1.5, 2]), \n", " ('0.5', '0.75', '1.0', '1.5', '2'))\n", " axs.set_ylabel('MRE (%)')\n", " axs.legend(loc=\"best\")\n", " fig.savefig(\"foo2.pdf\", bbox_inches='tight')\n", "\n", "from keras.utils import to_channels_first\n", "\n", "def conv1d_lkyrelu():\n", " \n", " #gpu_list = [\"gpu(%d)\" % i for i in range(num_gpus)]\n", " \n", " # create model\n", " model = Sequential()\n", " \n", " model.add(Dense(256, input_dim=8, kernel_initializer='normal', \n", " name='first'))\n", " #model.add(BatchNormalization())\n", " #model.add(Activation('relu')) \n", " model.add(PReLU(alpha_initializer='zeros', alpha_regularizer=None))\n", " \n", " model.add(Reshape((4, 64)))\n", " model.add(UpSampling1D(size=2))\n", " \n", " \n", " model.add(Conv1D(filters=64, kernel_size=3, strides=1, padding='same', \n", " dilation_rate=1, \n", " kernel_initializer='normal'))\n", " model.add(PReLU(alpha_initializer='zeros', alpha_regularizer=None))\n", " \n", "# model.add(Conv1D(filters=32, kernel_size=3, strides=1, padding='same', \n", "# dilation_rate=1, kernel_initializer='normal'))\n", "# model.add(PReLU(alpha_initializer='zeros', alpha_regularizer=None))\n", " \n", " model.add(Conv1D(filters=32, kernel_size=3, strides=1, padding='same', \n", " dilation_rate=1, \n", " kernel_initializer='normal'))\n", " model.add(PReLU(alpha_initializer='zeros', alpha_regularizer=None))\n", "\n", "\n", "# model.add(Conv1D(filters=32, kernel_size=3, strides=1, padding='same', \n", "# dilation_rate=1, kernel_initializer='normal'))\n", "# model.add(PReLU(alpha_initializer='ones', alpha_regularizer=None)) \n", " \n", "# model.add(Conv1D(filters=32, kernel_size=3, strides=1, padding='same', \n", "# dilation_rate=1, kernel_initializer='normal'))\n", "# model.add(PReLU(alpha_initializer='ones', alpha_regularizer=None)) \n", " \n", " model.add(Flatten())\n", " # Compile model\n", " model.compile(loss=naive_percent_loss, optimizer='adam', metrics=[calc_mre_K])\n", " return model\n", " \n", " \n", "def conv1d_model_bnorm():\n", " \n", " #gpu_list = [\"gpu(%d)\" % i for i in range(num_gpus)]\n", " \n", " # create model\n", " model = Sequential()\n", " \n", " model.add(Dense(256, input_dim=8, kernel_initializer='normal', \n", " name='first' ))\n", " #model.add(BatchNormalization())\n", " model.add(Activation('relu'))\n", " #model.add(Dropout(0.2))\n", " \n", " model.add(Reshape((4, 64)))\n", " model.add(UpSampling1D(size=2))\n", " \n", " model.add(Conv1D(filters=64, kernel_size=3, strides=1, padding='same', \n", " dilation_rate=1, kernel_initializer='normal'))\n", " #model.add(BatchNormalization())\n", " model.add(Activation('relu'))\n", " #model.add(Dropout(0.2))\n", " #model.add(UpSampling1D(size=5))\n", "\n", " model.add(Conv1D(filters=32, kernel_size=3, strides=1, padding='same', \n", " dilation_rate=1, kernel_initializer='normal'))\n", " model.add(Activation('relu'))\n", " #model.add(Dropout(0.3))\n", " \n", "# model.add(Conv1D(filters=64, kernel_size=3, strides=1, padding='same', \n", "# dilation_rate=1, kernel_initializer='normal'))\n", "# model.add(Activation('relu')) \n", "# model.add(Dropout(0.3))\n", "\n", "# model.add(Conv1D(filters=32, kernel_size=3, strides=1, padding='same', \n", "# dilation_rate=1, kernel_initializer='normal'))\n", "# model.add(Activation('relu')) \n", " \n", "# model.add(Conv1D(filters=32, kernel_size=3, strides=1, padding='same', \n", "# dilation_rate=1, kernel_initializer='normal'))\n", "# model.add(Activation('relu')) \n", " \n", "# model.add(Conv1D(filters=32, kernel_size=3, strides=1, padding='same', \n", "# dilation_rate=1, kernel_initializer='normal'))\n", "# model.add(Activation('relu')) \n", " \n", "# model.add(Conv1D(filters=32, kernel_size=3, strides=1, padding='same', \n", "# dilation_rate=1, kernel_initializer='normal'))\n", "# model.add(Activation('relu')) \n", " \n", " model.add(Conv1D(filters=32, kernel_size=3, strides=1, padding='same', \n", " dilation_rate=1, kernel_initializer='normal'))\n", " model.add(Activation('relu')) \n", " \n", " \n", " \n", " model.add(Flatten())\n", " # Compile model\n", "# if num_gpus == 1:\n", " model.compile(loss=naive_percent_loss, optimizer='adam', metrics=[calc_mre_K])\n", "# else:\n", "# model.compile(loss=naive_percent_loss, optimizer='adam', metrics=['accuracy'], context = gpu_list)\n", " \n", " \n", " return model \n", "\n", "def resnetb():\n", " model = Sequential()\n", " \n", " #first layer6\n", " model.add(Dense(256, input_dim=8, kernel_initializer='normal', \n", " name='first'))\n", " model.add(PReLU(alpha_initializer='zeros', alpha_regularizer=None))\n", " model.add(Reshape((8, 32)))\n", " \n", " \n", " #resnet block\n", "# model.add(Conv1D(filters=32, kernel_size=3, strides=1, padding='same', \n", "# dilation_rate=1, \n", "# kernel_initializer='normal'))\n", "# model.add(PReLU(alpha_initializer='zeros', alpha_regularizer=None))\n", "# model.add(Conv1D(filters=32, kernel_size=3, strides=1, padding='same', \n", "# dilation_rate=1, \n", "# kernel_initializer='normal'))\n", " \n", " \n", " \n", " #Last layer\n", " model.add(Flatten())\n", " \n", " #compile model\n", " model.compile(loss=naive_percent_loss, optimizer='adam', metrics=[calc_mre_K])\n", " \n", " return model\n" ] }, { "cell_type": "markdown", "metadata": {}, "source": [ "### Model testing" ] }, { "cell_type": "code", "execution_count": 4, "metadata": { "ExecuteTime": { "end_time": "2018-10-18T17:34:32.197328Z", "start_time": "2018-10-18T16:11:29.533784Z" }, "code_folding": [], "scrolled": true }, "outputs": [ { "name": "stderr", "output_type": "stream", "text": [ "/home/hegder/anaconda3/lib/python3.7/site-packages/keras/backend/mxnet_backend.py:89: UserWarning: MXNet Backend performs best with `channels_first` format. Using `channels_last` will significantly reduce performance due to the Transpose operations. For performance improvement, please use this API`keras.utils.to_channels_first(x_input)`to transform `channels_last` data to `channels_first` format and also please change the `image_data_format` in `keras.json` to `channels_first`.Note: `x_input` is a Numpy tensor or a list of Numpy tensorRefer to: https://github.com/awslabs/keras-apache-mxnet/tree/master/docs/mxnet_backend/performance_guide.md\n", " train_symbol = func(*args, **kwargs)\n", "/home/hegder/anaconda3/lib/python3.7/site-packages/keras/backend/mxnet_backend.py:92: UserWarning: MXNet Backend performs best with `channels_first` format. Using `channels_last` will significantly reduce performance due to the Transpose operations. For performance improvement, please use this API`keras.utils.to_channels_first(x_input)`to transform `channels_last` data to `channels_first` format and also please change the `image_data_format` in `keras.json` to `channels_first`.Note: `x_input` is a Numpy tensor or a list of Numpy tensorRefer to: https://github.com/awslabs/keras-apache-mxnet/tree/master/docs/mxnet_backend/performance_guide.md\n", " test_symbol = func(*args, **kwargs)\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "_________________________________________________________________\n", "Layer (type) Output Shape Param # \n", "=================================================================\n", "first (Dense) (None, 256) 2304 \n", "_________________________________________________________________\n", "p_re_lu_1 (PReLU) (None, 256) 256 \n", "_________________________________________________________________\n", "Reshape1 (Reshape) (None, 4, 64) 0 \n", "_________________________________________________________________\n", "Up1 (UpSampling1D) (None, 8, 64) 0 \n", "_________________________________________________________________\n", "Conv1 (Conv1D) (None, 8, 64) 12352 \n", "_________________________________________________________________\n", "p_re_lu_2 (PReLU) (None, 8, 64) 512 \n", "_________________________________________________________________\n", "Conv2 (Conv1D) (None, 8, 32) 6176 \n", "_________________________________________________________________\n", "p_re_lu_3 (PReLU) (None, 8, 32) 256 \n", "_________________________________________________________________\n", "Conv3 (Conv1D) (None, 8, 32) 3104 \n", "_________________________________________________________________\n", "p_re_lu_4 (PReLU) (None, 8, 32) 256 \n", "_________________________________________________________________\n", "Conv4 (Conv1D) (None, 8, 32) 3104 \n", "_________________________________________________________________\n", "p_re_lu_5 (PReLU) (None, 8, 32) 256 \n", "_________________________________________________________________\n", "Conv5 (Conv1D) (None, 8, 32) 3104 \n", "_________________________________________________________________\n", "p_re_lu_6 (PReLU) (None, 8, 32) 256 \n", "_________________________________________________________________\n", "flatten_1 (Flatten) (None, 256) 0 \n", "=================================================================\n", "Total params: 31,936\n", "Trainable params: 31,936\n", "Non-trainable params: 0\n", "_________________________________________________________________\n", "Train on 48000 samples, validate on 12000 samples\n", "Epoch 1/2000\n", " 1856/48000 [>.............................] - ETA: 35s - loss: 12264.2708 - calc_mre_K: 74.8552 " ] }, { "name": "stderr", "output_type": "stream", "text": [ "/home/hegder/anaconda3/lib/python3.7/site-packages/mxnet/module/bucketing_module.py:408: UserWarning: Optimizer created manually outside Module but rescale_grad is not normalized to 1.0/batch_size/num_workers (1.0 vs. 0.015625). Is this intended?\n", " force_init=force_init)\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "48000/48000 [==============================] - 4s 77us/step - loss: 1902.7889 - calc_mre_K: 11.6137 - val_loss: 775.0015 - val_calc_mre_K: 4.7364\n", "Epoch 2/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 636.6252 - calc_mre_K: 3.8857 - val_loss: 520.5962 - val_calc_mre_K: 3.1815\n", "Epoch 3/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 498.0257 - calc_mre_K: 3.0397 - val_loss: 450.9856 - val_calc_mre_K: 2.7561\n", "Epoch 4/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 444.5925 - calc_mre_K: 2.7136 - val_loss: 440.9772 - val_calc_mre_K: 2.6950\n", "Epoch 5/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 415.3938 - calc_mre_K: 2.5354 - val_loss: 428.2262 - val_calc_mre_K: 2.6171\n", "Epoch 6/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 408.2033 - calc_mre_K: 2.4915 - val_loss: 371.1729 - val_calc_mre_K: 2.2685\n", "Epoch 7/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 378.7598 - calc_mre_K: 2.3118 - val_loss: 450.6219 - val_calc_mre_K: 2.7540\n", "Epoch 8/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 358.7133 - calc_mre_K: 2.1894 - val_loss: 368.1862 - val_calc_mre_K: 2.2503\n", "Epoch 9/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 349.7285 - calc_mre_K: 2.1346 - val_loss: 357.1287 - val_calc_mre_K: 2.1827\n", "Epoch 10/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 332.7542 - calc_mre_K: 2.0310 - val_loss: 319.7993 - val_calc_mre_K: 1.9545\n", "Epoch 11/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 339.1140 - calc_mre_K: 2.0698 - val_loss: 316.9333 - val_calc_mre_K: 1.9370\n", "Epoch 12/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 316.8977 - calc_mre_K: 1.9342 - val_loss: 293.5551 - val_calc_mre_K: 1.7941\n", "Epoch 13/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 319.0048 - calc_mre_K: 1.9471 - val_loss: 303.8378 - val_calc_mre_K: 1.8570\n", "Epoch 14/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 308.0769 - calc_mre_K: 1.8804 - val_loss: 288.3407 - val_calc_mre_K: 1.7623\n", "Epoch 15/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 294.4646 - calc_mre_K: 1.7973 - val_loss: 273.2725 - val_calc_mre_K: 1.6702\n", "Epoch 16/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 286.2194 - calc_mre_K: 1.7469 - val_loss: 251.6242 - val_calc_mre_K: 1.5378\n", "Epoch 17/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 278.4631 - calc_mre_K: 1.6996 - val_loss: 255.8949 - val_calc_mre_K: 1.5640\n", "Epoch 18/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 271.9824 - calc_mre_K: 1.6600 - val_loss: 262.8639 - val_calc_mre_K: 1.6066\n", "Epoch 19/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 270.1624 - calc_mre_K: 1.6489 - val_loss: 271.6579 - val_calc_mre_K: 1.6603\n", "Epoch 20/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 264.1223 - calc_mre_K: 1.6121 - val_loss: 247.7746 - val_calc_mre_K: 1.5144\n", "Epoch 21/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 258.1547 - calc_mre_K: 1.5757 - val_loss: 280.2415 - val_calc_mre_K: 1.7127\n", "Epoch 22/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 254.6563 - calc_mre_K: 1.5543 - val_loss: 260.0399 - val_calc_mre_K: 1.5893\n", "Epoch 23/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 249.0284 - calc_mre_K: 1.5199 - val_loss: 237.0375 - val_calc_mre_K: 1.4487\n", "Epoch 24/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 253.8864 - calc_mre_K: 1.5496 - val_loss: 216.8740 - val_calc_mre_K: 1.3255\n", "Epoch 25/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 249.4401 - calc_mre_K: 1.5225 - val_loss: 249.3108 - val_calc_mre_K: 1.5237\n", "Epoch 26/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 238.7435 - calc_mre_K: 1.4572 - val_loss: 266.2927 - val_calc_mre_K: 1.6276\n", "Epoch 27/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 233.4114 - calc_mre_K: 1.4246 - val_loss: 208.9641 - val_calc_mre_K: 1.2771\n", "Epoch 28/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 230.0547 - calc_mre_K: 1.4041 - val_loss: 219.7557 - val_calc_mre_K: 1.3431\n", "Epoch 29/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 229.0038 - calc_mre_K: 1.3977 - val_loss: 214.2410 - val_calc_mre_K: 1.3094\n", "Epoch 30/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 220.4138 - calc_mre_K: 1.3453 - val_loss: 230.4205 - val_calc_mre_K: 1.4082\n", "Epoch 31/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 216.8139 - calc_mre_K: 1.3233 - val_loss: 208.5558 - val_calc_mre_K: 1.2747\n", "Epoch 32/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 217.0290 - calc_mre_K: 1.3246 - val_loss: 224.9118 - val_calc_mre_K: 1.3746\n", "Epoch 33/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 212.8776 - calc_mre_K: 1.2993 - val_loss: 199.0815 - val_calc_mre_K: 1.2167\n", "Epoch 34/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 213.6978 - calc_mre_K: 1.3043 - val_loss: 202.1036 - val_calc_mre_K: 1.2352\n", "Epoch 35/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 206.5128 - calc_mre_K: 1.2605 - val_loss: 194.1579 - val_calc_mre_K: 1.1866\n", "Epoch 36/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 201.3757 - calc_mre_K: 1.2291 - val_loss: 185.7135 - val_calc_mre_K: 1.1350\n", "Epoch 37/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 199.3132 - calc_mre_K: 1.2165 - val_loss: 217.8825 - val_calc_mre_K: 1.3316\n", "Epoch 38/2000\n", "48000/48000 [==============================] - 2s 49us/step - loss: 204.8789 - calc_mre_K: 1.2505 - val_loss: 176.7090 - val_calc_mre_K: 1.0800\n", "Epoch 39/2000\n", "48000/48000 [==============================] - 2s 49us/step - loss: 190.6736 - calc_mre_K: 1.1638 - val_loss: 184.2697 - val_calc_mre_K: 1.1262\n", "Epoch 40/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 193.0775 - calc_mre_K: 1.1785 - val_loss: 205.6794 - val_calc_mre_K: 1.2571\n", "Epoch 41/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 189.9372 - calc_mre_K: 1.1593 - val_loss: 209.7716 - val_calc_mre_K: 1.2821\n", "Epoch 42/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 188.6529 - calc_mre_K: 1.1514 - val_loss: 203.9531 - val_calc_mre_K: 1.2465\n", "Epoch 43/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 186.0370 - calc_mre_K: 1.1355 - val_loss: 178.8398 - val_calc_mre_K: 1.0930\n", "Epoch 44/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 188.7334 - calc_mre_K: 1.1519 - val_loss: 183.8436 - val_calc_mre_K: 1.1236\n", "Epoch 45/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 179.1147 - calc_mre_K: 1.0932 - val_loss: 185.7229 - val_calc_mre_K: 1.1351\n", "Epoch 46/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 181.5452 - calc_mre_K: 1.1081 - val_loss: 172.8600 - val_calc_mre_K: 1.0565\n", "Epoch 47/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 184.4521 - calc_mre_K: 1.1258 - val_loss: 175.4671 - val_calc_mre_K: 1.0724\n", "Epoch 48/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 175.1196 - calc_mre_K: 1.0688 - val_loss: 260.5045 - val_calc_mre_K: 1.5921\n", "Epoch 49/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 178.2553 - calc_mre_K: 1.0880 - val_loss: 157.0619 - val_calc_mre_K: 0.9599\n", "Epoch 50/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 175.1995 - calc_mre_K: 1.0693 - val_loss: 174.5686 - val_calc_mre_K: 1.0669\n", "Epoch 51/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 170.8937 - calc_mre_K: 1.0431 - val_loss: 203.3367 - val_calc_mre_K: 1.2428\n", "Epoch 52/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 170.9820 - calc_mre_K: 1.0436 - val_loss: 181.7487 - val_calc_mre_K: 1.1108\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "Epoch 53/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 173.1484 - calc_mre_K: 1.0568 - val_loss: 157.6922 - val_calc_mre_K: 0.9638\n", "Epoch 54/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 163.4685 - calc_mre_K: 0.9977 - val_loss: 152.8588 - val_calc_mre_K: 0.9343\n", "Epoch 55/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 171.4014 - calc_mre_K: 1.0462 - val_loss: 177.3427 - val_calc_mre_K: 1.0839\n", "Epoch 56/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 180.4526 - calc_mre_K: 1.1014 - val_loss: 150.5242 - val_calc_mre_K: 0.9200\n", "Epoch 57/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 161.5351 - calc_mre_K: 0.9859 - val_loss: 157.3193 - val_calc_mre_K: 0.9615\n", "Epoch 58/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 167.1473 - calc_mre_K: 1.0202 - val_loss: 148.9960 - val_calc_mre_K: 0.9106\n", "Epoch 59/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 159.0393 - calc_mre_K: 0.9707 - val_loss: 150.3719 - val_calc_mre_K: 0.9191\n", "Epoch 60/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 159.8650 - calc_mre_K: 0.9757 - val_loss: 165.1586 - val_calc_mre_K: 1.0095\n", "Epoch 61/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 159.1210 - calc_mre_K: 0.9712 - val_loss: 153.4727 - val_calc_mre_K: 0.9380\n", "Epoch 62/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 161.0900 - calc_mre_K: 0.9832 - val_loss: 167.4209 - val_calc_mre_K: 1.0233\n", "Epoch 63/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 155.8522 - calc_mre_K: 0.9512 - val_loss: 158.3514 - val_calc_mre_K: 0.9678\n", "Epoch 64/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 155.9146 - calc_mre_K: 0.9516 - val_loss: 146.5702 - val_calc_mre_K: 0.8958\n", "Epoch 65/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 157.0267 - calc_mre_K: 0.9584 - val_loss: 175.3835 - val_calc_mre_K: 1.0719\n", "Epoch 66/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 153.1568 - calc_mre_K: 0.9348 - val_loss: 148.1009 - val_calc_mre_K: 0.9052\n", "Epoch 67/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 157.6021 - calc_mre_K: 0.9619 - val_loss: 141.5301 - val_calc_mre_K: 0.8650\n", "Epoch 68/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 148.5306 - calc_mre_K: 0.9066 - val_loss: 142.0405 - val_calc_mre_K: 0.8681\n", "Epoch 69/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 149.4642 - calc_mre_K: 0.9123 - val_loss: 143.8907 - val_calc_mre_K: 0.8794\n", "Epoch 70/2000\n", "48000/48000 [==============================] - 2s 49us/step - loss: 153.6104 - calc_mre_K: 0.9376 - val_loss: 139.9726 - val_calc_mre_K: 0.8555\n", "Epoch 71/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 144.5853 - calc_mre_K: 0.8825 - val_loss: 155.7326 - val_calc_mre_K: 0.9518\n", "Epoch 72/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 146.7727 - calc_mre_K: 0.8958 - val_loss: 161.6986 - val_calc_mre_K: 0.9883\n", "Epoch 73/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 147.3205 - calc_mre_K: 0.8992 - val_loss: 153.2712 - val_calc_mre_K: 0.9368\n", "Epoch 74/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 145.9961 - calc_mre_K: 0.8911 - val_loss: 140.4347 - val_calc_mre_K: 0.8583\n", "Epoch 75/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 146.3559 - calc_mre_K: 0.8933 - val_loss: 140.6456 - val_calc_mre_K: 0.8596\n", "Epoch 76/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 142.3604 - calc_mre_K: 0.8689 - val_loss: 131.5680 - val_calc_mre_K: 0.8041\n", "Epoch 77/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 146.1083 - calc_mre_K: 0.8918 - val_loss: 134.6769 - val_calc_mre_K: 0.8231\n", "Epoch 78/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 139.1126 - calc_mre_K: 0.8491 - val_loss: 142.9218 - val_calc_mre_K: 0.8735\n", "Epoch 79/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 141.7618 - calc_mre_K: 0.8652 - val_loss: 133.6090 - val_calc_mre_K: 0.8166\n", "Epoch 80/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 140.3587 - calc_mre_K: 0.8567 - val_loss: 129.7303 - val_calc_mre_K: 0.7929\n", "Epoch 81/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 137.2947 - calc_mre_K: 0.8380 - val_loss: 133.8552 - val_calc_mre_K: 0.8181\n", "Epoch 82/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 137.8744 - calc_mre_K: 0.8415 - val_loss: 130.9207 - val_calc_mre_K: 0.8002\n", "Epoch 83/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 140.9715 - calc_mre_K: 0.8604 - val_loss: 141.9069 - val_calc_mre_K: 0.8673\n", "Epoch 84/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 140.7600 - calc_mre_K: 0.8591 - val_loss: 134.0286 - val_calc_mre_K: 0.8191\n", "Epoch 85/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 133.9786 - calc_mre_K: 0.8177 - val_loss: 151.1420 - val_calc_mre_K: 0.9237\n", "Epoch 86/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 136.8074 - calc_mre_K: 0.8350 - val_loss: 124.2670 - val_calc_mre_K: 0.7595\n", "Epoch 87/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 134.7955 - calc_mre_K: 0.8227 - val_loss: 135.7303 - val_calc_mre_K: 0.8296\n", "Epoch 88/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 136.5116 - calc_mre_K: 0.8332 - val_loss: 134.9733 - val_calc_mre_K: 0.8249\n", "Epoch 89/2000\n", "48000/48000 [==============================] - 3s 55us/step - loss: 134.3332 - calc_mre_K: 0.8199 - val_loss: 124.1983 - val_calc_mre_K: 0.7591\n", "Epoch 90/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 133.1409 - calc_mre_K: 0.8126 - val_loss: 122.0805 - val_calc_mre_K: 0.7461\n", "Epoch 91/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 132.1829 - calc_mre_K: 0.8068 - val_loss: 148.8014 - val_calc_mre_K: 0.9094\n", "Epoch 92/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 132.4076 - calc_mre_K: 0.8082 - val_loss: 130.7575 - val_calc_mre_K: 0.7992\n", "Epoch 93/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 133.0999 - calc_mre_K: 0.8124 - val_loss: 122.6661 - val_calc_mre_K: 0.7497\n", "Epoch 94/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 133.6497 - calc_mre_K: 0.8157 - val_loss: 136.5403 - val_calc_mre_K: 0.8345\n", "Epoch 95/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 134.1756 - calc_mre_K: 0.8189 - val_loss: 125.8499 - val_calc_mre_K: 0.7692\n", "Epoch 96/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 128.1829 - calc_mre_K: 0.7824 - val_loss: 120.9619 - val_calc_mre_K: 0.7393\n", "Epoch 97/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 131.4088 - calc_mre_K: 0.8021 - val_loss: 150.9638 - val_calc_mre_K: 0.9227\n", "Epoch 98/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 129.2636 - calc_mre_K: 0.7890 - val_loss: 126.4415 - val_calc_mre_K: 0.7728\n", "Epoch 99/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 131.9970 - calc_mre_K: 0.8056 - val_loss: 128.3836 - val_calc_mre_K: 0.7846\n", "Epoch 100/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 127.8191 - calc_mre_K: 0.7801 - val_loss: 119.3535 - val_calc_mre_K: 0.7295\n", "Epoch 101/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 135.2708 - calc_mre_K: 0.8256 - val_loss: 127.8778 - val_calc_mre_K: 0.7816\n", "Epoch 102/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 127.4603 - calc_mre_K: 0.7780 - val_loss: 121.4175 - val_calc_mre_K: 0.7421\n", "Epoch 103/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 124.8511 - calc_mre_K: 0.7620 - val_loss: 118.8897 - val_calc_mre_K: 0.7267\n", "Epoch 104/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 132.6419 - calc_mre_K: 0.8096 - val_loss: 145.5274 - val_calc_mre_K: 0.8894\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "Epoch 105/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 124.2600 - calc_mre_K: 0.7584 - val_loss: 129.8107 - val_calc_mre_K: 0.7934\n", "Epoch 106/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 125.3140 - calc_mre_K: 0.7649 - val_loss: 122.2460 - val_calc_mre_K: 0.7472\n", "Epoch 107/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 126.5669 - calc_mre_K: 0.7725 - val_loss: 132.0504 - val_calc_mre_K: 0.8071\n", "Epoch 108/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 125.6400 - calc_mre_K: 0.7668 - val_loss: 130.3404 - val_calc_mre_K: 0.7967\n", "Epoch 109/2000\n", "48000/48000 [==============================] - 2s 48us/step - loss: 126.6342 - calc_mre_K: 0.7729 - val_loss: 131.8918 - val_calc_mre_K: 0.8061\n", "Epoch 110/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 122.4863 - calc_mre_K: 0.7476 - val_loss: 120.7282 - val_calc_mre_K: 0.7379\n", "Epoch 111/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 122.8275 - calc_mre_K: 0.7497 - val_loss: 121.5979 - val_calc_mre_K: 0.7432\n", "Epoch 112/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 122.9411 - calc_mre_K: 0.7504 - val_loss: 114.0247 - val_calc_mre_K: 0.6969\n", "Epoch 113/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 122.4875 - calc_mre_K: 0.7476 - val_loss: 142.0574 - val_calc_mre_K: 0.8682\n", "Epoch 114/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 123.1204 - calc_mre_K: 0.7515 - val_loss: 112.6512 - val_calc_mre_K: 0.6885\n", "Epoch 115/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 127.4112 - calc_mre_K: 0.7777 - val_loss: 114.8082 - val_calc_mre_K: 0.7017\n", "Epoch 116/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 120.3769 - calc_mre_K: 0.7347 - val_loss: 110.6840 - val_calc_mre_K: 0.6765\n", "Epoch 117/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 121.2108 - calc_mre_K: 0.7398 - val_loss: 130.7038 - val_calc_mre_K: 0.7988\n", "Epoch 118/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 120.0784 - calc_mre_K: 0.7329 - val_loss: 109.2587 - val_calc_mre_K: 0.6678\n", "Epoch 119/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 121.6401 - calc_mre_K: 0.7424 - val_loss: 119.3679 - val_calc_mre_K: 0.7296\n", "Epoch 120/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 120.3946 - calc_mre_K: 0.7348 - val_loss: 120.5291 - val_calc_mre_K: 0.7367\n", "Epoch 121/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 124.7291 - calc_mre_K: 0.7613 - val_loss: 120.4486 - val_calc_mre_K: 0.7362\n", "Epoch 122/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 118.3275 - calc_mre_K: 0.7222 - val_loss: 121.6486 - val_calc_mre_K: 0.7435\n", "Epoch 123/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 118.0430 - calc_mre_K: 0.7205 - val_loss: 119.0856 - val_calc_mre_K: 0.7278\n", "Epoch 124/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 124.0017 - calc_mre_K: 0.7568 - val_loss: 131.8052 - val_calc_mre_K: 0.8056\n", "Epoch 125/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 120.0169 - calc_mre_K: 0.7325 - val_loss: 111.0155 - val_calc_mre_K: 0.6785\n", "Epoch 126/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 117.1291 - calc_mre_K: 0.7149 - val_loss: 108.4612 - val_calc_mre_K: 0.6629\n", "Epoch 127/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 118.8360 - calc_mre_K: 0.7253 - val_loss: 112.0743 - val_calc_mre_K: 0.6850\n", "Epoch 128/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 119.1295 - calc_mre_K: 0.7271 - val_loss: 121.3199 - val_calc_mre_K: 0.7415\n", "Epoch 129/2000\n", "48000/48000 [==============================] - 2s 49us/step - loss: 119.2631 - calc_mre_K: 0.7279 - val_loss: 118.9910 - val_calc_mre_K: 0.7273\n", "Epoch 130/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 115.8822 - calc_mre_K: 0.7073 - val_loss: 112.1258 - val_calc_mre_K: 0.6853\n", "Epoch 131/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 119.0329 - calc_mre_K: 0.7265 - val_loss: 108.2800 - val_calc_mre_K: 0.6618\n", "Epoch 132/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 119.1341 - calc_mre_K: 0.7271 - val_loss: 119.5808 - val_calc_mre_K: 0.7309\n", "Epoch 133/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 116.2830 - calc_mre_K: 0.7097 - val_loss: 124.4681 - val_calc_mre_K: 0.7607\n", "Epoch 134/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 113.7827 - calc_mre_K: 0.6945 - val_loss: 127.7704 - val_calc_mre_K: 0.7809\n", "Epoch 135/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 114.8143 - calc_mre_K: 0.7008 - val_loss: 134.7457 - val_calc_mre_K: 0.8235\n", "Epoch 136/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 116.3069 - calc_mre_K: 0.7099 - val_loss: 119.7766 - val_calc_mre_K: 0.7321\n", "Epoch 137/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 117.6356 - calc_mre_K: 0.7180 - val_loss: 106.6275 - val_calc_mre_K: 0.6517\n", "Epoch 138/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 113.3611 - calc_mre_K: 0.6919 - val_loss: 107.9170 - val_calc_mre_K: 0.6596\n", "Epoch 139/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 112.4439 - calc_mre_K: 0.6863 - val_loss: 109.9092 - val_calc_mre_K: 0.6718\n", "Epoch 140/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 114.3005 - calc_mre_K: 0.6976 - val_loss: 110.1872 - val_calc_mre_K: 0.6734\n", "Epoch 141/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 113.2376 - calc_mre_K: 0.6911 - val_loss: 116.3723 - val_calc_mre_K: 0.7113\n", "Epoch 142/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 118.4617 - calc_mre_K: 0.7230 - val_loss: 106.1509 - val_calc_mre_K: 0.6488\n", "Epoch 143/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 114.0538 - calc_mre_K: 0.6961 - val_loss: 116.3342 - val_calc_mre_K: 0.7110\n", "Epoch 144/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 113.7670 - calc_mre_K: 0.6944 - val_loss: 123.4880 - val_calc_mre_K: 0.7547\n", "Epoch 145/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 111.7316 - calc_mre_K: 0.6820 - val_loss: 106.0218 - val_calc_mre_K: 0.6480\n", "Epoch 146/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 112.8772 - calc_mre_K: 0.6889 - val_loss: 117.8516 - val_calc_mre_K: 0.7203\n", "Epoch 147/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 111.4188 - calc_mre_K: 0.6800 - val_loss: 107.6938 - val_calc_mre_K: 0.6582\n", "Epoch 148/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 111.8744 - calc_mre_K: 0.6828 - val_loss: 109.0731 - val_calc_mre_K: 0.6666\n", "Epoch 149/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 110.2834 - calc_mre_K: 0.6731 - val_loss: 107.0590 - val_calc_mre_K: 0.6543\n", "Epoch 150/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 110.1874 - calc_mre_K: 0.6725 - val_loss: 108.9695 - val_calc_mre_K: 0.6660\n", "Epoch 151/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 111.9836 - calc_mre_K: 0.6835 - val_loss: 110.3534 - val_calc_mre_K: 0.6745\n", "Epoch 152/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 109.7145 - calc_mre_K: 0.6696 - val_loss: 110.8572 - val_calc_mre_K: 0.6775\n", "Epoch 153/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 110.2422 - calc_mre_K: 0.6729 - val_loss: 105.4677 - val_calc_mre_K: 0.6446\n", "Epoch 154/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 110.8918 - calc_mre_K: 0.6768 - val_loss: 107.8672 - val_calc_mre_K: 0.6593\n", "Epoch 155/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 114.9676 - calc_mre_K: 0.7017 - val_loss: 104.9910 - val_calc_mre_K: 0.6417\n", "Epoch 156/2000\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "48000/48000 [==============================] - 3s 53us/step - loss: 108.5068 - calc_mre_K: 0.6623 - val_loss: 112.1881 - val_calc_mre_K: 0.6857\n", "Epoch 157/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 109.4276 - calc_mre_K: 0.6679 - val_loss: 108.4361 - val_calc_mre_K: 0.6627\n", "Epoch 158/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 110.0511 - calc_mre_K: 0.6717 - val_loss: 108.6786 - val_calc_mre_K: 0.6642\n", "Epoch 159/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 109.6806 - calc_mre_K: 0.6694 - val_loss: 126.1993 - val_calc_mre_K: 0.7713\n", "Epoch 160/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 108.2778 - calc_mre_K: 0.6609 - val_loss: 107.8562 - val_calc_mre_K: 0.6592\n", "Epoch 161/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 111.2709 - calc_mre_K: 0.6791 - val_loss: 104.5160 - val_calc_mre_K: 0.6388\n", "Epoch 162/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 110.2340 - calc_mre_K: 0.6728 - val_loss: 107.8390 - val_calc_mre_K: 0.6591\n", "Epoch 163/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 106.9347 - calc_mre_K: 0.6527 - val_loss: 107.4190 - val_calc_mre_K: 0.6565\n", "Epoch 164/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 109.4915 - calc_mre_K: 0.6683 - val_loss: 107.4226 - val_calc_mre_K: 0.6566\n", "Epoch 165/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 107.3223 - calc_mre_K: 0.6550 - val_loss: 107.0650 - val_calc_mre_K: 0.6544\n", "Epoch 166/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 112.1783 - calc_mre_K: 0.6847 - val_loss: 106.1413 - val_calc_mre_K: 0.6487\n", "Epoch 167/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 106.3065 - calc_mre_K: 0.6488 - val_loss: 102.2936 - val_calc_mre_K: 0.6252\n", "Epoch 168/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 107.4463 - calc_mre_K: 0.6558 - val_loss: 106.5010 - val_calc_mre_K: 0.6509\n", "Epoch 169/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 108.4077 - calc_mre_K: 0.6617 - val_loss: 124.3627 - val_calc_mre_K: 0.7601\n", "Epoch 170/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 106.6253 - calc_mre_K: 0.6508 - val_loss: 98.6109 - val_calc_mre_K: 0.6027\n", "Epoch 171/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 109.1354 - calc_mre_K: 0.6661 - val_loss: 109.4292 - val_calc_mre_K: 0.6688\n", "Epoch 172/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 106.5438 - calc_mre_K: 0.6503 - val_loss: 114.0416 - val_calc_mre_K: 0.6970\n", "Epoch 173/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 106.3723 - calc_mre_K: 0.6492 - val_loss: 108.4122 - val_calc_mre_K: 0.6626\n", "Epoch 174/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 106.6236 - calc_mre_K: 0.6508 - val_loss: 113.9005 - val_calc_mre_K: 0.6961\n", "Epoch 175/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 105.8314 - calc_mre_K: 0.6459 - val_loss: 109.5198 - val_calc_mre_K: 0.6694\n", "Epoch 176/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 106.2126 - calc_mre_K: 0.6483 - val_loss: 109.9304 - val_calc_mre_K: 0.6719\n", "Epoch 177/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 105.5392 - calc_mre_K: 0.6442 - val_loss: 104.6905 - val_calc_mre_K: 0.6398\n", "Epoch 178/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 104.9443 - calc_mre_K: 0.6405 - val_loss: 107.6219 - val_calc_mre_K: 0.6578\n", "Epoch 179/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 105.6134 - calc_mre_K: 0.6446 - val_loss: 122.3203 - val_calc_mre_K: 0.7476\n", "Epoch 180/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 104.4942 - calc_mre_K: 0.6378 - val_loss: 104.7684 - val_calc_mre_K: 0.6403\n", "Epoch 181/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 106.2836 - calc_mre_K: 0.6487 - val_loss: 104.4754 - val_calc_mre_K: 0.6386\n", "Epoch 182/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 105.7145 - calc_mre_K: 0.6452 - val_loss: 122.5208 - val_calc_mre_K: 0.7488\n", "Epoch 183/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 106.0086 - calc_mre_K: 0.6470 - val_loss: 114.4517 - val_calc_mre_K: 0.6995\n", "Epoch 184/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 104.0758 - calc_mre_K: 0.6352 - val_loss: 100.9431 - val_calc_mre_K: 0.6169\n", "Epoch 185/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 105.6322 - calc_mre_K: 0.6447 - val_loss: 102.2812 - val_calc_mre_K: 0.6251\n", "Epoch 186/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 104.7726 - calc_mre_K: 0.6395 - val_loss: 104.8845 - val_calc_mre_K: 0.6410\n", "Epoch 187/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 105.3529 - calc_mre_K: 0.6430 - val_loss: 100.5026 - val_calc_mre_K: 0.6142\n", "Epoch 188/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 102.7930 - calc_mre_K: 0.6274 - val_loss: 112.3629 - val_calc_mre_K: 0.6867\n", "Epoch 189/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 103.6800 - calc_mre_K: 0.6328 - val_loss: 97.7943 - val_calc_mre_K: 0.5977\n", "Epoch 190/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 105.8988 - calc_mre_K: 0.6464 - val_loss: 100.6578 - val_calc_mre_K: 0.6152\n", "Epoch 191/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 103.1117 - calc_mre_K: 0.6293 - val_loss: 105.1782 - val_calc_mre_K: 0.6428\n", "Epoch 192/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 103.8336 - calc_mre_K: 0.6338 - val_loss: 101.5947 - val_calc_mre_K: 0.6209\n", "Epoch 193/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 104.3174 - calc_mre_K: 0.6367 - val_loss: 101.4290 - val_calc_mre_K: 0.6199\n", "Epoch 194/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 103.0445 - calc_mre_K: 0.6289 - val_loss: 100.2179 - val_calc_mre_K: 0.6125\n", "Epoch 195/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 103.9741 - calc_mre_K: 0.6346 - val_loss: 104.0520 - val_calc_mre_K: 0.6360\n", "Epoch 196/2000\n", "48000/48000 [==============================] - 2s 48us/step - loss: 101.8330 - calc_mre_K: 0.6215 - val_loss: 104.9072 - val_calc_mre_K: 0.6412\n", "Epoch 197/2000\n", "48000/48000 [==============================] - 3s 56us/step - loss: 102.3311 - calc_mre_K: 0.6246 - val_loss: 105.4793 - val_calc_mre_K: 0.6447\n", "Epoch 198/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 104.4874 - calc_mre_K: 0.6377 - val_loss: 98.8774 - val_calc_mre_K: 0.6043\n", "Epoch 199/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 100.9658 - calc_mre_K: 0.6162 - val_loss: 102.1155 - val_calc_mre_K: 0.6241\n", "Epoch 200/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 105.3388 - calc_mre_K: 0.6429 - val_loss: 97.5254 - val_calc_mre_K: 0.5961\n", "Epoch 201/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 101.6143 - calc_mre_K: 0.6202 - val_loss: 118.7985 - val_calc_mre_K: 0.7261\n", "Epoch 202/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 101.5383 - calc_mre_K: 0.6197 - val_loss: 98.2693 - val_calc_mre_K: 0.6006\n", "Epoch 203/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 103.1716 - calc_mre_K: 0.6297 - val_loss: 99.8743 - val_calc_mre_K: 0.6104\n", "Epoch 204/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 103.0693 - calc_mre_K: 0.6291 - val_loss: 103.7322 - val_calc_mre_K: 0.6340\n", "Epoch 205/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 102.3542 - calc_mre_K: 0.6247 - val_loss: 103.5409 - val_calc_mre_K: 0.6328\n", "Epoch 206/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 100.7468 - calc_mre_K: 0.6149 - val_loss: 95.1989 - val_calc_mre_K: 0.5818\n", "Epoch 207/2000\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "48000/48000 [==============================] - 2s 51us/step - loss: 103.7325 - calc_mre_K: 0.6331 - val_loss: 96.6600 - val_calc_mre_K: 0.5908\n", "Epoch 208/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 106.8788 - calc_mre_K: 0.6523 - val_loss: 121.7051 - val_calc_mre_K: 0.7439\n", "Epoch 209/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 103.1602 - calc_mre_K: 0.6296 - val_loss: 107.8801 - val_calc_mre_K: 0.6593\n", "Epoch 210/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 100.1636 - calc_mre_K: 0.6114 - val_loss: 97.2253 - val_calc_mre_K: 0.5942\n", "Epoch 211/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 100.6832 - calc_mre_K: 0.6145 - val_loss: 109.2051 - val_calc_mre_K: 0.6674\n", "Epoch 212/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 102.0518 - calc_mre_K: 0.6229 - val_loss: 97.3057 - val_calc_mre_K: 0.5947\n", "Epoch 213/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 100.5632 - calc_mre_K: 0.6138 - val_loss: 97.9803 - val_calc_mre_K: 0.5989\n", "Epoch 214/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 100.7240 - calc_mre_K: 0.6148 - val_loss: 100.0181 - val_calc_mre_K: 0.6113\n", "Epoch 215/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 99.8555 - calc_mre_K: 0.6095 - val_loss: 102.1535 - val_calc_mre_K: 0.6243\n", "Epoch 216/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 100.1928 - calc_mre_K: 0.6115 - val_loss: 97.4651 - val_calc_mre_K: 0.5957\n", "Epoch 217/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 99.7825 - calc_mre_K: 0.6090 - val_loss: 100.3381 - val_calc_mre_K: 0.6132\n", "Epoch 218/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 100.9318 - calc_mre_K: 0.6160 - val_loss: 95.7448 - val_calc_mre_K: 0.5852\n", "Epoch 219/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 102.1479 - calc_mre_K: 0.6235 - val_loss: 99.6073 - val_calc_mre_K: 0.6088\n", "Epoch 220/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 99.5629 - calc_mre_K: 0.6077 - val_loss: 95.5550 - val_calc_mre_K: 0.5840\n", "Epoch 221/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 99.3377 - calc_mre_K: 0.6063 - val_loss: 94.9927 - val_calc_mre_K: 0.5806\n", "Epoch 222/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 101.3473 - calc_mre_K: 0.6186 - val_loss: 99.4693 - val_calc_mre_K: 0.6079\n", "Epoch 223/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 100.9005 - calc_mre_K: 0.6158 - val_loss: 99.8020 - val_calc_mre_K: 0.6100\n", "Epoch 224/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 99.5964 - calc_mre_K: 0.6079 - val_loss: 107.1228 - val_calc_mre_K: 0.6547\n", "Epoch 225/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 99.6001 - calc_mre_K: 0.6079 - val_loss: 93.1532 - val_calc_mre_K: 0.5693\n", "Epoch 226/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 99.7653 - calc_mre_K: 0.6089 - val_loss: 100.9328 - val_calc_mre_K: 0.6169\n", "Epoch 227/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 99.0875 - calc_mre_K: 0.6048 - val_loss: 101.7336 - val_calc_mre_K: 0.6218\n", "Epoch 228/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 99.6773 - calc_mre_K: 0.6084 - val_loss: 108.7712 - val_calc_mre_K: 0.6648\n", "Epoch 229/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 98.6927 - calc_mre_K: 0.6024 - val_loss: 97.7183 - val_calc_mre_K: 0.5972\n", "Epoch 230/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 99.7003 - calc_mre_K: 0.6085 - val_loss: 95.0020 - val_calc_mre_K: 0.5806\n", "Epoch 231/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 98.2644 - calc_mre_K: 0.5998 - val_loss: 95.0884 - val_calc_mre_K: 0.5812\n", "Epoch 232/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 98.8962 - calc_mre_K: 0.6036 - val_loss: 96.1243 - val_calc_mre_K: 0.5875\n", "Epoch 233/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 98.4440 - calc_mre_K: 0.6009 - val_loss: 97.7466 - val_calc_mre_K: 0.5974\n", "Epoch 234/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 99.8212 - calc_mre_K: 0.6093 - val_loss: 98.3566 - val_calc_mre_K: 0.6011\n", "Epoch 235/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 98.9774 - calc_mre_K: 0.6041 - val_loss: 92.3633 - val_calc_mre_K: 0.5645\n", "Epoch 236/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 99.3206 - calc_mre_K: 0.6062 - val_loss: 118.4091 - val_calc_mre_K: 0.7237\n", "Epoch 237/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 97.2623 - calc_mre_K: 0.5936 - val_loss: 96.0144 - val_calc_mre_K: 0.5868\n", "Epoch 238/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 98.7952 - calc_mre_K: 0.6030 - val_loss: 110.6610 - val_calc_mre_K: 0.6763\n", "Epoch 239/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 98.7706 - calc_mre_K: 0.6028 - val_loss: 109.1040 - val_calc_mre_K: 0.6668\n", "Epoch 240/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 98.3577 - calc_mre_K: 0.6003 - val_loss: 105.5355 - val_calc_mre_K: 0.6450\n", "Epoch 241/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 100.7646 - calc_mre_K: 0.6150 - val_loss: 94.6201 - val_calc_mre_K: 0.5783\n", "Epoch 242/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 97.2334 - calc_mre_K: 0.5935 - val_loss: 93.3718 - val_calc_mre_K: 0.5707\n", "Epoch 243/2000\n", "48000/48000 [==============================] - 2s 49us/step - loss: 97.4070 - calc_mre_K: 0.5945 - val_loss: 93.2243 - val_calc_mre_K: 0.5698\n", "Epoch 244/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 97.4968 - calc_mre_K: 0.5951 - val_loss: 97.4804 - val_calc_mre_K: 0.5958\n", "Epoch 245/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 97.0737 - calc_mre_K: 0.5925 - val_loss: 94.7559 - val_calc_mre_K: 0.5791\n", "Epoch 246/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 96.9094 - calc_mre_K: 0.5915 - val_loss: 100.6832 - val_calc_mre_K: 0.6154\n", "Epoch 247/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 98.0505 - calc_mre_K: 0.5985 - val_loss: 93.6904 - val_calc_mre_K: 0.5726\n", "Epoch 248/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 99.6141 - calc_mre_K: 0.6080 - val_loss: 92.8807 - val_calc_mre_K: 0.5677\n", "Epoch 249/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 97.5215 - calc_mre_K: 0.5952 - val_loss: 94.4616 - val_calc_mre_K: 0.5773\n", "Epoch 250/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 97.0416 - calc_mre_K: 0.5923 - val_loss: 97.2222 - val_calc_mre_K: 0.5942\n", "Epoch 251/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 97.3393 - calc_mre_K: 0.5941 - val_loss: 100.9278 - val_calc_mre_K: 0.6169\n", "Epoch 252/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 97.0328 - calc_mre_K: 0.5922 - val_loss: 93.1528 - val_calc_mre_K: 0.5693\n", "Epoch 253/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 96.6583 - calc_mre_K: 0.5900 - val_loss: 101.3051 - val_calc_mre_K: 0.6192\n", "Epoch 254/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 97.1259 - calc_mre_K: 0.5928 - val_loss: 97.0326 - val_calc_mre_K: 0.5930\n", "Epoch 255/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 95.3763 - calc_mre_K: 0.5821 - val_loss: 92.8040 - val_calc_mre_K: 0.5672\n", "Epoch 256/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 97.8630 - calc_mre_K: 0.5973 - val_loss: 101.3886 - val_calc_mre_K: 0.6197\n", "Epoch 257/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 95.7006 - calc_mre_K: 0.5841 - val_loss: 101.5791 - val_calc_mre_K: 0.6208\n", "Epoch 258/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 96.6158 - calc_mre_K: 0.5897 - val_loss: 91.6830 - val_calc_mre_K: 0.5603\n", "Epoch 259/2000\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "48000/48000 [==============================] - 2s 49us/step - loss: 96.2652 - calc_mre_K: 0.5876 - val_loss: 99.1111 - val_calc_mre_K: 0.6058\n", "Epoch 260/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 96.5169 - calc_mre_K: 0.5891 - val_loss: 90.8914 - val_calc_mre_K: 0.5555\n", "Epoch 261/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 96.0186 - calc_mre_K: 0.5861 - val_loss: 104.5424 - val_calc_mre_K: 0.6389\n", "Epoch 262/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 99.4370 - calc_mre_K: 0.6069 - val_loss: 92.0787 - val_calc_mre_K: 0.5628\n", "Epoch 263/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 95.1902 - calc_mre_K: 0.5810 - val_loss: 105.8537 - val_calc_mre_K: 0.6470\n", "Epoch 264/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 96.1991 - calc_mre_K: 0.5872 - val_loss: 90.6992 - val_calc_mre_K: 0.5543\n", "Epoch 265/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 95.0909 - calc_mre_K: 0.5804 - val_loss: 103.2948 - val_calc_mre_K: 0.6314\n", "Epoch 266/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 95.6114 - calc_mre_K: 0.5836 - val_loss: 94.6441 - val_calc_mre_K: 0.5785\n", "Epoch 267/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 95.4102 - calc_mre_K: 0.5823 - val_loss: 93.0793 - val_calc_mre_K: 0.5689\n", "Epoch 268/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 95.5656 - calc_mre_K: 0.5833 - val_loss: 94.1860 - val_calc_mre_K: 0.5756\n", "Epoch 269/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 95.9575 - calc_mre_K: 0.5857 - val_loss: 96.2557 - val_calc_mre_K: 0.5883\n", "Epoch 270/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 96.2432 - calc_mre_K: 0.5874 - val_loss: 97.2940 - val_calc_mre_K: 0.5946\n", "Epoch 271/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 95.6542 - calc_mre_K: 0.5838 - val_loss: 96.7887 - val_calc_mre_K: 0.5916\n", "Epoch 272/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 95.6902 - calc_mre_K: 0.5840 - val_loss: 93.7836 - val_calc_mre_K: 0.5732\n", "Epoch 273/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 95.9004 - calc_mre_K: 0.5853 - val_loss: 99.3904 - val_calc_mre_K: 0.6075\n", "Epoch 274/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 97.2285 - calc_mre_K: 0.5934 - val_loss: 94.4813 - val_calc_mre_K: 0.5775\n", "Epoch 275/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 93.6489 - calc_mre_K: 0.5716 - val_loss: 90.5012 - val_calc_mre_K: 0.5531\n", "Epoch 276/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 95.0668 - calc_mre_K: 0.5802 - val_loss: 96.9105 - val_calc_mre_K: 0.5923\n", "Epoch 277/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 95.5111 - calc_mre_K: 0.5830 - val_loss: 99.0192 - val_calc_mre_K: 0.6052\n", "Epoch 278/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 95.4069 - calc_mre_K: 0.5823 - val_loss: 95.2975 - val_calc_mre_K: 0.5824\n", "Epoch 279/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 95.5681 - calc_mre_K: 0.5833 - val_loss: 97.9956 - val_calc_mre_K: 0.5989\n", "Epoch 280/2000\n", "48000/48000 [==============================] - 2s 48us/step - loss: 94.5477 - calc_mre_K: 0.5771 - val_loss: 90.7232 - val_calc_mre_K: 0.5545\n", "Epoch 281/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 93.8116 - calc_mre_K: 0.5726 - val_loss: 90.1256 - val_calc_mre_K: 0.5508\n", "Epoch 282/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 94.2903 - calc_mre_K: 0.5755 - val_loss: 97.5963 - val_calc_mre_K: 0.5965\n", "Epoch 283/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 94.3194 - calc_mre_K: 0.5757 - val_loss: 94.6749 - val_calc_mre_K: 0.5786\n", "Epoch 284/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 95.2479 - calc_mre_K: 0.5813 - val_loss: 108.1763 - val_calc_mre_K: 0.6612\n", "Epoch 285/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 94.2920 - calc_mre_K: 0.5755 - val_loss: 101.6072 - val_calc_mre_K: 0.6210\n", "Epoch 286/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 94.1603 - calc_mre_K: 0.5747 - val_loss: 107.2049 - val_calc_mre_K: 0.6552\n", "Epoch 287/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 95.7233 - calc_mre_K: 0.5842 - val_loss: 90.6330 - val_calc_mre_K: 0.5539\n", "Epoch 288/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 93.9447 - calc_mre_K: 0.5734 - val_loss: 97.3053 - val_calc_mre_K: 0.5947\n", "Epoch 289/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 93.3811 - calc_mre_K: 0.5700 - val_loss: 94.6099 - val_calc_mre_K: 0.5782\n", "Epoch 290/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 94.1816 - calc_mre_K: 0.5748 - val_loss: 108.7794 - val_calc_mre_K: 0.6649\n", "Epoch 291/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 93.6786 - calc_mre_K: 0.5718 - val_loss: 92.0577 - val_calc_mre_K: 0.5626\n", "Epoch 292/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 93.9854 - calc_mre_K: 0.5736 - val_loss: 100.8902 - val_calc_mre_K: 0.6166\n", "Epoch 293/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 93.7021 - calc_mre_K: 0.5719 - val_loss: 88.5346 - val_calc_mre_K: 0.5411\n", "Epoch 294/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 92.6724 - calc_mre_K: 0.5656 - val_loss: 109.0596 - val_calc_mre_K: 0.6666\n", "Epoch 295/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 93.7623 - calc_mre_K: 0.5723 - val_loss: 91.2466 - val_calc_mre_K: 0.5577\n", "Epoch 296/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 93.9072 - calc_mre_K: 0.5732 - val_loss: 89.5144 - val_calc_mre_K: 0.5471\n", "Epoch 297/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 93.5050 - calc_mre_K: 0.5707 - val_loss: 102.5011 - val_calc_mre_K: 0.6265\n", "Epoch 298/2000\n", "48000/48000 [==============================] - 3s 55us/step - loss: 93.7864 - calc_mre_K: 0.5724 - val_loss: 90.2738 - val_calc_mre_K: 0.5517\n", "Epoch 299/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 93.0721 - calc_mre_K: 0.5681 - val_loss: 89.3962 - val_calc_mre_K: 0.5464\n", "Epoch 300/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 92.3281 - calc_mre_K: 0.5635 - val_loss: 91.8236 - val_calc_mre_K: 0.5612\n", "Epoch 301/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 92.6404 - calc_mre_K: 0.5654 - val_loss: 91.7917 - val_calc_mre_K: 0.5610\n", "Epoch 302/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 93.4328 - calc_mre_K: 0.5703 - val_loss: 94.9685 - val_calc_mre_K: 0.5804\n", "Epoch 303/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 93.3652 - calc_mre_K: 0.5699 - val_loss: 90.9325 - val_calc_mre_K: 0.5557\n", "Epoch 304/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 94.2544 - calc_mre_K: 0.5753 - val_loss: 93.2130 - val_calc_mre_K: 0.5697\n", "Epoch 305/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 92.3390 - calc_mre_K: 0.5636 - val_loss: 91.8178 - val_calc_mre_K: 0.5612\n", "Epoch 306/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 91.9068 - calc_mre_K: 0.5610 - val_loss: 86.5511 - val_calc_mre_K: 0.5290\n", "Epoch 307/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 94.5044 - calc_mre_K: 0.5768 - val_loss: 94.2398 - val_calc_mre_K: 0.5760\n", "Epoch 308/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 91.3235 - calc_mre_K: 0.5574 - val_loss: 92.3836 - val_calc_mre_K: 0.5646\n", "Epoch 309/2000\n", "48000/48000 [==============================] - 2s 48us/step - loss: 91.9471 - calc_mre_K: 0.5612 - val_loss: 91.6350 - val_calc_mre_K: 0.5600\n", "Epoch 310/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 93.1337 - calc_mre_K: 0.5684 - val_loss: 87.2588 - val_calc_mre_K: 0.5333\n", "Epoch 311/2000\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "48000/48000 [==============================] - 2s 51us/step - loss: 92.7290 - calc_mre_K: 0.5660 - val_loss: 96.9130 - val_calc_mre_K: 0.5923\n", "Epoch 312/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 91.2675 - calc_mre_K: 0.5571 - val_loss: 88.3411 - val_calc_mre_K: 0.5399\n", "Epoch 313/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 94.3594 - calc_mre_K: 0.5759 - val_loss: 86.7671 - val_calc_mre_K: 0.5303\n", "Epoch 314/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 90.7405 - calc_mre_K: 0.5538 - val_loss: 90.2045 - val_calc_mre_K: 0.5513\n", "Epoch 315/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 91.6234 - calc_mre_K: 0.5592 - val_loss: 104.7195 - val_calc_mre_K: 0.6400\n", "Epoch 316/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 91.6651 - calc_mre_K: 0.5595 - val_loss: 90.1575 - val_calc_mre_K: 0.5510\n", "Epoch 317/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 91.2141 - calc_mre_K: 0.5567 - val_loss: 90.3438 - val_calc_mre_K: 0.5522\n", "Epoch 318/2000\n", "48000/48000 [==============================] - 2s 49us/step - loss: 92.6005 - calc_mre_K: 0.5652 - val_loss: 85.5898 - val_calc_mre_K: 0.5231\n", "Epoch 319/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 90.9654 - calc_mre_K: 0.5552 - val_loss: 86.8029 - val_calc_mre_K: 0.5305\n", "Epoch 320/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 91.3799 - calc_mre_K: 0.5577 - val_loss: 87.8824 - val_calc_mre_K: 0.5371\n", "Epoch 321/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 90.8468 - calc_mre_K: 0.5545 - val_loss: 97.3825 - val_calc_mre_K: 0.5952\n", "Epoch 322/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 91.5128 - calc_mre_K: 0.5586 - val_loss: 89.9991 - val_calc_mre_K: 0.5501\n", "Epoch 323/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 91.4933 - calc_mre_K: 0.5584 - val_loss: 89.7807 - val_calc_mre_K: 0.5487\n", "Epoch 324/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 90.8670 - calc_mre_K: 0.5546 - val_loss: 106.4507 - val_calc_mre_K: 0.6506\n", "Epoch 325/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 91.2607 - calc_mre_K: 0.5570 - val_loss: 92.7949 - val_calc_mre_K: 0.5672\n", "Epoch 326/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 90.6262 - calc_mre_K: 0.5531 - val_loss: 97.2259 - val_calc_mre_K: 0.5943\n", "Epoch 327/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 91.0906 - calc_mre_K: 0.5560 - val_loss: 97.3802 - val_calc_mre_K: 0.5952\n", "Epoch 328/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 90.7029 - calc_mre_K: 0.5536 - val_loss: 92.6896 - val_calc_mre_K: 0.5665\n", "Epoch 329/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 91.3050 - calc_mre_K: 0.5573 - val_loss: 103.8897 - val_calc_mre_K: 0.6350\n", "Epoch 330/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 90.7138 - calc_mre_K: 0.5537 - val_loss: 88.4553 - val_calc_mre_K: 0.5406\n", "Epoch 331/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 90.9270 - calc_mre_K: 0.5550 - val_loss: 107.0133 - val_calc_mre_K: 0.6541\n", "Epoch 332/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 89.8437 - calc_mre_K: 0.5484 - val_loss: 93.0000 - val_calc_mre_K: 0.5684\n", "Epoch 333/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 90.0187 - calc_mre_K: 0.5494 - val_loss: 90.3632 - val_calc_mre_K: 0.5523\n", "Epoch 334/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 90.1588 - calc_mre_K: 0.5503 - val_loss: 85.3304 - val_calc_mre_K: 0.5215\n", "Epoch 335/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 90.3624 - calc_mre_K: 0.5515 - val_loss: 87.1972 - val_calc_mre_K: 0.5329\n", "Epoch 336/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 90.3481 - calc_mre_K: 0.5514 - val_loss: 90.4402 - val_calc_mre_K: 0.5528\n", "Epoch 337/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 89.6775 - calc_mre_K: 0.5473 - val_loss: 91.3080 - val_calc_mre_K: 0.5580\n", "Epoch 338/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 90.9744 - calc_mre_K: 0.5553 - val_loss: 94.9663 - val_calc_mre_K: 0.5804\n", "Epoch 339/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 90.6094 - calc_mre_K: 0.5530 - val_loss: 91.8675 - val_calc_mre_K: 0.5615\n", "Epoch 340/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 89.6018 - calc_mre_K: 0.5469 - val_loss: 91.4848 - val_calc_mre_K: 0.5591\n", "Epoch 341/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 90.0978 - calc_mre_K: 0.5499 - val_loss: 88.3955 - val_calc_mre_K: 0.5403\n", "Epoch 342/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 89.1659 - calc_mre_K: 0.5442 - val_loss: 84.4441 - val_calc_mre_K: 0.5161\n", "Epoch 343/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 90.4879 - calc_mre_K: 0.5523 - val_loss: 86.0728 - val_calc_mre_K: 0.5261\n", "Epoch 344/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 89.0772 - calc_mre_K: 0.5437 - val_loss: 91.1095 - val_calc_mre_K: 0.5568\n", "Epoch 345/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 88.9623 - calc_mre_K: 0.5430 - val_loss: 89.9519 - val_calc_mre_K: 0.5498\n", "Epoch 346/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 89.4066 - calc_mre_K: 0.5457 - val_loss: 92.4123 - val_calc_mre_K: 0.5648\n", "Epoch 347/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 89.6177 - calc_mre_K: 0.5470 - val_loss: 92.9177 - val_calc_mre_K: 0.5679\n", "Epoch 348/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 88.5136 - calc_mre_K: 0.5402 - val_loss: 91.5456 - val_calc_mre_K: 0.5595\n", "Epoch 349/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 88.6974 - calc_mre_K: 0.5414 - val_loss: 86.1277 - val_calc_mre_K: 0.5264\n", "Epoch 350/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 89.3247 - calc_mre_K: 0.5452 - val_loss: 85.3548 - val_calc_mre_K: 0.5217\n", "Epoch 351/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 88.5398 - calc_mre_K: 0.5404 - val_loss: 92.2465 - val_calc_mre_K: 0.5638\n", "Epoch 352/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 88.9964 - calc_mre_K: 0.5432 - val_loss: 87.4843 - val_calc_mre_K: 0.5347\n", "Epoch 353/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 88.8703 - calc_mre_K: 0.5424 - val_loss: 88.0605 - val_calc_mre_K: 0.5382\n", "Epoch 354/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 88.6168 - calc_mre_K: 0.5409 - val_loss: 84.2168 - val_calc_mre_K: 0.5147\n", "Epoch 355/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 89.1193 - calc_mre_K: 0.5439 - val_loss: 87.2859 - val_calc_mre_K: 0.5335\n", "Epoch 356/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 88.2197 - calc_mre_K: 0.5385 - val_loss: 94.0197 - val_calc_mre_K: 0.5746\n", "Epoch 357/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 87.7739 - calc_mre_K: 0.5357 - val_loss: 92.7090 - val_calc_mre_K: 0.5666\n", "Epoch 358/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 89.1492 - calc_mre_K: 0.5441 - val_loss: 87.5420 - val_calc_mre_K: 0.5350\n", "Epoch 359/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 88.5176 - calc_mre_K: 0.5403 - val_loss: 85.7028 - val_calc_mre_K: 0.5238\n", "Epoch 360/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 88.7922 - calc_mre_K: 0.5419 - val_loss: 87.0276 - val_calc_mre_K: 0.5319\n", "Epoch 361/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 89.2310 - calc_mre_K: 0.5446 - val_loss: 87.2908 - val_calc_mre_K: 0.5335\n", "Epoch 362/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 88.2247 - calc_mre_K: 0.5385 - val_loss: 86.7690 - val_calc_mre_K: 0.5303\n", "Epoch 363/2000\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "48000/48000 [==============================] - 3s 53us/step - loss: 88.4310 - calc_mre_K: 0.5397 - val_loss: 91.2048 - val_calc_mre_K: 0.5574\n", "Epoch 364/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 87.4593 - calc_mre_K: 0.5338 - val_loss: 84.1885 - val_calc_mre_K: 0.5145\n", "Epoch 365/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 87.9491 - calc_mre_K: 0.5368 - val_loss: 88.9228 - val_calc_mre_K: 0.5435\n", "Epoch 366/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 88.7196 - calc_mre_K: 0.5415 - val_loss: 87.1904 - val_calc_mre_K: 0.5329\n", "Epoch 367/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 87.0804 - calc_mre_K: 0.5315 - val_loss: 92.9627 - val_calc_mre_K: 0.5682\n", "Epoch 368/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 88.3304 - calc_mre_K: 0.5391 - val_loss: 85.2570 - val_calc_mre_K: 0.5211\n", "Epoch 369/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 88.5224 - calc_mre_K: 0.5403 - val_loss: 92.4780 - val_calc_mre_K: 0.5652\n", "Epoch 370/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 87.6138 - calc_mre_K: 0.5348 - val_loss: 85.7193 - val_calc_mre_K: 0.5239\n", "Epoch 371/2000\n", "48000/48000 [==============================] - 2s 48us/step - loss: 87.2656 - calc_mre_K: 0.5326 - val_loss: 95.0705 - val_calc_mre_K: 0.5810\n", "Epoch 372/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 87.2826 - calc_mre_K: 0.5327 - val_loss: 89.6439 - val_calc_mre_K: 0.5479\n", "Epoch 373/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 87.1010 - calc_mre_K: 0.5316 - val_loss: 89.1116 - val_calc_mre_K: 0.5446\n", "Epoch 374/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 87.2878 - calc_mre_K: 0.5328 - val_loss: 89.9719 - val_calc_mre_K: 0.5499\n", "Epoch 375/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 87.2965 - calc_mre_K: 0.5328 - val_loss: 85.4184 - val_calc_mre_K: 0.5221\n", "Epoch 376/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 87.6315 - calc_mre_K: 0.5349 - val_loss: 83.2941 - val_calc_mre_K: 0.5091\n", "Epoch 377/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 87.4826 - calc_mre_K: 0.5340 - val_loss: 87.7478 - val_calc_mre_K: 0.5363\n", "Epoch 378/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 88.4465 - calc_mre_K: 0.5398 - val_loss: 87.9570 - val_calc_mre_K: 0.5376\n", "Epoch 379/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 86.6256 - calc_mre_K: 0.5287 - val_loss: 83.4293 - val_calc_mre_K: 0.5099\n", "Epoch 380/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 86.3433 - calc_mre_K: 0.5270 - val_loss: 88.1756 - val_calc_mre_K: 0.5389\n", "Epoch 381/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 87.9194 - calc_mre_K: 0.5366 - val_loss: 94.9537 - val_calc_mre_K: 0.5803\n", "Epoch 382/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 88.1680 - calc_mre_K: 0.5381 - val_loss: 84.1755 - val_calc_mre_K: 0.5144\n", "Epoch 383/2000\n", "48000/48000 [==============================] - 2s 49us/step - loss: 86.5488 - calc_mre_K: 0.5283 - val_loss: 84.0552 - val_calc_mre_K: 0.5137\n", "Epoch 384/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 86.9769 - calc_mre_K: 0.5309 - val_loss: 91.9298 - val_calc_mre_K: 0.5619\n", "Epoch 385/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 87.0891 - calc_mre_K: 0.5315 - val_loss: 86.7897 - val_calc_mre_K: 0.5304\n", "Epoch 386/2000\n", "48000/48000 [==============================] - 2s 49us/step - loss: 87.7553 - calc_mre_K: 0.5356 - val_loss: 86.4020 - val_calc_mre_K: 0.5281\n", "Epoch 387/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 86.7116 - calc_mre_K: 0.5292 - val_loss: 87.4734 - val_calc_mre_K: 0.5346\n", "Epoch 388/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 87.9227 - calc_mre_K: 0.5366 - val_loss: 86.3937 - val_calc_mre_K: 0.5280\n", "Epoch 389/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 86.0884 - calc_mre_K: 0.5254 - val_loss: 92.5921 - val_calc_mre_K: 0.5659\n", "Epoch 390/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 88.5215 - calc_mre_K: 0.5403 - val_loss: 85.5472 - val_calc_mre_K: 0.5228\n", "Epoch 391/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 85.6733 - calc_mre_K: 0.5229 - val_loss: 81.8752 - val_calc_mre_K: 0.5004\n", "Epoch 392/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 86.9434 - calc_mre_K: 0.5307 - val_loss: 85.6642 - val_calc_mre_K: 0.5235\n", "Epoch 393/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 86.7052 - calc_mre_K: 0.5292 - val_loss: 84.6832 - val_calc_mre_K: 0.5176\n", "Epoch 394/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 86.6712 - calc_mre_K: 0.5290 - val_loss: 84.7161 - val_calc_mre_K: 0.5178\n", "Epoch 395/2000\n", "48000/48000 [==============================] - 2s 48us/step - loss: 86.3529 - calc_mre_K: 0.5271 - val_loss: 83.9662 - val_calc_mre_K: 0.5132\n", "Epoch 396/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 86.4067 - calc_mre_K: 0.5274 - val_loss: 82.0567 - val_calc_mre_K: 0.5015\n", "Epoch 397/2000\n", "48000/48000 [==============================] - 2s 49us/step - loss: 87.0825 - calc_mre_K: 0.5315 - val_loss: 83.7251 - val_calc_mre_K: 0.5117\n", "Epoch 398/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 85.4258 - calc_mre_K: 0.5214 - val_loss: 84.6355 - val_calc_mre_K: 0.5173\n", "Epoch 399/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 86.3481 - calc_mre_K: 0.5270 - val_loss: 91.6164 - val_calc_mre_K: 0.5599\n", "Epoch 400/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 85.8663 - calc_mre_K: 0.5241 - val_loss: 83.7492 - val_calc_mre_K: 0.5118\n", "Epoch 401/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 86.5218 - calc_mre_K: 0.5281 - val_loss: 81.4844 - val_calc_mre_K: 0.4980\n", "Epoch 402/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 85.1460 - calc_mre_K: 0.5197 - val_loss: 89.6465 - val_calc_mre_K: 0.5479\n", "Epoch 403/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 86.4484 - calc_mre_K: 0.5276 - val_loss: 82.0127 - val_calc_mre_K: 0.5012\n", "Epoch 404/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 85.3829 - calc_mre_K: 0.5211 - val_loss: 82.1146 - val_calc_mre_K: 0.5019\n", "Epoch 405/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 85.8741 - calc_mre_K: 0.5241 - val_loss: 82.2706 - val_calc_mre_K: 0.5028\n", "Epoch 406/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 86.5812 - calc_mre_K: 0.5284 - val_loss: 90.7608 - val_calc_mre_K: 0.5547\n", "Epoch 407/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 85.2983 - calc_mre_K: 0.5206 - val_loss: 86.4578 - val_calc_mre_K: 0.5284\n", "Epoch 408/2000\n", "48000/48000 [==============================] - 2s 49us/step - loss: 85.3998 - calc_mre_K: 0.5212 - val_loss: 82.2868 - val_calc_mre_K: 0.5029\n", "Epoch 409/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 85.3632 - calc_mre_K: 0.5210 - val_loss: 84.1493 - val_calc_mre_K: 0.5143\n", "Epoch 410/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 86.5933 - calc_mre_K: 0.5285 - val_loss: 90.1792 - val_calc_mre_K: 0.5512\n", "Epoch 411/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 85.6908 - calc_mre_K: 0.5230 - val_loss: 84.0880 - val_calc_mre_K: 0.5139\n", "Epoch 412/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 86.0339 - calc_mre_K: 0.5251 - val_loss: 87.3108 - val_calc_mre_K: 0.5336\n", "Epoch 413/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 85.4528 - calc_mre_K: 0.5216 - val_loss: 95.5429 - val_calc_mre_K: 0.5839\n", "Epoch 414/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 85.0066 - calc_mre_K: 0.5188 - val_loss: 86.8831 - val_calc_mre_K: 0.5310\n", "Epoch 415/2000\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "48000/48000 [==============================] - 2s 51us/step - loss: 85.7862 - calc_mre_K: 0.5236 - val_loss: 85.3674 - val_calc_mre_K: 0.5217\n", "Epoch 416/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 84.7693 - calc_mre_K: 0.5174 - val_loss: 87.4442 - val_calc_mre_K: 0.5344\n", "Epoch 417/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 85.3169 - calc_mre_K: 0.5207 - val_loss: 86.1009 - val_calc_mre_K: 0.5262\n", "Epoch 418/2000\n", "48000/48000 [==============================] - 2s 49us/step - loss: 85.3710 - calc_mre_K: 0.5211 - val_loss: 85.1696 - val_calc_mre_K: 0.5205\n", "Epoch 419/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 85.3657 - calc_mre_K: 0.5210 - val_loss: 87.4651 - val_calc_mre_K: 0.5346\n", "Epoch 420/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 85.5295 - calc_mre_K: 0.5220 - val_loss: 83.0808 - val_calc_mre_K: 0.5078\n", "Epoch 421/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 85.3757 - calc_mre_K: 0.5211 - val_loss: 85.9652 - val_calc_mre_K: 0.5254\n", "Epoch 422/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 85.2467 - calc_mre_K: 0.5203 - val_loss: 84.0196 - val_calc_mre_K: 0.5135\n", "Epoch 423/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 84.4289 - calc_mre_K: 0.5153 - val_loss: 87.2073 - val_calc_mre_K: 0.5330\n", "Epoch 424/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 86.0139 - calc_mre_K: 0.5250 - val_loss: 82.8559 - val_calc_mre_K: 0.5064\n", "Epoch 425/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 85.6031 - calc_mre_K: 0.5225 - val_loss: 84.9770 - val_calc_mre_K: 0.5194\n", "Epoch 426/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 84.8861 - calc_mre_K: 0.5181 - val_loss: 81.9947 - val_calc_mre_K: 0.5011\n", "Epoch 427/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 84.2758 - calc_mre_K: 0.5144 - val_loss: 88.7953 - val_calc_mre_K: 0.5427\n", "Epoch 428/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 85.3233 - calc_mre_K: 0.5208 - val_loss: 80.3762 - val_calc_mre_K: 0.4912\n", "Epoch 429/2000\n", "48000/48000 [==============================] - 3s 56us/step - loss: 85.5491 - calc_mre_K: 0.5222 - val_loss: 81.5289 - val_calc_mre_K: 0.4983\n", "Epoch 430/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 84.8320 - calc_mre_K: 0.5178 - val_loss: 81.2536 - val_calc_mre_K: 0.4966\n", "Epoch 431/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 84.4989 - calc_mre_K: 0.5157 - val_loss: 81.2871 - val_calc_mre_K: 0.4968\n", "Epoch 432/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 84.4364 - calc_mre_K: 0.5154 - val_loss: 86.1410 - val_calc_mre_K: 0.5265\n", "Epoch 433/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 84.8892 - calc_mre_K: 0.5181 - val_loss: 88.4534 - val_calc_mre_K: 0.5406\n", "Epoch 434/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 85.0727 - calc_mre_K: 0.5192 - val_loss: 86.5933 - val_calc_mre_K: 0.5292\n", "Epoch 435/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 84.4553 - calc_mre_K: 0.5155 - val_loss: 81.1042 - val_calc_mre_K: 0.4957\n", "Epoch 436/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 83.9461 - calc_mre_K: 0.5124 - val_loss: 84.5066 - val_calc_mre_K: 0.5165\n", "Epoch 437/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 83.9861 - calc_mre_K: 0.5126 - val_loss: 89.4804 - val_calc_mre_K: 0.5469\n", "Epoch 438/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 85.3391 - calc_mre_K: 0.5209 - val_loss: 88.9494 - val_calc_mre_K: 0.5436\n", "Epoch 439/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 83.3708 - calc_mre_K: 0.5089 - val_loss: 89.8847 - val_calc_mre_K: 0.5494\n", "Epoch 440/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 84.6968 - calc_mre_K: 0.5169 - val_loss: 81.3988 - val_calc_mre_K: 0.4975\n", "Epoch 441/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 85.0076 - calc_mre_K: 0.5188 - val_loss: 102.9444 - val_calc_mre_K: 0.6292\n", "Epoch 442/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 84.2531 - calc_mre_K: 0.5142 - val_loss: 80.0978 - val_calc_mre_K: 0.4895\n", "Epoch 443/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 84.5339 - calc_mre_K: 0.5160 - val_loss: 80.8076 - val_calc_mre_K: 0.4939\n", "Epoch 444/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 84.4733 - calc_mre_K: 0.5156 - val_loss: 80.8112 - val_calc_mre_K: 0.4939\n", "Epoch 445/2000\n", "48000/48000 [==============================] - 3s 55us/step - loss: 84.3277 - calc_mre_K: 0.5147 - val_loss: 87.2315 - val_calc_mre_K: 0.5331\n", "Epoch 446/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 83.8744 - calc_mre_K: 0.5119 - val_loss: 79.6366 - val_calc_mre_K: 0.4867\n", "Epoch 447/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 84.2440 - calc_mre_K: 0.5142 - val_loss: 81.4231 - val_calc_mre_K: 0.4976\n", "Epoch 448/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 83.9071 - calc_mre_K: 0.5121 - val_loss: 87.3383 - val_calc_mre_K: 0.5338\n", "Epoch 449/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 84.2936 - calc_mre_K: 0.5145 - val_loss: 92.3833 - val_calc_mre_K: 0.5646\n", "Epoch 450/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 83.8846 - calc_mre_K: 0.5120 - val_loss: 82.6085 - val_calc_mre_K: 0.5049\n", "Epoch 451/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 84.3401 - calc_mre_K: 0.5148 - val_loss: 80.3663 - val_calc_mre_K: 0.4912\n", "Epoch 452/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 83.4264 - calc_mre_K: 0.5092 - val_loss: 90.5961 - val_calc_mre_K: 0.5537\n", "Epoch 453/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 84.2152 - calc_mre_K: 0.5140 - val_loss: 80.6025 - val_calc_mre_K: 0.4926\n", "Epoch 454/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 83.8412 - calc_mre_K: 0.5117 - val_loss: 79.6390 - val_calc_mre_K: 0.4867\n", "Epoch 455/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 83.6216 - calc_mre_K: 0.5104 - val_loss: 83.2902 - val_calc_mre_K: 0.5091\n", "Epoch 456/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 84.1772 - calc_mre_K: 0.5138 - val_loss: 82.0643 - val_calc_mre_K: 0.5016\n", "Epoch 457/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 83.2540 - calc_mre_K: 0.5081 - val_loss: 84.2058 - val_calc_mre_K: 0.5146\n", "Epoch 458/2000\n", "48000/48000 [==============================] - 2s 49us/step - loss: 83.9543 - calc_mre_K: 0.5124 - val_loss: 82.4693 - val_calc_mre_K: 0.5040\n", "Epoch 459/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 83.0160 - calc_mre_K: 0.5067 - val_loss: 91.5357 - val_calc_mre_K: 0.5594\n", "Epoch 460/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 83.4808 - calc_mre_K: 0.5095 - val_loss: 85.0632 - val_calc_mre_K: 0.5199\n", "Epoch 461/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 83.2469 - calc_mre_K: 0.5081 - val_loss: 91.0277 - val_calc_mre_K: 0.5563\n", "Epoch 462/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 85.1645 - calc_mre_K: 0.5198 - val_loss: 87.3782 - val_calc_mre_K: 0.5340\n", "Epoch 463/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 84.0383 - calc_mre_K: 0.5129 - val_loss: 80.8261 - val_calc_mre_K: 0.4940\n", "Epoch 464/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 83.2139 - calc_mre_K: 0.5079 - val_loss: 81.1273 - val_calc_mre_K: 0.4958\n", "Epoch 465/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 82.8814 - calc_mre_K: 0.5059 - val_loss: 80.8322 - val_calc_mre_K: 0.4940\n", "Epoch 466/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 83.5532 - calc_mre_K: 0.5100 - val_loss: 79.1967 - val_calc_mre_K: 0.4840\n", "Epoch 467/2000\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "48000/48000 [==============================] - 2s 51us/step - loss: 83.5453 - calc_mre_K: 0.5099 - val_loss: 80.4774 - val_calc_mre_K: 0.4918\n", "Epoch 468/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 83.4901 - calc_mre_K: 0.5096 - val_loss: 79.9028 - val_calc_mre_K: 0.4883\n", "Epoch 469/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 83.2311 - calc_mre_K: 0.5080 - val_loss: 81.5334 - val_calc_mre_K: 0.4983\n", "Epoch 470/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 83.9753 - calc_mre_K: 0.5125 - val_loss: 81.3385 - val_calc_mre_K: 0.4971\n", "Epoch 471/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 82.7272 - calc_mre_K: 0.5049 - val_loss: 84.2941 - val_calc_mre_K: 0.5152\n", "Epoch 472/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 84.1182 - calc_mre_K: 0.5134 - val_loss: 83.8071 - val_calc_mre_K: 0.5122\n", "Epoch 473/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 83.0383 - calc_mre_K: 0.5068 - val_loss: 80.9138 - val_calc_mre_K: 0.4945\n", "Epoch 474/2000\n", "48000/48000 [==============================] - 2s 49us/step - loss: 82.9969 - calc_mre_K: 0.5066 - val_loss: 81.7797 - val_calc_mre_K: 0.4998\n", "Epoch 475/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 82.7071 - calc_mre_K: 0.5048 - val_loss: 85.8101 - val_calc_mre_K: 0.5244\n", "Epoch 476/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 83.0378 - calc_mre_K: 0.5068 - val_loss: 82.6114 - val_calc_mre_K: 0.5049\n", "Epoch 477/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 82.5705 - calc_mre_K: 0.5040 - val_loss: 80.9737 - val_calc_mre_K: 0.4949\n", "Epoch 478/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 84.3006 - calc_mre_K: 0.5145 - val_loss: 84.5703 - val_calc_mre_K: 0.5169\n", "Epoch 479/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 81.7938 - calc_mre_K: 0.4992 - val_loss: 77.5600 - val_calc_mre_K: 0.4740\n", "Epoch 480/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 83.7278 - calc_mre_K: 0.5110 - val_loss: 79.6321 - val_calc_mre_K: 0.4867\n", "Epoch 481/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 82.2662 - calc_mre_K: 0.5021 - val_loss: 79.1453 - val_calc_mre_K: 0.4837\n", "Epoch 482/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 83.4399 - calc_mre_K: 0.5093 - val_loss: 81.3283 - val_calc_mre_K: 0.4971\n", "Epoch 483/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 82.8574 - calc_mre_K: 0.5057 - val_loss: 82.5899 - val_calc_mre_K: 0.5048\n", "Epoch 484/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 82.7034 - calc_mre_K: 0.5048 - val_loss: 85.5266 - val_calc_mre_K: 0.5227\n", "Epoch 485/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 82.8947 - calc_mre_K: 0.5059 - val_loss: 85.9528 - val_calc_mre_K: 0.5253\n", "Epoch 486/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 82.4989 - calc_mre_K: 0.5035 - val_loss: 81.5871 - val_calc_mre_K: 0.4986\n", "Epoch 487/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 82.5148 - calc_mre_K: 0.5036 - val_loss: 87.6682 - val_calc_mre_K: 0.5358\n", "Epoch 488/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 82.0550 - calc_mre_K: 0.5008 - val_loss: 83.2247 - val_calc_mre_K: 0.5086\n", "Epoch 489/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 82.5240 - calc_mre_K: 0.5037 - val_loss: 82.2563 - val_calc_mre_K: 0.5027\n", "Epoch 490/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 82.3304 - calc_mre_K: 0.5025 - val_loss: 94.2303 - val_calc_mre_K: 0.5759\n", "Epoch 491/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 82.3285 - calc_mre_K: 0.5025 - val_loss: 79.8531 - val_calc_mre_K: 0.4880\n", "Epoch 492/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 83.2595 - calc_mre_K: 0.5082 - val_loss: 82.8625 - val_calc_mre_K: 0.5064\n", "Epoch 493/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 82.4453 - calc_mre_K: 0.5032 - val_loss: 81.3549 - val_calc_mre_K: 0.4972\n", "Epoch 494/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 83.4691 - calc_mre_K: 0.5095 - val_loss: 78.8862 - val_calc_mre_K: 0.4821\n", "Epoch 495/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 81.6053 - calc_mre_K: 0.4981 - val_loss: 80.5431 - val_calc_mre_K: 0.4923\n", "Epoch 496/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 82.1985 - calc_mre_K: 0.5017 - val_loss: 85.6439 - val_calc_mre_K: 0.5234\n", "Epoch 497/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 82.9599 - calc_mre_K: 0.5063 - val_loss: 81.3984 - val_calc_mre_K: 0.4975\n", "Epoch 498/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 81.1375 - calc_mre_K: 0.4952 - val_loss: 79.6726 - val_calc_mre_K: 0.4869\n", "Epoch 499/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 82.1804 - calc_mre_K: 0.5016 - val_loss: 83.3916 - val_calc_mre_K: 0.5097\n", "Epoch 500/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 81.9634 - calc_mre_K: 0.5003 - val_loss: 84.9553 - val_calc_mre_K: 0.5192\n", "Epoch 501/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 83.1247 - calc_mre_K: 0.5074 - val_loss: 81.4493 - val_calc_mre_K: 0.4978\n", "Epoch 502/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 81.9616 - calc_mre_K: 0.5003 - val_loss: 82.5306 - val_calc_mre_K: 0.5044\n", "Epoch 503/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 82.0147 - calc_mre_K: 0.5006 - val_loss: 80.1587 - val_calc_mre_K: 0.4899\n", "Epoch 504/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 82.8134 - calc_mre_K: 0.5055 - val_loss: 80.7115 - val_calc_mre_K: 0.4933\n", "Epoch 505/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 81.6962 - calc_mre_K: 0.4986 - val_loss: 78.3324 - val_calc_mre_K: 0.4787\n", "Epoch 506/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 82.2313 - calc_mre_K: 0.5019 - val_loss: 79.3802 - val_calc_mre_K: 0.4852\n", "Epoch 507/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 82.2760 - calc_mre_K: 0.5022 - val_loss: 94.6301 - val_calc_mre_K: 0.5784\n", "Epoch 508/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 82.5076 - calc_mre_K: 0.5036 - val_loss: 81.0320 - val_calc_mre_K: 0.4952\n", "Epoch 509/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 80.8098 - calc_mre_K: 0.4932 - val_loss: 81.3360 - val_calc_mre_K: 0.4971\n", "Epoch 510/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 82.6015 - calc_mre_K: 0.5042 - val_loss: 85.4223 - val_calc_mre_K: 0.5221\n", "Epoch 511/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 81.8906 - calc_mre_K: 0.4998 - val_loss: 83.7758 - val_calc_mre_K: 0.5120\n", "Epoch 512/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 81.0035 - calc_mre_K: 0.4944 - val_loss: 80.6667 - val_calc_mre_K: 0.4930\n", "Epoch 513/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 81.3512 - calc_mre_K: 0.4965 - val_loss: 82.0099 - val_calc_mre_K: 0.5012\n", "Epoch 514/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 81.9954 - calc_mre_K: 0.5005 - val_loss: 83.1390 - val_calc_mre_K: 0.5081\n", "Epoch 515/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 83.6785 - calc_mre_K: 0.5107 - val_loss: 79.4485 - val_calc_mre_K: 0.4856\n", "Epoch 516/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 81.3299 - calc_mre_K: 0.4964 - val_loss: 81.3948 - val_calc_mre_K: 0.4975\n", "Epoch 517/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 81.1250 - calc_mre_K: 0.4951 - val_loss: 81.8768 - val_calc_mre_K: 0.5004\n", "Epoch 518/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 81.2683 - calc_mre_K: 0.4960 - val_loss: 87.4565 - val_calc_mre_K: 0.5345\n", "Epoch 519/2000\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "48000/48000 [==============================] - 3s 53us/step - loss: 81.9307 - calc_mre_K: 0.5001 - val_loss: 77.4843 - val_calc_mre_K: 0.4736\n", "Epoch 520/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 82.6650 - calc_mre_K: 0.5045 - val_loss: 80.0393 - val_calc_mre_K: 0.4892\n", "Epoch 521/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 80.1645 - calc_mre_K: 0.4893 - val_loss: 79.4120 - val_calc_mre_K: 0.4853\n", "Epoch 522/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 81.4566 - calc_mre_K: 0.4972 - val_loss: 81.4097 - val_calc_mre_K: 0.4975\n", "Epoch 523/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 80.9915 - calc_mre_K: 0.4943 - val_loss: 78.4441 - val_calc_mre_K: 0.4794\n", "Epoch 524/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 81.9312 - calc_mre_K: 0.5001 - val_loss: 77.2055 - val_calc_mre_K: 0.4719\n", "Epoch 525/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 81.4020 - calc_mre_K: 0.4968 - val_loss: 79.8432 - val_calc_mre_K: 0.4880\n", "Epoch 526/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 80.7703 - calc_mre_K: 0.4930 - val_loss: 82.1272 - val_calc_mre_K: 0.5019\n", "Epoch 527/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 81.5344 - calc_mre_K: 0.4976 - val_loss: 80.7815 - val_calc_mre_K: 0.4937\n", "Epoch 528/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 80.9556 - calc_mre_K: 0.4941 - val_loss: 79.5202 - val_calc_mre_K: 0.4860\n", "Epoch 529/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 81.6222 - calc_mre_K: 0.4982 - val_loss: 79.4104 - val_calc_mre_K: 0.4853\n", "Epoch 530/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 81.0558 - calc_mre_K: 0.4947 - val_loss: 79.4576 - val_calc_mre_K: 0.4856\n", "Epoch 531/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 80.3335 - calc_mre_K: 0.4903 - val_loss: 78.7548 - val_calc_mre_K: 0.4813\n", "Epoch 532/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 80.2822 - calc_mre_K: 0.4900 - val_loss: 83.2442 - val_calc_mre_K: 0.5088\n", "Epoch 533/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 81.0205 - calc_mre_K: 0.4945 - val_loss: 87.6607 - val_calc_mre_K: 0.5357\n", "Epoch 534/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 80.9163 - calc_mre_K: 0.4939 - val_loss: 76.4528 - val_calc_mre_K: 0.4672\n", "Epoch 535/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 80.7187 - calc_mre_K: 0.4927 - val_loss: 79.4184 - val_calc_mre_K: 0.4854\n", "Epoch 536/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 80.4491 - calc_mre_K: 0.4910 - val_loss: 79.0348 - val_calc_mre_K: 0.4830\n", "Epoch 537/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 80.0500 - calc_mre_K: 0.4886 - val_loss: 79.4705 - val_calc_mre_K: 0.4857\n", "Epoch 538/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 80.4897 - calc_mre_K: 0.4913 - val_loss: 82.8092 - val_calc_mre_K: 0.5061\n", "Epoch 539/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 80.3116 - calc_mre_K: 0.4902 - val_loss: 78.5460 - val_calc_mre_K: 0.4800\n", "Epoch 540/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 80.8021 - calc_mre_K: 0.4932 - val_loss: 78.3198 - val_calc_mre_K: 0.4787\n", "Epoch 541/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 79.9388 - calc_mre_K: 0.4879 - val_loss: 76.3456 - val_calc_mre_K: 0.4666\n", "Epoch 542/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 80.2659 - calc_mre_K: 0.4899 - val_loss: 80.1394 - val_calc_mre_K: 0.4898\n", "Epoch 543/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 80.5261 - calc_mre_K: 0.4915 - val_loss: 79.9564 - val_calc_mre_K: 0.4887\n", "Epoch 544/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 79.8455 - calc_mre_K: 0.4873 - val_loss: 78.0595 - val_calc_mre_K: 0.4771\n", "Epoch 545/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 80.3372 - calc_mre_K: 0.4903 - val_loss: 76.4945 - val_calc_mre_K: 0.4675\n", "Epoch 546/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 79.5684 - calc_mre_K: 0.4856 - val_loss: 77.4809 - val_calc_mre_K: 0.4735\n", "Epoch 547/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 79.4013 - calc_mre_K: 0.4846 - val_loss: 79.6596 - val_calc_mre_K: 0.4869\n", "Epoch 548/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 79.7703 - calc_mre_K: 0.4869 - val_loss: 77.9238 - val_calc_mre_K: 0.4762\n", "Epoch 549/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 79.3975 - calc_mre_K: 0.4846 - val_loss: 80.0276 - val_calc_mre_K: 0.4891\n", "Epoch 550/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 79.7128 - calc_mre_K: 0.4865 - val_loss: 77.4195 - val_calc_mre_K: 0.4732\n", "Epoch 551/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 80.8197 - calc_mre_K: 0.4933 - val_loss: 88.9530 - val_calc_mre_K: 0.5436\n", "Epoch 552/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 79.3593 - calc_mre_K: 0.4844 - val_loss: 77.9631 - val_calc_mre_K: 0.4765\n", "Epoch 553/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 79.5088 - calc_mre_K: 0.4853 - val_loss: 82.0954 - val_calc_mre_K: 0.5017\n", "Epoch 554/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 82.1282 - calc_mre_K: 0.5013 - val_loss: 78.2913 - val_calc_mre_K: 0.4785\n", "Epoch 555/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 78.5980 - calc_mre_K: 0.4797 - val_loss: 79.0503 - val_calc_mre_K: 0.4831\n", "Epoch 556/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 78.6166 - calc_mre_K: 0.4798 - val_loss: 80.1970 - val_calc_mre_K: 0.4901\n", "Epoch 557/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 79.8482 - calc_mre_K: 0.4874 - val_loss: 77.2790 - val_calc_mre_K: 0.4723\n", "Epoch 558/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 79.5353 - calc_mre_K: 0.4854 - val_loss: 84.9852 - val_calc_mre_K: 0.5194\n", "Epoch 559/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 79.3965 - calc_mre_K: 0.4846 - val_loss: 76.1430 - val_calc_mre_K: 0.4654\n", "Epoch 560/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 79.7088 - calc_mre_K: 0.4865 - val_loss: 82.3685 - val_calc_mre_K: 0.5034\n", "Epoch 561/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 78.6799 - calc_mre_K: 0.4802 - val_loss: 92.0983 - val_calc_mre_K: 0.5629\n", "Epoch 562/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 78.5667 - calc_mre_K: 0.4795 - val_loss: 78.2633 - val_calc_mre_K: 0.4783\n", "Epoch 563/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 78.8523 - calc_mre_K: 0.4813 - val_loss: 78.8908 - val_calc_mre_K: 0.4822\n", "Epoch 564/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 79.2787 - calc_mre_K: 0.4839 - val_loss: 81.4291 - val_calc_mre_K: 0.4977\n", "Epoch 565/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 79.2967 - calc_mre_K: 0.4840 - val_loss: 77.1630 - val_calc_mre_K: 0.4716\n", "Epoch 566/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 78.2869 - calc_mre_K: 0.4778 - val_loss: 77.6243 - val_calc_mre_K: 0.4744\n", "Epoch 567/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 78.2733 - calc_mre_K: 0.4777 - val_loss: 75.9188 - val_calc_mre_K: 0.4640\n", "Epoch 568/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 79.0621 - calc_mre_K: 0.4826 - val_loss: 76.6390 - val_calc_mre_K: 0.4684\n", "Epoch 569/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 78.9082 - calc_mre_K: 0.4816 - val_loss: 76.9885 - val_calc_mre_K: 0.4705\n", "Epoch 570/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 78.7515 - calc_mre_K: 0.4807 - val_loss: 78.9202 - val_calc_mre_K: 0.4823\n", "Epoch 571/2000\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "48000/48000 [==============================] - 2s 52us/step - loss: 79.1038 - calc_mre_K: 0.4828 - val_loss: 77.3020 - val_calc_mre_K: 0.4724\n", "Epoch 572/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 78.5288 - calc_mre_K: 0.4793 - val_loss: 80.5382 - val_calc_mre_K: 0.4922\n", "Epoch 573/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 79.0160 - calc_mre_K: 0.4823 - val_loss: 77.1178 - val_calc_mre_K: 0.4713\n", "Epoch 574/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 77.7734 - calc_mre_K: 0.4747 - val_loss: 78.0108 - val_calc_mre_K: 0.4768\n", "Epoch 575/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 79.0328 - calc_mre_K: 0.4824 - val_loss: 99.5002 - val_calc_mre_K: 0.6081\n", "Epoch 576/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 79.1704 - calc_mre_K: 0.4832 - val_loss: 81.2237 - val_calc_mre_K: 0.4964\n", "Epoch 577/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 78.4922 - calc_mre_K: 0.4791 - val_loss: 77.0759 - val_calc_mre_K: 0.4711\n", "Epoch 578/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 78.7552 - calc_mre_K: 0.4807 - val_loss: 79.2664 - val_calc_mre_K: 0.4844\n", "Epoch 579/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 78.6874 - calc_mre_K: 0.4803 - val_loss: 77.5483 - val_calc_mre_K: 0.4740\n", "Epoch 580/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 78.3672 - calc_mre_K: 0.4783 - val_loss: 76.1830 - val_calc_mre_K: 0.4656\n", "Epoch 581/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 77.9002 - calc_mre_K: 0.4755 - val_loss: 77.2363 - val_calc_mre_K: 0.4720\n", "Epoch 582/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 78.2323 - calc_mre_K: 0.4775 - val_loss: 75.5331 - val_calc_mre_K: 0.4616\n", "Epoch 583/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 78.6896 - calc_mre_K: 0.4803 - val_loss: 83.8380 - val_calc_mre_K: 0.5124\n", "Epoch 584/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 78.3468 - calc_mre_K: 0.4782 - val_loss: 76.1053 - val_calc_mre_K: 0.4651\n", "Epoch 585/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 77.6933 - calc_mre_K: 0.4742 - val_loss: 74.8827 - val_calc_mre_K: 0.4577\n", "Epoch 586/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 78.0269 - calc_mre_K: 0.4762 - val_loss: 89.1461 - val_calc_mre_K: 0.5448\n", "Epoch 587/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 78.1975 - calc_mre_K: 0.4773 - val_loss: 75.5256 - val_calc_mre_K: 0.4616\n", "Epoch 588/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 79.8542 - calc_mre_K: 0.4874 - val_loss: 85.3280 - val_calc_mre_K: 0.5215\n", "Epoch 589/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 77.5551 - calc_mre_K: 0.4734 - val_loss: 78.1732 - val_calc_mre_K: 0.4778\n", "Epoch 590/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 78.3916 - calc_mre_K: 0.4785 - val_loss: 76.3000 - val_calc_mre_K: 0.4663\n", "Epoch 591/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 78.5718 - calc_mre_K: 0.4796 - val_loss: 75.3156 - val_calc_mre_K: 0.4603\n", "Epoch 592/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 78.4115 - calc_mre_K: 0.4786 - val_loss: 75.9747 - val_calc_mre_K: 0.4643\n", "Epoch 593/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 77.2905 - calc_mre_K: 0.4717 - val_loss: 77.3951 - val_calc_mre_K: 0.4730\n", "Epoch 594/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 78.2779 - calc_mre_K: 0.4778 - val_loss: 79.0500 - val_calc_mre_K: 0.4831\n", "Epoch 595/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 80.0868 - calc_mre_K: 0.4888 - val_loss: 78.6989 - val_calc_mre_K: 0.4810\n", "Epoch 596/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 77.2207 - calc_mre_K: 0.4713 - val_loss: 75.7702 - val_calc_mre_K: 0.4631\n", "Epoch 597/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 77.9375 - calc_mre_K: 0.4757 - val_loss: 80.1789 - val_calc_mre_K: 0.4900\n", "Epoch 598/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 77.7830 - calc_mre_K: 0.4747 - val_loss: 75.6367 - val_calc_mre_K: 0.4623\n", "Epoch 599/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 77.8996 - calc_mre_K: 0.4755 - val_loss: 75.5354 - val_calc_mre_K: 0.4617\n", "Epoch 600/2000\n", "48000/48000 [==============================] - 2s 49us/step - loss: 77.5005 - calc_mre_K: 0.4730 - val_loss: 76.2051 - val_calc_mre_K: 0.4657\n", "Epoch 601/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 77.7848 - calc_mre_K: 0.4748 - val_loss: 81.0561 - val_calc_mre_K: 0.4954\n", "Epoch 602/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 78.6516 - calc_mre_K: 0.4801 - val_loss: 76.3303 - val_calc_mre_K: 0.4665\n", "Epoch 603/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 77.4334 - calc_mre_K: 0.4726 - val_loss: 75.5365 - val_calc_mre_K: 0.4617\n", "Epoch 604/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 78.6044 - calc_mre_K: 0.4798 - val_loss: 85.8872 - val_calc_mre_K: 0.5249\n", "Epoch 605/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 77.6670 - calc_mre_K: 0.4740 - val_loss: 76.3514 - val_calc_mre_K: 0.4666\n", "Epoch 606/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 77.5818 - calc_mre_K: 0.4735 - val_loss: 79.2760 - val_calc_mre_K: 0.4845\n", "Epoch 607/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 78.1415 - calc_mre_K: 0.4769 - val_loss: 75.9864 - val_calc_mre_K: 0.4644\n", "Epoch 608/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 77.8271 - calc_mre_K: 0.4750 - val_loss: 85.6233 - val_calc_mre_K: 0.5233\n", "Epoch 609/2000\n", "48000/48000 [==============================] - 3s 55us/step - loss: 77.6805 - calc_mre_K: 0.4741 - val_loss: 75.5377 - val_calc_mre_K: 0.4617\n", "Epoch 610/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 78.2192 - calc_mre_K: 0.4774 - val_loss: 76.5768 - val_calc_mre_K: 0.4680\n", "Epoch 611/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 76.5556 - calc_mre_K: 0.4673 - val_loss: 79.8458 - val_calc_mre_K: 0.4880\n", "Epoch 612/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 77.5932 - calc_mre_K: 0.4736 - val_loss: 76.4385 - val_calc_mre_K: 0.4672\n", "Epoch 613/2000\n", "48000/48000 [==============================] - 2s 49us/step - loss: 77.8906 - calc_mre_K: 0.4754 - val_loss: 79.3098 - val_calc_mre_K: 0.4847\n", "Epoch 614/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 77.4078 - calc_mre_K: 0.4725 - val_loss: 80.7717 - val_calc_mre_K: 0.4936\n", "Epoch 615/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 77.6592 - calc_mre_K: 0.4740 - val_loss: 76.3392 - val_calc_mre_K: 0.4666\n", "Epoch 616/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 77.0155 - calc_mre_K: 0.4701 - val_loss: 76.0185 - val_calc_mre_K: 0.4646\n", "Epoch 617/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 79.0962 - calc_mre_K: 0.4828 - val_loss: 77.7860 - val_calc_mre_K: 0.4754\n", "Epoch 618/2000\n", "48000/48000 [==============================] - 2s 48us/step - loss: 77.0872 - calc_mre_K: 0.4705 - val_loss: 80.7127 - val_calc_mre_K: 0.4933\n", "Epoch 619/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 77.0748 - calc_mre_K: 0.4704 - val_loss: 80.8325 - val_calc_mre_K: 0.4940\n", "Epoch 620/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 76.8771 - calc_mre_K: 0.4692 - val_loss: 80.3949 - val_calc_mre_K: 0.4913\n", "Epoch 621/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 77.4156 - calc_mre_K: 0.4725 - val_loss: 74.7927 - val_calc_mre_K: 0.4571\n", "Epoch 622/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 77.7287 - calc_mre_K: 0.4744 - val_loss: 76.7294 - val_calc_mre_K: 0.4689\n", "Epoch 623/2000\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "48000/48000 [==============================] - 2s 50us/step - loss: 77.7628 - calc_mre_K: 0.4746 - val_loss: 76.4259 - val_calc_mre_K: 0.4671\n", "Epoch 624/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 76.4415 - calc_mre_K: 0.4666 - val_loss: 77.8054 - val_calc_mre_K: 0.4755\n", "Epoch 625/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 76.7878 - calc_mre_K: 0.4687 - val_loss: 75.8927 - val_calc_mre_K: 0.4638\n", "Epoch 626/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 77.4494 - calc_mre_K: 0.4727 - val_loss: 75.8223 - val_calc_mre_K: 0.4634\n", "Epoch 627/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 77.9056 - calc_mre_K: 0.4755 - val_loss: 79.2699 - val_calc_mre_K: 0.4845\n", "Epoch 628/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 77.4744 - calc_mre_K: 0.4729 - val_loss: 82.6949 - val_calc_mre_K: 0.5054\n", "Epoch 629/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 76.4943 - calc_mre_K: 0.4669 - val_loss: 75.6243 - val_calc_mre_K: 0.4622\n", "Epoch 630/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 77.2699 - calc_mre_K: 0.4716 - val_loss: 75.8055 - val_calc_mre_K: 0.4633\n", "Epoch 631/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 77.2239 - calc_mre_K: 0.4713 - val_loss: 75.8226 - val_calc_mre_K: 0.4634\n", "Epoch 632/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 76.5761 - calc_mre_K: 0.4674 - val_loss: 80.4214 - val_calc_mre_K: 0.4915\n", "Epoch 633/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 77.6081 - calc_mre_K: 0.4737 - val_loss: 74.2968 - val_calc_mre_K: 0.4541\n", "Epoch 634/2000\n", "48000/48000 [==============================] - 2s 49us/step - loss: 77.4080 - calc_mre_K: 0.4725 - val_loss: 75.3571 - val_calc_mre_K: 0.4606\n", "Epoch 635/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 76.4933 - calc_mre_K: 0.4669 - val_loss: 82.4477 - val_calc_mre_K: 0.5039\n", "Epoch 636/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 77.3498 - calc_mre_K: 0.4721 - val_loss: 75.6026 - val_calc_mre_K: 0.4621\n", "Epoch 637/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 77.0594 - calc_mre_K: 0.4703 - val_loss: 75.4148 - val_calc_mre_K: 0.4609\n", "Epoch 638/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 77.0159 - calc_mre_K: 0.4701 - val_loss: 74.9684 - val_calc_mre_K: 0.4582\n", "Epoch 639/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 77.4576 - calc_mre_K: 0.4728 - val_loss: 78.2255 - val_calc_mre_K: 0.4781\n", "Epoch 640/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 76.7792 - calc_mre_K: 0.4686 - val_loss: 81.5454 - val_calc_mre_K: 0.4984\n", "Epoch 641/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 76.7995 - calc_mre_K: 0.4687 - val_loss: 76.7738 - val_calc_mre_K: 0.4692\n", "Epoch 642/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 76.8567 - calc_mre_K: 0.4691 - val_loss: 74.0150 - val_calc_mre_K: 0.4523\n", "Epoch 643/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 76.4782 - calc_mre_K: 0.4668 - val_loss: 75.5087 - val_calc_mre_K: 0.4615\n", "Epoch 644/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 76.6800 - calc_mre_K: 0.4680 - val_loss: 74.9136 - val_calc_mre_K: 0.4578\n", "Epoch 645/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 77.0162 - calc_mre_K: 0.4701 - val_loss: 75.7475 - val_calc_mre_K: 0.4630\n", "Epoch 646/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 77.1098 - calc_mre_K: 0.4706 - val_loss: 76.3452 - val_calc_mre_K: 0.4666\n", "Epoch 647/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 76.4622 - calc_mre_K: 0.4667 - val_loss: 73.9322 - val_calc_mre_K: 0.4518\n", "Epoch 648/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 77.0078 - calc_mre_K: 0.4700 - val_loss: 75.9888 - val_calc_mre_K: 0.4644\n", "Epoch 649/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 76.2060 - calc_mre_K: 0.4651 - val_loss: 85.3601 - val_calc_mre_K: 0.5217\n", "Epoch 650/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 77.6062 - calc_mre_K: 0.4737 - val_loss: 76.1147 - val_calc_mre_K: 0.4652\n", "Epoch 651/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 75.9817 - calc_mre_K: 0.4638 - val_loss: 77.5800 - val_calc_mre_K: 0.4741\n", "Epoch 652/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 77.4938 - calc_mre_K: 0.4730 - val_loss: 84.1910 - val_calc_mre_K: 0.5145\n", "Epoch 653/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 76.1384 - calc_mre_K: 0.4647 - val_loss: 77.4940 - val_calc_mre_K: 0.4736\n", "Epoch 654/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 76.5940 - calc_mre_K: 0.4675 - val_loss: 81.7681 - val_calc_mre_K: 0.4997\n", "Epoch 655/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 76.6279 - calc_mre_K: 0.4677 - val_loss: 78.9758 - val_calc_mre_K: 0.4827\n", "Epoch 656/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 76.5988 - calc_mre_K: 0.4675 - val_loss: 74.3497 - val_calc_mre_K: 0.4544\n", "Epoch 657/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 76.5490 - calc_mre_K: 0.4672 - val_loss: 76.2299 - val_calc_mre_K: 0.4659\n", "Epoch 658/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 76.7622 - calc_mre_K: 0.4685 - val_loss: 79.2833 - val_calc_mre_K: 0.4845\n", "Epoch 659/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 76.5077 - calc_mre_K: 0.4670 - val_loss: 75.1735 - val_calc_mre_K: 0.4594\n", "Epoch 660/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 76.5935 - calc_mre_K: 0.4675 - val_loss: 84.0199 - val_calc_mre_K: 0.5135\n", "Epoch 661/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 76.6214 - calc_mre_K: 0.4677 - val_loss: 74.3638 - val_calc_mre_K: 0.4545\n", "Epoch 662/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 75.9871 - calc_mre_K: 0.4638 - val_loss: 73.8487 - val_calc_mre_K: 0.4513\n", "Epoch 663/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 76.0754 - calc_mre_K: 0.4643 - val_loss: 75.2188 - val_calc_mre_K: 0.4597\n", "Epoch 664/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 76.7228 - calc_mre_K: 0.4683 - val_loss: 75.0648 - val_calc_mre_K: 0.4588\n", "Epoch 665/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 76.6878 - calc_mre_K: 0.4681 - val_loss: 82.8250 - val_calc_mre_K: 0.5062\n", "Epoch 666/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 76.6697 - calc_mre_K: 0.4680 - val_loss: 75.2639 - val_calc_mre_K: 0.4600\n", "Epoch 667/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 75.7617 - calc_mre_K: 0.4624 - val_loss: 81.9704 - val_calc_mre_K: 0.5010\n", "Epoch 668/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 76.0024 - calc_mre_K: 0.4639 - val_loss: 77.0529 - val_calc_mre_K: 0.4709\n", "Epoch 669/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 76.6019 - calc_mre_K: 0.4675 - val_loss: 74.7767 - val_calc_mre_K: 0.4570\n", "Epoch 670/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 76.0860 - calc_mre_K: 0.4644 - val_loss: 76.0334 - val_calc_mre_K: 0.4647\n", "Epoch 671/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 75.7303 - calc_mre_K: 0.4622 - val_loss: 76.9445 - val_calc_mre_K: 0.4702\n", "Epoch 672/2000\n", "48000/48000 [==============================] - 2s 49us/step - loss: 76.3476 - calc_mre_K: 0.4660 - val_loss: 74.3722 - val_calc_mre_K: 0.4545\n", "Epoch 673/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 76.1754 - calc_mre_K: 0.4649 - val_loss: 83.2348 - val_calc_mre_K: 0.5087\n", "Epoch 674/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 75.8796 - calc_mre_K: 0.4631 - val_loss: 77.3340 - val_calc_mre_K: 0.4726\n", "Epoch 675/2000\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "48000/48000 [==============================] - 3s 52us/step - loss: 76.2967 - calc_mre_K: 0.4657 - val_loss: 73.3028 - val_calc_mre_K: 0.4480\n", "Epoch 676/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 76.3319 - calc_mre_K: 0.4659 - val_loss: 75.1129 - val_calc_mre_K: 0.4590\n", "Epoch 677/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 76.0374 - calc_mre_K: 0.4641 - val_loss: 77.5969 - val_calc_mre_K: 0.4742\n", "Epoch 678/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 76.7287 - calc_mre_K: 0.4683 - val_loss: 74.6554 - val_calc_mre_K: 0.4563\n", "Epoch 679/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 76.0910 - calc_mre_K: 0.4644 - val_loss: 73.9943 - val_calc_mre_K: 0.4522\n", "Epoch 680/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 75.8340 - calc_mre_K: 0.4629 - val_loss: 93.0210 - val_calc_mre_K: 0.5685\n", "Epoch 681/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 75.8063 - calc_mre_K: 0.4627 - val_loss: 74.3475 - val_calc_mre_K: 0.4544\n", "Epoch 682/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 76.1020 - calc_mre_K: 0.4645 - val_loss: 77.4519 - val_calc_mre_K: 0.4733\n", "Epoch 683/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 75.8312 - calc_mre_K: 0.4628 - val_loss: 82.7156 - val_calc_mre_K: 0.5055\n", "Epoch 684/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 75.8341 - calc_mre_K: 0.4629 - val_loss: 77.0233 - val_calc_mre_K: 0.4707\n", "Epoch 685/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 76.1792 - calc_mre_K: 0.4650 - val_loss: 74.5906 - val_calc_mre_K: 0.4559\n", "Epoch 686/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 76.4742 - calc_mre_K: 0.4668 - val_loss: 74.8668 - val_calc_mre_K: 0.4576\n", "Epoch 687/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 75.4463 - calc_mre_K: 0.4605 - val_loss: 74.0643 - val_calc_mre_K: 0.4527\n", "Epoch 688/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 75.8331 - calc_mre_K: 0.4628 - val_loss: 75.3094 - val_calc_mre_K: 0.4603\n", "Epoch 689/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 75.8157 - calc_mre_K: 0.4627 - val_loss: 78.4351 - val_calc_mre_K: 0.4794\n", "Epoch 690/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 75.8586 - calc_mre_K: 0.4630 - val_loss: 80.3384 - val_calc_mre_K: 0.4910\n", "Epoch 691/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 77.3010 - calc_mre_K: 0.4718 - val_loss: 75.4843 - val_calc_mre_K: 0.4613\n", "Epoch 692/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 75.0688 - calc_mre_K: 0.4582 - val_loss: 74.7796 - val_calc_mre_K: 0.4570\n", "Epoch 693/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 75.4165 - calc_mre_K: 0.4603 - val_loss: 74.1000 - val_calc_mre_K: 0.4529\n", "Epoch 694/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 75.5394 - calc_mre_K: 0.4611 - val_loss: 74.6463 - val_calc_mre_K: 0.4562\n", "Epoch 695/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 76.2745 - calc_mre_K: 0.4655 - val_loss: 75.2185 - val_calc_mre_K: 0.4597\n", "Epoch 696/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 75.3003 - calc_mre_K: 0.4596 - val_loss: 80.0858 - val_calc_mre_K: 0.4894\n", "Epoch 697/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 76.6806 - calc_mre_K: 0.4680 - val_loss: 74.5250 - val_calc_mre_K: 0.4555\n", "Epoch 698/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 76.0320 - calc_mre_K: 0.4641 - val_loss: 81.9585 - val_calc_mre_K: 0.5009\n", "Epoch 699/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 75.2457 - calc_mre_K: 0.4593 - val_loss: 78.7075 - val_calc_mre_K: 0.4810\n", "Epoch 700/2000\n", "48000/48000 [==============================] - 3s 55us/step - loss: 75.7782 - calc_mre_K: 0.4625 - val_loss: 72.7481 - val_calc_mre_K: 0.4446\n", "Epoch 701/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 75.7799 - calc_mre_K: 0.4625 - val_loss: 77.8811 - val_calc_mre_K: 0.4760\n", "Epoch 702/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 75.4130 - calc_mre_K: 0.4603 - val_loss: 76.1580 - val_calc_mre_K: 0.4654\n", "Epoch 703/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 75.2641 - calc_mre_K: 0.4594 - val_loss: 76.5979 - val_calc_mre_K: 0.4681\n", "Epoch 704/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 75.3956 - calc_mre_K: 0.4602 - val_loss: 78.0766 - val_calc_mre_K: 0.4772\n", "Epoch 705/2000\n", "48000/48000 [==============================] - 3s 55us/step - loss: 75.7739 - calc_mre_K: 0.4625 - val_loss: 74.4438 - val_calc_mre_K: 0.4550\n", "Epoch 706/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 74.9387 - calc_mre_K: 0.4574 - val_loss: 73.9057 - val_calc_mre_K: 0.4517\n", "Epoch 707/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 76.0042 - calc_mre_K: 0.4639 - val_loss: 77.8957 - val_calc_mre_K: 0.4761\n", "Epoch 708/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 75.3819 - calc_mre_K: 0.4601 - val_loss: 75.2639 - val_calc_mre_K: 0.4600\n", "Epoch 709/2000\n", "48000/48000 [==============================] - 3s 55us/step - loss: 75.5852 - calc_mre_K: 0.4613 - val_loss: 72.6963 - val_calc_mre_K: 0.4443\n", "Epoch 710/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 75.6962 - calc_mre_K: 0.4620 - val_loss: 79.6773 - val_calc_mre_K: 0.4870\n", "Epoch 711/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 75.4561 - calc_mre_K: 0.4605 - val_loss: 78.2928 - val_calc_mre_K: 0.4785\n", "Epoch 712/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 75.3355 - calc_mre_K: 0.4598 - val_loss: 82.0169 - val_calc_mre_K: 0.5013\n", "Epoch 713/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 75.6424 - calc_mre_K: 0.4617 - val_loss: 75.8993 - val_calc_mre_K: 0.4638\n", "Epoch 714/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 76.0899 - calc_mre_K: 0.4644 - val_loss: 75.4356 - val_calc_mre_K: 0.4610\n", "Epoch 715/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 75.0133 - calc_mre_K: 0.4578 - val_loss: 75.1615 - val_calc_mre_K: 0.4594\n", "Epoch 716/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 75.0025 - calc_mre_K: 0.4578 - val_loss: 79.8304 - val_calc_mre_K: 0.4879\n", "Epoch 717/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 75.2140 - calc_mre_K: 0.4591 - val_loss: 75.2749 - val_calc_mre_K: 0.4601\n", "Epoch 718/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 75.4905 - calc_mre_K: 0.4608 - val_loss: 74.6027 - val_calc_mre_K: 0.4559\n", "Epoch 719/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 75.4244 - calc_mre_K: 0.4604 - val_loss: 74.8684 - val_calc_mre_K: 0.4576\n", "Epoch 720/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 75.9031 - calc_mre_K: 0.4633 - val_loss: 75.0529 - val_calc_mre_K: 0.4587\n", "Epoch 721/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 75.3857 - calc_mre_K: 0.4601 - val_loss: 80.4717 - val_calc_mre_K: 0.4918\n", "Epoch 722/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 75.3779 - calc_mre_K: 0.4601 - val_loss: 89.1777 - val_calc_mre_K: 0.5450\n", "Epoch 723/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 74.9998 - calc_mre_K: 0.4578 - val_loss: 80.5357 - val_calc_mre_K: 0.4922\n", "Epoch 724/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 75.5422 - calc_mre_K: 0.4611 - val_loss: 73.9738 - val_calc_mre_K: 0.4521\n", "Epoch 725/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 75.3028 - calc_mre_K: 0.4596 - val_loss: 73.8335 - val_calc_mre_K: 0.4512\n", "Epoch 726/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 75.2767 - calc_mre_K: 0.4595 - val_loss: 72.7816 - val_calc_mre_K: 0.4448\n", "Epoch 727/2000\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "48000/48000 [==============================] - 2s 52us/step - loss: 75.4082 - calc_mre_K: 0.4603 - val_loss: 73.3399 - val_calc_mre_K: 0.4482\n", "Epoch 728/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 74.8492 - calc_mre_K: 0.4568 - val_loss: 76.1310 - val_calc_mre_K: 0.4653\n", "Epoch 729/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 75.2312 - calc_mre_K: 0.4592 - val_loss: 72.0959 - val_calc_mre_K: 0.4406\n", "Epoch 730/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 75.8252 - calc_mre_K: 0.4628 - val_loss: 72.6540 - val_calc_mre_K: 0.4440\n", "Epoch 731/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 74.9913 - calc_mre_K: 0.4577 - val_loss: 73.9318 - val_calc_mre_K: 0.4518\n", "Epoch 732/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 75.3091 - calc_mre_K: 0.4597 - val_loss: 77.6359 - val_calc_mre_K: 0.4745\n", "Epoch 733/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 74.3615 - calc_mre_K: 0.4539 - val_loss: 74.8717 - val_calc_mre_K: 0.4576\n", "Epoch 734/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 75.4607 - calc_mre_K: 0.4606 - val_loss: 74.6581 - val_calc_mre_K: 0.4563\n", "Epoch 735/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 75.4732 - calc_mre_K: 0.4607 - val_loss: 75.2694 - val_calc_mre_K: 0.4600\n", "Epoch 736/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 75.0357 - calc_mre_K: 0.4580 - val_loss: 72.4986 - val_calc_mre_K: 0.4431\n", "Epoch 737/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 75.0608 - calc_mre_K: 0.4581 - val_loss: 80.1526 - val_calc_mre_K: 0.4899\n", "Epoch 738/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 74.2541 - calc_mre_K: 0.4532 - val_loss: 76.8176 - val_calc_mre_K: 0.4695\n", "Epoch 739/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 75.0672 - calc_mre_K: 0.4582 - val_loss: 73.6833 - val_calc_mre_K: 0.4503\n", "Epoch 740/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 74.9701 - calc_mre_K: 0.4576 - val_loss: 71.6015 - val_calc_mre_K: 0.4376\n", "Epoch 741/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 74.9095 - calc_mre_K: 0.4572 - val_loss: 75.1805 - val_calc_mre_K: 0.4595\n", "Epoch 742/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 74.8450 - calc_mre_K: 0.4568 - val_loss: 74.9865 - val_calc_mre_K: 0.4583\n", "Epoch 743/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 74.9314 - calc_mre_K: 0.4573 - val_loss: 74.9089 - val_calc_mre_K: 0.4578\n", "Epoch 744/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 74.1358 - calc_mre_K: 0.4525 - val_loss: 77.2990 - val_calc_mre_K: 0.4724\n", "Epoch 745/2000\n", "48000/48000 [==============================] - 2s 47us/step - loss: 74.6135 - calc_mre_K: 0.4554 - val_loss: 80.0200 - val_calc_mre_K: 0.4891\n", "Epoch 746/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 77.3440 - calc_mre_K: 0.4721 - val_loss: 71.9123 - val_calc_mre_K: 0.4395\n", "Epoch 747/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 74.0000 - calc_mre_K: 0.4517 - val_loss: 77.9461 - val_calc_mre_K: 0.4764\n", "Epoch 748/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 74.9836 - calc_mre_K: 0.4577 - val_loss: 79.9926 - val_calc_mre_K: 0.4889\n", "Epoch 749/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 74.6073 - calc_mre_K: 0.4554 - val_loss: 72.3906 - val_calc_mre_K: 0.4424\n", "Epoch 750/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 74.7259 - calc_mre_K: 0.4561 - val_loss: 74.1051 - val_calc_mre_K: 0.4529\n", "Epoch 751/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 74.7908 - calc_mre_K: 0.4565 - val_loss: 73.1957 - val_calc_mre_K: 0.4474\n", "Epoch 752/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 74.5642 - calc_mre_K: 0.4551 - val_loss: 74.3625 - val_calc_mre_K: 0.4545\n", "Epoch 753/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 74.9071 - calc_mre_K: 0.4572 - val_loss: 72.0751 - val_calc_mre_K: 0.4405\n", "Epoch 754/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 74.3306 - calc_mre_K: 0.4537 - val_loss: 74.1870 - val_calc_mre_K: 0.4534\n", "Epoch 755/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 75.1224 - calc_mre_K: 0.4585 - val_loss: 74.7287 - val_calc_mre_K: 0.4567\n", "Epoch 756/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 74.6367 - calc_mre_K: 0.4555 - val_loss: 75.6637 - val_calc_mre_K: 0.4624\n", "Epoch 757/2000\n", "48000/48000 [==============================] - 3s 55us/step - loss: 74.6445 - calc_mre_K: 0.4556 - val_loss: 75.9234 - val_calc_mre_K: 0.4640\n", "Epoch 758/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 74.7333 - calc_mre_K: 0.4561 - val_loss: 75.1347 - val_calc_mre_K: 0.4592\n", "Epoch 759/2000\n", "48000/48000 [==============================] - 2s 49us/step - loss: 74.2251 - calc_mre_K: 0.4530 - val_loss: 79.5329 - val_calc_mre_K: 0.4861\n", "Epoch 760/2000\n", "48000/48000 [==============================] - 2s 49us/step - loss: 74.6202 - calc_mre_K: 0.4554 - val_loss: 76.1007 - val_calc_mre_K: 0.4651\n", "Epoch 761/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 75.0420 - calc_mre_K: 0.4580 - val_loss: 73.4458 - val_calc_mre_K: 0.4489\n", "Epoch 762/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 74.2286 - calc_mre_K: 0.4531 - val_loss: 73.6368 - val_calc_mre_K: 0.4500\n", "Epoch 763/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 74.4116 - calc_mre_K: 0.4542 - val_loss: 85.6614 - val_calc_mre_K: 0.5235\n", "Epoch 764/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 74.1835 - calc_mre_K: 0.4528 - val_loss: 79.3563 - val_calc_mre_K: 0.4850\n", "Epoch 765/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 74.9649 - calc_mre_K: 0.4575 - val_loss: 74.2464 - val_calc_mre_K: 0.4538\n", "Epoch 766/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 74.6912 - calc_mre_K: 0.4559 - val_loss: 79.0476 - val_calc_mre_K: 0.4831\n", "Epoch 767/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 74.2809 - calc_mre_K: 0.4534 - val_loss: 73.2997 - val_calc_mre_K: 0.4480\n", "Epoch 768/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 74.0416 - calc_mre_K: 0.4519 - val_loss: 73.4893 - val_calc_mre_K: 0.4491\n", "Epoch 769/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 74.0864 - calc_mre_K: 0.4522 - val_loss: 73.1112 - val_calc_mre_K: 0.4468\n", "Epoch 770/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 74.4906 - calc_mre_K: 0.4547 - val_loss: 78.8349 - val_calc_mre_K: 0.4818\n", "Epoch 771/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 74.7870 - calc_mre_K: 0.4565 - val_loss: 74.3724 - val_calc_mre_K: 0.4545\n", "Epoch 772/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 74.6441 - calc_mre_K: 0.4556 - val_loss: 74.9538 - val_calc_mre_K: 0.4581\n", "Epoch 773/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 74.0657 - calc_mre_K: 0.4521 - val_loss: 73.0585 - val_calc_mre_K: 0.4465\n", "Epoch 774/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 74.3941 - calc_mre_K: 0.4541 - val_loss: 75.1781 - val_calc_mre_K: 0.4594\n", "Epoch 775/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 75.1627 - calc_mre_K: 0.4588 - val_loss: 73.9757 - val_calc_mre_K: 0.4521\n", "Epoch 776/2000\n", "48000/48000 [==============================] - 2s 49us/step - loss: 73.8627 - calc_mre_K: 0.4508 - val_loss: 71.6793 - val_calc_mre_K: 0.4381\n", "Epoch 777/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 73.4363 - calc_mre_K: 0.4482 - val_loss: 76.6429 - val_calc_mre_K: 0.4684\n", "Epoch 778/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 74.7309 - calc_mre_K: 0.4561 - val_loss: 75.4290 - val_calc_mre_K: 0.4610\n", "Epoch 779/2000\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "48000/48000 [==============================] - 3s 52us/step - loss: 73.7800 - calc_mre_K: 0.4503 - val_loss: 83.0831 - val_calc_mre_K: 0.5078\n", "Epoch 780/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 74.6032 - calc_mre_K: 0.4553 - val_loss: 73.0489 - val_calc_mre_K: 0.4465\n", "Epoch 781/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 74.0856 - calc_mre_K: 0.4522 - val_loss: 77.1541 - val_calc_mre_K: 0.4715\n", "Epoch 782/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 74.2647 - calc_mre_K: 0.4533 - val_loss: 72.9310 - val_calc_mre_K: 0.4457\n", "Epoch 783/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 74.5035 - calc_mre_K: 0.4547 - val_loss: 75.0990 - val_calc_mre_K: 0.4590\n", "Epoch 784/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 74.5927 - calc_mre_K: 0.4553 - val_loss: 71.8580 - val_calc_mre_K: 0.4392\n", "Epoch 785/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 74.1939 - calc_mre_K: 0.4528 - val_loss: 73.8645 - val_calc_mre_K: 0.4514\n", "Epoch 786/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 73.5034 - calc_mre_K: 0.4486 - val_loss: 71.8902 - val_calc_mre_K: 0.4394\n", "Epoch 787/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 74.2841 - calc_mre_K: 0.4534 - val_loss: 74.2730 - val_calc_mre_K: 0.4539\n", "Epoch 788/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 74.0483 - calc_mre_K: 0.4520 - val_loss: 74.9240 - val_calc_mre_K: 0.4579\n", "Epoch 789/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 74.1791 - calc_mre_K: 0.4528 - val_loss: 71.7750 - val_calc_mre_K: 0.4387\n", "Epoch 790/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 74.7028 - calc_mre_K: 0.4559 - val_loss: 78.3300 - val_calc_mre_K: 0.4787\n", "Epoch 791/2000\n", "48000/48000 [==============================] - 3s 56us/step - loss: 74.2225 - calc_mre_K: 0.4530 - val_loss: 73.9722 - val_calc_mre_K: 0.4521\n", "Epoch 792/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 74.0969 - calc_mre_K: 0.4523 - val_loss: 72.9880 - val_calc_mre_K: 0.4461\n", "Epoch 793/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 74.3278 - calc_mre_K: 0.4537 - val_loss: 72.6552 - val_calc_mre_K: 0.4440\n", "Epoch 794/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 74.6591 - calc_mre_K: 0.4557 - val_loss: 76.1540 - val_calc_mre_K: 0.4654\n", "Epoch 795/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 73.7137 - calc_mre_K: 0.4499 - val_loss: 75.0091 - val_calc_mre_K: 0.4584\n", "Epoch 796/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 74.4077 - calc_mre_K: 0.4541 - val_loss: 74.7757 - val_calc_mre_K: 0.4570\n", "Epoch 797/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 74.5289 - calc_mre_K: 0.4549 - val_loss: 72.9924 - val_calc_mre_K: 0.4461\n", "Epoch 798/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 73.6129 - calc_mre_K: 0.4493 - val_loss: 73.0393 - val_calc_mre_K: 0.4464\n", "Epoch 799/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 73.9509 - calc_mre_K: 0.4514 - val_loss: 77.5738 - val_calc_mre_K: 0.4741\n", "Epoch 800/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 73.8114 - calc_mre_K: 0.4505 - val_loss: 75.0443 - val_calc_mre_K: 0.4586\n", "Epoch 801/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 73.4369 - calc_mre_K: 0.4482 - val_loss: 73.0538 - val_calc_mre_K: 0.4465\n", "Epoch 802/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 74.2801 - calc_mre_K: 0.4534 - val_loss: 72.2197 - val_calc_mre_K: 0.4414\n", "Epoch 803/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 74.2997 - calc_mre_K: 0.4535 - val_loss: 76.3351 - val_calc_mre_K: 0.4665\n", "Epoch 804/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 73.9089 - calc_mre_K: 0.4511 - val_loss: 73.8602 - val_calc_mre_K: 0.4514\n", "Epoch 805/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 73.8382 - calc_mre_K: 0.4507 - val_loss: 69.9543 - val_calc_mre_K: 0.4275\n", "Epoch 806/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 74.1991 - calc_mre_K: 0.4529 - val_loss: 77.6906 - val_calc_mre_K: 0.4748\n", "Epoch 807/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 73.8551 - calc_mre_K: 0.4508 - val_loss: 74.9291 - val_calc_mre_K: 0.4579\n", "Epoch 808/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 73.7520 - calc_mre_K: 0.4501 - val_loss: 74.8208 - val_calc_mre_K: 0.4573\n", "Epoch 809/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 73.7322 - calc_mre_K: 0.4500 - val_loss: 73.0671 - val_calc_mre_K: 0.4466\n", "Epoch 810/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 74.0196 - calc_mre_K: 0.4518 - val_loss: 75.1728 - val_calc_mre_K: 0.4594\n", "Epoch 811/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 73.4215 - calc_mre_K: 0.4481 - val_loss: 75.8579 - val_calc_mre_K: 0.4636\n", "Epoch 812/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 73.9050 - calc_mre_K: 0.4511 - val_loss: 71.8264 - val_calc_mre_K: 0.4390\n", "Epoch 813/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 74.0506 - calc_mre_K: 0.4520 - val_loss: 83.6221 - val_calc_mre_K: 0.5111\n", "Epoch 814/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 74.2211 - calc_mre_K: 0.4530 - val_loss: 72.8566 - val_calc_mre_K: 0.4453\n", "Epoch 815/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 73.2942 - calc_mre_K: 0.4474 - val_loss: 72.4141 - val_calc_mre_K: 0.4426\n", "Epoch 816/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 73.9607 - calc_mre_K: 0.4514 - val_loss: 71.5780 - val_calc_mre_K: 0.4375\n", "Epoch 817/2000\n", "48000/48000 [==============================] - 2s 49us/step - loss: 74.2618 - calc_mre_K: 0.4533 - val_loss: 73.0413 - val_calc_mre_K: 0.4464\n", "Epoch 818/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 72.5796 - calc_mre_K: 0.4430 - val_loss: 71.4914 - val_calc_mre_K: 0.4369\n", "Epoch 819/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 74.4357 - calc_mre_K: 0.4543 - val_loss: 73.2314 - val_calc_mre_K: 0.4475\n", "Epoch 820/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 73.9295 - calc_mre_K: 0.4512 - val_loss: 70.8854 - val_calc_mre_K: 0.4332\n", "Epoch 821/2000\n", "48000/48000 [==============================] - 2s 49us/step - loss: 73.4227 - calc_mre_K: 0.4481 - val_loss: 70.9848 - val_calc_mre_K: 0.4338\n", "Epoch 822/2000\n", "48000/48000 [==============================] - 2s 48us/step - loss: 73.2530 - calc_mre_K: 0.4471 - val_loss: 78.9386 - val_calc_mre_K: 0.4824\n", "Epoch 823/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 73.7941 - calc_mre_K: 0.4504 - val_loss: 72.0770 - val_calc_mre_K: 0.4405\n", "Epoch 824/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 73.8983 - calc_mre_K: 0.4510 - val_loss: 72.4022 - val_calc_mre_K: 0.4425\n", "Epoch 825/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 73.0460 - calc_mre_K: 0.4458 - val_loss: 77.2732 - val_calc_mre_K: 0.4723\n", "Epoch 826/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 73.7738 - calc_mre_K: 0.4503 - val_loss: 74.3087 - val_calc_mre_K: 0.4541\n", "Epoch 827/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 74.6716 - calc_mre_K: 0.4558 - val_loss: 70.0566 - val_calc_mre_K: 0.4282\n", "Epoch 828/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 73.0989 - calc_mre_K: 0.4462 - val_loss: 77.5132 - val_calc_mre_K: 0.4737\n", "Epoch 829/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 73.4447 - calc_mre_K: 0.4483 - val_loss: 73.2082 - val_calc_mre_K: 0.4474\n", "Epoch 830/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 74.0976 - calc_mre_K: 0.4523 - val_loss: 74.7760 - val_calc_mre_K: 0.4570\n", "Epoch 831/2000\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "48000/48000 [==============================] - 2s 50us/step - loss: 73.0431 - calc_mre_K: 0.4458 - val_loss: 73.4316 - val_calc_mre_K: 0.4488\n", "Epoch 832/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 73.0619 - calc_mre_K: 0.4459 - val_loss: 83.7474 - val_calc_mre_K: 0.5118\n", "Epoch 833/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 74.0158 - calc_mre_K: 0.4518 - val_loss: 80.3359 - val_calc_mre_K: 0.4910\n", "Epoch 834/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 74.5575 - calc_mre_K: 0.4551 - val_loss: 70.4314 - val_calc_mre_K: 0.4305\n", "Epoch 835/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 72.8543 - calc_mre_K: 0.4447 - val_loss: 77.6904 - val_calc_mre_K: 0.4748\n", "Epoch 836/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 73.0979 - calc_mre_K: 0.4462 - val_loss: 75.1460 - val_calc_mre_K: 0.4593\n", "Epoch 837/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 73.7399 - calc_mre_K: 0.4501 - val_loss: 70.0308 - val_calc_mre_K: 0.4280\n", "Epoch 838/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 72.9468 - calc_mre_K: 0.4452 - val_loss: 72.6761 - val_calc_mre_K: 0.4442\n", "Epoch 839/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 73.4140 - calc_mre_K: 0.4481 - val_loss: 72.6749 - val_calc_mre_K: 0.4442\n", "Epoch 840/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 73.2349 - calc_mre_K: 0.4470 - val_loss: 74.2645 - val_calc_mre_K: 0.4539\n", "Epoch 841/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 74.3029 - calc_mre_K: 0.4535 - val_loss: 75.8798 - val_calc_mre_K: 0.4638\n", "Epoch 842/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 73.4507 - calc_mre_K: 0.4483 - val_loss: 72.5665 - val_calc_mre_K: 0.4435\n", "Epoch 843/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 73.2979 - calc_mre_K: 0.4474 - val_loss: 72.2444 - val_calc_mre_K: 0.4415\n", "Epoch 844/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 73.2323 - calc_mre_K: 0.4470 - val_loss: 71.8140 - val_calc_mre_K: 0.4389\n", "Epoch 845/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 72.8278 - calc_mre_K: 0.4445 - val_loss: 77.1932 - val_calc_mre_K: 0.4718\n", "Epoch 846/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 73.8738 - calc_mre_K: 0.4509 - val_loss: 74.0537 - val_calc_mre_K: 0.4526\n", "Epoch 847/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 73.2000 - calc_mre_K: 0.4468 - val_loss: 69.6751 - val_calc_mre_K: 0.4258\n", "Epoch 848/2000\n", "48000/48000 [==============================] - 2s 47us/step - loss: 73.6094 - calc_mre_K: 0.4493 - val_loss: 73.8997 - val_calc_mre_K: 0.4516\n", "Epoch 849/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 72.9308 - calc_mre_K: 0.4451 - val_loss: 73.7549 - val_calc_mre_K: 0.4508\n", "Epoch 850/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 73.0960 - calc_mre_K: 0.4461 - val_loss: 71.9407 - val_calc_mre_K: 0.4397\n", "Epoch 851/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 73.6931 - calc_mre_K: 0.4498 - val_loss: 72.8921 - val_calc_mre_K: 0.4455\n", "Epoch 852/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 73.6994 - calc_mre_K: 0.4498 - val_loss: 69.7524 - val_calc_mre_K: 0.4263\n", "Epoch 853/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 73.1155 - calc_mre_K: 0.4463 - val_loss: 71.8051 - val_calc_mre_K: 0.4388\n", "Epoch 854/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 72.9046 - calc_mre_K: 0.4450 - val_loss: 75.8857 - val_calc_mre_K: 0.4638\n", "Epoch 855/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 73.1575 - calc_mre_K: 0.4465 - val_loss: 71.7284 - val_calc_mre_K: 0.4384\n", "Epoch 856/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 72.7116 - calc_mre_K: 0.4438 - val_loss: 70.2751 - val_calc_mre_K: 0.4295\n", "Epoch 857/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 73.2542 - calc_mre_K: 0.4471 - val_loss: 70.1802 - val_calc_mre_K: 0.4289\n", "Epoch 858/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 73.2882 - calc_mre_K: 0.4473 - val_loss: 75.1835 - val_calc_mre_K: 0.4595\n", "Epoch 859/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 73.6406 - calc_mre_K: 0.4495 - val_loss: 72.4577 - val_calc_mre_K: 0.4428\n", "Epoch 860/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 72.8781 - calc_mre_K: 0.4448 - val_loss: 71.5541 - val_calc_mre_K: 0.4373\n", "Epoch 861/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 72.6925 - calc_mre_K: 0.4437 - val_loss: 73.8626 - val_calc_mre_K: 0.4514\n", "Epoch 862/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 72.8597 - calc_mre_K: 0.4447 - val_loss: 70.7630 - val_calc_mre_K: 0.4325\n", "Epoch 863/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 73.1689 - calc_mre_K: 0.4466 - val_loss: 73.0344 - val_calc_mre_K: 0.4463\n", "Epoch 864/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 73.3326 - calc_mre_K: 0.4476 - val_loss: 73.4474 - val_calc_mre_K: 0.4489\n", "Epoch 865/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 72.6202 - calc_mre_K: 0.4432 - val_loss: 70.4354 - val_calc_mre_K: 0.4305\n", "Epoch 866/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 72.9237 - calc_mre_K: 0.4451 - val_loss: 71.1424 - val_calc_mre_K: 0.4348\n", "Epoch 867/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 73.1100 - calc_mre_K: 0.4462 - val_loss: 72.6170 - val_calc_mre_K: 0.4438\n", "Epoch 868/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 72.9277 - calc_mre_K: 0.4451 - val_loss: 71.1453 - val_calc_mre_K: 0.4348\n", "Epoch 869/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 72.8218 - calc_mre_K: 0.4445 - val_loss: 79.9064 - val_calc_mre_K: 0.4884\n", "Epoch 870/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 73.2126 - calc_mre_K: 0.4469 - val_loss: 72.0436 - val_calc_mre_K: 0.4403\n", "Epoch 871/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 73.1982 - calc_mre_K: 0.4468 - val_loss: 71.8420 - val_calc_mre_K: 0.4391\n", "Epoch 872/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 73.1332 - calc_mre_K: 0.4464 - val_loss: 72.1057 - val_calc_mre_K: 0.4407\n", "Epoch 873/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 72.5791 - calc_mre_K: 0.4430 - val_loss: 71.7445 - val_calc_mre_K: 0.4385\n", "Epoch 874/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 72.6875 - calc_mre_K: 0.4436 - val_loss: 73.9389 - val_calc_mre_K: 0.4519\n", "Epoch 875/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 72.4498 - calc_mre_K: 0.4422 - val_loss: 74.1482 - val_calc_mre_K: 0.4532\n", "Epoch 876/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 72.8335 - calc_mre_K: 0.4445 - val_loss: 70.0421 - val_calc_mre_K: 0.4281\n", "Epoch 877/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 73.5146 - calc_mre_K: 0.4487 - val_loss: 72.9177 - val_calc_mre_K: 0.4456\n", "Epoch 878/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 71.9098 - calc_mre_K: 0.4389 - val_loss: 70.4277 - val_calc_mre_K: 0.4304\n", "Epoch 879/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 72.9239 - calc_mre_K: 0.4451 - val_loss: 75.1720 - val_calc_mre_K: 0.4594\n", "Epoch 880/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 72.9498 - calc_mre_K: 0.4453 - val_loss: 73.1351 - val_calc_mre_K: 0.4470\n", "Epoch 881/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 71.9448 - calc_mre_K: 0.4391 - val_loss: 76.1734 - val_calc_mre_K: 0.4655\n", "Epoch 882/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 72.7901 - calc_mre_K: 0.4443 - val_loss: 72.8919 - val_calc_mre_K: 0.4455\n", "Epoch 883/2000\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "48000/48000 [==============================] - 3s 53us/step - loss: 72.6625 - calc_mre_K: 0.4435 - val_loss: 72.5712 - val_calc_mre_K: 0.4435\n", "Epoch 884/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 73.0796 - calc_mre_K: 0.4460 - val_loss: 74.1943 - val_calc_mre_K: 0.4534\n", "Epoch 885/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 72.8259 - calc_mre_K: 0.4445 - val_loss: 72.5252 - val_calc_mre_K: 0.4432\n", "Epoch 886/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 72.2531 - calc_mre_K: 0.4410 - val_loss: 76.3423 - val_calc_mre_K: 0.4666\n", "Epoch 887/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 72.4299 - calc_mre_K: 0.4421 - val_loss: 73.4423 - val_calc_mre_K: 0.4488\n", "Epoch 888/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 72.5955 - calc_mre_K: 0.4431 - val_loss: 71.3515 - val_calc_mre_K: 0.4361\n", "Epoch 889/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 72.9617 - calc_mre_K: 0.4453 - val_loss: 75.9629 - val_calc_mre_K: 0.4643\n", "Epoch 890/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 72.6754 - calc_mre_K: 0.4436 - val_loss: 85.2748 - val_calc_mre_K: 0.5212\n", "Epoch 891/2000\n", "48000/48000 [==============================] - 2s 49us/step - loss: 73.0085 - calc_mre_K: 0.4456 - val_loss: 70.8243 - val_calc_mre_K: 0.4329\n", "Epoch 892/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 71.9620 - calc_mre_K: 0.4392 - val_loss: 72.1269 - val_calc_mre_K: 0.4408\n", "Epoch 893/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 72.0540 - calc_mre_K: 0.4398 - val_loss: 68.8221 - val_calc_mre_K: 0.4206\n", "Epoch 894/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 73.1036 - calc_mre_K: 0.4462 - val_loss: 70.2157 - val_calc_mre_K: 0.4291\n", "Epoch 895/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 72.9770 - calc_mre_K: 0.4454 - val_loss: 71.9663 - val_calc_mre_K: 0.4398\n", "Epoch 896/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 71.9662 - calc_mre_K: 0.4392 - val_loss: 70.4448 - val_calc_mre_K: 0.4305\n", "Epoch 897/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 72.4647 - calc_mre_K: 0.4423 - val_loss: 72.1677 - val_calc_mre_K: 0.4411\n", "Epoch 898/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 71.8626 - calc_mre_K: 0.4386 - val_loss: 70.3476 - val_calc_mre_K: 0.4299\n", "Epoch 899/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 72.1744 - calc_mre_K: 0.4405 - val_loss: 79.0658 - val_calc_mre_K: 0.4832\n", "Epoch 900/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 73.0322 - calc_mre_K: 0.4458 - val_loss: 69.3548 - val_calc_mre_K: 0.4239\n", "Epoch 901/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 72.2562 - calc_mre_K: 0.4410 - val_loss: 70.9149 - val_calc_mre_K: 0.4334\n", "Epoch 902/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 72.3430 - calc_mre_K: 0.4415 - val_loss: 71.3529 - val_calc_mre_K: 0.4361\n", "Epoch 903/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 72.0164 - calc_mre_K: 0.4396 - val_loss: 70.1190 - val_calc_mre_K: 0.4285\n", "Epoch 904/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 72.5536 - calc_mre_K: 0.4428 - val_loss: 72.4095 - val_calc_mre_K: 0.4425\n", "Epoch 905/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 71.9868 - calc_mre_K: 0.4394 - val_loss: 69.8664 - val_calc_mre_K: 0.4270\n", "Epoch 906/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 72.1781 - calc_mre_K: 0.4405 - val_loss: 74.0787 - val_calc_mre_K: 0.4527\n", "Epoch 907/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 72.3159 - calc_mre_K: 0.4414 - val_loss: 70.4471 - val_calc_mre_K: 0.4305\n", "Epoch 908/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 71.8884 - calc_mre_K: 0.4388 - val_loss: 73.8511 - val_calc_mre_K: 0.4513\n", "Epoch 909/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 72.4439 - calc_mre_K: 0.4422 - val_loss: 72.5625 - val_calc_mre_K: 0.4435\n", "Epoch 910/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 72.1621 - calc_mre_K: 0.4404 - val_loss: 72.3865 - val_calc_mre_K: 0.4424\n", "Epoch 911/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 71.9880 - calc_mre_K: 0.4394 - val_loss: 70.8922 - val_calc_mre_K: 0.4333\n", "Epoch 912/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 71.5619 - calc_mre_K: 0.4368 - val_loss: 73.0477 - val_calc_mre_K: 0.4464\n", "Epoch 913/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 71.6234 - calc_mre_K: 0.4372 - val_loss: 69.3621 - val_calc_mre_K: 0.4239\n", "Epoch 914/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 73.3054 - calc_mre_K: 0.4474 - val_loss: 70.8996 - val_calc_mre_K: 0.4333\n", "Epoch 915/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 72.2917 - calc_mre_K: 0.4412 - val_loss: 69.8175 - val_calc_mre_K: 0.4267\n", "Epoch 916/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 71.5330 - calc_mre_K: 0.4366 - val_loss: 72.0239 - val_calc_mre_K: 0.4402\n", "Epoch 917/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 71.9143 - calc_mre_K: 0.4389 - val_loss: 71.9672 - val_calc_mre_K: 0.4398\n", "Epoch 918/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 72.2261 - calc_mre_K: 0.4408 - val_loss: 71.9405 - val_calc_mre_K: 0.4397\n", "Epoch 919/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 72.1139 - calc_mre_K: 0.4401 - val_loss: 73.5077 - val_calc_mre_K: 0.4492\n", "Epoch 920/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 72.1678 - calc_mre_K: 0.4405 - val_loss: 71.3662 - val_calc_mre_K: 0.4362\n", "Epoch 921/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 72.3274 - calc_mre_K: 0.4415 - val_loss: 70.0718 - val_calc_mre_K: 0.4282\n", "Epoch 922/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 71.8315 - calc_mre_K: 0.4384 - val_loss: 71.9886 - val_calc_mre_K: 0.4400\n", "Epoch 923/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 72.4446 - calc_mre_K: 0.4422 - val_loss: 67.4889 - val_calc_mre_K: 0.4125\n", "Epoch 924/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 71.6839 - calc_mre_K: 0.4375 - val_loss: 83.6090 - val_calc_mre_K: 0.5110\n", "Epoch 925/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 72.2264 - calc_mre_K: 0.4408 - val_loss: 73.3333 - val_calc_mre_K: 0.4482\n", "Epoch 926/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 72.0682 - calc_mre_K: 0.4399 - val_loss: 71.3794 - val_calc_mre_K: 0.4362\n", "Epoch 927/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 71.5717 - calc_mre_K: 0.4368 - val_loss: 73.1700 - val_calc_mre_K: 0.4472\n", "Epoch 928/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 72.2864 - calc_mre_K: 0.4412 - val_loss: 72.0560 - val_calc_mre_K: 0.4404\n", "Epoch 929/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 71.0249 - calc_mre_K: 0.4335 - val_loss: 71.9705 - val_calc_mre_K: 0.4399\n", "Epoch 930/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 72.4680 - calc_mre_K: 0.4423 - val_loss: 70.2004 - val_calc_mre_K: 0.4290\n", "Epoch 931/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 71.9499 - calc_mre_K: 0.4391 - val_loss: 74.5807 - val_calc_mre_K: 0.4558\n", "Epoch 932/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 72.1783 - calc_mre_K: 0.4405 - val_loss: 74.9445 - val_calc_mre_K: 0.4580\n", "Epoch 933/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 71.7595 - calc_mre_K: 0.4380 - val_loss: 72.0812 - val_calc_mre_K: 0.4405\n", "Epoch 934/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 72.0420 - calc_mre_K: 0.4397 - val_loss: 70.4034 - val_calc_mre_K: 0.4303\n", "Epoch 935/2000\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "48000/48000 [==============================] - 3s 52us/step - loss: 71.8713 - calc_mre_K: 0.4387 - val_loss: 72.3075 - val_calc_mre_K: 0.4419\n", "Epoch 936/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 71.5063 - calc_mre_K: 0.4364 - val_loss: 71.1165 - val_calc_mre_K: 0.4346\n", "Epoch 937/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 72.0541 - calc_mre_K: 0.4398 - val_loss: 72.7008 - val_calc_mre_K: 0.4443\n", "Epoch 938/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 71.6675 - calc_mre_K: 0.4374 - val_loss: 71.4258 - val_calc_mre_K: 0.4365\n", "Epoch 939/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 71.4339 - calc_mre_K: 0.4360 - val_loss: 74.9233 - val_calc_mre_K: 0.4579\n", "Epoch 940/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 72.4590 - calc_mre_K: 0.4423 - val_loss: 74.1253 - val_calc_mre_K: 0.4530\n", "Epoch 941/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 71.3304 - calc_mre_K: 0.4354 - val_loss: 69.6579 - val_calc_mre_K: 0.4257\n", "Epoch 942/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 71.7210 - calc_mre_K: 0.4378 - val_loss: 69.5940 - val_calc_mre_K: 0.4253\n", "Epoch 943/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 72.0737 - calc_mre_K: 0.4399 - val_loss: 71.2022 - val_calc_mre_K: 0.4352\n", "Epoch 944/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 71.5817 - calc_mre_K: 0.4369 - val_loss: 71.0556 - val_calc_mre_K: 0.4343\n", "Epoch 945/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 71.5547 - calc_mre_K: 0.4367 - val_loss: 69.4996 - val_calc_mre_K: 0.4247\n", "Epoch 946/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 71.0234 - calc_mre_K: 0.4335 - val_loss: 69.5443 - val_calc_mre_K: 0.4250\n", "Epoch 947/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 72.1470 - calc_mre_K: 0.4404 - val_loss: 69.8591 - val_calc_mre_K: 0.4269\n", "Epoch 948/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 71.2693 - calc_mre_K: 0.4350 - val_loss: 71.1697 - val_calc_mre_K: 0.4350\n", "Epoch 949/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 71.6938 - calc_mre_K: 0.4376 - val_loss: 70.8749 - val_calc_mre_K: 0.4332\n", "Epoch 950/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 70.7961 - calc_mre_K: 0.4321 - val_loss: 70.5474 - val_calc_mre_K: 0.4311\n", "Epoch 951/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 72.2471 - calc_mre_K: 0.4410 - val_loss: 75.2263 - val_calc_mre_K: 0.4597\n", "Epoch 952/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 71.8752 - calc_mre_K: 0.4387 - val_loss: 68.2813 - val_calc_mre_K: 0.4173\n", "Epoch 953/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 70.9406 - calc_mre_K: 0.4330 - val_loss: 71.0173 - val_calc_mre_K: 0.4340\n", "Epoch 954/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 70.9086 - calc_mre_K: 0.4328 - val_loss: 71.0774 - val_calc_mre_K: 0.4344\n", "Epoch 955/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 71.3510 - calc_mre_K: 0.4355 - val_loss: 74.1342 - val_calc_mre_K: 0.4531\n", "Epoch 956/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 71.7490 - calc_mre_K: 0.4379 - val_loss: 75.6656 - val_calc_mre_K: 0.4624\n", "Epoch 957/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 71.1012 - calc_mre_K: 0.4340 - val_loss: 70.4997 - val_calc_mre_K: 0.4309\n", "Epoch 958/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 71.7428 - calc_mre_K: 0.4379 - val_loss: 73.6446 - val_calc_mre_K: 0.4501\n", "Epoch 959/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 71.5524 - calc_mre_K: 0.4367 - val_loss: 67.6303 - val_calc_mre_K: 0.4133\n", "Epoch 960/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 71.2123 - calc_mre_K: 0.4346 - val_loss: 68.9148 - val_calc_mre_K: 0.4212\n", "Epoch 961/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 71.0790 - calc_mre_K: 0.4338 - val_loss: 73.0634 - val_calc_mre_K: 0.4465\n", "Epoch 962/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 71.4993 - calc_mre_K: 0.4364 - val_loss: 70.1159 - val_calc_mre_K: 0.4285\n", "Epoch 963/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 70.8454 - calc_mre_K: 0.4324 - val_loss: 69.2248 - val_calc_mre_K: 0.4231\n", "Epoch 964/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 71.9449 - calc_mre_K: 0.4391 - val_loss: 70.3683 - val_calc_mre_K: 0.4301\n", "Epoch 965/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 70.8856 - calc_mre_K: 0.4327 - val_loss: 68.8444 - val_calc_mre_K: 0.4207\n", "Epoch 966/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 70.4476 - calc_mre_K: 0.4300 - val_loss: 69.5032 - val_calc_mre_K: 0.4248\n", "Epoch 967/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 71.3859 - calc_mre_K: 0.4357 - val_loss: 69.0565 - val_calc_mre_K: 0.4220\n", "Epoch 968/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 71.4308 - calc_mre_K: 0.4360 - val_loss: 69.4411 - val_calc_mre_K: 0.4244\n", "Epoch 969/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 71.5515 - calc_mre_K: 0.4367 - val_loss: 69.7290 - val_calc_mre_K: 0.4262\n", "Epoch 970/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 70.5988 - calc_mre_K: 0.4309 - val_loss: 71.6331 - val_calc_mre_K: 0.4378\n", "Epoch 971/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 71.4212 - calc_mre_K: 0.4359 - val_loss: 68.2783 - val_calc_mre_K: 0.4173\n", "Epoch 972/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 70.8157 - calc_mre_K: 0.4322 - val_loss: 69.5347 - val_calc_mre_K: 0.4250\n", "Epoch 973/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 70.8060 - calc_mre_K: 0.4322 - val_loss: 69.3833 - val_calc_mre_K: 0.4240\n", "Epoch 974/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 71.5117 - calc_mre_K: 0.4365 - val_loss: 74.3812 - val_calc_mre_K: 0.4546\n", "Epoch 975/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 70.9862 - calc_mre_K: 0.4333 - val_loss: 72.0374 - val_calc_mre_K: 0.4403\n", "Epoch 976/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 71.0124 - calc_mre_K: 0.4334 - val_loss: 72.6798 - val_calc_mre_K: 0.4442\n", "Epoch 977/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 70.6009 - calc_mre_K: 0.4309 - val_loss: 71.1741 - val_calc_mre_K: 0.4350\n", "Epoch 978/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 72.6719 - calc_mre_K: 0.4436 - val_loss: 70.4232 - val_calc_mre_K: 0.4304\n", "Epoch 979/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 71.0686 - calc_mre_K: 0.4338 - val_loss: 81.6629 - val_calc_mre_K: 0.4991\n", "Epoch 980/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 71.0265 - calc_mre_K: 0.4335 - val_loss: 69.3473 - val_calc_mre_K: 0.4238\n", "Epoch 981/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 71.1099 - calc_mre_K: 0.4340 - val_loss: 72.8353 - val_calc_mre_K: 0.4451\n", "Epoch 982/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 71.2195 - calc_mre_K: 0.4347 - val_loss: 68.5832 - val_calc_mre_K: 0.4192\n", "Epoch 983/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 71.4536 - calc_mre_K: 0.4361 - val_loss: 70.2968 - val_calc_mre_K: 0.4296\n", "Epoch 984/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 70.1857 - calc_mre_K: 0.4284 - val_loss: 70.9116 - val_calc_mre_K: 0.4334\n", "Epoch 985/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 70.7512 - calc_mre_K: 0.4318 - val_loss: 72.8994 - val_calc_mre_K: 0.4455\n", "Epoch 986/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 70.5823 - calc_mre_K: 0.4308 - val_loss: 68.9610 - val_calc_mre_K: 0.4215\n", "Epoch 987/2000\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "48000/48000 [==============================] - 3s 52us/step - loss: 70.7230 - calc_mre_K: 0.4317 - val_loss: 70.2102 - val_calc_mre_K: 0.4291\n", "Epoch 988/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 70.5966 - calc_mre_K: 0.4309 - val_loss: 70.5693 - val_calc_mre_K: 0.4313\n", "Epoch 989/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 71.3090 - calc_mre_K: 0.4352 - val_loss: 67.6903 - val_calc_mre_K: 0.4137\n", "Epoch 990/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 71.0700 - calc_mre_K: 0.4338 - val_loss: 69.7356 - val_calc_mre_K: 0.4262\n", "Epoch 991/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 70.2682 - calc_mre_K: 0.4289 - val_loss: 68.1112 - val_calc_mre_K: 0.4163\n", "Epoch 992/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 71.3900 - calc_mre_K: 0.4357 - val_loss: 69.5300 - val_calc_mre_K: 0.4249\n", "Epoch 993/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 69.8333 - calc_mre_K: 0.4262 - val_loss: 68.1558 - val_calc_mre_K: 0.4165\n", "Epoch 994/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 71.9087 - calc_mre_K: 0.4389 - val_loss: 72.8606 - val_calc_mre_K: 0.4453\n", "Epoch 995/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 70.9336 - calc_mre_K: 0.4329 - val_loss: 70.9065 - val_calc_mre_K: 0.4333\n", "Epoch 996/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 70.6740 - calc_mre_K: 0.4314 - val_loss: 72.4510 - val_calc_mre_K: 0.4428\n", "Epoch 997/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 70.7245 - calc_mre_K: 0.4317 - val_loss: 71.1159 - val_calc_mre_K: 0.4346\n", "Epoch 998/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 70.4275 - calc_mre_K: 0.4299 - val_loss: 69.6573 - val_calc_mre_K: 0.4257\n", "Epoch 999/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 70.3441 - calc_mre_K: 0.4293 - val_loss: 74.0621 - val_calc_mre_K: 0.4526\n", "Epoch 1000/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 71.0901 - calc_mre_K: 0.4339 - val_loss: 69.9184 - val_calc_mre_K: 0.4273\n", "Epoch 1001/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 70.8879 - calc_mre_K: 0.4327 - val_loss: 70.4233 - val_calc_mre_K: 0.4304\n", "Epoch 1002/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 70.4844 - calc_mre_K: 0.4302 - val_loss: 74.8889 - val_calc_mre_K: 0.4577\n", "Epoch 1003/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 71.4583 - calc_mre_K: 0.4361 - val_loss: 69.1719 - val_calc_mre_K: 0.4227\n", "Epoch 1004/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 70.3164 - calc_mre_K: 0.4292 - val_loss: 72.2739 - val_calc_mre_K: 0.4417\n", "Epoch 1005/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 70.0930 - calc_mre_K: 0.4278 - val_loss: 71.8983 - val_calc_mre_K: 0.4394\n", "Epoch 1006/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 70.4035 - calc_mre_K: 0.4297 - val_loss: 75.3601 - val_calc_mre_K: 0.4606\n", "Epoch 1007/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 70.8734 - calc_mre_K: 0.4326 - val_loss: 75.1343 - val_calc_mre_K: 0.4592\n", "Epoch 1008/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 70.5005 - calc_mre_K: 0.4303 - val_loss: 67.9517 - val_calc_mre_K: 0.4153\n", "Epoch 1009/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 70.4530 - calc_mre_K: 0.4300 - val_loss: 70.7565 - val_calc_mre_K: 0.4324\n", "Epoch 1010/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 70.5984 - calc_mre_K: 0.4309 - val_loss: 68.5135 - val_calc_mre_K: 0.4187\n", "Epoch 1011/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 70.4860 - calc_mre_K: 0.4302 - val_loss: 69.7735 - val_calc_mre_K: 0.4264\n", "Epoch 1012/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 70.4194 - calc_mre_K: 0.4298 - val_loss: 71.7467 - val_calc_mre_K: 0.4385\n", "Epoch 1013/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 70.2600 - calc_mre_K: 0.4288 - val_loss: 70.7556 - val_calc_mre_K: 0.4324\n", "Epoch 1014/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 70.6919 - calc_mre_K: 0.4315 - val_loss: 69.7708 - val_calc_mre_K: 0.4264\n", "Epoch 1015/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 70.4776 - calc_mre_K: 0.4302 - val_loss: 70.7394 - val_calc_mre_K: 0.4323\n", "Epoch 1016/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 70.7276 - calc_mre_K: 0.4317 - val_loss: 73.8429 - val_calc_mre_K: 0.4513\n", "Epoch 1017/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 70.3046 - calc_mre_K: 0.4291 - val_loss: 69.3777 - val_calc_mre_K: 0.4240\n", "Epoch 1018/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 69.9870 - calc_mre_K: 0.4272 - val_loss: 68.9234 - val_calc_mre_K: 0.4212\n", "Epoch 1019/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 70.5828 - calc_mre_K: 0.4308 - val_loss: 69.6931 - val_calc_mre_K: 0.4259\n", "Epoch 1020/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 70.1028 - calc_mre_K: 0.4279 - val_loss: 69.6611 - val_calc_mre_K: 0.4257\n", "Epoch 1021/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 70.7045 - calc_mre_K: 0.4315 - val_loss: 73.2916 - val_calc_mre_K: 0.4479\n", "Epoch 1022/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 70.2601 - calc_mre_K: 0.4288 - val_loss: 77.8687 - val_calc_mre_K: 0.4759\n", "Epoch 1023/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 70.2667 - calc_mre_K: 0.4289 - val_loss: 68.5669 - val_calc_mre_K: 0.4190\n", "Epoch 1024/2000\n", "48000/48000 [==============================] - 3s 55us/step - loss: 69.9753 - calc_mre_K: 0.4271 - val_loss: 67.3590 - val_calc_mre_K: 0.4117\n", "Epoch 1025/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 70.8640 - calc_mre_K: 0.4325 - val_loss: 69.7374 - val_calc_mre_K: 0.4262\n", "Epoch 1026/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 70.5195 - calc_mre_K: 0.4304 - val_loss: 69.6832 - val_calc_mre_K: 0.4259\n", "Epoch 1027/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 70.1142 - calc_mre_K: 0.4279 - val_loss: 69.7113 - val_calc_mre_K: 0.4260\n", "Epoch 1028/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 70.6219 - calc_mre_K: 0.4310 - val_loss: 75.4117 - val_calc_mre_K: 0.4609\n", "Epoch 1029/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 70.4924 - calc_mre_K: 0.4303 - val_loss: 67.9497 - val_calc_mre_K: 0.4153\n", "Epoch 1030/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 70.7311 - calc_mre_K: 0.4317 - val_loss: 68.1387 - val_calc_mre_K: 0.4164\n", "Epoch 1031/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 70.0355 - calc_mre_K: 0.4275 - val_loss: 70.5484 - val_calc_mre_K: 0.4312\n", "Epoch 1032/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 70.3172 - calc_mre_K: 0.4292 - val_loss: 70.1160 - val_calc_mre_K: 0.4285\n", "Epoch 1033/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 70.8039 - calc_mre_K: 0.4322 - val_loss: 69.6534 - val_calc_mre_K: 0.4257\n", "Epoch 1034/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 69.8505 - calc_mre_K: 0.4263 - val_loss: 67.1216 - val_calc_mre_K: 0.4102\n", "Epoch 1035/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 70.0575 - calc_mre_K: 0.4276 - val_loss: 73.8971 - val_calc_mre_K: 0.4516\n", "Epoch 1036/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 71.1920 - calc_mre_K: 0.4345 - val_loss: 80.6219 - val_calc_mre_K: 0.4927\n", "Epoch 1037/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 70.6492 - calc_mre_K: 0.4312 - val_loss: 75.2320 - val_calc_mre_K: 0.4598\n", "Epoch 1038/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 69.7936 - calc_mre_K: 0.4260 - val_loss: 68.4783 - val_calc_mre_K: 0.4185\n", "Epoch 1039/2000\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "48000/48000 [==============================] - 2s 52us/step - loss: 69.8232 - calc_mre_K: 0.4262 - val_loss: 68.9723 - val_calc_mre_K: 0.4215\n", "Epoch 1040/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 70.5563 - calc_mre_K: 0.4306 - val_loss: 71.2723 - val_calc_mre_K: 0.4356\n", "Epoch 1041/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 69.5393 - calc_mre_K: 0.4244 - val_loss: 69.3127 - val_calc_mre_K: 0.4236\n", "Epoch 1042/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 70.4038 - calc_mre_K: 0.4297 - val_loss: 73.4934 - val_calc_mre_K: 0.4492\n", "Epoch 1043/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 69.9778 - calc_mre_K: 0.4271 - val_loss: 73.0066 - val_calc_mre_K: 0.4462\n", "Epoch 1044/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 70.3670 - calc_mre_K: 0.4295 - val_loss: 69.4487 - val_calc_mre_K: 0.4244\n", "Epoch 1045/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 70.4220 - calc_mre_K: 0.4298 - val_loss: 67.8402 - val_calc_mre_K: 0.4146\n", "Epoch 1046/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 70.2254 - calc_mre_K: 0.4286 - val_loss: 67.7767 - val_calc_mre_K: 0.4142\n", "Epoch 1047/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 70.0297 - calc_mre_K: 0.4274 - val_loss: 70.3601 - val_calc_mre_K: 0.4300\n", "Epoch 1048/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 69.6926 - calc_mre_K: 0.4254 - val_loss: 69.8745 - val_calc_mre_K: 0.4270\n", "Epoch 1049/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 70.8161 - calc_mre_K: 0.4322 - val_loss: 71.8206 - val_calc_mre_K: 0.4389\n", "Epoch 1050/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 70.0508 - calc_mre_K: 0.4276 - val_loss: 71.9298 - val_calc_mre_K: 0.4396\n", "Epoch 1051/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 69.4402 - calc_mre_K: 0.4238 - val_loss: 71.7751 - val_calc_mre_K: 0.4386\n", "Epoch 1052/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 69.8835 - calc_mre_K: 0.4265 - val_loss: 69.1638 - val_calc_mre_K: 0.4227\n", "Epoch 1053/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 70.6915 - calc_mre_K: 0.4315 - val_loss: 69.6459 - val_calc_mre_K: 0.4256\n", "Epoch 1054/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 70.2897 - calc_mre_K: 0.4290 - val_loss: 70.6561 - val_calc_mre_K: 0.4318\n", "Epoch 1055/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 69.1743 - calc_mre_K: 0.4222 - val_loss: 75.8424 - val_calc_mre_K: 0.4635\n", "Epoch 1056/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 70.8272 - calc_mre_K: 0.4323 - val_loss: 73.8736 - val_calc_mre_K: 0.4515\n", "Epoch 1057/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 70.1674 - calc_mre_K: 0.4283 - val_loss: 68.8439 - val_calc_mre_K: 0.4207\n", "Epoch 1058/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 69.9872 - calc_mre_K: 0.4272 - val_loss: 68.4235 - val_calc_mre_K: 0.4182\n", "Epoch 1059/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 69.8071 - calc_mre_K: 0.4261 - val_loss: 69.1859 - val_calc_mre_K: 0.4228\n", "Epoch 1060/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 70.2497 - calc_mre_K: 0.4288 - val_loss: 69.9561 - val_calc_mre_K: 0.4275\n", "Epoch 1061/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 69.5751 - calc_mre_K: 0.4247 - val_loss: 68.1809 - val_calc_mre_K: 0.4167\n", "Epoch 1062/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 70.3733 - calc_mre_K: 0.4295 - val_loss: 68.6720 - val_calc_mre_K: 0.4197\n", "Epoch 1063/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 69.4654 - calc_mre_K: 0.4240 - val_loss: 71.1043 - val_calc_mre_K: 0.4345\n", "Epoch 1064/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 70.2949 - calc_mre_K: 0.4290 - val_loss: 68.8428 - val_calc_mre_K: 0.4207\n", "Epoch 1065/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 69.3362 - calc_mre_K: 0.4232 - val_loss: 70.6620 - val_calc_mre_K: 0.4318\n", "Epoch 1066/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 69.7457 - calc_mre_K: 0.4257 - val_loss: 71.2523 - val_calc_mre_K: 0.4355\n", "Epoch 1067/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 69.4479 - calc_mre_K: 0.4239 - val_loss: 69.0941 - val_calc_mre_K: 0.4223\n", "Epoch 1068/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 69.7487 - calc_mre_K: 0.4257 - val_loss: 67.9939 - val_calc_mre_K: 0.4155\n", "Epoch 1069/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 69.9541 - calc_mre_K: 0.4270 - val_loss: 71.3046 - val_calc_mre_K: 0.4358\n", "Epoch 1070/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 69.7865 - calc_mre_K: 0.4259 - val_loss: 67.7306 - val_calc_mre_K: 0.4139\n", "Epoch 1071/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 69.4748 - calc_mre_K: 0.4240 - val_loss: 67.6739 - val_calc_mre_K: 0.4136\n", "Epoch 1072/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 69.9463 - calc_mre_K: 0.4269 - val_loss: 68.6621 - val_calc_mre_K: 0.4196\n", "Epoch 1073/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 70.0543 - calc_mre_K: 0.4276 - val_loss: 66.7885 - val_calc_mre_K: 0.4082\n", "Epoch 1074/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 69.8140 - calc_mre_K: 0.4261 - val_loss: 66.6415 - val_calc_mre_K: 0.4073\n", "Epoch 1075/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 70.0506 - calc_mre_K: 0.4276 - val_loss: 72.0959 - val_calc_mre_K: 0.4406\n", "Epoch 1076/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 69.6602 - calc_mre_K: 0.4252 - val_loss: 68.9369 - val_calc_mre_K: 0.4213\n", "Epoch 1077/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 69.7505 - calc_mre_K: 0.4257 - val_loss: 70.0057 - val_calc_mre_K: 0.4278\n", "Epoch 1078/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 70.6741 - calc_mre_K: 0.4314 - val_loss: 68.7793 - val_calc_mre_K: 0.4203\n", "Epoch 1079/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 69.2488 - calc_mre_K: 0.4227 - val_loss: 68.8233 - val_calc_mre_K: 0.4206\n", "Epoch 1080/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 69.6473 - calc_mre_K: 0.4251 - val_loss: 68.2664 - val_calc_mre_K: 0.4172\n", "Epoch 1081/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 69.1108 - calc_mre_K: 0.4218 - val_loss: 67.3994 - val_calc_mre_K: 0.4119\n", "Epoch 1082/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 69.9055 - calc_mre_K: 0.4267 - val_loss: 68.8595 - val_calc_mre_K: 0.4209\n", "Epoch 1083/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 70.7395 - calc_mre_K: 0.4318 - val_loss: 69.0309 - val_calc_mre_K: 0.4219\n", "Epoch 1084/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 69.5231 - calc_mre_K: 0.4243 - val_loss: 69.9069 - val_calc_mre_K: 0.4272\n", "Epoch 1085/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 69.4395 - calc_mre_K: 0.4238 - val_loss: 69.7392 - val_calc_mre_K: 0.4262\n", "Epoch 1086/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 69.3822 - calc_mre_K: 0.4235 - val_loss: 68.9089 - val_calc_mre_K: 0.4211\n", "Epoch 1087/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 69.7304 - calc_mre_K: 0.4256 - val_loss: 71.5975 - val_calc_mre_K: 0.4376\n", "Epoch 1088/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 70.1735 - calc_mre_K: 0.4283 - val_loss: 68.4424 - val_calc_mre_K: 0.4183\n", "Epoch 1089/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 69.3147 - calc_mre_K: 0.4231 - val_loss: 70.1504 - val_calc_mre_K: 0.4287\n", "Epoch 1090/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 69.0401 - calc_mre_K: 0.4214 - val_loss: 70.4309 - val_calc_mre_K: 0.4304\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "Epoch 1091/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 70.1700 - calc_mre_K: 0.4283 - val_loss: 69.9686 - val_calc_mre_K: 0.4276\n", "Epoch 1092/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 70.1071 - calc_mre_K: 0.4279 - val_loss: 67.0333 - val_calc_mre_K: 0.4097\n", "Epoch 1093/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 68.5828 - calc_mre_K: 0.4186 - val_loss: 67.0875 - val_calc_mre_K: 0.4100\n", "Epoch 1094/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 69.5715 - calc_mre_K: 0.4246 - val_loss: 69.7216 - val_calc_mre_K: 0.4261\n", "Epoch 1095/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 69.0423 - calc_mre_K: 0.4214 - val_loss: 66.3138 - val_calc_mre_K: 0.4053\n", "Epoch 1096/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 69.7840 - calc_mre_K: 0.4259 - val_loss: 71.5168 - val_calc_mre_K: 0.4371\n", "Epoch 1097/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 69.1781 - calc_mre_K: 0.4222 - val_loss: 77.4056 - val_calc_mre_K: 0.4731\n", "Epoch 1098/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 69.9773 - calc_mre_K: 0.4271 - val_loss: 71.8817 - val_calc_mre_K: 0.4393\n", "Epoch 1099/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 69.8153 - calc_mre_K: 0.4261 - val_loss: 66.7119 - val_calc_mre_K: 0.4077\n", "Epoch 1100/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 69.1336 - calc_mre_K: 0.4220 - val_loss: 71.5814 - val_calc_mre_K: 0.4375\n", "Epoch 1101/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 69.3750 - calc_mre_K: 0.4234 - val_loss: 70.4741 - val_calc_mre_K: 0.4307\n", "Epoch 1102/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 69.5503 - calc_mre_K: 0.4245 - val_loss: 70.0869 - val_calc_mre_K: 0.4283\n", "Epoch 1103/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 69.3877 - calc_mre_K: 0.4235 - val_loss: 68.0460 - val_calc_mre_K: 0.4159\n", "Epoch 1104/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 69.3585 - calc_mre_K: 0.4233 - val_loss: 70.0856 - val_calc_mre_K: 0.4283\n", "Epoch 1105/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 69.7449 - calc_mre_K: 0.4257 - val_loss: 65.4627 - val_calc_mre_K: 0.4001\n", "Epoch 1106/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 69.6551 - calc_mre_K: 0.4251 - val_loss: 69.8744 - val_calc_mre_K: 0.4270\n", "Epoch 1107/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 69.0012 - calc_mre_K: 0.4212 - val_loss: 68.7323 - val_calc_mre_K: 0.4201\n", "Epoch 1108/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 69.6574 - calc_mre_K: 0.4252 - val_loss: 70.1000 - val_calc_mre_K: 0.4284\n", "Epoch 1109/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 69.0348 - calc_mre_K: 0.4214 - val_loss: 68.8612 - val_calc_mre_K: 0.4208\n", "Epoch 1110/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 69.2029 - calc_mre_K: 0.4224 - val_loss: 69.5487 - val_calc_mre_K: 0.4250\n", "Epoch 1111/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 69.3187 - calc_mre_K: 0.4231 - val_loss: 74.7486 - val_calc_mre_K: 0.4568\n", "Epoch 1112/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 69.1188 - calc_mre_K: 0.4219 - val_loss: 73.6748 - val_calc_mre_K: 0.4503\n", "Epoch 1113/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 68.7808 - calc_mre_K: 0.4198 - val_loss: 70.9176 - val_calc_mre_K: 0.4334\n", "Epoch 1114/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 69.5569 - calc_mre_K: 0.4245 - val_loss: 83.1334 - val_calc_mre_K: 0.5081\n", "Epoch 1115/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 70.2641 - calc_mre_K: 0.4289 - val_loss: 70.5441 - val_calc_mre_K: 0.4312\n", "Epoch 1116/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 69.1114 - calc_mre_K: 0.4218 - val_loss: 71.2170 - val_calc_mre_K: 0.4353\n", "Epoch 1117/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 69.4902 - calc_mre_K: 0.4241 - val_loss: 79.0691 - val_calc_mre_K: 0.4832\n", "Epoch 1118/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 69.7708 - calc_mre_K: 0.4258 - val_loss: 74.7265 - val_calc_mre_K: 0.4567\n", "Epoch 1119/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 69.2425 - calc_mre_K: 0.4226 - val_loss: 68.7702 - val_calc_mre_K: 0.4203\n", "Epoch 1120/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 68.9551 - calc_mre_K: 0.4209 - val_loss: 68.4453 - val_calc_mre_K: 0.4183\n", "Epoch 1121/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 70.0263 - calc_mre_K: 0.4274 - val_loss: 67.2291 - val_calc_mre_K: 0.4109\n", "Epoch 1122/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 69.0675 - calc_mre_K: 0.4216 - val_loss: 68.6343 - val_calc_mre_K: 0.4195\n", "Epoch 1123/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 69.5271 - calc_mre_K: 0.4244 - val_loss: 67.5734 - val_calc_mre_K: 0.4130\n", "Epoch 1124/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 69.1204 - calc_mre_K: 0.4219 - val_loss: 67.5488 - val_calc_mre_K: 0.4128\n", "Epoch 1125/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 69.6451 - calc_mre_K: 0.4251 - val_loss: 68.9972 - val_calc_mre_K: 0.4217\n", "Epoch 1126/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 69.1642 - calc_mre_K: 0.4221 - val_loss: 69.8741 - val_calc_mre_K: 0.4271\n", "Epoch 1127/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 69.6436 - calc_mre_K: 0.4251 - val_loss: 67.9891 - val_calc_mre_K: 0.4155\n", "Epoch 1128/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 69.3185 - calc_mre_K: 0.4231 - val_loss: 82.4123 - val_calc_mre_K: 0.5037\n", "Epoch 1129/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 69.2907 - calc_mre_K: 0.4229 - val_loss: 83.7950 - val_calc_mre_K: 0.5121\n", "Epoch 1130/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 68.7579 - calc_mre_K: 0.4197 - val_loss: 76.7940 - val_calc_mre_K: 0.4693\n", "Epoch 1131/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 68.9652 - calc_mre_K: 0.4209 - val_loss: 73.0939 - val_calc_mre_K: 0.4467\n", "Epoch 1132/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 69.0230 - calc_mre_K: 0.4213 - val_loss: 71.9508 - val_calc_mre_K: 0.4397\n", "Epoch 1133/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 69.4298 - calc_mre_K: 0.4238 - val_loss: 78.3869 - val_calc_mre_K: 0.4791\n", "Epoch 1134/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 69.0054 - calc_mre_K: 0.4212 - val_loss: 68.3033 - val_calc_mre_K: 0.4174\n", "Epoch 1135/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 68.9783 - calc_mre_K: 0.4210 - val_loss: 67.9004 - val_calc_mre_K: 0.4150\n", "Epoch 1136/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 68.9162 - calc_mre_K: 0.4206 - val_loss: 69.4378 - val_calc_mre_K: 0.4244\n", "Epoch 1137/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 69.2219 - calc_mre_K: 0.4225 - val_loss: 68.9968 - val_calc_mre_K: 0.4217\n", "Epoch 1138/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 69.5769 - calc_mre_K: 0.4247 - val_loss: 66.2817 - val_calc_mre_K: 0.4051\n", "Epoch 1139/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 70.5526 - calc_mre_K: 0.4306 - val_loss: 67.1283 - val_calc_mre_K: 0.4103\n", "Epoch 1140/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 68.4881 - calc_mre_K: 0.4180 - val_loss: 68.7977 - val_calc_mre_K: 0.4205\n", "Epoch 1141/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 69.0100 - calc_mre_K: 0.4212 - val_loss: 68.4168 - val_calc_mre_K: 0.4181\n", "Epoch 1142/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 69.0068 - calc_mre_K: 0.4212 - val_loss: 68.9446 - val_calc_mre_K: 0.4214\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "Epoch 1143/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 69.3399 - calc_mre_K: 0.4232 - val_loss: 67.8261 - val_calc_mre_K: 0.4145\n", "Epoch 1144/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 68.6141 - calc_mre_K: 0.4188 - val_loss: 67.1965 - val_calc_mre_K: 0.4107\n", "Epoch 1145/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 69.2195 - calc_mre_K: 0.4225 - val_loss: 66.8148 - val_calc_mre_K: 0.4084\n", "Epoch 1146/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 69.9473 - calc_mre_K: 0.4269 - val_loss: 69.7592 - val_calc_mre_K: 0.4263\n", "Epoch 1147/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 68.6751 - calc_mre_K: 0.4192 - val_loss: 66.2992 - val_calc_mre_K: 0.4052\n", "Epoch 1148/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 68.6645 - calc_mre_K: 0.4191 - val_loss: 75.8719 - val_calc_mre_K: 0.4637\n", "Epoch 1149/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 69.0097 - calc_mre_K: 0.4212 - val_loss: 68.3799 - val_calc_mre_K: 0.4179\n", "Epoch 1150/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 68.6543 - calc_mre_K: 0.4190 - val_loss: 66.4349 - val_calc_mre_K: 0.4060\n", "Epoch 1151/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 69.0696 - calc_mre_K: 0.4216 - val_loss: 76.2473 - val_calc_mre_K: 0.4660\n", "Epoch 1152/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 69.2379 - calc_mre_K: 0.4226 - val_loss: 67.7717 - val_calc_mre_K: 0.4142\n", "Epoch 1153/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 68.8064 - calc_mre_K: 0.4200 - val_loss: 68.3263 - val_calc_mre_K: 0.4176\n", "Epoch 1154/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 68.3591 - calc_mre_K: 0.4172 - val_loss: 72.4296 - val_calc_mre_K: 0.4427\n", "Epoch 1155/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 69.5824 - calc_mre_K: 0.4247 - val_loss: 70.3992 - val_calc_mre_K: 0.4303\n", "Epoch 1156/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 68.7513 - calc_mre_K: 0.4196 - val_loss: 74.0631 - val_calc_mre_K: 0.4527\n", "Epoch 1157/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 68.7962 - calc_mre_K: 0.4199 - val_loss: 68.1635 - val_calc_mre_K: 0.4166\n", "Epoch 1158/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 69.2222 - calc_mre_K: 0.4225 - val_loss: 69.2257 - val_calc_mre_K: 0.4231\n", "Epoch 1159/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 69.6506 - calc_mre_K: 0.4251 - val_loss: 70.3724 - val_calc_mre_K: 0.4301\n", "Epoch 1160/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 68.2460 - calc_mre_K: 0.4165 - val_loss: 67.7916 - val_calc_mre_K: 0.4143\n", "Epoch 1161/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 69.0960 - calc_mre_K: 0.4217 - val_loss: 68.7213 - val_calc_mre_K: 0.4200\n", "Epoch 1162/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 68.4262 - calc_mre_K: 0.4176 - val_loss: 70.8736 - val_calc_mre_K: 0.4331\n", "Epoch 1163/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 69.2470 - calc_mre_K: 0.4226 - val_loss: 67.6896 - val_calc_mre_K: 0.4137\n", "Epoch 1164/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 69.1609 - calc_mre_K: 0.4221 - val_loss: 70.7054 - val_calc_mre_K: 0.4321\n", "Epoch 1165/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 69.0185 - calc_mre_K: 0.4213 - val_loss: 73.3802 - val_calc_mre_K: 0.4485\n", "Epoch 1166/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 68.7615 - calc_mre_K: 0.4197 - val_loss: 69.5717 - val_calc_mre_K: 0.4252\n", "Epoch 1167/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 68.1805 - calc_mre_K: 0.4161 - val_loss: 68.1207 - val_calc_mre_K: 0.4163\n", "Epoch 1168/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 68.3415 - calc_mre_K: 0.4171 - val_loss: 71.4388 - val_calc_mre_K: 0.4366\n", "Epoch 1169/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 69.2258 - calc_mre_K: 0.4225 - val_loss: 76.0259 - val_calc_mre_K: 0.4646\n", "Epoch 1170/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 68.8144 - calc_mre_K: 0.4200 - val_loss: 69.9447 - val_calc_mre_K: 0.4275\n", "Epoch 1171/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 68.3907 - calc_mre_K: 0.4174 - val_loss: 66.3633 - val_calc_mre_K: 0.4056\n", "Epoch 1172/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 69.0074 - calc_mre_K: 0.4212 - val_loss: 69.3908 - val_calc_mre_K: 0.4241\n", "Epoch 1173/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 68.9826 - calc_mre_K: 0.4210 - val_loss: 68.7334 - val_calc_mre_K: 0.4201\n", "Epoch 1174/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 69.1570 - calc_mre_K: 0.4221 - val_loss: 68.3334 - val_calc_mre_K: 0.4176\n", "Epoch 1175/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 68.4711 - calc_mre_K: 0.4179 - val_loss: 65.8687 - val_calc_mre_K: 0.4026\n", "Epoch 1176/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 68.6486 - calc_mre_K: 0.4190 - val_loss: 66.6644 - val_calc_mre_K: 0.4074\n", "Epoch 1177/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 68.8572 - calc_mre_K: 0.4203 - val_loss: 68.8820 - val_calc_mre_K: 0.4210\n", "Epoch 1178/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 69.0423 - calc_mre_K: 0.4214 - val_loss: 68.9076 - val_calc_mre_K: 0.4211\n", "Epoch 1179/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 68.2045 - calc_mre_K: 0.4163 - val_loss: 67.0540 - val_calc_mre_K: 0.4098\n", "Epoch 1180/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 68.4900 - calc_mre_K: 0.4180 - val_loss: 66.6276 - val_calc_mre_K: 0.4072\n", "Epoch 1181/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 69.3795 - calc_mre_K: 0.4235 - val_loss: 64.6327 - val_calc_mre_K: 0.3950\n", "Epoch 1182/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 68.5082 - calc_mre_K: 0.4181 - val_loss: 68.7717 - val_calc_mre_K: 0.4203\n", "Epoch 1183/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 68.1975 - calc_mre_K: 0.4162 - val_loss: 67.2372 - val_calc_mre_K: 0.4109\n", "Epoch 1184/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 68.5808 - calc_mre_K: 0.4186 - val_loss: 68.1015 - val_calc_mre_K: 0.4162\n", "Epoch 1185/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 68.5599 - calc_mre_K: 0.4185 - val_loss: 66.6697 - val_calc_mre_K: 0.4075\n", "Epoch 1186/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 68.4179 - calc_mre_K: 0.4176 - val_loss: 66.7320 - val_calc_mre_K: 0.4078\n", "Epoch 1187/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 69.0613 - calc_mre_K: 0.4215 - val_loss: 70.5096 - val_calc_mre_K: 0.4309\n", "Epoch 1188/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 68.6579 - calc_mre_K: 0.4191 - val_loss: 70.0627 - val_calc_mre_K: 0.4282\n", "Epoch 1189/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 68.7891 - calc_mre_K: 0.4199 - val_loss: 68.1191 - val_calc_mre_K: 0.4163\n", "Epoch 1190/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 68.4428 - calc_mre_K: 0.4177 - val_loss: 70.4421 - val_calc_mre_K: 0.4305\n", "Epoch 1191/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 69.2663 - calc_mre_K: 0.4228 - val_loss: 70.1144 - val_calc_mre_K: 0.4285\n", "Epoch 1192/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 68.7354 - calc_mre_K: 0.4195 - val_loss: 69.4270 - val_calc_mre_K: 0.4243\n", "Epoch 1193/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 68.7700 - calc_mre_K: 0.4197 - val_loss: 67.2696 - val_calc_mre_K: 0.4111\n", "Epoch 1194/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 68.5585 - calc_mre_K: 0.4184 - val_loss: 67.7122 - val_calc_mre_K: 0.4138\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "Epoch 1195/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 68.8510 - calc_mre_K: 0.4202 - val_loss: 69.7172 - val_calc_mre_K: 0.4261\n", "Epoch 1196/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 68.8730 - calc_mre_K: 0.4204 - val_loss: 68.8875 - val_calc_mre_K: 0.4210\n", "Epoch 1197/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 68.6519 - calc_mre_K: 0.4190 - val_loss: 69.1337 - val_calc_mre_K: 0.4225\n", "Epoch 1198/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 68.4710 - calc_mre_K: 0.4179 - val_loss: 69.3441 - val_calc_mre_K: 0.4238\n", "Epoch 1199/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 68.9151 - calc_mre_K: 0.4206 - val_loss: 68.2469 - val_calc_mre_K: 0.4171\n", "Epoch 1200/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 68.0587 - calc_mre_K: 0.4154 - val_loss: 66.5998 - val_calc_mre_K: 0.4070\n", "Epoch 1201/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 68.3450 - calc_mre_K: 0.4171 - val_loss: 67.1057 - val_calc_mre_K: 0.4101\n", "Epoch 1202/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 68.6138 - calc_mre_K: 0.4188 - val_loss: 65.2183 - val_calc_mre_K: 0.3986\n", "Epoch 1203/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 68.7394 - calc_mre_K: 0.4196 - val_loss: 68.4300 - val_calc_mre_K: 0.4182\n", "Epoch 1204/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 68.0550 - calc_mre_K: 0.4154 - val_loss: 68.4434 - val_calc_mre_K: 0.4183\n", "Epoch 1205/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 69.1176 - calc_mre_K: 0.4219 - val_loss: 69.6264 - val_calc_mre_K: 0.4255\n", "Epoch 1206/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 68.4618 - calc_mre_K: 0.4179 - val_loss: 75.8005 - val_calc_mre_K: 0.4633\n", "Epoch 1207/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 67.9487 - calc_mre_K: 0.4147 - val_loss: 68.9143 - val_calc_mre_K: 0.4212\n", "Epoch 1208/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 68.2354 - calc_mre_K: 0.4165 - val_loss: 68.7336 - val_calc_mre_K: 0.4201\n", "Epoch 1209/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 68.0392 - calc_mre_K: 0.4153 - val_loss: 69.9741 - val_calc_mre_K: 0.4277\n", "Epoch 1210/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 69.0111 - calc_mre_K: 0.4212 - val_loss: 69.4340 - val_calc_mre_K: 0.4244\n", "Epoch 1211/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 68.1493 - calc_mre_K: 0.4160 - val_loss: 72.7917 - val_calc_mre_K: 0.4449\n", "Epoch 1212/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 68.7388 - calc_mre_K: 0.4195 - val_loss: 73.1793 - val_calc_mre_K: 0.4472\n", "Epoch 1213/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 68.6863 - calc_mre_K: 0.4192 - val_loss: 66.7924 - val_calc_mre_K: 0.4082\n", "Epoch 1214/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 67.5822 - calc_mre_K: 0.4125 - val_loss: 76.7264 - val_calc_mre_K: 0.4689\n", "Epoch 1215/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 69.1225 - calc_mre_K: 0.4219 - val_loss: 65.1058 - val_calc_mre_K: 0.3979\n", "Epoch 1216/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 68.3529 - calc_mre_K: 0.4172 - val_loss: 68.3137 - val_calc_mre_K: 0.4175\n", "Epoch 1217/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 68.1487 - calc_mre_K: 0.4159 - val_loss: 67.4072 - val_calc_mre_K: 0.4120\n", "Epoch 1218/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 68.7519 - calc_mre_K: 0.4196 - val_loss: 73.8118 - val_calc_mre_K: 0.4511\n", "Epoch 1219/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 68.4827 - calc_mre_K: 0.4180 - val_loss: 68.9176 - val_calc_mre_K: 0.4212\n", "Epoch 1220/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 67.9255 - calc_mre_K: 0.4146 - val_loss: 68.1389 - val_calc_mre_K: 0.4164\n", "Epoch 1221/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 68.7598 - calc_mre_K: 0.4197 - val_loss: 67.7387 - val_calc_mre_K: 0.4140\n", "Epoch 1222/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 67.6685 - calc_mre_K: 0.4130 - val_loss: 66.3060 - val_calc_mre_K: 0.4052\n", "Epoch 1223/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 69.0593 - calc_mre_K: 0.4215 - val_loss: 77.7867 - val_calc_mre_K: 0.4754\n", "Epoch 1224/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 68.0240 - calc_mre_K: 0.4152 - val_loss: 70.3932 - val_calc_mre_K: 0.4302\n", "Epoch 1225/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 68.5484 - calc_mre_K: 0.4184 - val_loss: 75.5238 - val_calc_mre_K: 0.4616\n", "Epoch 1226/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 68.8395 - calc_mre_K: 0.4202 - val_loss: 69.9162 - val_calc_mre_K: 0.4273\n", "Epoch 1227/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 67.9878 - calc_mre_K: 0.4150 - val_loss: 67.9914 - val_calc_mre_K: 0.4155\n", "Epoch 1228/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 67.9549 - calc_mre_K: 0.4148 - val_loss: 68.8882 - val_calc_mre_K: 0.4210\n", "Epoch 1229/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 68.2066 - calc_mre_K: 0.4163 - val_loss: 68.8194 - val_calc_mre_K: 0.4206\n", "Epoch 1230/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 68.0357 - calc_mre_K: 0.4153 - val_loss: 66.8479 - val_calc_mre_K: 0.4085\n", "Epoch 1231/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 68.6057 - calc_mre_K: 0.4187 - val_loss: 70.6760 - val_calc_mre_K: 0.4319\n", "Epoch 1232/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 68.3210 - calc_mre_K: 0.4170 - val_loss: 79.4565 - val_calc_mre_K: 0.4856\n", "Epoch 1233/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 67.7352 - calc_mre_K: 0.4134 - val_loss: 67.5078 - val_calc_mre_K: 0.4126\n", "Epoch 1234/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 68.4297 - calc_mre_K: 0.4177 - val_loss: 68.3576 - val_calc_mre_K: 0.4178\n", "Epoch 1235/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 67.5976 - calc_mre_K: 0.4126 - val_loss: 68.1046 - val_calc_mre_K: 0.4162\n", "Epoch 1236/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 68.4889 - calc_mre_K: 0.4180 - val_loss: 69.4292 - val_calc_mre_K: 0.4243\n", "Epoch 1237/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 68.2147 - calc_mre_K: 0.4163 - val_loss: 65.3031 - val_calc_mre_K: 0.3991\n", "Epoch 1238/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 67.8572 - calc_mre_K: 0.4142 - val_loss: 68.6924 - val_calc_mre_K: 0.4198\n", "Epoch 1239/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 69.2400 - calc_mre_K: 0.4226 - val_loss: 66.7111 - val_calc_mre_K: 0.4077\n", "Epoch 1240/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 67.5248 - calc_mre_K: 0.4121 - val_loss: 66.9758 - val_calc_mre_K: 0.4093\n", "Epoch 1241/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 68.6479 - calc_mre_K: 0.4190 - val_loss: 66.4292 - val_calc_mre_K: 0.4060\n", "Epoch 1242/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 67.7585 - calc_mre_K: 0.4136 - val_loss: 69.9408 - val_calc_mre_K: 0.4274\n", "Epoch 1243/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 68.0461 - calc_mre_K: 0.4153 - val_loss: 65.4315 - val_calc_mre_K: 0.3999\n", "Epoch 1244/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 68.5226 - calc_mre_K: 0.4182 - val_loss: 68.0102 - val_calc_mre_K: 0.4157\n", "Epoch 1245/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 68.0689 - calc_mre_K: 0.4155 - val_loss: 72.0231 - val_calc_mre_K: 0.4402\n", "Epoch 1246/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 67.7755 - calc_mre_K: 0.4137 - val_loss: 70.4876 - val_calc_mre_K: 0.4308\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "Epoch 1247/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 67.9382 - calc_mre_K: 0.4147 - val_loss: 74.6998 - val_calc_mre_K: 0.4565\n", "Epoch 1248/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 67.9330 - calc_mre_K: 0.4146 - val_loss: 66.6137 - val_calc_mre_K: 0.4071\n", "Epoch 1249/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 68.2488 - calc_mre_K: 0.4166 - val_loss: 71.1461 - val_calc_mre_K: 0.4348\n", "Epoch 1250/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 67.9483 - calc_mre_K: 0.4147 - val_loss: 64.9762 - val_calc_mre_K: 0.3971\n", "Epoch 1251/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 68.3008 - calc_mre_K: 0.4169 - val_loss: 65.9712 - val_calc_mre_K: 0.4032\n", "Epoch 1252/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 68.0482 - calc_mre_K: 0.4153 - val_loss: 70.0406 - val_calc_mre_K: 0.4281\n", "Epoch 1253/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 68.2261 - calc_mre_K: 0.4164 - val_loss: 70.6578 - val_calc_mre_K: 0.4318\n", "Epoch 1254/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 68.1332 - calc_mre_K: 0.4159 - val_loss: 69.5267 - val_calc_mre_K: 0.4249\n", "Epoch 1255/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 67.9684 - calc_mre_K: 0.4148 - val_loss: 68.9373 - val_calc_mre_K: 0.4213\n", "Epoch 1256/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 68.1252 - calc_mre_K: 0.4158 - val_loss: 66.9600 - val_calc_mre_K: 0.4092\n", "Epoch 1257/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 68.1211 - calc_mre_K: 0.4158 - val_loss: 66.2710 - val_calc_mre_K: 0.4050\n", "Epoch 1258/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 67.7423 - calc_mre_K: 0.4135 - val_loss: 71.3141 - val_calc_mre_K: 0.4359\n", "Epoch 1259/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 67.6214 - calc_mre_K: 0.4127 - val_loss: 71.1746 - val_calc_mre_K: 0.4350\n", "Epoch 1260/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 67.4382 - calc_mre_K: 0.4116 - val_loss: 70.6548 - val_calc_mre_K: 0.4318\n", "Epoch 1261/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 68.0198 - calc_mre_K: 0.4152 - val_loss: 67.3710 - val_calc_mre_K: 0.4118\n", "Epoch 1262/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 68.4251 - calc_mre_K: 0.4176 - val_loss: 67.3533 - val_calc_mre_K: 0.4116\n", "Epoch 1263/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 66.8178 - calc_mre_K: 0.4078 - val_loss: 73.0205 - val_calc_mre_K: 0.4463\n", "Epoch 1264/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 68.7167 - calc_mre_K: 0.4194 - val_loss: 76.1155 - val_calc_mre_K: 0.4652\n", "Epoch 1265/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 67.5064 - calc_mre_K: 0.4120 - val_loss: 70.7939 - val_calc_mre_K: 0.4327\n", "Epoch 1266/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 68.4372 - calc_mre_K: 0.4177 - val_loss: 65.9905 - val_calc_mre_K: 0.4033\n", "Epoch 1267/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 68.1655 - calc_mre_K: 0.4160 - val_loss: 67.8521 - val_calc_mre_K: 0.4147\n", "Epoch 1268/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 67.3399 - calc_mre_K: 0.4110 - val_loss: 66.8863 - val_calc_mre_K: 0.4088\n", "Epoch 1269/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 68.3294 - calc_mre_K: 0.4170 - val_loss: 66.7632 - val_calc_mre_K: 0.4080\n", "Epoch 1270/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 67.8940 - calc_mre_K: 0.4144 - val_loss: 73.0008 - val_calc_mre_K: 0.4462\n", "Epoch 1271/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 68.1238 - calc_mre_K: 0.4158 - val_loss: 68.7431 - val_calc_mre_K: 0.4201\n", "Epoch 1272/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 67.3232 - calc_mre_K: 0.4109 - val_loss: 69.6501 - val_calc_mre_K: 0.4257\n", "Epoch 1273/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 67.6575 - calc_mre_K: 0.4129 - val_loss: 72.3383 - val_calc_mre_K: 0.4421\n", "Epoch 1274/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 67.7810 - calc_mre_K: 0.4137 - val_loss: 71.3557 - val_calc_mre_K: 0.4361\n", "Epoch 1275/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 67.7414 - calc_mre_K: 0.4135 - val_loss: 65.8105 - val_calc_mre_K: 0.4022\n", "Epoch 1276/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 68.1786 - calc_mre_K: 0.4161 - val_loss: 70.6323 - val_calc_mre_K: 0.4317\n", "Epoch 1277/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 68.1278 - calc_mre_K: 0.4158 - val_loss: 67.2178 - val_calc_mre_K: 0.4108\n", "Epoch 1278/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 68.2187 - calc_mre_K: 0.4164 - val_loss: 66.6059 - val_calc_mre_K: 0.4071\n", "Epoch 1279/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 67.4855 - calc_mre_K: 0.4119 - val_loss: 66.5339 - val_calc_mre_K: 0.4066\n", "Epoch 1280/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 67.3356 - calc_mre_K: 0.4110 - val_loss: 69.1443 - val_calc_mre_K: 0.4226\n", "Epoch 1281/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 67.4608 - calc_mre_K: 0.4117 - val_loss: 66.7278 - val_calc_mre_K: 0.4078\n", "Epoch 1282/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 67.9562 - calc_mre_K: 0.4148 - val_loss: 68.9028 - val_calc_mre_K: 0.4211\n", "Epoch 1283/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 67.7687 - calc_mre_K: 0.4136 - val_loss: 67.4074 - val_calc_mre_K: 0.4120\n", "Epoch 1284/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 67.9472 - calc_mre_K: 0.4147 - val_loss: 66.6759 - val_calc_mre_K: 0.4075\n", "Epoch 1285/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 67.8672 - calc_mre_K: 0.4142 - val_loss: 69.5201 - val_calc_mre_K: 0.4249\n", "Epoch 1286/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 67.6652 - calc_mre_K: 0.4130 - val_loss: 66.4343 - val_calc_mre_K: 0.4060\n", "Epoch 1287/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 67.6717 - calc_mre_K: 0.4130 - val_loss: 66.7787 - val_calc_mre_K: 0.4081\n", "Epoch 1288/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 67.3412 - calc_mre_K: 0.4110 - val_loss: 70.5935 - val_calc_mre_K: 0.4314\n", "Epoch 1289/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 68.1083 - calc_mre_K: 0.4157 - val_loss: 66.2180 - val_calc_mre_K: 0.4047\n", "Epoch 1290/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 68.1149 - calc_mre_K: 0.4157 - val_loss: 65.4666 - val_calc_mre_K: 0.4001\n", "Epoch 1291/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 67.0680 - calc_mre_K: 0.4094 - val_loss: 70.7042 - val_calc_mre_K: 0.4321\n", "Epoch 1292/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 68.0608 - calc_mre_K: 0.4154 - val_loss: 66.7864 - val_calc_mre_K: 0.4082\n", "Epoch 1293/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 67.0232 - calc_mre_K: 0.4091 - val_loss: 67.0523 - val_calc_mre_K: 0.4098\n", "Epoch 1294/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 68.1324 - calc_mre_K: 0.4158 - val_loss: 71.3281 - val_calc_mre_K: 0.4359\n", "Epoch 1295/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 68.8402 - calc_mre_K: 0.4202 - val_loss: 67.7819 - val_calc_mre_K: 0.4143\n", "Epoch 1296/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 67.1400 - calc_mre_K: 0.4098 - val_loss: 67.2390 - val_calc_mre_K: 0.4109\n", "Epoch 1297/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 67.5276 - calc_mre_K: 0.4122 - val_loss: 65.3529 - val_calc_mre_K: 0.3994\n", "Epoch 1298/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 68.2201 - calc_mre_K: 0.4164 - val_loss: 65.2305 - val_calc_mre_K: 0.3987\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "Epoch 1299/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 67.8230 - calc_mre_K: 0.4140 - val_loss: 67.6410 - val_calc_mre_K: 0.4134\n", "Epoch 1300/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 67.0176 - calc_mre_K: 0.4090 - val_loss: 67.4258 - val_calc_mre_K: 0.4121\n", "Epoch 1301/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 68.1043 - calc_mre_K: 0.4157 - val_loss: 67.3994 - val_calc_mre_K: 0.4119\n", "Epoch 1302/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 67.6428 - calc_mre_K: 0.4129 - val_loss: 68.1917 - val_calc_mre_K: 0.4168\n", "Epoch 1303/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 67.3734 - calc_mre_K: 0.4112 - val_loss: 67.2184 - val_calc_mre_K: 0.4108\n", "Epoch 1304/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 67.2413 - calc_mre_K: 0.4104 - val_loss: 73.0029 - val_calc_mre_K: 0.4462\n", "Epoch 1305/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 67.3972 - calc_mre_K: 0.4114 - val_loss: 65.3503 - val_calc_mre_K: 0.3994\n", "Epoch 1306/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 68.0402 - calc_mre_K: 0.4153 - val_loss: 64.6219 - val_calc_mre_K: 0.3950\n", "Epoch 1307/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 67.9463 - calc_mre_K: 0.4147 - val_loss: 69.5408 - val_calc_mre_K: 0.4250\n", "Epoch 1308/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 67.0080 - calc_mre_K: 0.4090 - val_loss: 69.8648 - val_calc_mre_K: 0.4270\n", "Epoch 1309/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 68.7100 - calc_mre_K: 0.4194 - val_loss: 67.0077 - val_calc_mre_K: 0.4095\n", "Epoch 1310/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 67.2456 - calc_mre_K: 0.4104 - val_loss: 69.1055 - val_calc_mre_K: 0.4223\n", "Epoch 1311/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 67.7192 - calc_mre_K: 0.4133 - val_loss: 68.2620 - val_calc_mre_K: 0.4172\n", "Epoch 1312/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 67.5722 - calc_mre_K: 0.4124 - val_loss: 65.6032 - val_calc_mre_K: 0.4010\n", "Epoch 1313/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 67.3291 - calc_mre_K: 0.4109 - val_loss: 67.7474 - val_calc_mre_K: 0.4141\n", "Epoch 1314/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 67.5472 - calc_mre_K: 0.4123 - val_loss: 69.5164 - val_calc_mre_K: 0.4249\n", "Epoch 1315/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 67.9914 - calc_mre_K: 0.4150 - val_loss: 67.1314 - val_calc_mre_K: 0.4103\n", "Epoch 1316/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 67.7255 - calc_mre_K: 0.4134 - val_loss: 68.6530 - val_calc_mre_K: 0.4196\n", "Epoch 1317/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 67.2588 - calc_mre_K: 0.4105 - val_loss: 68.5108 - val_calc_mre_K: 0.4187\n", "Epoch 1318/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 67.8168 - calc_mre_K: 0.4139 - val_loss: 65.3438 - val_calc_mre_K: 0.3994\n", "Epoch 1319/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 67.3118 - calc_mre_K: 0.4108 - val_loss: 67.4391 - val_calc_mre_K: 0.4122\n", "Epoch 1320/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 67.3912 - calc_mre_K: 0.4113 - val_loss: 69.4774 - val_calc_mre_K: 0.4246\n", "Epoch 1321/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 67.9533 - calc_mre_K: 0.4148 - val_loss: 66.0737 - val_calc_mre_K: 0.4038\n", "Epoch 1322/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 67.4901 - calc_mre_K: 0.4119 - val_loss: 72.7110 - val_calc_mre_K: 0.4444\n", "Epoch 1323/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 67.4589 - calc_mre_K: 0.4117 - val_loss: 69.3068 - val_calc_mre_K: 0.4236\n", "Epoch 1324/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 67.7203 - calc_mre_K: 0.4133 - val_loss: 68.5997 - val_calc_mre_K: 0.4193\n", "Epoch 1325/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 67.8003 - calc_mre_K: 0.4138 - val_loss: 66.2680 - val_calc_mre_K: 0.4050\n", "Epoch 1326/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 67.6527 - calc_mre_K: 0.4129 - val_loss: 71.4530 - val_calc_mre_K: 0.4367\n", "Epoch 1327/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 67.2233 - calc_mre_K: 0.4103 - val_loss: 66.8192 - val_calc_mre_K: 0.4084\n", "Epoch 1328/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 68.3418 - calc_mre_K: 0.4171 - val_loss: 77.1745 - val_calc_mre_K: 0.4717\n", "Epoch 1329/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 67.6388 - calc_mre_K: 0.4128 - val_loss: 66.9949 - val_calc_mre_K: 0.4095\n", "Epoch 1330/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 67.7553 - calc_mre_K: 0.4135 - val_loss: 71.2331 - val_calc_mre_K: 0.4354\n", "Epoch 1331/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 66.8275 - calc_mre_K: 0.4079 - val_loss: 65.1969 - val_calc_mre_K: 0.3985\n", "Epoch 1332/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 67.3247 - calc_mre_K: 0.4109 - val_loss: 67.5965 - val_calc_mre_K: 0.4131\n", "Epoch 1333/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 67.1362 - calc_mre_K: 0.4098 - val_loss: 69.4039 - val_calc_mre_K: 0.4242\n", "Epoch 1334/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 68.4950 - calc_mre_K: 0.4181 - val_loss: 65.3769 - val_calc_mre_K: 0.3996\n", "Epoch 1335/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 67.1184 - calc_mre_K: 0.4097 - val_loss: 67.2134 - val_calc_mre_K: 0.4108\n", "Epoch 1336/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 67.3167 - calc_mre_K: 0.4109 - val_loss: 65.7573 - val_calc_mre_K: 0.4019\n", "Epoch 1337/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 67.7074 - calc_mre_K: 0.4133 - val_loss: 72.9270 - val_calc_mre_K: 0.4457\n", "Epoch 1338/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 67.5914 - calc_mre_K: 0.4125 - val_loss: 67.4574 - val_calc_mre_K: 0.4123\n", "Epoch 1339/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 68.7049 - calc_mre_K: 0.4193 - val_loss: 68.7401 - val_calc_mre_K: 0.4201\n", "Epoch 1340/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 67.2015 - calc_mre_K: 0.4102 - val_loss: 69.4232 - val_calc_mre_K: 0.4243\n", "Epoch 1341/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 68.0928 - calc_mre_K: 0.4156 - val_loss: 68.0255 - val_calc_mre_K: 0.4157\n", "Epoch 1342/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 66.8100 - calc_mre_K: 0.4078 - val_loss: 65.9748 - val_calc_mre_K: 0.4032\n", "Epoch 1343/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 67.9307 - calc_mre_K: 0.4146 - val_loss: 69.8166 - val_calc_mre_K: 0.4267\n", "Epoch 1344/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 67.7577 - calc_mre_K: 0.4136 - val_loss: 66.7899 - val_calc_mre_K: 0.4082\n", "Epoch 1345/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 67.2523 - calc_mre_K: 0.4105 - val_loss: 68.1840 - val_calc_mre_K: 0.4167\n", "Epoch 1346/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 67.0465 - calc_mre_K: 0.4092 - val_loss: 66.9424 - val_calc_mre_K: 0.4091\n", "Epoch 1347/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 67.4941 - calc_mre_K: 0.4120 - val_loss: 69.2420 - val_calc_mre_K: 0.4232\n", "Epoch 1348/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 66.9384 - calc_mre_K: 0.4086 - val_loss: 66.8487 - val_calc_mre_K: 0.4086\n", "Epoch 1349/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 68.0017 - calc_mre_K: 0.4150 - val_loss: 67.4233 - val_calc_mre_K: 0.4121\n", "Epoch 1350/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 67.7530 - calc_mre_K: 0.4135 - val_loss: 65.4707 - val_calc_mre_K: 0.4001\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "Epoch 1351/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 67.5864 - calc_mre_K: 0.4125 - val_loss: 65.6424 - val_calc_mre_K: 0.4012\n", "Epoch 1352/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 67.1571 - calc_mre_K: 0.4099 - val_loss: 71.1639 - val_calc_mre_K: 0.4349\n", "Epoch 1353/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 68.5447 - calc_mre_K: 0.4184 - val_loss: 71.4424 - val_calc_mre_K: 0.4366\n", "Epoch 1354/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 66.9744 - calc_mre_K: 0.4088 - val_loss: 66.4154 - val_calc_mre_K: 0.4059\n", "Epoch 1355/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 66.6420 - calc_mre_K: 0.4068 - val_loss: 67.9334 - val_calc_mre_K: 0.4152\n", "Epoch 1356/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 67.8663 - calc_mre_K: 0.4142 - val_loss: 66.9332 - val_calc_mre_K: 0.4091\n", "Epoch 1357/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 67.3758 - calc_mre_K: 0.4112 - val_loss: 68.3316 - val_calc_mre_K: 0.4176\n", "Epoch 1358/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 66.9694 - calc_mre_K: 0.4087 - val_loss: 67.2720 - val_calc_mre_K: 0.4112\n", "Epoch 1359/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 67.4033 - calc_mre_K: 0.4114 - val_loss: 67.0479 - val_calc_mre_K: 0.4098\n", "Epoch 1360/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 67.2637 - calc_mre_K: 0.4105 - val_loss: 67.0340 - val_calc_mre_K: 0.4097\n", "Epoch 1361/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 66.9806 - calc_mre_K: 0.4088 - val_loss: 67.0556 - val_calc_mre_K: 0.4098\n", "Epoch 1362/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 68.1159 - calc_mre_K: 0.4157 - val_loss: 67.9435 - val_calc_mre_K: 0.4153\n", "Epoch 1363/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 66.7135 - calc_mre_K: 0.4072 - val_loss: 70.4680 - val_calc_mre_K: 0.4307\n", "Epoch 1364/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 67.6518 - calc_mre_K: 0.4129 - val_loss: 68.5262 - val_calc_mre_K: 0.4188\n", "Epoch 1365/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 67.1552 - calc_mre_K: 0.4099 - val_loss: 66.9622 - val_calc_mre_K: 0.4093\n", "Epoch 1366/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 67.1544 - calc_mre_K: 0.4099 - val_loss: 68.6093 - val_calc_mre_K: 0.4193\n", "Epoch 1367/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 67.3760 - calc_mre_K: 0.4112 - val_loss: 64.3368 - val_calc_mre_K: 0.3932\n", "Epoch 1368/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 67.7061 - calc_mre_K: 0.4132 - val_loss: 74.8968 - val_calc_mre_K: 0.4577\n", "Epoch 1369/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 66.8896 - calc_mre_K: 0.4083 - val_loss: 65.1666 - val_calc_mre_K: 0.3983\n", "Epoch 1370/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 67.5378 - calc_mre_K: 0.4122 - val_loss: 65.5392 - val_calc_mre_K: 0.4006\n", "Epoch 1371/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 67.7479 - calc_mre_K: 0.4135 - val_loss: 71.0983 - val_calc_mre_K: 0.4345\n", "Epoch 1372/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 67.4837 - calc_mre_K: 0.4119 - val_loss: 66.2721 - val_calc_mre_K: 0.4050\n", "Epoch 1373/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 67.1419 - calc_mre_K: 0.4098 - val_loss: 68.6179 - val_calc_mre_K: 0.4194\n", "Epoch 1374/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 66.6571 - calc_mre_K: 0.4068 - val_loss: 64.3385 - val_calc_mre_K: 0.3932\n", "Epoch 1375/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 67.3613 - calc_mre_K: 0.4111 - val_loss: 69.7326 - val_calc_mre_K: 0.4262\n", "Epoch 1376/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 67.3071 - calc_mre_K: 0.4108 - val_loss: 66.1587 - val_calc_mre_K: 0.4043\n", "Epoch 1377/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 67.1868 - calc_mre_K: 0.4101 - val_loss: 67.4118 - val_calc_mre_K: 0.4120\n", "Epoch 1378/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 67.5178 - calc_mre_K: 0.4121 - val_loss: 67.7400 - val_calc_mre_K: 0.4140\n", "Epoch 1379/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 67.5298 - calc_mre_K: 0.4122 - val_loss: 66.9898 - val_calc_mre_K: 0.4094\n", "Epoch 1380/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 67.2002 - calc_mre_K: 0.4102 - val_loss: 74.4311 - val_calc_mre_K: 0.4549\n", "Epoch 1381/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 67.3207 - calc_mre_K: 0.4109 - val_loss: 64.8931 - val_calc_mre_K: 0.3966\n", "Epoch 1382/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 67.2968 - calc_mre_K: 0.4107 - val_loss: 66.6345 - val_calc_mre_K: 0.4073\n", "Epoch 1383/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 66.4887 - calc_mre_K: 0.4058 - val_loss: 66.7751 - val_calc_mre_K: 0.4081\n", "Epoch 1384/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 67.6345 - calc_mre_K: 0.4128 - val_loss: 67.4674 - val_calc_mre_K: 0.4123\n", "Epoch 1385/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 67.2148 - calc_mre_K: 0.4102 - val_loss: 71.1416 - val_calc_mre_K: 0.4348\n", "Epoch 1386/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 67.1902 - calc_mre_K: 0.4101 - val_loss: 67.0961 - val_calc_mre_K: 0.4101\n", "Epoch 1387/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 66.7182 - calc_mre_K: 0.4072 - val_loss: 64.3022 - val_calc_mre_K: 0.3930\n", "Epoch 1388/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 66.6682 - calc_mre_K: 0.4069 - val_loss: 65.6591 - val_calc_mre_K: 0.4013\n", "Epoch 1389/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 66.9022 - calc_mre_K: 0.4083 - val_loss: 65.4326 - val_calc_mre_K: 0.3999\n", "Epoch 1390/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 67.1890 - calc_mre_K: 0.4101 - val_loss: 67.9142 - val_calc_mre_K: 0.4151\n", "Epoch 1391/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 67.2273 - calc_mre_K: 0.4103 - val_loss: 66.4240 - val_calc_mre_K: 0.4060\n", "Epoch 1392/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 67.1227 - calc_mre_K: 0.4097 - val_loss: 65.9242 - val_calc_mre_K: 0.4029\n", "Epoch 1393/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 66.7371 - calc_mre_K: 0.4073 - val_loss: 67.5858 - val_calc_mre_K: 0.4131\n", "Epoch 1394/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 67.8947 - calc_mre_K: 0.4144 - val_loss: 65.9269 - val_calc_mre_K: 0.4029\n", "Epoch 1395/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 66.8232 - calc_mre_K: 0.4079 - val_loss: 63.7094 - val_calc_mre_K: 0.3894\n", "Epoch 1396/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 67.2835 - calc_mre_K: 0.4107 - val_loss: 66.4421 - val_calc_mre_K: 0.4061\n", "Epoch 1397/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 67.0901 - calc_mre_K: 0.4095 - val_loss: 66.5681 - val_calc_mre_K: 0.4068\n", "Epoch 1398/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 66.7252 - calc_mre_K: 0.4073 - val_loss: 66.7317 - val_calc_mre_K: 0.4078\n", "Epoch 1399/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 66.9507 - calc_mre_K: 0.4086 - val_loss: 67.4888 - val_calc_mre_K: 0.4125\n", "Epoch 1400/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 66.6800 - calc_mre_K: 0.4070 - val_loss: 69.3475 - val_calc_mre_K: 0.4238\n", "Epoch 1401/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 67.4295 - calc_mre_K: 0.4116 - val_loss: 69.8861 - val_calc_mre_K: 0.4271\n", "Epoch 1402/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 67.0235 - calc_mre_K: 0.4091 - val_loss: 67.8037 - val_calc_mre_K: 0.4144\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "Epoch 1403/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 66.9354 - calc_mre_K: 0.4085 - val_loss: 70.9195 - val_calc_mre_K: 0.4334\n", "Epoch 1404/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 67.2161 - calc_mre_K: 0.4103 - val_loss: 68.5032 - val_calc_mre_K: 0.4187\n", "Epoch 1405/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 66.8105 - calc_mre_K: 0.4078 - val_loss: 65.6143 - val_calc_mre_K: 0.4010\n", "Epoch 1406/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 66.7764 - calc_mre_K: 0.4076 - val_loss: 64.6388 - val_calc_mre_K: 0.3951\n", "Epoch 1407/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 67.3079 - calc_mre_K: 0.4108 - val_loss: 66.6651 - val_calc_mre_K: 0.4074\n", "Epoch 1408/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 66.7752 - calc_mre_K: 0.4076 - val_loss: 70.0736 - val_calc_mre_K: 0.4283\n", "Epoch 1409/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 67.2300 - calc_mre_K: 0.4103 - val_loss: 64.2261 - val_calc_mre_K: 0.3925\n", "Epoch 1410/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 67.1384 - calc_mre_K: 0.4098 - val_loss: 67.1811 - val_calc_mre_K: 0.4106\n", "Epoch 1411/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 66.8875 - calc_mre_K: 0.4082 - val_loss: 65.9098 - val_calc_mre_K: 0.4028\n", "Epoch 1412/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 67.4341 - calc_mre_K: 0.4116 - val_loss: 69.3678 - val_calc_mre_K: 0.4240\n", "Epoch 1413/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 66.8198 - calc_mre_K: 0.4078 - val_loss: 67.6806 - val_calc_mre_K: 0.4136\n", "Epoch 1414/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 67.1252 - calc_mre_K: 0.4097 - val_loss: 67.8357 - val_calc_mre_K: 0.4146\n", "Epoch 1415/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 67.0519 - calc_mre_K: 0.4093 - val_loss: 69.9496 - val_calc_mre_K: 0.4275\n", "Epoch 1416/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 67.6782 - calc_mre_K: 0.4131 - val_loss: 68.2907 - val_calc_mre_K: 0.4174\n", "Epoch 1417/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 67.0114 - calc_mre_K: 0.4090 - val_loss: 65.8495 - val_calc_mre_K: 0.4025\n", "Epoch 1418/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 66.4995 - calc_mre_K: 0.4059 - val_loss: 68.3433 - val_calc_mre_K: 0.4177\n", "Epoch 1419/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 66.9401 - calc_mre_K: 0.4086 - val_loss: 68.4110 - val_calc_mre_K: 0.4181\n", "Epoch 1420/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 67.7827 - calc_mre_K: 0.4137 - val_loss: 65.1886 - val_calc_mre_K: 0.3984\n", "Epoch 1421/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 66.6666 - calc_mre_K: 0.4069 - val_loss: 66.1716 - val_calc_mre_K: 0.4044\n", "Epoch 1422/2000\n", "48000/48000 [==============================] - 3s 57us/step - loss: 67.4270 - calc_mre_K: 0.4115 - val_loss: 67.4051 - val_calc_mre_K: 0.4120\n", "Epoch 1423/2000\n", "48000/48000 [==============================] - 2s 48us/step - loss: 67.4934 - calc_mre_K: 0.4119 - val_loss: 65.1219 - val_calc_mre_K: 0.3980\n", "Epoch 1424/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 66.3716 - calc_mre_K: 0.4051 - val_loss: 65.5696 - val_calc_mre_K: 0.4007\n", "Epoch 1425/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 66.9440 - calc_mre_K: 0.4086 - val_loss: 66.1811 - val_calc_mre_K: 0.4045\n", "Epoch 1426/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 66.5846 - calc_mre_K: 0.4064 - val_loss: 66.4976 - val_calc_mre_K: 0.4064\n", "Epoch 1427/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 67.4119 - calc_mre_K: 0.4114 - val_loss: 66.4809 - val_calc_mre_K: 0.4063\n", "Epoch 1428/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 66.6011 - calc_mre_K: 0.4065 - val_loss: 64.9998 - val_calc_mre_K: 0.3973\n", "Epoch 1429/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 67.5407 - calc_mre_K: 0.4122 - val_loss: 71.3360 - val_calc_mre_K: 0.4360\n", "Epoch 1430/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 66.6412 - calc_mre_K: 0.4067 - val_loss: 69.4984 - val_calc_mre_K: 0.4248\n", "Epoch 1431/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 66.3351 - calc_mre_K: 0.4049 - val_loss: 67.1164 - val_calc_mre_K: 0.4102\n", "Epoch 1432/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 67.3852 - calc_mre_K: 0.4113 - val_loss: 71.5450 - val_calc_mre_K: 0.4373\n", "Epoch 1433/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 66.8128 - calc_mre_K: 0.4078 - val_loss: 63.2000 - val_calc_mre_K: 0.3863\n", "Epoch 1434/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 66.9854 - calc_mre_K: 0.4088 - val_loss: 67.5138 - val_calc_mre_K: 0.4126\n", "Epoch 1435/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 67.4200 - calc_mre_K: 0.4115 - val_loss: 67.6262 - val_calc_mre_K: 0.4133\n", "Epoch 1436/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 66.5952 - calc_mre_K: 0.4065 - val_loss: 63.9881 - val_calc_mre_K: 0.3911\n", "Epoch 1437/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 67.3581 - calc_mre_K: 0.4111 - val_loss: 67.2472 - val_calc_mre_K: 0.4110\n", "Epoch 1438/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 66.8473 - calc_mre_K: 0.4080 - val_loss: 65.6209 - val_calc_mre_K: 0.4011\n", "Epoch 1439/2000\n", "48000/48000 [==============================] - 2s 48us/step - loss: 66.8724 - calc_mre_K: 0.4082 - val_loss: 69.6773 - val_calc_mre_K: 0.4259\n", "Epoch 1440/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 66.9574 - calc_mre_K: 0.4087 - val_loss: 67.2167 - val_calc_mre_K: 0.4108\n", "Epoch 1441/2000\n", "48000/48000 [==============================] - 2s 49us/step - loss: 66.7310 - calc_mre_K: 0.4073 - val_loss: 66.7138 - val_calc_mre_K: 0.4077\n", "Epoch 1442/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 67.2614 - calc_mre_K: 0.4105 - val_loss: 78.2311 - val_calc_mre_K: 0.4781\n", "Epoch 1443/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 66.6519 - calc_mre_K: 0.4068 - val_loss: 65.2286 - val_calc_mre_K: 0.3987\n", "Epoch 1444/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 67.1794 - calc_mre_K: 0.4100 - val_loss: 63.6752 - val_calc_mre_K: 0.3892\n", "Epoch 1445/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 66.8831 - calc_mre_K: 0.4082 - val_loss: 65.3000 - val_calc_mre_K: 0.3991\n", "Epoch 1446/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 66.9847 - calc_mre_K: 0.4088 - val_loss: 66.2487 - val_calc_mre_K: 0.4049\n", "Epoch 1447/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 67.0515 - calc_mre_K: 0.4092 - val_loss: 66.5438 - val_calc_mre_K: 0.4067\n", "Epoch 1448/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 66.4356 - calc_mre_K: 0.4055 - val_loss: 72.8968 - val_calc_mre_K: 0.4455\n", "Epoch 1449/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 67.2229 - calc_mre_K: 0.4103 - val_loss: 70.9056 - val_calc_mre_K: 0.4333\n", "Epoch 1450/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 67.3540 - calc_mre_K: 0.4111 - val_loss: 68.3023 - val_calc_mre_K: 0.4175\n", "Epoch 1451/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 66.6016 - calc_mre_K: 0.4065 - val_loss: 64.8244 - val_calc_mre_K: 0.3962\n", "Epoch 1452/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 66.6275 - calc_mre_K: 0.4067 - val_loss: 65.1693 - val_calc_mre_K: 0.3983\n", "Epoch 1453/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 67.2779 - calc_mre_K: 0.4106 - val_loss: 64.6337 - val_calc_mre_K: 0.3950\n", "Epoch 1454/2000\n", "48000/48000 [==============================] - 2s 49us/step - loss: 66.5128 - calc_mre_K: 0.4060 - val_loss: 67.1368 - val_calc_mre_K: 0.4103\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "Epoch 1455/2000\n", "48000/48000 [==============================] - 3s 55us/step - loss: 67.0196 - calc_mre_K: 0.4091 - val_loss: 65.0229 - val_calc_mre_K: 0.3974\n", "Epoch 1456/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 66.6190 - calc_mre_K: 0.4066 - val_loss: 70.0301 - val_calc_mre_K: 0.4280\n", "Epoch 1457/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 67.4071 - calc_mre_K: 0.4114 - val_loss: 66.2515 - val_calc_mre_K: 0.4049\n", "Epoch 1458/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 66.4030 - calc_mre_K: 0.4053 - val_loss: 67.3289 - val_calc_mre_K: 0.4115\n", "Epoch 1459/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 67.3565 - calc_mre_K: 0.4111 - val_loss: 70.7631 - val_calc_mre_K: 0.4325\n", "Epoch 1460/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 67.2574 - calc_mre_K: 0.4105 - val_loss: 67.8466 - val_calc_mre_K: 0.4147\n", "Epoch 1461/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 66.7357 - calc_mre_K: 0.4073 - val_loss: 66.6880 - val_calc_mre_K: 0.4076\n", "Epoch 1462/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 66.7019 - calc_mre_K: 0.4071 - val_loss: 64.3763 - val_calc_mre_K: 0.3935\n", "Epoch 1463/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 66.1729 - calc_mre_K: 0.4039 - val_loss: 65.8847 - val_calc_mre_K: 0.4027\n", "Epoch 1464/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 66.5912 - calc_mre_K: 0.4064 - val_loss: 66.3411 - val_calc_mre_K: 0.4055\n", "Epoch 1465/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 66.6873 - calc_mre_K: 0.4070 - val_loss: 67.5956 - val_calc_mre_K: 0.4131\n", "Epoch 1466/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 66.7732 - calc_mre_K: 0.4076 - val_loss: 65.3600 - val_calc_mre_K: 0.3995\n", "Epoch 1467/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 67.5541 - calc_mre_K: 0.4123 - val_loss: 66.1377 - val_calc_mre_K: 0.4042\n", "Epoch 1468/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 66.3218 - calc_mre_K: 0.4048 - val_loss: 66.3255 - val_calc_mre_K: 0.4054\n", "Epoch 1469/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 66.9008 - calc_mre_K: 0.4083 - val_loss: 71.3952 - val_calc_mre_K: 0.4364\n", "Epoch 1470/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 66.6698 - calc_mre_K: 0.4069 - val_loss: 67.0085 - val_calc_mre_K: 0.4095\n", "Epoch 1471/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 66.9593 - calc_mre_K: 0.4087 - val_loss: 66.3973 - val_calc_mre_K: 0.4058\n", "Epoch 1472/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 66.0710 - calc_mre_K: 0.4033 - val_loss: 64.6952 - val_calc_mre_K: 0.3954\n", "Epoch 1473/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 67.5112 - calc_mre_K: 0.4121 - val_loss: 70.8127 - val_calc_mre_K: 0.4328\n", "Epoch 1474/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 66.8758 - calc_mre_K: 0.4082 - val_loss: 70.9166 - val_calc_mre_K: 0.4334\n", "Epoch 1475/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 66.6269 - calc_mre_K: 0.4067 - val_loss: 71.6717 - val_calc_mre_K: 0.4380\n", "Epoch 1476/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 67.3015 - calc_mre_K: 0.4108 - val_loss: 67.1179 - val_calc_mre_K: 0.4102\n", "Epoch 1477/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 66.0661 - calc_mre_K: 0.4032 - val_loss: 66.5226 - val_calc_mre_K: 0.4066\n", "Epoch 1478/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 66.5340 - calc_mre_K: 0.4061 - val_loss: 67.5654 - val_calc_mre_K: 0.4129\n", "Epoch 1479/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 66.7612 - calc_mre_K: 0.4075 - val_loss: 66.7074 - val_calc_mre_K: 0.4077\n", "Epoch 1480/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 67.6983 - calc_mre_K: 0.4132 - val_loss: 66.2587 - val_calc_mre_K: 0.4050\n", "Epoch 1481/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 66.8712 - calc_mre_K: 0.4081 - val_loss: 64.5485 - val_calc_mre_K: 0.3945\n", "Epoch 1482/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 66.7566 - calc_mre_K: 0.4074 - val_loss: 65.2015 - val_calc_mre_K: 0.3985\n", "Epoch 1483/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 66.7140 - calc_mre_K: 0.4072 - val_loss: 65.1287 - val_calc_mre_K: 0.3981\n", "Epoch 1484/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 66.6342 - calc_mre_K: 0.4067 - val_loss: 66.8103 - val_calc_mre_K: 0.4083\n", "Epoch 1485/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 67.4672 - calc_mre_K: 0.4118 - val_loss: 65.2898 - val_calc_mre_K: 0.3990\n", "Epoch 1486/2000\n", "48000/48000 [==============================] - 2s 48us/step - loss: 66.5969 - calc_mre_K: 0.4065 - val_loss: 65.1490 - val_calc_mre_K: 0.3982\n", "Epoch 1487/2000\n", "48000/48000 [==============================] - 3s 55us/step - loss: 66.2440 - calc_mre_K: 0.4043 - val_loss: 70.0348 - val_calc_mre_K: 0.4280\n", "Epoch 1488/2000\n", "48000/48000 [==============================] - 3s 55us/step - loss: 66.9567 - calc_mre_K: 0.4087 - val_loss: 64.1535 - val_calc_mre_K: 0.3921\n", "Epoch 1489/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 66.9388 - calc_mre_K: 0.4086 - val_loss: 69.2092 - val_calc_mre_K: 0.4230\n", "Epoch 1490/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 66.6466 - calc_mre_K: 0.4068 - val_loss: 68.9792 - val_calc_mre_K: 0.4216\n", "Epoch 1491/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 66.4134 - calc_mre_K: 0.4054 - val_loss: 65.5930 - val_calc_mre_K: 0.4009\n", "Epoch 1492/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 66.3573 - calc_mre_K: 0.4050 - val_loss: 67.2382 - val_calc_mre_K: 0.4110\n", "Epoch 1493/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 66.7016 - calc_mre_K: 0.4071 - val_loss: 66.2076 - val_calc_mre_K: 0.4046\n", "Epoch 1494/2000\n", "48000/48000 [==============================] - 2s 47us/step - loss: 66.5961 - calc_mre_K: 0.4065 - val_loss: 69.4124 - val_calc_mre_K: 0.4242\n", "Epoch 1495/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 66.3151 - calc_mre_K: 0.4048 - val_loss: 70.5698 - val_calc_mre_K: 0.4313\n", "Epoch 1496/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 66.8533 - calc_mre_K: 0.4080 - val_loss: 68.0773 - val_calc_mre_K: 0.4161\n", "Epoch 1497/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 66.6762 - calc_mre_K: 0.4070 - val_loss: 69.7507 - val_calc_mre_K: 0.4263\n", "Epoch 1498/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 65.8608 - calc_mre_K: 0.4020 - val_loss: 66.1455 - val_calc_mre_K: 0.4043\n", "Epoch 1499/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 67.1415 - calc_mre_K: 0.4098 - val_loss: 62.7482 - val_calc_mre_K: 0.3835\n", "Epoch 1500/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 66.4539 - calc_mre_K: 0.4056 - val_loss: 66.7140 - val_calc_mre_K: 0.4077\n", "Epoch 1501/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 66.7057 - calc_mre_K: 0.4071 - val_loss: 64.6436 - val_calc_mre_K: 0.3951\n", "Epoch 1502/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 66.5795 - calc_mre_K: 0.4064 - val_loss: 66.0650 - val_calc_mre_K: 0.4038\n", "Epoch 1503/2000\n", "48000/48000 [==============================] - 3s 57us/step - loss: 66.3948 - calc_mre_K: 0.4052 - val_loss: 70.5974 - val_calc_mre_K: 0.4315\n", "Epoch 1504/2000\n", "48000/48000 [==============================] - 3s 57us/step - loss: 66.8454 - calc_mre_K: 0.4080 - val_loss: 68.6244 - val_calc_mre_K: 0.4194\n", "Epoch 1505/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 66.5259 - calc_mre_K: 0.4060 - val_loss: 65.2731 - val_calc_mre_K: 0.3989\n", "Epoch 1506/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 65.9402 - calc_mre_K: 0.4025 - val_loss: 70.2351 - val_calc_mre_K: 0.4293\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "Epoch 1507/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 66.6556 - calc_mre_K: 0.4068 - val_loss: 65.1495 - val_calc_mre_K: 0.3982\n", "Epoch 1508/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 66.4463 - calc_mre_K: 0.4056 - val_loss: 66.2637 - val_calc_mre_K: 0.4050\n", "Epoch 1509/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 66.1080 - calc_mre_K: 0.4035 - val_loss: 66.2814 - val_calc_mre_K: 0.4051\n", "Epoch 1510/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 66.7207 - calc_mre_K: 0.4072 - val_loss: 70.0827 - val_calc_mre_K: 0.4283\n", "Epoch 1511/2000\n", "48000/48000 [==============================] - 3s 56us/step - loss: 66.0743 - calc_mre_K: 0.4033 - val_loss: 62.6982 - val_calc_mre_K: 0.3832\n", "Epoch 1512/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 66.4921 - calc_mre_K: 0.4058 - val_loss: 67.1770 - val_calc_mre_K: 0.4106\n", "Epoch 1513/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 66.1999 - calc_mre_K: 0.4041 - val_loss: 63.6641 - val_calc_mre_K: 0.3891\n", "Epoch 1514/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 66.2750 - calc_mre_K: 0.4045 - val_loss: 65.0346 - val_calc_mre_K: 0.3975\n", "Epoch 1515/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 66.4013 - calc_mre_K: 0.4053 - val_loss: 69.2834 - val_calc_mre_K: 0.4234\n", "Epoch 1516/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 66.7551 - calc_mre_K: 0.4074 - val_loss: 63.4327 - val_calc_mre_K: 0.3877\n", "Epoch 1517/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 66.0688 - calc_mre_K: 0.4033 - val_loss: 67.4071 - val_calc_mre_K: 0.4120\n", "Epoch 1518/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 66.1383 - calc_mre_K: 0.4037 - val_loss: 65.9858 - val_calc_mre_K: 0.4033\n", "Epoch 1519/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 66.5820 - calc_mre_K: 0.4064 - val_loss: 66.1430 - val_calc_mre_K: 0.4043\n", "Epoch 1520/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 66.2842 - calc_mre_K: 0.4046 - val_loss: 65.7125 - val_calc_mre_K: 0.4016\n", "Epoch 1521/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 66.0707 - calc_mre_K: 0.4033 - val_loss: 65.2387 - val_calc_mre_K: 0.3987\n", "Epoch 1522/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 66.1916 - calc_mre_K: 0.4040 - val_loss: 67.2127 - val_calc_mre_K: 0.4108\n", "Epoch 1523/2000\n", "48000/48000 [==============================] - 3s 55us/step - loss: 66.3347 - calc_mre_K: 0.4049 - val_loss: 64.0242 - val_calc_mre_K: 0.3913\n", "Epoch 1524/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 66.2928 - calc_mre_K: 0.4046 - val_loss: 68.9994 - val_calc_mre_K: 0.4217\n", "Epoch 1525/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 67.9661 - calc_mre_K: 0.4148 - val_loss: 66.7288 - val_calc_mre_K: 0.4078\n", "Epoch 1526/2000\n", "48000/48000 [==============================] - 3s 55us/step - loss: 65.4969 - calc_mre_K: 0.3998 - val_loss: 68.8549 - val_calc_mre_K: 0.4208\n", "Epoch 1527/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 66.5516 - calc_mre_K: 0.4062 - val_loss: 63.4690 - val_calc_mre_K: 0.3879\n", "Epoch 1528/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 65.6623 - calc_mre_K: 0.4008 - val_loss: 64.4923 - val_calc_mre_K: 0.3942\n", "Epoch 1529/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 66.5572 - calc_mre_K: 0.4062 - val_loss: 65.2605 - val_calc_mre_K: 0.3989\n", "Epoch 1530/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 66.4056 - calc_mre_K: 0.4053 - val_loss: 68.7542 - val_calc_mre_K: 0.4202\n", "Epoch 1531/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 66.1643 - calc_mre_K: 0.4038 - val_loss: 69.9806 - val_calc_mre_K: 0.4277\n", "Epoch 1532/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 66.3019 - calc_mre_K: 0.4047 - val_loss: 66.5140 - val_calc_mre_K: 0.4065\n", "Epoch 1533/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 66.3888 - calc_mre_K: 0.4052 - val_loss: 65.3117 - val_calc_mre_K: 0.3992\n", "Epoch 1534/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 66.2738 - calc_mre_K: 0.4045 - val_loss: 65.5231 - val_calc_mre_K: 0.4005\n", "Epoch 1535/2000\n", "48000/48000 [==============================] - 2s 49us/step - loss: 66.2206 - calc_mre_K: 0.4042 - val_loss: 71.9082 - val_calc_mre_K: 0.4395\n", "Epoch 1536/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 66.1774 - calc_mre_K: 0.4039 - val_loss: 70.5574 - val_calc_mre_K: 0.4312\n", "Epoch 1537/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 67.0341 - calc_mre_K: 0.4091 - val_loss: 64.0874 - val_calc_mre_K: 0.3917\n", "Epoch 1538/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 66.1498 - calc_mre_K: 0.4037 - val_loss: 65.9674 - val_calc_mre_K: 0.4032\n", "Epoch 1539/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 66.1579 - calc_mre_K: 0.4038 - val_loss: 68.1482 - val_calc_mre_K: 0.4165\n", "Epoch 1540/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 66.4388 - calc_mre_K: 0.4055 - val_loss: 66.6035 - val_calc_mre_K: 0.4071\n", "Epoch 1541/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 66.0321 - calc_mre_K: 0.4030 - val_loss: 66.5372 - val_calc_mre_K: 0.4067\n", "Epoch 1542/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 66.2363 - calc_mre_K: 0.4043 - val_loss: 66.5850 - val_calc_mre_K: 0.4070\n", "Epoch 1543/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 66.3376 - calc_mre_K: 0.4049 - val_loss: 63.9356 - val_calc_mre_K: 0.3907\n", "Epoch 1544/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 66.2940 - calc_mre_K: 0.4046 - val_loss: 63.0232 - val_calc_mre_K: 0.3852\n", "Epoch 1545/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 66.3053 - calc_mre_K: 0.4047 - val_loss: 67.1201 - val_calc_mre_K: 0.4102\n", "Epoch 1546/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 65.6813 - calc_mre_K: 0.4009 - val_loss: 64.9178 - val_calc_mre_K: 0.3968\n", "Epoch 1547/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 66.2074 - calc_mre_K: 0.4041 - val_loss: 63.4481 - val_calc_mre_K: 0.3878\n", "Epoch 1548/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 66.4775 - calc_mre_K: 0.4057 - val_loss: 71.4104 - val_calc_mre_K: 0.4364\n", "Epoch 1549/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 66.4848 - calc_mre_K: 0.4058 - val_loss: 65.5733 - val_calc_mre_K: 0.4008\n", "Epoch 1550/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 66.4507 - calc_mre_K: 0.4056 - val_loss: 66.2802 - val_calc_mre_K: 0.4051\n", "Epoch 1551/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 66.2163 - calc_mre_K: 0.4042 - val_loss: 67.4316 - val_calc_mre_K: 0.4121\n", "Epoch 1552/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 67.4413 - calc_mre_K: 0.4116 - val_loss: 64.5792 - val_calc_mre_K: 0.3947\n", "Epoch 1553/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 66.3908 - calc_mre_K: 0.4052 - val_loss: 64.4120 - val_calc_mre_K: 0.3937\n", "Epoch 1554/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 65.7817 - calc_mre_K: 0.4015 - val_loss: 66.3937 - val_calc_mre_K: 0.4058\n", "Epoch 1555/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 65.9200 - calc_mre_K: 0.4023 - val_loss: 69.4836 - val_calc_mre_K: 0.4247\n", "Epoch 1556/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 66.0072 - calc_mre_K: 0.4029 - val_loss: 71.6205 - val_calc_mre_K: 0.4377\n", "Epoch 1557/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 66.3973 - calc_mre_K: 0.4053 - val_loss: 66.5349 - val_calc_mre_K: 0.4066\n", "Epoch 1558/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 66.3109 - calc_mre_K: 0.4047 - val_loss: 65.1255 - val_calc_mre_K: 0.3980\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "Epoch 1559/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 65.7849 - calc_mre_K: 0.4015 - val_loss: 64.9095 - val_calc_mre_K: 0.3967\n", "Epoch 1560/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 66.0574 - calc_mre_K: 0.4032 - val_loss: 63.4745 - val_calc_mre_K: 0.3879\n", "Epoch 1561/2000\n", "48000/48000 [==============================] - 2s 48us/step - loss: 65.9471 - calc_mre_K: 0.4025 - val_loss: 64.2533 - val_calc_mre_K: 0.3927\n", "Epoch 1562/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 67.1296 - calc_mre_K: 0.4097 - val_loss: 63.8291 - val_calc_mre_K: 0.3901\n", "Epoch 1563/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 65.9713 - calc_mre_K: 0.4027 - val_loss: 65.3609 - val_calc_mre_K: 0.3995\n", "Epoch 1564/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 66.2184 - calc_mre_K: 0.4042 - val_loss: 68.3779 - val_calc_mre_K: 0.4179\n", "Epoch 1565/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 66.9179 - calc_mre_K: 0.4084 - val_loss: 64.5169 - val_calc_mre_K: 0.3943\n", "Epoch 1566/2000\n", "48000/48000 [==============================] - 2s 49us/step - loss: 65.7183 - calc_mre_K: 0.4011 - val_loss: 66.0045 - val_calc_mre_K: 0.4034\n", "Epoch 1567/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 66.4177 - calc_mre_K: 0.4054 - val_loss: 65.0090 - val_calc_mre_K: 0.3973\n", "Epoch 1568/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 65.9800 - calc_mre_K: 0.4027 - val_loss: 66.2335 - val_calc_mre_K: 0.4048\n", "Epoch 1569/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 66.7761 - calc_mre_K: 0.4076 - val_loss: 71.7707 - val_calc_mre_K: 0.4386\n", "Epoch 1570/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 65.9455 - calc_mre_K: 0.4025 - val_loss: 63.5269 - val_calc_mre_K: 0.3883\n", "Epoch 1571/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 65.9425 - calc_mre_K: 0.4025 - val_loss: 64.7338 - val_calc_mre_K: 0.3956\n", "Epoch 1572/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 66.2443 - calc_mre_K: 0.4043 - val_loss: 70.5395 - val_calc_mre_K: 0.4311\n", "Epoch 1573/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 66.5002 - calc_mre_K: 0.4059 - val_loss: 67.7839 - val_calc_mre_K: 0.4143\n", "Epoch 1574/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 65.7002 - calc_mre_K: 0.4010 - val_loss: 67.7010 - val_calc_mre_K: 0.4138\n", "Epoch 1575/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 65.7613 - calc_mre_K: 0.4014 - val_loss: 63.6660 - val_calc_mre_K: 0.3891\n", "Epoch 1576/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 66.4608 - calc_mre_K: 0.4056 - val_loss: 64.7732 - val_calc_mre_K: 0.3959\n", "Epoch 1577/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 66.0102 - calc_mre_K: 0.4029 - val_loss: 66.1020 - val_calc_mre_K: 0.4040\n", "Epoch 1578/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 66.5862 - calc_mre_K: 0.4064 - val_loss: 62.7669 - val_calc_mre_K: 0.3836\n", "Epoch 1579/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 66.0472 - calc_mre_K: 0.4031 - val_loss: 67.6637 - val_calc_mre_K: 0.4135\n", "Epoch 1580/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 66.0599 - calc_mre_K: 0.4032 - val_loss: 69.4000 - val_calc_mre_K: 0.4242\n", "Epoch 1581/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 66.0496 - calc_mre_K: 0.4031 - val_loss: 65.6813 - val_calc_mre_K: 0.4014\n", "Epoch 1582/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 65.7242 - calc_mre_K: 0.4011 - val_loss: 67.8671 - val_calc_mre_K: 0.4148\n", "Epoch 1583/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 66.1327 - calc_mre_K: 0.4036 - val_loss: 62.4826 - val_calc_mre_K: 0.3819\n", "Epoch 1584/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 65.9514 - calc_mre_K: 0.4025 - val_loss: 63.9251 - val_calc_mre_K: 0.3907\n", "Epoch 1585/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 65.7716 - calc_mre_K: 0.4014 - val_loss: 70.7387 - val_calc_mre_K: 0.4323\n", "Epoch 1586/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 66.0200 - calc_mre_K: 0.4030 - val_loss: 62.9697 - val_calc_mre_K: 0.3849\n", "Epoch 1587/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 66.0520 - calc_mre_K: 0.4031 - val_loss: 65.4911 - val_calc_mre_K: 0.4003\n", "Epoch 1588/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 66.2358 - calc_mre_K: 0.4043 - val_loss: 64.3441 - val_calc_mre_K: 0.3932\n", "Epoch 1589/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 66.0613 - calc_mre_K: 0.4032 - val_loss: 64.3395 - val_calc_mre_K: 0.3932\n", "Epoch 1590/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 65.8613 - calc_mre_K: 0.4020 - val_loss: 63.9209 - val_calc_mre_K: 0.3907\n", "Epoch 1591/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 66.1908 - calc_mre_K: 0.4040 - val_loss: 65.3886 - val_calc_mre_K: 0.3996\n", "Epoch 1592/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 66.4900 - calc_mre_K: 0.4058 - val_loss: 65.3654 - val_calc_mre_K: 0.3995\n", "Epoch 1593/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 65.6483 - calc_mre_K: 0.4007 - val_loss: 69.9748 - val_calc_mre_K: 0.4277\n", "Epoch 1594/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 65.8584 - calc_mre_K: 0.4020 - val_loss: 63.0032 - val_calc_mre_K: 0.3851\n", "Epoch 1595/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 66.0367 - calc_mre_K: 0.4031 - val_loss: 65.0608 - val_calc_mre_K: 0.3976\n", "Epoch 1596/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 66.0990 - calc_mre_K: 0.4034 - val_loss: 64.3026 - val_calc_mre_K: 0.3930\n", "Epoch 1597/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 65.9375 - calc_mre_K: 0.4025 - val_loss: 62.1823 - val_calc_mre_K: 0.3800\n", "Epoch 1598/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 66.1562 - calc_mre_K: 0.4038 - val_loss: 70.2031 - val_calc_mre_K: 0.4291\n", "Epoch 1599/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 65.8256 - calc_mre_K: 0.4018 - val_loss: 70.0586 - val_calc_mre_K: 0.4282\n", "Epoch 1600/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 66.0051 - calc_mre_K: 0.4029 - val_loss: 70.0688 - val_calc_mre_K: 0.4283\n", "Epoch 1601/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 65.8838 - calc_mre_K: 0.4021 - val_loss: 64.1881 - val_calc_mre_K: 0.3923\n", "Epoch 1602/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 65.9045 - calc_mre_K: 0.4022 - val_loss: 65.2870 - val_calc_mre_K: 0.3990\n", "Epoch 1603/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 66.1386 - calc_mre_K: 0.4037 - val_loss: 64.5132 - val_calc_mre_K: 0.3943\n", "Epoch 1604/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 66.9729 - calc_mre_K: 0.4088 - val_loss: 63.4325 - val_calc_mre_K: 0.3877\n", "Epoch 1605/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 65.4630 - calc_mre_K: 0.3996 - val_loss: 66.3160 - val_calc_mre_K: 0.4053\n", "Epoch 1606/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 65.8036 - calc_mre_K: 0.4016 - val_loss: 71.0933 - val_calc_mre_K: 0.4345\n", "Epoch 1607/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 66.1984 - calc_mre_K: 0.4040 - val_loss: 64.6423 - val_calc_mre_K: 0.3951\n", "Epoch 1608/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 66.3438 - calc_mre_K: 0.4049 - val_loss: 74.1584 - val_calc_mre_K: 0.4532\n", "Epoch 1609/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 65.5569 - calc_mre_K: 0.4001 - val_loss: 65.5629 - val_calc_mre_K: 0.4007\n", "Epoch 1610/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 66.2778 - calc_mre_K: 0.4045 - val_loss: 70.8717 - val_calc_mre_K: 0.4331\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "Epoch 1611/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 66.3014 - calc_mre_K: 0.4047 - val_loss: 70.7680 - val_calc_mre_K: 0.4325\n", "Epoch 1612/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 65.5333 - calc_mre_K: 0.4000 - val_loss: 65.3351 - val_calc_mre_K: 0.3993\n", "Epoch 1613/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 65.6899 - calc_mre_K: 0.4009 - val_loss: 63.9474 - val_calc_mre_K: 0.3908\n", "Epoch 1614/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 65.8348 - calc_mre_K: 0.4018 - val_loss: 64.7168 - val_calc_mre_K: 0.3955\n", "Epoch 1615/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 66.0606 - calc_mre_K: 0.4032 - val_loss: 67.7576 - val_calc_mre_K: 0.4141\n", "Epoch 1616/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 66.2980 - calc_mre_K: 0.4047 - val_loss: 66.0612 - val_calc_mre_K: 0.4037\n", "Epoch 1617/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 66.1568 - calc_mre_K: 0.4038 - val_loss: 63.3433 - val_calc_mre_K: 0.3871\n", "Epoch 1618/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 65.8982 - calc_mre_K: 0.4022 - val_loss: 65.1566 - val_calc_mre_K: 0.3982\n", "Epoch 1619/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 65.9400 - calc_mre_K: 0.4025 - val_loss: 64.3854 - val_calc_mre_K: 0.3935\n", "Epoch 1620/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 65.6786 - calc_mre_K: 0.4009 - val_loss: 66.4675 - val_calc_mre_K: 0.4062\n", "Epoch 1621/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 65.6584 - calc_mre_K: 0.4007 - val_loss: 71.5273 - val_calc_mre_K: 0.4372\n", "Epoch 1622/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 65.9001 - calc_mre_K: 0.4022 - val_loss: 66.1279 - val_calc_mre_K: 0.4042\n", "Epoch 1623/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 65.6045 - calc_mre_K: 0.4004 - val_loss: 65.2083 - val_calc_mre_K: 0.3985\n", "Epoch 1624/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 65.9236 - calc_mre_K: 0.4024 - val_loss: 64.2456 - val_calc_mre_K: 0.3927\n", "Epoch 1625/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 65.8295 - calc_mre_K: 0.4018 - val_loss: 69.3136 - val_calc_mre_K: 0.4236\n", "Epoch 1626/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 65.9638 - calc_mre_K: 0.4026 - val_loss: 75.4927 - val_calc_mre_K: 0.4614\n", "Epoch 1627/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 65.9738 - calc_mre_K: 0.4027 - val_loss: 64.6760 - val_calc_mre_K: 0.3953\n", "Epoch 1628/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 65.9829 - calc_mre_K: 0.4027 - val_loss: 66.4516 - val_calc_mre_K: 0.4061\n", "Epoch 1629/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 65.8120 - calc_mre_K: 0.4017 - val_loss: 63.9222 - val_calc_mre_K: 0.3907\n", "Epoch 1630/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 66.1492 - calc_mre_K: 0.4037 - val_loss: 66.4442 - val_calc_mre_K: 0.4061\n", "Epoch 1631/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 65.4385 - calc_mre_K: 0.3994 - val_loss: 64.4980 - val_calc_mre_K: 0.3942\n", "Epoch 1632/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 66.0046 - calc_mre_K: 0.4029 - val_loss: 67.4967 - val_calc_mre_K: 0.4125\n", "Epoch 1633/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 65.5460 - calc_mre_K: 0.4001 - val_loss: 63.5514 - val_calc_mre_K: 0.3884\n", "Epoch 1634/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 65.2719 - calc_mre_K: 0.3984 - val_loss: 65.6661 - val_calc_mre_K: 0.4013\n", "Epoch 1635/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 66.4024 - calc_mre_K: 0.4053 - val_loss: 65.0737 - val_calc_mre_K: 0.3977\n", "Epoch 1636/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 65.3610 - calc_mre_K: 0.3989 - val_loss: 66.3884 - val_calc_mre_K: 0.4057\n", "Epoch 1637/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 66.4215 - calc_mre_K: 0.4054 - val_loss: 71.7503 - val_calc_mre_K: 0.4385\n", "Epoch 1638/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 66.3982 - calc_mre_K: 0.4053 - val_loss: 62.9377 - val_calc_mre_K: 0.3847\n", "Epoch 1639/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 66.0221 - calc_mre_K: 0.4030 - val_loss: 64.1458 - val_calc_mre_K: 0.3920\n", "Epoch 1640/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 65.0950 - calc_mre_K: 0.3973 - val_loss: 65.2789 - val_calc_mre_K: 0.3990\n", "Epoch 1641/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 65.8340 - calc_mre_K: 0.4018 - val_loss: 65.1097 - val_calc_mre_K: 0.3979\n", "Epoch 1642/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 65.7494 - calc_mre_K: 0.4013 - val_loss: 66.8396 - val_calc_mre_K: 0.4085\n", "Epoch 1643/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 66.2423 - calc_mre_K: 0.4043 - val_loss: 68.0521 - val_calc_mre_K: 0.4159\n", "Epoch 1644/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 66.1177 - calc_mre_K: 0.4036 - val_loss: 68.6359 - val_calc_mre_K: 0.4195\n", "Epoch 1645/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 65.5884 - calc_mre_K: 0.4003 - val_loss: 65.4999 - val_calc_mre_K: 0.4003\n", "Epoch 1646/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 66.0438 - calc_mre_K: 0.4031 - val_loss: 62.7217 - val_calc_mre_K: 0.3833\n", "Epoch 1647/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 65.3076 - calc_mre_K: 0.3986 - val_loss: 66.1051 - val_calc_mre_K: 0.4040\n", "Epoch 1648/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 66.2803 - calc_mre_K: 0.4045 - val_loss: 69.9275 - val_calc_mre_K: 0.4274\n", "Epoch 1649/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 65.8615 - calc_mre_K: 0.4020 - val_loss: 66.4553 - val_calc_mre_K: 0.4062\n", "Epoch 1650/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 65.6741 - calc_mre_K: 0.4008 - val_loss: 66.1879 - val_calc_mre_K: 0.4045\n", "Epoch 1651/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 65.6776 - calc_mre_K: 0.4009 - val_loss: 66.3649 - val_calc_mre_K: 0.4056\n", "Epoch 1652/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 65.7276 - calc_mre_K: 0.4012 - val_loss: 68.5033 - val_calc_mre_K: 0.4187\n", "Epoch 1653/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 66.0434 - calc_mre_K: 0.4031 - val_loss: 67.8065 - val_calc_mre_K: 0.4144\n", "Epoch 1654/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 65.8867 - calc_mre_K: 0.4021 - val_loss: 66.6162 - val_calc_mre_K: 0.4072\n", "Epoch 1655/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 66.2000 - calc_mre_K: 0.4041 - val_loss: 68.9705 - val_calc_mre_K: 0.4215\n", "Epoch 1656/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 66.7629 - calc_mre_K: 0.4075 - val_loss: 65.3290 - val_calc_mre_K: 0.3993\n", "Epoch 1657/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 65.5138 - calc_mre_K: 0.3999 - val_loss: 66.0797 - val_calc_mre_K: 0.4039\n", "Epoch 1658/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 65.9306 - calc_mre_K: 0.4024 - val_loss: 66.0692 - val_calc_mre_K: 0.4038\n", "Epoch 1659/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 64.9879 - calc_mre_K: 0.3967 - val_loss: 62.8201 - val_calc_mre_K: 0.3840\n", "Epoch 1660/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 65.9206 - calc_mre_K: 0.4023 - val_loss: 65.1087 - val_calc_mre_K: 0.3979\n", "Epoch 1661/2000\n", "48000/48000 [==============================] - 2s 49us/step - loss: 65.4625 - calc_mre_K: 0.3996 - val_loss: 66.8515 - val_calc_mre_K: 0.4086\n", "Epoch 1662/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 66.6477 - calc_mre_K: 0.4068 - val_loss: 64.5679 - val_calc_mre_K: 0.3946\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "Epoch 1663/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 65.3807 - calc_mre_K: 0.3991 - val_loss: 65.7133 - val_calc_mre_K: 0.4016\n", "Epoch 1664/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 65.9800 - calc_mre_K: 0.4027 - val_loss: 63.2490 - val_calc_mre_K: 0.3866\n", "Epoch 1665/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 65.3110 - calc_mre_K: 0.3986 - val_loss: 73.4893 - val_calc_mre_K: 0.4491\n", "Epoch 1666/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 66.3354 - calc_mre_K: 0.4049 - val_loss: 65.0823 - val_calc_mre_K: 0.3978\n", "Epoch 1667/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 65.7222 - calc_mre_K: 0.4011 - val_loss: 65.2625 - val_calc_mre_K: 0.3989\n", "Epoch 1668/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 66.1318 - calc_mre_K: 0.4036 - val_loss: 70.5070 - val_calc_mre_K: 0.4309\n", "Epoch 1669/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 65.6611 - calc_mre_K: 0.4008 - val_loss: 66.9402 - val_calc_mre_K: 0.4091\n", "Epoch 1670/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 65.4188 - calc_mre_K: 0.3993 - val_loss: 69.9615 - val_calc_mre_K: 0.4276\n", "Epoch 1671/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 66.0385 - calc_mre_K: 0.4031 - val_loss: 64.4930 - val_calc_mre_K: 0.3942\n", "Epoch 1672/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 65.7754 - calc_mre_K: 0.4015 - val_loss: 69.2012 - val_calc_mre_K: 0.4229\n", "Epoch 1673/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 65.8463 - calc_mre_K: 0.4019 - val_loss: 62.3192 - val_calc_mre_K: 0.3809\n", "Epoch 1674/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 65.2322 - calc_mre_K: 0.3981 - val_loss: 67.0399 - val_calc_mre_K: 0.4097\n", "Epoch 1675/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 65.5469 - calc_mre_K: 0.4001 - val_loss: 69.5741 - val_calc_mre_K: 0.4252\n", "Epoch 1676/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 66.1115 - calc_mre_K: 0.4035 - val_loss: 65.5192 - val_calc_mre_K: 0.4004\n", "Epoch 1677/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 65.3384 - calc_mre_K: 0.3988 - val_loss: 61.4111 - val_calc_mre_K: 0.3753\n", "Epoch 1678/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 65.5199 - calc_mre_K: 0.3999 - val_loss: 63.9915 - val_calc_mre_K: 0.3911\n", "Epoch 1679/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 66.0721 - calc_mre_K: 0.4033 - val_loss: 67.5598 - val_calc_mre_K: 0.4129\n", "Epoch 1680/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 65.8631 - calc_mre_K: 0.4020 - val_loss: 64.1874 - val_calc_mre_K: 0.3923\n", "Epoch 1681/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 65.6614 - calc_mre_K: 0.4008 - val_loss: 64.2577 - val_calc_mre_K: 0.3927\n", "Epoch 1682/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 65.7264 - calc_mre_K: 0.4012 - val_loss: 68.6462 - val_calc_mre_K: 0.4195\n", "Epoch 1683/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 65.7708 - calc_mre_K: 0.4014 - val_loss: 64.5286 - val_calc_mre_K: 0.3944\n", "Epoch 1684/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 65.3814 - calc_mre_K: 0.3991 - val_loss: 70.6349 - val_calc_mre_K: 0.4317\n", "Epoch 1685/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 65.9093 - calc_mre_K: 0.4023 - val_loss: 66.3136 - val_calc_mre_K: 0.4053\n", "Epoch 1686/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 65.1849 - calc_mre_K: 0.3979 - val_loss: 68.9366 - val_calc_mre_K: 0.4213\n", "Epoch 1687/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 65.7866 - calc_mre_K: 0.4015 - val_loss: 64.6873 - val_calc_mre_K: 0.3953\n", "Epoch 1688/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 65.8162 - calc_mre_K: 0.4017 - val_loss: 67.6264 - val_calc_mre_K: 0.4133\n", "Epoch 1689/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 65.3558 - calc_mre_K: 0.3989 - val_loss: 70.5498 - val_calc_mre_K: 0.4312\n", "Epoch 1690/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 65.6467 - calc_mre_K: 0.4007 - val_loss: 65.9529 - val_calc_mre_K: 0.4031\n", "Epoch 1691/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 65.5733 - calc_mre_K: 0.4002 - val_loss: 62.7556 - val_calc_mre_K: 0.3836\n", "Epoch 1692/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 66.0537 - calc_mre_K: 0.4032 - val_loss: 70.3702 - val_calc_mre_K: 0.4301\n", "Epoch 1693/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 65.5853 - calc_mre_K: 0.4003 - val_loss: 68.7949 - val_calc_mre_K: 0.4205\n", "Epoch 1694/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 65.2973 - calc_mre_K: 0.3985 - val_loss: 65.2003 - val_calc_mre_K: 0.3985\n", "Epoch 1695/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 65.3050 - calc_mre_K: 0.3986 - val_loss: 67.2069 - val_calc_mre_K: 0.4107\n", "Epoch 1696/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 65.5943 - calc_mre_K: 0.4004 - val_loss: 64.3645 - val_calc_mre_K: 0.3934\n", "Epoch 1697/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 65.7340 - calc_mre_K: 0.4012 - val_loss: 64.8475 - val_calc_mre_K: 0.3963\n", "Epoch 1698/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 65.9058 - calc_mre_K: 0.4023 - val_loss: 64.2058 - val_calc_mre_K: 0.3924\n", "Epoch 1699/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 65.3562 - calc_mre_K: 0.3989 - val_loss: 68.9435 - val_calc_mre_K: 0.4214\n", "Epoch 1700/2000\n", "48000/48000 [==============================] - 2s 49us/step - loss: 66.0164 - calc_mre_K: 0.4029 - val_loss: 65.2497 - val_calc_mre_K: 0.3988\n", "Epoch 1701/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 65.8284 - calc_mre_K: 0.4018 - val_loss: 66.6008 - val_calc_mre_K: 0.4071\n", "Epoch 1702/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 65.3157 - calc_mre_K: 0.3987 - val_loss: 78.1176 - val_calc_mre_K: 0.4774\n", "Epoch 1703/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 66.4829 - calc_mre_K: 0.4058 - val_loss: 69.0828 - val_calc_mre_K: 0.4222\n", "Epoch 1704/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 65.4325 - calc_mre_K: 0.3994 - val_loss: 65.9520 - val_calc_mre_K: 0.4031\n", "Epoch 1705/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 65.5543 - calc_mre_K: 0.4001 - val_loss: 62.2417 - val_calc_mre_K: 0.3804\n", "Epoch 1706/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 65.5490 - calc_mre_K: 0.4001 - val_loss: 67.8387 - val_calc_mre_K: 0.4146\n", "Epoch 1707/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 65.4966 - calc_mre_K: 0.3998 - val_loss: 64.9305 - val_calc_mre_K: 0.3968\n", "Epoch 1708/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 66.0069 - calc_mre_K: 0.4029 - val_loss: 64.7667 - val_calc_mre_K: 0.3958\n", "Epoch 1709/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 64.9010 - calc_mre_K: 0.3961 - val_loss: 64.0543 - val_calc_mre_K: 0.3915\n", "Epoch 1710/2000\n", "48000/48000 [==============================] - 2s 49us/step - loss: 65.8296 - calc_mre_K: 0.4018 - val_loss: 63.9315 - val_calc_mre_K: 0.3907\n", "Epoch 1711/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 65.9844 - calc_mre_K: 0.4027 - val_loss: 65.4456 - val_calc_mre_K: 0.4000\n", "Epoch 1712/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 65.1561 - calc_mre_K: 0.3977 - val_loss: 65.3732 - val_calc_mre_K: 0.3995\n", "Epoch 1713/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 65.6040 - calc_mre_K: 0.4004 - val_loss: 64.4636 - val_calc_mre_K: 0.3940\n", "Epoch 1714/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 65.8342 - calc_mre_K: 0.4018 - val_loss: 64.1450 - val_calc_mre_K: 0.3920\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "Epoch 1715/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 65.5313 - calc_mre_K: 0.4000 - val_loss: 66.0216 - val_calc_mre_K: 0.4035\n", "Epoch 1716/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 66.5894 - calc_mre_K: 0.4064 - val_loss: 65.7539 - val_calc_mre_K: 0.4019\n", "Epoch 1717/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 65.2667 - calc_mre_K: 0.3984 - val_loss: 62.9366 - val_calc_mre_K: 0.3847\n", "Epoch 1718/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 65.9347 - calc_mre_K: 0.4024 - val_loss: 63.3227 - val_calc_mre_K: 0.3870\n", "Epoch 1719/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 65.6295 - calc_mre_K: 0.4006 - val_loss: 66.4028 - val_calc_mre_K: 0.4058\n", "Epoch 1720/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 65.4326 - calc_mre_K: 0.3994 - val_loss: 68.3214 - val_calc_mre_K: 0.4176\n", "Epoch 1721/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 65.7198 - calc_mre_K: 0.4011 - val_loss: 64.7704 - val_calc_mre_K: 0.3959\n", "Epoch 1722/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 65.7105 - calc_mre_K: 0.4011 - val_loss: 67.3920 - val_calc_mre_K: 0.4119\n", "Epoch 1723/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 65.3911 - calc_mre_K: 0.3991 - val_loss: 65.2184 - val_calc_mre_K: 0.3986\n", "Epoch 1724/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 65.3632 - calc_mre_K: 0.3989 - val_loss: 65.3099 - val_calc_mre_K: 0.3992\n", "Epoch 1725/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 65.0207 - calc_mre_K: 0.3969 - val_loss: 62.7400 - val_calc_mre_K: 0.3835\n", "Epoch 1726/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 65.4581 - calc_mre_K: 0.3995 - val_loss: 64.9064 - val_calc_mre_K: 0.3967\n", "Epoch 1727/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 65.8948 - calc_mre_K: 0.4022 - val_loss: 75.1485 - val_calc_mre_K: 0.4593\n", "Epoch 1728/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 65.5142 - calc_mre_K: 0.3999 - val_loss: 64.7961 - val_calc_mre_K: 0.3960\n", "Epoch 1729/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 66.1596 - calc_mre_K: 0.4038 - val_loss: 67.4126 - val_calc_mre_K: 0.4120\n", "Epoch 1730/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 65.1321 - calc_mre_K: 0.3975 - val_loss: 62.0029 - val_calc_mre_K: 0.3789\n", "Epoch 1731/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 64.9564 - calc_mre_K: 0.3965 - val_loss: 62.7184 - val_calc_mre_K: 0.3833\n", "Epoch 1732/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 65.5102 - calc_mre_K: 0.3998 - val_loss: 65.2900 - val_calc_mre_K: 0.3990\n", "Epoch 1733/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 65.1414 - calc_mre_K: 0.3976 - val_loss: 65.3051 - val_calc_mre_K: 0.3991\n", "Epoch 1734/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 65.0951 - calc_mre_K: 0.3973 - val_loss: 64.8582 - val_calc_mre_K: 0.3964\n", "Epoch 1735/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 65.8427 - calc_mre_K: 0.4019 - val_loss: 66.1390 - val_calc_mre_K: 0.4042\n", "Epoch 1736/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 65.2440 - calc_mre_K: 0.3982 - val_loss: 66.5277 - val_calc_mre_K: 0.4066\n", "Epoch 1737/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 65.7946 - calc_mre_K: 0.4016 - val_loss: 67.1235 - val_calc_mre_K: 0.4102\n", "Epoch 1738/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 65.2143 - calc_mre_K: 0.3980 - val_loss: 68.2756 - val_calc_mre_K: 0.4173\n", "Epoch 1739/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 65.7237 - calc_mre_K: 0.4011 - val_loss: 65.8377 - val_calc_mre_K: 0.4024\n", "Epoch 1740/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 65.4950 - calc_mre_K: 0.3997 - val_loss: 65.6662 - val_calc_mre_K: 0.4013\n", "Epoch 1741/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 65.6411 - calc_mre_K: 0.4006 - val_loss: 66.6351 - val_calc_mre_K: 0.4073\n", "Epoch 1742/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 65.1562 - calc_mre_K: 0.3977 - val_loss: 64.7529 - val_calc_mre_K: 0.3957\n", "Epoch 1743/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 65.4353 - calc_mre_K: 0.3994 - val_loss: 68.8150 - val_calc_mre_K: 0.4206\n", "Epoch 1744/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 66.0442 - calc_mre_K: 0.4031 - val_loss: 64.1250 - val_calc_mre_K: 0.3919\n", "Epoch 1745/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 65.4427 - calc_mre_K: 0.3994 - val_loss: 67.1995 - val_calc_mre_K: 0.4107\n", "Epoch 1746/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 64.9321 - calc_mre_K: 0.3963 - val_loss: 65.2787 - val_calc_mre_K: 0.3990\n", "Epoch 1747/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 65.6494 - calc_mre_K: 0.4007 - val_loss: 62.6272 - val_calc_mre_K: 0.3828\n", "Epoch 1748/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 65.4162 - calc_mre_K: 0.3993 - val_loss: 64.7356 - val_calc_mre_K: 0.3956\n", "Epoch 1749/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 64.6886 - calc_mre_K: 0.3948 - val_loss: 65.2767 - val_calc_mre_K: 0.3990\n", "Epoch 1750/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 65.8311 - calc_mre_K: 0.4018 - val_loss: 63.3068 - val_calc_mre_K: 0.3869\n", "Epoch 1751/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 65.7909 - calc_mre_K: 0.4016 - val_loss: 63.5502 - val_calc_mre_K: 0.3884\n", "Epoch 1752/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 65.3115 - calc_mre_K: 0.3986 - val_loss: 68.9958 - val_calc_mre_K: 0.4217\n", "Epoch 1753/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 65.5872 - calc_mre_K: 0.4003 - val_loss: 67.2742 - val_calc_mre_K: 0.4112\n", "Epoch 1754/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 65.9625 - calc_mre_K: 0.4026 - val_loss: 66.0956 - val_calc_mre_K: 0.4040\n", "Epoch 1755/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 65.6072 - calc_mre_K: 0.4004 - val_loss: 65.6789 - val_calc_mre_K: 0.4014\n", "Epoch 1756/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 65.1969 - calc_mre_K: 0.3979 - val_loss: 67.5341 - val_calc_mre_K: 0.4128\n", "Epoch 1757/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 66.0794 - calc_mre_K: 0.4033 - val_loss: 63.9441 - val_calc_mre_K: 0.3908\n", "Epoch 1758/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 64.9354 - calc_mre_K: 0.3963 - val_loss: 64.8174 - val_calc_mre_K: 0.3961\n", "Epoch 1759/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 65.7744 - calc_mre_K: 0.4015 - val_loss: 63.9099 - val_calc_mre_K: 0.3906\n", "Epoch 1760/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 64.9832 - calc_mre_K: 0.3966 - val_loss: 65.8996 - val_calc_mre_K: 0.4028\n", "Epoch 1761/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 66.0225 - calc_mre_K: 0.4030 - val_loss: 65.2655 - val_calc_mre_K: 0.3989\n", "Epoch 1762/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 64.4931 - calc_mre_K: 0.3936 - val_loss: 65.6778 - val_calc_mre_K: 0.4014\n", "Epoch 1763/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 65.6475 - calc_mre_K: 0.4007 - val_loss: 66.2912 - val_calc_mre_K: 0.4051\n", "Epoch 1764/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 65.4615 - calc_mre_K: 0.3995 - val_loss: 65.0992 - val_calc_mre_K: 0.3979\n", "Epoch 1765/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 65.5590 - calc_mre_K: 0.4001 - val_loss: 68.8651 - val_calc_mre_K: 0.4209\n", "Epoch 1766/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 65.6618 - calc_mre_K: 0.4008 - val_loss: 67.2830 - val_calc_mre_K: 0.4112\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "Epoch 1767/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 65.3526 - calc_mre_K: 0.3989 - val_loss: 64.9750 - val_calc_mre_K: 0.3971\n", "Epoch 1768/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 65.3750 - calc_mre_K: 0.3990 - val_loss: 65.4216 - val_calc_mre_K: 0.3998\n", "Epoch 1769/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 65.0710 - calc_mre_K: 0.3972 - val_loss: 63.2707 - val_calc_mre_K: 0.3867\n", "Epoch 1770/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 66.0215 - calc_mre_K: 0.4030 - val_loss: 66.6913 - val_calc_mre_K: 0.4076\n", "Epoch 1771/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 64.7841 - calc_mre_K: 0.3954 - val_loss: 65.7313 - val_calc_mre_K: 0.4017\n", "Epoch 1772/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 65.3673 - calc_mre_K: 0.3990 - val_loss: 66.4022 - val_calc_mre_K: 0.4058\n", "Epoch 1773/2000\n", "48000/48000 [==============================] - 3s 55us/step - loss: 65.6170 - calc_mre_K: 0.4005 - val_loss: 64.5896 - val_calc_mre_K: 0.3947\n", "Epoch 1774/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 64.8250 - calc_mre_K: 0.3957 - val_loss: 66.4323 - val_calc_mre_K: 0.4060\n", "Epoch 1775/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 65.3160 - calc_mre_K: 0.3987 - val_loss: 62.3366 - val_calc_mre_K: 0.3810\n", "Epoch 1776/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 65.6408 - calc_mre_K: 0.4006 - val_loss: 66.5367 - val_calc_mre_K: 0.4067\n", "Epoch 1777/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 65.3703 - calc_mre_K: 0.3990 - val_loss: 63.0461 - val_calc_mre_K: 0.3853\n", "Epoch 1778/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 65.4195 - calc_mre_K: 0.3993 - val_loss: 71.0105 - val_calc_mre_K: 0.4340\n", "Epoch 1779/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 65.4050 - calc_mre_K: 0.3992 - val_loss: 64.2355 - val_calc_mre_K: 0.3926\n", "Epoch 1780/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 65.3087 - calc_mre_K: 0.3986 - val_loss: 65.8743 - val_calc_mre_K: 0.4026\n", "Epoch 1781/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 65.6230 - calc_mre_K: 0.4005 - val_loss: 66.0363 - val_calc_mre_K: 0.4036\n", "Epoch 1782/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 64.9270 - calc_mre_K: 0.3963 - val_loss: 65.6749 - val_calc_mre_K: 0.4014\n", "Epoch 1783/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 65.6236 - calc_mre_K: 0.4005 - val_loss: 67.8491 - val_calc_mre_K: 0.4147\n", "Epoch 1784/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 65.3553 - calc_mre_K: 0.3989 - val_loss: 64.8675 - val_calc_mre_K: 0.3964\n", "Epoch 1785/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 65.5266 - calc_mre_K: 0.3999 - val_loss: 66.1349 - val_calc_mre_K: 0.4042\n", "Epoch 1786/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 65.3173 - calc_mre_K: 0.3987 - val_loss: 64.3195 - val_calc_mre_K: 0.3931\n", "Epoch 1787/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 65.2782 - calc_mre_K: 0.3984 - val_loss: 70.9320 - val_calc_mre_K: 0.4335\n", "Epoch 1788/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 65.4551 - calc_mre_K: 0.3995 - val_loss: 63.6540 - val_calc_mre_K: 0.3890\n", "Epoch 1789/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 65.3116 - calc_mre_K: 0.3986 - val_loss: 63.8346 - val_calc_mre_K: 0.3901\n", "Epoch 1790/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 65.0524 - calc_mre_K: 0.3970 - val_loss: 66.3747 - val_calc_mre_K: 0.4057\n", "Epoch 1791/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 65.2704 - calc_mre_K: 0.3984 - val_loss: 63.5462 - val_calc_mre_K: 0.3884\n", "Epoch 1792/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 65.0513 - calc_mre_K: 0.3970 - val_loss: 63.6379 - val_calc_mre_K: 0.3889\n", "Epoch 1793/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 64.8824 - calc_mre_K: 0.3960 - val_loss: 62.5395 - val_calc_mre_K: 0.3822\n", "Epoch 1794/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 65.6583 - calc_mre_K: 0.4007 - val_loss: 65.2933 - val_calc_mre_K: 0.3991\n", "Epoch 1795/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 65.1575 - calc_mre_K: 0.3977 - val_loss: 64.3303 - val_calc_mre_K: 0.3932\n", "Epoch 1796/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 65.4660 - calc_mre_K: 0.3996 - val_loss: 75.6097 - val_calc_mre_K: 0.4621\n", "Epoch 1797/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 65.6283 - calc_mre_K: 0.4006 - val_loss: 67.9302 - val_calc_mre_K: 0.4152\n", "Epoch 1798/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 64.9007 - calc_mre_K: 0.3961 - val_loss: 67.3290 - val_calc_mre_K: 0.4115\n", "Epoch 1799/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 65.4382 - calc_mre_K: 0.3994 - val_loss: 66.5502 - val_calc_mre_K: 0.4067\n", "Epoch 1800/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 65.2441 - calc_mre_K: 0.3982 - val_loss: 66.3230 - val_calc_mre_K: 0.4054\n", "Epoch 1801/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 65.3794 - calc_mre_K: 0.3990 - val_loss: 66.7303 - val_calc_mre_K: 0.4078\n", "Epoch 1802/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 65.2494 - calc_mre_K: 0.3983 - val_loss: 62.9193 - val_calc_mre_K: 0.3845\n", "Epoch 1803/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 65.1313 - calc_mre_K: 0.3975 - val_loss: 65.1127 - val_calc_mre_K: 0.3980\n", "Epoch 1804/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 65.2504 - calc_mre_K: 0.3983 - val_loss: 67.5573 - val_calc_mre_K: 0.4129\n", "Epoch 1805/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 64.9017 - calc_mre_K: 0.3961 - val_loss: 64.2626 - val_calc_mre_K: 0.3928\n", "Epoch 1806/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 65.6567 - calc_mre_K: 0.4007 - val_loss: 66.4045 - val_calc_mre_K: 0.4058\n", "Epoch 1807/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 64.7047 - calc_mre_K: 0.3949 - val_loss: 65.1310 - val_calc_mre_K: 0.3981\n", "Epoch 1808/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 65.5698 - calc_mre_K: 0.4002 - val_loss: 65.6130 - val_calc_mre_K: 0.4010\n", "Epoch 1809/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 65.0199 - calc_mre_K: 0.3968 - val_loss: 66.0527 - val_calc_mre_K: 0.4037\n", "Epoch 1810/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 64.9476 - calc_mre_K: 0.3964 - val_loss: 62.9842 - val_calc_mre_K: 0.3849\n", "Epoch 1811/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 65.0459 - calc_mre_K: 0.3970 - val_loss: 70.6080 - val_calc_mre_K: 0.4315\n", "Epoch 1812/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 65.3638 - calc_mre_K: 0.3989 - val_loss: 73.3858 - val_calc_mre_K: 0.4485\n", "Epoch 1813/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 65.0092 - calc_mre_K: 0.3968 - val_loss: 66.6587 - val_calc_mre_K: 0.4074\n", "Epoch 1814/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 65.1173 - calc_mre_K: 0.3974 - val_loss: 63.9096 - val_calc_mre_K: 0.3906\n", "Epoch 1815/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 65.8196 - calc_mre_K: 0.4017 - val_loss: 63.9010 - val_calc_mre_K: 0.3905\n", "Epoch 1816/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 64.8591 - calc_mre_K: 0.3959 - val_loss: 72.4478 - val_calc_mre_K: 0.4428\n", "Epoch 1817/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 65.5923 - calc_mre_K: 0.4003 - val_loss: 64.8627 - val_calc_mre_K: 0.3964\n", "Epoch 1818/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 64.9236 - calc_mre_K: 0.3963 - val_loss: 61.9262 - val_calc_mre_K: 0.3785\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "Epoch 1819/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 65.4976 - calc_mre_K: 0.3998 - val_loss: 65.1999 - val_calc_mre_K: 0.3985\n", "Epoch 1820/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 65.1849 - calc_mre_K: 0.3979 - val_loss: 62.6498 - val_calc_mre_K: 0.3829\n", "Epoch 1821/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 65.3825 - calc_mre_K: 0.3991 - val_loss: 64.0559 - val_calc_mre_K: 0.3915\n", "Epoch 1822/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 64.9981 - calc_mre_K: 0.3967 - val_loss: 66.0733 - val_calc_mre_K: 0.4038\n", "Epoch 1823/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 65.4073 - calc_mre_K: 0.3992 - val_loss: 66.9652 - val_calc_mre_K: 0.4093\n", "Epoch 1824/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 64.8740 - calc_mre_K: 0.3960 - val_loss: 63.1894 - val_calc_mre_K: 0.3862\n", "Epoch 1825/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 65.9161 - calc_mre_K: 0.4023 - val_loss: 70.0321 - val_calc_mre_K: 0.4280\n", "Epoch 1826/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 65.0822 - calc_mre_K: 0.3972 - val_loss: 65.9426 - val_calc_mre_K: 0.4030\n", "Epoch 1827/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 65.8584 - calc_mre_K: 0.4020 - val_loss: 67.8978 - val_calc_mre_K: 0.4150\n", "Epoch 1828/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 64.4239 - calc_mre_K: 0.3932 - val_loss: 65.7707 - val_calc_mre_K: 0.4020\n", "Epoch 1829/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 66.2018 - calc_mre_K: 0.4041 - val_loss: 65.0069 - val_calc_mre_K: 0.3973\n", "Epoch 1830/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 65.2585 - calc_mre_K: 0.3983 - val_loss: 63.7787 - val_calc_mre_K: 0.3898\n", "Epoch 1831/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 64.9164 - calc_mre_K: 0.3962 - val_loss: 64.7679 - val_calc_mre_K: 0.3958\n", "Epoch 1832/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 64.9993 - calc_mre_K: 0.3967 - val_loss: 64.9675 - val_calc_mre_K: 0.3971\n", "Epoch 1833/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 65.3586 - calc_mre_K: 0.3989 - val_loss: 66.2865 - val_calc_mre_K: 0.4051\n", "Epoch 1834/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 64.9113 - calc_mre_K: 0.3962 - val_loss: 69.6469 - val_calc_mre_K: 0.4257\n", "Epoch 1835/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 65.2954 - calc_mre_K: 0.3985 - val_loss: 62.8358 - val_calc_mre_K: 0.3840\n", "Epoch 1836/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 64.3970 - calc_mre_K: 0.3930 - val_loss: 62.7279 - val_calc_mre_K: 0.3834\n", "Epoch 1837/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 65.1833 - calc_mre_K: 0.3978 - val_loss: 64.9928 - val_calc_mre_K: 0.3972\n", "Epoch 1838/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 65.2946 - calc_mre_K: 0.3985 - val_loss: 66.3288 - val_calc_mre_K: 0.4054\n", "Epoch 1839/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 64.8257 - calc_mre_K: 0.3957 - val_loss: 65.8338 - val_calc_mre_K: 0.4024\n", "Epoch 1840/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 65.6210 - calc_mre_K: 0.4005 - val_loss: 68.3718 - val_calc_mre_K: 0.4179\n", "Epoch 1841/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 65.0505 - calc_mre_K: 0.3970 - val_loss: 64.8413 - val_calc_mre_K: 0.3963\n", "Epoch 1842/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 65.2240 - calc_mre_K: 0.3981 - val_loss: 64.0453 - val_calc_mre_K: 0.3914\n", "Epoch 1843/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 64.6483 - calc_mre_K: 0.3946 - val_loss: 64.8229 - val_calc_mre_K: 0.3962\n", "Epoch 1844/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 65.3242 - calc_mre_K: 0.3987 - val_loss: 64.8705 - val_calc_mre_K: 0.3965\n", "Epoch 1845/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 65.3779 - calc_mre_K: 0.3990 - val_loss: 66.6347 - val_calc_mre_K: 0.4073\n", "Epoch 1846/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 64.5241 - calc_mre_K: 0.3938 - val_loss: 61.8594 - val_calc_mre_K: 0.3781\n", "Epoch 1847/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 65.2874 - calc_mre_K: 0.3985 - val_loss: 65.8339 - val_calc_mre_K: 0.4024\n", "Epoch 1848/2000\n", "48000/48000 [==============================] - 2s 49us/step - loss: 65.7353 - calc_mre_K: 0.4012 - val_loss: 67.3898 - val_calc_mre_K: 0.4119\n", "Epoch 1849/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 64.9531 - calc_mre_K: 0.3964 - val_loss: 65.8642 - val_calc_mre_K: 0.4025\n", "Epoch 1850/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 65.2534 - calc_mre_K: 0.3983 - val_loss: 63.0357 - val_calc_mre_K: 0.3853\n", "Epoch 1851/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 64.9506 - calc_mre_K: 0.3964 - val_loss: 63.4088 - val_calc_mre_K: 0.3875\n", "Epoch 1852/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 65.0124 - calc_mre_K: 0.3968 - val_loss: 64.1943 - val_calc_mre_K: 0.3924\n", "Epoch 1853/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 65.0181 - calc_mre_K: 0.3968 - val_loss: 65.1308 - val_calc_mre_K: 0.3981\n", "Epoch 1854/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 65.2177 - calc_mre_K: 0.3981 - val_loss: 66.7847 - val_calc_mre_K: 0.4082\n", "Epoch 1855/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 65.2559 - calc_mre_K: 0.3983 - val_loss: 64.7500 - val_calc_mre_K: 0.3957\n", "Epoch 1856/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 65.3886 - calc_mre_K: 0.3991 - val_loss: 69.4873 - val_calc_mre_K: 0.4247\n", "Epoch 1857/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 64.3884 - calc_mre_K: 0.3930 - val_loss: 63.4921 - val_calc_mre_K: 0.3881\n", "Epoch 1858/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 65.1726 - calc_mre_K: 0.3978 - val_loss: 63.8392 - val_calc_mre_K: 0.3902\n", "Epoch 1859/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 65.4354 - calc_mre_K: 0.3994 - val_loss: 64.5430 - val_calc_mre_K: 0.3945\n", "Epoch 1860/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 65.2336 - calc_mre_K: 0.3982 - val_loss: 66.3286 - val_calc_mre_K: 0.4054\n", "Epoch 1861/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 65.5464 - calc_mre_K: 0.4001 - val_loss: 62.0588 - val_calc_mre_K: 0.3793\n", "Epoch 1862/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 65.7472 - calc_mre_K: 0.4013 - val_loss: 67.3519 - val_calc_mre_K: 0.4116\n", "Epoch 1863/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 64.9428 - calc_mre_K: 0.3964 - val_loss: 73.5142 - val_calc_mre_K: 0.4493\n", "Epoch 1864/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 64.8838 - calc_mre_K: 0.3960 - val_loss: 69.5179 - val_calc_mre_K: 0.4248\n", "Epoch 1865/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 64.8004 - calc_mre_K: 0.3955 - val_loss: 74.9216 - val_calc_mre_K: 0.4579\n", "Epoch 1866/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 65.0532 - calc_mre_K: 0.3971 - val_loss: 63.2827 - val_calc_mre_K: 0.3868\n", "Epoch 1867/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 64.7778 - calc_mre_K: 0.3954 - val_loss: 73.2480 - val_calc_mre_K: 0.4477\n", "Epoch 1868/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 65.4006 - calc_mre_K: 0.3992 - val_loss: 61.5852 - val_calc_mre_K: 0.3764\n", "Epoch 1869/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 64.7789 - calc_mre_K: 0.3954 - val_loss: 64.0765 - val_calc_mre_K: 0.3916\n", "Epoch 1870/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 64.7400 - calc_mre_K: 0.3951 - val_loss: 65.5086 - val_calc_mre_K: 0.4004\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "Epoch 1871/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 65.0975 - calc_mre_K: 0.3973 - val_loss: 67.8927 - val_calc_mre_K: 0.4149\n", "Epoch 1872/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 64.5137 - calc_mre_K: 0.3938 - val_loss: 64.8116 - val_calc_mre_K: 0.3961\n", "Epoch 1873/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 65.0733 - calc_mre_K: 0.3972 - val_loss: 66.1619 - val_calc_mre_K: 0.4044\n", "Epoch 1874/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 64.9034 - calc_mre_K: 0.3961 - val_loss: 68.7757 - val_calc_mre_K: 0.4203\n", "Epoch 1875/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 65.1324 - calc_mre_K: 0.3975 - val_loss: 63.1492 - val_calc_mre_K: 0.3859\n", "Epoch 1876/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 65.3669 - calc_mre_K: 0.3990 - val_loss: 67.6134 - val_calc_mre_K: 0.4132\n", "Epoch 1877/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 64.9768 - calc_mre_K: 0.3966 - val_loss: 65.4588 - val_calc_mre_K: 0.4001\n", "Epoch 1878/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 65.1637 - calc_mre_K: 0.3977 - val_loss: 66.9839 - val_calc_mre_K: 0.4094\n", "Epoch 1879/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 64.6763 - calc_mre_K: 0.3948 - val_loss: 66.3023 - val_calc_mre_K: 0.4052\n", "Epoch 1880/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 64.9628 - calc_mre_K: 0.3965 - val_loss: 62.8743 - val_calc_mre_K: 0.3843\n", "Epoch 1881/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 65.1161 - calc_mre_K: 0.3974 - val_loss: 64.8523 - val_calc_mre_K: 0.3964\n", "Epoch 1882/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 65.5391 - calc_mre_K: 0.4000 - val_loss: 69.9100 - val_calc_mre_K: 0.4273\n", "Epoch 1883/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 64.8749 - calc_mre_K: 0.3960 - val_loss: 64.3897 - val_calc_mre_K: 0.3935\n", "Epoch 1884/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 64.5195 - calc_mre_K: 0.3938 - val_loss: 66.3085 - val_calc_mre_K: 0.4053\n", "Epoch 1885/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 64.7510 - calc_mre_K: 0.3952 - val_loss: 66.6967 - val_calc_mre_K: 0.4076\n", "Epoch 1886/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 65.2411 - calc_mre_K: 0.3982 - val_loss: 65.7075 - val_calc_mre_K: 0.4016\n", "Epoch 1887/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 64.8556 - calc_mre_K: 0.3958 - val_loss: 64.2851 - val_calc_mre_K: 0.3929\n", "Epoch 1888/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 64.7938 - calc_mre_K: 0.3955 - val_loss: 66.2214 - val_calc_mre_K: 0.4047\n", "Epoch 1889/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 65.0578 - calc_mre_K: 0.3971 - val_loss: 65.1529 - val_calc_mre_K: 0.3982\n", "Epoch 1890/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 65.0645 - calc_mre_K: 0.3971 - val_loss: 71.0470 - val_calc_mre_K: 0.4342\n", "Epoch 1891/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 65.8372 - calc_mre_K: 0.4018 - val_loss: 63.8396 - val_calc_mre_K: 0.3902\n", "Epoch 1892/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 64.7578 - calc_mre_K: 0.3953 - val_loss: 65.7221 - val_calc_mre_K: 0.4017\n", "Epoch 1893/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 64.5486 - calc_mre_K: 0.3940 - val_loss: 65.7659 - val_calc_mre_K: 0.4019\n", "Epoch 1894/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 64.9147 - calc_mre_K: 0.3962 - val_loss: 66.9702 - val_calc_mre_K: 0.4093\n", "Epoch 1895/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 65.2507 - calc_mre_K: 0.3983 - val_loss: 64.9685 - val_calc_mre_K: 0.3971\n", "Epoch 1896/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 65.1339 - calc_mre_K: 0.3975 - val_loss: 65.1370 - val_calc_mre_K: 0.3981\n", "Epoch 1897/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 64.9827 - calc_mre_K: 0.3966 - val_loss: 63.9733 - val_calc_mre_K: 0.3910\n", "Epoch 1898/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 64.7171 - calc_mre_K: 0.3950 - val_loss: 65.1375 - val_calc_mre_K: 0.3981\n", "Epoch 1899/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 65.2803 - calc_mre_K: 0.3984 - val_loss: 64.4070 - val_calc_mre_K: 0.3936\n", "Epoch 1900/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 64.7418 - calc_mre_K: 0.3952 - val_loss: 66.9222 - val_calc_mre_K: 0.4090\n", "Epoch 1901/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 64.9627 - calc_mre_K: 0.3965 - val_loss: 66.8911 - val_calc_mre_K: 0.4088\n", "Epoch 1902/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 65.2603 - calc_mre_K: 0.3983 - val_loss: 65.8420 - val_calc_mre_K: 0.4024\n", "Epoch 1903/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 64.9010 - calc_mre_K: 0.3961 - val_loss: 66.6023 - val_calc_mre_K: 0.4071\n", "Epoch 1904/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 64.9385 - calc_mre_K: 0.3964 - val_loss: 65.6966 - val_calc_mre_K: 0.4015\n", "Epoch 1905/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 65.6294 - calc_mre_K: 0.4006 - val_loss: 64.9375 - val_calc_mre_K: 0.3969\n", "Epoch 1906/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 63.9322 - calc_mre_K: 0.3902 - val_loss: 65.8387 - val_calc_mre_K: 0.4024\n", "Epoch 1907/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 65.5111 - calc_mre_K: 0.3998 - val_loss: 65.0614 - val_calc_mre_K: 0.3976\n", "Epoch 1908/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 64.7108 - calc_mre_K: 0.3950 - val_loss: 63.7603 - val_calc_mre_K: 0.3897\n", "Epoch 1909/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 65.4194 - calc_mre_K: 0.3993 - val_loss: 64.6762 - val_calc_mre_K: 0.3953\n", "Epoch 1910/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 64.4385 - calc_mre_K: 0.3933 - val_loss: 63.6611 - val_calc_mre_K: 0.3891\n", "Epoch 1911/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 65.1881 - calc_mre_K: 0.3979 - val_loss: 65.5279 - val_calc_mre_K: 0.4005\n", "Epoch 1912/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 64.8110 - calc_mre_K: 0.3956 - val_loss: 62.6723 - val_calc_mre_K: 0.3830\n", "Epoch 1913/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 65.0372 - calc_mre_K: 0.3970 - val_loss: 63.3562 - val_calc_mre_K: 0.3872\n", "Epoch 1914/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 64.1257 - calc_mre_K: 0.3914 - val_loss: 63.7928 - val_calc_mre_K: 0.3899\n", "Epoch 1915/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 64.7306 - calc_mre_K: 0.3951 - val_loss: 64.2633 - val_calc_mre_K: 0.3928\n", "Epoch 1916/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 64.5228 - calc_mre_K: 0.3938 - val_loss: 62.2852 - val_calc_mre_K: 0.3807\n", "Epoch 1917/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 65.1749 - calc_mre_K: 0.3978 - val_loss: 67.3684 - val_calc_mre_K: 0.4117\n", "Epoch 1918/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 64.4574 - calc_mre_K: 0.3934 - val_loss: 67.7467 - val_calc_mre_K: 0.4141\n", "Epoch 1919/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 64.7685 - calc_mre_K: 0.3953 - val_loss: 66.7331 - val_calc_mre_K: 0.4078\n", "Epoch 1920/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 64.8270 - calc_mre_K: 0.3957 - val_loss: 64.6652 - val_calc_mre_K: 0.3952\n", "Epoch 1921/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 65.3846 - calc_mre_K: 0.3991 - val_loss: 66.0281 - val_calc_mre_K: 0.4035\n", "Epoch 1922/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 64.3930 - calc_mre_K: 0.3930 - val_loss: 68.2195 - val_calc_mre_K: 0.4169\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "Epoch 1923/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 65.0306 - calc_mre_K: 0.3969 - val_loss: 63.0897 - val_calc_mre_K: 0.3856\n", "Epoch 1924/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 64.4102 - calc_mre_K: 0.3931 - val_loss: 65.1242 - val_calc_mre_K: 0.3980\n", "Epoch 1925/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 64.8245 - calc_mre_K: 0.3957 - val_loss: 68.3922 - val_calc_mre_K: 0.4180\n", "Epoch 1926/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 64.9380 - calc_mre_K: 0.3963 - val_loss: 64.2983 - val_calc_mre_K: 0.3930\n", "Epoch 1927/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 64.8977 - calc_mre_K: 0.3961 - val_loss: 64.0972 - val_calc_mre_K: 0.3917\n", "Epoch 1928/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 64.4850 - calc_mre_K: 0.3936 - val_loss: 64.8519 - val_calc_mre_K: 0.3964\n", "Epoch 1929/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 64.7448 - calc_mre_K: 0.3952 - val_loss: 71.0334 - val_calc_mre_K: 0.4341\n", "Epoch 1930/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 65.0503 - calc_mre_K: 0.3970 - val_loss: 65.9256 - val_calc_mre_K: 0.4029\n", "Epoch 1931/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 64.8241 - calc_mre_K: 0.3957 - val_loss: 63.8655 - val_calc_mre_K: 0.3903\n", "Epoch 1932/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 64.7254 - calc_mre_K: 0.3951 - val_loss: 68.5311 - val_calc_mre_K: 0.4188\n", "Epoch 1933/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 65.2509 - calc_mre_K: 0.3983 - val_loss: 65.3597 - val_calc_mre_K: 0.3995\n", "Epoch 1934/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 65.0490 - calc_mre_K: 0.3970 - val_loss: 67.4809 - val_calc_mre_K: 0.4124\n", "Epoch 1935/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 64.5290 - calc_mre_K: 0.3939 - val_loss: 61.9407 - val_calc_mre_K: 0.3786\n", "Epoch 1936/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 64.7466 - calc_mre_K: 0.3952 - val_loss: 65.7295 - val_calc_mre_K: 0.4017\n", "Epoch 1937/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 64.7570 - calc_mre_K: 0.3952 - val_loss: 65.2822 - val_calc_mre_K: 0.3990\n", "Epoch 1938/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 64.2429 - calc_mre_K: 0.3921 - val_loss: 62.1068 - val_calc_mre_K: 0.3796\n", "Epoch 1939/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 65.0165 - calc_mre_K: 0.3968 - val_loss: 65.7231 - val_calc_mre_K: 0.4017\n", "Epoch 1940/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 65.4748 - calc_mre_K: 0.3996 - val_loss: 68.8741 - val_calc_mre_K: 0.4209\n", "Epoch 1941/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 64.3331 - calc_mre_K: 0.3927 - val_loss: 64.8883 - val_calc_mre_K: 0.3966\n", "Epoch 1942/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 64.2869 - calc_mre_K: 0.3924 - val_loss: 64.2304 - val_calc_mre_K: 0.3926\n", "Epoch 1943/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 65.0164 - calc_mre_K: 0.3968 - val_loss: 64.4897 - val_calc_mre_K: 0.3942\n", "Epoch 1944/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 64.8040 - calc_mre_K: 0.3955 - val_loss: 65.9034 - val_calc_mre_K: 0.4028\n", "Epoch 1945/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 64.6577 - calc_mre_K: 0.3946 - val_loss: 72.7664 - val_calc_mre_K: 0.4447\n", "Epoch 1946/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 66.0083 - calc_mre_K: 0.4029 - val_loss: 66.1953 - val_calc_mre_K: 0.4046\n", "Epoch 1947/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 64.3194 - calc_mre_K: 0.3926 - val_loss: 61.9385 - val_calc_mre_K: 0.3785\n", "Epoch 1948/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 64.2186 - calc_mre_K: 0.3920 - val_loss: 64.2978 - val_calc_mre_K: 0.3930\n", "Epoch 1949/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 64.8634 - calc_mre_K: 0.3959 - val_loss: 65.6327 - val_calc_mre_K: 0.4011\n", "Epoch 1950/2000\n", "48000/48000 [==============================] - 2s 50us/step - loss: 65.3391 - calc_mre_K: 0.3988 - val_loss: 64.3174 - val_calc_mre_K: 0.3931\n", "Epoch 1951/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 65.4707 - calc_mre_K: 0.3996 - val_loss: 67.6212 - val_calc_mre_K: 0.4133\n", "Epoch 1952/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 64.4824 - calc_mre_K: 0.3936 - val_loss: 64.5926 - val_calc_mre_K: 0.3948\n", "Epoch 1953/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 64.3030 - calc_mre_K: 0.3925 - val_loss: 62.1645 - val_calc_mre_K: 0.3799\n", "Epoch 1954/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 65.0125 - calc_mre_K: 0.3968 - val_loss: 63.3230 - val_calc_mre_K: 0.3870\n", "Epoch 1955/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 64.6999 - calc_mre_K: 0.3949 - val_loss: 65.5683 - val_calc_mre_K: 0.4007\n", "Epoch 1956/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 64.6980 - calc_mre_K: 0.3949 - val_loss: 64.1663 - val_calc_mre_K: 0.3922\n", "Epoch 1957/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 64.9735 - calc_mre_K: 0.3966 - val_loss: 71.1550 - val_calc_mre_K: 0.4349\n", "Epoch 1958/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 64.0486 - calc_mre_K: 0.3909 - val_loss: 64.2085 - val_calc_mre_K: 0.3924\n", "Epoch 1959/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 65.2468 - calc_mre_K: 0.3982 - val_loss: 67.4153 - val_calc_mre_K: 0.4120\n", "Epoch 1960/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 64.4559 - calc_mre_K: 0.3934 - val_loss: 66.0387 - val_calc_mre_K: 0.4036\n", "Epoch 1961/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 64.5490 - calc_mre_K: 0.3940 - val_loss: 64.2007 - val_calc_mre_K: 0.3924\n", "Epoch 1962/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 64.6010 - calc_mre_K: 0.3943 - val_loss: 62.8507 - val_calc_mre_K: 0.3841\n", "Epoch 1963/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 64.6335 - calc_mre_K: 0.3945 - val_loss: 65.1943 - val_calc_mre_K: 0.3985\n", "Epoch 1964/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 65.1797 - calc_mre_K: 0.3978 - val_loss: 65.1413 - val_calc_mre_K: 0.3981\n", "Epoch 1965/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 64.5230 - calc_mre_K: 0.3938 - val_loss: 65.7411 - val_calc_mre_K: 0.4018\n", "Epoch 1966/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 65.2894 - calc_mre_K: 0.3985 - val_loss: 63.6479 - val_calc_mre_K: 0.3890\n", "Epoch 1967/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 64.9057 - calc_mre_K: 0.3962 - val_loss: 63.7689 - val_calc_mre_K: 0.3897\n", "Epoch 1968/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 64.5067 - calc_mre_K: 0.3937 - val_loss: 64.7211 - val_calc_mre_K: 0.3956\n", "Epoch 1969/2000\n", "48000/48000 [==============================] - 3s 55us/step - loss: 64.3319 - calc_mre_K: 0.3927 - val_loss: 64.4382 - val_calc_mre_K: 0.3938\n", "Epoch 1970/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 64.4013 - calc_mre_K: 0.3931 - val_loss: 62.8073 - val_calc_mre_K: 0.3839\n", "Epoch 1971/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 64.6310 - calc_mre_K: 0.3945 - val_loss: 63.4368 - val_calc_mre_K: 0.3877\n", "Epoch 1972/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 64.6530 - calc_mre_K: 0.3946 - val_loss: 60.1953 - val_calc_mre_K: 0.3679\n", "Epoch 1973/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 64.0654 - calc_mre_K: 0.3910 - val_loss: 64.0274 - val_calc_mre_K: 0.3913\n", "Epoch 1974/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 64.5154 - calc_mre_K: 0.3938 - val_loss: 68.3069 - val_calc_mre_K: 0.4175\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "Epoch 1975/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 65.6875 - calc_mre_K: 0.4009 - val_loss: 63.4805 - val_calc_mre_K: 0.3880\n", "Epoch 1976/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 64.1815 - calc_mre_K: 0.3917 - val_loss: 62.0555 - val_calc_mre_K: 0.3793\n", "Epoch 1977/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 64.7846 - calc_mre_K: 0.3954 - val_loss: 63.6942 - val_calc_mre_K: 0.3893\n", "Epoch 1978/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 64.6462 - calc_mre_K: 0.3946 - val_loss: 62.5292 - val_calc_mre_K: 0.3822\n", "Epoch 1979/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 64.2272 - calc_mre_K: 0.3920 - val_loss: 61.9759 - val_calc_mre_K: 0.3788\n", "Epoch 1980/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 64.6669 - calc_mre_K: 0.3947 - val_loss: 65.1547 - val_calc_mre_K: 0.3982\n", "Epoch 1981/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 64.3271 - calc_mre_K: 0.3926 - val_loss: 65.4312 - val_calc_mre_K: 0.3999\n", "Epoch 1982/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 65.0594 - calc_mre_K: 0.3971 - val_loss: 64.8108 - val_calc_mre_K: 0.3961\n", "Epoch 1983/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 64.4500 - calc_mre_K: 0.3934 - val_loss: 64.2908 - val_calc_mre_K: 0.3929\n", "Epoch 1984/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 64.7859 - calc_mre_K: 0.3954 - val_loss: 62.0633 - val_calc_mre_K: 0.3793\n", "Epoch 1985/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 64.4451 - calc_mre_K: 0.3933 - val_loss: 64.7696 - val_calc_mre_K: 0.3959\n", "Epoch 1986/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 64.7181 - calc_mre_K: 0.3950 - val_loss: 64.2110 - val_calc_mre_K: 0.3924\n", "Epoch 1987/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 64.5921 - calc_mre_K: 0.3942 - val_loss: 65.9276 - val_calc_mre_K: 0.4029\n", "Epoch 1988/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 64.7386 - calc_mre_K: 0.3951 - val_loss: 63.1569 - val_calc_mre_K: 0.3860\n", "Epoch 1989/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 65.2469 - calc_mre_K: 0.3982 - val_loss: 62.9826 - val_calc_mre_K: 0.3849\n", "Epoch 1990/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 64.6962 - calc_mre_K: 0.3949 - val_loss: 63.5048 - val_calc_mre_K: 0.3881\n", "Epoch 1991/2000\n", "48000/48000 [==============================] - 2s 52us/step - loss: 64.3860 - calc_mre_K: 0.3930 - val_loss: 64.1145 - val_calc_mre_K: 0.3918\n", "Epoch 1992/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 64.4244 - calc_mre_K: 0.3932 - val_loss: 67.0799 - val_calc_mre_K: 0.4100\n", "Epoch 1993/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 64.5164 - calc_mre_K: 0.3938 - val_loss: 64.1491 - val_calc_mre_K: 0.3921\n", "Epoch 1994/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 64.7775 - calc_mre_K: 0.3954 - val_loss: 67.9672 - val_calc_mre_K: 0.4154\n", "Epoch 1995/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 64.9730 - calc_mre_K: 0.3966 - val_loss: 64.6774 - val_calc_mre_K: 0.3953\n", "Epoch 1996/2000\n", "48000/48000 [==============================] - 3s 54us/step - loss: 64.1815 - calc_mre_K: 0.3917 - val_loss: 66.2016 - val_calc_mre_K: 0.4046\n", "Epoch 1997/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 64.5767 - calc_mre_K: 0.3941 - val_loss: 73.3735 - val_calc_mre_K: 0.4484\n", "Epoch 1998/2000\n", "48000/48000 [==============================] - 3s 52us/step - loss: 65.0487 - calc_mre_K: 0.3970 - val_loss: 65.1389 - val_calc_mre_K: 0.3981\n", "Epoch 1999/2000\n", "48000/48000 [==============================] - 3s 53us/step - loss: 64.3865 - calc_mre_K: 0.3930 - val_loss: 64.6418 - val_calc_mre_K: 0.3951\n", "Epoch 2000/2000\n", "48000/48000 [==============================] - 2s 51us/step - loss: 64.7916 - calc_mre_K: 0.3955 - val_loss: 63.1296 - val_calc_mre_K: 0.3858\n" ] } ], "source": [ "%autoreload\n", "# import warnings\n", "# warnings.filterwarnings('ignore')\n", "\n", "\n", "# model = resnetb()\n", "# #model = conv1d_lkyrelu()\n", "# #model = conv1d_model_bnorm()\n", "# #model = conv1d_model(1)\n", "\n", "# #model = naiveploss_model()\n", "# model.summary()\n", "\n", " \n", "# from IPython.display import SVG\n", "# from keras.utils.vis_utils import model_to_dot\n", "\n", "# #SVG(model_to_dot(model).create(prog='dot', format='svg'))\n", " \n", "\n", "\n", " \n", " \n", "import scnets as scn\n", "\n", "# model = scn.resnet(in_size=8, \n", "# out_size=256,\n", "# num_units=3,\n", "# red_dim=16,\n", "# batch_size=64,\n", "# ker_size=3)\n", "\n", "# model = scn.conv1dmodel(in_size=8, \n", "# out_size=256,\n", "# batch_size=64,\n", "# c1_nf=64,\n", "# clayers=4,\n", "# ker_size=5)\n", "\n", "model = scn.convprel(in_size=8, \n", " out_size=256,\n", " batch_size=64,\n", " c1_nf=64,\n", " clayers=4,\n", " ker_size=3)\n", "\n", "# model = scn.fullycon(in_size=8, \n", "# out_size=256, \n", "# batch_size=64,\n", "# N_hidden=2, \n", "# N_neurons=256, \n", "# N_gpus=1)\n", "\n", "\n", "\n", "\n", "\n", "# # from keras import optimizers\n", "# # sgd = optimizers.SGD(lr=0.01, decay=1e-6, momentum=0.9, nesterov=True)\n", "# #model.compile(loss=naive_percent_loss, optimizer='nadam', metrics=[calc_mre_K])\n", "\n", "# from IPython.display import SVG\n", "# from keras.utils.vis_utils import model_to_dot\n", "# from keras.utils.vis_utils import plot_model\n", "\n", "# SVG(model_to_dot(model, show_shapes=False, show_layer_names=False).create(prog='dot', format='svg'))\n", "\n", "\n", "# #plot_model(model, show_shapes=False, show_layer_names=False, to_file='model_resnet.svg')\n", "\n", "\n", "\n", "model.summary() \n", " \n", " \n", "x_t, x_v, y_t, y_v = train_test_split(x_train, y_train, test_size=0.2, random_state=42)\n", "# model = naiveploss_mgpu_model()\n", "# model.summary() \n", "history = model.fit(x_t, y_t,\n", " batch_size=64,\n", " epochs=2000, \n", " verbose=1,\n", " validation_data=(x_v, y_v))\n", "\n", "\n", "\n", "\n", "\n", "# y_pred = model.predict(x_test)\n", "# print(calc_mre(y_test, y_pred))\n", "\n" ] }, { "cell_type": "code", "execution_count": 4, "metadata": { "ExecuteTime": { "end_time": "2018-09-30T16:23:59.012349Z", "start_time": "2018-09-30T16:23:58.711234Z" } }, "outputs": [ { "name": "stderr", "output_type": "stream", "text": [ "/home/hegder/anaconda3/envs/dp2/lib/python3.5/site-packages/keras/backend/mxnet_backend.py:89: UserWarning: MXNet Backend performs best with `channels_first` format. Using `channels_last` will significantly reduce performance due to the Transpose operations. For performance improvement, please use this API`keras.utils.to_channels_first(x_input)`to transform `channels_last` data to `channels_first` format and also please change the `image_data_format` in `keras.json` to `channels_first`.Note: `x_input` is a Numpy tensor or a list of Numpy tensorRefer to: https://github.com/awslabs/keras-apache-mxnet/tree/master/docs/mxnet_backend/performance_guide.md\n", " train_symbol = func(*args, **kwargs)\n", "/home/hegder/anaconda3/envs/dp2/lib/python3.5/site-packages/keras/backend/mxnet_backend.py:92: UserWarning: MXNet Backend performs best with `channels_first` format. Using `channels_last` will significantly reduce performance due to the Transpose operations. For performance improvement, please use this API`keras.utils.to_channels_first(x_input)`to transform `channels_last` data to `channels_first` format and also please change the `image_data_format` in `keras.json` to `channels_first`.Note: `x_input` is a Numpy tensor or a list of Numpy tensorRefer to: https://github.com/awslabs/keras-apache-mxnet/tree/master/docs/mxnet_backend/performance_guide.md\n", " test_symbol = func(*args, **kwargs)\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "__________________________________________________________________________________________________\n", "Layer (type) Output Shape Param # Connected to \n", "==================================================================================================\n", "input_1 (InputLayer) (None, 8) 0 \n", "__________________________________________________________________________________________________\n", "dense_1 (Dense) (None, 256) 2304 input_1[0][0] \n", "__________________________________________________________________________________________________\n", "p_re_lu_1 (PReLU) (None, 256) 256 dense_1[0][0] \n", "__________________________________________________________________________________________________\n", "reshape_1 (Reshape) (None, 8, 32) 0 p_re_lu_1[0][0] \n", "__________________________________________________________________________________________________\n", "conv1d_1 (Conv1D) (None, 8, 32) 3104 reshape_1[0][0] \n", "__________________________________________________________________________________________________\n", "activation_1 (Activation) (None, 8, 32) 0 conv1d_1[0][0] \n", "__________________________________________________________________________________________________\n", "conv1d_2 (Conv1D) (None, 8, 32) 3104 activation_1[0][0] \n", "__________________________________________________________________________________________________\n", "add_1 (Add) (None, 8, 32) 0 conv1d_2[0][0] \n", " reshape_1[0][0] \n", "__________________________________________________________________________________________________\n", "conv1d_3 (Conv1D) (None, 8, 32) 3104 add_1[0][0] \n", "__________________________________________________________________________________________________\n", "activation_2 (Activation) (None, 8, 32) 0 conv1d_3[0][0] \n", "__________________________________________________________________________________________________\n", "conv1d_4 (Conv1D) (None, 8, 32) 3104 activation_2[0][0] \n", "__________________________________________________________________________________________________\n", "add_2 (Add) (None, 8, 32) 0 conv1d_4[0][0] \n", " add_1[0][0] \n", "__________________________________________________________________________________________________\n", "flatten_1 (Flatten) (None, 256) 0 add_2[0][0] \n", "==================================================================================================\n", "Total params: 14,976\n", "Trainable params: 14,976\n", "Non-trainable params: 0\n", "__________________________________________________________________________________________________\n" ] } ], "source": [ "import scnets as scn\n", "from keras.models import load_model\n", "from scnets import relerr_loss, calc_mre_K\n", "\n", "#Creates a HDF5 file 'my_model.h5'\n", "#odel.save('res15k_model.h5')\n", "#odel = load_model('model/multi_task/try.h5', custom_objects={'loss_max': loss_max})\n", "model = load_model('res15k_model.h5', custom_objects={'relerr_loss': relerr_loss, \n", " 'calc_mre_K' :calc_mre_K})\n", "model.summary()" ] }, { "cell_type": "code", "execution_count": 20, "metadata": { "ExecuteTime": { "end_time": "2018-10-10T13:19:04.315718Z", "start_time": "2018-10-10T13:19:04.280764Z" }, "scrolled": true }, "outputs": [ { "name": "stdout", "output_type": "stream", "text": [ "MXNet Backend: Successfully exported the model as MXNet model!\n", "MXNet symbol file - my_mod_convprel-symbol.json\n", "MXNet params file - my_mod_convprel-0000.params\n", "\n", "\n", "Model input data_names and data_shapes are: \n", "data_names : ['/first_input6']\n", "data_shapes : [DataDesc[/first_input6,(64, 8),float32,NCHW]]\n", "\n", "\n", "Note: In the above data_shapes, the first dimension represent the batch_size used for model training. \n", "You can change the batch_size for binding the module based on your inference batch_size.\n" ] }, { "data": { "text/plain": [ "(['/first_input6'], [DataDesc[/first_input6,(64, 8),float32,NCHW]])" ] }, "execution_count": 20, "metadata": {}, "output_type": "execute_result" } ], "source": [ "# Import the save_mxnet_model API\n", "from keras.models import save_mxnet_model\n", "save_mxnet_model(model=model, prefix='my_mod_convprel', epoch=0)" ] }, { "cell_type": "code", "execution_count": null, "metadata": { "ExecuteTime": { "end_time": "2018-09-30T10:30:53.029361Z", "start_time": "2018-09-30T10:30:52.979098Z" } }, "outputs": [], "source": [ "import numpy as np\n", "import mxnet as mx\n", "\n", "# Step1: Load the model in MXNet\n", "\n", "# Use the same prefix and epoch parameters we used in save_mxnet_model API.\n", "sym, arg_params, aux_params = mx.model.load_checkpoint(prefix='my_mod', epoch=0)\n", "\n", "# We use the data_names and data_shapes returned by save_mxnet_model API.\n", "mod = mx.mod.Module(symbol=sym, \n", " data_names=['/input_12'], \n", " context=mx.gpu(), \n", " label_names=None)\n", "mod.bind(for_training=False, \n", " data_shapes=[('/input_12', (1,8))], \n", " label_shapes=mod._label_shapes)\n", "mod.set_params(arg_params, aux_params, allow_missing=True)\n", "\n", "data_iter = mx.io.NDArrayIter(x_test, None, 1)" ] }, { "cell_type": "code", "execution_count": null, "metadata": { "ExecuteTime": { "end_time": "2018-09-30T10:40:42.350586Z", "start_time": "2018-09-30T10:40:42.321089Z" } }, "outputs": [], "source": [ "size = np.random.randint(30, 71, (1,8))\n", "#spec_ac = snlay.calc_spectrum(size, mats, lams)\n", "size = (size - 50.0)/20.0\n", "#size = np.expand_dims(size, axis = 0)\n", "size.shape\n", "res1 = model.predict(size)\n", "data_iter = mx.io.NDArrayIter(size, None, 1)\n", "res2 = mod.predict(data_iter)\n", "\n", "# #y_pred = result.asnumpy()\n", "# print(calc_mre(y_test, y_pred))" ] }, { "cell_type": "code", "execution_count": null, "metadata": { "ExecuteTime": { "end_time": "2018-09-30T11:18:00.192453Z", "start_time": "2018-09-30T11:18:00.168150Z" } }, "outputs": [], "source": [ "import mxnet as mx\n", "from mxnet import nd\n", "mx.random.seed(1)\n", "\n" ] }, { "cell_type": "code", "execution_count": 5, "metadata": { "ExecuteTime": { "end_time": "2018-09-30T16:24:20.697887Z", "start_time": "2018-09-30T16:24:18.314292Z" } }, "outputs": [ { "name": "stderr", "output_type": "stream", "text": [ "/home/hegder/anaconda3/envs/dp2/lib/python3.5/site-packages/mxnet/module/bucketing_module.py:408: UserWarning: Optimizer created manually outside Module but rescale_grad is not normalized to 1.0/batch_size/num_workers (1.0 vs. 0.03125). Is this intended?\n", " force_init=force_init)\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "0.8106580427940897\n" ] } ], "source": [ "# idx = 102\n", "# print(\"Predicted - \", np.argmax(result[idx].asnumpy()))\n", "# print(\"Actual - \", y_test[idx])\n", "\n", "#result.asnumpy()[1] - y_test[1]\n", "\n", "\n", "y_pred = model.predict(x_test)\n", "print(calc_mre(y_test, y_pred))\n", "#print(calc_mre(y_test, result.asnumpy()))\n", "#history60 = history\n", "plot_training_history(history, 64*2.56)" ] }, { "cell_type": "code", "execution_count": null, "metadata": {}, "outputs": [], "source": [] }, { "cell_type": "code", "execution_count": 4, "metadata": { "ExecuteTime": { "end_time": "2018-10-01T16:13:05.287947Z", "start_time": "2018-10-01T16:13:04.943850Z" } }, "outputs": [ { "name": "stdout", "output_type": "stream", "text": [ "[42 52 67 48 57 66 35 64]\n" ] }, { "data": { "image/png": "iVBORw0KGgoAAAANSUhEUgAAAxAAAADUCAYAAAAMaMXCAAAABHNCSVQICAgIfAhkiAAAAAlwSFlzAAALEgAACxIB0t1+/AAAADl0RVh0U29mdHdhcmUAbWF0cGxvdGxpYiB2ZXJzaW9uIDIuMi4zLCBodHRwOi8vbWF0cGxvdGxpYi5vcmcvIxREBQAAIABJREFUeJzs3Xd4k+X6wPFvZme6NxRogUKh7CEohChiFMW99/a49/4dj4p63HvvfVQcuA2IhijIBqG0UFZLC3SvdGf9/ghFZpuWpGnT+3NdvUjfvu/z3E9D2/d+n6VwuVwIIYQQQgghhCeU/g5ACCGEEEII0XNIAiGEEEIIIYTwmCQQQgghhBBCCI9JAiGEEEIIIYTwmCQQQgghhBBCCI9JAiGEEEIIIYTwmCQQQgghhBBCCI+pO3KyQW+MBAYB68wWU4tvQhJCCCGEEEJ0Vx4lEAa9MRR4DbgIcAGDga0GvfF1oMhsMT3iuxCFEEIIIYQQ3YWnQ5hmA6OAaUDjXsd/AU73dlBCCCGEEEKI7snTBOJ04AazxfQH7h6IVjlAutejEkIIIYQQQnRLns6BSAIKD3Jc04EyADDojbOAh4AMwAo8a7aYnupIGUIIIYQQbZH7DSF8x9Ob/w3AUUDBfsdPBf72tDKD3ng88DpwMbAQCAX6eXq9EEIIIUR75H5DCN/yNIF4AnjFoDeGAwrAYNAb/wXcBJzdgfpmA4+YLaYFuz+vBbI7cL0QQgghRHvkfkMIH1K4XK72zwIMeuOlwH+A/rsPFQL3my2mjz28Pgx3F+KdwBVALLAYuNlsMW3f//zr73hAAfTF/UMvhBBCiJ4pAih65emHPbvhOEwdud+Qew3RC3nl59Hj+Qtmi+l94H2D3hgHKM0WU2kH64rG3XtxCTATKAWeA74CJhzk/L7AAYmFEEIIIXqcfhx8LqUvdOR+Q+41RG902D+PHicQBr0xCxgC2ICNuH8gO8K6+98XzRZT/u4y7wfKDHpjitli2rnf+bUAD9x1I2GhoR2sqntzOB3kbcwjY0gGKqXK3+F4XSC3L5DbBoHdvkBuGwR2+wK5bRDY7atvaODhJ1+Crn3C35H7jVqAB++5hfCwwLrXaI/D4SA3N5fMzExUqsD6f9eW3tpugLr6Bh58/Hnwws9juwmEQW8cCbyPex8Ixe7DLoPeuBy42Gwx5e0+L9psMVUdqhyzxVRj0BsL2HcZ2Ha7T8JCQwkPD2vvtB7F4XCg1WoIDwsLyP+8gdy+QG4bBHb7ArltENjtC+S2QeC3r6t15n4jJDiIkJBg3wbWzbT+vwsJCe5V/+96a7vB3XZvaTOBMOiNaYAZKMLdFbh+95eygLuAxQa9cThwIpACtLcj9evAzQa9cR5QhnuS04qD9D4IIYQQQnSW3G8I4UPt9UA8AKwATjBbTHunLasNeuOnwE+4d6MeDpzmQX1P4h6buAr3JnZ/Amd0NGghhBBCiDbI/YYQPtReAnEccM5+yQMAZovJYdAbH8G9vvJVZovpx/YqM1tMTuDu3R9CCB+r21nMX3N+pTwnD1tjMzalGsWQEYw7dyaj0iL8HZ4QQviE3G8I4VvtJRBxQH4bX98GOMwW0ztei0gIcdhKfv2Nrd/+hHP7tgN2T6qvDOG+D/M4ZVIiV8zoi0qpOFQxQgghhBAHaC+BKAcG4J4DcTBpdHw1JiG8YkdFExGhanQhHi8m1mv8/VcuYdu34dIGEX/EBKLHjUEdFoat1kpQfBwfpA3lgU828dxTP3DRBB1xR01GFRTk77CFEEII0QO0d+c1H3jAoDcev7s7cA+D3qgE/g+Y56vghDgUl8vFne9tRKVU8Mq/hhER2nuTCKfdTuH/PkeXOZSY8eN4e14hy2wZ3PavLAYfM+mQicELV2Xy/ZXvsHlxAUWff8GoZ59CEdy7ViERQgghRMe1d9f1MO5J1CsNeuOTQM7u41m4d3jsB4zzXXhCHFxuUT0NzQ4MI2K48c0cXvnXMMKDe2cSsfXNtykxzUcTFcX6c+9n3tpaXrtzGrE6bZvXqVUKRp8+g5yPvyK6uIQdX8+l7/nndlHUQgghhOiplG190WwxbQWm7z7vE9yrGawCPtp9yjFmi2mbTyMU4iC+X1bKrIkJ3DyrP/ERWuatrvB3SH5RuWIlJab5KNRqymZewg9ranj12mHtJg8ACoWC9JNPIO2mmwDY8d0PtFRU+jpkIYQQQvRwbSYQAGaLabXZYhoFjAbOBc4DRpstptFmi2mNrwMUYn9NNgfz11Rw0vh4FAoFU4dHs3pLV25y2j3Yaq1sfvlVAFTHn8ZLOaE8e+VQ4iLaTx72NnzKSKrSRuFqaaHwsy98EaoQQgghAojHYz7MFtNaYK0PYxHCI7+vrWR4v3ASo91j+8ekR/CWqQiXy4VCEfgrCrlcLiqXLGPrW+9gq6omaHAG/y5K5+ELBtM/PqRTZU659XLW3nwrZb/9TsjggZCV5eWohRBCCBEo2u2BEKK7Ma+rxDgmbs/nqXHBOJwudlY2+zGqrtNYWMSmF16ipaKCkPR0Xos8niuM/ZiYEdnpMqP690Vx5HSsWh3OhkYvRiuEEEKIQNM7Z52KHm399jpumtV/z+cKhYKxAyNYvaWWPrGBv4pQaL9URr/4LJWr1/JEfhJD40M588ikwy538g0Xc90bExinhDFeiFMIIYQQgUl6IESPUlLdTIvdSUrMvkuTjkmPYPXW3jMPIjghgc8aB+JwKbn1lDSvlKkODeWqE9P5bo0Nm93Z/gVCCCGE6JUkgRA9yvrtdRwVUUvZQss+xwM9gXA5nWx57Q12fv8jLqeTuUtKWLyhmscuzkCt8t68j7EDI+gXVI/56Xconjffa+UKIYQQInB4NITJoDe+C7xptpiW+DgeIdqUty6f6YteZ8tSNfH6qSiU7hw4LTEEa6OD0ppmEiIDa0dll9PJ5ldep/TXBSiDgtgRNYDXf23kzRuG+2QDvRPjytD9OJ/CratJONqAUqPxeh1CCCGE6Lk87YFIAxYZ9MZ1Br3xRoPeGOXLoIQ4FPWSBSicTlRhYTib3ZOm7XV14HQwcoCOtfl1fo7Q+7b/73N38qDVEnfT7fx7QSOzLxxMv06uuNSevhOGUhUWT0t5OeV//OmTOoQQQgjRc3mUQJgtpqOBDOAH4F5gp0Fv/MigN071ZXBC7K2pto70olUADH/w36hCQqhe8zerbriFoi++JDM1jA2FgZVAVCxdTtEXX4JSSb9bb+f+RSquPK4vEwZ3fsWl9iiUCnTGEwEo+mouLpfLZ3UJIYQQoufxeA6E2WLaYraY7gVSgfOBaGCBQW/MNeiNtxj0Rp2vghQCYMPXPxHktBE5aiRhA/5ZhclWVUXh53PI3PoHuUX1fozQu5qKi9n0/IsA9DnvPB5eHcykoVFeWXGpPZPOnkGdVkdjURG1OTk+r08IIYQQPUdnJlE7AfvufxVAI3AHsN2gN57ixdiE2MPlcFC7YB4AKSeftOd41OhRDLzuGlAo4OeviFpjDpgn5tveeR9HQwPRR0zkzcbhxOi03LLX8rW+pNZqUE04CoAS069dUqcQQgghegaPZ2Aa9Mb+wBXAZUAk8Bkw2WwxrTDojSrg/4BXgG8Pcf37uHsuWvY6fLTZYlreudBFb1K1eg2q2irsMQlEj913l4Ik43Eo1Go2v/gKhtLFFJVcTGpSuJ8i9Z60qy5HGRzM/NRjKS+z8dLVw1Aqu26n7UkXnMSaRb9Qtugv0q66HI1OOhmFEEII4fkqTCZgOrAe+C/wkdlisrZ+3WwxOQx64xvAf9op6lWzxXRLZ4MVvVdY//4sTZ3K9KmD96y8tLeEY45m53c/QH4Bm377i9TzZ/ghSu9wNDejCgoiOCGBVRPO4Y8lJbx5w3CCNF276nJ4n2SqppzMRlUSR4b3/IRMCCGEEN7haQ/ELmBKO8u4luBercmrHE4HDofD28X6VWt7Aq1drXzRvqYQHd/qJnP9aeMOWW68YRoF739Iw19/4DjnGK/VvTdfvneOxia2vPIaLRWVDH/sYeYuLePThTt57dpMwoOUXfL/Zf/2TbvuPF58/G/OLmvo8bt8y89dzxXIbYPAbp/DGXhtEkJ4mECYLaZLPTjHBRS0c9rFBr3xYtwJybvAc2aLqc0tb/M25qHVBuY69Lm5uf4Owae82b6cnQ76RMGG3PWHPMeZEId16CgWaoeTmp3ttboPxhfvXdP3P2JfvhI0at5571e+KYzi9uOCqNixiYodXq+uTXu378h0eHluDuePU6JQe3/fia4mP3c9VyC3DQKzfS0tNn+HIITwAU+HML0AbDVbTC/sd/wmYIDZYrrNg2JeBO4EKoGJwOe4J2I/19ZFGUMyCA8L8yTMHsPhcJCbm0tmZiYqlcrf4XidN9vncrnY9NSzBCsTmTRoEllZg9o8vyprHE88uZbHhg33yXwBX7131WvWsmH5ShRqFSXn386P61W8ccNQ+nbxU/+DtS82pIj5D72MKz+UEbMf6NJ4vEl+7nquQG4bBHb76uoDZ2U8IcQ/PH2ceAZw+kGO/4V7BaZ2EwizxbRq7+sMeuPjwMW0k0ColKqA+4XaSqUK3LaBd9rXULSDyiVLidSEEDHj+HbLi4tUERGiZme1jf4+2mgNvPve2Wpq2Prq6wDUTJ7JezkqXrtuuF+HDO3dvqSUONKbimhY30JjQQHh6el+i8sb5Oeu5wrktkFgtk+l7Pr2yKItQviep7My44CygxyvABI6WXebQ5eEAKha6c4788LTyUrzbAP0Y1Xb2Pbfx2ncudOXoXlFc0UF6+57gJbychri+/F2cxavXevf5GF/6vAwIg3uOSWFXx90kTUhhOhuXjVbTOF7fUjyIIQXedoDUQRMBrbtd3wy4NFdmkFvPBv4BbAC44B7cC/7KsQhtSYQWyIHkhCl9eiaodYtaLZkU/7HIlLPOcuX4R22EtN8GouKsEYk8sOQs3j72lFEh3e/OT9DzzmF5QvmUfHnYpovuYig+Dh/hySEEF4TiAu2tCeQJ++3pbe2G7y7qIGnCcSHwLMGvbERmL/72HHAM8AbHpZxA/Dm7jp3AK/uvl6Ig3I0NlK7PgcXCjTDRqBQeDanIeXoKVSvX0qZ5U/6nn2mx9f5g+6k01i1vJyyYUfx9KUjCdZ0z+ELwQkJaMdMwL5qKTu+/4H0yy/1d0hCCNGWDi3aEsgLtrQnECfve6I3ttubixp4mkA8CgwEvgJat/lVAJ8AD3tSgNli0nc4ul7E5XIxf/58jjvuOH+H0m1Ur12Hy26nLr4/QwYnenzdkGkT+O21ECgqoqFgO2EDumb3Zk81l1egDg9jU7mdez/M45jps3hwZr8u3SSuM4ZfcAZ/r1rKzl/m0++cs1AH2OIGQoiA0eFFWwJxwZb2BPLk/bb01naDdxc18HQZVwdwiUFvnA20bgO8ymwxbfFaJL2c3W7n2muvZfPmzd36iXlXqlqxEoCc0DSOTY/w+Dq1VkNJ3yzSC5ZT/sef3SqBsDc0sP6h2dQ2w4uxJ3Pd2SOZMaZnDAcKHzQQZ9oQnAWbqd2YR8x+O4ILIUR30JlFWwJ5wZb2BOLkfU/0xnZ7c1GDDi3qbraYNgObvVa72MNutwNgs9nQaj0b6x/oosePo6mukaUlA7ixb8eeDAWNnwQFyyn740/6XXh+t0jKnHY7uU89R+P2QqwhcTxz3yjSBsT6O6wOGXPbdVz3/jZujkgnxt/BCCGEZ2TRFiG8zOMEwqA3TsM97yGR/VZvMltMl3s5rl6nNYGw2+2SQOwWe8REcnSDibPsQqP2dMEwt6FTx7Drex2UlFK3aTO6jME+itIzTpuNFY88jW3NKlq0oUx94gGi+ves5AEgvF9fzjpeyzvzixg3KNLf4QghxAFk0RYhfM+juzKD3ngn8DvuvSAGAKn7fYjDtHcCIf6xeksNYwd6PnypVWaqjt9jjiDlkksJTvJ8/oQv1BXuYN6tD2JbswJXcAjjH32AqP59/BrT4ThhXBy7KhpZMceEo7nZ3+EIIcT+bgC2404gPkEWbRHC6zztgbgRuNlsMb3ky2B6s9bEoaWlpZ0ze4eCjz9FlzGYNZs13HByWoev16iVNE8wkDcgnrSIjicg3rJpUzG77rwVncuBMiyMEQ//h/BBA/0WjzeoVUpuqDPR/PEqirXN9DnlZH+HJIQQe8iiLUL4nqfjQqKAH3wZSG9ns7mX1pIeCGjctYuiOV+x8bkXyS9pYHg/XafKmT4qlvlryr0cnWda7E7eNBVy/UeFNI84goQZ0xnz/NM9PnloNfI092ph+V98g6Opyc/RCCGEEKIredoDMRc4BnjHh7H0ajKE6R+VS5YB4BwykkHxEQRpOjb/oZUhK4a3v1hL3jvLCArW0P+C87wZ5kG5XC6Wv/wenxbHoR6YwUe3jSQhcny3mMTtTXETx5OTMgD1znx2/fQLfU8/1d8hCSGEEKKLeJpA/AU8YtAbs4C/gX3G2Zgtpk+9HVhvs/cqTF3pM8suhqWGMzKtc0/5faFiqTuB2BozhHGdmP/QKjxEzZi+Wsq++xZlcDB9TjsVdWiIt8I8gLXRztwnPiB99U+cGRzCEfe8jkYX5LP6/EmhUJB1xQVsmP0o+XO+IfkEI6oQ331vhRBCCNF9ePpo9xXcqy/djHtHx4/3+vjIN6H1Lv5KIJZsrCZ7u7VL62xLS3U11g0bUWg0mOoSOSoz+rDKO2JqJmUxA3A2NVG20OKlKPflcrn4bW0FT935DumrfwKFgqE3XotGF+6T+rqLmHFjcKWmo2ioY8f3P/o7HCGEEEJ0EU83kuvcGBLhMX8NYaqss1Fe27VJS1sqly0Hl4vgzOGUt6gYlnp4N+FTh8VwX+gIzq7Mp8Q0j6Tjj/PqcKKSqmae+mYbEdmLMRbOAyD9miuJm3KU1+rorhQKBcOvvJCc/zzM9s/mkDBtKsGJ/l3xSgghhBC+16GN5ITv+KsHoqrORoW1+6z81Dr/oShxGFPio1EqD+9mPyxYRd9pR2L76jfqt+VTl7cJ3ZCMTpfXUlmFbdUaqjQhfLczlP8t3MVVcdvps/UnANKuvIzkE44/rJh7kujRowieNoMvt6jICo8m2N8BCSGEEMLnPO5ZMOiNlxn0xtUGvbHWoDem7T52l0FvPMN34fUerYlDVyYQLpeLqm7WA6EbOoTQ/v0w2/sydfjhDV9qddnxA1iqywJg23sf4Ozk97hx1y7W3XEXzXO/4/2XfyS3sI63L0sl9c8vweWi/yUXkTLrJK/E3JOMu+1fRB0znbdMRf4ORXjA0dREfX4BLoeDlupq6vPzadi+neaKCn+HJoQQoofwqAfCoDdeDTwOPAfcB7Q+Fi7DvWHLVz6JrhdxOBxA1w5hqm920GJ3UV7bfXogUs8+k+hZp7J69ioeHeydnY7jIrREH38iTV/mwsY8atfnEDV6lMfXO202GgoLyX70SRzVNewKTWTyMSM56syhAETcexf1W7f16pWI/nVCKuc88Tcz06F/uCtglqvtyVwuF3V5m6hctpyQ1L4kGKYBUJO9ntzZj4FCAS7XPtcEpyQz4tHZaGPcybu9rh6lVoNSq+3y+IUQQnRfHdlI7hqzxTTHoDfetdfxlcAT3g+r9/FHD0R1nZ2wYBUV3SiBAPhrQzWj03WEBKm8VuZ5Jw7hxmWnc/WUaI+TB3tDA9veepdS80JwOgFoSuzPgMvPZcyEcXvOix47huixY7wWa08UGarh1tF2yh66l6akeMa88KzcdHYxl9OJo7iY4u1FNBUVUZ27kaaCAgCqBo1l+c4Uiquaidmay1RNBJG2WuqVwdSqw1ACEfY6rGVWLnorn5TYYlJigpn0v/9D6bARNmQocRPGEj1mNKED+gfcssRCCCE6xtMEYhCw7CDH6wH/bfMbQPwxibqyzsaAhBDydtTTbHN2er8Fb3A5HOz62UTU6FGYVls5ZmSsV8vXhai59dpp3PneRp4dZSWrvw6n3Y5S7f4RsG7Mo2zhH2hjoomZdAShfftQ2eCkYEU2QU4nNl0M8VlDGXfN5Wzcvt2rsQWKY08cy7wvo9Hs3EXh53Pof9EF/g4pYDWVlLLj62+IP9pAxNAh2OxOVr/4NjbLr+TvfZ4mhNL0cSgGD2PC4EiSooOIjxhIWPDZaLETGhaCWqXAZnfS2GTDurOUx3Sx7KxsYkdFMzZtCNq6Zupz1lOfs56CDz7CGR5B7Lgx9D1+BhHDMv31LRBCCOFHniYQu3AnEQX7HZ8MbO1opQa9MQRYB8SZLaaojl4fiPwxibqqzkZMuIYYnYYKawspMf6bAmvN28S2t95Bm5TMqriLeej8wV6vY1RaBA+cO4g73t3IDUPrSfr9M5KON+Kor6fo67l7ehl2qaP58Y9G/lxfyVmGC5k1JZWUIf2Bf4aaiQOptFqG3XIdhY88RNFX3xA9fhwRmUP9HVZAcTQ2UjzvV7Z/+hnOpibWqFJZ8LudnMI6TmhQMC5Yh2pwJvGZGSQPHUD0iOFt9AT9c1yjVqIJDyIiIxWAgcmh7i/o38ZmtVK5ei2Fi5bTkL0WVV0NVQsX8mNTP4Y1xDEtK5pQNXuScSGEEIHP09/4HwLPGPTGswEXEGLQG2fiHr70XCfqfRgoAuI6cW1A8lcCER2uIS5CS3mtza8JRNWq1QBUJA9BPySGsGDvDV/a25Rh0Tx75VBWP/IU0SWlFHzg3sbEhYKCtCPYZXWQt07BtKmh3DyrP9HhGp/EEaj6TxhJ/pTjUf/5CxuffYExLzzr0837eoOWyiqq166lfMVqKpcuQ9HSDMDG6KHUBiVy3vhkRqdHEKwYzfoNkxgxYgQqlfd+fjQ6HYn6o0jUH4XL5aJheyFla7LJ6jOeBWsreG7uNm6u/JbkuFBG33Uz6vDA3v9ECCGE5wnEbGAAkIt7AvXa3cffA57pSIUGvXEsMBO4Dfi8vfMdTkfAPfVtbc/e7Wppadnzb1e1t6K2hagwFbE6NWXVTTgcoV4p92Dta09rArGwJZlTx8b69HswJCWEQc/fS/YPv1OxfiPNVdU0jNETnTWcY9IiSIj658ns/nF0pm09iTfad+SNF/Pb+rWEl+5k4zPPkXHn7Sg1/n863VPeO3t9PQq1GlVQEPVNDla+8C7aNYsB9y/fppR0kk89mYunT95nLoLDoUChUPi8fcF9+5Datw+pwHFjYiktr2fT9c/TvKOZeVffTsQ1NzPpqAyvzpPoKe9dZwVy+xzOwGuTEMLzjeQcwKUGvfEhYBzu5V9Xmi2mLR2pzKA3qoG3gOs9vSZvYx5abWA+Bc7Nzd3zetu2bQDk5+eTnZ3dJfVv3t5CXJgChc3Fuo31xCl2erX8vdvXFld9A/VbtuBSqchWJHB2YwHZ2b6fpKnKSCEhI2WvI8WUFhVT6sFqpJ62rac63PZFn38KtW+8h33NOpotC1F1ow3mutt753K5cBYXY8/ZgGPzFpw7d1GoP5nflEPZWOxgujKBiX0GEpGZRsjwIYTHxmID1q9ff9Dy/NE+3Q3X0PjJ/4gsLcP17L/5+u1UgidPpP9Rw71aT3d777wtENvX0tJ9lgkXQnhPhx4Lmi2mbcC2w6jvdmCt2WIyG/RGgycXZAzJIDws7DCq7H4cDge5ublkZmbuGWqQl5cHQEJCAllZWV0Sh3LtZjIHRaGraMbucJGV1dcr5R6sfW0p/+NPNrugKj6NWVMHMHJEH6/E4QsdbVtP47X2ZUFRfAr//WwTp8SM5rgs706K74zu9t45W1ooW2hh13c/0rRjx57jdoWSqpIGTjq9H08Oj0YXMtGj8vzdPvu4seS//S4Vi/6iT20hmAr5e90Wxt59E5mHuaO8v9vma4Hcvrr6en+HIITwgUMmEAa98T7gObPF1Lj79SGZLabH2qvIoDcOxN3z0KH1LlVKVcD9Qm2lUv3TNufuCbxOp7PL2ltd7yA2IogWB6wrsHq93r3b15aaNe4RcctI5bYpyT3i/fa0bT2VN9rXf8ww7kjqz/Wv5xCt0zJcWU54ehoKP3/fust7lz37v9Tt7kWwqsMoThlG4uSJTDp+AtNidJ0u11/tU+l0DLn1ZuxXX0npb79T8NGnJKWlcMf7mzBkxXDtzFQiQw+vN7m7vHe+EojtUykDqz1CCLe2eiCuAt4AGne/PhQX0G4CAUwF4oH1Br0R3EuARBj0xmLgdLPFtNijiAOUP5ZxraqzEa3T0OJwUeHH3ag1ETqcKjXxE8bKpOUAMyg5jCcuGcL7L3zLWdu/J3bSRAbfciOqoCB/h9ZlXC4X1atWU/T1XJLPOYd1iiQWrKmAigSOiKhCbTAy6YzpREcFxmRzdVgYKbNOImbCeLRxcUxqgTd+2c6D937OjPGJGM+fjkrlvyWjhRBCHL5DJhBmiyntYK8Pw+fAL3t9fiTuSdijgUovlN+jta6+1DqZuitU7l7G1eHw727USRdcxM2b03n5lN69GVugGp0ewQUzh1D7+s9ULP6L5tJS+l1wHpFZbS0x2rO5HA6sG/OoWLKUiiXLaC4pAeDL8t8oHX0MM0bHMe3kq4iNDNxEKjgpCYAINdx6fDLLvl6A68ta5i74mTHXX0H6BO/OjxBCCNF1umxpFLPF1Ii7NwMAg95YCbjMFlNxV8XQnXV1D4TT6aKm3k5UmBqXC8qt/uuB+HpJCUMzEukXHxhPYMWBJhgnsi7qAbY//SRs3kLOQ4+gCg1lyJ23Bdwu3gUff0rxvF+x19TsOdYQpMM+aTrXnT+LxKTet/WNUqNmwDlnUvjZFyRVFbDrkQfYkjWJafdci1Yny74KIURP41ECYdAbXwC2mi2mF/Y7fhMwwGwx3dbRis0WkxnofX9JD6F1+b6uSiBqG+2EBatQq5REh2uoqbdjd7hQq3y/+lEra94manYU89HvWl65dkSX1Sv8Y8QRQ4l86gm+eupjMq2b0dUUs/GpZxn19BOE9Elpv4Burr7JweLcKsqWFdC/pgZrSDTKrLEMNU4hdexwv8/98CelRkPKrBNJONpA0dffsOPbHwjNXoL5yhwG33g9aVPG+ztEIYQQHeBpD8TpwBkHOf4XcAfuPR3EYWgdwtRVG8m17kINoFYpiAxTU1VnIz6y64aUbP+FcRNzAAAgAElEQVTf51SvWs25Y05kYPIRXVav8J9+A+L517M38dTXW+k/732GxKlQ734CbW9owG61EtyNlnw9FEdzM+V/LKJ40VKqt+9gZZqB7xv6MTJNx7EnnUJa4qkkjxzq1b0QAoE6PIwBF19IwjEG8p5/CTZtZt2Lb1AQ3h/D6Hh/hycCjEFvDAHWAXFmi0keWArhRZ4mEPFA2UGOVwAJ3gun97Lb7YSEhHRpAhEV/s/bH6fTUF7b0mUJRPXfa6letZpmpZbjLp/VJXWK7iEkSMUD5w1m/sDrmf1TEfr55Vw1I4iSl16kNieXIXfeRtToUf4O8wB1W7Zi3byFohVrafh7DermBsD9S3Sc9i+ufOZUIkL9v2FeTxDaty+jnniMoq/nEhqfzoM/FrJsi5WbTx5AkEYmWAuveRgoAuL8HYgQgcbTv3ZFwGQO3ANiMuDd3cd6qdYEoquGMFXW2ff0QADERWi7bB6Ey+Fg27vvA1Ax7lhS+smTx95oxsQUJmbF8+YvRVz4xEpusjagq6tj/UOPkHDM0aSecybqcB1KrQal2vc35i6Xi8YdO1CFhKCNiqJuy1aCEhMpsWlYubmWhi++ov+WJYD7F6eib38GzDqBiIyBBCclo5bkoUMUKhWpZ51BKvDxEXYem7OVZ+9+h3OuMpKe2c/f4YkezqA3jgVm4h4h8Xl75zucjoDcCbwtgbwDelt6a7vBuzvDe/oX70PgWYPe2AjM333sOOAZ3Eu9isNkt9sJDg7uuh4Iq22fJVNjIzRUdNFKTCULfqMhv4AabSQzbz6/S+oU3VNkqIY7T0/jXH0S782PQ9n4PdPKllD66wJKf10AgDIoiFFPP0Fov1SfxNBcVkbjzl0UfvEltdnufRlcSiUKp5Of+s9kTewoxg2K5MghAwmLU5AwYihRo0YSkpoqQ5S8RBei5pa+O9j81S9su+93dpxxAVMumCnfX9EpBr1RDbyFe+8pj+RtzEOr7Z3LiAfiDuie6I3t9ubO8J4mEI8CA4GvcO/7AKAAPsHdRSgOU1cPYXKsWUrf4DAgHdjdA9EFCYSttpaCDz8BQHfa2YTpZOUlAalxITxwXgblJ97IT78cjePnrxhQvQWVUgHR8djjkvacW1+wHZfNhlKrJSghHlVw8CHLbSgsoqmkBJfNjr25Cdu2fKrtDiIGDUQRrqOovJn8Vz9As+YvABqVQaBUEmJvxB4ZywXTknnk9HG7b2QH+/rb0KvFTpxA5RETqVy6DOa8y/c5G5n58M2o1b138rnotNuBtWaLyWzQGw2eXJAxJIPwsDDfRtXNBPIO6G3pre0G7+4M71ECYbaYHMAlBr1xNv/sJL3KbDFt8VokvVxrD0RXDGGybswjdf6HuBQKSgeqSTjaQJxOy5biBp/XvfP7H7FbrZTFpXPKeUaf1yd6lrgILRefPRbXWWPYUFTPnzlVrNpQRs7sVcRFaBkX04L+56dRuJx7rlHpdGjj4wmKjyftkovQJCZS39DM1scfp3F99gF1bJj7HX8njuXT6OkkRGo5uTGIfnEphIwcw/gLziAyNgJHYxOqkGB5At6F1OHhDL33LkoX/MaW198idv0iPru5gZMevwPdYe5gLXoPg944EHfPQ4fWh1YpA28XcE8F4g7onuiN7fbmzvAdGrRrtpg2A5u9VrvYw2azdVkPRNGXXwOgcLnY8vpbRI8dQ1yEhqV5vu+BqD7ieH79s4KrbjlJbs7EISkUCjJTw8lMDQdjKnaHi/zSBjb/9DuN0YnYnApoaSassRqsVhqtVhq3buW04pHUBkUSolVx4Y4W0pRayqL7oQrSog7S4nTaiXM1oT9mPFfOmohWrQTGHlC/OlR6xvxBoVCQeOx0guLjyX30cdKKVvPztfcx4l5Z6E94bCruhV/WG/RGAC0QYdAbi4HTzRbTYn8GJ0SgOGQCYdAb7wOeM1tMjbtfH5LZYnrM65F1IYfTRX2Tw68rqDgcji6ZRF2fX0DlsuXYFGpiTphJv0lj0ERGEhthpcLHk6gbmx08PCef6689j5QhsT6tSwQWtUrBoOQwBl1xElxx0p7jLqcTW3UNzWVlNJWU8tPkI1Cq1SgUCprL0lCFhqLePSzB4XCQnZ1NVlZWr3vq1NNEjRrJ8IceYMPjT9FX08LdnxZw6oRgsrL8HZnoAT4Hftnr8yOB94DRQKVfIhIiALV1x3wV7gnSjbtfH4oL6NEJhCW7ks//LOb164b7LQabzdYlk6iLvvoGgGWRI7jliot2P4GFuAiNT+dANO7cyYsLKsjqH84xIyV5EN6hUCrRxkSjjYlGNyRjn68FxcvqXj1ZROZQRj//NI6GRp5x6rj1rfXYNNv51wkDUCql91IcnNliasR93wKAQW+sBFxmi6nYf1EJEXgOmUCYLaa0g70ORLlF9awvsGJ3OFGr/LMGeesk6pqaGp/V4Whqomr5ClAqWdfvqD3JA0B0sIJKqw2Xy+WToUXLH3+JiYX5ZN5/j9fLFkIEJm10NERHM9jh4L4Tgtj84Sd89X0zM/7veqL6JrVfgOj1zBaTGZBN5ITwskPeLRv0RodBb0zY/fpdg96o67qwutamnfU0211s3uX7ScSH0hX7QKiCg5n40Xsob32IiD7u3X5r1uew8prrKXjxRUK0SmrqvV//lqXrUBXkERSkJm6YrGQjhOi4MGs5QytySNmVw5obb2bTLwv9HZIQQvRabT1ubwTCd7++BDj0Wok93KadDYwfFEF2QZ3fYuiqjeSUGg07VDH0jQ0CQBsTTVNxMTXZ64nVaSi3encYU2OzneWvfghAnxOPRx0a6tXyhRC9gyoxgVEvPU/skZPROO2UvvYiy9/5HJfL1f7FQgghvKqtORBLgW8MeuMy3Hs+tG4kdwCzxXS1L4LrCtX1NuqbHRw3Jo7VW2s58yj/dIu3JhAtLb6bh+ByOlEolRRVNNEn1p0PBicloY2NoaWikkGKKsprbQxK9lJ9LhfvPv8do6u3og4PJ+WUWd4pWAjRKwXFxTLkrtvZOfc78t//kJbvvuC3naUcc991KGRivBBCdJm2eiAuARYBA3BPlE7HvZPS/h+DfBui7zQUFbFhzg+MjLaT1T/c7z0QvtwHwmmzseziy8n+94PsKK0nNc6dQCgUCiJHuJc2SW/Y7tWJ1N8sLGTgyu8B6HfheWgiIrxWthCid1IoFPQ57RQy7rgN1GqCV5j57PGPcDilJ0IIIbpKW5Ooi4DrAAx6oxM4zWwxlXZVYL7mcjjIfeS/OHYVc5ZCid01GWvNEVTX24gK6/pNi1r3gfBVAlG3ZSt2q5WWyioKI2z0iftnRFrkiCzKzBaSKrdSXuudVaC2lDnYOOdbjm6uJWxgOknHzfBKuUIIARA/9Si0MdEUfPkNf0WNYek7G3jkwsGEh/hvOW4hhOgt2ppEvdWgN7aut/kQ4L/H8z5QZvmDpl3F2DTBgIvKJUs5KryC9dv908zWfSB8tYyrNXcDALrMoeyoaKZv7D8JRNSoUaBQEFm0keqyw18mu9Jq442FLUy5+ES0MTEMvPYaGV4ghPC6yOHDGPmf+3n2XyPoExvEDU8vYdOiVf4OSwghAl5bj2qSgVCgAngAeA04rGWKDHrjq8AsIBKwAnOAu8wWk++3QN6Ly+Gg8IuvALCkGTnllDEMSo1kzXrILrByVGZ0x8t0uahYtJiQPn0ISxvQ4etbhzD5KoGo3Z1AKAYMIqxcRWjQPzf0QfFxRI8fS9XylQTnrARGdLqe+pJy/j1nF0cOVDF1UjqKia+i1HR9j44QovdQq5TceXo6vy77nNInV1A2/iiOuPM6VMEBu/aHEEL4VVsJxFrgHYPeuBD3JOpbDHrjQR/Pd2An6peBO80WU71Bb4wHvgDuBmZ3IObDVr5oMU07dxKUmMgC5SDumjiMYK2KYbWVfLu0pFNllv66gM0vv4Y2Npbxb73W4SfuvhzC5HK5sG5wJxDVcf1JjT0wX0s952zqRk9jxTYd13SynobCIlbcfg8Tk0cx+uJjASR5EEJ0CZfLRcbogRTmrcG1YhGWGwuY/Oj/EZwgGwoKIYS3tZVAXAE8CVyKexL1hYDjIOd5vBO12WLKOci1bU7CdjgdOBwHq7bzwocPI/nUk7FGJpGwKQSNyj2EqG9IM+rc1dQVhBLSt6/H5TUVl7D17fcAaKmoYNHXvzPg6CNJjg466Pmt7dm7XXa7Ha1Wi81m83p767ZsxVZTizYmhiKXjpRY6wF1hKanEReeRPnavE7V73I6WfHf51E3NzI2ugUUeL0d3cHB3rtAEsjtC+S2QWC3z9O29TnrDGImH8H6R56A0iIW33gHox7+N1GDuvdeqAH93jkDr01CiLYnUWcDM2HPJOrx3phEbdAb7wHux73HRAVwV1vn523MQ6v1wVPssaOx5NlJDreRnZ0NQNMv8zl181+sm1NC6IxjPC7Knp2D02ZDER6Gq66ebXN/4qnccO45IZhQ7aF3dc7Nzd3zur6+npKSEpqamvbE4y3Nv5kBcA5MY2VuEUFqxUHraGxxUVbTzNqlS1GGhXWojrI/VxGyYxvOsHA4YQYKlXKf9gWaQG4bBHb7ArltENjt87RtQZdfjOOzOQTl57Pm3gdQn38eUYP7+TY4LwjE966lxTfDcoUQ/uXpchVpQJk3KjRbTI8Djxv0xkzgIqC4rfMzhmQQ3sGbWU/NzdnK1FHhZGUlAFDZ0ETe4r9QFpeTlZXleUFZWTRMnoQySMuam+8gva6A6anwyUoNz1yegVq171x1h8NBbm4umZmZqHYPdVKpVAwdOhSXy9Wxuj3QkpJC1ZAhhKYP4OlvG7n/rDSy+h+4sbjL5eLcFx+i4emNjHr+GUL6pHhUfk15NaW/PQ1AxpWXET1m9AHtCxQHe+8CSSC3L5DbBoHdvs60zTlmNJuefZ6qpcv5a14uM8bqyeof3v6FfhDI711dfb2/QxAHUV7bQn2TJHei89pMIAx6423Aq2aLqWD350cAq1snPRv0xnDgMbPFdFNHKzZbTLkGvXEN8D5w7KHOUylVXvuF6nI4yJn9GFFjRpM883jW5tdx/rSUPeVHDc8EwLZtCwqns93x+47mZlRB7mFKunR3F3lt5kSaa6xcNr0vN35dybLNdUwddvBJ2SrVP22z2+2EhYVhs9m8/gckJD6ekJnHs6GoDrsjj1FpkSgUB+8ZUQdpoNZB2YLfSLvsknbLdjhd/PDo2wy0NxAxLJOEo6fhdDqBfdsXaAK5bRDY7QvktkFgt68jbVOpVGTefSc7v/8RZ/ok7vogjxtO7M+siQk+jrLzAvG9UykDqz2B4rWftlNY3sT1U/0dieip2tpIDuApYO/dv+YDe08OCAOuP4z6Vbg3o+sS1X+vpXr1Gop//oXqJheVdTbSk0L3fF0TEUFTdCIKu436rdsOuH5bSQOzP9/MCQ+u4N9vrWblNddTvmjxPuf8kHIcMVffgK5fX44YEkXOdqtHsdntdkJDQ322DwTAL6vKMY6JO2TyALA9fSIApQt+x+nBrtgfvGNmYMEyFGoV6ddc1WbZQgjRlRQqFX1OPZmpIxN4/brhfPbLJube+igN5RX+Dk0Iv2mxO1mYXUl+aSM7qpz+Dkf0UO0lEPvfDXb67tCgN4Yb9MbLDHpjlEFvVBj0xhHAvwFTZ8vsqJJ5vwKQcLSBdQV1ZPULR6Xct0nKNHc+07rs6d7+++VWIkPVvHbtcPoVrsFWVcWOn/4Jv6HZwfrCOsYNigRgWGo4OYWedd/uvRO1y+W9HVU3v/o62955j8bySuatLsc4Nq7N8/uMHEpjbB/sVisVS5a2ee7PK8uYt0tLlOFo+p55BmED+nstbiGE8Ka0xFDuC11O/NZVLLrudgpXeneumRA9xV8bqhmUEsbpkxNYmOe7h5YisLWXQHiTCzgf2Ip7D4hvgR+BW7qi8oaiHVQsWYpCrSbh2GNYm29lVNqB8wBistzDmGpz9p3Mtq7ASklVC9fN7E//hGDGVa0FIC91wp5zVm6pJTM1nBCVk4qly+mzbQU52+s8SgjsdjvV9U7UarXX9oKwNzRSuuB3dn7/I2u31xOr05KWGNrmNUePimNxxEgAdv3w8yFjzy6w8ty3+Tx61RiG33oDqeee7ZWYhRDCVwZfcQm6YZmENlvZNvshln0416sPbIToCeatLmfG6FhOmZjA0m12GpplpSzRcZ5Ooj5sZoupHpjRVfXtb8fXc8HlImH60QTFxrJ22y6uMh64VGv/CSNZ9yHYG/bdM+/j33dy/rRk1CoF1o15NG7fjitMxx/Ofpy1+5y/cqs4cmgUjTt2sOGxx1FHRBA08Fp2VjbTJ7btDY1sNjuXvLgetVq9Z0nXw1W9Zg0uux3d0CG8tqSWkz0Y+5vZN4w1UcM5NuwPrBs3UrMum6iR+24sV7BpB8tmv8z/3XgdA5PdCYkMXRJCdHfa6GiyZj9I/vsfsuv7H7F99RGmnI0c++DNqGXTOdELNDQ7WLyhmjtPS0MXoiQjUcW81RWccVSyv0MTPYwnCcQEg95Ytfu1Ahhr0BuTdn8e45uwvKu5rIwy80JQKulz2qm02J1s3FnP8H4H9kBEpibz4pjbeO6GcXuOFZQ1snprLQ+e596yonjefAASjjGwdmsjNrsTtUqBZX0VL16dSWhiCCF9UmjcsZOpwaXkFNa1m0DY7TZsTtWeBMIbKpetAKAofijNdienH5nY7jUKhYIpo5Mo1eqJX/4L9Vu37ZNAlJRUs/rfjzCksZSkFT/BhM5uOxdYXC5XQKzh7nA4cLlcXh9K1x0EUttUKpUk7Z2kVKtJv/JydBmD2fTSq+hyl2G6+g6mvPQkkZFt99AK0dP9sb6KUQN0RIVrcDgcGIao+XpJKacfmSS/U0SHeJJAfMe+cx++2O/rPv1L3JCfjyo2luCE+EPu7mytqaeyqpHGgnzsOwuJUDtRBweRcLQBdXg4FUuX43I4iNNPISQ5iS/+3MWgpFDCgg8sT6FQkJwSTX5pI33j3Df9n5p3csaRiYQEqbBZrZT/sQiA1BOOo98X5eQW1qNWKwjWKklLDEGhUBA35SgKP59DVk0uuYVZzBjd9twDu8MBChVqtYYWDyYvt8flcFC1YiUAH5XE8/hdgw6Y73EohhGxvLwxi1dfP4ngxH+SjuKKBhbc/V8GNJYSnJJCvwvPP+w4ezqn00lzczMqlQqlsitHBPqGUqlk4MCBAdGW/QVS21paWlAqlWhkp/dOi9dPJbRfKrn/fZL62CFc+soGHrsog8zU7rnUqxCd5WhupsQ0H3VEBIv+VjHjyH/WrslMVjJntYPs7XWMOMjy7kIcSnsJhN+371z/7wfROp04FUrsmiAcKi1N6hBMQ05nuyKamgYb522fS5Y1b881Nbv/3TT3F0Y8/ADJJ56Ay2YjesI4fl5Zxoe/7eT164Yfss4BCSHklzSS1bKdqpJKfv07jDl3jwGgzLwQZ3MzUaNHEdInhTHpLazcUkOTzYkhK2ZPBh+nn0Lh53OI2ryavOTpwIBD1udyuXA6HKBQolR5pwfCujEPu9VKTXA0p588mv7xIR5fOypNx84GFdnWYMbvzh/+Ni0m98PPGVBXhFqnY9gD96HRyS+b5uZmQkI8/952dy6XC5XK3RMWaE+jAqltarWapqYmSSAOU9iAAYx+7mnGarVEr6vm5rdyuTLTzinHZxIUHeXv8IQ4bE6bjY1PPEXVytUAGAHVch05GYOInToVZUwUpx6RwFeLSySBEB3SZgLRuv+DP9ljklA1N6CyVqNtaQQaCaGGa05KJ7pfMpFhagqft1C9OpjQ1L6EpafjCAqm7K/lqMt28dyD/2N92lQSogZR/nEZ9U0OXvnXsD29CwczMDmU5QtWk7b0LezaEE445W5idO4/1MknziQoLg5tnLtHYdzACOYsKqastoX7zx64p4zQvn2JGJZJbU4uITkrcDjHHLIHoHXvB4VC4bUEomLpcgBq+g/n3ClJ7Zy9L5VSwf3nDOQ/n2xi/OBIwjau4qg1X5AKaKKjybz3LkKSZbxk6w2pEP6gVCpxOp0B0aPiT+pQ97ClGWPiGBKnYMPtt/Pn1w7Sr76KtBl6P0cnxOHZ8tqbVK1cjVqnoz5pAGzfSpDVStXqv0k+9RRwOTlpQhxnPbGW6nobUWHyUEJ4pssmUXfW5BceJzw8DKfNhqOxEUdTE7aaWsLSUlGq3eFn3nPnAdeln3U6JfN/5YFTTmZHRQulNc3ERWhJiQ0iWNP2Td/McfEsyhlASfwgEss2M6PsD8A9D0ChVBI7edKec0enR3DfR3noQtQM36/rO9E4g9qcXCbV/k1BaeM+e07szW63o1S526Ly0ipMyzRp1PSdyJmXntCpp61Th0Uz5q7RfPbHLgZWOVGoVUSPHcugG69DExHRfgG9gMPhkJs34TeSQHhfYqiCmiEDqF27jp0vv8COhYs44s7r0ERG+js0ITolYbqB6jV/M/Teu7hnQQunX5jAuGQnLZWVhGUMhuxsosI0TB+o4cflZVxgSPF3yKKH6PYJRCulRoNSo0ETEUFwQvurCWkidPQ94zQABiSGMCDR82EmGrWSRy/K4AnricT/9hJ1lt/ZlZGGMkhL4rHT9zk3IlRNWmIow/qFo9yvhyF28iR2ff8jNcGDyCmoPWQC4XA49iQQCuXhJxDZBVbe3RTMh4/cQkx0UKfLCQ9Rc+VxqXDceTivPrPdnbmFEKInC4qNIevh/1D8yzy2vvsBinUrWHT1jQy95XqSJh/h7/BEBxj0xleBWUAk7qXj5wB3mS2mw59k2INEDh/OuDdeobLRRW7hGo66YgjBGhXBiQl7Fv6oWrUaw0/PMC91Bq5pl/l9iGfNumzyP/iYhKP1JJ8406+xiEOTR1eHoFEruf/GqSSfYASnk61vvs2W196koajogHMvnd6Hs446cIUjVVAQo555kqDjTiKnqOGAr7ey2Wwole5eEYVSdVhDmErWrOehj3K5+4x0Eg8jedifJA9CiN5AoVCQfIKRcS8/R9iw4Wia6tny+JOseHf/9UNEN/cyMNRsMUUAo4FRwN3+DalrFP8yb88wZnD//V7wdwVTh0cfdARGbfZ6sLVw3NYfyZ47vytDPYCtpoac2Y9Rt2kTW996l6pVq/0ajzi0HtMD4Q8KhYL+559D5aLF2KxWUs8566Bj/6ePim2znGGp4cxfXX7Ir9vtdmhNIFSd74Foqaxi4+xHuUIXy7TBT3aqDCGEEBCcmMioRx9kxzffkv/xp7y/zsni7/O5+vjUdofBCv8zW0w5+x1yAYMOdb7D6QiIpbjrt+Wz9a23wQWjXnqe4CT3w03TqjIum95nnza2vu57wXloIiPZ/sFHlH07F/usY/zWC6EMD6f/pRdT/sefWHNyyXvmeUY8/QSbm0JYvdXK2VMS0aoP79l3a7sD4f3uKIfTe22WBKIdGp2O0c8/g9Nu82jo1MEMUFlJzf6dqg0hRA8dfMDX7XY7CoWKmHANjQpVpxOIdW+8h8reTEp6CqoAWhlIHJ7jjafsed3S0oJSoUStcf/ojxyZxZNPPeqv0ITo1hRKJX3POI24KUcxPCyaJ7/exsXPruOeLCujj5skPbPdnEFvvAe4HwgHKoC7DnVu3sY8tNqe/X66bHYa33gLl92BesJ4NpeXQXkZ2yuc7ChvJtRWSHb2gaMocnNzcQ3ohz1MR1BVCX/P/Rb14EPmWl7nKNqBq74e9ZAM94E+ybjOPhPVJ//Dvmkzfz/8CM+nnItTpeGrRUVcNElLRuLhJ/G5ubmHXUZP09Jy+HNsW3mcQBj0xmnAcUAi+w19MltMl3stom5IGxN9WNfXmH/j+DILW34OYvxBEgibzQZKFUnRWrYqOjeEqXbjRpqWLMKlUjP4qssOK14RWH4xfbvn9f33PcigQQO57PKLDnl+b3wqI0RbghMTCAb+e3EGv32/jMY3n+f3z+IZcfM1JIwf4+/wxCGYLabHgccNemMmcBFQfKhzM4ZkEB4W1mWx+UL+ex9QX1pGcEoyI267GVWQexjz3C+2cu60OEaP3HeCtMPhIDc3l8zMTFQqFdtPOYmdn/4P5Zpssk471efxOm12dsz5kh1fz0Wp1TL0mScJTv5n1Uh7Whrr7ryb5p27ONZl4bKX72JhdhXPflfApIwIrj8xlcjQjj8H37/dvUldfb3XyvKoH8igN94J/A6cjntDg9T9PkQbWldtali5/KA74Dp2byKXFB0EnZhE7Whu5u9nXwWgzymzCEnu2LKtoncrKtqBQW/k559MXHDeZVxwnjsBPfOMC1i8eMme81asWMXJs87a87ndbufddz7g/HMvZdaJZ3D3nf9HcXFJl8cvRFeakBGFNiWF4NoyNs1+hEUPPEFLVZW/wxJtMFtMucAa4P1DnaNSqlCpeu6HNXs9xd//CEolGbfdgjY0FJVKRW2jE8v6Kk6blHTQ64A9r/vONGJXaWhYt5bmHTt9Gm/T9kLW33MfO778GlwukowzCElM2OecoKhIBt95O43qEEYcMwG1Ws300fF8ftdogrUqLn4umyYbnap/73b3lo+mwkKvTnz2NHW7EbjZbDG95MW6e43wQQNx6KJQWaup27yF0PR99+ez2Wy4UJIYFYSLjvVAuFwuNr/yOsriIlwx8fQ7+wxvhy96iUWLlvD6my963APx5hvvsnnTFl58+RkiIyP44P1PeOD/Hub1N1+SpUVFwNINyWDci8+y89vvKfhsDvy9jD+vzmbAtdeSfsyR/g5PHJoKOHAIQACw19Wx6QX37VnqOWeh22v40dwlJRw9Ipao8PaHZ2l0Ouxjp7CrpJqxmsMb4e5wuthZ2UR+aROV1haOHhFLRKgal8vFjq/nsv3Tz3DZ7QQnJTLophuIHD7soOUsqIhg4dF38MLZo/ccCw9Wc+fpaZTWNDN/TTmnTjpwERuxL1ttLa6M/BsAACAASURBVOvu+zfExEKwd5bi9/R/SBTwg1dq7IUUSiWhYyfQvHA+FYv/OiCBsNvtOFGRHB2ES6HsUA9EzbpsyhdasCk1jP/PvR2e+2C1WgkODpYdbb1s4u1/dUk9y56Z7LWyLr/iEv6fvfOOi7L+A/j7Bhzr2DJFQBTcq9RciKbi3iPLke1hpY1fWZZZalMtR+W2bKhppmaGI5HS3FtxMWTvDcfB3T2/P05OkAMPPBz4vF8vXq+75zuez/e4B76f72cplUpUKtUt++p0Orb+/gfffvc1rq76JAJTnprIxg2biYuLx8/P12xyiYjca0gtLGg4eiSuPbpzefE3cPYsyV/P50xkKoNfGIZcJirQd5OQ4FA7YAywBcgFWgHvA2F3U666ojg1DUEnoAwKxGfMjUNEjVbH5oOpLHimmclztX31WcZ+dppRDWq3KQ87kcHavQkkZBTjrLTEz90aKwspS/6IY1RXd0K1kST98CMAHgP64zd5QpX7luISLat2xfPFlBvyq9MzsHR1QSKRMLSzO2v2JIgKhAnEr9+ItrAIq4AAUOnMMqepCsTvQG9glVnu+gDSuF8wkft3k7p3H97jxlRoK1Mg3J0s0VEzFyaH1q3YHziYbh28savFpu2jjz6iW7duDB9e9/6ODxLm3NjfKTw8TE8SkJ2VjVqt5pVX3qjUlp6WISoQIg8EVu5utP7oA5K2/UHc5i38KzRkw4IzTBvqR3t3sHR0vNsiPqgIwOPAfMASSAM2A7PuplB1hV1AY9ovWohOXYxEdsOnf9/ZLBq6WhHoZXpsh6u9Aj83a05G59EpsGbf30OXcli4NZa5E5vSwscOa8UNWRIzi1m7N5F5/+XxhI0S/0lP4Dugb7Xzbfg3hdZ+Slo00hfpTd6xk9i1PxDw0vO49QqhS5Ajn/4aRVRKEQFV1NkSgaKEBJJ3hoFUSqPHx8Oqn8wyr6kKxH/AnJDg0FbAaaBCIZbwiLCfzSJNPcapZXOy7Nxxzk0l8+B/4OJsaFOXlCJIZLjaW6JDWiMXpoMXczjn3oF3xrWplVx5eXmkpop+6yIAFdP2WVtbUawqNrzPzMg0vHZ0csTS0pJlyxfTsKH3HZNQROReQyKV4j18KB4DQnnE0pLws1l88esVXji7FMegprR+5XkUDVzvtpgPFOERYYVA9bvTeoAqOcUQ82hhrwSUFdo3/pPC4z0rp56/Fd1bOHB+5z/Y/xZJ4BvTTco2djmpkFk/X+HTyUG0b1zZRcbbxYr3xgaQ3KchX21siqvKgbeqmS+vSMNP4UmsmNrKcE1mbYWupISYVWtx6tAeCwcHBndyY9vhNKYP86vxOk2lVKNDIpEgl93dAnu1JXbND6DT4d6vD9Y+Dc02r6l21qXosy+9BqwGfiz3s86UCUKCQxUhwaErQoJDo0OCQwtCgkOvhASHTquN0PcjEomEwod7k+PfBmvvihuu3AI1cpkMW4UMrWBaGldNQSHFaWn8sj+ZJ0K8KlXBNhW1Wk1mZuatO4o8cAQGNmH37r9Rq0tISUnl11+3GNpkMhnDhg9myeLvSE1NA/TucOH7ItBoxCxOIg8eMoUCiURCrzYurBrvikJbgvr0CQ6/8ApRW3YYTaAhIlJbck6d5uTU1/SxBEa+W5HxBaTmqunR0tnI6Orp1sIZp/92kPnfYTIOHLxl/9RsNa+vvMibI/yNKg8ARfH69LGezgreerwFYSfSyS6oeq+zbl8iPVs54+t2w72pQa8QHNq0RpOfT8zq7wEY2smNncfTKdGYxy3HGEv/jGPzwSoTeN3TZB8/Qfax48hsbGj0xHizzm2SAhEeESat5sfUHFhy9CnU+gH26P0TZ4QEh46rnej3H4FD+/Kr3wjsmgRUuJ5fqEZuIb+uQJgWA5G0bTvHn5+K85n99G1XfSG76iguLhYVCBGjPPPsFAoKChg2dAyzPphD//4VD/See/4pWrRozhvT32FA6HCenvIiBw8e4i7VHxIRuWdwCPCn07LF2HfqhExTQsra1ex+dSbZ8cl3WzSRekBBVDSRn3yOoNGgVauNFn3b+G8Ko7t61OrUPNDLlqNuHQGI/2UD6vSqC+HmqzS8tjKS8cGe9G1n3NKWHvEPJ1+ZRvzGTQC42lvyaFsXNv5r/HlIzCzm90NpPNOv4mm5RCIh4KXnkVpakh6+n5xTp/F2saKJpw37z2XVeJ2mci1NRXK2us7mryt0paXErFoD6IPrLR0d+auaosY15Y4VkrtuUny/3KVTIcGhO4BuwIaqxtWX6pAAzRvaEJumIrtA/0UsW1duQTFyuQVWFqARpJSUlFS7Zq1aTfKOnaDTEvhwM2QSodafkUqlIiMjw6yfcX2u8lh+bYIgIJVK76uTxTlz9S7A5WX29vZi3/6/KlwXBAEPD3cWLZ5fYfyo0cMNfeRyORMnjWfipMqnGvfiZ1J+bfUBQRDQ6XSGzcOD8tzdL8gdHWnxzptkHjjI1e9WYBt3kdOvTCM3eBh9XhyNleWNs7f7cX2mYs7KtyJQlJDIhY/moisuxjW4B36TK9f0ic9QEXE+i2nDalejRCKR4NC9O+rC05CSyJl33qPl7A+wucldtVSj4+21l3i4iUOVrlKqpCSuLv0OBAEL+xvWiQm9vHh60TkmhHhja1XxHPqrbbE83tMTd0dFpfmsPT3xGTeGa+t+IurbZbRbtJBhnd3ZdjitSgXmdknOVleS8X5AIpXiNWwIafv24xIayie/RnHiSia1K4lcmZoUkuuFvqJjS/QBSueBueERYeG1uXFIcKgc6A58Xl2/+lAdsjz+zgL71v5Go7w4Luh0SKRSoq8lIJVA9JVItEiJi4/n3LlzVc5Reuw4moIC4q088G3uWG3fW5Gbm0thYeFtzVEV9bnKY2RkJIIgEBAQYMgpXZ8oLi6+daf7lPqyNq1WS1RUVKXTx/r+3N13ONhj/dJzlITthjPnuJBRypK5JxjQSk4PPx0WVpaG3+F9ub5bYM7Kt3WJKiERrbYUTV4Bpfn5SC3kyJVKFC4u2Pr7AdeV9pISpHI5SKUVnj1BECjRCKhLdRSXailITqMoMYlirQS1ToJKo//RFRaApoTixq0QBLC0kOK+ax2KUhWWMr2fvVPHh3B/tDdSS8sb8+t0JO/4k2s//ISupASHNq1p+urLSMqlzC5QaVi7N5HfD6fx4oBGONjUfu/UrbUbvyRN5LmkLeRfusTZGTNpOWumwYNCEAQ+3RyNtaWM6cP8jFpBSnJyuTBbr+y4dOuCe+gNK7aPqzUPN3Hg98OpPNHzRoG7Q5dyiEouYs6EwCpl8xo+lPR//qUo9hpxP68nZMJEvtwSQ1JWMV7OVrVeszEEQSA5S42j7f23D5XIZHiE9kPdvjsvLr+Mm4MlS59vxuxPzJNU1SQFIiQ4dDz6eIetwKfooy17AntCgkOfCI8Iq9KCUA2L0KdX+6G6TvWhOmR5QtKTcF69Ek1hNk2HD8WhRXN2/HsRa2sr2rRpjVxmgYODM61atTI6XhAEzizXJ8PKateTMZ1qFzxdnsLCwirvVxvqc5XH8msrs0DI5XfMkFfnCIJAcXExVlZWRv8h3M/Ut7VpNBqaN29u+P49KM/dfbu2Ll0ojI6hs78fw5KKWLkrkWPL1tNMmkmTx8eQ6exAi5Yt7t/1VYE5K9/WJedmvIelrrIffZq9N1sfegadANLSEp498ImhrURqgVqqoFhqQbHEkq3uj5Lm4IOVhZReKRF0STpg6Gt9/QegxNqey2M/QCIBdakO+/NnQV1AmZNM9vETXPxhA5pu/fAdHIqvtyOpmzcR/4t+q9WgZzCNX3jOENys0er4/VAaK3cn0K2ZIz+/2QY3h8qn9zXh4SYOzFwn4PPODJIWf03OiZNc+nIBHZYuQiKTsW5fEpcSClk+tRUyIzGYWpWKyDnzKE5JwbaxP02mvlTp7+7k3t68sfoiY7p5YCmXUqrRMX9LDNOH+aGwqNrDXiqX0/SVl4leuRrPgf1RWEgJ7eDKH0fSea6/eWsb5xZqUJXoyMgruXXnewhNURESK2vWRySzdm8iT/dtyLgeHhQWFpntHqbufGYCM8Mjwj4pd+2rkODQd9G7JdVIgQgJDp2P3vrQOzwirNrfSll1yPpC52ZO7LFvQvvCo+SeOIlz61YUFZdiaWGBTCbD0tKC4hJNlWvOPnkKVUIiBRZKuo3pe9ufjVqtJisrC4lEYvbiX+UrPtY3ZDKZwRWmPmxGb0YikdTLdUH9WZtEIjH6jNX35+5+Xpv99QJfLRrZM3+KLYcOpqLLSCdu0SLyHd1xeHwMfn2CK6TivN+RSe+PtQgu7mBri8TGDomNLRKtBooK8Hb34s0R/kilEiTFxWQdtgCtBgQBS10plrpSQ+6jr4Y54t6nEwAZB0pJ2ZmLIAgIpRp0Gg2CphS5rS0KtwaEjLhxap8d8BqCVgsSKTnpOSRv245lcjyWuzexNg72448jbjxr48aZJo+Sbd+SwjVR5BdpyFdpyVNpaO2rZNGzzQn0Ns+Bq8JCSqdABz7bnsSrL05H8esPeAzoj0QmY9+ZTDb+m8KqV1tho6j8+9VpNFz6Yj4FV66icHejxQfvIbepnGY1qKEtjT2s+et4BkM7u7HhnxS8Xazo3sLplvLZNQmg9SdzDJ/hsM5uvL7qIk/3a2hUoaktydlqvJwVZObdH5Y00MfInHn3fQ57d+dy02DWvNaahq7mtcyA6QpEE+BXI9c3Ah/U5IYhwaFfAY+iVx7MF81xF1myZAlt27alR48et+wb6GXLQqsA2nOU7GPH8Z88kUJVCQqF/iTBwsIClbpqnSrlT72v+lmvh+kXcPs5xouLi9HpdOTl5eEo5iwXERERuSNIZDI6L1tM6p6/if1lI8qcVJK+WULM9z/iOXgQTYf3N7rpEqkbOn31GXZ2t9p828Hm9QAIWi1atRqtSnX9pxiFy42EJq7duuDazbR6QE4PdTC8dgb8B/Qm68gxMv79l7nTx4FEQmGxloz8TvgVaMgv1mKrkKK0lqO0lmNnLcPOyvyW8PcfC+Cn8CQmLb7AwIf6M8WtIRfiC/hkUzQLhjgYjVEAKM3OpvBaPHJ7e1rOeh9Lp6oVgsm9vfl0UzSPNHPg+78TWflqK5MPeMr6CYKAfeRhGtg4cuRyDl2a3VoBMZXkbDVNPG04fCkHlVpbobbFvYg6O4fDny3GqriYVs4C015oUWcHZqZ+49KBNsDVm663u95mEiHBoYvQF6TrFR4RZvK4e50LFy4glUpNUiCkUgkNO7ZCk6BAFRdPcWoqRcUlKK7HeSgUFqhUxhUIQRAQBAGtVIbfwL5m+VIUFxfj7u5OZmamqECIiIiI3EGkFhZ4DgjFtWcwp3/+Bd3R45CSQtaGn/gix51BA9rQ2s+uXljM6hsSmQy5jU2dKHkSiQSXzh1x6dzRcM3OWo6dtRyzRcCagJ2VnOf7N2J0Nw/W7ElkzGenkEslvBeQQu7cz0icMhnvYUMM/bVqNTKFAkWDBrT5fB6l2TlYe3tVcwfoEGCP0lrOy99eYGhnN3wbGK9KXR0xK1eT/MefjG/fg62Hbc2rQGSp8XKxwsXekoz8EnwUNZfvTnEt4hBXF3+DVUkhMkdHek2fXKd/O0xVIH4EloUEhzYA/kEfRN0T+BhYYcoEIcGhvsArgBqICQkOLWv6JzwibEBNhL7XKCoqIikpyeT+j3Zw59IePwKzLpHx70GKVOUUiGosEBKJBI/X3uT1uYf4tWeA0T41QafTUVpaipeXF5mZmQQE3P6cIiIiIiI1Q6qwxKLjQ7ScPJG8U6fJvHAZXz9/Zq+/irWFhMmaw3QY0hOXls3utqgiDyAuSkveHOHP+GBP4tJV+F5JJUYQiF29Fk1eHlZenqTt+RttcTHtFn4JgMLFpYJFpiokEglT+njz+W8xPNWndkXO3Hr3IuWvXdic/IfiDEeyRjbGWWmeoOekLDU+rla42luSkVeKj+u9p0BoVSoOL1yOcDgCK0DZqiWBr01FXsfxwzWJgZABXwMW6IOo1egDoU1yYQqPCLvGzaVu6wk1VSAeCrDnR9vmBGZdIu6XDZQ6tcRGccMCUVxSUOXY7UfT6NbOE6X17Zsr1Wo1lpaWuLq6irUgRERERO4yEqkU544P49zxYZoCj/Xw5NBfhxGW7eLif7vId/fDb+Qwgvp2q1dxEiL3B94uVni7WEGzQchtbbmyeCkJm34ztEsVCtQZmShca1abqkdLZx4JcsRCXrs4TLuAxvg/PYXoZSsYkbKLsIhHGD/IPMp2craaToEOuNpb3JOB1DqdwM4PF+N08TDI5PhNfgKvIYMrZOeqK0zahYZHhGmAt0KCQz9AHw8BcDU8IkxVZ5LdRxQVFZGXl2dyf7lMgk3LIHKcetCha3PyNp/ARaFP12ZlaUlxceUvqa6khLzLV9j6n4qPJgSZRe6yjDQuLi5kZNSLcBQRERGReoNUKuHhR4JISh9O8l+7UKbGkvnt14St/QFZ1950HPUojt7ud1tMkQcQt94hyGxtubp4KdYNvXF/tBcu3boht6ndCX1tlYcyPAaEkn3sGNnHT1K4ZQPCwA/M4r6TnFWMp7MCF6XlPRdIXarR8dH6KPJcu/Nki3wCX3gWW99Gd+z+NTrGvq4wnK0jWe5bCgsLycqqWRXEh/3lrFf1YtCAtqh+Pox1g+sKhJUl6sLKX9KcU6eJnPspY5wa07JRtaUzTKZMgRAtECK3S3JyCuPHTWb7js0olXbM//Jr7Gxtef7FZ+7I/T+Z9yV2dra88uqLd+R+IiJ3CktnJ/wmT8Rn7GhS9+4jaet2SEuDvZs5GLGLE6NnMrSzuxgrIXLH0cdprL3bYgB6Vyj/Z58h58w0mqef4eTe43To8/BtzSkIAinZJXg6KWjgYHlPWSAyIq8we38JCksZn03rjJVF1zsuQ5UKREhw6HLg9fCIsILrr6skPCLsObNLdh9RVFREQUEB+fn5KJXKWw8AgtylbDyhZdH2WFTFpVjJZAhaLVYKC/KNFN7JOPAfAE6tW5rtn0RxcTEKhQIXFxcuXbpkljlFRADeePM1k/rdrHiIiIgYR2ZtjdfggXgOCCXr6DHSI/7ByakB6W7WfLzhKpbqAp5M3IJ7j64EDeyNwsX5bossInJHsfb0oOHI4cRv+JXY3/+4bQUiT6VBIgGltRxXpQUxqearoXA7xOzYTcKKZXRp1pOxc6cil92dg4PqLBBNy7UHog+cFjFCUVERnp6eJCUlERRkmnuRVCphxdQWzNkYQ6OMSAoTkskdMxIbKwXqkoparqaggIwDBwHoOKqP2eQu78IkWiBEQH/iotVq77YYIiIiVSCRyXB5pDMuj3QGoAUwqZcXpzbsoOjENfI3XePopl9Q+QTRqF9PAvr0ENPBijwweI8aQYmVHbOPudCvWHNb6W2Ts9R4OutT1brY330XJkGrJfKHDWT/vhkp0K2VC7K6D3Wokio/2fCIsF7lXofcEWnuQ3Q6HcXFxXTo0KFGCgSAo60F859qxjv77NBFQ8aBg9hY2VFSqqnQL23ffoTSUrI9muLiV7ssBcYQFYgHh3FjJzF06EAiIg4QH5dAYFAT3n77DTy9PBg3dhJDhgzgwIFDREfF8NnnH9OmbWt+37KdLb9tIzMrC39/P16b9jJNm+ozdRUUFLJg/iIOHz6Ko4MDj40fU+F+N7sUJcQn8s3S5Zw/H4lOp6Vdu7Z8PPcDXnz+VQDGjH4CgFdeeZFBg/uTnJTCksXfcv58JBYWFoT278OTUyYhl+sDR//55yDLvl1JVlY23bp3EZUekQcaiURCmxF9yfZxIW1/BNnHjmMTf4mMVZdIWbOagmYP0fy1qfi7W4tuTiL1GplCQZORg2mXd4k9pzIZ/kjtY4SSs9R4OukVCFd7CzLy754LU1FcHJELFlMcEw2A/7NP4zV44F2TB0yMgbgePP1leERY0U3XrYG3wiPCPqoL4e4HytyAvL29SU5OrvF4qVSCo78PudFXyTx4COtO/SktvaHlCoJA8l+7AGg8rL/Z5AZRgahLDgwbVWVbwEvP4xHaD4CUsF1EfbOsyr7dtm42vD71+lsURkUbbTOF7dv+5NPPP8bT04OlS5bxwfsfs3zlEgB2/rmLeZ/MpqGPNwX5BWzf9idbt/7BnHmz8Pb24q+du3n7rZn8+PNqbGysWbzoW3Jzc/ll/Vo0Gi0fvP9xlfctKlLx+vS36duvN+/PmoFcLuPcuQsAfLtsEePHTebXTT8ZXJjUajWvT3+bYcMHM2v2e+Tn5TNjxiyUSiXjHhtNQkIiH82ex4cfvkfnRzrx9959fPbpAoaPGFKlDCIi9R2ZQmEoXlaan0/mwf9IC48g/0IkZGXy6vJIFBZSQlo58kjyfwQGP4x90wAxm5NIvWRYZzd+/OMij+SdwaNf31rNkXS9CjVgSON6pxF0OhJ/+524XzYgaDSU2DnS/q1XcWzX9o7LcjOmGj9mAcYclG2utz2wFBUVYW1tjZeXV41SuZZHYmuDlaMDmvx8lKnXKigQ+RcvoUpKRqWwo1m/7uYSG6gYAyFmYar/DBs+GD8/XxQKBS+8+CzR0THExFwDYPiIIfj6NUIqlWJhacFvv23l6acn06iRDzKZjEGD+6NU2nHyxCm0Wi1/7w3nmWeexN7eHmdnJyZOGl/lfQ/9dxi5hZxnnp2CtbUVFhYWtG9f9R+//w4exsbGmsfGj8HS0hIXVxfGjx/Dnt37APh77346tG9Ht+5dkMtl9AvtQ2BQU/N+WCIi9zEWSiUeof1o88kcHlrxLT2mTWH7+x2YM6EpDmmxlOzYzLm3Z7B//JP8M/NT4nbuQS0eIonUIzr52zDs8HdELf2OrGPHazVHSrYaT2crABxs5BSptZRodOYU89ZIJOScj0TQaEhp2pnuyxfdE8oDmJ6FSYLxGIimQI75xLn/KCwsxNbWFi8vL/79999azaHRaHAIDITYOBT/7cE+Nw9BEJBIJFg38iHN1gP3h9silZu3VL1arcbKygp7e3uKi4tRq9UoFMZL04vUDFOtAx6h/QzWiFvRbsEXtyMS7u43Spja2Fhj72BvUBw9PCqaeVNTUpk39wukn843XNOUasjIyCA3N4/S0lLcy425eXyFuVLT8PbyMtl1IiUlldjYOAYNHGm4JugEbK6nB8zIyMDdo2I51uruLyLyIGPl5oaVm/55ae5jh2+/piSX9CP92EnISIezR4k/e5T470Dt4onX/94jKMhDdHUSua+xsFJQ0CEE5eE/uLzgK9p++RnWXtVXxb6Z5Cw1HQLsAb23iLOdBZl5pYa4iDuBVifws2tf3Hq05dXXByGV3jvPZbU70pDg0Bj0ioMAHAsJDi3vaCwDPIBf6068e5+ioiJsbGxuywKh0Whwad8W3x49uLByFdISFQWXr6AMCuTwtRKuOgcx6Mkxt56ohpS5MEmlUpydncnKysLT09Ps9xG5N0hNTTO8LipSkZebh6urq/7CTZsFd3d3Xpr6HJ07d6w0j1arxcLCgtSUVJydnQD9pr8q3N3dSExKMijF5ZEaKXbj7u5GYFBTvv3ua6Pzubq6cv66C1QZKeVkERERqRobn4YEvPg8jQWB4qRksk+eIvP4SfLOnkOiKmTmlmQK1Ik8EuRIz5gwvJt40/CRDlh7e4vuTiL3FV2eG8fOyEs0z7tC5LzPaPP5pzWqU5GUpcbLycrwviwOoq4VCK1aTdLW7XgNH8rczXEUWSqZ+sLD95TyALd2YVoJrEZvgfgZWFXuZwnwBPBkHcpXJ+h0OuLj480yV2FhITY2NoYsTLVBo9FgYWlJw9EjcXnmZTKsnLB0daVUo+OrbbF0f/FxFI72ZpF39+7dHD+uN+eVKRCAGAfxALB9259cuxaHWl3C8mWr8G/sh7+/r9G+w0cMYdXK77kWG4cgCBQVqTh86Cg5OTnIZDJ69Q5m9aofyM/PJzs7h5/Wra/yvo906UxpSSmrV/2ASlVMaWkpJ0+eBsDR0QGpVEpiQmKF/rk5ufy2eStqtRqdTkdiYhLHjp0AoHfvnpw4eYqDBw+h0WjZvetvLl+6YsZPSkSk/iORSLD29sJr8EBaz3qPLr/8QOcvPmbTjA6seqUVrd3A4vA+0n9ax8lXpvPv6MeJeH4a5xd9R/r+fyjNzb3bSxARqRZvVxtS+k5E7eSOKj6Bq4sWIwimJRQVBIHkbHUFZUEfB1G3gdQ6jYZLn88n7qdf2D3ra64kFfLJpEDkdzPdUhVUa4EIjwibCxASHBoPrA+PCFPfEanqmEOHDjF37lx27Nhx23OpVCpsbW3x8PAgPT0drVaLrIanNBqNBvl19yQH9wYUSBQoXJz5aX8SPq5WdG1uvpPVtWvX0qVLFx566CFRgXjAGDgolHlzvyDuWjyBgU2Y/dHMKt0Uhg0fjEwmY9asOaSlpmNtbUXLls0JaqaPNXjl1ZdY8OXXPDZuMo4ODox7bDRnzpwzOpeNjTXzF37KN0uWM27MBARBoH2HdrRv3xaFQsHkJ59gxjsfUFqq4eWpzzFgYCjzF3zKd9+tZN26XyhRq/Hw9GDUqOEANPTxZubMt1m6eBlzsj+jW/cuBPc0b3yQiMiDhtTCApuG+ix/3i5WDO/WkEzpVLJPnCLr3AXIyoSUeHJS4snZu5vzPSfjG9KVNn5KnCVFWNjaIrW0vMur0BMSHKpAf8j5KOAGJANLwyPCvrqrgonccSYNaMz/zg3jteJ1ZP53mIRNv+E1cvgtx+Wr9A43Susb+7m6TuUqaLVc+Xox2ceOg40dP+la8+VTzbBW3JuWP1Od6i8A7YDD5S+GBId2BrThEWHHzC1YXXLp0iVSUlLMMleZC5NCocDBwYH09HQ8PDxqNEd5BcLeVoFGU8rL313gUkIBq15tbRY5QR/zcPLkSVq1amV4LyoQDw6+fr5MmFg52HnDxh8qimrcCwAAIABJREFUXZNIJAwdNoihwwYZnUuptGPW7PcqXCvfd8a7b1Zoa9TIh08/N56pafKTE5j85IQK1zy9PJj90UzjCwF6hvSgZ0iPKttFRERuD5mVFW69QnDrFQKApqiIwpgYci9cJPXkeRxaNGPP6QwWbotleOw2WuReotTdB6W/L54tm+DUohm2vo3ulvhyIAXoB0QDbYCwkODQ5PCIsA13SyiRO0+jBtYEtWvMtSYT8Nm5Cky0QKRk61O4lj9kc1VakF6HFoiYNd+TEfEvEisrljccxRvPd8XD6d6NSzVVgVgMLOQmBQLwAd4AuphTqLrm8uXL5OTk1DhoOCoqCoCAgADDtTIXJsAQB1FTBaK0tPSGBcLOCncHGcM7WLP53Ap83TrVaK7qOHv2LGq1mvz8fKCyC5OYiUlERERExBhyGxscWrbEoWVLGo0ZRUdgHHpXj5Mzd6DK1iBLjqE0OYa4g+HEAVq5guJOPe+4rOERYYXA++UunQoJDt0BdAOMKhBanfaBqydTtt76vu7JvTx54dscfpq/AGdfL5PWnZChwtPJskIfZzs55+ML6uTzSt29h+TtO5DIZGxsPJqhQzrS2tfW7PfS6sw3n6kKREvAmJXhBPpCmPcVV67o/aVTU1Np1Mj0E5LvvvsOW1tbPvzwQ8O1MgsEYIiD6NChQ43kKQtKBbCwsEAu0ZF4YT/79/6FICwwWzaMw4cP4+7uXkGBcHBwAEQLhIiIiIhIzZFIJHSYO4vSvHwKo6MpvBZH7uWr5Fy4iCwrg6iMu1u9FyAkOFQOdAc+r6rP5UuXsbS0uHNC3UNERkbebRHqnEA3gRUHUxmUnwXoFd/q1n3yQikWOoFz52645hZma4hN0nDuXFGV42qDNj4B1aq1AEQE9EXh35AA2zTOnUurfmAtKCkx3/NoqgKhA4xF8Tphei2JewJBELhy5QqNGzcmLS2tRgrE8ePHK/Uvb4Fo0KAB6enpNZZJo9EY4ibkcjkajYYtW7ZQUlKCSqUyzH+7HDlyhEcffZSEhARAr0C4XU/v5+LiYlCsROofxtyURERERMyFhb0Sx3ZtcWzXFu/r10qys2lRVMzMhd/eVdmARUAuUOUfwsCgQOxsbe+cRPcAWq2WyMhImjdvXuPYzfuN6Q1UvPRdJFNHNKfwyH/E/vQLrT/+EGu3Bkb7746+RisPS1q1upGZUu5YSNilGIMbuLnQBjTh6umzXC22Jtq9C4ufDKqzoOmCwkKzzWWqAnEQvavShJuuvwH8ZzZp7gCJiYnY2toSGBhIamrVqSdvJiMjg4SEBDQaTYXrRUVF2F7/o+Pu7k5aWs01xtLS0goWiNTUVDQaDW5ubuTk5JhFgSgtLeXUqVNMnjyZb775BhCzMImIiIiI1B2WTk6UWJhvw1IbQoJD56O3PvQOjwir0oFdJpXV+010Vchk9X/tAV52dGzqwJZD6XQ6chhdejqJ6zcQNP01o/1Tskt4KMChwufi5mhFZn6p2T8rmZ0t+aOeZ9n6K3w/KRBFHVrCZFLzyW6qAjET2B8SHHoS2Iu+LkQf9IXkQky9WUhw6MvAZPQBTYfCI8JMHmsuLl++TGBgIG5ubjXa7B8/fpwuXbpw9OjRChtvlUplOMVv0KABV69erbFM5YOoyywQw4YNIyIigpycHLxqWPzEGOfOncPHxwdvb28KCgoAMYhaRERERKT+EhIc+hX6TEy9wyPCxCC/B5wpfRvy0rfnGfTkeLKOHiMjPAKvwYNQNm1SqW/KTSlcAZzsLMgt1KDRCshlt+daLggCKX/uxK3Po2QVS5i9IYrZTwThan9vZDIzBZNsJOERYceBzsA5YAAwEDgLPFLDDEzJwKfoA7LvCmUKhLu7e40sECdOnKBjx440atSI2NhYw/XyLky1tUCUVyDKLBEjRozA0dGR7OzsGs9njCNHjtCpUyeUSqXRIGpXV1dRgRARERERqReEBIcuQn/Q2Ts8IqzmvsUi9Y4ADxseCXLkxzMaLLp0BiBm1RqjtSFurgEBIJdJcLCVk11w+3EEmQf/I3r5KiLnfcbMdZcY1dWDjk0dbnveO4mpFgjCI8LOAxNv52bhEWG/AYQEh5oceGDuzAiXLl2iS5cuyGQyLl68aPLcx48fZ/r06Zw9e5aoqCiaNtXnwy8sLMTKygqtVouLiwtpaWm3nPPmDAClpaVIJBK0Wi22trYsWLCAgIAAHBwcyMrKMsv6L168SNeuXbGxsSE/Px+tVotKpcLCwgKtVouDgwOZmZloNJrbDtquz5kdyq9NEASkUqnJhWnuB8rWUp/WVEZ9W5sgCOh0OsPz+qA8d/WR+rw+c2Z9MZWQ4FBf4BVADcSEBIeWNf0THhE24I4LJHLP8NoQP8Z/eYqAR7rhduYc+ZEXyTxwENfu3Qx98lUadDqwt668RXZVWpCRV0IDh9pbCjRFKmJWrQHgnDIQSwsZUx71vsWoew+TFQiAkOBQS/RFWSpYLsIjwuLMKVR5zJ0Z4cyZM3Tr1o2ioiJiYmIqRNhXRUlJCZGRkcjlcpRKJYcOHaLh9YI7KSkpZGRkcO7cOfLy8khOTjZpTriR+SA/P5+4uDisrfUl1v38/Dh37hw6nY4LFy7UKNC7Ks6fP0+XLl2IiYlBpVJx+vRpsrKySElJMcgrl8s5cuSIIabjdqnPmR0iIyMRBIGAgIB66TtaXFx8t0WoM+rL2rRaLVFRUZUU/vr+3NVn6uP6zJn1xVTCI8KuAeZJXyhSr3BWWvDyQB++3xPLV4+N49ryFUSvXIMyKAhFA1dAb33wcLI0epjqam9JRv7tfacTNv5KSWYWNPTjp8IAfnihKVLp/fd1NUmBCAkObQysAnpQ8aGUoI+HqLMdlDkzI2g0GpKTkwkNDSUpKYnvv//epGj6Y8eO0aRJEx5++GGuXbvGoUOHDONkMhnNmjWjVatWCIJASUkJTZo0wcrKinXr1tG3b99KdSFuznxgYWFBUFBQJVkaN26MjY3NbUf8C4JASkoKffr0wcHBARsbG/z9/SvIDuDm5oabmxv+/v63db/6nNmh/NrKLBBl7mf1AUEQDK5t5koffK9Q39am0Who3ry54fv3oDx39W1tUL/XZ86sLyIi5mDgQ65sjrjG34oWtAkIQFtYiNxeaWhPzlLj6WxldKyLvQWZt1FMrigunqRtf4BEwkqbnnw8MQgnu/szfbCpO5+VgAPwBJCIXmm4I5gzM8K1a9dwc3NDqVTi5eVFWlqaSXOfPXuW9u3bI5PJCAgIYP369YZxKpUKpVJpeN+gQQOysrLw8fFhzZo1uLq6MnjwYONru575QKvVolAoKsni7OxMamrqba8/JSUFa2trnJ2dAVAqlRQVFaFWq7GxsTHM7+LiQk5Ojtk+7/qc2UEmkxlcYerDZvRmJBLJXV3XuLGTmPrKC/To0bVW4+d/+TV2trY8/+Izldru9trMhUQiMfqM1ffnrr6uDern+syZ9UVExBxIJBImPGLJ57tSWT39HVyEAmTXiwpriorIPH8JTyfjBYFd7S1Jz62dAiEIAlHLViBotUR6P0TvgQ/RrrGxCgn3B6YqEJ2ALuERYWfrUpi65tq1a/j6+gJgZ2eHTqejoKAAOzu7asfFxcUZqk/7+/sTExODIAhIJJIKheRAf4qfmpqKq6sriYmJxMVV9u6KiIjA0vKG/1z5OhDlcXR05OLFi7Vaa3mioqIqVM+2s7MjPz8flUplCKIGMRNTfWbLb9sI+2s3UdExtGjRnK8XfVGhvahIxYL5izh44BByCzmDBvXnueefqtVGe+fOXXzx2UIsFZZIkODs7MTIUcMYNXq4uZZTgU/mfYmdnS2vvPqi4dobbxpPzSciIiIiIuJmL+WJnh58/mcKi55rbrh+7ft1uP21G2n/iUBlbwxXpSVXU2pXSC7n5Cnyzp2nVGHD1db9mdfz9jNs3k1MrVQRjxnclEKCQ+UhwaFW6BUXaUhwqNX1uIo7QkJCAj4+PoBeAy3LxBQTE8PPP/9c7biymAdnZ2ekUqlho12+DgToFYj09HSDkhEfH19hLkEQeO2115g2bRp79+4F9EHUxtxgHB0dycnJub1FAzExMTRu3NjwXqlUUlBQUCGNK4gKRH3GxcWZx58Yx5gxI422L170Dbk5uazf+D1fff05+/f/w2+bt9b6fv6N/fgrbCs7w35nxrtvsmL5ak6eOFXr+e5XBEEwGhir0dQ8sLQ2Y0REREREjDM+2IOsglJ+3p+MRqvTuyVbWSFBwDX8V0qyKmfBdL0NFybH9u3QjHmGnQ1DmTG59X0Z91AeUy0QbwPzQoJDJ4RHhGXdxv1mArPKvVcB+6lBLYnbIT4+3qBAAAYFYs+ePZw8eZLHH3/c6LikpCS8vW9EyJdZIVxdXY1aINLS0tBoNDg6OlayQOTl5SGRSJg+fTrvvfcevr6+aDQaQ/rW8phLgYiKiqqkQOTn51NcXIxCcSNNmYuLCxkZYqrs+khwz+4ApKZWTjNcXFzMnj37WLxkAUqlErlczrjHRrPt9z8YNXo4ly9dYfq0//HVoi9p2jSAoqIinnn6JcaOHcXwEUNuee9WrVvi6+fLpUtXaN+hHaC3uv24bj27du0hPy+fZs2DeP31V/H0qmw2Tk1N44vPFnLlShRarZZmzQKZNn0qDX282bzpd/bs/huJRMKOHX/h7OTEz+vXVrBKvDdjFgFNGjPlqUmGOZcs/o6cnFxmvv82giDw+5btbPltG5lZWfj7+/HatJdp2jSgkiymyD5u7CSGDBnAgQOHiI6K4etFX7Bly3ZkUilFqiKOHjnOxEmP89j4Meze9Tc/rvuFjIwMGjXy4eWpz9OqdUtAb1kxNkZERERE5PaRy6R89HhTPv8tmnXhSQzp2IChg8dyLSIS36wrRK9cRbP/vVlhjIu9JRm1VCASM9XMu+LKgmndcbC5P+MeymOqAjEf8ARSQoJDk4AKn154RFigKZOER4R9CHxYA/nMSnx8PO3atTO8d3NzIykpiR07diCVGjfGCIJAQkKCUQWiY8eOFepAgD4GIi0tjezsbIKDgzl2rGKZjLK5WrZsybhx4/j111/RarVGLRBOTk5mUSCio6Pp2bOn4X15BeJmC0R0dPRt3+9Oc/ToUZo1a4ZSqbx15ztEeZexuiQqKuq254iPT0RTqqFJkxsyBzZtQkzsNbRaLYFBTXlyykQ+mj2P5SuWsuDLRTRu7G+S8iAIAmfPnic25hreE288Q6tWfs+FCxf5etGXODk5su6HX5g5czYrVi6t9CzqdDpGjxlB+w7t0Gq1fL1wCXPnfMa3yxbpFZzLVyu5MJWnX2gfvv1mBU9O0Weh1mq17N0bzrvvvgXAtm072Lr1D+bMm4W3txd/7dzN22/N5MefV2NjY11pPlNk3/nnLuZ9MhufRg0N1ev37NnHRx+/zwez3qW0tJQzp88yf/7XfPLJbFq3aU3YX7v53/X7Ojs7GR0jIiIiImI+AjxtWPZyK2JTVfx+OJVnlpxDah/CB4VxZB74j6wjR3Hu1NHQ39Xegoy8mv0tzjjwH1I3d979o4Cn+jSkZaN7Z69yO5iqQPxYp1LcIeLj4w2uSKBXILZt24a7uzuXL1+mtLS0kiUgKysLS0vLCpvTxo0bEx0dTWlpKRqNpsIpvru7O0eOHKG4uJiQkBD++usvSkpKDDEPiYmJBmVk1KhRjB07tkIhufKY0wJRfkOrVCrJy8ur5MLk6urK0aNHb/t+d5JNmzbxzjvvMGfOHB577LG7LY4Bc2zs7xSqoiIUVgrk8huB4XZKO3RaHWp1CTY21owZO5Ljx08y9aVp5OcXsHL1N9XOGRMdy6CBIylRl1BaWsrESY/TvXsXAMOJ/6LFX9Lgetq8yU8+wYYNm4iJiSUgoHGFuTw9PfD0vGGZePKpiYwfNxmVqhhra+OZMsrTtdsjzP/ya86cOUdgYBOOHT2BTCrloYfbA7Bl81aefuZJGjXSWycHDe7Pxg2bOXniFN2uy1yGqbIPHzEEXz99+uWyZ79jp4d4pEsnABQKBWFhe+jbp7fBKjNocH+2b9tBxP5/DcrZzWNERERERMyPn7s104b68dLARlxOLMT5tIaYVWuI+m4F9q1aIb9+mORqb0lWQSk6nWCSC1JhbCxXvlpEqVZHs36vMbZ767peyh3DJAUiPCJsdl0LUteUxSOUr6ng7u7OqlWrmDlzJitXriQtLa2CpQH0G/7ySgfoFYjffvsNlUqFjY1NhUDTMgtEeno6zz//PB4eHiQmJhpSoyYmJuLlpQ+cadSoEYGBgRw6dMioAuHg4EBeXh46na5KC8mtKCwsJDs723BP0AdRZ2VlIZfLKwRvl4+B+OOPP+jevTuOjo61uu+dYMuWLSxcuJDx48ebJdj8QcXaxgZ1sRqNRotMpv+eFeQXIJVJUShuhCgNGz6YGW9/wFNPT8bevvrMEf6N/Vi1+ltKS0v5fu2PnDx52mBpy83NRaVSMe21t6Dcs6PT6khPy6ikQOTk5LJ0yTJOnTxNYWGRYUhubq5JCoSFhQW9evdkV9geAgObsCtsD3379jY8Uykpqcyb+wXST+cbxmhKNUbd+UyV3cPDvdJYD/eK19LTM2jXrk2Fa55enqSnp1c5RkRERESk7rCUS2nlq0RoOID0iH8ouHKVlJ1/0XDUCEO7jUJGbpHmlulXNUUqLn42H11JCVc82vH6053rRQbAMupPAvtbkJubC+g35WW4u7sjlUoZNGgQO3bsIDk52agCUX7zDXoXpujoaAoLCysVXSuLq4iLi6Nx48b4+PgQFxdXQYEof49x48ZVqUDI5XJsbGzIy8ur9UY+JiYGPz+/CoqCUqkkJSWl0olmmQKhUqmYMWMG8+bNY8iQW7up3C3Wrl3Ll19+CcDChQvvsjT3Lz4+3sgt5ERFRRMY2ASAK1ei8PP1NXxv8vML+GrhUoYMGciGDZvo3bsnDX1uXTnTwsKCp56ezHPPTuX3LdsZPWYEDg4OWFlb8c13X+Pre+siiSuWr6aosJDlK5fi5ORIcnIK48dNNlhLTDkF6hfah/+9+S4TJj7GgQOH+G75IkObu7s7L019js6dO1Yzgx6TZTf2T+KmSw0auJKcnFLhWkpyCm3btq5yjIiIiIhI3SORyWjy8ovknD6D15BBFdr0bkwl1SoQgiAQ9c23FCclkWrVgIGzX8PKon6lNK7yWDskOPRySHCo8/XXV66/N/pz58StPXFxcfj4+FTQ/po1a8Zjjz2Gm5sbnp6eJCcnVxpnzALh6+tLQkIC+fn5FeIfQG+BuHr1Ki4uLtjY2NCoUaMKgdQ3KxChoaF07ty5SveE242DiI6OrhBADXoFIiMjw1D5uowyBWLv3r0UFRXd06f6+fn5REdH06FDB5o3b86lS5fQ6XR3W6x7Fo1Gi1pdglarRdDpXZPKfOqtrKzo06cXq1aupSC/gNSUNDZu2Mygwf0N47/84iuaNQvkjbdeY8SIoXw0e57JPvlSqZQJE8fz04/rKS4uRiKRMHz4EJYuWUZKSiqgV1DCw/8xOmdRYREKKyvs7GzJz89n1Yq1FdqdnJxISko2KBTGaNmyOc7OTsz/YhGNfH3w9/cztI0YOZRVK7/nWmwcgiBQVKTi8KGjRp+7mspeHX37Pcqe3fs4feosGo2WnTt3ERsbR4/gbjWaR0RERETE/Nj6++E9fCiS6wdpSdv+IOvIUdxtJGTklSIIAur0dDIPHUZX7u9/esQ/XPpiARn/HEAttcR/+nQ8Pe7feg9VUZ0F4ieguNzrO1Y8ri4on8K1jCZNmvDxxx8D4OHhQUpKitFxZbUjylAoFLi7u3Pp0qVKCoSzs7Oh4ByAj49PhVSuN1s0FApFtSlkbzcO4ub4B9ArEOnp6RXiH8ruVVBQwG+//UZoaCiRkZG1vm9tEQSBVatWkZiYyKuvvoqTk5PRfsePH6d169YoFAoUCgVKpZKEhIQKLmoiN1j3w898v/ZGKFNo3yG0bdfGUA/ilVdfZMH8xYwbOwm5XMagwQMYOWoYANu3/cnFi5dZuUof9/DklImcOnma5ctW8/LU5026f3BwN1av/J4tv21j/ONjeebZJ1n/y6+8+foMMjOzUCrtaNuujSFOojxPPjWRTz+Zz5DBo3FxdmbipMfZs2efoX3Q4P7M/nAeQwaN1gc1/7TKqAz9QvuwauX3vPzKCxWuDxs+GKlUyqxZc0hLTcfa2oqWLZsT1Kyp0XlqInt1tGvXhmnTX2b+l1+TmZmJj09DPvv8Y1xcnGs0j4iIiIhI3aIpUhH7/ToEjYZRMgvUF3w4kpuOJj8fgDZffoayqd6Cn3vuPJkHDgKgHjKB9o8E3TW565IqFYjycQ/Xsyfd19ycwvVmPD09SUxMBCA1NZXU1FTatGlDYmIiXbtWrobr7+/PuXPnKikQUqkUV1dXw6bd19eXEydOGNrLLBpJSUkmye3g4EB29o1cxJGRkXz++ecsX77caOrXm4mOjqZv374VrpVZIG5WIKRSKY6Ojhw5coTNmzczZcoUk2Q0F2q1mhkzZhATE0Pbtm3p378/8+fPp3v37pX6Hj16lM6dOxveN2/enMjISFGBqIIpT01kylMTq2y3sbExpDRVqVRYW1sbrHVDhg5kyNCBhr5yuYwl31TtMjZgQD8GDOhX4ZpUKuWHH1eWm0POhInjmTBxvNE5Nmz8wfDa17cR3373dYX2/gNufKe9vb1YvmJJhfYZ71ZMvQcwYeJ4Ro0eXsnyJpFIGDpsEEOHDao0xhg1kb06eQBC+/cltH9fo21VjRERERERucPodPiMG0PW4SMUXI1ClhCNBpArldgFNIZyFnDFQ4/wb5Qcl0B/nnnq0bsncx1jUmRuSHDo3yHBoZWc8EOCQ+1DgkP/Nr9Y5udWCoSHh4fBhWnTpk28/bZ+M2XMhQn0gdTGFAjQZ3cyZoEoK95W1am6MW52Ydq6dSuHDh1i9erVJo03ZoGws7MjPT3dqNuUq6srvXr1IjAwkKKiIrKybqfsR81Yu3YtWVlZ/PLLL3z44YfMnz+fd955h6KiylUfjxw5QseON3zWmzVrdlcsJiIiIiIiIiL1G7mdLT5jR9N2/uckPvsxF/o8y8Mrl9Fp3Rpazv4AZaDeYp2cpWba3wJOfUN5+rned1nqusXU1D4hgLGK0Qqgh9mkqUPKV5M2RnkXpjNnznD58mUuXLhQpQLh7+/P+fPnKwVRgz49a5nVolGjRsTHxxuUEW9v7xpF4Zd3YRIEgbCwML766itWrFhBbGxstWO1Wi2xsbGGAO4ylEplpRoQZfj6+jJy5EgkEskd35Tv2bOHp59+2iBX9+7d6dChAytXrqzQT6VScfHiRdq3b2+41qJFC1GBEBEREREREalTnLzdiLL3R9HAtcJ+LiqliGeXnGNsdw+e6+9TrzIuGaPaLEwhwaHl/UEahgSHlt9xyoCBQOXAgXuQsiDqqvD09KygQAwfPpw1a9YgkUiMpqz09/c3GkQNMGHCBMNrpVKJQqEgMzOTpKSkShmdbkV5C8Tly5fRarX069eP+Ph4Zs2axdq1a6v8kiYmJhqCuctTVtPCmAKxdOlSw3zNmjXj4sWLdOtW90GdWVlZXL58mU6dOlW4/vbbbzN06FBGjx5t+OxOnjxJUFBQBVcU0QIhIiIiIiIiUte42FuSmKnmYkIBOgF0OoH0vBI+2xTDtKG+9H+owd0W8Y5wqzSuseiDpwXAWIUxHTDDzDKZHa1WS3JycrUWiAYNGpCVlUVCQgIajYapU6fSv39/mjRpYrR/WWYjYwrEzQQFBXHw4EFyc3MrpYm9FY6OjoaiZLt27aJfv35IJBImT57Mxo0b2bdvH717GzeTGXNfguoViPLKSPPmze9YYbnw8HC6du1aya3K29ubkSNH8vPPP/Pmm3qf8KNHj1ZSNHx9fcnJySEvL++WNQrMhUwmE6sDi9w1dDqd0fTPIiIiIiJ1h6+bFRIJzNkQhVQqQSoBuUzKB48F0LW56S7q9zu3+u/TA30m8ghgGFDeIb4EuBYeEZZWR7KZjdjYWJydnY1umMuQy+W4uLiwe/du2rRpg7+/P23atMHZ2XhGFA8PD6ytrY26MN3Myy+/zHvvvUfv3r2rVWKM4ejoaAii3rVrF++//z6gz6//7rvvMmfOHLp3726odlseYylcQR8DAbeubNu8eXPWrVtXI3lry759+3j0UePBRt27d2fZsmWG90eOHOHZZ5+t0EcqlRIUFMS5c+cM7mOCIJCQkEBcXBydO3c2+2ZLIpGg1WrNOqeIiKncToFJEREREZHa4aK05IfpbW7dsZ5T7Y4qPCLsAEBIcKg/EB8eEXZfJtr/8ccfGTZs2C37eXp6snv3bkN2nylTppCammq0r0Qiwd/fv1JGF2N07doVPz8/Nm7cyNy5c2ske1kMxIEDB0hNTeWhhx4ytIWEhLB27Vp++uknoxmToqKiaNmyZaXrFhYWWFlZVatQAQQGBhITE0NJSYlRBcVclJSU8O+//zJr1iyj7e3bt+fcuXNoNBq0Wi1nzpyp8DmU0bVrVyIiIujatSuCIDB27FgSEhKwsrJiwIAB/O9//zO77AqFApVKhUwmq5PNXElJCXl5eRQVFWFjY2MoLJifn4+Pj88tf4c1RRAEtFotGo2m3vlv1qe1abXaCsUhRURERERE7iSm7nhewEhN1JDgUIeQ4NBfzSuSecnNzWXr1q1MmjTpln09PT05evQobdroNcuBAwdWm8rU39/fJBcmgP/973+oVKpaWSDOnDnD9OnTWbhwYaVNwxtvvMGaNWuMFlGLjo6uFEBdhlKpvOXm08rKCh8fH44cOWK0XRAE9u7dW+UpvCAIJqWrPXDgAP7+/ri6uhptt7e3x8vLi4sXL3L27Fn8/f0NbljlefTRR9m3T18f4NKlS6SmpnLw4EE2b97M9u3b2bVYunlZAAAgAElEQVRr1y1lqSmRkZH8+uuvZrFuFBYW8sILL7B161YA/vnnH/r378/KlSuJjIxk5cqVTJo0iU2bNnHy5ElGjhxJVlYWSUlJ7Nq1yyzWEJ1OR1RUVL0sylef1mZpaWlSGmcREREREZG6wNRdz2QgOCQ4dHx4RFgcQEhwaFfgZyC72pF3mfXr19OrVy88PDxu2dfDwwOdTkfr1q1NmvvVV1812d++efPmrFixglatWpnUv4yGDRvSqlUrPvroI6PuSK1bt8be3p7Dhw/TpUvFQlZVxUCAaQoE6BWU999/n+3btxtcn8pYunQpixcv5vHHH+eDDz6odKq7c+dO3nrrLdasWVNJtjJiY2OZMWMGn3zySbVytG/fnpMnT1JQUFAp/qGMVq1akZOTw7Vr1/jjjz8YNGgQEokEZ2dnlixZwtNPP8327dtp06YNTz31VCVlTBCEGp9ML1q0iMOHDxMfH8+7775b65NtrVbLm2++ibW1NUuWLOHgwYNERESwYsUK2rVrB8CgQRXrFKSlpTF06FA0Gg2enp5s2LCBhQsXVul2ZwoSiQSJRIJcLq/TE+7k5GQWLFjAnDlzUCgUnD59mjfffJNvvvmGpk2NF3C7Xe7U2kREREREROo7plog2gEFwKmQ4NDRIcGh7wHhwDagc3UD7yYqlYoffviBZ555xqT+Hh4eeHt7V3kSfjNNmjTBzc3NZHl69epVY1cgZ2dnfvzxR6PKQxkjR45k8+bNFa5lZ2dTUlJSpXxl2aFuRd++fenUqROzZ89m1apVjBgxgnnz5rFu3TrWr1/Pn3/+yaFDh1i+fHmFcVqtlsWLF/Pss8/y2muvcfXq1QrtgiBw/vx5pkyZwrRp0+jVq1e1cnTo0IETJ05w5MiRKhUIqVRKr169+Pvvv/njjz8YMmSIoa1t27Zs3ryZvn37smPHDtavX19h7JkzZ+jevTvnz5+/5WdSRkJCAseOHWPnzp2cPHmSd999t8ZWgKtX/9/evcfXcOePH38lISVCg6ARErYhiWvLtijGpKumdlu6qlVatX205Vssartqt11r6Y/a0iqt+rXbVltKxbq069vvuLTT0St1bwSVr0sEDRHkQkJyvn985sQRuRxyPfF+Ph55cOYzt3cmZ+ZznwOsWLGC8ePHk52dzZw5c4iPjycnJ4eFCxcWFB6KMm7cOObOncumTZtYtWoVsbGxPPjgg5w6deqazqGy5eTkMHr0aL799lvefvttXC4X06dPp3379gwfPvyK2bRWr17N6NGjq31MQojqQ9eMMbpmbNY144KuGVZVn48QNZFXLRDOQGlD14zZwCdAHjDYss1Pr+VgumbUAuYAw1GFlxXAGMs2c65lPxkZGaxYsYIePXoQExPDzz//zIoVKxg+fHhBF6FLly4xfvx4evXqRWxsrFf7jYmJKTUjWx0NGDCAefPmkZmZWdBK4B5AXVyNeHBwsNf951988UUeeeQRLl26xIQJE/jhhx/45JNPWLhwIbfeeivvvfceDz74IJ07d6Z79+4AfP3119x8881MnDiR1q1bM2zYMEaOHElcXBzr1q1jzZo1ZGdnM3LkSB555JFSz6FLly7MmzePc+fOMWfOnGLXi4uLY9q0aQQFBV113SMjI4mMjCQ6OprHHnuM/v3706hRI3Jzc5k8eTK9e/fmqaee4uOPPy6269eZM2dITEykQ4cOfPzxxwwaNIiwsDA++OADRo0axYQJE5gzZ06pBcWNGzeyYMECjh8/Tvfu3enQoQNDhw4lMDCQ0NBQ3njjjRK3B1Vgcv++ASZPnkydOnUYOXIkS5Ys8Wp8DqjvSn5+fsE5Z2VlkZiYSGBgIJGRkV4VNEEVDGrXrn3VWBCXy8W5c+c4deoUJ0+eZOnSpbRo0YI33niDgQMHEhAQQG5uLq+++iqmafLoo4/y+9//nuDgYFatWkW/fv0YOHAgc+fOveLlgeXN5XKxatUq7rnnniK7yFUn+fn5LFmyhB49ehQ7U1xRtm7dyuLFixk9enSFtPSkp6eTn59P48aNyc/PZ/PmzWRnZ9O0aVOaNWtGo0aNSmz9SU5OplGjRtx000189tlnrFy5ktDQUMLCwjh06BANGzZk7NixhIWFFWzjcrlIS0sjJSWFOnXqEB0dXep5ulwudu/eTUZGBrGxsTRs2LBgUHpFjY/ZtWsXn3/+OQcPHmT06NEFXWWv1ZEjR5gyZQpZWVk8+uij9O/fv9jvqMvlIikpie+//x5N04iIiChyvRroOPAycAdQdPO3EKJMvO64rWvG71Bdmb4H2gITdM3YYtnm8Ws43l+BPkAH4CKqBWMmMNGbjY8ePcqGDRtYuHAht99+OwsXLqRly5YcPnyYvn37MmjQIKZOnUqzZs1YsWIFFy5cYPr06V6fXM+ePSvlnQflLTQ0lG7duhEfH8+gQYNITU1l2bJlJbZaeNuFyb3u2rVrCz736dPnigHJYWFhzJgxg0mTJrF27VpycnJYvnw5s2bNws/Pj0GDBtGuXTvmz5/PO++8Q79+/XjppZfo0qWL1wOPW7duTVZWFmFhYSV20enVqxenT5/m4YcfLjYjEB0dzYABA5g5cyYTJ05k2bJlhIeHM3PmTOLj4xk2bBgTJ06kX79+LF++nISEBMaOHUtQUBAjRozg5MmT7N69m3Xr1hEfr4YA1atXj3fffZdx48YxbNgwpk+ffkUB5tixYwWFg5UrVzJ79mymT59Onz59ynV2qPHjx3P48GFGjx7NtGnTaNmyJRkZGWzbto29e/eSlpZG48aNCQkJITAwkJ9//plVq1Zx6dIlHnroIfLz81m6dCnh4eHk5uaSm5vLCy+8UDB9sFtOTg4JCQm0atWKRo0a8eWXX/Lcc89x8eJFoqKiiI6OpmXLluzcuZPvvvsOPz8/QkNDadKkCa1atWLKlCkEBQXx1FNP8corr7B06VL8/f3p378/Xbt25Z133mHnzp3Ex8fTrFkzevXqxZgxYxg5ciRDhgxhzZo1HDhwgPDwcKKioujYseNVLYfbt2/nb3/7GykpKYSFhdG5c2eioqLIz88nMzOTXbt2sX37duLi4hg8eDBTp05lw4YNxMfH8/777xd8PxISEvj0008ZMmRIid8pUBm2HTt2sGXLFgICAqhbty5NmjTh7NmzmKZJeno6ffv2pX379pw/f54DBw6wadMmUlNTCQ4OpkWLFnTs2JGWLVvSoEEDIiIiiIiIuOp3P2nSJJKSknj99dcZN24cQ4YMuervyOVy8dlnn/HBBx/Qrl07QkJCWL58OYMHD2bYsGEMHDiQ5ORk9u7dS15eHqGhoYwZM4a+ffteVyY6KSmJxx9/nKysLG677TaSk5O56aabCAsLIzU1lV9++YVz587RoUMHevXqRXh4OPXq1aNevXrk5eXx4YcfkpCQQHZ2NrVr16Zdu3YF+zt27BgREREEBAQwYMAAhg4dytGjR/npp59ISUmhbt26hIeHc+LECQYMGMDEiRPx8/Nj+/btfPTRR+zYsQNQ39PIyEiOHj1KTk4Ot9xyC4mJiWRkZODn50d4eDi6rtOvXz+6d+9OQEAAeXl5rFy5knnz5hEYGEj79u1p3749sbGxhIaGkp+fz4oVK/jss89o3bo1PXr0YMSIEQV/j5mZmUydOpXNmzczYMAAevbsyahRo+jWrRtDhgyhY8eO/Pjjj3z//fccOXKErKwsNm7cyKFDhwgJCcHPz4/U1FT8/Pxo3rw5+/btY9SoUbRq1YrFixczY8YMHnroIRo3bkxiYiJnzpzh4sWLpKenc/z4cQIDA+natStz587liSeeID8/ny1bthT8vf3617+mY8eObNmyhW3btnHPPfdw1113sW/fPr766iuOHTtGXl4ezzzzTJHTkO/fv5/169cXTFjx8MMPEx0dTUpKCqdOneLixUvX/LdUVpZtroSr3mUlhChHfi6Xq9SVdM2YC4wGpgMvAc2BxUB74AnLNteWsLnnfo4AEy3bXOF8NoBlQOPCMzyNeW7KzcCZPdu/ITcnl0sXL+Lv788dd97B8OGPExsTw/kLF/jmm6+58447adCgAdt3bGfOnDkE+AcQ3iKcv/7lr1f1268O8vLz2L9vP22j2xLgXz59sbdu28r0adM5nX6a4OBg4uLiGPrI0GIHbc98+WXaxcZ6NTuVt2bOnMmexD2kHE2hx109mDZtWrnFBzBu/DjCwprzl8mTS1zvo48+4p5+/bilWbNi18nIyGDChAkcTUmhdu1avPvuuzRrqtbfvXs38+fPZ+euXfzmN3cTFRXFksVL8Pf3Z8SIEcTExqhM98VLzJo164r95rvyWb16NW8teIvomGhat25NUlISexNVJi0mNpZDhw6xcOFbtG5VdCtHWeVezOXdf71LfHw8kZGRHEg6QGxsO6Kj29K4UWPSz6STcS6D3Iu5NG3SlPvuv4/atWqzPH45eXl5dO/WnZ69ehLgH8CWH7cwe/YcDh06RL2gIOrUrUOdm+qQejKVFi1acPzY8YKYZr38MrfeeitJSUkcSDrAkSNHiI2JpXuP7jQMKXpu7NyLuXz37Xf06dOn1LiOHz/OpOdVxrl3r9507NiRE7+cICkpicQ9iYSGhhIXF0e9evXYuXMnexITefbZZ+nW7U5OnPiFLZs3Y64zAQgOrk/btm1o1649q1at4uf9+4mOiWbO7DnMmDmDM2fO0KFDR/bt3Uvi3r3cfXcc5v+Y/O6+39GgQQOyMrPIzMwkMyuTzMxMci7k4Ofvz8lUNat1z1498fPzIzsrm9OnTxMYGIgeF0fDhg358ssvOXz4EEF1g2ge3pwe3XvQPLw5WZlZJB89yp49CaT+ksrZc2c5fPgwF85fIDommtiY2ILCYFRUFNNfms6J4yeYMXMmCQk/EXVrFFlZWZy/cJ769euTl5dPrVoBjBo5iiNHjnDw4EGefvppwsLCSE5OZvXq1bSNbktsTCy1a9dm//79LHjrLS45hcCgoCBSUlLIycmhbdu2REREEBISAqiWhtOnT3M6/TR16wbRqlUrFi1axFinALLp6000bdKUzrd1xs9j/o2c3Bx27NjB5h82k5aWRnZ2NlnZWeTm5HJv/3u5//778ff3Jy0treD7CFfeM5OSklizZg1t27Qltl0sLcJbFExmkX4mnVmzZrF+3XpqB9YmIiKSwYMfpHev3vgH+JORkUFycjINQxoWnJsLV8H4pwMHDvDN19+wfsN6Tp1KIyQkhLRTp4iMjGT8+PHUC67Hvr37SNybyP79P6vMem4ud999Nw888ADHTxzniy++KPhbOZ99nu9/+J5ud3YrGOMEqpVv1apV/Gftfzh48BAd2renYcOG+Pv7UzuwNr17a7RpE8W5c+dwuVyENg5Vk1IcP07LFi2uuK8fOnSIlatWkpt7kbZt2xDaOJRatWoREhJC49DGNG3aFD/8OHr0KG8tfIvQ0FC6dulKTm4OR44cYevWrexJ2EOnzp3o1KkT69at52RqKkH1gujTpw8REZGknz5dUEHVokULAgMDSUtLU9vu2cO9/e8lIiKCtLQ0Vq9eXfA+niZNmpCXn0+LWzsChLw5e9pZ7+9kZadrxgTgAcs29eLWcec1pkz6I/W8nBSlpqiIvIgvuFHjBsjKzmbaP+dDOXwfvS1AHAEetWxzk8cyP+BF4EXLNkvt46BrRghqwHWUZZtJzrImQCrQyrLNw57rj3luSkvgyDXEIoQQQojqKeLN2dOSK/OAXhYgJK8hbkRl/j5623fiNss2PV8ih2WbLmC6rhkbvNyHu1OxZ4nnTKE0T0eBCOCcl/sXQgghRPXTAPVMr44kryFuNOXyffR2EPXpEpL3eXmsDOffmwH3lCohhdIKvDl7mguo1NoKIYQQQpS7Su26dC0kryFuQOXyfSxxBKuuGed0zQj1+LxG14xbPD43A056cyDLNs+gvqSe81LejmqFkC+vEEIIIcpM14xaumbUQVWS+uuaUUfXjGubQ10IUaLSWiCCubKQcTdQeJTRtUzX8S/gBV0zvkPNwjQVeL/wAGohhBBCiOv0IvB3j8/nga8AvUrORogaqDzmjyx9FPZlM4BQIAFVMIlHTe0qhBBCCFFmlm1ORVVQCiEqSPlNQO8FyzYvAeOcHyGEEEIIIYSPKe0tXi6ubmG4lhYHIYQQQgghRA1SWguEHxCva0au87kO8KGuGeedz9c1KEnXjAXA/agZmTJQXZkmWbaZq2tGLWAOMBxVwFkBjLFsM8fZtsT06kLXjLrAbiDUss0QZ5lPx6ZrxiJgGJDrsTjOss0tTrpPxwega8b9wD9Qb1vPAF61bPMVX49N14zMQotuAhIt2+zkpPt6fM2B+ag33fsBX6PO72gNiO1WYB7QAzV27D3gBcs2830tNl0zxgAjgE7A957z8+uaEQwsBAYAOcC7wF+cKcPLnF4ZSolvGjAQaAcssWzzD4W2DQPeAeJQk4vMtmzzNW/TK1pxsema0RR4DdBQMyseBmZYtvmxx7ZVcu1uhLxGcUqJfRFleJb7gpqYB/NGMXEvogKud2ktEB8AB4EU52cxcMDj80Hgw2uIze0NIMayzQaoWZk6A887aX9FZQI6AG1Qb7ue6bFtaenVxTSunme3JsS2wLLNYI+fLR5pPh2frhn3oh5if0Y9CKOBz51kn46t0DULBhJRb4F38+n4gAVAbaA10BLIQmW2wIdj0zUjAPgU+AkIA34N/Bb1Nwq+F9tx4GVUhrOweUATIBK4ExgM/LEc0ytDSfH9L+p6fFxEGsAS1KyGzYD7UBOODLiG9IpWXGzBwA7gLtT88uOAt3XN6OGxTlVduxshr1GckmKHsj3LfUFNzYOVpqi4oQKud4kFCMs2n/Dmx/u4Cva7x7LNLI9FLiDK+f9TwEuWbR6zbPMkaiDUE7pm+HuZXuV0zeiCesgXvgA+H1spfD2+6ajz22jZ5iXLNs9ZtvmTk+brsRXQNeNOVC3oIo/Fvh7fr4BPLNvMsGwzG5XZ6uik+XJs0c7PFMs2cyzbTEZl4P7LSfep2CzbXGnZ5krgF8/lumYEoWrIXrBsM92yzYPAK8CT5ZFeWYqLz0lbZNnmWiC9cJquGb9CPcAnW7aZadnmdlQB+Elv0itDcbFZtvm/lm2+YtlmsmWbLss2vwA2o1rMqvTa1fS8RklKib00Ph37jZoHKyHu0lxX3JU6iNqTrhmTgRdQtRdpwCRdM0JQtYfbPVbdhqoNbqlrxtmS0lFNp1XKaQp6BxhTaLnPx+Z4XNeMx1G1Ue8Br1mqK4VPx6drRj2gK7BM14w9QGPgW2A86g2lPhtbEZ4EPrds8xjUmL/NV4GHdc1YC+ShmmLX1oDYipom2w9opWtGBL4dm6e2qC6xOzyWbQPaO60wZUq3bDOvIk++HHQCjlm26Zk53wY84mV6taFrxs2odzzNcBZV6bWrqXkNbxQVu0fydT3Lqeax3wB5sCIVF7eHcr/eVVaqsmzzZcs266NqQt8GTgD1nWTPt+Sdcf6t70V6dfAnYJdlm1ah5TUhtnmo2tAmqBLreOcHfD++hqiM2QhUCb41kAr8G9+PrYBT2/cI6p0sbjUhvm+ARsBp1Lm1BSbj+7HtA5KA6bpm1NU1oxWXv3Pu/uG+Gpun+kC2pWbqczsDBAB1yyG9uqvP1W+HPcPl61RaerXgZGI+AmzLNjc4i6v02tXgvEapiokdyvYsr+5qch6sJMXFDRV0vau8WcayzURUzcMi1EAfUIN+3EKcfzO8SK9SzmDHMcBzRST7dGwAlm1us2zzpGWbeZZtfofqDzvESfb1+NznMM+yzUNON5gXUH3O3TVgvhqbp4eBbGCtxzKfvnZOM+t64DtUTVt9YANq/IpPx+Zkqu5H9U09DPwPqnuWC3C/gNMnYyskAwhyMqBuIajv3vlySK/uMrjyOoE6/wwv06uc87tfDNTjypaRanHtalJe41oVir2sz/Jqq6bnwYpTStwVdr2rvADhCADaWLZ5BkhGDfhxux1VGkouLb2SzrUkvVElvARdM04AK4EGzv/b4duxFaXgDeK+fu0s2zyLyqB5zvrh+X+fja2Qp4APPGv7fP3aoVoeIlGFv2yn8Dcf6IbqpunLsWHZ5n7LNn9r2WZTyzZjUAXALZZtpuDjsXnYj5ohpLPHstuBPU4XlrKmV3e7gObOjEZut6NmU/EmvUo5mf+lQFPgfss2PTP+1ena1ZS8xvUIQA2QLYrXz/IKO7vycaPlwdyKjVvXjLuKWL9crrefy1W5r3Vwpmt7CFiFajLpAHwCfG3Z5khdM6YAD6BmmbgIrEFNGTfR2b7E9KrkTJ/lWYq7C3gf1XR0GtWlwidjA9A142FUDWgGarzACuBNyzZfcdJ99tpBQX/RYcDvULOdvArcYdnmHb4eG4CuGdGo2ZdiLNvcXyjNp+PTNeNn1N/jP5xF04Chlm22rAGxdULN4HMBNYXnh8Djlm2u97XYnIxmLWAsasrOfkC+paaWfA8109RQVA3YOuANyzbnOduWKb0axFcblYl7xTm/p4E8yzYvOtt+gZrZcBwqo7ceeNKyzU+9Sa+q2FAVLctQ48Z+6xTgC29b6deuJuc1SuNF7GV6lldXNT0PVhwv4n6ACrjeVTGI2oXKpM1BDZxy9zP/u5M+AwgFElAtJPGoKabwMr3KOLUuBTUvumacBlyWbZ5wPvtsbI6xqH6UtVDT+C5AXUc3X4/vn6ixENtQ5/c18KCT5uuxgRo8valw4cHh6/ENRM1OlIIay7IDlckB34/tIWA06j08e4GnLNtc76T5WmwvcvleD+p++RWgo/rkvgUcQdVI/wvVkuRW1vTKUFJ876DGWLk9hpoq/Q/O50dR55yKyvTNKFQ4KC29ohUX29+BQagCbqquGe70xZZtumcLq4prV2PzGl4oLfayPsurpRsgD1YkL+KukOtd6S0QQgghhBBCCN9VXcZACCGEEEIIIXyAFCCEEEIIIYQQXpMChBBCCCGEEMJrUoAQQgghhBBCeE0KEEIIIYQQQgivSQFCCCGEEEII4bWqeA+EEMIH6ZqhA18CLS3bPFrFpwOArhlTgccs24y6jm2DgX3AA5Ztbinvcyt0rPrAz4Bh2ebOijyWEEKUJ7n3X7+afO+XAoQQ1YCuGU8CC4FGlm1meCzfBcQWs3y7ZZsjrtpZDaRrRgsgGYizbNMqp90+D/xY0Q8QAMs2M3TNeBX18p6+FX08IYRvkHt/yeTeX31JFyYhqocNqAJ9H/cCXTNCgfbA8SKWdwA2VvI51hi6ZtQBngH+fyUedhHQR9eMDpV4TCFE9Sb3/kok9/7yIy0QQlQDlm0e1jUjCfgN8B9n8d3AT4BdxHI/nIeIrhmtUbUb3YEQIAn4p2WbHznpTwOzgVucV97jLH8eGIdqls7XNSMKmOUcywVsAf5k2ebu4s67tG10zfgD8C/UQ3A+EAMkAP9l2eZWj/30BV4D2gD7gT8CFjDcss3FqBoogC91zQA4bNlmK4/tBzrn0RLYDDxl2WZScecN3AvUBdZ57KMVcBAYAgx3YjoB/N39u3TWczm/t17AfUAaMBFYDywABgAngT9btvlv93aWbabqmvEt8BgwuYRzE0LcIOTeL/d+XyUtEEJUHxtRNy63u4EvnJ/Cy/datpnifA52tr0X6Ai8Dbyva0ack74cCAQeKHS84cBi5wHSDPgaSAV6ox5I+wBL14wmRZ3sNWzjD8wExgNdgHRgua4ZtZz9hAOfAj846c8CrxY6XBfn3weBMOAOj7QwVI3So8BdqAfpe0Wds4c+qG4Al4pIexn4COiE+t29r2tGm0LrvAD8N9AZ9XD/EFiGepDcDqwFPtQ1o3Gh7X4A4hBCiMvk3i/3fp8jBQghqo+NQAddM5o6n+9GDVz7CmhXaPkG90aWbe62bPNNyzZ3WbaZZNnmfNRNbJiTfhZYAzzu3kbXjC6oJvIPnUXPAIcs23zG2d8+VE3LGdTNuSjebuMHTLBsc5Nlm3uBKcCvgFud9NGoB9Eoyzb3WLa5EXWT9nTS+fe0ZZsnLNs86ZF2E6q2aqszSG0W0Ntpqi5OayClmLQ3LNtcbtnmAeBF4ALqd+5pmWWbHzjr/B1Vo3XAss1FzrIpQBDQo9B2R53YhRDCTe79cu/3OdKFSYjq4wvn39/omrEJdbP5yrLNs87AOffyNsCf3RvpmhGEumndj6qRCUTdWL/02PeHwKe6Ztxi2eYJVA3UVss2E5z0O4CuumZkFjqnus7xiuLtNi7Ac/YJ9827GarWqh2wxbLNPI91vivmmEU5VuihkoJ6cDUFjhSzTV3gbDFpO9z/sWzzkq4Zvzjn6mmnxzondc3IA3Z5LEvXNSPXOQdPF5xjCyGEm9z7L5N7v4+QAoQQ1YRlm6d0zdiJarIOBLY5NUigHgju5XmoPqJurwADgT8Be4EsVL/Ymz3WMVE1OY/qmvE6MBSY4ZHuj6oFG1vEqRV3s/V2m/xCDwiXx/aFlxX3uSS5xWxbUgvrSaDRNeyv8L4uFrFd4WVFbdeIyzVqQggh9/4ryb3fR0gBQojqZSOqr2dtLtdKgXqIzHeW/+jxcAHQgCWWbX4CoGuGP9AW+MW9gmWbebpmfIxqyk5E3cyWeuzjR+APQIrnYLtSXM82RdkDDNM1I8DjYVO4+dd9Yw8ow3E8baPoh19F64j6vQkhhCe59yty7/cRUoAQonrZiKpNagwM9lhuo2aZ+D3qYeJpHzBQ14x/A5moWSGa4/EQcXzgpP0/4PNCTb9vAE8Cq3XNeAk180ULoD+w1rLNb4s41+vZpigLUIPn3tI14zXgFucc4XKN0ikntn66ZiQAOZZtpnu5/6J8DszRNaOlZZvJpa5dDnTN8EM98P9WGccTQvgUuffLvd+nyCBqIaoXG9UcWgc1ywUAlm2eA7YC9fEYROd4FjiMqqnaiOoHuqLwji3b3IXq45K+rYMAAAEySURBVHkblwfQudN+QdX8nAJWoh5MS4BI1FzkV7mebYrZTwpq+ru7nPN7HTWADVS/USzbzAfGAA+jHlbbvd1/McdMxJkqsCz7uUY6ataU5ZV4TCGEb5B7v9z7fYqfy3Ut3c2EEKLi6ZqhoWYg6VTSXORlPEZv1PR7bSzbzK6IYxQ63n+jBkbOquhjCSGEL5J7v++QLkxCiCqna8YzqNktjqFm5ngN+KGiHiAAlm1u0jXjH6hp/RJKW78sdM2oj5pdZG5FHkcIIXyJ3Pt9l7RACCGqnK4ZL6PmLm+GegPoeuB5yzbTqvTEhBBCVBi59/suKUAIIYQQQgghvCaDqIUQQgghhBBekwKEEEIIIYQQwmtSgBBCCCGEEEJ4TQoQQgghhBBCCK9JAUIIIYQQQgjhtf8DEApPhnIW5X8AAAAASUVORK5CYII=\n", "text/plain": [ "
" ] }, "metadata": { "needs_background": "light" }, "output_type": "display_data" } ], "source": [ "from matplotlib import gridspec\n", "import snlay as snlay\n", "#here we test edge cases where failure happens\n", "#size = np.array([70, 60, 50, 40, 30, 20, 10, 10])\n", "#size = np.array([70, 70, 70, 70, 70, 70, 70, 70])\n", "#size = np.array([30, 30, 30, 30, 30, 30, 30, 30])\n", "\n", "#size = np.array([65, 65, 65, 65, 55, 65, 35, 65])\n", "#size = np.array([65, 35, 45, 35, 45, 35, 45, 35])\n", "\n", "size = np.random.randint(30, 71, 8)\n", "mats = np.array([3, 4, 3, 4, 3, 4, 3, 4])\n", "spec_ac = snlay.calc_spectrum(size, mats, lams)\n", "\n", "print(size)\n", "\n", "\n", "size = (size - 50.0)/20.0\n", "\n", "\n", "\n", "spec = model.predict(np.expand_dims(size, axis = 0))\n", "\n", "spec = np.ravel(spec)\n", "\n", "\n", "fig1 = plt.figure(figsize=(11,3))\n", "gs = gridspec.GridSpec(1, 2, width_ratios=[8, 3]) \n", "\n", "ax = plt.subplot(gs[0])\n", "#ax = fig1.add_subplot(1,2,1)\n", "#ax.set_title('silica coated gold')\n", "ax.set_xlabel('Wavelength (nm)')\n", "ax.set_ylabel('Extinction Efficiency Qe')\n", "ax.set_ylim((0, 6))\n", "ax.set_xlim((300, 1200))\n", "plt.plot(lams, spec_ac,'b', linewidth=1, label='True')\n", "plt.plot(lams, spec, 'r--', linewidth=2, label='predicted')\n", "plt.plot(lams, 10*np.abs(spec_ac - spec)/spec_ac,'k', linewidth=1, label='10x Relative error')\n", "ax.legend(loc='best')\n", "\n", "ax2 = plt.subplot(gs[1])\n", "#fig2 = plt.figure(figsize=(3,3))\n", "#ax2 = fig1.add_subplot(1,2,2)\n", "#ax.set_title('silica coated gold')\n", "ax2.set_xlabel('Wavelength (nm)')\n", "#ax.set_ylabel('Extinction Efficiency Qe')\n", "#ax2.set_ylim((2, 6))\n", "ax2.set_xlim((300, 450))\n", "plt.plot(lams, spec_ac,'b', linewidth=1, label='True')\n", "plt.plot(lams, spec, 'r--', linewidth=2, label='predicted')\n", "#plt.plot(lams, 10*np.abs(spec_ac - spec)/spec_ac,'b', linewidth=1, label='10x Relative error')\n", "#ax.legend(loc='best')\n", "\n", "\n", "\n", "plt.tight_layout()\n", "plt.savefig('grid_figure.pdf')\n", "\n", "\n", "\n", "\n" ] }, { "cell_type": "code", "execution_count": 5, "metadata": { "ExecuteTime": { "end_time": "2018-10-01T16:13:32.441268Z", "start_time": "2018-10-01T16:13:32.415956Z" } }, "outputs": [ { "data": { "text/plain": [ "(8,)" ] }, "execution_count": 5, "metadata": {}, "output_type": "execute_result" } ], "source": [ "size.shape" ] }, { "cell_type": "code", "execution_count": 6, "metadata": { "ExecuteTime": { "end_time": "2018-09-30T12:48:05.013204Z", "start_time": "2018-09-30T12:48:04.790487Z" } }, "outputs": [ { "name": "stdout", "output_type": "stream", "text": [ "1.829695701599121\n" ] }, { "ename": "NameError", "evalue": "name 'result' is not defined", "output_type": "error", "traceback": [ "\u001b[0;31m---------------------------------------------------------------------------\u001b[0m", "\u001b[0;31mNameError\u001b[0m Traceback (most recent call last)", "\u001b[0;32m\u001b[0m in \u001b[0;36m\u001b[0;34m()\u001b[0m\n\u001b[1;32m 27\u001b[0m \u001b[0;34m\u001b[0m\u001b[0m\n\u001b[1;32m 28\u001b[0m \u001b[0;34m\u001b[0m\u001b[0m\n\u001b[0;32m---> 29\u001b[0;31m \u001b[0mresult\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n\u001b[0m", "\u001b[0;31mNameError\u001b[0m: name 'result' is not defined" ] } ], "source": [ "# loop for hundred runs \n", "import snlay as snlay\n", "import time\n", "\n", "mats = np.array([3, 4, 3, 4, 3, 4, 3, 4])\n", "#spec_ac = snlay.calc_spectrum(size, mats, lams)\n", "\n", "reps = 100\n", "\n", "\n", "start = time.time()\n", "for ind in np.arange(reps):\n", " size = np.random.randint(30, 71, 8)\n", " #size = np.random.randint(30, 71, (1,8))\n", " #spec_ac = snlay.calc_spectrum(size, mats, lams)\n", " size = (size - 50.0)/20.0\n", " size = np.expand_dims(size, axis=0)\n", " spec = model.predict(size)\n", " #data_iter = mx.io.NDArrayIter(size, None, 1)\n", " #result = mod.predict(size)\n", " #result = result.asnumpy()\n", "\n", " \n", "end = time.time()\n", "print(1000*(end - start)/reps) \n", " \n", "\n", "\n", "result\n", "\n" ] }, { "cell_type": "markdown", "metadata": { "heading_collapsed": true }, "source": [ "### Inverse scattering " ] }, { "cell_type": "code", "execution_count": null, "metadata": { "ExecuteTime": { "end_time": "2018-09-29T09:59:20.270577Z", "start_time": "2018-09-29T09:59:20.206497Z" }, "hidden": true }, "outputs": [], "source": [ "# from keras.models import Model\n", "# from keras.layers import Input, Add, AveragePooling1D, MaxPooling1D, Concatenate\n", "\n", "\n", "\n", "# a = Input(shape=(1,))\n", "# first = Dense(8, input_dim=1, kernel_initializer='normal', activation='linear', \n", "# name='dummy', use_bias=False)(a)\n", "# last = model(first)\n", "# #last = Dense(256, kernel_initializer='normal')(first)\n", "# #last = Flatten()(first)\n", "\n", "# model_d = Model(inputs=a, outputs=last)\n", "# model_d.compile(loss=naive_percent_loss, optimizer='nadam', metrics=[calc_mre_K], context=['gpu(0)'])\n", "\n", "model_d = Sequential()\n", "model_d.add(Dense(8, input_dim=1, kernel_initializer='normal', activation='linear', \n", " name='dummy', use_bias=False))\n", "\n", "for layer in model.layers[1:]:\n", " model_d.add(layer)\n", "\n", "# for layer in model_d.layers[1:]:\n", "# layer.trainable = False\n", "\n", "# for ind in range(1,len(model_d.layers)):\n", "# model_d.layers[ind].set_weights(model.layers[ind-1].get_weights())\n", "\n", "model_d.compile(loss=naive_percent_loss, optimizer='adam') \n", " \n", " \n", "model_d.summary()\n", "\n", "\n", "for layer in model.layers[1:]:\n", " print(layer)\n", "\n", "\n", "\n", "\n", "\n", "\n", "# # # let us create a target spectrum first\n", "# import snlay as snlay\n", "# #size = np.array([60, 65, 65, 65, 35, 35, 35, 35])\n", "# size = np.random.randint(30,70,8)\n", "# mats = np.array([3, 4, 3, 4, 3, 4, 3, 4])\n", "# target = snlay.calc_spectrum(size, mats, lams)\n", "\n", "# print(size)\n", "\n", "# # #do the training here\n", "# xd_t = np.ones((1,1))\n", "# yd_t = target.reshape(1,250)\n", "\n", "\n", "\n", "\n", "\n", "# history = model_d.fit(xd_t, yd_t,\n", "# batch_size=1,\n", "# epochs=5000, \n", "# verbose=0)\n", "\n", "# # #here is the final result\n", "# size_out = model_d.get_layer('dummy')\n", "# wts = size_out.get_weights()\n", "# wts = np.array(wts).ravel()\n", "# size_res= 0.5*wts*(size_max - size_min) + size_av\n", "# size_res_rounded = np.round(size_res)\n", "\n", "# spec_zer = model_d.predict(xd_t).ravel()\n", "# achieved = snlay.calc_spectrum(size_res_rounded, mats, lams)\n", "\n", "\n", "# fig1 = plt.figure(figsize=(22,5))\n", "# ax = fig1.add_subplot(1,1,1)\n", "# #plt.plot(lams, spec_zer, label='new model')\n", "# plt.plot(lams, target, linewidth=2, label='target')\n", "# plt.plot(lams, achieved, '--', linewidth=3, label='achieved')\n", "# plt.legend(bbox_to_anchor=(1.05, 1), loc=2, borderaxespad=0.)\n", "\n", "# print(size_res_rounded)\n" ] }, { "cell_type": "code", "execution_count": null, "metadata": { "ExecuteTime": { "end_time": "2018-09-26T10:28:37.316134Z", "start_time": "2018-09-26T10:27:22.339834Z" }, "hidden": true }, "outputs": [], "source": [ "from keras.models import Model\n", "from keras.layers import Input, Add, AveragePooling1D, MaxPooling1D, Concatenate\n", "\n", "a = Input(shape=(8,))\n", "first = Dense(256, kernel_initializer='normal')(a)\n", "#first = Dense(128, kernel_initializer='normal')(first)\n", "#first = BatchNormalization()(first)\n", "first= Activation('relu')(first)\n", "\n", "first = Reshape((256,1))(first)\n", "#first = UpSampling1D(size = 2)(first)\n", "first = Conv1D(filters=32, kernel_size=1, strides=1, padding='same', \n", " kernel_initializer='normal', activation='relu')(first)\n", "#first = UpSampling1D(size = 2)(first)\n", "\n", "\n", "\n", "\n", "# first_1 = Conv1D(filters=4, kernel_size=1, strides=1, padding='same', \n", "# kernel_initializer='normal')(first)\n", "# first_1 = Activation('relu')(first_1)\n", "# first_1 = Conv1D(filters=8, kernel_size=3, strides=1, padding='same', \n", "# kernel_initializer='normal')(first_1)\n", "first_2 = Conv1D(filters=4, kernel_size=1, strides=1, padding='same', \n", " kernel_initializer='normal')(first)\n", "#first_2 = Activation('relu')(first_2)\n", "first_2 = Conv1D(filters=8, kernel_size=3, strides=1, padding='same', \n", " kernel_initializer='normal')(first_2)\n", "first_3 = Conv1D(filters=4, kernel_size=1, strides=1, padding='same', \n", " kernel_initializer='normal')(first)\n", "first_3 = Activation('relu')(first_3)\n", "first_3 = Conv1D(filters=8, kernel_size=5, strides=1, padding='same', \n", " kernel_initializer='normal')(first_3)\n", "first_4 = Conv1D(filters=16, kernel_size=1, strides=1, padding='same', \n", " kernel_initializer='normal')(first)\n", "list_of_tensors = [first_2, first_3, first_4]\n", "conc = Concatenate()(list_of_tensors)\n", "first = Add()([first, conc])\n", "first= Activation('relu')(first)\n", "\n", "\n", "\n", "# first_1 = Conv1D(filters=4, kernel_size=1, strides=1, padding='same', \n", "# kernel_initializer='normal')(first)\n", "# first_1 = Activation('relu')(first_1)\n", "# first_1 = Conv1D(filters=8, kernel_size=3, strides=1, padding='same', \n", "# kernel_initializer='normal')(first_1)\n", "first_2 = Conv1D(filters=4, kernel_size=1, strides=1, padding='same', \n", " kernel_initializer='normal')(first)\n", "#first_2 = Activation('relu')(first_2)\n", "first_2 = Conv1D(filters=8, kernel_size=3, strides=1, padding='same', \n", " kernel_initializer='normal')(first_2)\n", "first_3 = Conv1D(filters=4, kernel_size=1, strides=1, padding='same', \n", " kernel_initializer='normal')(first)\n", "first_3 = Activation('relu')(first_3)\n", "first_3 = Conv1D(filters=8, kernel_size=5, strides=1, padding='same', \n", " kernel_initializer='normal')(first_3)\n", "first_4 = Conv1D(filters=16, kernel_size=1, strides=1, padding='same', \n", " kernel_initializer='normal')(first)\n", "list_of_tensors = [first_2, first_3, first_4]\n", "conc = Concatenate()(list_of_tensors)\n", "first = Add()([first, conc])\n", "first= Activation('relu')(first)\n", "\n", "\n", "\n", "\n", "# first_1 = Conv1D(filters=4, kernel_size=1, strides=1, padding='same', \n", "# kernel_initializer='normal')(first)\n", "# first_1 = Activation('relu')(first_1)\n", "# first_1 = Conv1D(filters=8, kernel_size=3, strides=1, padding='same', \n", "# kernel_initializer='normal')(first_1)\n", "first_2 = Conv1D(filters=4, kernel_size=1, strides=1, padding='same', \n", " kernel_initializer='normal')(first)\n", "#first_2 = Activation('relu')(first_2)\n", "first_2 = Conv1D(filters=8, kernel_size=3, strides=1, padding='same', \n", " kernel_initializer='normal')(first_2)\n", "first_3 = Conv1D(filters=4, kernel_size=1, strides=1, padding='same', \n", " kernel_initializer='normal')(first)\n", "first_3 = Activation('relu')(first_3)\n", "first_3 = Conv1D(filters=8, kernel_size=5, strides=1, padding='same', \n", " kernel_initializer='normal')(first_3)\n", "first_4 = Conv1D(filters=16, kernel_size=1, strides=1, padding='same', \n", " kernel_initializer='normal')(first)\n", "list_of_tensors = [first_2, first_3, first_4]\n", "conc = Concatenate()(list_of_tensors)\n", "first = Add()([first, conc])\n", "first= Activation('relu')(first)\n", "\n", "# first_1 = Conv1D(filters=4, kernel_size=1, strides=1, padding='same', \n", "# kernel_initializer='normal')(first)\n", "# first_1 = Activation('relu')(first_1)\n", "# first_1 = Conv1D(filters=8, kernel_size=3, strides=1, padding='same', \n", "# kernel_initializer='normal')(first_1)\n", "first_2 = Conv1D(filters=4, kernel_size=1, strides=1, padding='same', \n", " kernel_initializer='normal')(first)\n", "#first_2 = Activation('relu')(first_2)\n", "first_2 = Conv1D(filters=8, kernel_size=3, strides=1, padding='same', \n", " kernel_initializer='normal')(first_2)\n", "first_3 = Conv1D(filters=4, kernel_size=1, strides=1, padding='same', \n", " kernel_initializer='normal')(first)\n", "first_3 = Activation('relu')(first_3)\n", "first_3 = Conv1D(filters=8, kernel_size=5, strides=1, padding='same', \n", " kernel_initializer='normal')(first_3)\n", "first_4 = Conv1D(filters=16, kernel_size=1, strides=1, padding='same', \n", " kernel_initializer='normal')(first)\n", "list_of_tensors = [first_2, first_3, first_4]\n", "conc = Concatenate()(list_of_tensors)\n", "first = Add()([first, conc])\n", "first= Activation('relu')(first)\n", "\n", "\n", "\n", "\n", "\n", "first = Reshape((32,256))(first)\n", "first = MaxPooling1D(pool_size=32, strides=None, padding='same')(first)\n", "last = Flatten()(first)\n", "\n", "model = Model(inputs=a, outputs=last)\n", "model.compile(loss=naive_percent_loss, optimizer='nadam', metrics=[calc_mre_K], context=['gpu(0)'])\n", "\n", "#model.summary()\n", "\n", "from IPython.display import SVG\n", "from keras.utils.vis_utils import model_to_dot\n", "from keras.utils.vis_utils import plot_model\n", "\n", "#SVG(model_to_dot(model, show_shapes=True, show_layer_names=False).create(prog='dot', format='svg'))\n", "\n", "\n", "plot_model(model, show_shapes=False, show_layer_names=False, to_file='model.png')\n", "#plot_model(model, to_file='model.png', )\n", "\n", "x_t, x_v, y_t, y_v = train_test_split(x_train, y_train, test_size=0.2, random_state=42)\n", "# model = naiveploss_mgpu_model()\n", "# model.summary() \n", "history = model.fit(x_train, y_train,\n", " batch_size=64,\n", " epochs=2000, \n", " verbose=1,\n", " validation_data=(x_test, y_test))\n", "\n", "\n", "\n", "\n" ] }, { "cell_type": "markdown", "metadata": { "heading_collapsed": true }, "source": [ "### Model shipment" ] }, { "cell_type": "code", "execution_count": null, "metadata": { "ExecuteTime": { "end_time": "2018-09-06T10:51:07.997455Z", "start_time": "2018-09-06T10:49:19.326Z" }, "hidden": true }, "outputs": [], "source": [ "\n", " \n", "from keras.models import load_model\n", "\n", "#Creates a HDF5 file 'my_model.h5'\n", "model.save('my_model.h5')\n", "\n", "# Deletes the existing model\n", "#del model \n", "\n", "# Returns a compiled model identical to the previous one\n", "#model = load_model('my_model.h5')\n", "\n" ] }, { "cell_type": "code", "execution_count": null, "metadata": { "ExecuteTime": { "end_time": "2018-09-08T09:40:51.323101Z", "start_time": "2018-09-08T09:40:50.155098Z" }, "hidden": true }, "outputs": [], "source": [ "\n", "from keras.models import Model\n", "from keras.layers import Input, Add, Lambda, Dense\n", "import numpy as np\n", "\n", "\n", "def dropper(x):\n", " ms = 4\n", " #print(x.shape)\n", " \n", " \n", " \n", " return x**2\n", "# msk = np.array([1,1,1,1,0,0,0,0])\n", "\n", "\n", "a = Input(shape=(1,))\n", "b = Dense(8, input_dim=1, kernel_initializer='normal', activation='linear', \n", " name='dummy', use_bias=False)(a)\n", "b = Lambda(dropper)(b)\n", "b = Dense(256)(b)\n", "# #b = Lambda()\n", "# a = Lambda(dropper)(a)\n", "# \n", "\n", "model = Model(inputs=a, outputs=b)\n", "\n", "# model = Sequential()\n", "# model.add(Dense(256, input_dim=8))\n", "# #model.add(Lambda(lambda x: x**2))\n", "# model.add(Lambda(dropper))\n", "# model.add(Dense(256))\n", "\n", "\n", "\n", "model.compile(loss=naive_percent_loss, optimizer='adam', metrics=[calc_mre_K])\n", "#model.summary()" ] }, { "cell_type": "code", "execution_count": null, "metadata": { "ExecuteTime": { "end_time": "2018-09-28T10:32:16.469180Z", "start_time": "2018-09-28T10:32:16.452138Z" }, "hidden": true }, "outputs": [], "source": [ "2 +2 " ] }, { "cell_type": "code", "execution_count": null, "metadata": { "ExecuteTime": { "end_time": "2018-09-28T10:32:36.778144Z", "start_time": "2018-09-28T10:32:36.682162Z" }, "hidden": true }, "outputs": [], "source": [ "import matplotlib.pyplot as plt" ] }, { "cell_type": "code", "execution_count": null, "metadata": { "hidden": true }, "outputs": [], "source": [] } ], "metadata": { "kernelspec": { "display_name": "Python 3", "language": "python", "name": "python3" }, "language_info": { "codemirror_mode": { "name": "ipython", "version": 3 }, "file_extension": ".py", "mimetype": "text/x-python", "name": "python", "nbconvert_exporter": "python", "pygments_lexer": "ipython3", "version": "3.7.0" }, "latex_envs": { "LaTeX_envs_menu_present": true, "autoclose": false, "autocomplete": true, "bibliofile": "biblio.bib", "cite_by": "apalike", "current_citInitial": 1, "eqLabelWithNumbers": true, "eqNumInitial": 1, "hotkeys": { "equation": "Ctrl-E", "itemize": "Ctrl-I" }, "labels_anchors": false, "latex_user_defs": false, "report_style_numbering": false, "user_envs_cfg": false } }, "nbformat": 4, "nbformat_minor": 2 }