How to properly save and load an intermediate model in Keras?TensorFlow and Categorical variablesfeature extraction for a pretrained model in kerasKeras LSTM: use weights from Keras model to replicate predictions using numpyKeras shape error in applications Inception Resnet v2with tf.device(DEVICE): model = modellib.MaskRCNN(mode = “inference”, model_dir = LOGS_DIR, config = config)Using categorial_crossentropy to train a model in keras3 dimensional array as input with Embedding Layer and LSTM in KerasIN CIFAR 10 DATASETproblem of entry format for a simple model in KerasSetting input shape for an NLP task in R(Rstudio) using keras 1D convolution layer, when it expects 3 dimensional input (a tensor)

Have I damaged my car by attempting to reverse with hand/park brake up?

Why is "breaking the mould" positively connoted?

What is the closest airport to the center of the city it serves?

How I can I roll a number of non-digital dice to get a random number between 1 and 150?

What was Bran's plan to kill the Night King?

What was the first sci-fi story to feature the plot "the humans were the monsters all along"?

Why isn't nylon as strong as kevlar?

Can there be a single technologically advanced nation, in a continent full of non-technologically advanced nations?

Why wasn't the Night King naked in S08E03?

Which module had more 'comfort' in terms of living space, the Lunar Module or the Command module?

The number of days until the end of the month

Using column size much larger than necessary

Can my company stop me from working overtime?

How do I inject UserInterface into Access Control?

Building a list of products from the elements in another list

Why has the UK chosen to use Huawei infrastructure when Five Eyes allies haven't?

Is there an idiom that support the idea that "inflation is bad"?

Is there an official reason for not adding a post-credits scene?

Could the black hole photo be a gravastar?

What are the differences between credential stuffing and password spraying?

Can hackers enable the camera after the user disabled it?

Can I use a fetch land to shuffle my deck while the opponent has Ashiok, Dream Render in play?

Are there any of the Children of the Forest left, or are they extinct?

Where can I go to avoid planes overhead?



How to properly save and load an intermediate model in Keras?


TensorFlow and Categorical variablesfeature extraction for a pretrained model in kerasKeras LSTM: use weights from Keras model to replicate predictions using numpyKeras shape error in applications Inception Resnet v2with tf.device(DEVICE): model = modellib.MaskRCNN(mode = “inference”, model_dir = LOGS_DIR, config = config)Using categorial_crossentropy to train a model in keras3 dimensional array as input with Embedding Layer and LSTM in KerasIN CIFAR 10 DATASETproblem of entry format for a simple model in KerasSetting input shape for an NLP task in R(Rstudio) using keras 1D convolution layer, when it expects 3 dimensional input (a tensor)













1












$begingroup$


I'm working with a model that involves 3 stages of 'nesting' of models in Keras.



Conceptually the first is a transfer learning CNN model, for example MobileNetV2. (Model 1) This is then wrapped by a model that consists of a small DNN. (Model 2) Finally during training these are all wrapped by a model that concatenates multiple outputs from model 2, calculates loss, and then backpropagates into model 2 and in the future model 1. (Model 3)



For inference later I simply want to save the weights of models 1 and 2. I have had multiple issues with this, due to what appear to be bugs in some versions of Keras (I'm using 2.2.2) and also loading the weights more explicitly is appearing to result in randomized weights and so isn't working correctly. Instead of attempting to troubleshoot what is going wrong with whatever scenario I'm simply trying to determine what is the best practice for saving intermediate nested models.



def create_model_2(IN_DIM=(224, 224, 3), OUT_DIM=128):
# First define the transfer learning model
initial_img = Input(shape=(IN_DIM))

black_box = MobileNetV2(include_top=False, input_shape=IN_DIM, weights="imagenet", pooling="avg")(initial_img)

bb_model = Model(b_img, black_box)

# freeze layers for transfer learning model
for layer in bb_model.layers:
layer.trainable = False

#########################
###### TOWER BLOCK ######
#########################

img = Input(shape=(IN_DIM))

x = bb_model(img)

# add some layers to try to learn
x = Dense(64, activation='relu', name='new_fc0')(x)

x = Dense(OUT_DIM, activation='relu', name='new_fc1')(x)

# L2 norm to project to unit sphere
out = Lambda(lambda x: K.l2_normalize(x, axis=1), name='final_l2_norm')(x)

_model_2 = Model(img, out)

return _model_2


Then the structure of Model 3:



IN_DIM = (224, 224, 3) # mobilenetv2=(224, 224, 3) Iv3=(299, 299, 3)
OUT_DIM = 32

model_2 = create_model_2(IN_DIM, OUT_DIM)

# then define images for triplets
anchor_img = Input(shape=IN_DIM)
pos_img = Input(shape=IN_DIM)
neg_img = Input(shape=IN_DIM)

# create three vectors representing the images
anchor_in = model_2(anchor_img)
positive_in = model_2(pos_img)
negative_in = model_2(neg_img)

# concatenate the vectors into one large vector for input into the triplet loss "processor"
merged_vector = concatenate([anchor_in, positive_in, negative_in], axis=-1)

# actually define the model:
model_3 = Model(inputs=[anchor_img, pos_img, neg_img], outputs=merged_vector)


The model seems to run and train just fine:



OPTIMIZER = SGD(lr=learning_rate, momentum=0.9)

final_model.compile(optimizer=OPTIMIZER, loss=triplet_loss, metrics=[avg_AP_dist, avg_AN_dist])

history = final_model.fit_generator(generator=training_generator,
epochs=5, # short for debugging
use_multiprocessing=True,
workers=4)


But saving the model after training is unclear:



out_file = "../../models/:_epoch_:_weights.h5".format(MODEL_DESC, 5)
model_2.save_weights(out_file) # save the actual Tower weights, discard the "booster" wrapper
print("Saved: :".format(out_file))


Or:



out_file = "../../models/:_epoch_:_weights.h5".format(MODEL_DESC, 5)
model_2.save(out_file) # save the actual Tower weights, discard the "booster" wrapper
print("Saved: :".format(out_file))


Or something else?



The current failure modes seem to be if I try to load in just the weights into a newly instantiated model_2 instance I get:



ValueError: axes don't match array


Which from searching may be related to a bug in Keras. If I save the model (.save() rather than .save_weights() then it loads without complaint but the inference is not stable and appears to be horrible/random.)



Thank you.



Still getting the following traceback:



<snip>/src/notebooks/vectorizer.py in load_model()
65
66 # load the weights
---> 67 loaded_model.load_weights(weights_path)
68
69 print("Model ready")

/opt/conda/lib/python3.6/site-packages/keras/engine/network.py in load_weights(self, filepath, by_name, skip_mismatch, reshape)
1164 else:
1165 saving.load_weights_from_hdf5_group(
-> 1166 f, self.layers, reshape=reshape)
1167
1168 def _updated_config(self):

/opt/conda/lib/python3.6/site-packages/keras/engine/saving.py in load_weights_from_hdf5_group(f, layers, reshape)
1043 original_keras_version,
1044 original_backend,
-> 1045 reshape=reshape)
1046 if len(weight_values) != len(symbolic_weights):
1047 raise ValueError('Layer #' + str(k) +

/opt/conda/lib/python3.6/site-packages/keras/engine/saving.py in preprocess_weights_for_loading(layer, weights, original_keras_version, original_backend, reshape)
680 weights = convert_nested_time_distributed(weights)
681 elif layer.__class__.__name__ in ['Model', 'Sequential']:
--> 682 weights = convert_nested_model(weights)
683
684 if original_keras_version == '1':

/opt/conda/lib/python3.6/site-packages/keras/engine/saving.py in convert_nested_model(weights)
668 weights=weights[:num_weights],
669 original_keras_version=original_keras_version,
--> 670 original_backend=original_backend))
671 weights = weights[num_weights:]
672 return new_weights

/opt/conda/lib/python3.6/site-packages/keras/engine/saving.py in preprocess_weights_for_loading(layer, weights, original_keras_version, original_backend, reshape)
680 weights = convert_nested_time_distributed(weights)
681 elif layer.__class__.__name__ in ['Model', 'Sequential']:
--> 682 weights = convert_nested_model(weights)
683
684 if original_keras_version == '1':

/opt/conda/lib/python3.6/site-packages/keras/engine/saving.py in convert_nested_model(weights)
656 weights=weights[:num_weights],
657 original_keras_version=original_keras_version,
--> 658 original_backend=original_backend))
659 weights = weights[num_weights:]
660

/opt/conda/lib/python3.6/site-packages/keras/engine/saving.py in preprocess_weights_for_loading(layer, weights, original_keras_version, original_backend, reshape)
799 weights[0] = np.reshape(weights[0], layer_weights_shape)
800 elif layer_weights_shape != weights[0].shape:
--> 801 weights[0] = np.transpose(weights[0], (3, 2, 0, 1))
802 if layer.__class__.__name__ == 'ConvLSTM2D':
803 weights[1] = np.transpose(weights[1], (3, 2, 0, 1))

/opt/conda/lib/python3.6/site-packages/numpy/core/fromnumeric.py in transpose(a, axes)
596
597 """
--> 598 return _wrapfunc(a, 'transpose', axes)
599
600

/opt/conda/lib/python3.6/site-packages/numpy/core/fromnumeric.py in _wrapfunc(obj, method, *args, **kwds)
49 def _wrapfunc(obj, method, *args, **kwds):
50 try:
---> 51 return getattr(obj, method)(*args, **kwds)
52
53 # An AttributeError occurs if the object does not have

ValueError: axes don't match array









share|improve this question











$endgroup$
















    1












    $begingroup$


    I'm working with a model that involves 3 stages of 'nesting' of models in Keras.



    Conceptually the first is a transfer learning CNN model, for example MobileNetV2. (Model 1) This is then wrapped by a model that consists of a small DNN. (Model 2) Finally during training these are all wrapped by a model that concatenates multiple outputs from model 2, calculates loss, and then backpropagates into model 2 and in the future model 1. (Model 3)



    For inference later I simply want to save the weights of models 1 and 2. I have had multiple issues with this, due to what appear to be bugs in some versions of Keras (I'm using 2.2.2) and also loading the weights more explicitly is appearing to result in randomized weights and so isn't working correctly. Instead of attempting to troubleshoot what is going wrong with whatever scenario I'm simply trying to determine what is the best practice for saving intermediate nested models.



    def create_model_2(IN_DIM=(224, 224, 3), OUT_DIM=128):
    # First define the transfer learning model
    initial_img = Input(shape=(IN_DIM))

    black_box = MobileNetV2(include_top=False, input_shape=IN_DIM, weights="imagenet", pooling="avg")(initial_img)

    bb_model = Model(b_img, black_box)

    # freeze layers for transfer learning model
    for layer in bb_model.layers:
    layer.trainable = False

    #########################
    ###### TOWER BLOCK ######
    #########################

    img = Input(shape=(IN_DIM))

    x = bb_model(img)

    # add some layers to try to learn
    x = Dense(64, activation='relu', name='new_fc0')(x)

    x = Dense(OUT_DIM, activation='relu', name='new_fc1')(x)

    # L2 norm to project to unit sphere
    out = Lambda(lambda x: K.l2_normalize(x, axis=1), name='final_l2_norm')(x)

    _model_2 = Model(img, out)

    return _model_2


    Then the structure of Model 3:



    IN_DIM = (224, 224, 3) # mobilenetv2=(224, 224, 3) Iv3=(299, 299, 3)
    OUT_DIM = 32

    model_2 = create_model_2(IN_DIM, OUT_DIM)

    # then define images for triplets
    anchor_img = Input(shape=IN_DIM)
    pos_img = Input(shape=IN_DIM)
    neg_img = Input(shape=IN_DIM)

    # create three vectors representing the images
    anchor_in = model_2(anchor_img)
    positive_in = model_2(pos_img)
    negative_in = model_2(neg_img)

    # concatenate the vectors into one large vector for input into the triplet loss "processor"
    merged_vector = concatenate([anchor_in, positive_in, negative_in], axis=-1)

    # actually define the model:
    model_3 = Model(inputs=[anchor_img, pos_img, neg_img], outputs=merged_vector)


    The model seems to run and train just fine:



    OPTIMIZER = SGD(lr=learning_rate, momentum=0.9)

    final_model.compile(optimizer=OPTIMIZER, loss=triplet_loss, metrics=[avg_AP_dist, avg_AN_dist])

    history = final_model.fit_generator(generator=training_generator,
    epochs=5, # short for debugging
    use_multiprocessing=True,
    workers=4)


    But saving the model after training is unclear:



    out_file = "../../models/:_epoch_:_weights.h5".format(MODEL_DESC, 5)
    model_2.save_weights(out_file) # save the actual Tower weights, discard the "booster" wrapper
    print("Saved: :".format(out_file))


    Or:



    out_file = "../../models/:_epoch_:_weights.h5".format(MODEL_DESC, 5)
    model_2.save(out_file) # save the actual Tower weights, discard the "booster" wrapper
    print("Saved: :".format(out_file))


    Or something else?



    The current failure modes seem to be if I try to load in just the weights into a newly instantiated model_2 instance I get:



    ValueError: axes don't match array


    Which from searching may be related to a bug in Keras. If I save the model (.save() rather than .save_weights() then it loads without complaint but the inference is not stable and appears to be horrible/random.)



    Thank you.



    Still getting the following traceback:



    <snip>/src/notebooks/vectorizer.py in load_model()
    65
    66 # load the weights
    ---> 67 loaded_model.load_weights(weights_path)
    68
    69 print("Model ready")

    /opt/conda/lib/python3.6/site-packages/keras/engine/network.py in load_weights(self, filepath, by_name, skip_mismatch, reshape)
    1164 else:
    1165 saving.load_weights_from_hdf5_group(
    -> 1166 f, self.layers, reshape=reshape)
    1167
    1168 def _updated_config(self):

    /opt/conda/lib/python3.6/site-packages/keras/engine/saving.py in load_weights_from_hdf5_group(f, layers, reshape)
    1043 original_keras_version,
    1044 original_backend,
    -> 1045 reshape=reshape)
    1046 if len(weight_values) != len(symbolic_weights):
    1047 raise ValueError('Layer #' + str(k) +

    /opt/conda/lib/python3.6/site-packages/keras/engine/saving.py in preprocess_weights_for_loading(layer, weights, original_keras_version, original_backend, reshape)
    680 weights = convert_nested_time_distributed(weights)
    681 elif layer.__class__.__name__ in ['Model', 'Sequential']:
    --> 682 weights = convert_nested_model(weights)
    683
    684 if original_keras_version == '1':

    /opt/conda/lib/python3.6/site-packages/keras/engine/saving.py in convert_nested_model(weights)
    668 weights=weights[:num_weights],
    669 original_keras_version=original_keras_version,
    --> 670 original_backend=original_backend))
    671 weights = weights[num_weights:]
    672 return new_weights

    /opt/conda/lib/python3.6/site-packages/keras/engine/saving.py in preprocess_weights_for_loading(layer, weights, original_keras_version, original_backend, reshape)
    680 weights = convert_nested_time_distributed(weights)
    681 elif layer.__class__.__name__ in ['Model', 'Sequential']:
    --> 682 weights = convert_nested_model(weights)
    683
    684 if original_keras_version == '1':

    /opt/conda/lib/python3.6/site-packages/keras/engine/saving.py in convert_nested_model(weights)
    656 weights=weights[:num_weights],
    657 original_keras_version=original_keras_version,
    --> 658 original_backend=original_backend))
    659 weights = weights[num_weights:]
    660

    /opt/conda/lib/python3.6/site-packages/keras/engine/saving.py in preprocess_weights_for_loading(layer, weights, original_keras_version, original_backend, reshape)
    799 weights[0] = np.reshape(weights[0], layer_weights_shape)
    800 elif layer_weights_shape != weights[0].shape:
    --> 801 weights[0] = np.transpose(weights[0], (3, 2, 0, 1))
    802 if layer.__class__.__name__ == 'ConvLSTM2D':
    803 weights[1] = np.transpose(weights[1], (3, 2, 0, 1))

    /opt/conda/lib/python3.6/site-packages/numpy/core/fromnumeric.py in transpose(a, axes)
    596
    597 """
    --> 598 return _wrapfunc(a, 'transpose', axes)
    599
    600

    /opt/conda/lib/python3.6/site-packages/numpy/core/fromnumeric.py in _wrapfunc(obj, method, *args, **kwds)
    49 def _wrapfunc(obj, method, *args, **kwds):
    50 try:
    ---> 51 return getattr(obj, method)(*args, **kwds)
    52
    53 # An AttributeError occurs if the object does not have

    ValueError: axes don't match array









    share|improve this question











    $endgroup$














      1












      1








      1


      2



      $begingroup$


      I'm working with a model that involves 3 stages of 'nesting' of models in Keras.



      Conceptually the first is a transfer learning CNN model, for example MobileNetV2. (Model 1) This is then wrapped by a model that consists of a small DNN. (Model 2) Finally during training these are all wrapped by a model that concatenates multiple outputs from model 2, calculates loss, and then backpropagates into model 2 and in the future model 1. (Model 3)



      For inference later I simply want to save the weights of models 1 and 2. I have had multiple issues with this, due to what appear to be bugs in some versions of Keras (I'm using 2.2.2) and also loading the weights more explicitly is appearing to result in randomized weights and so isn't working correctly. Instead of attempting to troubleshoot what is going wrong with whatever scenario I'm simply trying to determine what is the best practice for saving intermediate nested models.



      def create_model_2(IN_DIM=(224, 224, 3), OUT_DIM=128):
      # First define the transfer learning model
      initial_img = Input(shape=(IN_DIM))

      black_box = MobileNetV2(include_top=False, input_shape=IN_DIM, weights="imagenet", pooling="avg")(initial_img)

      bb_model = Model(b_img, black_box)

      # freeze layers for transfer learning model
      for layer in bb_model.layers:
      layer.trainable = False

      #########################
      ###### TOWER BLOCK ######
      #########################

      img = Input(shape=(IN_DIM))

      x = bb_model(img)

      # add some layers to try to learn
      x = Dense(64, activation='relu', name='new_fc0')(x)

      x = Dense(OUT_DIM, activation='relu', name='new_fc1')(x)

      # L2 norm to project to unit sphere
      out = Lambda(lambda x: K.l2_normalize(x, axis=1), name='final_l2_norm')(x)

      _model_2 = Model(img, out)

      return _model_2


      Then the structure of Model 3:



      IN_DIM = (224, 224, 3) # mobilenetv2=(224, 224, 3) Iv3=(299, 299, 3)
      OUT_DIM = 32

      model_2 = create_model_2(IN_DIM, OUT_DIM)

      # then define images for triplets
      anchor_img = Input(shape=IN_DIM)
      pos_img = Input(shape=IN_DIM)
      neg_img = Input(shape=IN_DIM)

      # create three vectors representing the images
      anchor_in = model_2(anchor_img)
      positive_in = model_2(pos_img)
      negative_in = model_2(neg_img)

      # concatenate the vectors into one large vector for input into the triplet loss "processor"
      merged_vector = concatenate([anchor_in, positive_in, negative_in], axis=-1)

      # actually define the model:
      model_3 = Model(inputs=[anchor_img, pos_img, neg_img], outputs=merged_vector)


      The model seems to run and train just fine:



      OPTIMIZER = SGD(lr=learning_rate, momentum=0.9)

      final_model.compile(optimizer=OPTIMIZER, loss=triplet_loss, metrics=[avg_AP_dist, avg_AN_dist])

      history = final_model.fit_generator(generator=training_generator,
      epochs=5, # short for debugging
      use_multiprocessing=True,
      workers=4)


      But saving the model after training is unclear:



      out_file = "../../models/:_epoch_:_weights.h5".format(MODEL_DESC, 5)
      model_2.save_weights(out_file) # save the actual Tower weights, discard the "booster" wrapper
      print("Saved: :".format(out_file))


      Or:



      out_file = "../../models/:_epoch_:_weights.h5".format(MODEL_DESC, 5)
      model_2.save(out_file) # save the actual Tower weights, discard the "booster" wrapper
      print("Saved: :".format(out_file))


      Or something else?



      The current failure modes seem to be if I try to load in just the weights into a newly instantiated model_2 instance I get:



      ValueError: axes don't match array


      Which from searching may be related to a bug in Keras. If I save the model (.save() rather than .save_weights() then it loads without complaint but the inference is not stable and appears to be horrible/random.)



      Thank you.



      Still getting the following traceback:



      <snip>/src/notebooks/vectorizer.py in load_model()
      65
      66 # load the weights
      ---> 67 loaded_model.load_weights(weights_path)
      68
      69 print("Model ready")

      /opt/conda/lib/python3.6/site-packages/keras/engine/network.py in load_weights(self, filepath, by_name, skip_mismatch, reshape)
      1164 else:
      1165 saving.load_weights_from_hdf5_group(
      -> 1166 f, self.layers, reshape=reshape)
      1167
      1168 def _updated_config(self):

      /opt/conda/lib/python3.6/site-packages/keras/engine/saving.py in load_weights_from_hdf5_group(f, layers, reshape)
      1043 original_keras_version,
      1044 original_backend,
      -> 1045 reshape=reshape)
      1046 if len(weight_values) != len(symbolic_weights):
      1047 raise ValueError('Layer #' + str(k) +

      /opt/conda/lib/python3.6/site-packages/keras/engine/saving.py in preprocess_weights_for_loading(layer, weights, original_keras_version, original_backend, reshape)
      680 weights = convert_nested_time_distributed(weights)
      681 elif layer.__class__.__name__ in ['Model', 'Sequential']:
      --> 682 weights = convert_nested_model(weights)
      683
      684 if original_keras_version == '1':

      /opt/conda/lib/python3.6/site-packages/keras/engine/saving.py in convert_nested_model(weights)
      668 weights=weights[:num_weights],
      669 original_keras_version=original_keras_version,
      --> 670 original_backend=original_backend))
      671 weights = weights[num_weights:]
      672 return new_weights

      /opt/conda/lib/python3.6/site-packages/keras/engine/saving.py in preprocess_weights_for_loading(layer, weights, original_keras_version, original_backend, reshape)
      680 weights = convert_nested_time_distributed(weights)
      681 elif layer.__class__.__name__ in ['Model', 'Sequential']:
      --> 682 weights = convert_nested_model(weights)
      683
      684 if original_keras_version == '1':

      /opt/conda/lib/python3.6/site-packages/keras/engine/saving.py in convert_nested_model(weights)
      656 weights=weights[:num_weights],
      657 original_keras_version=original_keras_version,
      --> 658 original_backend=original_backend))
      659 weights = weights[num_weights:]
      660

      /opt/conda/lib/python3.6/site-packages/keras/engine/saving.py in preprocess_weights_for_loading(layer, weights, original_keras_version, original_backend, reshape)
      799 weights[0] = np.reshape(weights[0], layer_weights_shape)
      800 elif layer_weights_shape != weights[0].shape:
      --> 801 weights[0] = np.transpose(weights[0], (3, 2, 0, 1))
      802 if layer.__class__.__name__ == 'ConvLSTM2D':
      803 weights[1] = np.transpose(weights[1], (3, 2, 0, 1))

      /opt/conda/lib/python3.6/site-packages/numpy/core/fromnumeric.py in transpose(a, axes)
      596
      597 """
      --> 598 return _wrapfunc(a, 'transpose', axes)
      599
      600

      /opt/conda/lib/python3.6/site-packages/numpy/core/fromnumeric.py in _wrapfunc(obj, method, *args, **kwds)
      49 def _wrapfunc(obj, method, *args, **kwds):
      50 try:
      ---> 51 return getattr(obj, method)(*args, **kwds)
      52
      53 # An AttributeError occurs if the object does not have

      ValueError: axes don't match array









      share|improve this question











      $endgroup$




      I'm working with a model that involves 3 stages of 'nesting' of models in Keras.



      Conceptually the first is a transfer learning CNN model, for example MobileNetV2. (Model 1) This is then wrapped by a model that consists of a small DNN. (Model 2) Finally during training these are all wrapped by a model that concatenates multiple outputs from model 2, calculates loss, and then backpropagates into model 2 and in the future model 1. (Model 3)



      For inference later I simply want to save the weights of models 1 and 2. I have had multiple issues with this, due to what appear to be bugs in some versions of Keras (I'm using 2.2.2) and also loading the weights more explicitly is appearing to result in randomized weights and so isn't working correctly. Instead of attempting to troubleshoot what is going wrong with whatever scenario I'm simply trying to determine what is the best practice for saving intermediate nested models.



      def create_model_2(IN_DIM=(224, 224, 3), OUT_DIM=128):
      # First define the transfer learning model
      initial_img = Input(shape=(IN_DIM))

      black_box = MobileNetV2(include_top=False, input_shape=IN_DIM, weights="imagenet", pooling="avg")(initial_img)

      bb_model = Model(b_img, black_box)

      # freeze layers for transfer learning model
      for layer in bb_model.layers:
      layer.trainable = False

      #########################
      ###### TOWER BLOCK ######
      #########################

      img = Input(shape=(IN_DIM))

      x = bb_model(img)

      # add some layers to try to learn
      x = Dense(64, activation='relu', name='new_fc0')(x)

      x = Dense(OUT_DIM, activation='relu', name='new_fc1')(x)

      # L2 norm to project to unit sphere
      out = Lambda(lambda x: K.l2_normalize(x, axis=1), name='final_l2_norm')(x)

      _model_2 = Model(img, out)

      return _model_2


      Then the structure of Model 3:



      IN_DIM = (224, 224, 3) # mobilenetv2=(224, 224, 3) Iv3=(299, 299, 3)
      OUT_DIM = 32

      model_2 = create_model_2(IN_DIM, OUT_DIM)

      # then define images for triplets
      anchor_img = Input(shape=IN_DIM)
      pos_img = Input(shape=IN_DIM)
      neg_img = Input(shape=IN_DIM)

      # create three vectors representing the images
      anchor_in = model_2(anchor_img)
      positive_in = model_2(pos_img)
      negative_in = model_2(neg_img)

      # concatenate the vectors into one large vector for input into the triplet loss "processor"
      merged_vector = concatenate([anchor_in, positive_in, negative_in], axis=-1)

      # actually define the model:
      model_3 = Model(inputs=[anchor_img, pos_img, neg_img], outputs=merged_vector)


      The model seems to run and train just fine:



      OPTIMIZER = SGD(lr=learning_rate, momentum=0.9)

      final_model.compile(optimizer=OPTIMIZER, loss=triplet_loss, metrics=[avg_AP_dist, avg_AN_dist])

      history = final_model.fit_generator(generator=training_generator,
      epochs=5, # short for debugging
      use_multiprocessing=True,
      workers=4)


      But saving the model after training is unclear:



      out_file = "../../models/:_epoch_:_weights.h5".format(MODEL_DESC, 5)
      model_2.save_weights(out_file) # save the actual Tower weights, discard the "booster" wrapper
      print("Saved: :".format(out_file))


      Or:



      out_file = "../../models/:_epoch_:_weights.h5".format(MODEL_DESC, 5)
      model_2.save(out_file) # save the actual Tower weights, discard the "booster" wrapper
      print("Saved: :".format(out_file))


      Or something else?



      The current failure modes seem to be if I try to load in just the weights into a newly instantiated model_2 instance I get:



      ValueError: axes don't match array


      Which from searching may be related to a bug in Keras. If I save the model (.save() rather than .save_weights() then it loads without complaint but the inference is not stable and appears to be horrible/random.)



      Thank you.



      Still getting the following traceback:



      <snip>/src/notebooks/vectorizer.py in load_model()
      65
      66 # load the weights
      ---> 67 loaded_model.load_weights(weights_path)
      68
      69 print("Model ready")

      /opt/conda/lib/python3.6/site-packages/keras/engine/network.py in load_weights(self, filepath, by_name, skip_mismatch, reshape)
      1164 else:
      1165 saving.load_weights_from_hdf5_group(
      -> 1166 f, self.layers, reshape=reshape)
      1167
      1168 def _updated_config(self):

      /opt/conda/lib/python3.6/site-packages/keras/engine/saving.py in load_weights_from_hdf5_group(f, layers, reshape)
      1043 original_keras_version,
      1044 original_backend,
      -> 1045 reshape=reshape)
      1046 if len(weight_values) != len(symbolic_weights):
      1047 raise ValueError('Layer #' + str(k) +

      /opt/conda/lib/python3.6/site-packages/keras/engine/saving.py in preprocess_weights_for_loading(layer, weights, original_keras_version, original_backend, reshape)
      680 weights = convert_nested_time_distributed(weights)
      681 elif layer.__class__.__name__ in ['Model', 'Sequential']:
      --> 682 weights = convert_nested_model(weights)
      683
      684 if original_keras_version == '1':

      /opt/conda/lib/python3.6/site-packages/keras/engine/saving.py in convert_nested_model(weights)
      668 weights=weights[:num_weights],
      669 original_keras_version=original_keras_version,
      --> 670 original_backend=original_backend))
      671 weights = weights[num_weights:]
      672 return new_weights

      /opt/conda/lib/python3.6/site-packages/keras/engine/saving.py in preprocess_weights_for_loading(layer, weights, original_keras_version, original_backend, reshape)
      680 weights = convert_nested_time_distributed(weights)
      681 elif layer.__class__.__name__ in ['Model', 'Sequential']:
      --> 682 weights = convert_nested_model(weights)
      683
      684 if original_keras_version == '1':

      /opt/conda/lib/python3.6/site-packages/keras/engine/saving.py in convert_nested_model(weights)
      656 weights=weights[:num_weights],
      657 original_keras_version=original_keras_version,
      --> 658 original_backend=original_backend))
      659 weights = weights[num_weights:]
      660

      /opt/conda/lib/python3.6/site-packages/keras/engine/saving.py in preprocess_weights_for_loading(layer, weights, original_keras_version, original_backend, reshape)
      799 weights[0] = np.reshape(weights[0], layer_weights_shape)
      800 elif layer_weights_shape != weights[0].shape:
      --> 801 weights[0] = np.transpose(weights[0], (3, 2, 0, 1))
      802 if layer.__class__.__name__ == 'ConvLSTM2D':
      803 weights[1] = np.transpose(weights[1], (3, 2, 0, 1))

      /opt/conda/lib/python3.6/site-packages/numpy/core/fromnumeric.py in transpose(a, axes)
      596
      597 """
      --> 598 return _wrapfunc(a, 'transpose', axes)
      599
      600

      /opt/conda/lib/python3.6/site-packages/numpy/core/fromnumeric.py in _wrapfunc(obj, method, *args, **kwds)
      49 def _wrapfunc(obj, method, *args, **kwds):
      50 try:
      ---> 51 return getattr(obj, method)(*args, **kwds)
      52
      53 # An AttributeError occurs if the object does not have

      ValueError: axes don't match array






      neural-network keras convnet computer-vision image-recognition






      share|improve this question















      share|improve this question













      share|improve this question




      share|improve this question








      edited Oct 11 '18 at 0:15







      PixelatedBrian

















      asked Oct 10 '18 at 17:29









      PixelatedBrianPixelatedBrian

      63




      63




















          2 Answers
          2






          active

          oldest

          votes


















          0












          $begingroup$

          Try to save the model to JSON, and the weights in HDF5 format with save_weights().



          # save the model
          model_json = model_2.to_json()
          with open("model_2.json", "w") as j_file:
          j_file.write(model_json)

          # save the weights
          model.save_weights("model_2.h5")


          Later to load the model:



          # load the model
          j_file = open('model_2.json', 'r')
          loaded_json_model = j_file.read()
          j_file.close()
          loaded_model = model_from_json(loaded_json_model)


          # load the weights
          loaded_model.load_weights("model_2.h5")





          share|improve this answer









          $endgroup$












          • $begingroup$
            Edited with traceback of error. Same error as before.
            $endgroup$
            – PixelatedBrian
            Oct 11 '18 at 0:16


















          0












          $begingroup$

          After you've done training the model:



          final_model.save('model.h5')


          To reload the model, simply use:



          from keras import load_model
          model=load_model('model.h5')





          share|improve this answer









          $endgroup$













            Your Answer








            StackExchange.ready(function()
            var channelOptions =
            tags: "".split(" "),
            id: "557"
            ;
            initTagRenderer("".split(" "), "".split(" "), channelOptions);

            StackExchange.using("externalEditor", function()
            // Have to fire editor after snippets, if snippets enabled
            if (StackExchange.settings.snippets.snippetsEnabled)
            StackExchange.using("snippets", function()
            createEditor();
            );

            else
            createEditor();

            );

            function createEditor()
            StackExchange.prepareEditor(
            heartbeatType: 'answer',
            autoActivateHeartbeat: false,
            convertImagesToLinks: false,
            noModals: true,
            showLowRepImageUploadWarning: true,
            reputationToPostImages: null,
            bindNavPrevention: true,
            postfix: "",
            imageUploader:
            brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
            contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
            allowUrls: true
            ,
            onDemand: true,
            discardSelector: ".discard-answer"
            ,immediatelyShowMarkdownHelp:true
            );



            );













            draft saved

            draft discarded


















            StackExchange.ready(
            function ()
            StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fdatascience.stackexchange.com%2fquestions%2f39483%2fhow-to-properly-save-and-load-an-intermediate-model-in-keras%23new-answer', 'question_page');

            );

            Post as a guest















            Required, but never shown

























            2 Answers
            2






            active

            oldest

            votes








            2 Answers
            2






            active

            oldest

            votes









            active

            oldest

            votes






            active

            oldest

            votes









            0












            $begingroup$

            Try to save the model to JSON, and the weights in HDF5 format with save_weights().



            # save the model
            model_json = model_2.to_json()
            with open("model_2.json", "w") as j_file:
            j_file.write(model_json)

            # save the weights
            model.save_weights("model_2.h5")


            Later to load the model:



            # load the model
            j_file = open('model_2.json', 'r')
            loaded_json_model = j_file.read()
            j_file.close()
            loaded_model = model_from_json(loaded_json_model)


            # load the weights
            loaded_model.load_weights("model_2.h5")





            share|improve this answer









            $endgroup$












            • $begingroup$
              Edited with traceback of error. Same error as before.
              $endgroup$
              – PixelatedBrian
              Oct 11 '18 at 0:16















            0












            $begingroup$

            Try to save the model to JSON, and the weights in HDF5 format with save_weights().



            # save the model
            model_json = model_2.to_json()
            with open("model_2.json", "w") as j_file:
            j_file.write(model_json)

            # save the weights
            model.save_weights("model_2.h5")


            Later to load the model:



            # load the model
            j_file = open('model_2.json', 'r')
            loaded_json_model = j_file.read()
            j_file.close()
            loaded_model = model_from_json(loaded_json_model)


            # load the weights
            loaded_model.load_weights("model_2.h5")





            share|improve this answer









            $endgroup$












            • $begingroup$
              Edited with traceback of error. Same error as before.
              $endgroup$
              – PixelatedBrian
              Oct 11 '18 at 0:16













            0












            0








            0





            $begingroup$

            Try to save the model to JSON, and the weights in HDF5 format with save_weights().



            # save the model
            model_json = model_2.to_json()
            with open("model_2.json", "w") as j_file:
            j_file.write(model_json)

            # save the weights
            model.save_weights("model_2.h5")


            Later to load the model:



            # load the model
            j_file = open('model_2.json', 'r')
            loaded_json_model = j_file.read()
            j_file.close()
            loaded_model = model_from_json(loaded_json_model)


            # load the weights
            loaded_model.load_weights("model_2.h5")





            share|improve this answer









            $endgroup$



            Try to save the model to JSON, and the weights in HDF5 format with save_weights().



            # save the model
            model_json = model_2.to_json()
            with open("model_2.json", "w") as j_file:
            j_file.write(model_json)

            # save the weights
            model.save_weights("model_2.h5")


            Later to load the model:



            # load the model
            j_file = open('model_2.json', 'r')
            loaded_json_model = j_file.read()
            j_file.close()
            loaded_model = model_from_json(loaded_json_model)


            # load the weights
            loaded_model.load_weights("model_2.h5")






            share|improve this answer












            share|improve this answer



            share|improve this answer










            answered Oct 10 '18 at 20:10









            bakkabakka

            26816




            26816











            • $begingroup$
              Edited with traceback of error. Same error as before.
              $endgroup$
              – PixelatedBrian
              Oct 11 '18 at 0:16
















            • $begingroup$
              Edited with traceback of error. Same error as before.
              $endgroup$
              – PixelatedBrian
              Oct 11 '18 at 0:16















            $begingroup$
            Edited with traceback of error. Same error as before.
            $endgroup$
            – PixelatedBrian
            Oct 11 '18 at 0:16




            $begingroup$
            Edited with traceback of error. Same error as before.
            $endgroup$
            – PixelatedBrian
            Oct 11 '18 at 0:16











            0












            $begingroup$

            After you've done training the model:



            final_model.save('model.h5')


            To reload the model, simply use:



            from keras import load_model
            model=load_model('model.h5')





            share|improve this answer









            $endgroup$

















              0












              $begingroup$

              After you've done training the model:



              final_model.save('model.h5')


              To reload the model, simply use:



              from keras import load_model
              model=load_model('model.h5')





              share|improve this answer









              $endgroup$















                0












                0








                0





                $begingroup$

                After you've done training the model:



                final_model.save('model.h5')


                To reload the model, simply use:



                from keras import load_model
                model=load_model('model.h5')





                share|improve this answer









                $endgroup$



                After you've done training the model:



                final_model.save('model.h5')


                To reload the model, simply use:



                from keras import load_model
                model=load_model('model.h5')






                share|improve this answer












                share|improve this answer



                share|improve this answer










                answered Apr 10 at 15:37









                Tanay KarveTanay Karve

                1




                1



























                    draft saved

                    draft discarded
















































                    Thanks for contributing an answer to Data Science Stack Exchange!


                    • Please be sure to answer the question. Provide details and share your research!

                    But avoid


                    • Asking for help, clarification, or responding to other answers.

                    • Making statements based on opinion; back them up with references or personal experience.

                    Use MathJax to format equations. MathJax reference.


                    To learn more, see our tips on writing great answers.




                    draft saved


                    draft discarded














                    StackExchange.ready(
                    function ()
                    StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fdatascience.stackexchange.com%2fquestions%2f39483%2fhow-to-properly-save-and-load-an-intermediate-model-in-keras%23new-answer', 'question_page');

                    );

                    Post as a guest















                    Required, but never shown





















































                    Required, but never shown














                    Required, but never shown












                    Required, but never shown







                    Required, but never shown

































                    Required, but never shown














                    Required, but never shown












                    Required, but never shown







                    Required, but never shown







                    Popular posts from this blog

                    Adding axes to figuresAdding axes labels to LaTeX figuresLaTeX equivalent of ConTeXt buffersRotate a node but not its content: the case of the ellipse decorationHow to define the default vertical distance between nodes?TikZ scaling graphic and adjust node position and keep font sizeNumerical conditional within tikz keys?adding axes to shapesAlign axes across subfiguresAdding figures with a certain orderLine up nested tikz enviroments or how to get rid of themAdding axes labels to LaTeX figures

                    Tähtien Talli Jäsenet | Lähteet | NavigointivalikkoSuomen Hippos – Tähtien Talli

                    Do these cracks on my tires look bad? The Next CEO of Stack OverflowDry rot tire should I replace?Having to replace tiresFishtailed so easily? Bad tires? ABS?Filling the tires with something other than air, to avoid puncture hassles?Used Michelin tires safe to install?Do these tyre cracks necessitate replacement?Rumbling noise: tires or mechanicalIs it possible to fix noisy feathered tires?Are bad winter tires still better than summer tires in winter?Torque converter failure - Related to replacing only 2 tires?Why use snow tires on all 4 wheels on 2-wheel-drive cars?