Using VAE with Sequence to Sequence Approach Announcing the arrival of Valued Associate #679: Cesar Manara Planned maintenance scheduled April 17/18, 2019 at 00:00UTC (8:00pm US/Eastern) 2019 Moderator Election Q&A - Questionnaire 2019 Community Moderator Election ResultsWhy do we need to add START <s> + END </s> symbols when using Recurrent Neural Nets for Sequence-to-Sequence Models?Keras VAE example loss functionVery long sequence in neural networksTraining Encoder-Decoder using Decoder OutputsOn-the-fly seq2seq: starting translation before the input sequence endsEncoder-Decoder Sequence-to-Sequence Model for Translations in Both DirectionsCan Sequence to sequence models be used to convert code from one programming language to another?KL divergence in VAEVariational auto-encoders (VAE): why the random sample?Understanding ELBO Learning Dynamics for VAE?

3 doors, three guards, one stone

Why constant symbols in a language?

Bonus calculation: Am I making a mountain out of a molehill?

Were Kohanim forbidden from serving in King David's army?

When -s is used with third person singular. What's its use in this context?

LaTeX gives error undefined control sequence table

What is the role of the transistor and diode in a soft start circuit?

Why did the IBM 650 use bi-quinary?

What are the motives behind Cersei's orders given to Bronn?

How to recreate this effect in Photoshop?

Do you forfeit tax refunds/credits if you aren't required to and don't file by April 15?

Can a non-EU citizen traveling with me come with me through the EU passport line?

How to bypass password on Windows XP account?

What does the "x" in "x86" represent?

Check which numbers satisfy the condition [A*B*C = A! + B! + C!]

Why was the term "discrete" used in discrete logarithm?

Did Xerox really develop the first LAN?

Is above average number of years spent on PhD considered a red flag in future academia or industry positions?

How do I mention the quality of my school without bragging

What are the possible ways to detect skin while classifying diseases?

Echoing a tail command produces unexpected output?

Is there a concise way to say "all of the X, one of each"?

Stars Make Stars

I am not a queen, who am I?



Using VAE with Sequence to Sequence Approach



Announcing the arrival of Valued Associate #679: Cesar Manara
Planned maintenance scheduled April 17/18, 2019 at 00:00UTC (8:00pm US/Eastern)
2019 Moderator Election Q&A - Questionnaire
2019 Community Moderator Election ResultsWhy do we need to add START <s> + END </s> symbols when using Recurrent Neural Nets for Sequence-to-Sequence Models?Keras VAE example loss functionVery long sequence in neural networksTraining Encoder-Decoder using Decoder OutputsOn-the-fly seq2seq: starting translation before the input sequence endsEncoder-Decoder Sequence-to-Sequence Model for Translations in Both DirectionsCan Sequence to sequence models be used to convert code from one programming language to another?KL divergence in VAEVariational auto-encoders (VAE): why the random sample?Understanding ELBO Learning Dynamics for VAE?










0












$begingroup$


In the code below, I'm using a VAE with a seq-to-seq approach for translation. At the beginning I sarted only by using a simple seq-to-seq approach which implements a RNN-AE, until this step I had not errors. When I try to use a VAE by adding the two layers for mean and deviation , and by changing the loss function, I get this error :




tensorflow.python.framework.errors_impl.InvalidArgumentError: Incompatible shapes [32,153] vs [32]




In this line of code:




Traceback (most recent call last):
in module validation_steps = 1




I get this error even when I change th validation_steps value or the batch-size



# Train - Test Split
X, y = lines.eng, lines.fr
X_train, X_test, y_train, y_test = train_test_split(X, y, test_size = 0.1)
print(X_train.shape, X_test.shape)

"""fonction pour charger le data par lot (batch size)"""
def generate_batch(X = X_train, y = y_train, batch_size = 32):
''' Generate a batch of data '''
while True:
for j in range(0, len(X), batch_size):
encoder_input_data = np.zeros((batch_size,
max_len_eng),dtype='float32')
decoder_input_data =
np.zeros((batch_size,max_len_fr),dtype='float32')
decoder_target_data = np.zeros((batch_size,max_len_fr,
num_decoder_tokens),
dtype='float32')
#for i, (input_text, target_text) in enumerate(zip(lines.eng, lines.fr)):
for i, (input_text, target_text) in enumerate(zip(X[j:j+batch_size],
y[j:j+batch_size])):
for t, word in enumerate(input_text.split()):
encoder_input_data[i, t] = input_token_index[word]
for t, word in enumerate(target_text.split()):
# decoder_target_data is ahead of decoder_input_data by one timestep
if t<len(target_text.split())-1:
decoder_input_data[i, t] = target_token_index[word]
# decoder input seq
if t > 0:
# decoder_target_data will be ahead by one timestep
# and will not include the start character.
decoder_target_data[i, t - 1, target_token_index[word]] = 1
yield([encoder_input_data, decoder_input_data],
decoder_target_data)

encoder_inputs = Input(shape=(None,))
en_x= Embedding(num_encoder_tokens, embedding_size,mask_zero =
True)(encoder_inputs)
encoder = LSTM(50, return_state=True)
encoder_outputs, state_h, state_c = encoder(en_x) #initialisé à 0
# We discard `encoder_outputs` and only keep the states.
encoder_states = [state_h, state_c]
""" ---------------Here change for VAE----------------- """
""" ___________________________________________________ """
latent_dim =embedding_size
# output layer for mean and log variance
z_mu = Dense(latent_dim)(encoder_outputs) #remplacer h
z_log_var = Dense(latent_dim)(encoder_outputs)

def sampling(args):
batch_size=32
z_mean, z_log_sigma = args
epsilon = K.random_normal(shape=(batch_size, latent_dim),
mean=0., stddev=1.)
return z_mean + K.exp(z_log_sigma) * epsilon

# note that "output_shape" isn't necessary with the TensorFlow backend
# so you could write `Lambda(sampling)([z_mean, z_log_sigma])`
z = Lambda(sampling, output_shape=(latent_dim,))([z_mu, z_log_var])
z1 = Lambda(sampling, output_shape=(latent_dim,))([z_mu, z_log_var])

#loss function with VAE
def vae_loss(y_true, y_pred):
# E[log P(X|z)]
recon = K.sum(K.binary_crossentropy(y_pred, y_true), axis=1)
# D_KL(Q(z|X) || P(z|X)); calculate in closed form as both dist. are
Gaussian
kl = 0.5 * K.sum(K.exp(z_log_var) + K.square(z_mu) - 1. - z_log_var, axis=1)

return recon + kl
""" ----------------------------------------------------"""
""" ----------------------------------------------------"""
# Set up the decoder, using `encoder_states` as initial state.
decoder_inputs = Input(shape=(None,))
dex= Embedding(num_decoder_tokens, embedding_size,mask_zero = True)
final_dex= dex(decoder_inputs)
decoder_lstm = LSTM(50, return_sequences=True, return_state=True)

decoder_outputs, _, _ = decoder_lstm(final_dex,initial_state=[z, z1])
decoder_dense = Dense(num_decoder_tokens, activation='softmax')
decoder_outputs = decoder_dense(decoder_outputs)
model = Model([encoder_inputs, decoder_inputs], decoder_outputs)
model.compile(optimizer='rmsprop', loss=vae_loss, metrics=['acc'])
model.summary()
train_samples = len(X_train)
val_samples = len(X_test)
batch_size = 32
epochs = 10

model.fit_generator(generator = generate_batch(X_train, y_train, batch_size
= batch_size),
steps_per_epoch = train_samples//batch_size,
epochs=epochs,
validation_data = generate_batch(X_test, y_test, batch_size
= batch_size),
validation_steps = 1)

encoder_model = Model(encoder_inputs, encoder_states)


this is the model summary:
enter image description here



Thank you in advance for your help.










share|improve this question











$endgroup$
















    0












    $begingroup$


    In the code below, I'm using a VAE with a seq-to-seq approach for translation. At the beginning I sarted only by using a simple seq-to-seq approach which implements a RNN-AE, until this step I had not errors. When I try to use a VAE by adding the two layers for mean and deviation , and by changing the loss function, I get this error :




    tensorflow.python.framework.errors_impl.InvalidArgumentError: Incompatible shapes [32,153] vs [32]




    In this line of code:




    Traceback (most recent call last):
    in module validation_steps = 1




    I get this error even when I change th validation_steps value or the batch-size



    # Train - Test Split
    X, y = lines.eng, lines.fr
    X_train, X_test, y_train, y_test = train_test_split(X, y, test_size = 0.1)
    print(X_train.shape, X_test.shape)

    """fonction pour charger le data par lot (batch size)"""
    def generate_batch(X = X_train, y = y_train, batch_size = 32):
    ''' Generate a batch of data '''
    while True:
    for j in range(0, len(X), batch_size):
    encoder_input_data = np.zeros((batch_size,
    max_len_eng),dtype='float32')
    decoder_input_data =
    np.zeros((batch_size,max_len_fr),dtype='float32')
    decoder_target_data = np.zeros((batch_size,max_len_fr,
    num_decoder_tokens),
    dtype='float32')
    #for i, (input_text, target_text) in enumerate(zip(lines.eng, lines.fr)):
    for i, (input_text, target_text) in enumerate(zip(X[j:j+batch_size],
    y[j:j+batch_size])):
    for t, word in enumerate(input_text.split()):
    encoder_input_data[i, t] = input_token_index[word]
    for t, word in enumerate(target_text.split()):
    # decoder_target_data is ahead of decoder_input_data by one timestep
    if t<len(target_text.split())-1:
    decoder_input_data[i, t] = target_token_index[word]
    # decoder input seq
    if t > 0:
    # decoder_target_data will be ahead by one timestep
    # and will not include the start character.
    decoder_target_data[i, t - 1, target_token_index[word]] = 1
    yield([encoder_input_data, decoder_input_data],
    decoder_target_data)

    encoder_inputs = Input(shape=(None,))
    en_x= Embedding(num_encoder_tokens, embedding_size,mask_zero =
    True)(encoder_inputs)
    encoder = LSTM(50, return_state=True)
    encoder_outputs, state_h, state_c = encoder(en_x) #initialisé à 0
    # We discard `encoder_outputs` and only keep the states.
    encoder_states = [state_h, state_c]
    """ ---------------Here change for VAE----------------- """
    """ ___________________________________________________ """
    latent_dim =embedding_size
    # output layer for mean and log variance
    z_mu = Dense(latent_dim)(encoder_outputs) #remplacer h
    z_log_var = Dense(latent_dim)(encoder_outputs)

    def sampling(args):
    batch_size=32
    z_mean, z_log_sigma = args
    epsilon = K.random_normal(shape=(batch_size, latent_dim),
    mean=0., stddev=1.)
    return z_mean + K.exp(z_log_sigma) * epsilon

    # note that "output_shape" isn't necessary with the TensorFlow backend
    # so you could write `Lambda(sampling)([z_mean, z_log_sigma])`
    z = Lambda(sampling, output_shape=(latent_dim,))([z_mu, z_log_var])
    z1 = Lambda(sampling, output_shape=(latent_dim,))([z_mu, z_log_var])

    #loss function with VAE
    def vae_loss(y_true, y_pred):
    # E[log P(X|z)]
    recon = K.sum(K.binary_crossentropy(y_pred, y_true), axis=1)
    # D_KL(Q(z|X) || P(z|X)); calculate in closed form as both dist. are
    Gaussian
    kl = 0.5 * K.sum(K.exp(z_log_var) + K.square(z_mu) - 1. - z_log_var, axis=1)

    return recon + kl
    """ ----------------------------------------------------"""
    """ ----------------------------------------------------"""
    # Set up the decoder, using `encoder_states` as initial state.
    decoder_inputs = Input(shape=(None,))
    dex= Embedding(num_decoder_tokens, embedding_size,mask_zero = True)
    final_dex= dex(decoder_inputs)
    decoder_lstm = LSTM(50, return_sequences=True, return_state=True)

    decoder_outputs, _, _ = decoder_lstm(final_dex,initial_state=[z, z1])
    decoder_dense = Dense(num_decoder_tokens, activation='softmax')
    decoder_outputs = decoder_dense(decoder_outputs)
    model = Model([encoder_inputs, decoder_inputs], decoder_outputs)
    model.compile(optimizer='rmsprop', loss=vae_loss, metrics=['acc'])
    model.summary()
    train_samples = len(X_train)
    val_samples = len(X_test)
    batch_size = 32
    epochs = 10

    model.fit_generator(generator = generate_batch(X_train, y_train, batch_size
    = batch_size),
    steps_per_epoch = train_samples//batch_size,
    epochs=epochs,
    validation_data = generate_batch(X_test, y_test, batch_size
    = batch_size),
    validation_steps = 1)

    encoder_model = Model(encoder_inputs, encoder_states)


    this is the model summary:
    enter image description here



    Thank you in advance for your help.










    share|improve this question











    $endgroup$














      0












      0








      0





      $begingroup$


      In the code below, I'm using a VAE with a seq-to-seq approach for translation. At the beginning I sarted only by using a simple seq-to-seq approach which implements a RNN-AE, until this step I had not errors. When I try to use a VAE by adding the two layers for mean and deviation , and by changing the loss function, I get this error :




      tensorflow.python.framework.errors_impl.InvalidArgumentError: Incompatible shapes [32,153] vs [32]




      In this line of code:




      Traceback (most recent call last):
      in module validation_steps = 1




      I get this error even when I change th validation_steps value or the batch-size



      # Train - Test Split
      X, y = lines.eng, lines.fr
      X_train, X_test, y_train, y_test = train_test_split(X, y, test_size = 0.1)
      print(X_train.shape, X_test.shape)

      """fonction pour charger le data par lot (batch size)"""
      def generate_batch(X = X_train, y = y_train, batch_size = 32):
      ''' Generate a batch of data '''
      while True:
      for j in range(0, len(X), batch_size):
      encoder_input_data = np.zeros((batch_size,
      max_len_eng),dtype='float32')
      decoder_input_data =
      np.zeros((batch_size,max_len_fr),dtype='float32')
      decoder_target_data = np.zeros((batch_size,max_len_fr,
      num_decoder_tokens),
      dtype='float32')
      #for i, (input_text, target_text) in enumerate(zip(lines.eng, lines.fr)):
      for i, (input_text, target_text) in enumerate(zip(X[j:j+batch_size],
      y[j:j+batch_size])):
      for t, word in enumerate(input_text.split()):
      encoder_input_data[i, t] = input_token_index[word]
      for t, word in enumerate(target_text.split()):
      # decoder_target_data is ahead of decoder_input_data by one timestep
      if t<len(target_text.split())-1:
      decoder_input_data[i, t] = target_token_index[word]
      # decoder input seq
      if t > 0:
      # decoder_target_data will be ahead by one timestep
      # and will not include the start character.
      decoder_target_data[i, t - 1, target_token_index[word]] = 1
      yield([encoder_input_data, decoder_input_data],
      decoder_target_data)

      encoder_inputs = Input(shape=(None,))
      en_x= Embedding(num_encoder_tokens, embedding_size,mask_zero =
      True)(encoder_inputs)
      encoder = LSTM(50, return_state=True)
      encoder_outputs, state_h, state_c = encoder(en_x) #initialisé à 0
      # We discard `encoder_outputs` and only keep the states.
      encoder_states = [state_h, state_c]
      """ ---------------Here change for VAE----------------- """
      """ ___________________________________________________ """
      latent_dim =embedding_size
      # output layer for mean and log variance
      z_mu = Dense(latent_dim)(encoder_outputs) #remplacer h
      z_log_var = Dense(latent_dim)(encoder_outputs)

      def sampling(args):
      batch_size=32
      z_mean, z_log_sigma = args
      epsilon = K.random_normal(shape=(batch_size, latent_dim),
      mean=0., stddev=1.)
      return z_mean + K.exp(z_log_sigma) * epsilon

      # note that "output_shape" isn't necessary with the TensorFlow backend
      # so you could write `Lambda(sampling)([z_mean, z_log_sigma])`
      z = Lambda(sampling, output_shape=(latent_dim,))([z_mu, z_log_var])
      z1 = Lambda(sampling, output_shape=(latent_dim,))([z_mu, z_log_var])

      #loss function with VAE
      def vae_loss(y_true, y_pred):
      # E[log P(X|z)]
      recon = K.sum(K.binary_crossentropy(y_pred, y_true), axis=1)
      # D_KL(Q(z|X) || P(z|X)); calculate in closed form as both dist. are
      Gaussian
      kl = 0.5 * K.sum(K.exp(z_log_var) + K.square(z_mu) - 1. - z_log_var, axis=1)

      return recon + kl
      """ ----------------------------------------------------"""
      """ ----------------------------------------------------"""
      # Set up the decoder, using `encoder_states` as initial state.
      decoder_inputs = Input(shape=(None,))
      dex= Embedding(num_decoder_tokens, embedding_size,mask_zero = True)
      final_dex= dex(decoder_inputs)
      decoder_lstm = LSTM(50, return_sequences=True, return_state=True)

      decoder_outputs, _, _ = decoder_lstm(final_dex,initial_state=[z, z1])
      decoder_dense = Dense(num_decoder_tokens, activation='softmax')
      decoder_outputs = decoder_dense(decoder_outputs)
      model = Model([encoder_inputs, decoder_inputs], decoder_outputs)
      model.compile(optimizer='rmsprop', loss=vae_loss, metrics=['acc'])
      model.summary()
      train_samples = len(X_train)
      val_samples = len(X_test)
      batch_size = 32
      epochs = 10

      model.fit_generator(generator = generate_batch(X_train, y_train, batch_size
      = batch_size),
      steps_per_epoch = train_samples//batch_size,
      epochs=epochs,
      validation_data = generate_batch(X_test, y_test, batch_size
      = batch_size),
      validation_steps = 1)

      encoder_model = Model(encoder_inputs, encoder_states)


      this is the model summary:
      enter image description here



      Thank you in advance for your help.










      share|improve this question











      $endgroup$




      In the code below, I'm using a VAE with a seq-to-seq approach for translation. At the beginning I sarted only by using a simple seq-to-seq approach which implements a RNN-AE, until this step I had not errors. When I try to use a VAE by adding the two layers for mean and deviation , and by changing the loss function, I get this error :




      tensorflow.python.framework.errors_impl.InvalidArgumentError: Incompatible shapes [32,153] vs [32]




      In this line of code:




      Traceback (most recent call last):
      in module validation_steps = 1




      I get this error even when I change th validation_steps value or the batch-size



      # Train - Test Split
      X, y = lines.eng, lines.fr
      X_train, X_test, y_train, y_test = train_test_split(X, y, test_size = 0.1)
      print(X_train.shape, X_test.shape)

      """fonction pour charger le data par lot (batch size)"""
      def generate_batch(X = X_train, y = y_train, batch_size = 32):
      ''' Generate a batch of data '''
      while True:
      for j in range(0, len(X), batch_size):
      encoder_input_data = np.zeros((batch_size,
      max_len_eng),dtype='float32')
      decoder_input_data =
      np.zeros((batch_size,max_len_fr),dtype='float32')
      decoder_target_data = np.zeros((batch_size,max_len_fr,
      num_decoder_tokens),
      dtype='float32')
      #for i, (input_text, target_text) in enumerate(zip(lines.eng, lines.fr)):
      for i, (input_text, target_text) in enumerate(zip(X[j:j+batch_size],
      y[j:j+batch_size])):
      for t, word in enumerate(input_text.split()):
      encoder_input_data[i, t] = input_token_index[word]
      for t, word in enumerate(target_text.split()):
      # decoder_target_data is ahead of decoder_input_data by one timestep
      if t<len(target_text.split())-1:
      decoder_input_data[i, t] = target_token_index[word]
      # decoder input seq
      if t > 0:
      # decoder_target_data will be ahead by one timestep
      # and will not include the start character.
      decoder_target_data[i, t - 1, target_token_index[word]] = 1
      yield([encoder_input_data, decoder_input_data],
      decoder_target_data)

      encoder_inputs = Input(shape=(None,))
      en_x= Embedding(num_encoder_tokens, embedding_size,mask_zero =
      True)(encoder_inputs)
      encoder = LSTM(50, return_state=True)
      encoder_outputs, state_h, state_c = encoder(en_x) #initialisé à 0
      # We discard `encoder_outputs` and only keep the states.
      encoder_states = [state_h, state_c]
      """ ---------------Here change for VAE----------------- """
      """ ___________________________________________________ """
      latent_dim =embedding_size
      # output layer for mean and log variance
      z_mu = Dense(latent_dim)(encoder_outputs) #remplacer h
      z_log_var = Dense(latent_dim)(encoder_outputs)

      def sampling(args):
      batch_size=32
      z_mean, z_log_sigma = args
      epsilon = K.random_normal(shape=(batch_size, latent_dim),
      mean=0., stddev=1.)
      return z_mean + K.exp(z_log_sigma) * epsilon

      # note that "output_shape" isn't necessary with the TensorFlow backend
      # so you could write `Lambda(sampling)([z_mean, z_log_sigma])`
      z = Lambda(sampling, output_shape=(latent_dim,))([z_mu, z_log_var])
      z1 = Lambda(sampling, output_shape=(latent_dim,))([z_mu, z_log_var])

      #loss function with VAE
      def vae_loss(y_true, y_pred):
      # E[log P(X|z)]
      recon = K.sum(K.binary_crossentropy(y_pred, y_true), axis=1)
      # D_KL(Q(z|X) || P(z|X)); calculate in closed form as both dist. are
      Gaussian
      kl = 0.5 * K.sum(K.exp(z_log_var) + K.square(z_mu) - 1. - z_log_var, axis=1)

      return recon + kl
      """ ----------------------------------------------------"""
      """ ----------------------------------------------------"""
      # Set up the decoder, using `encoder_states` as initial state.
      decoder_inputs = Input(shape=(None,))
      dex= Embedding(num_decoder_tokens, embedding_size,mask_zero = True)
      final_dex= dex(decoder_inputs)
      decoder_lstm = LSTM(50, return_sequences=True, return_state=True)

      decoder_outputs, _, _ = decoder_lstm(final_dex,initial_state=[z, z1])
      decoder_dense = Dense(num_decoder_tokens, activation='softmax')
      decoder_outputs = decoder_dense(decoder_outputs)
      model = Model([encoder_inputs, decoder_inputs], decoder_outputs)
      model.compile(optimizer='rmsprop', loss=vae_loss, metrics=['acc'])
      model.summary()
      train_samples = len(X_train)
      val_samples = len(X_test)
      batch_size = 32
      epochs = 10

      model.fit_generator(generator = generate_batch(X_train, y_train, batch_size
      = batch_size),
      steps_per_epoch = train_samples//batch_size,
      epochs=epochs,
      validation_data = generate_batch(X_test, y_test, batch_size
      = batch_size),
      validation_steps = 1)

      encoder_model = Model(encoder_inputs, encoder_states)


      this is the model summary:
      enter image description here



      Thank you in advance for your help.







      python deep-learning tensorflow autoencoder sequence-to-sequence






      share|improve this question















      share|improve this question













      share|improve this question




      share|improve this question








      edited Apr 1 at 18:35







      Kikio

















      asked Apr 1 at 13:35









      KikioKikio

      638




      638




















          0






          active

          oldest

          votes












          Your Answer








          StackExchange.ready(function()
          var channelOptions =
          tags: "".split(" "),
          id: "557"
          ;
          initTagRenderer("".split(" "), "".split(" "), channelOptions);

          StackExchange.using("externalEditor", function()
          // Have to fire editor after snippets, if snippets enabled
          if (StackExchange.settings.snippets.snippetsEnabled)
          StackExchange.using("snippets", function()
          createEditor();
          );

          else
          createEditor();

          );

          function createEditor()
          StackExchange.prepareEditor(
          heartbeatType: 'answer',
          autoActivateHeartbeat: false,
          convertImagesToLinks: false,
          noModals: true,
          showLowRepImageUploadWarning: true,
          reputationToPostImages: null,
          bindNavPrevention: true,
          postfix: "",
          imageUploader:
          brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
          contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
          allowUrls: true
          ,
          onDemand: true,
          discardSelector: ".discard-answer"
          ,immediatelyShowMarkdownHelp:true
          );



          );













          draft saved

          draft discarded


















          StackExchange.ready(
          function ()
          StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fdatascience.stackexchange.com%2fquestions%2f48351%2fusing-vae-with-sequence-to-sequence-approach%23new-answer', 'question_page');

          );

          Post as a guest















          Required, but never shown

























          0






          active

          oldest

          votes








          0






          active

          oldest

          votes









          active

          oldest

          votes






          active

          oldest

          votes















          draft saved

          draft discarded
















































          Thanks for contributing an answer to Data Science Stack Exchange!


          • Please be sure to answer the question. Provide details and share your research!

          But avoid


          • Asking for help, clarification, or responding to other answers.

          • Making statements based on opinion; back them up with references or personal experience.

          Use MathJax to format equations. MathJax reference.


          To learn more, see our tips on writing great answers.




          draft saved


          draft discarded














          StackExchange.ready(
          function ()
          StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fdatascience.stackexchange.com%2fquestions%2f48351%2fusing-vae-with-sequence-to-sequence-approach%23new-answer', 'question_page');

          );

          Post as a guest















          Required, but never shown





















































          Required, but never shown














          Required, but never shown












          Required, but never shown







          Required, but never shown

































          Required, but never shown














          Required, but never shown












          Required, but never shown







          Required, but never shown







          Popular posts from this blog

          Adding axes to figuresAdding axes labels to LaTeX figuresLaTeX equivalent of ConTeXt buffersRotate a node but not its content: the case of the ellipse decorationHow to define the default vertical distance between nodes?TikZ scaling graphic and adjust node position and keep font sizeNumerical conditional within tikz keys?adding axes to shapesAlign axes across subfiguresAdding figures with a certain orderLine up nested tikz enviroments or how to get rid of themAdding axes labels to LaTeX figures

          Tähtien Talli Jäsenet | Lähteet | NavigointivalikkoSuomen Hippos – Tähtien Talli

          Do these cracks on my tires look bad? The Next CEO of Stack OverflowDry rot tire should I replace?Having to replace tiresFishtailed so easily? Bad tires? ABS?Filling the tires with something other than air, to avoid puncture hassles?Used Michelin tires safe to install?Do these tyre cracks necessitate replacement?Rumbling noise: tires or mechanicalIs it possible to fix noisy feathered tires?Are bad winter tires still better than summer tires in winter?Torque converter failure - Related to replacing only 2 tires?Why use snow tires on all 4 wheels on 2-wheel-drive cars?