Tag: sequence-to-sequence

9 Input for LSTM for financial time series directional prediction 2018-01-25T17:12:41.587

9 How do attention mechanisms in RNNs learn weights for a variable length input 2018-01-30T00:35:51.420

8 Why do we need to add START <s> + END </s> symbols when using Recurrent Neural Nets for Sequence-to-Sequence Models? 2018-01-23T09:39:17.013

7 How to determine feature importance in a neural network? 2019-01-27T14:01:53.513

4 How/What to initialize the hidden states in RNN sequence-to-sequence models? 2018-01-30T06:30:54.517

3 Encoder-Decoder Sequence-to-Sequence Model for Translations in Both Directions 2018-08-01T13:30:11.267

3 Is this a problem for a Seq2Seq model? 2018-08-19T19:45:13.700

3 Predict output sequence one at a time with feedback 2018-08-20T06:34:06.920

3 Why does Position Embeddings work? 2018-11-08T16:05:00.290

3 Does this encoder-decoder LSTM make sense for time series sequence to sequence? 2018-12-12T13:07:32.217

3 One-hot encode multi-class multi-label sequences 2019-02-26T12:32:50.310

3 SymbolicException: Inputs to eager execution function cannot be Keras symbolic tensors 2019-12-08T17:29:11.803

3 Answer to Question 2019-12-21T15:31:45.223

3 Sentences language translation with neural network, with a simple layer structure (if possible sequential) 2020-02-12T11:16:46.280

2 Very long sequence in neural networks 2018-05-03T08:17:13.427

2 Training Encoder-Decoder using Decoder Outputs 2018-07-07T14:20:51.903

2 Keras: Softmax output into embedding layer 2018-07-24T19:57:54.097

2 Give Variable Length input to LSTM 2018-11-04T15:21:12.363

2 Predicting next number in a sequence - data analysis 2018-12-14T14:05:22.767

2 Recommended model for univariate or multivariate multistep ahead time series forecasting 2018-12-19T11:59:44.973

2 Working ofLSTM with multiple Units - NER 2019-04-04T23:47:22.233

2 ValueError: Cannot convert a partially known TensorShape to a Tensor: (?, 256) 2019-04-28T23:24:10.627

2 IndexError: list index out of range 2019-05-02T21:08:31.693

2 Why isnt my seq2seq model reconising what the <END> tag is? 2019-07-21T21:56:34.787

2 Any good Implementations of Bi-LSTM bahdanau attention in Keras? 2019-12-02T21:22:22.810

2 Methods for learning with noisy labels 2020-02-16T10:13:27.887

2 Can I fine-tune BERT, ELMO or XLnet for Seq2Seq neural machine translation? 2020-02-24T08:40:38.953

2 LSTM low training/validation error but really bad predictions 2020-07-12T07:33:03.767

2 What is the best way for synthetic data generation while maintaining privacy? 2020-07-17T18:38:34.417

2 Does the output of the Sequence-to-Sequence encoder model exist in the same semantic space as the inputs (Word2vec)? 2020-10-19T15:47:36.293

2 How to train a model on top of a transformer to output a sequence? 2020-10-30T11:37:22.500

2 Build a corpus for machine translation 2020-12-29T22:07:10.357

1 How to pad real-valued sequences 2018-03-21T05:30:18.500

1 Group prediction 2018-04-18T05:40:49.820

1 What are the benefits and tradeoffs of a 1D conv vs a multi-input seq2seq LSTM model? 2018-07-16T13:47:53.210

1 Neural network outputting same result for all inputs 2018-07-26T16:39:05.017

1 On-the-fly seq2seq: starting translation before the input sequence ends 2018-07-30T12:18:15.053

1 Can Sequence to sequence models be used to convert code from one programming language to another? 2018-10-04T19:29:44.960

1 What should the size of the decoder output be in a sequence to sequence model 2019-02-02T11:53:50.710

1 Initialising states in a multilayer sequence to sequence model 2019-02-07T12:46:53.097

1 Adding context in a sequence to sequence problem 2019-04-01T11:17:57.347

1 How to create a seq2seq without specifying a fixed decoder length? 2019-10-18T17:48:54.390

1 Improving Performance of LSTM for time series prediction 2019-11-26T09:57:12.800

1 Weight matrices in transformers 2019-12-05T10:34:50.910

1 How can I feed BERT to neural machine translation? 2019-12-06T11:50:33.677

1 How can I do a sequence to sequence model (RNN / LSTM) with Keras with fixed length data? 2020-02-14T14:03:35.790

1 How can I finetune XLM-R for neural machine translation between the same language(Catalan to Catalan-with different grammar structure))? 2020-02-27T05:19:27.570

1 Which input to use when generating a new sequence 2020-03-04T14:20:10.340

1 Sequence labeling with partially known labels 2020-03-08T07:46:25.183

1 Seq2Seq for sentence correction 2020-03-08T08:45:40.717

1 Comparing Language Model of two corpora 2020-04-30T20:13:40.567

1 How to add attention mechanism to my sequence-to-sequence architecture in Keras? 2020-05-17T19:11:57.523

1 When using padding in sequence models, is Keras validation accuracy valid/ reliable? 2020-07-19T19:47:48.377

1 Pytorch LSTM not training 2020-09-21T22:34:53.170

1 Why does an attention layer in a transformer learn context? 2020-11-12T15:31:37.863

1 Long range forecasting with sequence-to-sequence models 2020-12-11T02:13:20.567

1 Timeseries LSTM: does test data need to come after training data? 2021-02-23T12:46:17.790

0 Do we really need `<unk>` tokens? 2018-06-20T21:46:57.343

0 What type of neural network could emulate a binary to HTML conversion tool? 2018-10-11T08:21:07.433

0 Strategy for "forcing" number of labels in seq2seq predictions with Keras? 2019-04-25T11:23:29.567

0 Seq2Seq Model training: Encoder vs. Decoder 2019-06-14T16:45:21.113

0 Confusion about Decoder labels for training seq-to-seq models 2019-08-05T22:57:30.490

0 How can I build a seq2seq model , which is topic aware 2019-10-01T12:45:09.990

0 two different attention methods for seq2seq 2019-11-05T02:44:41.240

0 How are Q, K, and V Vectors Trained in a Transformer Self-Attention? 2020-02-17T09:55:54.033

0 Is it possible feed BERT to seq2seq encoder/decoder NMT (for low resource language)? 2020-02-22T01:54:34.677

0 Seeking your advice on XLM-R for NMT 2020-03-07T10:31:35.323

0 Transformer seq2seq model and loading embeddings from XLM-RoBERTa 2020-03-11T18:01:15.257

0 Multiple time series sequence prediction for multiple multivariate time series 2020-03-17T13:24:00.130

0 Feeding XLM-R embeddings to neural machine translation? 2020-03-18T08:19:33.093

0 Tied embedding in Sequence to Sequence Task 2020-04-30T19:26:51.820

0 Concatenating Encoder hidden states in LSTM pytorch 2020-06-26T02:44:00.650

0 How to approach a text parsing problem 2020-07-30T22:55:38.827

0 Sequence-to-Sequence Autoencoder with asymmetric length of encoding sequence 2020-08-03T18:05:27.217

0 Why does the non autoregresive transfomer model in fairseq require the prev_output_tokens input? 2020-08-23T08:35:23.377

0 Case weights with sequence to sequence models? 2020-10-07T19:51:54.363

0 Learning conditional statements from natural language 2020-11-10T04:39:35.597

0 Using Transcoder Model for language to language conversion 2020-11-17T16:48:08.783

0 How to reduce dimensionality of encoder decoder output? 2020-11-19T12:42:49.737

0 Predict customer behaviour with Transformer(attention is all you need) 2020-12-05T11:24:21.987

0 How far into the future can I forecast a time-series with an LSTM and strongly seasonal data 2021-01-07T16:41:31.977

0 How to use BERT in seq2seq model? 2021-02-14T11:40:02.000

0 Difference between zero-padding and character-padding in Recurrent Neural Networks 2021-02-27T10:13:26.473