Is there any other reason why we make sequence length the same length using padding? Other than in order to do matrix multiplication (therefore doing parallel computation).
Is there any other reason why we make sequence length the same using padding?
195 Views Asked by aerin At
1
There are 1 best solutions below
Related Questions in TENSORFLOW
- (Tensorflow)Does the op assign change the gradient computation?
- Tensorflow Windows Accessing Folders Denied:"NewRandomAccessFile failed to Create/Open: Access is denied. ; Input/output error"
- Android App TensorFlow Google Cloud ML
- Convert Tensorflow model to Caffe model
- Google Tensorflow LSTMCell Variables Mapping to Hochreiter97_lstm.pdf paper
- additive Gaussian noise in Tensorflow
- TFlearn evaluate method results meaning
- Regularization losses Tensorflow - TRAINABLE_VARIABLES to Tensor Array
- feed picture to model tensorflow for training
- Fail to read the new format of tensorflow checkpoint?
- I got a error when running a github project in tensorflow
- Tensorflow R0.12 softmax_cross_entropy_with_logits ASSERT Error
- RuntimeError in run_one_batch of TensorFlowDataFrame in tensorflow
- Same output in neural network for each input after training
- ConvNet : Validation Loss not strongly decreasing but accuracy is improving
Related Questions in PYTORCH
- Pytorch install with anaconda error
- How should I save the model of PyTorch if I want it loadable by OpenCV dnn module
- PyTorch: memorize output from several layers of sequencial
- in Pytorch, restore the model parameters but the same initial loss
- Seq2seq pytorch Inference slow
- Why does autograd not produce gradient for intermediate variables?
- pytorch inception model outputs the wrong label for every input image
- "expected CPU tensor(got CUDA tensor)" error for PyTorch
- Float16 (HalfTensor) in pytorch + cuda
- Access parameter names in torch
- Efficient way of calculating sum of unequal sized chunks of tensor in Pytorch
- what is the equivalent of theano.tensor.clip in pytorch?
- How can I do scatter and gather operations in NumPy?
- How do I write a PyTorch sequential model?
- How to combine multiple models together?
Related Questions in RECURRENT-NEURAL-NETWORK
- RNN Cell not present in tf.get_collection
- How to choose layers in RNN (recurrent neural networks)?
- Get the last output of a dynamic_rnn in TensorFlow
- Dynamic tensor shape for tensorflow RNN
- Choosing the Length of Time Steps in Recurrent Neural Network
- How to use keras RNN for text classification in a dataset?
- How to Initialize LSTMCell with tuple
- What is the most efficient way to implement multi-layer RNNs in TensorFlow?
- Confused about weight and bias dependencies affecting learning
- Mxnet RNN Time Series Prediction
- Recurrent Neural Networks for Panel Data
- Multi-dimension dynamic rnn with tensorflow
- Tensorflow Network Save and Restore
- Wrong Number of LSTM Dimensions in Keras
- Tensorflow seq2seq `feed_previous' argument`
Related Questions in SEQUENCE-MODELING
- Is adding a FC Linear layer on top of seq2seq architecture a potential source of data leaking from future to past?
- how to create train - dev - test sets from a given dataset in sequence models
- How can I do a sequence-to-sequence model (RNN / LSTM) with Keras with fixed length data?
- How to implement Bi-Directional Conv LSTM in Pytorch
- Is there any other reason why we make sequence length the same using padding?
- 1D CNN predictions plot mismatch with actual time series plot
- TF-IDF vector vs a vector of tokens
- How the function nn.LSTM behaves within the batches/ seq_len?
- Transformer based decoding
Trending Questions
- UIImageView Frame Doesn't Reflect Constraints
- Is it possible to use adb commands to click on a view by finding its ID?
- How to create a new web character symbol recognizable by html/javascript?
- Why isn't my CSS3 animation smooth in Google Chrome (but very smooth on other browsers)?
- Heap Gives Page Fault
- Connect ffmpeg to Visual Studio 2008
- Both Object- and ValueAnimator jumps when Duration is set above API LvL 24
- How to avoid default initialization of objects in std::vector?
- second argument of the command line arguments in a format other than char** argv or char* argv[]
- How to improve efficiency of algorithm which generates next lexicographic permutation?
- Navigating to the another actvity app getting crash in android
- How to read the particular message format in android and store in sqlite database?
- Resetting inventory status after order is cancelled
- Efficiently compute powers of X in SSE/AVX
- Insert into an external database using ajax and php : POST 500 (Internal Server Error)
Popular Questions
- How do I undo the most recent local commits in Git?
- How can I remove a specific item from an array in JavaScript?
- How do I delete a Git branch locally and remotely?
- Find all files containing a specific text (string) on Linux?
- How do I revert a Git repository to a previous commit?
- How do I create an HTML button that acts like a link?
- How do I check out a remote Git branch?
- How do I force "git pull" to overwrite local files?
- How do I list all files of a directory?
- How to check whether a string contains a substring in JavaScript?
- How do I redirect to another webpage?
- How can I iterate over rows in a Pandas DataFrame?
- How do I convert a String to an int in Java?
- Does Python have a string 'contains' substring method?
- How do I check if a string contains a specific word?
It may depend on the specific situation you are dealing with. But in general, the only reason I would do zero padding or any kind of padding to RNN would be to make batch-wise computations work. Also, padding should be done in a way that it doesn't affect the results. So, it should not contribute to computing hidden state computation that you would be using for downstream tasks. For example, you may pad the end of the particular sequences from {t+1:T}, but then for further task or processing we should use only h{0:t}
However, if you are doing anything different than simple RNN (for eg. bidirectional-RNN), it can be complicated to do padding. For example: for the forward direction you would pad in the end and for the reverse direction, you would want to pad the front part of sequences.
Even for batching or doing parallel computations pytorch has
packed sequenceswhich should be faster than padding IMO.