I am building a chat-bot with a sequence to sequence encoder decoder model as in NMT. From the data given I can understand that when training they feed the decoder outputs into the decoder inputs along with the encoder cell states. I cannot figure out that when i am actually deploying a chatbot in real time, how what should I input into the decoder since that time is the output that i have to predict. Can someone help me out with this please?
Seq2Seq Models for Chatbots
651 Views Asked by Subham Mukherjee At
1
There are 1 best solutions below
Related Questions in TENSORFLOW
- (Tensorflow)Does the op assign change the gradient computation?
- Tensorflow Windows Accessing Folders Denied:"NewRandomAccessFile failed to Create/Open: Access is denied. ; Input/output error"
- Android App TensorFlow Google Cloud ML
- Convert Tensorflow model to Caffe model
- Google Tensorflow LSTMCell Variables Mapping to Hochreiter97_lstm.pdf paper
- additive Gaussian noise in Tensorflow
- TFlearn evaluate method results meaning
- Regularization losses Tensorflow - TRAINABLE_VARIABLES to Tensor Array
- feed picture to model tensorflow for training
- Fail to read the new format of tensorflow checkpoint?
- I got a error when running a github project in tensorflow
- Tensorflow R0.12 softmax_cross_entropy_with_logits ASSERT Error
- RuntimeError in run_one_batch of TensorFlowDataFrame in tensorflow
- Same output in neural network for each input after training
- ConvNet : Validation Loss not strongly decreasing but accuracy is improving
Related Questions in MACHINE-LEARNING
- How to cluster a set of strings?
- Enforcing that inputs sum to 1 and are contained in the unit interval in scikit-learn
- scikit-learn preperation
- Spark MLLib How to ignore features when training a classifier
- Increasing the efficiency of equipment using Amazon Machine Learning
- How to interpret scikit's learn confusion matrix and classification report?
- Amazon Machine Learning for sentiment analysis
- What Machine Learning algorithm would be appropriate?
- LDA generated topics
- Spectral clustering with Similarity matrix constructed by jaccard coefficient
- Speeding up Viterbi execution
- Memory Error with Classifier fit and partial_fit
- How to find algo type(regression,classification) in Caret in R for all algos at once?
- Difference between weka tool's correlation coefficient and scikit learn's coefficient of determination score
- What are the approaches to the Big-Data problems?
Related Questions in CHATBOT
- Android chatbot Gui link with Program Ab AIML interpreter
- How to classify a sentence based on a few templates?
- Make Kik Bot listen on port 8080 without ngrok
- Design custom entities for api.ai
- After I've built a conversation flow in a chatbot, how'd I get the chatbot to actually perform the desired actions?
- How to send attachment in slack bot using SlackClient for node.js
- Two responses coming when making a single ajax request to wit.ai
- Bot, microsoft framework and Self-signed certificate
- c# save dictionary with <string,List<string>> in .dat with Binaryformatter (or something else)
- mIRC chat bot doesn't acknowledge commands
- How do I connect my database to API.AI?
- How to stream .aac url to wit.ai speech api
- Save chat history from watson conversation service to Cloudant NoSQL DB
- Chat Application Using PHP updating previous messages
- cant send a Generic Template from messenger api (facebook)
Related Questions in MACHINE-TRANSLATION
- Is Speech-to-Text-to-Translation an Impossible Dream?
- Translate value before sending information
- State of the art language translation toolkit
- How to speedup mkcls step in mgiza++ or giza++, it taking up lots of time for word clustering?
- Tensorflow seq2seq `feed_previous' argument`
- How to differentiate between real improvement and random noise?
- What is the difference between ensembling and averaging models?
- What is the difference between mteval-v13a.pl and NLTK BLEU?
- finetuning tensorflow seq2seq model
- What do the counts in giza++ phrase-table mean?
- How to deal with punctuations in machine translation
- Aligning Parallel Sentences for Languages without a Written Form
- Sockeye WMT German to English news translation stopping criterium
- Why special characters like () "" : [] are often removed from data before training translation machine?
- Text language translation in android
Related Questions in SEQUENCE-TO-SEQUENCE
- Extracting attention matrix with TensorFlow's seq2seq example code during decoding
- Do I need the Sequence Length parameter for RNNCell in Tensorflow
- Seq2seq pytorch Inference slow
- What is the difference between ensembling and averaging models?
- Multiple issues with axes while implementing a Seq2Seq with attention in CNTK
- Is adding a FC Linear layer on top of seq2seq architecture a potential source of data leaking from future to past?
- Merging sequence embedding with Time Series Features
- Workaround / fallback value for tfp.distributions.Categorical.log_prob in tensorflow graph mode
- Loss function negative log likelihood giving loss despite perfect accuracy
- many to many sequence prediction variable length input/output inkeras
- LSTM to GRU sampling model issue
- Tensorflow RNN: how to infer a sequence without duplicates?
- How to modify the Tensorflow Sequence2Sequence model to implement Bidirectional LSTM rather than Unidirectional one?
- Tensor flow continuous text sequence-to-sequence. Why batch?
- Seq2seq multiple input features (Passing multiple word/word tokens as input)
Trending Questions
- UIImageView Frame Doesn't Reflect Constraints
- Is it possible to use adb commands to click on a view by finding its ID?
- How to create a new web character symbol recognizable by html/javascript?
- Why isn't my CSS3 animation smooth in Google Chrome (but very smooth on other browsers)?
- Heap Gives Page Fault
- Connect ffmpeg to Visual Studio 2008
- Both Object- and ValueAnimator jumps when Duration is set above API LvL 24
- How to avoid default initialization of objects in std::vector?
- second argument of the command line arguments in a format other than char** argv or char* argv[]
- How to improve efficiency of algorithm which generates next lexicographic permutation?
- Navigating to the another actvity app getting crash in android
- How to read the particular message format in android and store in sqlite database?
- Resetting inventory status after order is cancelled
- Efficiently compute powers of X in SSE/AVX
- Insert into an external database using ajax and php : POST 500 (Internal Server Error)
Popular Questions
- How do I undo the most recent local commits in Git?
- How can I remove a specific item from an array in JavaScript?
- How do I delete a Git branch locally and remotely?
- Find all files containing a specific text (string) on Linux?
- How do I revert a Git repository to a previous commit?
- How do I create an HTML button that acts like a link?
- How do I check out a remote Git branch?
- How do I force "git pull" to overwrite local files?
- How do I list all files of a directory?
- How to check whether a string contains a substring in JavaScript?
- How do I redirect to another webpage?
- How can I iterate over rows in a Pandas DataFrame?
- How do I convert a String to an int in Java?
- Does Python have a string 'contains' substring method?
- How do I check if a string contains a specific word?
The exact answer depends on which building blocks you take from Neural Machine Translation model (NMT) and which ones you would replace with your own. I assume the graph structure exactly as in NMT.
If so, at inference time, you can feed just a vector of zeros to the decoder.
Internal details: NMT uses the entity called
Helperto determine the next input in the decoder (seetf.contrib.seq2seq.Helperdocumentation).In particular,
tf.contrib.seq2seq.BasicDecoderrelies solely on helper when it performs a step: thenext_inputsthat the are fed in to the subsequent cell is exactly the return value ofHelper.next_inputs().There are different implementations of
Helperinterface, e.g.,tf.contrib.seq2seq.TrainingHelperis returning the next decoder input (which is usually ground truth). This helper is used in training as indicated in the tutorial.tf.contrib.seq2seq.GreedyEmbeddingHelperdiscards the inputs, and returns theargmaxsampled token from the previous output. NMT uses this helper in inference whensampling_temperaturehyper-parameter is 0.tf.contrib.seq2seq.SampleEmbeddingHelperdoes the same, but samples the token according to categorical (a.k.a. generalized Bernoulli) distribution. NMT uses this helper in inference whensampling_temperature > 0.The code is in
BaseModel._build_decodermethod. Note that bothGreedyEmbeddingHelperandSampleEmbeddingHelperdon't care what the decoder input is. So in fact you can feed anything, but the zero tensor is the standard choice.