I understand that the model uses previously trained Part of Speech tagging during its configuration stage. But what if most of the words are new, how would the parser decide its operation then?
How does a Transition-based Dependency parser decide which operation to do next in its configuration stage?
190 Views Asked by Akash At
2
There are 2 best solutions below
Related Questions in NLP
- Seeking Python Libraries for Removing Extraneous Characters and Spaces in Text
- Clarification on T5 Model Pre-training Objective and Denoising Process
- The training accuracy and the validation accuracy curves are almost parallel to each other. Is the model overfitting?
- Give Bert an input and ask him to predict. In this input, can Bert apply the first word prediction result to all subsequent predictions?
- Output of Cosine Similarity is not as expected
- Getting an error while using the open ai api to summarize news atricles
- SpanRuler on Retokenized tokens links back to original token text, not the token text with a split (space) introduced
- Should I use beam search on validation phase?
- Dialogflow failing to dectect the correct intent
- How to detect if two sentences are simmilar, not in meaning, but in syllables/words?
- Is BertForSequenceClassification using the CLS vector?
- Issue with memory when using spacy_universal_sentence_encoder for similarity detection
- Why does the Cloud Natural Language Model API return so many NULLs?
- Is there any OCR or technique that can recognize/identify radio buttons printed out in the form of pdf document?
- Model, lexicon to do fine grained emotions analysis on text in r
Related Questions in STANFORD-NLP
- Why are SST-2 and CoLA commonly used datasets for debiasing?
- How can I correctly change the upos of words in a sentence using Stanza?
- I wanted to evaluate and see the performance of Spider 1.0 dataset using llama-2-7B model, hugging g=face transformer, not working, how to fix it?
- Facing error to evaluate spider 1.0 dataset using orca-2-7B model, hugging face transformers
- java.lang.IllegalArgumentException using Stanford Parser and Jetpack Compose
- Displaying a graph for parsed sentences with Stanford-parser
- Displaying parser tree using Jetpack Compose
- sentences to clauses with Python
- Stanford Stanza sometimes splits a sentence into two sentences
- GloVe Nearest neighbors (NLP)
- How to use local files in an Azure Function hosted on the Linux Consumption plan?
- Sentences Annotation Class giving null value using Stanford Core NLP using c#
- How to make stanza lemmatizer to return just the lemma instead of a dictionary?
- GloVe algorithm: reading the coccurence.bin file contents in Python
- Stanford CoreNLP library doesn't tokenize new lines
Related Questions in DEPENDENCY-PARSING
- Are there method that can extract interaction between person in text?
- List of dependencies in Spacy
- Dependency Parsing in Spacy
- How can I parse the action that belongs to a Person using Spacy in Python?
- Is possible to get dependency/pos information for entities in Spacy?
- SpaCy Matcher - Restricting Potential Matches
- How many of "tags" and "predicted_dependencies" in the dependency parsing (allennlp.predictors.predictor)?
- Run dependency parser on pre-initialized doc object of spacy
- How to find whether a sentence contain a noun using spacy?
- Pytorch save/load model with lower dev set accuracy?
- Add known matches to Spacy document with character offsets
- Extract subtree (Phrase) for nsubj elements using spacy
- Issue in creating Semgrex patterns with relation names containing ":" colon
- Unable to use Allennlp biaffine parser model
- Stanford Stanza -- Dependency Parsing Module -- Output for document with more than one sentence
Trending Questions
- UIImageView Frame Doesn't Reflect Constraints
- Is it possible to use adb commands to click on a view by finding its ID?
- How to create a new web character symbol recognizable by html/javascript?
- Why isn't my CSS3 animation smooth in Google Chrome (but very smooth on other browsers)?
- Heap Gives Page Fault
- Connect ffmpeg to Visual Studio 2008
- Both Object- and ValueAnimator jumps when Duration is set above API LvL 24
- How to avoid default initialization of objects in std::vector?
- second argument of the command line arguments in a format other than char** argv or char* argv[]
- How to improve efficiency of algorithm which generates next lexicographic permutation?
- Navigating to the another actvity app getting crash in android
- How to read the particular message format in android and store in sqlite database?
- Resetting inventory status after order is cancelled
- Efficiently compute powers of X in SSE/AVX
- Insert into an external database using ajax and php : POST 500 (Internal Server Error)
Popular # Hahtags
Popular Questions
- How do I undo the most recent local commits in Git?
- How can I remove a specific item from an array in JavaScript?
- How do I delete a Git branch locally and remotely?
- Find all files containing a specific text (string) on Linux?
- How do I revert a Git repository to a previous commit?
- How do I create an HTML button that acts like a link?
- How do I check out a remote Git branch?
- How do I force "git pull" to overwrite local files?
- How do I list all files of a directory?
- How to check whether a string contains a substring in JavaScript?
- How do I redirect to another webpage?
- How can I iterate over rows in a Pandas DataFrame?
- How do I convert a String to an int in Java?
- Does Python have a string 'contains' substring method?
- How do I check if a string contains a specific word?
I'd like to flesh @Quantum's answer out into a detailed one as follows:
Before 2014 many parsers were depending on a manually designed set of feature templates, and such methods have two drawbacks: 1) they required a lot of expertise and are usually incomplete; 2) most of the runtime is consumed by the feature extraction part of the configuration stage. After Chen and Mannning published their paper, A Fast and Accurate Dependency Parser using Neural Networks, almost all parsers are relying on neural networks.
Let's see how Chen and Manning did the job.
As illustrated in the above diagram, the output of the neural network is a distribution after a softmax function, then it is a simple classification problem depending on some given information. The given information contains mainly three parts: the top 3 words on the stack and buffer, and the two leftmost/rightmost children of the top two words on the stack, and the leftmost and rightmost grandchildren; the POS tags of the above; and the arc labels of all children/grandchildren.
The inputs are embedded into a matrix and transformed by two matrices(and as shown in the picture a cube function) to become the logits and then the distribution of three elements atop of the network.
HTH :)
References: 1) A Fast and Accurate Dependency Parser using Neural Networks, 2) CMU Neural Nets for NLP 2017 (12): Transition-based Dependency Parsing