Science topic

Word Sense Disambiguation - Science topic

Explore the latest questions and answers in Word Sense Disambiguation, and find Word Sense Disambiguation experts.
Questions related to Word Sense Disambiguation
  • asked a question related to Word Sense Disambiguation
Question
3 answers
May I know, What are different algorithms/techniques for word sense disambiguation using knowledge based approach?
Also any resources where I can find the papers or data for my research?
Kindly help
Relevant answer
Answer
Dear Roopa,
I highly recommend you look at Bidirectional Encoder Representations from Transformers, otherwise known as BERT. Jacob Devlin at Google pretrained this algorithm in 2018, and it was a total game changer. Now a days, almost all papers are using at least this framework if not adding deep learning tools to it. The paper was cited >35,000 times in only a few years...
Please also take a look at this review published this year that discusses the recent developments in the field:
All the best,
Nathan
  • asked a question related to Word Sense Disambiguation
Question
7 answers
Hi , This is related to my final year project for bachelors degree. I was wondering if Reinforcement learning models can be applied to Word sense Disambiguation and that if it will give a better performance than existing methods? Thank you
Relevant answer
Answer
Reinforcement learning provide best or optimize path but supervise method provide only a method or path .
  • asked a question related to Word Sense Disambiguation
Question
5 answers
I did a word sense disambiguation project and now I need to calculate f-measure. So I have to calculate tp, tn, fp and fn. I calculated them but the problem is the fp in the first class equals to the fn in the other class and so when calculating micro average precision and recall, they have exactly same result and this is wrong. Because they should be different. How should I change the piece of code below to calculate fp and fn correctly. (this code is for a two sense wsd. t[1] is the tag (sense) of the sentence, argmax is the tag that the code has identified for the sentence).
the piece of code to calculte tp, tn, fp and fn:
if t[1] == clas1 and argmax == clas1:
tp1 += 1
elif t[1] == clas1 and argmax != clas1:
fn1 += 1
elif t[1] != clas1 and argmax == clas1:
fp1 += 1
elif t[1] != clas1 and argmax != clas1:
tn1 += 1
if t[1] == clas2 and argmax == clas2:
tp2 += 1
elif t[1] == clas2 and argmax != clas2:
fn2 += 1
elif t[1] != clas2 and argmax == clas2:
fp2 += 1
elif t[1] != clas2 and argmax != clas2:
tn2 += 1
The result:
tp1 176 tn 112 fp 49 fn 24
tp2 112 tn 176 fp 24 fn 49
MacroAvrgPrecision: 0.802875816993464
MacroAvrgRecall: 0.7878260869565217
MacroF-measure: 0.7952797587571994
true: 288 false: 73
accuracy: 79.77839335180056
weightedAvrPrecision: 0.7977839335180056
weightedAvrRecall: 0.7977839335180056
weightedF-measure: 0.7977839335180056
microAvrPrecision: 0.7977839335180056
microAvrRecall: 0.7977839335180056
microF_measure: 0.7977839335180056
Relevant answer
Answer
Bahadorreza,
Considering the sentence: " In binary classification, you will get fp_classX = fn_classY anyways. ", the problem is exactly here. Because for calculating micro average precision the formula is:
microAvrPrecision = (tp1+tp2)/(tp1+tp2+fp1+fp2)
and for calculating micro average recall the formula is:
microAvrRecall = (tp1+tp2)/(tp1+tp2+fn1+fn2)
So when fp in classX equals fn classY, naturally the results are same. Is this right and scientific? (I have 40 ambiguous words with same micro average precision and recall, and it doesn't seem correct).
  • asked a question related to Word Sense Disambiguation
Question
4 answers
For word sense disambiguation, there are supervised and unsupervised methods. The supervised method contains, statistical method, exemplar based methods and rule based methods. I know support vector machine is a supervised method, but which class of that? Is it statistical? (The answer should have a reference). Thanks.
Relevant answer
Answer
Thanks for your answer. Would you please write the reference of that, too. I have to refer to it in my thesis. I googled, but I couldn't find a reference.
  • asked a question related to Word Sense Disambiguation
Question
13 answers
I want to do WSD as part of my research. My dataset is not larg so I prefer to use a good tool instead of implementing a WSD algorithm. 
Relevant answer
Answer
You can use BabelFy, http://babelfy.org/ a tool for word sense disambiguation and entity linking. It's based on BabelNet which includes WordNet
  • asked a question related to Word Sense Disambiguation
Question
2 answers
speech perception tests such as Word and Non-word repetition tasks
Relevant answer
Answer
Its a nice question and actually a doubt comes to a researcher when it comes to the concept of 'test retest reliability". I feel to measure reliability you need to have the tests repeated soon after a first one and preferably within a span of two weeks or so. To make sure that it covers the inter session and intra session reliability. Also not giving much of the time for the individual to adjust for the test patterns. In one of my research I have repeated the second recording after half an hour of first recording and the third recording after a span of 6 days of first recording. Without allowing any training within these 6 days.
  • asked a question related to Word Sense Disambiguation
Question
6 answers
I would like to identify certain processes in my study and what leads up to each one. I would like to extract this information from email messages. In order to automate the process, I am using regular expression matching but would like to explore a more robust approach .
I think identifying sample sentences and trying to match its semantics with sentences in email messages would be a good start.
I have been through and ran sample java code for stemming, pos, lemmatizing and similarity. However I am now looking for some basic code which integrates all this
approaches in one program/project. I am hoping it also includes tokenisation and word sense disambiguation as well. Importantly it should have an approach of calculating sentence similarity as well. I am sure some one has done this.
I am particularly looking for java, standford core nlp and wordnet approach which integrates tokenising, stemming, pos, lemmatizing, word sense disambiguation and an approach of calculating sentence similarity as well
Relevant answer
Answer
  • asked a question related to Word Sense Disambiguation
Question
3 answers
I'm looking for interesting/original applications of CRFs in NLP/text-mining that would not be PoS Tagging or Named Entity Recognition. Things like reconstruction the structure of a document, word sense disambiguation or pattern recognition for instance, but that can be something else.
Do you know of some?
Relevant answer
Answer
Hi,
You may be interested in a lexicographic focus for CRF. For instance, spotting definitions in text (papers, text books, newswire, etc.). Or, given a definition, using CRF for identifying the hypernym, which then can be useful for automatic glossary creation, or learning semantic networks such as WordNet. I have published a couple of papers on this matter: http://link.springer.com/chapter/10.1007/978-3-319-18111-0_28 and http://lml.bas.bg/ranlp2015/docs/RANLP_main.pdf (Weakly Supervised Definition Extraction). There's an EMNLP 2013 papers that also works in this direction: http://www.aclweb.org/old_anthology/D/D13/D13-1073.pdf
Hope this helps!
  • asked a question related to Word Sense Disambiguation
Question
3 answers
My motivation is to somehow (blindly) learn (negative) patterns with plain text corpora.
For a bag of words {this, is, a, book}, once a corpus tells us there is no usage of "book is this a" for sure, and so on so forth, then hopefully by negation one may find some hidden rules to promote the bag-of-words model to something similar to LDA.
Relevant answer
Answer
If I understand your question in the right way, then it is more about the syntactical possibilities given by a language, especially an SVO one. If that is the case, then you could try to define the term via the function the specific wordcluster fills i.e. "determiner" or "premodification". You can find more information on this terminology in Quirk 1985 - 'A comprehensive grammar of the English Language'
  • asked a question related to Word Sense Disambiguation
Question
3 answers
As above.
Relevant answer
Answer
There is a wrapper program available from CPAN or Sourceforge called "SenseClusters" which does text disambiguation where words have the same form but different meanings.  In order to use it with Arabic, you may have to convert your text to a roman alphabet.
  • asked a question related to Word Sense Disambiguation
Question
9 answers
I am familiar with TLG and the Perseus Digital Project.  I want to do corpus linguistics on Hellenistic Greek.  Some of the things I need to do is search by POS, search by Lemma, search by morphological element (reduplication, particular morpheme, stem formation, etc.) and search for collocates. 
I am not sure either of the above will do all of that. I am considering developing my own corpora and using a tagger that does all of this to the corpora, as well as a search engine that will recognize what I tagged. 
Do I need to do this, or is there already a selection of tools that will get the job done?
Relevant answer
Answer
Do you know the PROIEL treebank? http://foni.uio.no:3000/ It has a considerable amount of Greek text from several periods (the core is the New Testament, but there is also Herodotus and some Byzantine chronicles). You can download fully lemmatised and tagged texts there and use them to train a morphological tagger. 
  • asked a question related to Word Sense Disambiguation
Question
4 answers
I'm working on a case in which I need a WSD algorithm to disambiguate all the words of a sentence for a step.
I want to know if there is any tool which can help to make it's implementation faster.
Of course if it'll be in Java it would be much more better.
Relevant answer
Answer
Yeah, I've read about BabelNet too. It works on a graph-based model for evaluating local and global-context measures for  WSD
  • asked a question related to Word Sense Disambiguation
Question
1 answer
Pagerank works on a graph which only has direct edges between vertices and there is no weight on edges of the graph but random walk's graph edges are not directed and they have weight. Then, how we should change pagerank which could be applied to these kinds of graphs.
Relevant answer
Answer
Use two things:
1) undirected edges are equivalent to edges in both directions
2) scale the probability to surf from a node to another according to edge weight
Here is a paper that uses pagerank on undirected, weighted graphs:
cf: Section on "continuous LexRank", and LexRank is a PageRank variant for undirected graphs.