Word Sense Disambiguation (WSD) is the task in Natural Language Processing (NLP) of determining which sense (meaning) of a word is used in a given context when the word has multiple meanings (i.e., it is polysemous).
Example:
Consider the word “bank” in the following sentences:
She sat by the bank of the river.
He went to the bank to deposit a check.
Here, “bank” has different meanings:
In (1), it refers to the side of a river.
In (2), it refers to a financial institution.
WSD helps algorithms decide the correct meaning based on surrounding words (context).
Why WSD is Important:
Machine Translation: Translating ambiguous words correctly.
Information Retrieval: Understanding query intent.
Chatbots / Virtual Assistants: Providing accurate responses.
Text Mining / Knowledge Extraction: Linking correct entities or facts.
Common Approaches to WSD:
Knowledge-based Methods:
Use lexical resources like WordNet.
Algorithms: Lesk algorithm, semantic similarity measures.
Supervised Learning:
Uses labeled corpora (sense-annotated text).
Algorithms: Decision Trees, SVMs, Neural Networks.
Limitation: Requires large, annotated datasets.
Unsupervised Learning:
Clusters word occurrences into different senses without labeled data.
Uses context similarity, co-occurrence statistics.
Neural / Deep Learning Methods:
Contextual embeddings (e.g., BERT, ELMo) have significantly improved WSD performance.
Fine-tuning models on WSD tasks gives state-of-the-art results.
Comments are closed.