✔ Scroll down and test yourself — answers are hidden under the “View Answer” button.
Attempt all questions first.
✔️ Click SUBMIT at the end to unlock VIEW ANSWER buttons.
Word Sense Disambiguation (WSD) MCQs with Answers and Explanations
Understanding the correct meaning of a word in context is one of the most fundamental challenges in Natural Language Processing (NLP). This collection of multiple-choice questions (MCQs) focuses on Word Sense Disambiguation (WSD), covering core concepts, algorithms, and real-world scenarios. These questions are designed for students, researchers, and professionals preparing for exams, interviews, or strengthening their NLP knowledge.
What is Word Sense Disambiguation (WSD)?
Word Sense Disambiguation (WSD) is the task of identifying the correct meaning (sense) of a word based on its context in a sentence or document. Many words in natural language are ambiguous (polysemous), meaning they have multiple meanings.
For example:
- "bank" → Financial institution OR river edge
- "bat" → Flying mammal OR sports equipment
- "plant" → Industrial facility OR living organism
WSD systems analyze surrounding words, syntax, semantics, and sometimes external knowledge bases to determine the intended meaning.
Why is WSD Important?
- Improves machine translation accuracy
- Enhances search engines and information retrieval
- Supports chatbots and virtual assistants
- Essential for text understanding and AI applications
Common Approaches to WSD
- Knowledge-based methods (e.g., Lesk algorithm, WordNet)
- Supervised learning (trained on labeled datasets)
- Unsupervised methods (Word Sense Induction)
- Neural approaches (contextual embeddings like transformers)
What You Will Learn from These MCQs
- Context-based ambiguity resolution
- Lesk algorithm and gloss overlap
- Most Frequent Sense (MFS) bias
- Graph-based and probabilistic WSD
- Transformer-based contextual understanding
- Domain adaptation and sense granularity issues
Each question includes a detailed explanation to help you clearly understand why a particular sense is correct, making this set ideal for both learning and revision.
📚 Table of Contents
- MCQ 1: Context Window Limitation
- MCQ 2: Lesk Algorithm – Gloss Overlap
- MCQ 3: Most Frequent Sense Bias
- MCQ 4: Graph-Based WSD (PageRank)
- MCQ 5: One Sense per Discourse
- MCQ 6: Word Sense Induction (WSI)
- MCQ 7: Contextual Embeddings (Transformers)
- MCQ 8: Domain Adaptation Problem
- MCQ 9: Gloss-Context Similarity
- MCQ 10: Context-Based Meaning (Plant)
- MCQ 11: Lack of Context Ambiguity
- MCQ 12: Context Clues (Python)
- MCQ 13: Contextual Representation Change
- MCQ 14: Accuracy Calculation
- MCQ 15: Collocations in WSD
- MCQ 16: Sense Granularity Problem
- MCQ 17: Idioms and Multiword Expressions
- MCQ 18: Probabilistic WSD
- MCQ 19: Attention Mechanism in WSD
- MCQ 20: Multilingual WSD
Explanation:
A very small context window may not include important words like “river” that indicate the intended meaning. Without enough surrounding context, the WSD algorithm cannot gather sufficient evidence to correctly determine the sense of the word.
Explanation:
The Lesk algorithm chooses the sense whose dictionary gloss has the largest overlap with the context words. In this sentence, words like “hear” and “concert” relate to music, making the musical sense the most appropriate interpretation.
Explanation:
Many WSD systems learn to predict the sense that appears most frequently in training data. This leads to bias toward the Most Frequent Sense (MFS), causing incorrect predictions when rarer senses appear in different contexts.
Explanation:
Graph-based WSD assumes that the correct sense will have strong semantic connections with other senses appearing in the same context. PageRank helps identify the sense that is most central or influential within this semantic network.
Explanation:
The Yarowsky algorithm relies on the principle "one sense per discourse," which suggests that a word appearing multiple times in the same document usually carries the same meaning throughout that discourse.
Explanation:
Word Sense Induction (WSI) automatically discovers senses by clustering contexts in which a word appears. Unlike traditional WSD, it does not rely on predefined sense inventories such as WordNet.
Explanation:
Transformer models generate contextual embeddings where the representation of a word changes depending on surrounding words. In this case, the presence of "clicked" and "icon" signals the computer-related sense.
Explanation:
Different domains often use words with different senses or frequency distributions. A model trained on news data may not recognize specialized meanings commonly used in biomedical literature.
Explanation:
Gloss-based neural WSD methods encode both the sentence context and dictionary glosses into vector representations. The sense whose gloss embedding is most similar to the context embedding is selected.
Explanation:
Words like "produces" and "turbines" strongly indicate an industrial context. Therefore, the correct interpretation of "plant" here refers to a power or industrial facility rather than vegetation.
Explanation:
The sentence provides no contextual indicators to determine whether "bank" refers to a financial institution or a river bank. Without surrounding words that provide semantic clues, WSD algorithms struggle to determine the intended sense.
Explanation:
Context words such as "spotted" and "forest" indicate a wildlife environment. These contextual cues strongly suggest that the word "python" refers to the snake rather than the programming language.
Explanation:
Contextual language models generate different vector representations for the same word depending on surrounding words. Thus, "book" in a reading context and "book" in a legal context produce different embeddings.
Explanation:
Accuracy is calculated as the number of correct predictions divided by the total number of instances. Here it is 70 / 100 = 0.70, meaning the system correctly disambiguated 70% of the test cases.
Explanation:
The word "code" strongly suggests a programming environment. Collocations with technical terms help WSD systems infer that "bug" refers to a software error rather than an insect.
Explanation:
Sense granularity refers to the mismatch between very fine-grained lexical senses in resources like WordNet and the broader senses actually used in real datasets or applications.
Explanation:
The phrase "kick the bucket" is an idiom meaning "to die." If the system analyzes each word independently instead of recognizing the phrase as a multiword expression, it may incorrectly interpret the literal meaning.
Explanation:
In probabilistic WSD models, the sense with the highest posterior probability given the context is selected. Since 0.7 is greater than 0.3, the river bank sense is chosen.
Explanation:
Attention mechanisms allow transformer models to identify and emphasize the most informative context words. In this case, "river" provides strong evidence that "bank" refers to the river edge sense.
Explanation:
Cross-lingual WSD relies on translations to resolve ambiguity. The Spanish word "murciélago" specifically refers to the flying mammal, eliminating the alternative sense of "bat" as a sports equipment.