Paper: WEBQUESTIONS paper (Semantic Parsing on Freebase from Question-Answer Pairs), authored by Berant et al. (2013)
The WEBQUESTIONS paper (Semantic Parsing on Freebase from Question-Answer Pairs), authored by Berant et al. (2013) and presented at EMNLP, introduced a benchmark dataset and a semantic parsing method for answering open-domain questions using Freebase, a large-scale knowledge base (KB). Below is a structured breakdown of the paper:
1. Key Contributions
Dataset Introduction:
Released WEBQUESTIONS, a dataset of 3,778 question-answer pairs (later expanded to 5,810) collected using Google Suggest API.
Questions are natural language queries (e.g., "Where was Einstein born?"), paired with answers from Freebase.
Task Definition:
Goal: Map natural language questions to Freebase queries (logical forms) to retrieve answers.
Challenge: Freebase contains ~2B facts, making manual query construction infeasible.
Methodology:
Proposed a log-linear semantic parser that learns to map questions to Freebase queries (using lambda calculus) from weak supervision (only QA pairs, not logical forms).
Leveraged feature-based learning (e.g., lexical, syntactic, KB-based features).
2. Semantic Parsing Approach
Pipeline
Candidate Generation:
For each question, generate possible Freebase entities (e.g., "Einstein" →
m.0b3fp9).Use n-gram matching and alias detection (Freebase’s
/type/object/key).
Query Construction:
Build logical forms (lambda-DCS expressions) like:
(lambda x (PlaceOfBirth (Person "Einstein")) → Returns "Ulm, Germany".
Supports compositionality (e.g., "Who directed Inception?" →
(Film Director "Inception")).
Learning & Inference:
Training: Learn feature weights using beam search + perceptron.
Features:
Lexical (question words + Freebase predicates).
Compositional (how predicates combine).
KB-coverage (answer recall).
Answer Extraction:
Execute the top-ranked logical form on Freebase to retrieve answers.
3. Key Innovations
Weak Supervision: Learned from QA pairs alone, without annotated logical forms.
Lambda-DCS: A simplified query language for Freebase, enabling efficient parsing.
Feature Engineering: Combined linguistic and KB-structure features.
4. Results & Impact
Achieved 39.9% F1 on WEBQUESTIONS (vs. 30.4% for IR baselines).
Limitations:
Struggled with complex questions (multi-hop, aggregation).
Dependency on Freebase (deprecated in 2016).
Legacy:
Pioneered KBQA (Knowledge-Based Question Answering).
Inspired later work like GraphQuery (Yih et al., 2015) and neural semantic parsers.
5. Dataset Details
| Statistic | Value |
|---|---|
| # Questions | 3,778 (train) + 2,032 (test) |
| Avg. Question Length | 4.5 words |
| Answer Sources | Freebase (entities, relations) |
| Example Question | "Who founded Microsoft?" → ["Bill Gates", "Paul Allen"] |
6. Comparison with Later Work
| Aspect | WEBQUESTIONS (2013) | Modern KBQA (e.g., SPARQL, BERT-based) |
|---|---|---|
| Supervision | Weak (QA pairs) | Strong (annotated queries) / Zero-shot |
| KB | Freebase | Wikidata, DBPedia, Custom KGs |
| Model | Log-linear parser | Neural models (Transformers, GNNs) |
| Complexity | Single-relation | Multi-hop, temporal, compositional |
7. Code Example (Simplified Lambda-DCS Query)
# Pseudo-code for "Where was Einstein born?" query = { "entity": "Albert_Einstein", "relation": "PlaceOfBirth", "target": "?city" } # Freebase equivalent: (PlaceOfBirth (Person "Albert_Einstein"))
8. Why This Paper Matters
Foundational Work: First large-scale QA dataset tied to Freebase.
Paradigm Shift: Showed semantic parsing could be learned without logical form annotations.
Benchmark: WEBQUESTIONS remains a standard evaluation set for KBQA.
Comments
Post a Comment