understanding hpsg
HPSG combines ideas from:
- Unification-based grammar
- Lexicalist theories
- Constraint-based syntax and semantics
It’s known for using typed feature structures and modeling grammar in a way that integrates syntax, semantics, and the lexicon in a highly modular and formal way. Let us see a quick visual schema of how HPSG organizes sentence structure-
```
[ S ]
|-- HEAD: verb
|-- SUBJ: < NP >
|-- COMPS: < NP >
|-- SYNSEM:
| |-- LOCAL:
| | |-- CAT: [HEAD verb, SUBJ <NP>, COMPS <NP>]
| | |-- CONTENT: [Semantic Representation]
|
|-- DTRS (Daughters):
|
|-- [ NP ] ← Subject (e.g., "John")
|
|-- [ V' ]
|
|-- HEAD: verb
|-- COMPS: < NP >
|-- DTRS:
|-- [ V ] ← Verb (e.g., "gave")
|-- [ NP ] ← Object (e.g., "a book")
```
1. [ S ]
– Sentence node
- The top-level phrase (S for Sentence).
- A headed phrase, meaning the verb determines the grammatical behavior.
2. HEAD: verb
- The head of the sentence is the verb (e.g., “gave”).
- The head determines the phrase’s syntactic and semantic properties.
3. SUBJ: <NP>
- Specifies the subject requirement of the verb (like “John”).
- It is a list of NPs that the verb expects in the SUBJ feature.
4. COMPS: <NP>
- Specifies the complements required by the verb (like objects).
- “Gave” requires two NPs (e.g., “a book” and “to Mary”).
5. SYNSEM
(Syntactic and Semantic Features)
- Merges syntax and semantics into a unified structure.
-
Contains:
- LOCAL:
-
CAT
: category features (HEAD, SUBJ, COMPS) -
CONTENT
: semantic representation (meaning of the phrase)
-
- LOCAL:
6. DTRS
(Daughters)
-
Immediate components of the phrase:
-
NP
– the subject (“John”) -
V'
– the verb phrase, which contains:-
V
– the verb (“gave”) -
NP
– the object (“a book”)
-
-
-
Follows the Head Feature Principle (explained below).
✅ Key Principles in the Schema
Principle | Description |
---|---|
Head Feature Principle | The head daughter passes up its features to the mother node. |
Valence Principle | SUBJ and COMPS lists are discharged as syntactic elements combine. |
Feature Structures | Every node is an Attribute-Value Matrix (AVM). |
Lexicalism | Most grammatical information is stored in the lexicon. For example, the verb “gave” encodes that it takes a subject and two complements. |
HPSG Tree Diagram with AVMs for “John gave a book”
```
[S]
+----------------+----------------+
| |
[NP] John [V'] (Verb Phrase)
|
+-------------+-------------+
| |
[V] "gave" [NP] "a book"
AVMs for Each Node:
[S]
⎡HEAD verb ⎤
⎢SUBJ ⟨NP⟩ ⎥
⎢COMPS ⟨NP⟩ ⎥
⎢CONT give'(x, y) ⎥
⎣ ⎦
[NP] → "John"
⎡HEAD noun ⎤
⎢INDEX j ⎥
⎢CONT john' ⎥
⎣ ⎦
[V']
⎡HEAD verb ⎤
⎢COMPS ⟨NP⟩ ⎥
⎢CONT λx. give'(j, x) ⎥
⎣ ⎦
[V] → "gave"
⎡HEAD verb ⎤
⎢SUBJ ⟨NP⟩ ⎥
⎢COMPS ⟨NP⟩ ⎥
⎢CONT λxλy. give'(y, x)⎥
⎣ ⎦
[NP] → "a book"
⎡HEAD noun ⎤
⎢INDEX b ⎥
⎢CONT book'(b) ⎥
⎣ ⎦
```
- The verb “gave” expects a subject (SUBJ ⟨NP⟩) and a complement (COMPS ⟨NP⟩), both noun phrases.
- The semantics (CONT) is built through lambda abstraction.
- “John” is the subject, indexed as j, and “a book” is the object, indexed as b.
- The semantic content of the sentence becomes give’(john’, book’).
Definitions
1. Unification-based Grammar
🔍 Idea: Grammar rules operate by matching and merging feature structures (attribute-value pairs). This merging is called unification.
🧠 Analogy: Think of two puzzle pieces (feature structures). You can join them (unify) if they match—but if there’s a mismatch (e.g., one says “singular”, the other says “plural”), unification fails.
📝 Example: Let’s look at subject-verb agreement: “She walks.” ✅ || “She walk.” ❌
```
[NP]
⎡ PER 3rd ⎤
⎢ NUM sg ⎥
⎣ GEN f ⎦
[V]
⎡ HEAD verb ⎤
⎢ AGR ⎡ PER 3rd ⎤
⎢ ⎢ NUM sg ⎥
⎣ ⎣ GEN f ⎦
```
✅ These unify! So the sentence is grammatical. In “She walk”, the verb’s AGR is plural (say: NUM pl), which fails to unify with the subject (NUM sg). So the structure is rejected.
2. Lexicalist Theories
🔍 Idea: Lexicon is rich. Words carry most of the syntactic and semantic information. Instead of complex grammar rules, each word specifies what it needs (like arguments, tense, etc.).
📝 Example: For the verb “give”:
```
Lexical Entry for "give":
⎡ HEAD verb ⎤
⎢ SUBJ <NP> ⎥
⎢ COMPS <NP, NP> ⎥
⎢ SEM λxλyλz. give'(z, y, x) ⎥
⎣ ⎦
```
“give” requires: (1) subject (e.g., “John”) and (2) complements (e.g., “a book”, “to Mary”). So:
- “John gave a book to Mary” ✅ (all arguments present)
- “John gave” ❌ (incomplete)
In HPSG, this rich lexical information drives the sentence structure, hence: Lexicalist.
3. Constraint-based Syntax and Semantics
🔍 Idea: Rather than transformational rules (like in Chomsky’s theories), HPSG uses constraints that must be satisfied. These constraints apply to feature structures. There are no derivations, only well-formedness constraints on structure.
📝 Example: You want to form a head-complement phrase like: “read a book”
- The phrase must satisfy constraints like:
- The verb (head) must specify it takes one NP as complement.
- The complement NP must match what the verb expects (e.g., a proper noun phrase).
- The semantic roles must link correctly: read’(x, y) where x is the subject, y is the object.
In HPSG, a Head-Complement Schema might look like:
```
[SYN] ⎡ HEAD verb ⎤
⎢ COMPS < > ⎥ ← Empty after all complements satisfied
⎣ ⎦
DTRs (daughters):
- Head Daughter: [HEAD verb, COMPS <NP>]
- Non-head Daughter: [NP]
```
The constraint says: “You can combine a verb phrase with an NP if that NP satisfies the verb’s COMPS list”. This is declarative (constraint-based), not procedural.
Summary Table
Concept | What it Means | HPSG’s Implementation |
---|---|---|
Unification-based Grammar | Merge of feature structures | Grammar rules succeed only if features match |
Lexicalist Theories | Lexicon is rich with syntactic/semantic info | Words define argument structure, not grammar rules |
Constraint-based Syntax/Sem | Well-formedness defined via feature constraints | Syntactic rules are constraints, not transformations |