understanding hpsg

HPSG combines ideas from:

  • Unification-based grammar
  • Lexicalist theories
  • Constraint-based syntax and semantics

It’s known for using typed feature structures and modeling grammar in a way that integrates syntax, semantics, and the lexicon in a highly modular and formal way. Let us see a quick visual schema of how HPSG organizes sentence structure-

```
[ S ]
 |-- HEAD: verb
 |-- SUBJ: < NP >
 |-- COMPS: < NP >
 |-- SYNSEM:
 |     |-- LOCAL:
 |     |     |-- CAT: [HEAD verb, SUBJ <NP>, COMPS <NP>]
 |     |     |-- CONTENT: [Semantic Representation]
 |
 |-- DTRS (Daughters):
       |
       |-- [ NP ]  ← Subject (e.g., "John")
       |
       |-- [ V' ]
             |
             |-- HEAD: verb
             |-- COMPS: < NP >
             |-- DTRS:
                   |-- [ V ]      ← Verb (e.g., "gave")
                   |-- [ NP ]     ← Object (e.g., "a book")
```

1. [ S ] – Sentence node

  • The top-level phrase (S for Sentence).
  • A headed phrase, meaning the verb determines the grammatical behavior.

2. HEAD: verb

  • The head of the sentence is the verb (e.g., “gave”).
  • The head determines the phrase’s syntactic and semantic properties.

3. SUBJ: <NP>

  • Specifies the subject requirement of the verb (like “John”).
  • It is a list of NPs that the verb expects in the SUBJ feature.

4. COMPS: <NP>

  • Specifies the complements required by the verb (like objects).
  • “Gave” requires two NPs (e.g., “a book” and “to Mary”).

5. SYNSEM (Syntactic and Semantic Features)

  • Merges syntax and semantics into a unified structure.
  • Contains:

    • LOCAL:
      • CAT: category features (HEAD, SUBJ, COMPS)
      • CONTENT: semantic representation (meaning of the phrase)

6. DTRS (Daughters)

  • Immediate components of the phrase:

    • NP – the subject (“John”)
    • V' – the verb phrase, which contains:
      • V – the verb (“gave”)
      • NP – the object (“a book”)
  • Follows the Head Feature Principle (explained below).




✅ Key Principles in the Schema

Principle Description
Head Feature Principle The head daughter passes up its features to the mother node.
Valence Principle SUBJ and COMPS lists are discharged as syntactic elements combine.
Feature Structures Every node is an Attribute-Value Matrix (AVM).
Lexicalism Most grammatical information is stored in the lexicon. For example, the verb “gave” encodes that it takes a subject and two complements.



HPSG Tree Diagram with AVMs for “John gave a book”

```
                                 [S]
                +----------------+----------------+
                |                                 |
           [NP] John                      [V'] (Verb Phrase)
                                              |
                                +-------------+-------------+
                                |                           |
                         [V] "gave"                   [NP] "a book"

AVMs for Each Node:

[S]
⎡HEAD   verb             ⎤
⎢SUBJ   ⟨NP⟩              ⎥
⎢COMPS  ⟨NP⟩              ⎥
⎢CONT   give'(x, y)      ⎥
⎣                        ⎦

[NP] → "John"
⎡HEAD   noun             ⎤
⎢INDEX  j                ⎥
⎢CONT   john'            ⎥
⎣                        ⎦

[V']
⎡HEAD   verb             ⎤
⎢COMPS  ⟨NP⟩              ⎥
⎢CONT   λx. give'(j, x)  ⎥
⎣                        ⎦

[V] → "gave"
⎡HEAD   verb             ⎤
⎢SUBJ   ⟨NP⟩              ⎥
⎢COMPS  ⟨NP⟩              ⎥
⎢CONT   λxλy. give'(y, x)⎥
⎣                        ⎦

[NP] → "a book"
⎡HEAD   noun             ⎤
⎢INDEX  b                ⎥
⎢CONT   book'(b)         ⎥
⎣                        ⎦

```
  • The verb “gave” expects a subject (SUBJ ⟨NP⟩) and a complement (COMPS ⟨NP⟩), both noun phrases.
  • The semantics (CONT) is built through lambda abstraction.
  • “John” is the subject, indexed as j, and “a book” is the object, indexed as b.
  • The semantic content of the sentence becomes give’(john’, book’).



Definitions

1. Unification-based Grammar

🔍 Idea: Grammar rules operate by matching and merging feature structures (attribute-value pairs). This merging is called unification.
🧠 Analogy: Think of two puzzle pieces (feature structures). You can join them (unify) if they match—but if there’s a mismatch (e.g., one says “singular”, the other says “plural”), unification fails.
📝 Example: Let’s look at subject-verb agreement: “She walks.” ✅ || “She walk.” ❌

```
[NP]
⎡ PER 3rd ⎤
⎢ NUM sg  ⎥
⎣ GEN f   ⎦

[V]
⎡ HEAD verb ⎤
⎢ AGR ⎡ PER 3rd ⎤
⎢     ⎢ NUM sg  ⎥
⎣     ⎣ GEN f   ⎦
```

✅ These unify! So the sentence is grammatical. In “She walk”, the verb’s AGR is plural (say: NUM pl), which fails to unify with the subject (NUM sg). So the structure is rejected.



2. Lexicalist Theories

🔍 Idea: Lexicon is rich. Words carry most of the syntactic and semantic information. Instead of complex grammar rules, each word specifies what it needs (like arguments, tense, etc.).
📝 Example: For the verb “give”:

```
Lexical Entry for "give":

⎡ HEAD   verb                   ⎤
⎢ SUBJ   <NP>                   ⎥
⎢ COMPS  <NP, NP>               ⎥
⎢ SEM    λxλyλz. give'(z, y, x) ⎥
⎣                               ⎦
```

“give” requires: (1) subject (e.g., “John”) and (2) complements (e.g., “a book”, “to Mary”). So:

  • “John gave a book to Mary” ✅ (all arguments present)
  • “John gave” ❌ (incomplete)

In HPSG, this rich lexical information drives the sentence structure, hence: Lexicalist.



3. Constraint-based Syntax and Semantics

🔍 Idea: Rather than transformational rules (like in Chomsky’s theories), HPSG uses constraints that must be satisfied. These constraints apply to feature structures. There are no derivations, only well-formedness constraints on structure.

📝 Example: You want to form a head-complement phrase like: “read a book”

  • The phrase must satisfy constraints like:
    • The verb (head) must specify it takes one NP as complement.
    • The complement NP must match what the verb expects (e.g., a proper noun phrase).
    • The semantic roles must link correctly: read’(x, y) where x is the subject, y is the object.

In HPSG, a Head-Complement Schema might look like:

```
[SYN] ⎡ HEAD verb ⎤
      ⎢ COMPS < > ⎥ ← Empty after all complements satisfied
      ⎣           ⎦

DTRs (daughters):
- Head Daughter: [HEAD verb, COMPS <NP>]
- Non-head Daughter: [NP]
```


The constraint says: “You can combine a verb phrase with an NP if that NP satisfies the verb’s COMPS list”. This is declarative (constraint-based), not procedural.



Summary Table


Concept What it Means HPSG’s Implementation
Unification-based Grammar Merge of feature structures Grammar rules succeed only if features match
Lexicalist Theories Lexicon is rich with syntactic/semantic info Words define argument structure, not grammar rules
Constraint-based Syntax/Sem Well-formedness defined via feature constraints Syntactic rules are constraints, not transformations