Semantic Parsing: transforming sentences to logical forms using ...

26 downloads 73 Views 181KB Size Report
1. COE'07. 1. Semantic Parsing: transforming sentences to logical forms using machine learning models. Minh Le Nguyen. School of Information Science.
Syntactic and Semantic Natural Language Learning

Semantic Parsing: transforming sentences to logical forms using machine learning models

• Most computational research in natural-language learning has addressed “low-level” syntactic processing. – – – –

Minh Le Nguyen School of Information Science Japan Advanced Institute of Science and Technology

Morphology (e.g. past-tense generation) Part-of-speech tagging Chunking Syntactic parsing

• Learning for semantic analysis has been restricted to relatively “shallow” meaning representations. – Word sense disambiguation (e.g. SENSEVAL) – Semantic role assignment (determining agent, patient, instrument, etc., e.g. FrameNet, PropBank) – Information extraction

COE’07

1

COE’07

2

Semantic Parsing

CLang: RoboCup Coach Language

• Semantic parsing is the process of mapping a naturallanguage sentence to a complete, detailed semantic representation: logical form or meaning representation (MR).

• In RoboCup Coach competition teams compete to coach simulated players • The coaching instructions are given in a formal language called CLang

• For many applications, the desired output is immediately executable by another program. • Application domains:

If the ball is in our penalty area, then all our players except player 4 should stay in our half.

Coach

– CLang: RoboCup Coach Language – GeoQuery: A Database Query Application – Legal domain

Simulated soccer field

Semantic Parsing CLang

((bpos (penalty-area our)) (do (player-except our{4}) (pos (half our)))

COE’07

3

4

Approach

GeoQuery: A Database Query Application

• Applying ML to the transforming problem • Motivations

• Query application for U.S. geography database containing about 800 facts [Zelle & Mooney, 1996]

– Robustness, reduction of development rules – Treating ambiguity – Handling with the difficulty of consistent rules

How many cities are there in the US?

User

COE’07

• Current work – ML for query database language / robocup controlled language – ML for legal domain

Semantic Parsing Query answer(A,

count(B, (city(B), loc(B, C), const(C, countryid(USA))),A))

COE’07

5

COE’07

6

1

Machine Learning Framework Learning phase

Semantic Tagging Transforming phase

NL sentence Structured ML Learner

Pos Tagging: Using our FlexCRF toolkit

Preprocessing

Semantic Tagging

our

player

PRP

NN

our

player

2

has

the

ball

CD

VP

DT

NN

unum

bowner

null

null

Training Examples

Semantic Tagging

Rules with weight Logical form

Semantic parsing

sentence

Generate Logical form

corpus

COE’07

7

corpus

CYK Parsing

Structured SVM

Semantic tagging

Decoding

Conditional Random Filed

model

COE’07

8

Example: Generate LF from a semantic tree

Semantic parsing Semantic tagging

Pos Tagging

Semantic tree

Rules With weights

Semantic tree

COE’07

9

Logical form Generation

Logical form

COE’07

10

Legal domain

Current Work • English Data (CLANG)

KNP, Cabocha

– Structured SVM (Robocup) • Precision: 85% • Recall: 74% – Maximum entropy model (DB query) • Precision 89% • Recall 51%

Dependency tree is suitable for the Japanese legal domain

W1

sentence

Dependency parsing

W2 W3 W4

Structured SVM Online-SCFG

Semantic parsing

W5 W6 W7 W8

• Japanese Data

∀x1 , x2 , ∃e1 , e2 , e3

– Splitting long sentences into a set of short sentences – Mapping NL Japanese sentence to logical form

P( w6' ( e1 ) ∧ agt (e1 , a ) ∧ obj ( e1 , e2 ) ∧ (( w3' ( e2 ) ∧ obj ( e2 , x1 ) ∧ w2' ( x1)) ∧ (w 5' ( e2 ) ∧ obj ( e2 , x2 ) ∧ w4' ( x2 )))

Logical form

W9

patterns

Correspondences between DT’s node and predicate in LF

COE’07

11

COE’07

12

2

Dependency Parsing

Online Structured Learning

• • •

Dependency tree

Dependency Parsing

Sentence

corpus

Patterns learning

Example:

Japanese unlabeled accuracy 93.9%

Dependency Model

Input: set of dependency trees and their logical forms Output: the correspondences of each node in DT with a predicate in LF Method: Using statistical machine translation to align a node in DT to a predicate a LF W1 W2 W3 W4 W5 W6 W7 W8 W9 ∀x1 , x2 , ∃e1 , e2 , e3

Japanese unlabeled accuracy 91.6%

P( w6' (e1 ) ∧ agt (e1 , a ) ∧ obj ( e1 , e2 ) ∧ (( w3' (e2 ) ∧ obj (e2 , x1 ) ∧ w2' ( x1)) ∧ (w '5 (e2 ) ∧ obj (e2 , x2 ) ∧ w4' ( x2 )))

very good results in shared task CONLL-2006 State of the art result in English data (Penn III)

including Japanese data

English unlabeled accuracy 91.6%

Plan to participate CONLL-2007 task

English unlabeled accuracy 90.7%

COE’07

13

Splitting clauses in legal domain

splitting NL

14

Online-SCFG Methods

• Using machine learning for splitting clauses mapping

COE’07

• Preprocessing LF

LF

– Generate a sequence of word tokens – Transform a logical form representation into a sequence of atomic logical form.

combining

mapping



mapping

Using GIZA++ to generate alignment between each word in NL to each token in LF

• Using synchronous grammar to estimate the model for generating logical form zenjou ni gaitou suru mono ga aru toki ha, kuchou ha, kore o kokuhatsu suru mono to suru

前条に該当する者があるときは、区長は、これを告発するものとする。

• Using online structured prediction learning to estimate the SCFG grammar

前条に該当する者があるときは、区長は これを告発するものとする。

• Some issues for Japanese data

¾ Collected 108 sentences and their logical form ¾ 9/10 for training and 1/10 for testing data ¾ The accuracy of the model is 93.10%!

– Require a formal grammar representation for LF – In the case there is no formal grammar it becomes phrase based SMT models

COE’07

15

• Developed by Aho & Ullman (1972) as a theory of compilers that combines syntax analysis and code generation in a single phase • Generates a pair of strings in a single derivation

QUERY

What

is

16

Synchronous Context-Free Grammars (SCFG)

Context-Free Semantic Grammar

QUERY Æ What is CITY

COE’07

CITY

CITY Æ the capital CITY CITY Æ of STATE STATE Æ Ohio

the

capital of

CITY STATE

Ohio

COE’07

17

COE’07

18

3

Synchronous Context-Free Grammars

Synchronous Context-Free Grammars Developed by Aho & Ullman (1972) as a theory of compilers that combines syntax analysis and code generation in a single phase Generates a pair of strings in a single derivation pattern

QUERY

template

What

is

QUERY

answer

CITY

(

CITY

)

QUERY Æ What is CITY / answer(CITY) the

capital

capital

CITY

of

STATE

Ohio

(

loc_2

CITY

(

)

STATE

stateid

(

)

'ohio'

)

CITY ÆOhio the CITY capital(CITY) CITY Æ What ofcapital loc_2(STATE) QUERY Æ is CITY // answer(CITY) STATE ÆSTATE Ohio answer(capital(loc_2(stateid('ohio')))) / /stateid('ohio') What is the capital of

COE’07

19

Probabilistic Parsing Model

d1 capital

(

loc_2

CITY (

) )

• • • •

CITY

capital

STATE

20

Parsing Model

d2

CITY

COE’07

(

loc_2

CITY (

)

RIVER

)

N (non-terminals) = {QUERY, CITY, STATE, …} S (start symbol) = QUERY Tm (MRL terminals) = {answer, capital, loc_2, (, ), …} Tn (NL words) = {What, is, the, capital, of, Ohio, …}

QUERY Æ What is CITY / answer(CITY) stateid

(

'ohio'

)

riverid

CITY Æ capital CITY / capital(CITY)

0.5

CITY Æ of STATE / loc_2(STATE)

0.3

STATE Æ Ohio / stateid('ohio')

0.5

+

λ

Pr(d1|capital of Ohio) = exp( 1.3 ) / Z

(

'ohio'

)

CITY Æ capital CITY / capital(CITY)

0.5

CITY Æ of RIVER / loc_2(RIVER)

0.05

RIVER Æ Ohio / riverid('ohio')

0.5

+

• L (lexicon) =

CITY Æ of STATE / loc_2(STATE)

λ

STATE Æ Ohio / stateid('ohio') • λ (parameters of probabilistic model) = ?

Pr(d2|capital of Ohio) = exp( 1.05 ) / Z

Online structured prediction learning

normalization constant COE’07

CITY Æ the capital CITY / capital(CITY)

21

COE’07

Online structured prediction learning

Results

• Extend the traditional Perceptron learning

• English data (CLANG)

Perceptron

Extend

22

– It is applicable for Robocup language • Precision 89.5 (best precision) • Recall 61.2 – The result is good because we do not need fully semantic tree annotation

Online structured Learning

• Japanese data (110 sentences) Two class labels learning problem Many class labels, tree structure

COE’07

– Because of spare data problem so the alignment of each word in NL sentence and each token in LF is not good. – It is need to verify this problem in detail for improving the accuracy of our model • Enlarge the training data ?

23

COE’07

24

4

Conclusions

Future work

• Learning is applicable for transforming NL to logical form

• Experiment on a larger corpus of Japanese data • Integrate with the rule-based models

• The number of training data should be enlarged to make sure the accuracy of the models

• Transforming a logical form representation to a NL sentence

• The splitting result for legal Japanese data is attractive • Semi-supervised learning models for semantic parsing We should integrate this model with the rule-based model COE’07

25

COE’07

26

5