SlideShare a Scribd company logo
Training a Natural Language Generator From
Unaligned Data
PHILIP ARTHUR
MT STUDY GROUP
9/3/2015 1
Paper Description
Title: Training a Natural Language Generator From Unaligned Data
Author: Ondrej Dusek and Filip Jurcicek
Meeting: ACL
Year: 2015
Type: Long Paper
Reference: https://aclweb.org/anthology/P/P15/P15-1044.pdf
9/3/2015 2
Natural Language Generation
• Generating a natural language from abstract representation.
9/3/2015 3
Motivation & Contribution
• Motivation
• Current NLG system requires separate training data alignment step.
• Using CFG or phrase based limit the ability to capture long-range syntactic dependencies.
• Contribution
• Novel method that integrates alignment step into sentence planner.
• Using deep-syntactic trees with rule based surface realization.
• Ability to learn from incomplete tree.
9/3/2015 4
System Overview
9/3/2015 5
Main contribution
About the data structure
• Each node has a lemma and formeme. (Dušek et al., 2012)
• Contain nodes for content words (nouns, full verbs, adjectives, adverbs) and coordinating
conjunctions.
• Treex toolkit is used to generate this dependency tree for the input.
9/3/2015 6
Dataset
• BAGEL: dataset of restaurant, consists of Dialogue Acts (DA) & list of slot-value pairs (SVPs),
contains information about restaurants.
• Where: http://farm2.user.srcf.net/research/bagel/ACL10-inform-training.txt
• Example:
9/3/2015 7
FULL_DA = inform(name="Green Man",eattype=restaurant)
ABSTRACT_DA = inform(name="X1",eattype=restaurant)
-> "[name+X]X []is a [eattype+restaurant]restaurant.";
Sentence Planner
• Building a deep syntactic tree based on A* search. (2 Hypotheses: Open + Closed)
• Based on: Candidate Generator + Scorer/Ranker
9/3/2015 8
Sentence Planner Algorithm
Init: Start from an open set with a single empty sentence plan tree and an empty closed set.
Loop:
1. Select the best-scoring candidate C from the open set. Add C to closed set.
2. The candidate generator generates C, a set of possible successors to C. These are trees that
have more nodes than C and are deemed viable. Note that C may be empty.
3. The scorer scores all successors in C and if they are not already in the closed set, it adds
them to the open set.
4. Check if the best successor in the open set scores better than the best candidate in the
closed set.
Stop: The algorithm finishes if the top score in the open set is lower than the top score in the
closed set for d consecutive iterations, or if there are no more candidates in the open set. It
returns the best-scoring candidate from both sets.
9/3/2015 9
Candidate Generator
• Adding one new node in all possible positions with all possible lemmas and formemes.
9/3/2015 10
CG + Expansion Filtering
1. Lemma-formeme compatibility: nodes with combination of lemma + formeme in Training.
2. Syntatic vialibility: new node generated compatible with parent (seen before, including
dependency of left and right child).
3. Number of children: For particular parent node, #child can’t exceed the max #child of same
node in Training.
4. Tree Size: #nodes in tree can’t exceed the max #nodes of tree in training data. For each level.
5. Weak semantic compatibility: Include node that appears in training, containing SVPs from
the current input.
6. Strong semantic compatibility: For each lemma + formeme, there is a compatibility list.
Node generation is allowed if all of SVPs are present in the current input DA.
9/3/2015 11
Scorer / Ranker
• Is a function that maps global features from sentence plan 𝑡 + input 𝑚 to a real value.
• Based on basic perceptron scorer:
9/3/2015 12
Training
Objective:
Init: all 𝑤 ∈ 𝐰 = 1
For each input MR in Training:
1. 𝑡𝑡𝑜𝑝 = Generate a sentence plan of the input based on current weight.
2. 𝑡 𝑔𝑜𝑙𝑑 = Parsed input by using automatic annotator (treex).
3. Update:
9/3/2015 13
Differing Subtrees Update
• Starting from the common subtree 𝑡 𝑐 of 𝑡𝑡𝑜𝑝and 𝑡 𝑔𝑜𝑙𝑑, pairs of differing subtrees 𝑡𝑡𝑜𝑝
𝑖
, 𝑡 𝑔𝑜𝑙𝑑
𝑖
are
created by gradually adding nodes from 𝑡𝑡𝑜𝑝into 𝑡𝑡𝑜𝑝
𝑖
and from 𝑡 𝑔𝑜𝑙𝑑 into 𝑡 𝑔𝑜𝑙𝑑
𝑖
.
9/3/2015 14
Algorithm Differing Subtree Update
• In the third step of training, substitute “full-tree” update with “” update:
• It is reported that if we don’t use the same size of subtree, performance will degrade.
9/3/2015 15
Future Promise Estimation
• The same idea of A* search where score = scorer(input, weight) + heuristic(input).
• Based on Expected number of children 𝐸𝑐(𝑛) of different node types.
• The future promise (fp) of a particular sentence plan 𝑡 is calculated based on its node 𝑛1 … 𝑛 𝑡:
• 𝑐(𝑛𝑖) is the current number of children.
• 𝜆 is a preset parameter.
• Not included in stop criterion check.
9/3/2015 16
Averaging Weight + Parallel Training
• Using Iterative mixing approrach (McDonald et al. 2010).
• Training data are splitted into several parts.
• Weights updated are averaged after each pass through the training data.
• Record weights after each training pass, take an average at the end  Final weight.
9/3/2015 17
Surface Realizer
• Built using Treex NLP toolkit (Ptacek, 2008).
• Rule based simple pipeline, outlining:
1. Agreement
2. Word Ordering
3. Compound verb forms
4. Grammatical words
5. Punctuation
6. Word Inflection
7. Phonetic Changes
• Round trip test using automatic analysis with subsequent generation reached 89.79% BLEU.
9/3/2015 18
Features
• current tree properties: depth, #nodes, #repeated_nodes
• tree and input DA: #nodes/SVP, #repeated_nodes/SVP
• node features: lemma, formeme, #children of all nodes in the current tree.
• input features: whole SVPs (slot+value), just slots, and pairs of slots in the DA
• combinations of node and input features
• repeat features: #repeated_lemma_formeme with #repeated_slots in the input DA.
• dependency features: parent-child pairs for lemmas + formeme, including left or right.
• sibling features: sibling pairs for lemmas + formeme, combined with SVP.
• bigram features: pairs of lemmas + formeme adjacent in tree left-right order, combined with
SVP.
9/3/2015 19
Setup
• #iterations do not improve:
• Training = 3
• Testing = 4
• Maximum 200 sentence planner iterations per input DA.
• 𝛼 = 0.1
• If fp is used then 𝛼 = 0.3
• 10 folds cross validation is used in the experiment.
9/3/2015 20
Results
• The Proposed method gives improvement on both BLEU and NIST with whole training portion.
• Compare to the previous work (67%) is still lower.  Task is harder, no alignment is used.
• Larger training data can demonstrate the effectiveness of the proposed method.
• Both improvement gain a 95% confidence of statistical significance testing (Koehn, 2004).
9/3/2015 21
Example of Generated pairs
9/3/2015 22
Discussion
+ Generator learns to produce meaningful utterances that correspond well to the input DA.
- Not all required information is always present.
- Some facts are sometimes repeated or irrelevant information Appears
◦ Occurs because of the data sparsity.
◦ Design a scorer features that discourage conflicting information.
- Repeated slots in input are not handled correctly.
9/3/2015 23
Conclusion
• The paper presented a NLG, capable of learning from unaligned pairs.
• The contribution consists of A* based sentence planner, rule-based surface realization from
Treex toolkit.
• The empiric results shows a promising result, although didn’t surpass the previous work, but
their work is substantially harder.
• Code: https://github.com/UFAL-DSG/tgen
9/3/2015 24

More Related Content

PPTX
Ensemble learning
PPTX
Ensemble methods
PDF
Ensemble modeling and Machine Learning
PPTX
Ensemble methods
PPTX
Machine Learning - Ensemble Methods
PPTX
Bag the model with bagging
PPTX
Ensemble learning
PPTX
Ensemble learning Techniques
Ensemble learning
Ensemble methods
Ensemble modeling and Machine Learning
Ensemble methods
Machine Learning - Ensemble Methods
Bag the model with bagging
Ensemble learning
Ensemble learning Techniques

What's hot (20)

PPTX
Lecture 6: Ensemble Methods
PPT
ensemble learning
PPTX
Semi supervised learning machine learning made simple
PDF
Machine Learning and Data Mining: 16 Classifiers Ensembles
PDF
Teaching calculus with Wolfram Alpha
PDF
A Neural Grammatical Error Correction built on Better Pre-training and Sequen...
PDF
H2O World - Ensembles with Erin LeDell
PPT
Ensemble Learning Featuring the Netflix Prize Competition and ...
PPT
Semi-supervised Learning
PDF
(SURVEY) Semi Supervised Learning
PDF
Lecture 9: Machine Learning in Practice (2)
PDF
Exposé Ontology
PPTX
Instance based learning
PDF
BigML Education - Feature Engineering with Flatline
PDF
Meta learning tutorial
PDF
Lecture 3: Basic Concepts of Machine Learning - Induction & Evaluation
PDF
Boosting Algorithms Omar Odibat
PDF
Learning loss for active learning
PDF
Feature Importance Analysis with XGBoost in Tax audit
ODP
Linear Regression Ex
Lecture 6: Ensemble Methods
ensemble learning
Semi supervised learning machine learning made simple
Machine Learning and Data Mining: 16 Classifiers Ensembles
Teaching calculus with Wolfram Alpha
A Neural Grammatical Error Correction built on Better Pre-training and Sequen...
H2O World - Ensembles with Erin LeDell
Ensemble Learning Featuring the Netflix Prize Competition and ...
Semi-supervised Learning
(SURVEY) Semi Supervised Learning
Lecture 9: Machine Learning in Practice (2)
Exposé Ontology
Instance based learning
BigML Education - Feature Engineering with Flatline
Meta learning tutorial
Lecture 3: Basic Concepts of Machine Learning - Induction & Evaluation
Boosting Algorithms Omar Odibat
Learning loss for active learning
Feature Importance Analysis with XGBoost in Tax audit
Linear Regression Ex
Ad

Viewers also liked (6)

PPT
Syntax analysis
PPT
NLTK: Natural Language Processing made easy
PDF
GPU Accelerated Natural Language Processing by Guillermo Molini
PPTX
Words and lexemes ppt
PDF
Practical Natural Language Processing
PDF
Natural Language Processing
Syntax analysis
NLTK: Natural Language Processing made easy
GPU Accelerated Natural Language Processing by Guillermo Molini
Words and lexemes ppt
Practical Natural Language Processing
Natural Language Processing
Ad

Similar to [Paper Introduction] Training a Natural Language Generator From Unaligned Data (20)

PPTX
AIML UNIT 4.pptx. IT contains syllabus and full subject
PPTX
Unit V -Multiple Learners.pptx for artificial intelligence
PPTX
Unit V -Multiple Learners in artificial intelligence and machine learning
PPT
5_Model for Predictions_Machine_Learning.ppt
PPT
PPT SLIDES
PPT
PPT SLIDES
PDF
Self training improves_nlu
PPTX
UNIT-II-Machine-Learning.pptx Machine Learning Different AI Models
PPTX
Presentation on supervised learning
PPT
activelearning.ppt
PPTX
Artificial Neural Networks , Recurrent networks , Perceptron's
PDF
Training Deep Networks with Backprop (D1L4 Insight@DCU Machine Learning Works...
PPTX
Bagging - Boosting-and-Stacking-ensemble.pptx
PDF
Concepts of Database Management 8th Edition Pratt Solutions Manual
PPTX
Lecture2_machine learning training+testing.pptx
PDF
Active learning for ranking through expected loss optimization
PDF
Modelling and evaluation
PDF
FLY-SMOTE.pdf
AIML UNIT 4.pptx. IT contains syllabus and full subject
Unit V -Multiple Learners.pptx for artificial intelligence
Unit V -Multiple Learners in artificial intelligence and machine learning
5_Model for Predictions_Machine_Learning.ppt
PPT SLIDES
PPT SLIDES
Self training improves_nlu
UNIT-II-Machine-Learning.pptx Machine Learning Different AI Models
Presentation on supervised learning
activelearning.ppt
Artificial Neural Networks , Recurrent networks , Perceptron's
Training Deep Networks with Backprop (D1L4 Insight@DCU Machine Learning Works...
Bagging - Boosting-and-Stacking-ensemble.pptx
Concepts of Database Management 8th Edition Pratt Solutions Manual
Lecture2_machine learning training+testing.pptx
Active learning for ranking through expected loss optimization
Modelling and evaluation
FLY-SMOTE.pdf

More from NAIST Machine Translation Study Group (14)

PDF
[Paper Introduction] Efficient Lattice Rescoring Using Recurrent Neural Netwo...
PDF
[Paper Introduction] Distant supervision for relation extraction without labe...
PDF
On using monolingual corpora in neural machine translation
PDF
RNN-based Translation Models (Japanese)
PDF
[Paper Introduction] Efficient top down btg parsing for machine translation p...
PDF
[Paper Introduction] Translating into Morphologically Rich Languages with Syn...
PDF
[Paper Introduction] Supervised Phrase Table Triangulation with Neural Word E...
PDF
[Paper Introduction] Evaluating MT Systems with Second Language Proficiency T...
PDF
[Paper Introduction] Bilingual word representations with monolingual quality ...
PDF
[Paper Introduction] A Context-Aware Topic Model for Statistical Machine Tran...
PDF
[Book Reading] 機械翻訳 - Section 3 No.1
PPTX
[Book Reading] 機械翻訳 - Section 5 No.2
PDF
[Book Reading] 機械翻訳 - Section 7 No.1
PDF
[Book Reading] 機械翻訳 - Section 2 No.2
[Paper Introduction] Efficient Lattice Rescoring Using Recurrent Neural Netwo...
[Paper Introduction] Distant supervision for relation extraction without labe...
On using monolingual corpora in neural machine translation
RNN-based Translation Models (Japanese)
[Paper Introduction] Efficient top down btg parsing for machine translation p...
[Paper Introduction] Translating into Morphologically Rich Languages with Syn...
[Paper Introduction] Supervised Phrase Table Triangulation with Neural Word E...
[Paper Introduction] Evaluating MT Systems with Second Language Proficiency T...
[Paper Introduction] Bilingual word representations with monolingual quality ...
[Paper Introduction] A Context-Aware Topic Model for Statistical Machine Tran...
[Book Reading] 機械翻訳 - Section 3 No.1
[Book Reading] 機械翻訳 - Section 5 No.2
[Book Reading] 機械翻訳 - Section 7 No.1
[Book Reading] 機械翻訳 - Section 2 No.2

Recently uploaded (20)

PDF
Communicating Health Policies to Diverse Populations (www.kiu.ac.ug)
PDF
Assessment of environmental effects of quarrying in Kitengela subcountyof Kaj...
PPTX
Substance Disorders- part different drugs change body
PPTX
Biomechanics of the Hip - Basic Science.pptx
PPT
Animal tissues, epithelial, muscle, connective, nervous tissue
PPTX
Hypertension_Training_materials_English_2024[1] (1).pptx
PPT
Mutation in dna of bacteria and repairss
PDF
Cosmic Outliers: Low-spin Halos Explain the Abundance, Compactness, and Redsh...
PPTX
gene cloning powerpoint for general biology 2
PDF
Worlds Next Door: A Candidate Giant Planet Imaged in the Habitable Zone of ↵ ...
PPTX
Welcome-grrewfefweg-students-of-2024.pptx
PPTX
SCIENCE 4 Q2W5 PPT.pptx Lesson About Plnts and animals and their habitat
PDF
lecture 2026 of Sjogren's syndrome l .pdf
PPTX
Understanding the Circulatory System……..
PDF
CHAPTER 2 The Chemical Basis of Life Lecture Outline.pdf
PPT
Heredity-grade-9 Heredity-grade-9. Heredity-grade-9.
PPTX
BIOMOLECULES PPT........................
PDF
CHAPTER 3 Cell Structures and Their Functions Lecture Outline.pdf
PPTX
Microbes in human welfare class 12 .pptx
PDF
Worlds Next Door: A Candidate Giant Planet Imaged in the Habitable Zone of ↵ ...
Communicating Health Policies to Diverse Populations (www.kiu.ac.ug)
Assessment of environmental effects of quarrying in Kitengela subcountyof Kaj...
Substance Disorders- part different drugs change body
Biomechanics of the Hip - Basic Science.pptx
Animal tissues, epithelial, muscle, connective, nervous tissue
Hypertension_Training_materials_English_2024[1] (1).pptx
Mutation in dna of bacteria and repairss
Cosmic Outliers: Low-spin Halos Explain the Abundance, Compactness, and Redsh...
gene cloning powerpoint for general biology 2
Worlds Next Door: A Candidate Giant Planet Imaged in the Habitable Zone of ↵ ...
Welcome-grrewfefweg-students-of-2024.pptx
SCIENCE 4 Q2W5 PPT.pptx Lesson About Plnts and animals and their habitat
lecture 2026 of Sjogren's syndrome l .pdf
Understanding the Circulatory System……..
CHAPTER 2 The Chemical Basis of Life Lecture Outline.pdf
Heredity-grade-9 Heredity-grade-9. Heredity-grade-9.
BIOMOLECULES PPT........................
CHAPTER 3 Cell Structures and Their Functions Lecture Outline.pdf
Microbes in human welfare class 12 .pptx
Worlds Next Door: A Candidate Giant Planet Imaged in the Habitable Zone of ↵ ...

[Paper Introduction] Training a Natural Language Generator From Unaligned Data

  • 1. Training a Natural Language Generator From Unaligned Data PHILIP ARTHUR MT STUDY GROUP 9/3/2015 1
  • 2. Paper Description Title: Training a Natural Language Generator From Unaligned Data Author: Ondrej Dusek and Filip Jurcicek Meeting: ACL Year: 2015 Type: Long Paper Reference: https://aclweb.org/anthology/P/P15/P15-1044.pdf 9/3/2015 2
  • 3. Natural Language Generation • Generating a natural language from abstract representation. 9/3/2015 3
  • 4. Motivation & Contribution • Motivation • Current NLG system requires separate training data alignment step. • Using CFG or phrase based limit the ability to capture long-range syntactic dependencies. • Contribution • Novel method that integrates alignment step into sentence planner. • Using deep-syntactic trees with rule based surface realization. • Ability to learn from incomplete tree. 9/3/2015 4
  • 6. About the data structure • Each node has a lemma and formeme. (Dušek et al., 2012) • Contain nodes for content words (nouns, full verbs, adjectives, adverbs) and coordinating conjunctions. • Treex toolkit is used to generate this dependency tree for the input. 9/3/2015 6
  • 7. Dataset • BAGEL: dataset of restaurant, consists of Dialogue Acts (DA) & list of slot-value pairs (SVPs), contains information about restaurants. • Where: http://farm2.user.srcf.net/research/bagel/ACL10-inform-training.txt • Example: 9/3/2015 7 FULL_DA = inform(name="Green Man",eattype=restaurant) ABSTRACT_DA = inform(name="X1",eattype=restaurant) -> "[name+X]X []is a [eattype+restaurant]restaurant.";
  • 8. Sentence Planner • Building a deep syntactic tree based on A* search. (2 Hypotheses: Open + Closed) • Based on: Candidate Generator + Scorer/Ranker 9/3/2015 8
  • 9. Sentence Planner Algorithm Init: Start from an open set with a single empty sentence plan tree and an empty closed set. Loop: 1. Select the best-scoring candidate C from the open set. Add C to closed set. 2. The candidate generator generates C, a set of possible successors to C. These are trees that have more nodes than C and are deemed viable. Note that C may be empty. 3. The scorer scores all successors in C and if they are not already in the closed set, it adds them to the open set. 4. Check if the best successor in the open set scores better than the best candidate in the closed set. Stop: The algorithm finishes if the top score in the open set is lower than the top score in the closed set for d consecutive iterations, or if there are no more candidates in the open set. It returns the best-scoring candidate from both sets. 9/3/2015 9
  • 10. Candidate Generator • Adding one new node in all possible positions with all possible lemmas and formemes. 9/3/2015 10
  • 11. CG + Expansion Filtering 1. Lemma-formeme compatibility: nodes with combination of lemma + formeme in Training. 2. Syntatic vialibility: new node generated compatible with parent (seen before, including dependency of left and right child). 3. Number of children: For particular parent node, #child can’t exceed the max #child of same node in Training. 4. Tree Size: #nodes in tree can’t exceed the max #nodes of tree in training data. For each level. 5. Weak semantic compatibility: Include node that appears in training, containing SVPs from the current input. 6. Strong semantic compatibility: For each lemma + formeme, there is a compatibility list. Node generation is allowed if all of SVPs are present in the current input DA. 9/3/2015 11
  • 12. Scorer / Ranker • Is a function that maps global features from sentence plan 𝑡 + input 𝑚 to a real value. • Based on basic perceptron scorer: 9/3/2015 12
  • 13. Training Objective: Init: all 𝑤 ∈ 𝐰 = 1 For each input MR in Training: 1. 𝑡𝑡𝑜𝑝 = Generate a sentence plan of the input based on current weight. 2. 𝑡 𝑔𝑜𝑙𝑑 = Parsed input by using automatic annotator (treex). 3. Update: 9/3/2015 13
  • 14. Differing Subtrees Update • Starting from the common subtree 𝑡 𝑐 of 𝑡𝑡𝑜𝑝and 𝑡 𝑔𝑜𝑙𝑑, pairs of differing subtrees 𝑡𝑡𝑜𝑝 𝑖 , 𝑡 𝑔𝑜𝑙𝑑 𝑖 are created by gradually adding nodes from 𝑡𝑡𝑜𝑝into 𝑡𝑡𝑜𝑝 𝑖 and from 𝑡 𝑔𝑜𝑙𝑑 into 𝑡 𝑔𝑜𝑙𝑑 𝑖 . 9/3/2015 14
  • 15. Algorithm Differing Subtree Update • In the third step of training, substitute “full-tree” update with “” update: • It is reported that if we don’t use the same size of subtree, performance will degrade. 9/3/2015 15
  • 16. Future Promise Estimation • The same idea of A* search where score = scorer(input, weight) + heuristic(input). • Based on Expected number of children 𝐸𝑐(𝑛) of different node types. • The future promise (fp) of a particular sentence plan 𝑡 is calculated based on its node 𝑛1 … 𝑛 𝑡: • 𝑐(𝑛𝑖) is the current number of children. • 𝜆 is a preset parameter. • Not included in stop criterion check. 9/3/2015 16
  • 17. Averaging Weight + Parallel Training • Using Iterative mixing approrach (McDonald et al. 2010). • Training data are splitted into several parts. • Weights updated are averaged after each pass through the training data. • Record weights after each training pass, take an average at the end  Final weight. 9/3/2015 17
  • 18. Surface Realizer • Built using Treex NLP toolkit (Ptacek, 2008). • Rule based simple pipeline, outlining: 1. Agreement 2. Word Ordering 3. Compound verb forms 4. Grammatical words 5. Punctuation 6. Word Inflection 7. Phonetic Changes • Round trip test using automatic analysis with subsequent generation reached 89.79% BLEU. 9/3/2015 18
  • 19. Features • current tree properties: depth, #nodes, #repeated_nodes • tree and input DA: #nodes/SVP, #repeated_nodes/SVP • node features: lemma, formeme, #children of all nodes in the current tree. • input features: whole SVPs (slot+value), just slots, and pairs of slots in the DA • combinations of node and input features • repeat features: #repeated_lemma_formeme with #repeated_slots in the input DA. • dependency features: parent-child pairs for lemmas + formeme, including left or right. • sibling features: sibling pairs for lemmas + formeme, combined with SVP. • bigram features: pairs of lemmas + formeme adjacent in tree left-right order, combined with SVP. 9/3/2015 19
  • 20. Setup • #iterations do not improve: • Training = 3 • Testing = 4 • Maximum 200 sentence planner iterations per input DA. • 𝛼 = 0.1 • If fp is used then 𝛼 = 0.3 • 10 folds cross validation is used in the experiment. 9/3/2015 20
  • 21. Results • The Proposed method gives improvement on both BLEU and NIST with whole training portion. • Compare to the previous work (67%) is still lower.  Task is harder, no alignment is used. • Larger training data can demonstrate the effectiveness of the proposed method. • Both improvement gain a 95% confidence of statistical significance testing (Koehn, 2004). 9/3/2015 21
  • 22. Example of Generated pairs 9/3/2015 22
  • 23. Discussion + Generator learns to produce meaningful utterances that correspond well to the input DA. - Not all required information is always present. - Some facts are sometimes repeated or irrelevant information Appears ◦ Occurs because of the data sparsity. ◦ Design a scorer features that discourage conflicting information. - Repeated slots in input are not handled correctly. 9/3/2015 23
  • 24. Conclusion • The paper presented a NLG, capable of learning from unaligned pairs. • The contribution consists of A* based sentence planner, rule-based surface realization from Treex toolkit. • The empiric results shows a promising result, although didn’t surpass the previous work, but their work is substantially harder. • Code: https://github.com/UFAL-DSG/tgen 9/3/2015 24