答案句子选择 孙亚伟.

Slides:



Advertisements
Similar presentations
广州市教育局教学研究室英语科 Module 1 Unit 2 Reading STANDARD ENGLISH AND DIALECTS.
Advertisements

Mark Twain. How much do you know about him? (1835—1910) The greatest humorist of the 19th century in American literature. Also one of the greatest writer.
Which TV program is the video? 中国达人秀 China’s Got Talent 选秀节目 talent show talent n. 天资;天赋.
高考英语专题复习 《 利用 21 世纪英语报提高阅读 理解能力技巧 》 晋江市第一中学 英语组 黄蓉蓉.
桂林市 2011 年高三第二次调研考 试质量分析暨备考教学建议 桂林市教育科学研究所 李陆桂. 二调平均分与一调、 2010 广西高考英语平均分的比较 科目 类别 英语 文科文科 2010 年广西 一调 二调 与 10 年广西相差
次丘镇中学 王正东. 1 .常见用法 定冠词 the 的用法 I saw a film last night.The film is very interesting. 昨天晚上我看了一部电影。这部电影非常有趣。 The bag on the desk isn't mine. 桌子上的那个书包不是我的。
1 )正确 2 )多词 3 )缺词 4 )错词 删除 补漏 更正 “1126” 原则 “1225” 原则 “1117” 原则.
高考英语短文改错 试题解析 内蒙古师范大学外国语学院 方芳 2011 年 3 月. 一、短文改错设疑方式 此 题要求改正所给短文中的错误。对标有 题号的每一行做出判断: 1) 如无错误,在该行右边横线上画一个 ( );如有错误(每行只有一个错误), 则按下列情况改正:
2 Learning procedures Comments on your homework ( 作业评讲) Tips on how to make your sentences better ( 技巧指导) 3 Practice makes perfect ( 巩固复习) 1.
短文改错解题技巧 1 )错词 2 )多词 3 )缺词 更正 删除 补漏 短文改错(共 10 小题,每小题 1 分,满分 10 分) 假定英语课上老师要求同桌之间交换修改作文,请你 修改你同桌写的以下作文。文中共有 10 处语言错误, 每句中最多有两处。错误涉及一个单词的增加、删除 或修改。 增加:在缺词处加一个漏字符号(
L5-L6 Review Oct. 30, Adj. as predicate 1.She is tall and pretty. 2.His house is pretty but very small. 3.Is your book expensive ? (try different.
1.Founder: Elvis Presley(The Hillbilly Cat)
春节的由来 chūn jiē de yóu lái Learn Chinese through Chinese Holidays
Click to edit Master title style
中职英语课程改革中 如何实践“以就业为导向,服务为宗旨”的办学理念
He said: What is a team? Team is not to let the other person failed, and do not let any team member fail!
广德二中2006届高考 英语专题复习 单项填空 答题指导.
数 学 与 工 程 的 对 话 中山大学 信息科学与技术学院 李硕彦教授演讲 (10月21, 24日) 李硕彦 ( Bob Li ) 简介:
3. 一般問題 部份資料來源: YAHOO網 及本校08年升中學生提供
Module 1 My classmates.
湖州中学微课程 走进定语从句的世界 --初识定语从句 湖州中学 朱筱杭 湖州中学微课程.
Unit 2 Warm-up 吴小凰.
American Music —— black music
Unit 3 A day out Welcome to the unit.
Welcome Welcome to my class Welcome to my class!.
Unit 9 What does he look like?
Festivals around the world
U5 people’s daily life 义务教育广州版小学英语教材 M3 Daily life 指导:市教育研究院 赵淑红
Module 12 Western music 第2课时.
Module 10 Life history Unit2 He decided to be an actor.
What do you think of game shows?
WRITNG Welcome to enjoy English..
Unit 4 I used to be afraid of the dark.
Adventure in Literature
下 页.
Life relies on sports 生命在于运动.
Reading Do you remember what you were doing? 学习目标 1、了解几个重要历史事件。
Unit 5.
In this lesson, you will review how to
Unit 12 What did you do last weekend?.
Nationality Objective
Unit 2 I think that mooncakes are delicious!
中国英语教师网.
学练优英语教学课件 八年级(上) it! for Go
Do you want to watch a game show?
動詞的時態 - by May.
词汇语义资源在中文关系抽取中的应用 报告人:钱龙华 刘丹丹 胡亚楠 钱龙华 周国栋
住棚節 oct 4, 2015.
Nationality Objective
Nationality Objective
Unit 5 Why do you like pandas?
Unit title: 假期 – Holiday
以弗所書 1:1-23 神永世的計畫 2:1-10 個人得救的步驟 2:11-22 猶太人與外邦人 成為一體.
10/26 今天的学习目标 (Today’s Learning Objectives)
Uses of “It” I、 用作人称代词的 it II. 用作先行词的 it III. 用在强调句型中的 it
Lesson 44:Popular Sayings
Chapter 3 Nationality Objectives:
Who is he? He is my idol. Jack Ma.
Answering aggregation question over knowledge base
Word Revision. Word Revision 1.灾难 2.洪水 3.经历 4.引起,导致 5.猛烈的,激烈的 6.埋葬 7.发生 8.袭击 旋转的,循环的 飓风 雷暴 龙卷风 柱状物 热带的 墓地 棺材.
中央社新聞— <LTTC:台灣學生英語聽說提升 讀寫相對下降>
中考英语阅读理解 完成句子命题与备考 宝鸡市教育局教研室 任军利
中学英语教学中如何培养核心素养? ---基于学科关键问题的思考与实践
高中书面表达训练.
钱炘祺 一种面向实体浏览中属性融合的人机交互的设计与实现 Designing Human-Computer Interaction of Property Consolidation for Entity Browsing 钱炘祺
高考英语短文改错答题技巧 砀山中学 黄东亚.
Why do you like pandas? Section B 1a-2c.
多维阅读第13级.
Sun-Star第六届全国青少年英语口语大赛 全国总决赛 2015年2月 北京
Train Track and Children
Presentation transcript:

答案句子选择 孙亚伟

目录 问题 评测 方法 Tree Match Word Alignment Deep Learning 挑战 小结

问题 背景 问题 定义:给一问句和候选句子列表,找出正确句子(答案+证据) Q:Who won the best actor Oscar in 1973? S1:Jack Lemmon won the Academy Award for Best Actor for Save the Tiger (1973). S2:Oscar winner Kevin Spacey said that Jack Lemmon is remembered as always making time for other people.

评测 数据集QASent 指标 𝑀𝑅𝑅= 1 |𝑄| 𝑖=1 |𝑄| 1 𝑟𝑎𝑛𝑘 𝑖 𝑀𝐴𝑃= 𝑞=1 |𝑄| 𝐴𝑣𝑒𝑃(𝑞) |𝑄| 𝑀𝑅𝑅= 1 |𝑄| 𝑖=1 |𝑄| 1 𝑟𝑎𝑛𝑘 𝑖 𝑀𝐴𝑃= 𝑞=1 |𝑄| 𝐴𝑣𝑒𝑃(𝑞) |𝑄|   QASent 问句 来源 TREC 8-13 QA track 特性 人工编辑过; 每一问句均有正确答案; 句子 参加TREC的系统检索返回的句子 问句与句子有共同的非停用词 Class QASent Location 37 (16%) Human 65 (29%) Numeric 70 (31%) Abbreviation 2 (1%) Entity Description 16 (7%) QASent # of questions 227 # of sentences 8478 # of answers 928 Average length of questions 9.59 Average length of sentences 28.85

方法汇总 相关工作 MAP MRR Vasin Punyakanok (2004) 0.419 0.494 Hang Cui (2005) 0.427 0.526 Mengqiu Wang(2007) 0.603 0.685 Michael Heilman (2010) 0.609 0.692 Mengqiu Wang (2010) 0.595 0.695 Xuchen Yao (2013) 0.631 0.748 Aliaksei Severyn (2013) 0.678 0.736 Eyal Shnarch (2013) 0.686 0.754 Wen-tau Yih (2013) 0.709 0.770 Lei Yu (2014) 0.711 0.785 Di Wang (2015) 0.713 0.791 Minwei Feng (2015) 0.800 Aliaksei Severyn (2015) 0.746 0.808 Kateryna Tymoshenko (2015) 0.726 0.813 Zhiguo Wang (2015) 0.820 Ming Tan (2015) 0.728 0.832 Cicero dos Santos (2016) 0.753 0.851 Zhiguo Wang (2016) 0.771 0.845 http://aclweb.org/aclwiki/index.php?title=Question_Answering_(State_of_the_art)

方法归类 方法归类 方法抽象 相关工作 Tree Match Vasin Punyakanok (2004) Hang Cui (2005) Mengqiu Wang(2007) Michael Heilman (2010) Mengqiu Wang (2010) Xuchen Yao (2013) Aliaksei Severyn (2013) Kateryna Tymoshenko (2015) Word Alignment Eyal Shnarch (2013) Wen-tau Yih (2013) Deep Learning Lei Yu (2014) Di Wang (2015) Minwei Feng (2015) Aliaksei Severyn (2015) Zhiguo Wang (2015) Ming Tan (2015) Cicero dos Santos (2016) Zhiguo Wang (2016)

Tree Match Vasin Punyakanok (2004) 思想:计算问句依存树与句子依存树编辑距离。两棵树越像,越像正确句子 Hang Cui (2005): 思想:选择正确句子时候,不仅考虑问句与句子之间的lexical matching,还考虑问句中的关系与句子中关系的匹配程度,即Tree Path Match。两棵树关系路径越像,越像正确句子 方法:基于统计模型的fuzzy relation matching方法 Mengqiu Wang(2007) 思想:把句子和问句之间的相关性,看成是两棵依存树的语法转换(即answer (Semantic + Syntactic) transformation question) 方法:采用概率的准同步上下文无关文法 Michael Heilman (2010) 思想:计算问句依存树与句子依存树编辑距离后,考虑树编辑序列的特征,而不是总体长度 方法:首先计算源树到目标树编辑序列;然后从该序列提33个语法特征;最后转化为分类问题 Mengqiu Wang (2010) 思想:学习句子间的对齐结构(Structured Latent Variables),用于计算句子间的相关性 方法:提出一概率模型(Tree-edit CRF model) Xuchen Yao (2013) 思想: “TED特征+词汇语义特征”相结合方法来计算问句与句子之间的相关性 方法:特征= TED序列特征 + 词汇语义特征, 模型=逻辑回归 Kateryna Tymoshenko (2015) 思想:计算问句与句子相关性时,考虑了字符串中相似性以及字符串中概念实体的类属关系 方法:把问句和句子分别转换成树结构;利用外部资源给两棵树加关联边;利用Tree-Kernel自动抽取特征,转化为L2R问题。

Tree Match——Vasin Punyakanok (2004) 假设:正确句子依存树与问句依存树编辑距离(TED)最小,即“两棵树越像” 方法: 用树匹配近似算法来计算问句依存树与句子依存的编辑距离 Text Dependency Tree Tree Edit Distance(TED)Distance Score 实验:效果胜过词袋方法 补充: 有些句子局部能回答问句,不必分析全句 Lexical Semantic,需考虑更多语义关系

Tree Match——Hang Cui (2005) 假设:正确句子与问句,不仅词汇匹配,而且关系匹配 方法:基于aligned words在依存树上的路径,来算问句与句子之间的关系匹配。具体步骤: 从问句依存树和句子依存树上,抽取所有满足一定约束的路径; 借用机器翻译中对齐模型思想,来计算问句与句子关系匹配程度: 上式中 𝑃 𝑡 ( 𝑅𝑒𝑙 𝑖 S | 𝑅𝑒𝑙 𝐴 𝑖 Q ):采用最大期望和互信息来估计 实验:比词密度方法效果好 补充:

Tree Match——Michael Heilman (2010) 假设: TED过程,应考虑树编辑序列的特征,而不是总体长度 方法: 树转换树: 拓展树编辑操作:增加复杂操作,如重排序、移动 搜索最小编辑序列:启发式贪心搜索+Tree Kernel方法 特征表示:依据编辑操作,总结33个语法特征 分类器:逻辑回归算法 实验:在三个任务(RTE、Paraphrases、QA),效果提升 树转换树 特征表示 分类器 Premise: Pierce built the home for his daughter off Rossville Blvd, as he lives nearby. Hypothesis: Pierce lives near Rossville Blvd.

Word Alignment——Eyal Shnarch (2013) 假设:句子级推理可由词汇级推理结果推出。 方法:提出了一个基于Markovian的句子级概率推理模型。该模型分两层: Term-Level: Text与Hypothesis词汇对齐 Sentence-Level:依据Term-level结果,运用Markovian(考虑邻近的词汇因素)来估算从Text推导Hypothesis的概率 实验:在两个任务(RTE、QA),效果提升 补充:Lexical inference rules资源(如harmfuldangerous) 𝑥 𝑖 ∈ 0,1 ,1代表 ℎ 𝑖 inferred from Text; 𝑦 𝑖 ∈ 0,1 :推理决策变量, 𝑦 𝑛 :最终结果

Word Alignment——Wen-tau Yih (2013) 假设:正确句子与问句之间有一种潜在结构 方法:利用丰富的词汇语义关系和潜在结构,算出问句与句子之间的匹配程度,具体分两部分: 词汇语义关系:同义/反义;上下位;词语义相似 计算问句与句子匹配(两种模型): Bag-of-Words Model Learning Latent Structures 实验:Learning Latent Structures效果好过Bag-of-Words Model;丰富的词汇语义关系有必要 𝑞: What is the fastest car in the world? 𝑠: The Jaguar XJ220 is the dearest, fastest and most sought after car on the planet. 𝒉 𝑓 𝑞,𝑠 = 𝜃 𝑇 Φ(ℎ)

Deep Learning——Di Wang (2015) 假设:决定正确句子因素match the same words ; words meaning 方法:结合关键词匹配和堆双向长短词语记忆模型(BLSTM,考虑单词序列上下文),来计算问句和句子之间相关性 特征:BLSTM相关性 首先把问句和句子中的words借用word2vec转换为向量表示; 把这些向量先后送入双向的BLSTM模型; 最后模型输出它们的相关性结果 特征: BM25关键词匹配 模型: Gradient boosted regression tree (GBDT) 实验:三层BLSTM+BM25方案效果最好 补充:该方法没用句法分析和外在语义资源 What sport does Jennifer Capriati play? Positive Sentence: “Capriati, 19, who has not played competitive tennis since November 1994, has been given a wild card to take part in the Paris tournament which starts on February 13.” Negative Sentence: “Capriati also was playing in the U.S. Open semifinals in ’91, one year before Davenport won the junior title on those same courts.”

思考 方法归类 方法抽象 思考 Tree Match Word Alignment Deep Learning 如何表示树结构? 如何度量两棵树相似? 从两棵树相似程度,如何映射为问句与正确句子相关程度? 在树匹配时候,借用知识库资源给树节点赋予语义知识? 两棵树相似就代表是正确句子? Word Alignment 问句与句子对齐就代表是正确句子? 怎么挖掘词汇之间潜在关系? 如何利用知识库资源? 怎么利用词汇对齐推导句子之间的相关性? 两句间“最佳”对齐结构是什么? Deep Learning 句子级向量怎么表示? 怎么学习词汇语义矩阵? 怎么由词表示句? 两个句子向量的相似程度就代表是正确句子?

简单抽象模型 Q: S: [Q]: Who received the will rogers award ? 答案类型 Focus/实体 谓词 约束限定 共指/匹配 Type Match 转述/推理 Match S: 实体 谓词 答案 约束限定 冗余部分 [Q]: Who received the will rogers award ? [A]: kudos , frank -- frank sinatra , that is , who wednesday night was honored with the will rogers award at the beverly hills diamond jubilee gala .

挑战 谓词方面 问句 正确句子 wife married the author of by constructed built end late received honored came up with dubbed associated with related die killed started founded released spread top official or ceo executive director name at birth real name revenue sales original name a.k.a . affiliation member members performers [Q]: what is the name of durst 's group ? [A]: limp bizkit lead singer fred durst did a lot before he hit the big time . [Q]: what is florence nightingale famous for ? [A]: the newly named `` corner shop '' where asprey has traded since <num> occupies what were once nine Georgian houses ,one of which was the last london home of nursing pioneer florence nightingale . [A]: in <num> , the founder of modern nursing , florence nightingale , was born in florence , italy .

挑战 答案类型方面 特殊类问题方面(包含Non-factoid) how far is yaroslavl from moscow yaroslavl , <num> miles northeast of moscow How often….occurs once every <num> years Why is XXX famous How is cataract treated ?  he had surgery for cataracts . 类型 答案 style of music rap artists industry chemical ethnic background jewish kind of business the clothing retailer profession ophthalmologist

小结 问题:答案句子选择 评测:数据集QASent 方法: Tree Match Word Alignment Deep Learning 挑战 谓词方面 答案类型方面 特殊类问题求解方面

参考文献 Vasin Punyakanok, Dan Roth, and Wen-Tau Yih. Mapping dependencies trees: An application to question answering. In Proceedings of the 8th International Symposium on Artificial Intelligence and Mathematics, Fort Lauderdale, FL, USA. 2004. Hang Cui, Renxu Sun, Keya Li, Min-Yen Kan, and Tat-Seng Chua. Question answering passage retrieval using dependency relations. In ACM-SIGIR 2005. Heilman, Michael and Smith, Noah A. 2010. Tree Edit Models for Recognizing Textual Entailments, Paraphrases, and Answers to Questions. In NAACL-HLT 2010. Shnarch, Eyal, Ido Dagan, and Jacob Goldberger. A Probabilistic Lexical Model for Ranking Textual Inferences //Proceedings of the First Joint Conference on Lexical and Computational Semantics 2012. Yih, Wen-tau and Chang, Ming-Wei and Meek, Christopher and Pastusiak, Andrzej. 2013. Question Answering Using Enhanced Lexical Semantic Models. In ACL 2013. Di Wang and Eric Nyberg. 2015. A Long Short-Term Memory Model for Answer Sentence Selection in Question Answering. In ACL 2015. More… http://aclweb.org/aclwiki/index.php?title=Question_Answering_(State_of_the_art)

致谢 欢迎老师和同学提问!

示例 Q: Which was the first movie that James Dean was in? S: James Dean, who began as an actor on TV dramas, didn’t make his screen debut until 1951’s “Fixed Bayonet.” Q: What was the GE building in rockefeller plaza called before ? S: Known as the RCA Building until 1988 , it is most famous for housing the headquarters of the television network NBC . Q: how long was I love lucy on the air ? S: The black-and-white series originally ran from October 15, 1951, to May 6, 1957, on the Columbia Broadcasting System ( CBS ) Q: In what film is Gordon Gekko the main character? S: He received a best actor Oscar in 1987 for his role as Gordon Gekko in “Wall Street”. Q: What is the name of Durst’s group? S: Limp Bizkit lead singer Fred Durst did a lot before he hit the big time.