Constrained, Causal, and Knowledge-Grounded Reasoning for Neural Language Generation- [electronic resource]
Constrained, Causal, and Knowledge-Grounded Reasoning for Neural Language Generation- [electronic resource]
- 자료유형
- 학위논문파일 국외
- 최종처리일시
- 20240214101236
- ISBN
- 9798379912741
- DDC
- 620
- 저자명
- Qin, Lianhui.
- 서명/저자
- Constrained, Causal, and Knowledge-Grounded Reasoning for Neural Language Generation - [electronic resource]
- 발행사항
- [S.l.]: : University of Washington., 2023
- 발행사항
- Ann Arbor : : ProQuest Dissertations & Theses,, 2023
- 형태사항
- 1 online resource(162 p.)
- 주기사항
- Source: Dissertations Abstracts International, Volume: 85-01, Section: B.
- 주기사항
- Advisor: Choi, Yejin.
- 학위논문주기
- Thesis (Ph.D.)--University of Washington, 2023.
- 사용제한주기
- This item must not be sold to any third party vendors.
- 초록/해제
- 요약This thesis aims to establish a connection between reasoning and language generation. Today's language models (LMs, such as GPT-3), despite producing human-like fluent text, essentially act like "a mouth without a brain" -- They generate without grounding on the world knowledge, and lack the ability to flexibly reason about everyday situations and events, including counterfactual ("what if?") and abductive ("what might explain the observations?") reasoning. This thesis bridges the gap from three angles: (1) Differentiable reasoning with constraints: Humans can incorporate any constraints from the context on the fly and conduct reasoning in new situations without the need of specific training. I develop a unified inference framework that endows the LMs with the flexibility and efficiency, through a differentiable process to reason over the vast space of discrete language, combined with arbitrary neural and symbolic constraints; (2) Counterfactual and nonmonotonic reasoning in natural language: I establish the first formulation of counterfactual reasoning in language, and used my inference tool to enable the common monotonic LMs for the capabilities of nonmonotonic reasoning ranging from counterfactual, abductive, and temporal reasoning in complex context; (3) Integration of knowledge and logic in neural language models: I develop mechanisms of integrating rich external knowledge and structures with the neural LMs, to ground and boost the reasoning abilities.
- 일반주제명
- Engineering.
- 일반주제명
- Computer science.
- 키워드
- Text generation
- 기타저자
- University of Washington Computer Science and Engineering
- 기본자료저록
- Dissertations Abstracts International. 85-01B.
- 기본자료저록
- Dissertation Abstract International
- 전자적 위치 및 접속
- 로그인 후 원문을 볼 수 있습니다.