서브메뉴
검색
Training and Architecting Sequence to Sequence Language Models for Applications in Varied Domains
Training and Architecting Sequence to Sequence Language Models for Applications in Varied Domains
- 자료유형
- 학위논문
- Control Number
- 0015490885
- International Standard Book Number
- 9781085558693
- Dewey Decimal Classification Number
- 004
- Main Entry-Personal Name
- Li, Congrui.
- Publication, Distribution, etc. (Imprint
- [Sl] : Rensselaer Polytechnic Institute, 2019
- Publication, Distribution, etc. (Imprint
- Ann Arbor : ProQuest Dissertations & Theses, 2019
- Physical Description
- 142 p
- General Note
- Source: Dissertations Abstracts International, Volume: 81-02, Section: B.
- General Note
- Advisor: Fox, Peter.
- Dissertation Note
- Thesis (Ph.D.)--Rensselaer Polytechnic Institute, 2019.
- Restrictions on Access Note
- This item must not be sold to any third party vendors.
- Restrictions on Access Note
- This item must not be added to any third party search indexes.
- Summary, Etc.
- 요약Lots of challenges exist while dealing with language text sequence data directly on document level. The sequence-to-sequence (seq2seq) model is an ideal tool for this task. A basic sequence-to-sequence model consists of two recurrent networks: an encoder that processes the input and a decoder that generates the output. To allow the decoder's more direct access to the input, an attention mechanism was introduced by researchers so that the decoder can peek into the input at every decoding step. To improve the long-term dependencies, more sophisticated neuron cell structures, such as Long Short-Term Memory and Gated Recurrent Unit, were also developed by researchers. The task of Neural Machine Translation was the very first testbed for seq2seq models with wild success, and then followed by the task of chatbot applications in various domains.혻This thesis introduces three innovative case studies using variants of seq2seq model, and each of them focuses on a different stage of the model's training process. The first case study focuses on the stage before the training of seq2seq model. We introduce a generative chatbot in Chinese language trained with data on a finer level of granularity. Based on the evaluation of A/B testing results by multiple human evaluators, we conclude that the character-level model can still maintain the performance of the word-level benchmark.The second case study focuses on the stage during the training of seq2seq model. We introduce an unsupervised information retrieval (IR) model using sequence autoencoder which is competitive with multiple existing techniques, including Jaccard similarity, bag-of-words cosine similarity, tf-idf cosine similarity, as well as the recent neural network approaches such as Doc2Vec and Skip-Thoughts. The third case study focuses on the stage after the training of seq2seq model. We explore mergers and acquisitions in the domain of business analytics. We further demonstrate the effectiveness of the IR model in the previous case study for measuring business proximity, and also investigate the capability of the IR model's output as pre-trained input for a downstream supervised task, to prediction acquisitions. For the subsequent task, we compare the variations of models with two different types of inputs as well as three different types of network structure. Sophisticated data preprocessing techniques are carried out for each experiment to improve the quality of the training data. Bidirectional seq2seq models with GRU cells and Luong attention are used for all tasks.In conclusion, research is conducted before, during, and after the training of seq2seq model so that improvements or discoveries are made in each case study to more effectively encode natural language text sequence data at the document level to obtain responses/answers/trends for various training corpora.
- Subject Added Entry-Topical Term
- Computer science
- Added Entry-Corporate Name
- Rensselaer Polytechnic Institute Computer Science
- Host Item Entry
- Dissertations Abstracts International. 81-02B.
- Host Item Entry
- Dissertation Abstract International
- Electronic Location and Access
- 로그인을 한후 보실 수 있는 자료입니다.
- Control Number
- joongbu:566855
Подробнее информация.
- Бронирование
- 캠퍼스간 도서대출
- 서가에 없는 책 신고
- моя папка