본문

서브메뉴

Training and Architecting Sequence to Sequence Language Models for Applications in Varied Domains
Содержание
Training and Architecting Sequence to Sequence Language Models for Applications in Varied Domains
자료유형  
 학위논문
Control Number  
0015490885
International Standard Book Number  
9781085558693
Dewey Decimal Classification Number  
004
Main Entry-Personal Name  
Li, Congrui.
Publication, Distribution, etc. (Imprint  
[Sl] : Rensselaer Polytechnic Institute, 2019
Publication, Distribution, etc. (Imprint  
Ann Arbor : ProQuest Dissertations & Theses, 2019
Physical Description  
142 p
General Note  
Source: Dissertations Abstracts International, Volume: 81-02, Section: B.
General Note  
Advisor: Fox, Peter.
Dissertation Note  
Thesis (Ph.D.)--Rensselaer Polytechnic Institute, 2019.
Restrictions on Access Note  
This item must not be sold to any third party vendors.
Restrictions on Access Note  
This item must not be added to any third party search indexes.
Summary, Etc.  
요약Lots of challenges exist while dealing with language text sequence data directly on document level. The sequence-to-sequence (seq2seq) model is an ideal tool for this task. A basic sequence-to-sequence model consists of two recurrent networks: an encoder that processes the input and a decoder that generates the output. To allow the decoder's more direct access to the input, an attention mechanism was introduced by researchers so that the decoder can peek into the input at every decoding step. To improve the long-term dependencies, more sophisticated neuron cell structures, such as Long Short-Term Memory and Gated Recurrent Unit, were also developed by researchers. The task of Neural Machine Translation was the very first testbed for seq2seq models with wild success, and then followed by the task of chatbot applications in various domains.혻This thesis introduces three innovative case studies using variants of seq2seq model, and each of them focuses on a different stage of the model's training process. The first case study focuses on the stage before the training of seq2seq model. We introduce a generative chatbot in Chinese language trained with data on a finer level of granularity. Based on the evaluation of A/B testing results by multiple human evaluators, we conclude that the character-level model can still maintain the performance of the word-level benchmark.The second case study focuses on the stage during the training of seq2seq model. We introduce an unsupervised information retrieval (IR) model using sequence autoencoder which is competitive with multiple existing techniques, including Jaccard similarity, bag-of-words cosine similarity, tf-idf cosine similarity, as well as the recent neural network approaches such as Doc2Vec and Skip-Thoughts. The third case study focuses on the stage after the training of seq2seq model. We explore mergers and acquisitions in the domain of business analytics. We further demonstrate the effectiveness of the IR model in the previous case study for measuring business proximity, and also investigate the capability of the IR model's output as pre-trained input for a downstream supervised task, to prediction acquisitions. For the subsequent task, we compare the variations of models with two different types of inputs as well as three different types of network structure. Sophisticated data preprocessing techniques are carried out for each experiment to improve the quality of the training data. Bidirectional seq2seq models with GRU cells and Luong attention are used for all tasks.In conclusion, research is conducted before, during, and after the training of seq2seq model so that improvements or discoveries are made in each case study to more effectively encode natural language text sequence data at the document level to obtain responses/answers/trends for various training corpora.
Subject Added Entry-Topical Term  
Computer science
Added Entry-Corporate Name  
Rensselaer Polytechnic Institute Computer Science
Host Item Entry  
Dissertations Abstracts International. 81-02B.
Host Item Entry  
Dissertation Abstract International
Electronic Location and Access  
로그인을 한후 보실 수 있는 자료입니다.
Control Number  
joongbu:566855
New Books MORE
최근 3년간 통계입니다.

Подробнее информация.

  • Бронирование
  • 캠퍼스간 도서대출
  • 서가에 없는 책 신고
  • моя папка
материал
Reg No. Количество платежных Местоположение статус Ленд информации
TQ0006876 T   원문자료 열람가능/출력가능 열람가능/출력가능
마이폴더 부재도서신고

* Бронирование доступны в заимствований книги. Чтобы сделать предварительный заказ, пожалуйста, нажмите кнопку бронирование

해당 도서를 다른 이용자가 함께 대출한 도서

Related books

Related Popular Books

도서위치