서브메뉴
검색
Beyond the Black Box: Optimization Within Latent Spaces.
Beyond the Black Box: Optimization Within Latent Spaces.
- 자료유형
- 학위논문
- Control Number
- 0017163503
- International Standard Book Number
- 9798384053736
- Dewey Decimal Classification Number
- 004
- Main Entry-Personal Name
- Kishore, Varsha.
- Publication, Distribution, etc. (Imprint
- [S.l.] : Cornell University., 2024
- Publication, Distribution, etc. (Imprint
- Ann Arbor : ProQuest Dissertations & Theses, 2024
- Physical Description
- 193 p.
- General Note
- Source: Dissertations Abstracts International, Volume: 86-03, Section: B.
- General Note
- Advisor: Weinberger, Kilian.
- Dissertation Note
- Thesis (Ph.D.)--Cornell University, 2024.
- Summary, Etc.
- 요약In the past decade, neural networks have evolved into extraordinarily powerful tools, with wide-ranging applications across many different domains. These models allow us to use immense computational power to learn low-level features and high-level abstract concepts from vast datasets.Neural networks embed data of different forms (text, image, audio, etc.) into high dimensional latent spaces that encode salient features of the data and capture complex relationships between data points. This thesis aims to probe model parameters and latent spaces-to understand not only how information is stored and processed in networks but also how the encoded knowledge can be extracted and harnessed. We leverage these insights to develop novel methods that optimize specific parameters or representations from trained models to perform various downstream tasks. We present three specific methods in this thesis. First, we introduce BERTScore, an algorithm that utilizes representations from pre-trained language models to measure the similarity between two pieces of text. BERTScore approximates a form of transport distance to match tokens in the texts. Then, we focus on an information retrieval setting, where transformers are trained end-to-end to map search queries to corresponding documents. In this setting, we introduce IncDSI, a method to add new documents to a trained retrieval system by solving a constrained convex optimization problem to obtain new document representations. Finally, we present Fixed Neural Network Steganography (FNNS), a technique for image steganography that hides information by exploiting a neural network's sensitivity to imperceptible perturbations.
- Subject Added Entry-Topical Term
- Computer science.
- Subject Added Entry-Topical Term
- Computer engineering.
- Subject Added Entry-Topical Term
- Systems science.
- Subject Added Entry-Topical Term
- Information technology.
- Index Term-Uncontrolled
- BERTScore
- Index Term-Uncontrolled
- Fixed Neural Network Steganography
- Index Term-Uncontrolled
- Image steganography
- Index Term-Uncontrolled
- Neural networks
- Index Term-Uncontrolled
- Data points
- Added Entry-Corporate Name
- Cornell University Computer Science
- Host Item Entry
- Dissertations Abstracts International. 86-03B.
- Electronic Location and Access
- 로그인을 한후 보실 수 있는 자료입니다.
- Control Number
- joongbu:658541