서브메뉴
검색
Pushing Transport Layer Latency Down Towards its Physical Limits in Data Centers with Programmable Architectures and Algorithms.
Pushing Transport Layer Latency Down Towards its Physical Limits in Data Centers with Programmable Architectures and Algorithms.
- 자료유형
- 학위논문
- Control Number
- 0017162190
- International Standard Book Number
- 9798384201717
- Dewey Decimal Classification Number
- 001
- Main Entry-Personal Name
- Arslan, Serhat.
- Publication, Distribution, etc. (Imprint
- [S.l.] : Stanford University., 2024
- Publication, Distribution, etc. (Imprint
- Ann Arbor : ProQuest Dissertations & Theses, 2024
- Physical Description
- 129 p.
- General Note
- Source: Dissertations Abstracts International, Volume: 86-03, Section: B.
- General Note
- Advisor: McKeown, Nick;Katti, Sachin.
- Dissertation Note
- Thesis (Ph.D.)--Stanford University, 2024.
- Summary, Etc.
- 요약Data center applications keep scaling horizontally across many machines to accommodate more users and data. This makes the communication performance requirements even more stringent, i.e., higher bandwidth and lower latency. The increasing link capacities address the bandwidth demands, but the latency requirements necessitate more sophisticated solutions.In this thesis, I observe that the transport layer is the only layer in the networking stack to impact latency both at the end-hosts and the network. The way it handles packets sets the end-hosts processing delay. And its congestion control determines the queuing delay in the network. Hence, I study transport layer designs to push both latencies down to their physical limits.First, I argue that end-host latency can be minimized by offloading the transport layer to NIC hardware, but fixed-function chips prohibit custom solutions for diversified environments. As a solution, I introduce nanoTransport, a programmable NIC architecture for message-based Remote Procedure Calls. It is programmed using the P4 language, making it easy to modify (or create) transport protocols while the packets are processed orders of magnitude faster than traditional software stacks. It identifies common events and primitive operations for a streamlined, modular, and programmable pipeline; including packetization, reassembly, timeouts, and packet generation, all expressed by the programmer.Next, I argue that network latency can only be minimized with quick and accurate congestion control decisions, which require precise congestion signals and the shortest control loop delay. I present Bolt to address these requirements and push congestion control to its theoretical limits. Bolt is based on three core ideas, (I) Sub-RTT Control (SRC) reacts to congestion faster than one RTT, (II) Proactive Ramp-up (PRU) foresees flow completions to promptly occupy released bandwidth, and (III) Supply matching (SM) matches bandwidth demand with supply to maximize utilization. I show that these mechanisms reduce 99th-p latency by 80% and improve 99th-p flow completion time by up to 3X compared to Swift and HPCC even at 400Gb/s.
- Subject Added Entry-Topical Term
- Software.
- Subject Added Entry-Topical Term
- Control algorithms.
- Subject Added Entry-Topical Term
- Response time.
- Subject Added Entry-Topical Term
- Protocol.
- Subject Added Entry-Topical Term
- Communication.
- Subject Added Entry-Topical Term
- Queuing.
- Subject Added Entry-Topical Term
- Design.
- Subject Added Entry-Topical Term
- Network interface cards.
- Subject Added Entry-Topical Term
- Web studies.
- Added Entry-Corporate Name
- Stanford University.
- Host Item Entry
- Dissertations Abstracts International. 86-03B.
- Electronic Location and Access
- 로그인을 한후 보실 수 있는 자료입니다.
- Control Number
- joongbu:657918
Info Détail de la recherche.
- Réservation
- 캠퍼스간 도서대출
- 서가에 없는 책 신고
- My Folder