본문

서브메뉴

Improving Articulated Pose Tracking and Contact Force Estimation for Qualitative Assessment of Human Actions.
Improving Articulated Pose Tracking and Contact Force Estimation for Qualitative Assessmen...
Contents Info
Improving Articulated Pose Tracking and Contact Force Estimation for Qualitative Assessment of Human Actions.
Material Type  
 학위논문
 
0017162762
Date and Time of Latest Transaction  
20250211152051
ISBN  
9798382738338
DDC  
004
Author  
Louis, Nathan.
Title/Author  
Improving Articulated Pose Tracking and Contact Force Estimation for Qualitative Assessment of Human Actions.
Publish Info  
[S.l.] : University of Michigan., 2024
Publish Info  
Ann Arbor : ProQuest Dissertations & Theses, 2024
Material Info  
121 p.
General Note  
Source: Dissertations Abstracts International, Volume: 85-12, Section: B.
General Note  
Advisor: Corso, Jason J.;Owens, Andrew.
학위논문주기  
Thesis (Ph.D.)--University of Michigan, 2024.
Abstracts/Etc  
요약Using video to automate human performance metrics or skill analysis is an important but under-explored task. Currently, measuring the quality of an action can be highly subjective, where even assessments from experts are affected by bias and inter-rater reliability. In contrast, Computer vision and AI have the potential to provide real-time non-intrusive solutions with increased objectivity, scalability, and repeatability across various domains. From video alone, we can automatically provide supplemental objective scoring of Olympic sports, evaluate the technical skill of surgeons for training purposes, or monitor the physical rehabilitation progress of a patient. Today we solve these problems with supervised learning, obtaining features that represent high correlation with our desired point of analysis. Supervised learning is powerful, data-driven, and sometimes the best available option. However alone, it may be sub-optimal in the presence of scarce data and insufficient when needed to generalize to varying conditions or to truly understand the target task.In this dissertation, the bases of our human analysis understanding are skeletal poses, namely hand poses and full body poses. For articulated hand poses, we improve tracking using our CondPose network to integrate prior detection confidences and encourage tracking consistency. While for human poses, we propose two physical simulation-based metrics for evaluating physical plausibility and perform external force estimation through predicted ground reaction forces (GRFs). However, in the human analysis domain, collecting and annotating data at the scale of other deep learning tasks is a recurring challenge. This limits our generalizability to different environments, procedures, and motions. We address this by exploring semi-supervised learning methods, such as contrastive pre-training and multi-task learning.We apply articulated hand pose tracking in the surgical environment for assessing surgical skill. By applying a time-shifted sampling augmentation, we introduce clip-contrastive pre-training on embedded hand features as an unsupervised learning step. We show that this contrastive pre-training improves performance when fine-tuned on surgical skill classification and assessment task. Unlike most prior work, we evaluate on open surgery videos rather than solely simulated environments. Specifically, we use videos of non-laparoscopic, collected through collaboration with the Cardiac Surgery department at Michigan Medicine. We use full body poses and contact force estimation to bridge the gap between visual observations and the physical world. This physically-grounded component is vital for understanding actions involving sports or physical rehab where humans interact with their environment. We leverage multi-task learning to perform 2D-to-3D human pose estimation and integrate other abundant sources of motion capture data, without requiring additional force plate supervision. Our experiments shows that this improves GRF estimation on unseen motions. To address data limitations, we also collect two novel datasets SurgicalHands and ForcePose. We use SurgicalHands in the surgical domain as a multi-instance articulated hand pose tracking dataset. It encompasses a high degree of complexity in appearance and movement, not present in prior datasets. ForcePose is a multi-view GRF dataset of tracked human poses and time-synchronized force plates, to our knowledge the largest and most varied of its kind. This dataset serves as a benchmark for mapping human body motion and physical forces, enabling physical grounding of specific actions.
Subject Added Entry-Topical Term  
Computer science.
Subject Added Entry-Topical Term  
Electrical engineering.
Subject Added Entry-Topical Term  
Surgery.
Subject Added Entry-Topical Term  
Bioinformatics.
Index Term-Uncontrolled  
CondPose network
Index Term-Uncontrolled  
Physical rehabilitation
Index Term-Uncontrolled  
Multi-task learning
Index Term-Uncontrolled  
SurgicalHands
Index Term-Uncontrolled  
ForcePose
Added Entry-Corporate Name  
University of Michigan Electrical and Computer Engineering
Host Item Entry  
Dissertations Abstracts International. 85-12B.
Electronic Location and Access  
로그인을 한후 보실 수 있는 자료입니다.
Control Number  
joongbu:658647
New Books MORE
최근 3년간 통계입니다.

Detail Info.

  • Reservation
  • 캠퍼스간 도서대출
  • 서가에 없는 책 신고
  • My Folder
Material
Reg No. Call No. Location Status Lend Info
TQ0034965 T   원문자료 열람가능/출력가능 열람가능/출력가능
마이폴더 부재도서신고

* Reservations are available in the borrowing book. To make reservations, Please click the reservation button

해당 도서를 다른 이용자가 함께 대출한 도서

Related books

Related Popular Books

도서위치