Please use this identifier to cite or link to this item: http://theses.iitj.ac.in:8080/jspui/handle/123456789/197
Full metadata record
DC FieldValueLanguage
dc.contributor.advisorHarit, Gaurav-
dc.date.accessioned2022-05-26T13:17:13Z-
dc.date.available2022-05-26T13:17:13Z-
dc.date.issued2020-07-
dc.identifier.citationJain, Hitesh. (2020). Assessment of Human Actions in Videos (Doctor’s thesis). Indian Institute of Technology Jodhpur, Jodhpur.en_US
dc.identifier.urihttp://theses.iitj.ac.in:8080/jspui/handle/123456789/197-
dc.description.abstractHumans have the desire to achieve better performance and outcome of their efforts. In performances such as sports, yoga or professions such as surgery, humans strive to attain perfection and efficiency. They mprove their performances by comparing themselves with others or seeking feedback from experts. In this regard the question arises here; can we make a human action assessment system which can support the users with a flexibility of using the system at preferred time and place, thereby avoiding the dependency on a trainer. The key challenge in action quality assessment is the lack of an appropriate metric as it is quite a subjective task with a significant influence of human (expert) bias. The labels like scores or skill levels provided by human judges lack interpretability. Thus, it will be interesting to develop an utomated action assessment system that can bring more interpretability and objectivity to this domain. Action assessment system can be made more objective by considering a set of reference high- cision/expert-level action videos. In this thesis, we propose techniques to assess human actions, where the problem of action assessment is transformed into the problem of comparing a given action video with a reference ideo. We consider two use-cases : Yoga and Sports, and assess actions like Sun Salutation, Diving and Gymnastic Vaults. We contribute a new Sun Salutation Assessment dataset that includes expert and non-expert performances and their respective ground truth judgments. Sun Salutation is a long term complex action with many possible errors and is a good test-bed for human action assessment techniques. Physical exercises like Sun Salutation and Aerobics are repetitive in nature. Such exercises require the performer to achieve consistent dynamic poses and smooth transitions. The quality check for such exercises involves the analysis of attributes such as pace, consistency, smoothness, etc. of the performers. We propose a framework to identify jerky and inconsistent movements in a performance that involves action segmentation using Hidden Markov Model, followed by an inter-pose timing analysis. Based on the performance skill, humans can be divided into experts, mid-level, and amateurs. Amateur performers are prone to miss or wrongly perform a part of some action. We propose a template matching-based approach where the pose sequence of the test performer is compared with an expert sequence and the action segments that were missed or anomalously performed are reported. Template-based matching is a fairly simple approach to identify differences and has difficulty in scaling when the number of templates tends to grow. In practice, actions can be performed correctly in multiples ways. Thus there is a need for comparison of the test sequence to all expert renderings possible for an action sequence. Towards this, we propose an unsupervised sequence-to-sequence autoencoder-based model that learns to reconstruct all expert videos. The skill level of a test performance is judged based on how well the learned model can reconstruct the test sequence. The closer the test performance is to an expert, the more accurately it gets reconstructed. Actions like diving or gymnastic vaults are constrained with limited availability of expert performances, which makes it difficult to train an autoencoder-based assessment model. To address this limitation, we propose a Deep Metric Learning based framework, where a Long Short Term Memory (LSTM)-based Siamese network learns to predict if two videos in a pair are similar or dissimilar based on the difference in their scores. The learned model is utilized for action scoring where the performances are compared with the reference expert performance to determine the score.en_US
dc.description.statementofresponsibilityby Hiteshi Jainen_US
dc.format.extentxvii, 106p.en_US
dc.language.isoenen_US
dc.publisherIndian Institute of Technology Jodhpuren_US
dc.rightsIIT Jodhpuren_US
dc.subject.ddcComputer Science and Engineeringen_US
dc.subject.ddcHuman action evaluationen_US
dc.titleAssessment of Human Actions in Videos.en_US
dc.typeThesisen_US
dc.creator.researcherJain, Hiteshi-
dc.date.registered2014-07-
dc.date.awarded2020-07-
dc.publisher.placeJodhpuren_US
dc.publisher.departmentComputer Science and Engineeringen_US
dc.type.degreeDoctor of Philosophyen_US
dc.format.accompanyingmaterialCDen_US
dc.description.notecol. ill.; including bibliographyen_US
dc.identifier.accessionTP00068-
Appears in Collections:Ph. D. Theses

Files in This Item:
File Description SizeFormat 
TP00068.pdf12.65 MBAdobe PDFView/Open    Request a copy
01_title.pdf140.11 kBAdobe PDFView/Open
02_declaration.pdf61.9 kBAdobe PDFView/Open
03_certificate.pdf61.88 kBAdobe PDFView/Open
04_abstract.pdf48.31 kBAdobe PDFView/Open
05_acknowledgements.pdf62.77 kBAdobe PDFView/Open
06_contents.pdf67.49 kBAdobe PDFView/Open
07_list_of_figures.pdf92.57 kBAdobe PDFView/Open
08_list_of_tables.pdf78.36 kBAdobe PDFView/Open
09_list_of_abbreviations.pdf69.87 kBAdobe PDFView/Open
10_chapter 1.pdf109.21 kBAdobe PDFView/Open
11_chapter 2.pdf1.66 MBAdobe PDFView/Open
12_chapter 3.pdf151.36 kBAdobe PDFView/Open
13_chapter 4.pdf1.89 MBAdobe PDFView/Open
14_chapter 5.pdf1.13 MBAdobe PDFView/Open
15_chapter 6.pdf1.11 MBAdobe PDFView/Open
16_chapter 7.pdf815.56 kBAdobe PDFView/Open
17_chapter 8.pdf6.21 MBAdobe PDFView/Open
18_chapter 6 conclusion and future scope.pdf118.96 kBAdobe PDFView/Open
19_list_of_publications.pdf62.81 kBAdobe PDFView/Open
20_references.pdf106.64 kBAdobe PDFView/Open


Items in IIT Jodhpur Theses Repository are protected by copyright, with all rights reserved, unless otherwise indicated.