Publications by authors named "Shaokai Ye"

Article Synopsis
  • Keypoint tracking algorithms can analyze animal movement in various settings but struggle to convert continuous keypoint data into distinct actions due to noise and jitter.
  • Keypoint-MoSeq is a machine learning platform that autonomously identifies behavioral modules (or 'syllables') from keypoint data by distinguishing useful behavior from noise.
  • This platform outperforms traditional clustering methods in recognizing behavioral transitions, correlating behavior with neural activity, and is applicable across different species and behaviors from video recordings.
View Article and Find Full Text PDF

Quantification of behavior is critical in diverse applications from neuroscience, veterinary medicine to animal conservation. A common key step for behavioral analysis is first extracting relevant keypoints on animals, known as pose estimation. However, reliable inference of poses currently requires domain knowledge and manual labeling effort to build supervised models.

View Article and Find Full Text PDF

Keypoint tracking algorithms have revolutionized the analysis of animal behavior, enabling investigators to flexibly quantify behavioral dynamics from conventional video recordings obtained in a wide variety of settings. However, it remains unclear how to parse continuous keypoint data into the modules out of which behavior is organized. This challenge is particularly acute because keypoint data is susceptible to high frequency jitter that clustering algorithms can mistake for transitions between behavioral modules.

View Article and Find Full Text PDF
Article Synopsis
  • Estimating animal pose in computer vision is tricky due to occlusions and the high similarity between animals, making it hard to identify individual animals during interactions.
  • We enhance DeepLabCut, an existing pose estimation tool, by adding features for tracking multiple animals and predicting their identities, which helps during occlusions.
  • Our framework is tested on four diverse datasets that we are releasing to support future research in this area.
View Article and Find Full Text PDF

Large deep neural network (DNN) models pose the key challenge to energy efficiency due to the significantly higher energy consumption of off-chip DRAM accesses than arithmetic or SRAM operations. It motivates the intensive research on model compression with two main approaches. Weight pruning leverages the redundancy in the number of weights and can be performed in a non-structured, which has higher flexibility and pruning rate but incurs index accesses due to irregular weights, or structured manner, which preserves the full matrix structure with a lower pruning rate.

View Article and Find Full Text PDF

Weight pruning methods of deep neural networks (DNNs) have been demonstrated to achieve a good model pruning rate without loss of accuracy, thereby alleviating the significant computation/storage requirements of large-scale DNNs. Structured weight pruning methods have been proposed to overcome the limitation of irregular network structure and demonstrated actual GPU acceleration. However, in prior work, the pruning rate (degree of sparsity) and GPU acceleration are limited (to less than 50%) when accuracy needs to be maintained.

View Article and Find Full Text PDF