<?xml version="1.0" encoding="UTF-8"?><xml><records><record><source-app name="Biblio" version="7.x">Drupal-Biblio</source-app><ref-type>47</ref-type><contributors><authors><author><style face="normal" font="default" size="100%">Lin, J. F. S.</style></author><author><style face="normal" font="default" size="100%">Bonnet, V.</style></author><author><style face="normal" font="default" size="100%">Joukov, V.</style></author><author><style face="normal" font="default" size="100%">Venture, G.</style></author><author><style face="normal" font="default" size="100%">Kulić, D.</style></author></authors></contributors><titles><title><style face="normal" font="default" size="100%">Comparison of Kinematic and Dynamic Sensor Modalities and Derived Features for Human Motion Segmentation</style></title><secondary-title><style face="normal" font="default" size="100%">Proceedings of the IEEE/NIH Strategic Conference on Healthcare Innovations and Point of Care Technologies</style></secondary-title></titles><dates><year><style  face="normal" font="default" size="100%">2016</style></year></dates><pages><style face="normal" font="default" size="100%">109–112</style></pages><language><style face="normal" font="default" size="100%">eng</style></language></record><record><source-app name="Biblio" version="7.x">Drupal-Biblio</source-app><ref-type>47</ref-type><contributors><authors><author><style face="normal" font="default" size="100%">Lin, J. F. S.</style></author><author><style face="normal" font="default" size="100%">Bonnet, V.</style></author><author><style face="normal" font="default" size="100%">Panchea, A. M.</style></author><author><style face="normal" font="default" size="100%">Ramdani, N.</style></author><author><style face="normal" font="default" size="100%">Venture, G.</style></author><author><style face="normal" font="default" size="100%">Kulić, D.</style></author></authors></contributors><titles><title><style face="normal" font="default" size="100%">Human Motion Segmentation using Cost Weights Recovered from Inverse Optimal Control</style></title><secondary-title><style face="normal" font="default" size="100%">Proceedings of the IEEE/RAS International Conference on Humanoid Robots</style></secondary-title></titles><dates><year><style  face="normal" font="default" size="100%">2016</style></year></dates><pages><style face="normal" font="default" size="100%">1107–1113</style></pages><language><style face="normal" font="default" size="100%">eng</style></language></record><record><source-app name="Biblio" version="7.x">Drupal-Biblio</source-app><ref-type>17</ref-type><contributors><authors><author><style face="normal" font="default" size="100%">Lin, J. F. S.</style></author><author><style face="normal" font="default" size="100%">Karg, M. E.</style></author><author><style face="normal" font="default" size="100%">Kulić, D.</style></author></authors></contributors><titles><title><style face="normal" font="default" size="100%">Movement Primitive Segmentation for Human Motion Modeling: A Framework for Analysis</style></title><secondary-title><style face="normal" font="default" size="100%">IEEE Transactions on Human-Machine Systems</style></secondary-title></titles><keywords><keyword><style  face="normal" font="default" size="100%">Algorithm design and analysis</style></keyword><keyword><style  face="normal" font="default" size="100%">algorithm mechanics</style></keyword><keyword><style  face="normal" font="default" size="100%">application-specific requirements</style></keyword><keyword><style  face="normal" font="default" size="100%">Cameras</style></keyword><keyword><style  face="normal" font="default" size="100%">classification algorithms</style></keyword><keyword><style  face="normal" font="default" size="100%">Data collection</style></keyword><keyword><style  face="normal" font="default" size="100%">data sources</style></keyword><keyword><style  face="normal" font="default" size="100%">Databases</style></keyword><keyword><style  face="normal" font="default" size="100%">distance-based methods</style></keyword><keyword><style  face="normal" font="default" size="100%">exercise monitoring</style></keyword><keyword><style  face="normal" font="default" size="100%">gesture recognition</style></keyword><keyword><style  face="normal" font="default" size="100%">human motion modeling</style></keyword><keyword><style  face="normal" font="default" size="100%">human movement observation data sequences</style></keyword><keyword><style  face="normal" font="default" size="100%">human-machine interaction</style></keyword><keyword><style  face="normal" font="default" size="100%">image motion analysis</style></keyword><keyword><style  face="normal" font="default" size="100%">image segmentation</style></keyword><keyword><style  face="normal" font="default" size="100%">machine learning algorithms</style></keyword><keyword><style  face="normal" font="default" size="100%">Manuals</style></keyword><keyword><style  face="normal" font="default" size="100%">Motion segmentation</style></keyword><keyword><style  face="normal" font="default" size="100%">movement identification</style></keyword><keyword><style  face="normal" font="default" size="100%">movement learning</style></keyword><keyword><style  face="normal" font="default" size="100%">movement modeling</style></keyword><keyword><style  face="normal" font="default" size="100%">movement primitive segmentation</style></keyword><keyword><style  face="normal" font="default" size="100%">physiology</style></keyword><keyword><style  face="normal" font="default" size="100%">robot imitation learning</style></keyword><keyword><style  face="normal" font="default" size="100%">segment definitions</style></keyword><keyword><style  face="normal" font="default" size="100%">segmentation algorithms</style></keyword><keyword><style  face="normal" font="default" size="100%">stochastic dynamic models</style></keyword><keyword><style  face="normal" font="default" size="100%">time series analysis</style></keyword><keyword><style  face="normal" font="default" size="100%">validation techniques</style></keyword></keywords><dates><year><style  face="normal" font="default" size="100%">2016</style></year></dates><volume><style face="normal" font="default" size="100%">46</style></volume><pages><style face="normal" font="default" size="100%">325–339</style></pages><language><style face="normal" font="default" size="100%">eng</style></language><abstract><style face="normal" font="default" size="100%">Movement primitive segmentation enables long sequences of human movement observation data to be segmented into smaller components, termed movement primitives, to facilitate movement identification, modeling, and learning. It has been applied to exercise monitoring, gesture recognition, human-machine interaction, and robot imitation learning. This paper proposes a segmentation framework to categorize and compare different segmentation algorithms considering segment definitions, data sources, application-specific requirements, algorithm mechanics, and validation techniques. The framework is applied to human motion segmentation methods by grouping them into online, semionline, and offline approaches. Among the online approaches, distance-based methods provide the best performance, while stochastic dynamic models work best in the semionline and offline settings. However, most algorithms to date are tested with small datasets, and algorithm generalization across participants and to movement changes remains largely untested.</style></abstract></record><record><source-app name="Biblio" version="7.x">Drupal-Biblio</source-app><ref-type>47</ref-type><contributors><authors><author><style face="normal" font="default" size="100%">Lin, J. F. S.</style></author><author><style face="normal" font="default" size="100%">Samadani, A.</style></author><author><style face="normal" font="default" size="100%">Kulić, D.</style></author></authors></contributors><titles><title><style face="normal" font="default" size="100%">Segmentation by Data Point Classification Applied to Forearm Surface EMG</style></title><secondary-title><style face="normal" font="default" size="100%">Proceedings of the EAI International Summit of Smart City 360</style></secondary-title></titles><dates><year><style  face="normal" font="default" size="100%">2016</style></year></dates><publisher><style face="normal" font="default" size="100%">Springer International Publishing</style></publisher><volume><style face="normal" font="default" size="100%">166</style></volume><pages><style face="normal" font="default" size="100%">153–165</style></pages><language><style face="normal" font="default" size="100%">eng</style></language></record><record><source-app name="Biblio" version="7.x">Drupal-Biblio</source-app><ref-type>47</ref-type><contributors><authors><author><style face="normal" font="default" size="100%">Lin, J. F. S.</style></author><author><style face="normal" font="default" size="100%">Joukov, V.</style></author><author><style face="normal" font="default" size="100%">Kulić, D.</style></author></authors></contributors><titles><title><style face="normal" font="default" size="100%">Full-body Multi-primitive Segmentation using Classifiers</style></title><secondary-title><style face="normal" font="default" size="100%">Proceedings of the IEEE/RAS International Conference on Humanoid Robots</style></secondary-title></titles><dates><year><style  face="normal" font="default" size="100%">2014</style></year></dates><pages><style face="normal" font="default" size="100%">874–880</style></pages><language><style face="normal" font="default" size="100%">eng</style></language></record><record><source-app name="Biblio" version="7.x">Drupal-Biblio</source-app><ref-type>47</ref-type><contributors><authors><author><style face="normal" font="default" size="100%">Lin, J. F. S.</style></author><author><style face="normal" font="default" size="100%">Joukov, V.</style></author><author><style face="normal" font="default" size="100%">Kulić, D.</style></author></authors></contributors><titles><title><style face="normal" font="default" size="100%">Human Motion Segmentation by Data Point Classification</style></title><secondary-title><style face="normal" font="default" size="100%">Proceedings of the International Conference of the IEEE Engineering in Medicine and Biology Society</style></secondary-title></titles><dates><year><style  face="normal" font="default" size="100%">2014</style></year></dates><pages><style face="normal" font="default" size="100%">9–13</style></pages><language><style face="normal" font="default" size="100%">eng</style></language></record><record><source-app name="Biblio" version="7.x">Drupal-Biblio</source-app><ref-type>17</ref-type><contributors><authors><author><style face="normal" font="default" size="100%">Lin, J. F. S.</style></author><author><style face="normal" font="default" size="100%">Kulić, D.</style></author></authors></contributors><titles><title><style face="normal" font="default" size="100%">On-line Segmentation of Human Motion for Automated Rehabilitation Exercise Analysis</style></title><secondary-title><style face="normal" font="default" size="100%">IEEE Transactions on Neural Systems and Rehabilitation Engineering</style></secondary-title></titles><dates><year><style  face="normal" font="default" size="100%">2014</style></year></dates><volume><style face="normal" font="default" size="100%">22</style></volume><pages><style face="normal" font="default" size="100%">168–180</style></pages><language><style face="normal" font="default" size="100%">eng</style></language><abstract><style face="normal" font="default" size="100%">To enable automated analysis of rehabilitation movements, an approach for accurately identifying and segmenting movement repetitions is required. This paper proposes an approach for on-line, automated segmentation and identification of movement segments from continuous time-series data of human movement. The proposed approach uses a two stage identification and recognition process, based on velocity features and stochastic modeling of each motion to be identified. In the first stage, motion segment candidates are identified based on a characteristic sequence of velocity features such as velocity peaks and zero velocity crossings. In the second stage, Hidden Markov models are used to accurately identify segment locations from the identified candidates. The proposed approach is capable of on-line segmentation and identification, enabling interactive feedback in rehabilitation applications. The approach is validated on a 20 person rehabilitation movement dataset.</style></abstract></record><record><source-app name="Biblio" version="7.x">Drupal-Biblio</source-app><ref-type>47</ref-type><contributors><authors><author><style face="normal" font="default" size="100%">Lin, J. F. S.</style></author><author><style face="normal" font="default" size="100%">Kulić, D.</style></author></authors></contributors><titles><title><style face="normal" font="default" size="100%">Human Pose Recovery for Rehabilitation Using Ambulatory Sensors</style></title><secondary-title><style face="normal" font="default" size="100%">Proceedings of the International Conference of the IEEE Engineering in Medicine and Biology Society</style></secondary-title></titles><dates><year><style  face="normal" font="default" size="100%">2013</style></year></dates><pub-location><style face="normal" font="default" size="100%">Osaka, Japan</style></pub-location><pages><style face="normal" font="default" size="100%">4799–4802</style></pages><language><style face="normal" font="default" size="100%">eng</style></language></record><record><source-app name="Biblio" version="7.x">Drupal-Biblio</source-app><ref-type>47</ref-type><contributors><authors><author><style face="normal" font="default" size="100%">Zhang, T.</style></author><author><style face="normal" font="default" size="100%">Karg, M. E.</style></author><author><style face="normal" font="default" size="100%">Lin, J. F. S.</style></author><author><style face="normal" font="default" size="100%">Kulić, D.</style></author><author><style face="normal" font="default" size="100%">Venture, G.</style></author></authors></contributors><titles><title><style face="normal" font="default" size="100%">IMU based Single Stride Identification of Humans</style></title><secondary-title><style face="normal" font="default" size="100%">Proceedings of the IEEE International Symposium on Robot and Human Interactive Communication</style></secondary-title></titles><dates><year><style  face="normal" font="default" size="100%">2013</style></year></dates><pages><style face="normal" font="default" size="100%">220-225</style></pages><language><style face="normal" font="default" size="100%">eng</style></language></record><record><source-app name="Biblio" version="7.x">Drupal-Biblio</source-app><ref-type>32</ref-type><contributors><authors><author><style face="normal" font="default" size="100%">Lin, J. F. S.</style></author></authors></contributors><titles><title><style face="normal" font="default" size="100%">Automated Rehabilitation Exercise Motion Tracking</style></title></titles><dates><year><style  face="normal" font="default" size="100%">2012</style></year></dates><urls><web-urls><url><style face="normal" font="default" size="100%">hdl.handle.net/10012/7191</style></url></web-urls></urls><publisher><style face="normal" font="default" size="100%">University of Waterloo</style></publisher><language><style face="normal" font="default" size="100%">eng</style></language><abstract><style face="normal" font="default" size="100%">Current physiotherapy practice relies on visual observation of the patient for diagnosis and assessment. The assessment process can potentially be automated to improve accuracy and reliability. This thesis proposes a method to recover patient joint angles and automatically extract movement profiles utilizing small and lightweight body-worn sensors. Joint angles are estimated from sensor measurements via the extended Kalman filter (EKF). Constant-acceleration kinematics is employed as the state evolution model. The forward kinematics of the body is utilized as the measurement model. The state and measurement models are used to estimate the position, velocity and acceleration of each joint, updated based on the sensor inputs from inertial measurement units (IMUs). Additional joint limit constraints are imposed to reduce drift, and an automated approach is developed for estimating and adapting the process noise during on-line estimation. Once joint angles are determined, the exercise data is segmented to identify each of the repetitions. This process of identifying when a particular repetition begins and ends allows the physiotherapist to obtain useful metrics such as the number of repetitions performed, or the time required to complete each repetition. A feature-guided hidden Markov model (HMM) based algorithm is developed for performing the segmentation. In a sequence of unlabelled data, motion segment candidates are found by scanning the data for velocity-based features, such as velocity peaks and zero crossings, which match the pre-determined motion templates. These segment potentials are passed into the HMM for template matching. This two-tier approach combines the speed of a velocity feature based approach, which only requires the data to be differentiated, with the accuracy of the more computationally-heavy HMM, allowing for fast and accurate segmentation. The proposed algorithms were verified experimentally on a dataset consisting of 20 healthy subjects performing rehabilitation exercises. The movement data was collected by IMUs strapped onto the hip, thigh and calf. The joint angle estimation system achieves an overall average RMS error of 4.27 cm, when compared against motion capture data. The segmentation algorithm reports 78% accuracy when the template training data comes from the same participant, and 74% for a generic template.</style></abstract><work-type><style face="normal" font="default" size="100%">masters</style></work-type></record><record><source-app name="Biblio" version="7.x">Drupal-Biblio</source-app><ref-type>17</ref-type><contributors><authors><author><style face="normal" font="default" size="100%">Lin, J. F. S.</style></author><author><style face="normal" font="default" size="100%">Kulić, D.</style></author></authors></contributors><titles><title><style face="normal" font="default" size="100%">Human Pose Recovery using Wireless Inertial Measurement Units</style></title><secondary-title><style face="normal" font="default" size="100%">Physiological Measurement</style></secondary-title></titles><dates><year><style  face="normal" font="default" size="100%">2012</style></year></dates><volume><style face="normal" font="default" size="100%">33</style></volume><pages><style face="normal" font="default" size="100%">2099–2115</style></pages><language><style face="normal" font="default" size="100%">eng</style></language></record><record><source-app name="Biblio" version="7.x">Drupal-Biblio</source-app><ref-type>47</ref-type><contributors><authors><author><style face="normal" font="default" size="100%">Lin, J. F. S.</style></author><author><style face="normal" font="default" size="100%">Kulić, D.</style></author></authors></contributors><titles><title><style face="normal" font="default" size="100%">Segmenting Human Motion for Automated Rehabilitation Exercise Analysis</style></title><secondary-title><style face="normal" font="default" size="100%">Proceedings of the International Conference of the IEEE Engineering in Medicine and Biology Society</style></secondary-title></titles><dates><year><style  face="normal" font="default" size="100%">2012</style></year></dates><pub-location><style face="normal" font="default" size="100%">San Diego, USA</style></pub-location><pages><style face="normal" font="default" size="100%">2881–2884</style></pages><language><style face="normal" font="default" size="100%">eng</style></language><abstract><style face="normal" font="default" size="100%">This paper proposes an approach for the automated segmentation and identification of movement segments from continuous time series data of human movement, collected through motion capture of ambulatory sensors. The proposed approach uses a two stage identification and recognition process, based on velocity and stochastic modeling of each motion to be identified. In the first stage, motion segment candidates are identified based on a unique sequence of velocity features such as velocity peaks and zero velocity crossings. In the second stage, Hidden Markov models are used to accurately identify segment locations from the identified candidates. The approach is capable of on-line segmentation and identification, enabling interactive feedback in rehabilitation applications. The approach is validated on a rehabilitation movement dataset, and achieves a segmentation accuracy of 89%.</style></abstract></record><record><source-app name="Biblio" version="7.x">Drupal-Biblio</source-app><ref-type>47</ref-type><contributors><authors><author><style face="normal" font="default" size="100%">Lin, J. F. S.</style></author><author><style face="normal" font="default" size="100%">Kulić, D.</style></author></authors></contributors><titles><title><style face="normal" font="default" size="100%">Automatic Human Motion Segmentation and Identification using Feature Guided HMM for Physical Rehabilitation Exercises</style></title><secondary-title><style face="normal" font="default" size="100%">Proceedings of the IEEE International Conference on Intelligent Robots and Systems, Workshop on Robotics for Neurology and Rehabilitation</style></secondary-title></titles><keywords><keyword><style  face="normal" font="default" size="100%">motion primitive</style></keyword><keyword><style  face="normal" font="default" size="100%">Motion segmentation</style></keyword></keywords><dates><year><style  face="normal" font="default" size="100%">2011</style></year></dates><pub-location><style face="normal" font="default" size="100%">San Francisco, CA</style></pub-location><pages><style face="normal" font="default" size="100%">33–36</style></pages><language><style face="normal" font="default" size="100%">eng</style></language><abstract><style face="normal" font="default" size="100%">Fast and accurate motion segmentation and identification methods are required to enable real-time assessment and feedback for physical rehabilitation. Exercise motions exhibit cyclic patterns that can be characterized by simple features, such as zero-velocity crossings or velocity peaks. In this paper, these features are used as framing windows for simultaneous motion segmentation and identification via Hidden Markov models. Comparisons to other segmentation methods show that feature guiding increases the segmentation accuracy and greatly reduces the runtime needed to perform the segmentation.</style></abstract></record><record><source-app name="Biblio" version="7.x">Drupal-Biblio</source-app><ref-type>47</ref-type><contributors><authors><author><style face="normal" font="default" size="100%">Lin, J. F. S.</style></author><author><style face="normal" font="default" size="100%">Kulić, D.</style></author></authors></contributors><titles><title><style face="normal" font="default" size="100%">Motion Segmentation with Inertial Measurements using Feature Guided HMM</style></title><secondary-title><style face="normal" font="default" size="100%">Proceedings of the Symposium on Advanced Intelligent Systems</style></secondary-title></titles><dates><year><style  face="normal" font="default" size="100%">2011</style></year></dates><language><style face="normal" font="default" size="100%">eng</style></language><abstract><style face="normal" font="default" size="100%">Motion segmentation and identification methods tend to assume that joint angle data are readily available for processing. The standard data acquisition method, motion capture, is not feasible for real-world deployment, such as in physical rehabilitation clinics, due to space, environment and cost constraints. In this paper, a feature-guided Hidden Markov model segmentation algorithm is adapted to segment and identify human motion from acceleration and gyroscopic data instead of position data. Acceleration and gyroscopic data are readily obtainable through small lightweight sensors, and would serve as a more appropriate sensor system for a rehabilitation clinic than a motion capture system.</style></abstract></record></records></xml>