Redirigiendo al acceso original de articulo en 16 segundos...
Inicio  /  Algorithms  /  Vol: 12 Par: 10 (2019)  /  Artículo
ARTÍCULO
TITULO

Can People Really Do Nothing? Handling Annotation Gaps in ADL Sensor Data

Alaa E. Abdel Hakim and Wael Deabes    

Resumen

In supervised Activities of Daily Living (ADL) recognition systems, annotating collected sensor readings is an essential, yet exhaustive, task. Readings are collected from activity-monitoring sensors in a 24/7 manner. The size of the produced dataset is so huge that it is almost impossible for a human annotator to give a certain label to every single instance in the dataset. This results in annotation gaps in the input data to the adopting learning system. The performance of the recognition system is negatively affected by these gaps. In this work, we propose and investigate three different paradigms to handle these gaps. In the first paradigm, the gaps are taken out by dropping all unlabeled readings. A single ?Unknown? or ?Do-Nothing? label is given to the unlabeled readings within the operation of the second paradigm. The last paradigm handles these gaps by giving every set of them a unique label identifying the encapsulating certain labels. Also, we propose a semantic preprocessing method of annotation gaps by constructing a hybrid combination of some of these paradigms for further performance improvement. The performance of the proposed three paradigms and their hybrid combination is evaluated using an ADL benchmark dataset containing more than 2.5×106" role="presentation" style="position: relative;">2.5×1062.5×106 2.5 × 10 6 sensor readings that had been collected over more than nine months. The evaluation results emphasize the performance contrast under the operation of each paradigm and support a specific gap handling approach for better performance.

 Artículos similares