With the emergence of high-resolution fingerprint sensors, there has been a lot of focus on level-3 fingerprint features, especially the pores, for the next generation automated fingerprint recognition systems (AFRS). Following the success of deep learning in various computer vision tasks, researchers have developed learning-based approaches for detection of pores in high-resolution fingerprint images. Generally, learning-based approaches provide better performance than handcrafted feature-based approaches. However, domain adaptability of the existing learning-based pore detection methods has never been studied. In this paper, we study this aspect and propose an approach for pore detection in cross-sensor scenarios. For this purpose, we have generated an in-house 1000 dpi fingerprint dataset with ground truth pore coordinates (referred to as IITI-HRFP-GT), and evaluated the performance of the existing learning-based pore detection approaches. The core of the proposed approach for detection of pores in cross-sensor scenarios is DeepDomainPore, which is a residual learning-based convolutional neural network(CNN) trained for pore detection. The domain adaptability in DeepDomainPore is achieved by embedding a gradient reversal layer between the CNN and a domain classifier network. The proposed approach achieves state-of-the-art performance in a cross-sensor scenario involving public high-resolution fingerprint datasets with 88.12% true detection rate and 83.82% F-score.
Detecting when eating occurs is an essential step toward automatic dietary monitoring, medication adherence assessment, and diet-related health interventions. Wearable technologies play a central role in designing unubtrusive diet monitoring solutions by leveraging machine learning algorithms that work on time-series sensor data to detect eating moments. While much research has been done on developing activity recognition and eating moment detection algorithms, the performance of the detection algorithms drops substantially when the model trained with one user is utilized by a new user. To facilitate development of personalized models, we propose PALS, Proximity-based Active Learning on Streaming data, a novel proximity-based model for recognizing eating gestures with the goal of significantly decreasing the need for labeled data with new users. Particularly, we propose an optimization problem to perform active learning under limited query budget by leveraging unlabeled data. Our extensive analysis on data collected in both controlled and uncontrolled settings indicates that the F-score of PLAS ranges from 22% to 39% for a budget that varies from 10 to 60 query. Furthermore, compared to the state-of-the-art approaches, off-line PALS, on average, achieves to 40% higher recall and 12\% higher f-score in detecting eating gestures.