研究目的
Investigating the effectiveness of RGB-D static gesture recognition using a fine-tuned Inception V3 convolutional neural network to eliminate the need for gesture segmentation and feature extraction in traditional algorithms.
研究成果
The proposed RGB-D static gesture recognition method based on fine-tuning Inception V3 achieves a high accuracy of 91.35% on the ASL dataset, outperforming traditional machine learning methods and other CNN models. The method effectively utilizes depth information to enhance recognition performance, particularly for similar gestures.
研究不足
The study is limited to static gesture recognition and does not address dynamic gestures. The performance under varying lighting and background conditions, although improved, may still face challenges in extreme scenarios.
1:Experimental Design and Method Selection:
The study employs a fine-tuned Inception V3 model for RGB-D static gesture recognition, utilizing a two-stage training strategy to adapt the model to the classification task.
2:Sample Selection and Data Sources:
The ASL dataset, collected using Kinect, is used, comprising RGB and depth images of 24 English letters' gestures.
3:List of Experimental Equipment and Materials:
The experiment uses a GPU GTX 1060, with Keras as the framework and TensorFlow as the backend.
4:Experimental Procedures and Operational Workflow:
The process includes data cleaning, depth image processing, data augmentation, and a two-stage training strategy for model fine-tuning.
5:Data Analysis Methods:
The performance is evaluated based on recognition accuracy, with comparisons made against traditional machine learning methods and other CNN models.
独家科研数据包,助您复现前沿成果,加速创新突破
获取完整内容