研究目的
To explore the use of deep transfer learning based on the VGG-16 network for accurately and automatically classifying age-related macular degeneration (AMD) and diabetic macular edema (DME) in optical coherence tomography (OCT) images to assist in medical diagnosis and reduce clinician workload.
研究成果
The deep transfer learning method based on the VGG-16 network demonstrates high effectiveness in classifying retinal OCT images with a small dataset, achieving accuracy comparable to human experts. It provides a fully automated approach for pre-diagnosis, reducing clinician workload and offering potential applications in clinics for assisting diagnostic decisions. Future work should focus on translating this approach into practical software for medical use.
研究不足
The pre-trained VGG-16 network cannot offer transparent interpretations of results, limiting its use for treatment instructions without clinical support. Performance is slightly inferior to training a CNN from scratch with a large dataset, which is difficult due to data collection challenges and longer training times.
1:Experimental Design and Method Selection:
The study employed a deep transfer learning approach by fine-tuning the pre-trained VGG-16 network on the ImageNet dataset for classifying retinal OCT images into categories such as choroidal neovascularization (CNV), diabetic macular edema (DME), drusen, and normal. The method involved image preprocessing, dataset division, and network retraining with backpropagation.
2:Sample Selection and Data Sources:
A total of 207,130 retinal OCT images from 5319 adult patients were retrospectively collected from multiple institutions (Shiley Eye Institute, California Retinal Research Foundation, Medical Center Ophthalmology Associates, Shanghai First People’s Hospital, Beijing Tongren Eye Center) between 2013 and 2017. After preprocessing, 109,312 images were used, with 1000 images (250 per category) from 633 patients as the validation dataset and the rest from 4686 patients as the training dataset.
3:After preprocessing, 109,312 images were used, with 1000 images (250 per category) from 633 patients as the validation dataset and the rest from 4686 patients as the training dataset.
List of Experimental Equipment and Materials:
3. List of Experimental Equipment and Materials: The primary equipment included an SD-OCT device (Heidelberg Spectralis, Heidelberg Engineering, Germany) for image acquisition. Computational resources included an Ubuntu 16.04 system with Intel Core i7-2700K CPU, 256 GB RAM, Dual AMD Filepro flash storage, and NVIDIA GTX 1080 GPU. Software tools included Python with numpy and scikit-learn modules.
4:04 system with Intel Core i7-2700K CPU, 256 GB RAM, Dual AMD Filepro flash storage, and NVIDIA GTX 1080 GPU. Software tools included Python with numpy and scikit-learn modules.
Experimental Procedures and Operational Workflow:
4. Experimental Procedures and Operational Workflow: Images were acquired, labeled by retinal specialists, preprocessed (normalization and resizing to 224x224 pixels), and divided into datasets. The VGG-16 network was fine-tuned by initializing convolutional layers with pre-trained weights and retraining fully connected layers for four output classes. Training was performed using SGD optimizer with a learning rate of 0.001, momentum of 0.9, weight decay of 10^-4, and batches of 1000 images over 200 epochs.
5:001, momentum of 9, weight decay of 10^-4, and batches of 1000 images over 200 epochs.
Data Analysis Methods:
5. Data Analysis Methods: Performance was evaluated using accuracy, sensitivity, specificity, and receiver-operating characteristic (ROC) curves. Statistical analysis was conducted using Python with numpy and scikit-learn, calculating metrics based on true positives, false positives, true negatives, and false negatives.
独家科研数据包,助您复现前沿成果,加速创新突破
获取完整内容