Medicine image classification using deep learning: highlighting the MedNet-MoBiL hybrid model
Deep learning has transformed image classification tasks across many domains, including medical diagnostics. Medicine wrappers, boxes, and strips often contain valuable but complex information that can be difficult to read and comprehend manually. This complexity drives users to seek additional know...
Saved in:
Published in | Discover Computing Vol. 28; no. 1; p. 103 |
---|---|
Main Authors | , , , , , |
Format | Journal Article |
Language | English |
Published |
Dordrecht
Springer Nature B.V
01.12.2025
|
Subjects | |
Online Access | Get full text |
Cover
Loading…
Summary: | Deep learning has transformed image classification tasks across many domains, including medical diagnostics. Medicine wrappers, boxes, and strips often contain valuable but complex information that can be difficult to read and comprehend manually. This complexity drives users to seek additional knowledge online. However, traditional search engines often present a large volume of results, requiring users to manually filter through multiple links to find relevant information, which can be time-consuming. To address this issue, we propose a lightweight pipeline that leverages MobileNetV2 for image classification and Optical Character Recognition (OCR) for extracting text content from medicine packaging. The extracted text is processed using the RAKE algorithm to identify significant keywords, which are then matched with relevant URLs through a Google Search API. To ensure relevance, retrieved links are ranked using ROUGE scores. Performance metrics demonstrate the model's efficiency, with ROUGE-1 achieving 90% Recall, 95% F1-Score, and 90% Accuracy, and ROUGE-L achieving 83% Recall, 91% F1-Score, and 83% Accuracy. The pipeline was trained and validated on a curated dataset of 3,000 real-world medicine packaging images, publicly available on GitHub. These results highlight the novelty and practicality of our solution for automating medical information retrieval from packaging, using an interpretable and scalable deep learning-driven approach. |
---|---|
Bibliography: | ObjectType-Article-1 SourceType-Scholarly Journals-1 ObjectType-Feature-2 content type line 14 |
ISSN: | 2948-2992 1386-4564 2948-2992 1573-7659 |
DOI: | 10.1007/s10791-025-09619-w |