1
|
Jagadesh BN, Mantena SV, Sathe AP, Prabhakara Rao T, Lella KK, Pabboju SS, Vatambeti R. Enhancing food recognition accuracy using hybrid transformer models and image preprocessing techniques. Sci Rep 2025; 15:5591. [PMID: 39955332 PMCID: PMC11829996 DOI: 10.1038/s41598-025-90244-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/22/2024] [Accepted: 02/11/2025] [Indexed: 02/17/2025] Open
Abstract
This study presents a robust approach for continuous food recognition essential for nutritional research, leveraging advanced computer vision techniques. The proposed method integrates Mutually Guided Image Filtering (MuGIF) to enhance dataset quality and minimize noise, followed by feature extraction using the Visual Geometry Group (VGG) architecture for intricate visual analysis. A hybrid transformer model, combining Vision Transformer and Swin Transformer variants, is introduced to capitalize on their complementary strengths. Hyperparameter optimization is performed using the Improved Discrete Bat Algorithm (IDBA), resulting in a highly accurate and efficient classification system. Experimental results highlight the superior performance of the proposed model, achieving a classification accuracy of 99.83%, significantly outperforming existing methods. This study underscores the potential of hybrid transformer architectures and advanced preprocessing techniques in advancing food recognition systems, offering enhanced accuracy and efficiency for practical applications in dietary monitoring and personalized nutrition recommendations.
Collapse
Affiliation(s)
- B N Jagadesh
- School of Computer Science and Engineering, VIT-AP University, Vijayawada, India
| | - Srihari Varma Mantena
- Department of Computer Science and Engineering, SRKR Engineering College, Bhimavaram, 534204, India
| | - Asha P Sathe
- Department of Computer Engineering, Army Institute of Technology, Pune, India
| | - T Prabhakara Rao
- Department of Computer Science and Engineering, Aditya University, Surampalem, India
| | - Kranthi Kumar Lella
- School of Computer Science and Engineering, VIT-AP University, Vijayawada, India
| | - Shyam Sunder Pabboju
- Department of Computer Science and Engineering, Mahatma Gandhi Institute of Technology, Hyderabad, India
| | - Ramesh Vatambeti
- School of Computer Science and Engineering, VIT-AP University, Vijayawada, India.
| |
Collapse
|
2
|
Zhao Z, Wang R, Liu M, Bai L, Sun Y. Application of machine vision in food computing: A review. Food Chem 2025; 463:141238. [PMID: 39368204 DOI: 10.1016/j.foodchem.2024.141238] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/05/2024] [Revised: 09/03/2024] [Accepted: 09/09/2024] [Indexed: 10/07/2024]
Abstract
With global intelligence advancing and the awareness of sustainable development growing, artificial intelligence technology is increasingly being applied to the food industry. This paper, grounded in practical application cases, reviews the current research status and prospects of machine vision-based image recognition technology in food computing. It explores the general workflow of image recognition, applications based on traditional machine learning and deep learning methods. The paper covers areas including food safety detection, dietary nutrition analysis, process monitoring, and enterprise management model optimization. The aim is to provide a solid theoretical foundation and technical guidance for the integration and cross-fertilization of the food industry with artificial intelligence technology.
Collapse
Affiliation(s)
- Zhiyao Zhao
- School of Computer and Artificial Intelligence, School of Light Industry Science and Engineering, Beijing Technology and Business University, Beijing 100048, China.
| | - Rong Wang
- School of Computer and Artificial Intelligence, School of Light Industry Science and Engineering, Beijing Technology and Business University, Beijing 100048, China.
| | - Minghao Liu
- School of Computer and Artificial Intelligence, School of Light Industry Science and Engineering, Beijing Technology and Business University, Beijing 100048, China.
| | - Lin Bai
- School of Computer and Artificial Intelligence, School of Light Industry Science and Engineering, Beijing Technology and Business University, Beijing 100048, China.
| | - Ying Sun
- School of Computer and Artificial Intelligence, School of Light Industry Science and Engineering, Beijing Technology and Business University, Beijing 100048, China.
| |
Collapse
|
3
|
Crystal AA, Valero M, Nino V, Ingram KH. Empowering Diabetics: Advancements in Smartphone-Based Food Classification, Volume Measurement, and Nutritional Estimation. SENSORS (BASEL, SWITZERLAND) 2024; 24:4089. [PMID: 39000868 PMCID: PMC11244259 DOI: 10.3390/s24134089] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 04/23/2024] [Revised: 06/06/2024] [Accepted: 06/14/2024] [Indexed: 07/16/2024]
Abstract
Diabetes has emerged as a worldwide health crisis, affecting approximately 537 million adults. Maintaining blood glucose requires careful observation of diet, physical activity, and adherence to medications if necessary. Diet monitoring historically involves keeping food diaries; however, this process can be labor-intensive, and recollection of food items may introduce errors. Automated technologies such as food image recognition systems (FIRS) can make use of computer vision and mobile cameras to reduce the burden of keeping diaries and improve diet tracking. These tools provide various levels of diet analysis, and some offer further suggestions for improving the nutritional quality of meals. The current study is a systematic review of mobile computer vision-based approaches for food classification, volume estimation, and nutrient estimation. Relevant articles published over the last two decades are evaluated, and both future directions and issues related to FIRS are explored.
Collapse
Affiliation(s)
- Afnan Ahmed Crystal
- Department of Computer Science, Kennesaw State University, Kennesaw, GA 30060, USA
| | - Maria Valero
- Department of Information Technology, Kennesaw State University, Kennesaw, GA 30060, USA
| | - Valentina Nino
- Departement of Industrial and Systems Engineering, Kennesaw State University, Kennesaw, GA 30060, USA
| | - Katherine H Ingram
- Department of Exercise Science and Sport Management, Kennesaw State University, Kennesaw, GA 30060, USA
| |
Collapse
|
4
|
Lin N, Zhou X, Chen W, He C, Wang X, Wei Y, Long Z, Shen T, Zhong L, Yang C, Dai T, Zhang H, Shi H, Ma X. Development and validation of a point-of-care nursing mobile tool to guide the diagnosis of malnutrition in hospitalized adult patients: a multicenter, prospective cohort study. MedComm (Beijing) 2024; 5:e526. [PMID: 38606361 PMCID: PMC11006711 DOI: 10.1002/mco2.526] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/11/2023] [Revised: 03/07/2024] [Accepted: 03/10/2024] [Indexed: 04/13/2024] Open
Abstract
Malnutrition is a prevalent and severe issue in hospitalized patients with chronic diseases. However, malnutrition screening is often overlooked or inaccurate due to lack of awareness and experience among health care providers. This study aimed to develop and validate a novel digital smartphone-based self-administered tool that uses facial features, especially the ocular area, as indicators of malnutrition in inpatient patients with chronic diseases. Facial photographs and malnutrition screening scales were collected from 619 patients in four different hospitals. A machine learning model based on back propagation neural network was trained, validated, and tested using these data. The model showed a significant correlation (p < 0.05) and a high accuracy (area under the curve 0.834-0.927) in different patient groups. The point-of-care mobile tool can be used to screen malnutrition with good accuracy and accessibility, showing its potential for screening malnutrition in patients with chronic diseases.
Collapse
Affiliation(s)
- Nan Lin
- Department of BiotherapyCancer CenterWest China Hospital, Sichuan UniversityChengduChina
| | - Xueyan Zhou
- Department of BiotherapyState Key Laboratory of Biotherapy, Frontiers Science Center for Disease‐related Molecular Network, West China Hospital, and Key Laboratory of Bio‐Resource and Eco‐Environment of Ministry of Education, College of Life Sciences, Sichuan UniversityChengduSichuanChina
| | - Weichang Chen
- State Key Laboratory of Oral DiseasesNational Clinical Research Center for Oral Diseases, Sichuan UniversityChengduChina
| | | | - Xiaoxuan Wang
- Department of BiotherapyCancer CenterWest China Hospital, Sichuan UniversityChengduChina
| | - Yuhao Wei
- Department of BiotherapyCancer CenterWest China Hospital, Sichuan UniversityChengduChina
| | | | - Tao Shen
- Department of Colorectal SurgeryThe Third Affiliated Hospital of Kunming Medical University/Yunnan Tumor HospitalKunmingChina
| | - Lingyu Zhong
- Department of Clinical NutritionHospital of Chengdu Office of People’s Government of Tibetan Autonomous RegionChengduChina
| | - Chan Yang
- Division of Endocrinology and MetabolismState Key Laboratory of Biotherapy, West China Hospital, Sichuan UniversityChengduChina
| | - Tingting Dai
- Department of Clinical NutritionWest China Hospital, Sichuan UniversityChengduChina
| | - Hao Zhang
- Division of Pancreatic SurgeryDepartment of General SurgeryWest China Hospital, Sichuan UniversityChengduChina
| | - Hubing Shi
- Laboratory of Integrative MedicineClinical Research Center for Breast, State Key Laboratory of Biotherapy, West China Hospital, Sichuan University and Collaborative Innovation CenterChengduSichuanChina
| | - Xuelei Ma
- Department of BiotherapyCancer CenterWest China Hospital, Sichuan UniversityChengduChina
| |
Collapse
|
5
|
Shonkoff E, Cara KC, Pei X(A, Chung M, Kamath S, Panetta K, Hennessy E. AI-based digital image dietary assessment methods compared to humans and ground truth: a systematic review. Ann Med 2023; 55:2273497. [PMID: 38060823 PMCID: PMC10836267 DOI: 10.1080/07853890.2023.2273497] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 03/24/2023] [Accepted: 10/16/2023] [Indexed: 12/18/2023] Open
Abstract
OBJECTIVE Human error estimating food intake is a major source of bias in nutrition research. Artificial intelligence (AI) methods may reduce bias, but the overall accuracy of AI estimates is unknown. This study was a systematic review of peer-reviewed journal articles comparing fully automated AI-based (e.g. deep learning) methods of dietary assessment from digital images to human assessors and ground truth (e.g. doubly labelled water). MATERIALS AND METHODS Literature was searched through May 2023 in four electronic databases plus reference mining. Eligible articles reported AI estimated volume, energy, or nutrients. Independent investigators screened articles and extracted data. Potential sources of bias were documented in absence of an applicable risk of bias assessment tool. RESULTS Database and hand searches identified 14,059 unique publications; fifty-two papers (studies) published from 2010 to 2023 were retained. For food detection and classification, 79% of papers used a convolutional neural network. Common ground truth sources were calculation using nutrient tables (51%) and weighed food (27%). Included papers varied widely in food image databases and results reported, so meta-analytic synthesis could not be conducted. Relative errors were extracted or calculated from 69% of papers. Average overall relative errors (AI vs. ground truth) ranged from 0.10% to 38.3% for calories and 0.09% to 33% for volume, suggesting similar performance. Ranges of relative error were lower when images had single/simple foods. CONCLUSIONS Relative errors for volume and calorie estimations suggest that AI methods align with - and have the potential to exceed - accuracy of human estimations. However, variability in food image databases and results reported prevented meta-analytic synthesis. The field can advance by testing AI architectures on a limited number of large-scale food image and nutrition databases that the field determines to be adequate for training and testing and by reporting accuracy of at least absolute and relative error for volume or calorie estimations.
Collapse
Affiliation(s)
- Eleanor Shonkoff
- School of Health Sciences, Merrimack College, North Andover, MA, USA
| | - Kelly Copeland Cara
- Friedman School of Nutrition Science and Policy, Tufts University, Boston, MA, USA
| | - Xuechen (Anna) Pei
- Friedman School of Nutrition Science and Policy, Tufts University, Boston, MA, USA
| | - Mei Chung
- Friedman School of Nutrition Science and Policy, Tufts University, Boston, MA, USA
| | - Shreyas Kamath
- School of Engineering, Tufts University, Medford, MA, USA
| | - Karen Panetta
- School of Engineering, Tufts University, Medford, MA, USA
| | - Erin Hennessy
- Friedman School of Nutrition Science and Policy, Tufts University, Boston, MA, USA
- ChildObesity180, Tufts University, Boston, MA, USA
| |
Collapse
|
6
|
Hiraguchi H, Perone P, Toet A, Camps G, Brouwer AM. Technology to Automatically Record Eating Behavior in Real Life: A Systematic Review. SENSORS (BASEL, SWITZERLAND) 2023; 23:7757. [PMID: 37765812 PMCID: PMC10534458 DOI: 10.3390/s23187757] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/14/2023] [Revised: 09/04/2023] [Accepted: 09/05/2023] [Indexed: 09/29/2023]
Abstract
To monitor adherence to diets and to design and evaluate nutritional interventions, it is essential to obtain objective knowledge about eating behavior. In most research, measures of eating behavior are based on self-reporting, such as 24-h recalls, food records (food diaries) and food frequency questionnaires. Self-reporting is prone to inaccuracies due to inaccurate and subjective recall and other biases. Recording behavior using nonobtrusive technology in daily life would overcome this. Here, we provide an up-to-date systematic overview encompassing all (close-to) publicly or commercially available technologies to automatically record eating behavior in real-life settings. A total of 1328 studies were screened and, after applying defined inclusion and exclusion criteria, 122 studies were included for in-depth evaluation. Technologies in these studies were categorized by what type of eating behavior they measure and which type of sensor technology they use. In general, we found that relatively simple sensors are often used. Depending on the purpose, these are mainly motion sensors, microphones, weight sensors and photo cameras. While several of these technologies are commercially available, there is still a lack of publicly available algorithms that are needed to process and interpret the resulting data. We argue that future work should focus on developing robust algorithms and validating these technologies in real-life settings. Combining technologies (e.g., prompting individuals for self-reports at sensed, opportune moments) is a promising route toward ecologically valid studies of eating behavior.
Collapse
Affiliation(s)
- Haruka Hiraguchi
- TNO Human Factors, Netherlands Organization for Applied Scientific Research, Kampweg 55, 3769 DE Soesterberg, The Netherlands (A.-M.B.)
- Kikkoman Europe R&D Laboratory B.V., Nieuwe Kanaal 7G, 6709 PA Wageningen, The Netherlands
| | - Paola Perone
- TNO Human Factors, Netherlands Organization for Applied Scientific Research, Kampweg 55, 3769 DE Soesterberg, The Netherlands (A.-M.B.)
| | - Alexander Toet
- TNO Human Factors, Netherlands Organization for Applied Scientific Research, Kampweg 55, 3769 DE Soesterberg, The Netherlands (A.-M.B.)
| | - Guido Camps
- Division of Human Nutrition and Health, Wageningen University & Research, Stippeneng 4, 6708 WE Wageningen, The Netherlands
- OnePlanet Research Center, Plus Ultra II, Bronland 10, 6708 WE Wageningen, The Netherlands
| | - Anne-Marie Brouwer
- TNO Human Factors, Netherlands Organization for Applied Scientific Research, Kampweg 55, 3769 DE Soesterberg, The Netherlands (A.-M.B.)
- Department of Artificial Intelligence, Radboud University, Thomas van Aquinostraat 4, 6525 GD Nijmegen, The Netherlands
| |
Collapse
|