{"id":503,"date":"2026-04-10T06:37:13","date_gmt":"2026-04-10T06:37:13","guid":{"rendered":"https:\/\/nutrelino.com\/blog\/?p=503"},"modified":"2026-04-10T06:37:15","modified_gmt":"2026-04-10T06:37:15","slug":"ai-isnt-actually-counting-your-calories-why-10-problems-you-should-know-about","status":"publish","type":"post","link":"https:\/\/nutrelino.com\/blog\/en\/ai-isnt-actually-counting-your-calories-why-10-problems-you-should-know-about\/","title":{"rendered":"AI isn\u2019t actually counting your calories. Why? 10 problems you should know about"},"content":{"rendered":"\n<p>All it takes is a photo of your food, and within seconds you get calories, macros, and an \u201canalysis.\u201d<\/p>\n\n\n\n<p>Sounds simple. And more importantly, accurate.<\/p>\n\n\n\n<p>But here\u2019s the problem: AI doesn\u2019t actually calculate calories. It estimates them.<\/p>\n\n\n\n<p>It doesn\u2019t know the exact amount of ingredients, it doesn\u2019t understand how the food was prepared, and it doesn\u2019t see everything that\u2019s really in the meal. It works with what it \u201csees\u201d and even more with what it assumes is probably there based on data.<\/p>\n\n\n\n<p>And that\u2019s where the gap starts, between what looks precise and what actually is.<\/p>\n\n\n\n<p>The goal of this article isn\u2019t to criticize AI. The goal is to show its limits.<\/p>\n\n\n\n<p>Because if you understand them, you can use AI as a tool. If you don\u2019t, it can easily become a source of misleading decisions.<\/p>\n\n\n\n<p>Here are the 10 most common problems that currently affect the accuracy of AI in nutrition.<\/p>\n\n\n\n<p><\/p>\n\n\n\n<p><\/p>\n\n\n\n<h2 class=\"wp-block-heading\"><strong>1. The Illusion of the Third Dimension: Pixels Don\u2019t See Real Volume<\/strong><\/h2>\n\n\n\n<p>One of the fundamental technical limitations of AI when evaluating food from a photo is that it <strong>processes images as a two-dimensional signal.<\/strong> It does not perceive true depth or portion volume the way humans do in real space.<\/p>\n\n\n\n<p class=\"has-background\" style=\"background-color:#ffe9da\"><strong>Without depth-sensing technologies, such as LiDAR, AI cannot reliably determine whether food on a plate is spread flat or stacked vertically. As a result, it tends to systematically underestimate the amount of food and this issue becomes more pronounced as portion size increases.<\/strong><\/p>\n\n\n\n<p><\/p>\n\n\n\n<p>The data shows this quite clearly:<\/p>\n\n\n\n<ul class=\"wp-block-list\">\n<li>ChatGPT-4 underestimated food weight in 76.3% of cases<\/li>\n\n\n\n<li>The average error in weight estimation was around 36% for models like ChatGPT and Claude, while Gemini reached errors of up to 65%<\/li>\n\n\n\n<li>In one test, AI estimated a portion of curry at 255 g, while the actual weight was 480 g\u00a0 meaning the model \u201cmissed\u201d nearly half of the food<\/li>\n<\/ul>\n\n\n\n<p><\/p>\n\n\n\n<p>Importantly, this issue doesn\u2019t affect only weight. <strong>If AI misjudges volume, it will also miscalculate calories and nutrients.<\/strong> The error doesn\u2019t happen at the calculation stage it starts at the very beginning.<\/p>\n\n\n\n<p>While accuracy can be relatively acceptable for small portions, it breaks down significantly for medium and large portions (p &lt; 0.001, meaning it is extremely unlikely that this result is due to chance).<\/p>\n\n\n\n<p>In practice, this means these tools should not be used as a precise \u201cdigital scale.\u201d They can serve as a rough estimate, but they tend to underestimate reality. <strong>For users who require accurate energy and nutrient intake such as in clinical nutrition or performance-focused settings this limitation is critical.<\/strong><\/p>\n\n\n\n<p><\/p>\n\n\n\n<p><\/p>\n\n\n\n<h2 class=\"wp-block-heading\"><strong>2. Fat Blindness: Hidden Calories in Oils and Sauces<\/strong><\/h2>\n\n\n\n<p>One of the biggest limitations of AI when analyzing food from a photo is its inability to work with non-visible components. The model evaluates only what it sees on the surface, while ignoring ingredients that are absorbed into the food or hidden within its structure, such as oils, butter, or dressings. <strong>This issue is referred to as \u201cnon-visible component blindness.\u201d<\/strong><\/p>\n\n\n\n<p class=\"has-background\" style=\"background-color:#ffe9da\"><strong>From a nutritional perspective, this is a critical problem. Fat is the most energy-dense macronutrient, and even small amounts can significantly impact total calorie intake. If AI fails to detect it, the entire calculation becomes systematically underestimated.<\/strong><\/p>\n\n\n\n<p><\/p>\n\n\n\n<p>Research findings highlight this clearly:<\/p>\n\n\n\n<ul class=\"wp-block-list\">\n<li>In the ChatGPT-4o model, fat estimation error reached as high as 76.5%<\/li>\n\n\n\n<li>Even with simple foods like hazelnuts, the model misestimated fat content by approximately 75%<\/li>\n\n\n\n<li>With more complex meals, the issue becomes even more pronounced, for example, in a tuna salad, AI initially identified only about 24% of the actual fat content<\/li>\n<\/ul>\n\n\n\n<p><\/p>\n\n\n\n<p class=\"has-background\" style=\"background-color:#ffe9da\"><strong>The problem isn\u2019t that the model makes random mistakes. It simply doesn\u2019t \u201csee\u201d fat. If oil is absorbed into the food or mixed into a sauce, the model has no visual signal that it\u2019s there.<\/strong><\/p>\n\n\n\n<p><\/p>\n\n\n\n<p><strong>Interestingly, accuracy improves significantly when a text description is added to the image. When researchers included details such as \u201c2 tablespoons of oil,\u201d the accuracy of energy estimation (R\u00b2) increased from 0.59 to 0.94. This shows that the issue is not in the calculation itself, but in missing input data.<\/strong><\/p>\n\n\n\n<p>(R\u00b2&nbsp; a measure of model accuracy, increased from 0.59 to 0.94, indicating a substantial improvement in agreement with reality)<\/p>\n\n\n\n<p>In practice, this means that estimates based solely on a photo are insufficient for accurate nutrition tracking. <strong>Without information about preparation methods and added fats, actual intake can be higher by hundreds of calories per day. These tools should not be seen as a complete solution, <\/strong>but rather as a rough aid that requires manual input of key details.<\/p>\n\n\n\n<p><\/p>\n\n\n\n<p><\/p>\n\n\n\n<h2 class=\"wp-block-heading\"><strong>3. Nutritional Hallucinations: Probability of Words Instead of Real Data<\/strong><\/h2>\n\n\n\n<p>AI models in nutrition suffer from a phenomenon known as <strong>\u201cnutritional hallucinations.\u201d Unlike specialized nutrition software, they do not perform precise calculations based on real chemical databases. Instead, they generate responses based on the statistical probability of words and patterns learned from text.<\/strong><\/p>\n\n\n\n<p>This means AI does not work with actual nutritional values in real time. It produces answers that sound correct. It lacks a true understanding of relationships between nutrients, which leads to incorrect combinations, even when the output appears convincing.<\/p>\n\n\n\n<p><strong>Research shows that these errors are not rare, but repeatable.<\/strong><br>When food is misidentified, the deviations can be extreme:<\/p>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Claude 3.5 Sonnet confused scrambled eggs with pasta, resulting in a 1788% overestimation of carbohydrates<\/li>\n\n\n\n<li>Gemini 1.5 Pro identified falafel as meatballs, leading to a 360% overestimation of protein<\/li>\n<\/ul>\n\n\n\n<p><\/p>\n\n\n\n<p class=\"has-background\" style=\"background-color:#ffe9da\"><strong>Even more advanced models are not immune. ChatGPT-4 showed statistically significant inaccuracies in 10 out of 16 monitored nutrients and tended to systematically underestimate as many as 11 of them. For 13 nutrients, including potassium, fiber, and vitamin D, the deviation from reality exceeded 10%.<\/strong><\/p>\n\n\n\n<p><\/p>\n\n\n\n<p>The problem is not just the error itself, but how it is presented.<\/p>\n\n\n\n<p><strong>AI delivers outputs in a smooth, authoritative tone, often supported by tables and numbers that appear scientific. For the average user, it is practically impossible to distinguish between a precise calculation and a \u201cprobable estimate.\u201d<\/strong><\/p>\n\n\n\n<p><strong>In practice, this means AI should not be treated as a reliable calculation tool, but rather as a supportive text-based tool.<\/strong><\/p>\n\n\n\n<p>In situations where accuracy matters, especially in health conditions or clinical diets, human oversight is essential. Misinterpreted or \u201challucinated\u201d data can represent a real risk to health.<\/p>\n\n\n\n<p><\/p>\n\n\n\n<p><\/p>\n\n\n\n<h2 class=\"wp-block-heading\"><strong>4. Clinical Risk: When Inaccuracy Becomes a Health Threat<\/strong><\/h2>\n\n\n\n<p>In clinical nutrition, AI inaccuracy is no longer just a statistical deviation it becomes a real risk. <strong>Generative models are not able to apply medical guidelines with the required precision or account for individual patient limits, where exact nutrient amounts matter.<\/strong><\/p>\n\n\n\n<p class=\"has-background\" style=\"background-color:#ffe9da\"><strong>In chronic conditions such as kidney disease, diabetes, or cardiovascular disorders, even relatively small deviations can lead to a worsening of health. In these cases, the issue isn\u2019t that AI isn\u2019t perfect, it\u2019s that its errors have real consequences.<\/strong><\/p>\n\n\n\n<p><\/p>\n\n\n\n<p>Research shows that these deviations can be significant even in critical parameters:<\/p>\n\n\n\n<ul class=\"wp-block-list\">\n<li>When generating meals for dialysis patients, ChatGPT-4 underestimated:<ul><li>potassium by 49%<\/li><\/ul><ul><li>energy by 36%<\/li><\/ul>\n<ul class=\"wp-block-list\">\n<li>protein by 28%<\/li>\n<\/ul>\n<\/li>\n\n\n\n<li>The app Fastic reported sodium levels up to 34 times higher than reality<\/li>\n\n\n\n<li>Fitbit reported approximately 20 times higher iron content<\/li>\n<\/ul>\n\n\n\n<p><\/p>\n\n\n\n<p class=\"has-background\" style=\"background-color:#ffe9da\"><strong>These are not just theoretical errors. For patients who must monitor specific minerals or macronutrients, even a ~30% deviation can pose a health risk.<\/strong><\/p>\n\n\n\n<p><\/p>\n\n\n\n<p>It\u2019s also important to understand how \u201cgood\u201d results are interpreted. In one evaluation, 97% of ChatGPT\u2019s energy estimates fell within \u00b140% of USDA reference values. At first glance, this may seem like high accuracy. <strong>In practice, however, a 40% deviation means that a meal estimated at 500 kcal could actually be anywhere between 300 and 700 kcal a difference that significantly impacts any dietary plan.<\/strong><\/p>\n\n\n\n<p>The issue is not only accuracy, but also context. AI can generate recommendations that are not aligned with specific diagnoses, for example:<\/p>\n\n\n\n<ul class=\"wp-block-list\">\n<li>suggesting sugary drinks when trying to manage sugar intake<\/li>\n\n\n\n<li>presenting processed foods as \u201csuitable\u201d sources of nutrients<\/li>\n<\/ul>\n\n\n\n<p><\/p>\n\n\n\n<p><strong>In such cases, the model does not demonstrate clinical judgment it simply generates a probable answer without considering risk.<\/strong><\/p>\n\n\n\n<p>In practice, this means AI should not be used to independently manage diet in medical conditions. <strong>It can serve as a supportive tool, but final decisions must remain under professional supervision. <\/strong>Without this oversight, what appears to be a helpful assistant can become a source of inaccurate and sometimes inappropriate recommendations.<\/p>\n\n\n\n<p><\/p>\n\n\n\n<p><\/p>\n\n\n\n<h2 class=\"wp-block-heading\"><strong>5. Cultural Blindness: Algorithms Trained on a \u201cWestern\u201d Plate<\/strong><\/h2>\n\n\n\n<p>Current nutrition apps and AI models show a significant limitation known as \u201ccultural blindness.\u201d Most systems are trained primarily on Western databases, especially the U.S.-based USDA, where meals are typically represented as clearly separated components on a plate.<\/p>\n\n\n\n<p class=\"has-background\" style=\"background-color:#ffe9da\"><strong>This approach does not work well for complex, mixed, or layered dishes that are common in Asian, Mediterranean, or Middle Eastern cuisines. In these cases, AI often fails to identify individual components and their proportions, leading to significantly distorted estimates.<\/strong><\/p>\n\n\n\n<p><\/p>\n\n\n\n<p>Results from large-scale testing (Li et al., 2024) show that this is not an isolated issue, but a systematic bias:<\/p>\n\n\n\n<ul class=\"wp-block-list\">\n<li>for Western diets, apps overestimated energy intake by an average of 1040 kJ<\/li>\n\n\n\n<li>for Asian diets, they underestimated it by \u22121520 kJ (approx. \u2212360 kcal), (95% CI: \u2212874 to \u22122165 kJ)<\/li>\n<\/ul>\n\n\n\n<p><\/p>\n\n\n\n<p><strong>For specific foods, the discrepancies can be even larger. For example, AI underestimated the energy content of Pearl Milk Tea by up to 76%. For dishes like pho or stir-fry, systems often failed to correctly identify individual ingredients, resulting in highly inaccurate calculations.<\/strong><\/p>\n\n\n\n<p>These deviations are not limited to total energy, but also affect macronutrient composition:<\/p>\n\n\n\n<ul class=\"wp-block-list\">\n<li>in Western diets, some apps reported carbohydrate intake higher by 7\u20138% of total energy<\/li>\n\n\n\n<li>in Asian diets, fat intake was on average 6% higher than reference values<\/li>\n\n\n\n<li>across multiple diet types, carbohydrate intake was systematically overestimated<\/li>\n<\/ul>\n\n\n\n<p><\/p>\n\n\n\n<p>These differences highlight that models do not operate with a universal understanding of food, but rather rely on data shaped by a specific cultural context. When that context does not match the user\u2019s dietary habits, the results can be significantly skewed.<\/p>\n\n\n\n<p><strong>In practice, this means users receive data that may appear precise, but does not reflect the actual food they consumed. This issue is especially pronounced with mixed dishes, where AI cannot accurately separate components or estimate their quantities.<\/strong><\/p>\n\n\n\n<p>For this reason, relying solely on AI for analyzing complex or regional dishes without manual verification is not recommended. A more accurate approach is to search for specific foods in a database or input the meal ingredient by ingredient. Visual recognition in these cases often fails due to limited diversity in training data.<\/p>\n\n\n\n<p><\/p>\n\n\n\n<p><\/p>\n\n\n\n<h2 class=\"wp-block-heading\"><strong>6. Nutritional Imbalance: Healthy Foods in the Wrong Proportions<\/strong><\/h2>\n\n\n\n<p>At first glance, AI seems to get things right. It can build a meal plan that includes \u201chealthy\u201d foods: vegetables, yogurt, fish, whole grains. <strong>The problem is that nutrition isn\u2019t just about what you eat, but especially about the proportions in which you eat it.<\/strong><\/p>\n\n\n\n<p>And this is where a major limitation appears.<\/p>\n\n\n\n<p class=\"has-background\" style=\"background-color:#ffe9da\"><strong>AI models do not operate with a real understanding of physiology or biochemical relationships between nutrients. They do not optimize meal plans the way a professional would. Instead, they generate them based on probability, what commonly appears together, not what makes nutritional sense together.<\/strong><\/p>\n\n\n\n<p><\/p>\n\n\n\n<p><strong>The result is a meal plan that may look \u201cclean,\u201d but doesn\u2019t function properly underneath.<\/strong><\/p>\n\n\n\n<p>The data confirms this. In a study (Kaya Ka\u00e7ar et al., 2025), where AI generated 30 weight-loss meal plans (1400\u20131800 kcal), the models achieved relatively solid overall quality scores (around 71 points on the DQI-I scale). They showed sufficient variety and included all major food groups. <\/p>\n\n\n\n<p><strong>However, when nutritional balance was evaluated, meaning the ratio of macronutrients and fatty acids, the results essentially failed:<\/strong><\/p>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>the average balance score was only 0.27 out of 10<\/strong><\/li>\n\n\n\n<li><strong>ChatGPT 4.0 scored 0.0 out of 10<\/strong><\/li>\n\n\n\n<li><strong>other models were around 0.4 out of 10<\/strong><\/li>\n<\/ul>\n\n\n\n<p>In other words, AI can choose \u201cgood foods,\u201d but it cannot combine them correctly.<\/p>\n\n\n\n<p><\/p>\n\n\n\n<p>The biggest issue lies in proportions, especially between:<\/p>\n\n\n\n<ul class=\"wp-block-list\">\n<li>protein, fat, and carbohydrates<\/li>\n\n\n\n<li>different types of fats (saturated vs. unsaturated)<\/li>\n\n\n\n<li>omega-6 and omega-3 fatty acids<\/li>\n<\/ul>\n\n\n\n<p><\/p>\n\n\n\n<p><strong>These are not minor details. They are fundamental to how the body functions and directly affect inflammation, cardiovascular health, hormonal balance, and overall metabolism.<\/strong><\/p>\n\n\n\n<p>With incorrect ratios, a meal plan can look \u201chealthy on paper,\u201d but be harmful in the long term.<\/p>\n\n\n\n<p>It\u2019s also worth noting how these plans look in practice. They tend to follow repetitive patterns:<\/p>\n\n\n\n<ul class=\"wp-block-list\">\n<li>very low food variety<\/li>\n\n\n\n<li>limited protein sources (e.g., fish only as salmon or cod)<\/li>\n\n\n\n<li>complete omission of certain food groups (e.g., red meat)<\/li>\n\n\n\n<li>ignoring details like dressings or added fats<\/li>\n<\/ul>\n\n\n\n<p><\/p>\n\n\n\n<p>This suggests that models are not working with real nutritional logic, but rather with \u201csafe templates\u201d that look healthy but are not truly optimized.<\/p>\n\n\n\n<p><strong>The core issue is that AI cannot handle complexity. Creating a balanced diet requires simultaneously optimizing energy, macronutrients, micronutrients, and fat quality a combination that current models are not capable of managing.<\/strong><\/p>\n\n\n\n<p class=\"has-background\" style=\"background-color:#ffe9da\"><strong>The biggest risk is the illusion of expertise. The user sees a well-structured plan full of healthy foods and assumes it is correct. In reality, it may just be a random combination of foods without deeper nutritional logic.<\/strong><\/p>\n\n\n\n<p><\/p>\n\n\n\n<p>That\u2019s why one simple rule applies: AI can be a useful source of inspiration, but not a reliable tool for building a diet plan. Especially in weight-loss diets or medical conditions, professional oversight remains essential.<\/p>\n\n\n\n<p><\/p>\n\n\n\n<p><\/p>\n\n\n\n<h2 class=\"wp-block-heading\"><strong>7. The Large Portion Paradox: The More You Eat, the More AI Lies<\/strong><\/h2>\n\n\n\n<p>At first glance, it may seem like AI makes random errors. In reality, there is a clear and repeatable pattern: <strong>the larger the portion, the bigger the error.<\/strong><\/p>\n\n\n\n<p>AI models tend to \u201cnormalize\u201d what they see. Instead of estimating the actual volume, they gravitate toward an average version of a given dish. This may work reasonably well for small portions, but the error increases significantly with larger meals.<\/p>\n\n\n\n<p>The result is systematic underestimation.<\/p>\n\n\n\n<p class=\"has-background\" style=\"background-color:#ffe9da\"><strong>Data confirms this. In weight estimation, models like ChatGPT and Claude showed average errors around 36%, while Gemini ranged much higher, between 64% and 109%. The key point is that the error does not grow randomly, but increases directly with portion size.<\/strong><\/p>\n\n\n\n<p><\/p>\n\n\n\n<p>For small meals, accuracy was relatively good. For medium and large portions, it dropped significantly.<\/p>\n\n\n\n<p><strong>Concrete measurements show the trend clearly:<\/strong><\/p>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>small portions: 408 g vs. 430 g (minimal difference)<\/strong><\/li>\n\n\n\n<li><strong>medium portions: 580 g vs. 426 g<\/strong><\/li>\n\n\n\n<li><strong>large portions: 798 g vs. 530 g<\/strong><\/li>\n<\/ul>\n\n\n\n<p><\/p>\n\n\n\n<p>In other words: the larger the portion, the more calories \u201cdisappear.\u201d<\/p>\n\n\n\n<p>The average deviation was approximately 27.8%, and overall, AI underestimated food weight in 76.3% of cases.<\/p>\n\n\n\n<p><strong>A practical example:<br>lentil curry: AI estimated 255 g vs. the actual 480 g. Nearly half of the food and therefore calories simply \u201cdisappeared.\u201d<\/strong><\/p>\n\n\n\n<p><strong>This trend is consistent:<\/strong><\/p>\n\n\n\n<ul class=\"wp-block-list\">\n<li>accuracy for large portions is 20\u201330% lower than for small ones<\/li>\n\n\n\n<li>all models show systematic underestimation<\/li>\n\n\n\n<li>the more food on the plate, the greater the deviation<\/li>\n<\/ul>\n\n\n\n<p><\/p>\n\n\n\n<p>This is not a flaw of a specific model. It is a property of the system.<\/p>\n\n\n\n<p>The problem is that users are not aware of this error. They see a number that looks precise and naturally trust it. But if AI consistently \u201cremoves\u201d hundreds of calories from larger meals, actual intake ends up significantly higher than the data suggests.<\/p>\n\n\n\n<p class=\"has-background\" style=\"background-color:#ffe9da\"><strong>This leads to a common frustration:<br>\u201cI eat less, I track everything, but I\u2019m not losing weight.\u201d<\/strong><\/p>\n\n\n\n<p><\/p>\n\n\n\n<p>In reality, this is not a failure of the person, it\u2019s a systematic error of the tool.<\/p>\n\n\n\n<p>AI today is not a precise measurement tool. It is an estimate. And that estimate has a clear direction: underestimation, especially with larger portions.<\/p>\n\n\n\n<p>If your goal is general orientation, it may be sufficient. If your goal is accuracy, for example in weight loss, it becomes a risk.<\/p>\n\n\n\n<p>That\u2019s why one simple rule applies: when accuracy matters, a scale beats a camera.<\/p>\n\n\n\n<p><\/p>\n\n\n\n<p><\/p>\n\n\n\n<h2 class=\"wp-block-heading\"><strong>8. The Allergy Trap: When an Error Isn\u2019t Just an Error<\/strong><\/h2>\n\n\n\n<p>When it comes to food allergies, there is no room for \u201cclose enough.\u201d A meal is either safe or it isn\u2019t. <strong>And this is exactly where one of AI\u2019s most dangerous limitations becomes clear.<\/strong><\/p>\n\n\n\n<p class=\"has-background\" style=\"background-color:#ffe9da\"><strong>AI models do not operate with a medical understanding of risk. They cannot assess when an error may have real health consequences. Instead, they generate responses based on probability, what \u201csounds right\u201d, not what is actually safe.<\/strong><\/p>\n\n\n\n<p><\/p>\n\n\n\n<p><strong>This means that even if AI is given clear information about an allergy, it cannot reliably guarantee compliance.<\/strong><\/p>\n\n\n\n<p>Testing confirms this very clearly. <strong>When generating 56 meal plans for a person with food allergies, ChatGPT failed in 7% of cases.<\/strong> In practice, this means that 4 meals included an allergen that should not have been there.<\/p>\n\n\n\n<p>A specific example: in a nut-free diet, the model included almond milk without hesitation.<\/p>\n\n\n\n<p>This is not a minor inaccuracy. It is a potential health risk.<\/p>\n\n\n\n<p class=\"has-background\" style=\"background-color:#ffe9da\"><strong>Even more concerning is that AI does not recognize its own mistakes. It does not flag them, does not express uncertainty, and delivers the answer in a confident, authoritative tone.<\/strong><\/p>\n\n\n\n<p><\/p>\n\n\n\n<p>A similar issue appears with dangerously low-calorie diets. When AI was intentionally asked to create an extremely low-calorie plan, it produced no warning. Instead, it presented the plan as a valid solution, even though it could lead to serious health consequences in practice.<\/p>\n\n\n\n<p>It\u2019s important to understand that AI can sometimes produce correct answers. In some cases, it generated meal plans aligned with guidelines (e.g., for diabetes or dialysis). The problem is consistency.<\/p>\n\n\n\n<p>When the same request was repeated, the model often produced completely different and sometimes incorrect, results.<\/p>\n\n\n\n<p>This means it is not a reliable system, but a tool with high variability.<\/p>\n\n\n\n<p>The core issue is the absence of accountability. When AI makes a mistake, there is no mechanism to stop it or label it as dangerous. At the same time, it is unclear who is responsible for the consequences.<\/p>\n\n\n\n<p class=\"has-background\" style=\"background-color:#ffe9da\"><strong>For the user, this creates a false sense of safety. The response looks professional, reads smoothly, and often includes seemingly logical explanations. Without expertise, it is nearly impossible to recognize that something is wrong.<\/strong><\/p>\n\n\n\n<p><strong>With allergies, this becomes critical.<\/strong><\/p>\n\n\n\n<p>Even a small error can have serious consequences, from acute reactions to anaphylactic shock. Long-term risks also exist, such as nutritional deficiencies caused by poorly designed elimination diets.<\/p>\n\n\n\n<p><strong>That\u2019s why one simple rule applies:<br>AI can help with guidance, but it must not manage diet in medical conditions.<\/strong><\/p>\n\n\n\n<p>Especially with allergies, professional supervision should always be present. Otherwise, a seemingly helpful tool becomes a hidden risk.<\/p>\n\n\n\n<p><\/p>\n\n\n\n<p><\/p>\n\n\n\n<h2 class=\"wp-block-heading\"><strong>9. Blindness to Fortification and Brands<\/strong><\/h2>\n\n\n\n<p>AI can recognize what\u2019s on the plate. But it cannot understand what is actually in the food.<\/p>\n\n\n\n<p><strong>When analyzing a meal, it relies entirely on visual input. Anything that isn\u2019t visible on the surface effectively doesn\u2019t exist to the model. This becomes a major issue, especially with processed foods.<\/strong><\/p>\n\n\n\n<p>AI cannot identify fortification (added vitamins and minerals), nor can it distinguish between different brands unless it has access to the exact product name or visible packaging.<\/p>\n\n\n\n<p>For a human, the difference between two cereals can be significant. One may be fortified with iron and B vitamins, while the other is not. For AI, however, they are simply \u201cflakes.\u201d<\/p>\n\n\n\n<p>As a result, the model relies on average database values rather than specific data. The output may appear precise, but it is only an estimate based on a \u201ctypical version\u201d of the food.<\/p>\n\n\n\n<p>Even the models themselves acknowledge this limitation. ChatGPT-4, for example, stated that it cannot determine whether cornflakes are fortified with vitamins and minerals even though this significantly affects their nutritional profile.<\/p>\n\n\n\n<p class=\"has-background\" style=\"background-color:#ffe9da\"><strong>Data supports this quantitatively. In an analysis of 114 meals, the average deviation was approximately 26.9%. For most nutrients, the error exceeded 10%, and in 11 out of 16 cases, AI systematically underestimated values.<\/strong><\/p>\n\n\n\n<p><\/p>\n\n\n\n<p>This means that while energy and basic macronutrients may appear relatively accurate, micronutrients are often far from reality.<\/p>\n\n\n\n<p>The biggest issue arises when tracking micronutrients. A person with anemia may believe they are consuming enough iron, while in reality they are not. Similarly, for sodium, sugar, or other critical nutrients, AI can systematically distort intake and the user has no way to detect it.<\/p>\n\n\n\n<p>That\u2019s why one simple rule applies: the more processed the food, the less reliable AI analysis from a photo becomes.<\/p>\n\n\n\n<p><strong>If accuracy matters, it is essential to work with specific data:<\/strong><\/p>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>the exact product name<\/strong><\/li>\n\n\n\n<li><strong>the nutrition label<\/strong><\/li>\n\n\n\n<li><strong>or a database linked to a specific brand<\/strong><\/li>\n<\/ul>\n\n\n\n<p><\/p>\n\n\n\n<p>Without that, AI always works with averages. And in this case, \u201caverage\u201d often means a deviation that can be nutritionally significant.<\/p>\n\n\n\n<p><\/p>\n\n\n\n<p><\/p>\n\n\n\n<h2 class=\"wp-block-heading\"><strong>10. Stochastic Variability: Different Answer for the Same Photo<\/strong><\/h2>\n\n\n\n<p>Unlike a calculator or a laboratory scale, AI does not function as a precise, repeatable tool. It is a probabilistic system. This means that the same input does not always produce the same output.<\/p>\n\n\n\n<p><strong>In practice, it\u2019s simple: the same photo of a meal can return different values even when nothing about the image has changed.<\/strong><\/p>\n\n\n\n<p class=\"has-background\" style=\"background-color:#ffe9da\"><strong>The reason is what\u2019s known as stochastic variability. The model does not generate answers through fixed calculations, but by selecting the most probable output based on learned data. And this selection can vary slightly each time.<\/strong><\/p>\n\n\n\n<p><\/p>\n\n\n\n<p>The result is that AI is not fully consistent.<\/p>\n\n\n\n<p><strong>A simple example:<br>one day, AI estimates your meal at 500 kcal; the next day: using the exact same photo, it estimates 600 kcal. Not because the food changed, but because the model\u2019s output did.<\/strong><\/p>\n\n\n\n<p>A similar issue appears in diet planning. When the same request was repeated for the same health profile (e.g., a person with diabetes), the models generated different and sometimes inconsistent, recommendations. This means AI does not create a stable reference point.<\/p>\n\n\n\n<p class=\"has-background\" style=\"background-color:#ffe9da\"><strong>And that\u2019s a major problem when tracking progress. If the numbers change not because of your behavior, but because of tool variability, you lose the ability to evaluate what actually works.<\/strong><\/p>\n\n\n\n<p class=\"has-background\" style=\"background-color:#ffe9da\"><strong>In simple terms:<br>you don\u2019t know whether your body is changing or just the algorithm\u2019s response.<\/strong><\/p>\n\n\n\n<p><strong>From a scientific perspective, this is an even bigger issue. A reliable tool must be reproducible, the same input should lead to the same result. With AI, this is not currently the case.<\/strong><\/p>\n\n\n\n<p>That\u2019s why experts suggest working with ranges instead of single values. Instead of relying on one number, multiple outputs should be generated and averaged, or expressed as a confidence interval.<\/p>\n\n\n\n<p>However, this significantly complicates everyday use.<\/p>\n\n\n\n<p>The main risk, once again, is hidden. The output appears precise, clear, and definitive. The user has no reason to question it. In reality, it is just the \u201cbest current estimate,\u201d not a stable result.<\/p>\n\n\n\n<p><strong>That\u2019s why one simple rule applies:<br>AI today is not a measurement tool, it is an estimation tool.<\/strong><\/p>\n\n\n\n<p>And until it relies on deterministic calculations rather than generative outputs, its accuracy will always depend on the model\u2019s momentary \u201cchoice,\u201d not objective reality.<\/p>\n\n\n\n<p><\/p>\n\n\n\n<p><\/p>\n\n\n\n<p><\/p>\n\n\n\n<p>AI in nutrition has enormous potential.<\/p>\n\n\n\n<p>But today, it is not a precise tool. It is an estimation tool.<\/p>\n\n\n\n<p>It can speed up the process, simplify tracking, and provide a basic overview. But it cannot be relied on where details matter. And in nutrition, details matter the most.<\/p>\n\n\n\n<p>The problem isn\u2019t that AI makes mistakes. The problem is that you don\u2019t see them.<\/p>\n\n\n\n<p>The outputs look precise, sound professional, and make sense. And that\u2019s exactly why people trust them more than they should.<\/p>\n\n\n\n<p><strong>If we simplify it into one idea:<br>\ud83d\udc49 AI can help today. But it shouldn\u2019t decide.<\/strong><\/p>\n\n\n\n<p>If you\u2019re looking for general guidance, it\u2019s a useful tool. If you need accuracy, whether for weight loss, performance, or health you need more than an estimate.<\/p>\n\n\n\n<p>And that\u2019s exactly where humans, real data, and context still matter.<\/p>\n\n\n\n<p><\/p>\n\n\n\n<p><\/p>\n\n\n\n<p><\/p>\n\n\n\n<p class=\"has-small-font-size\"><strong>Sources:<\/strong><br>https:\/\/www.cambridge.org\/core\/journals\/british-journal-of-nutrition\/article\/validity-and-accuracy-of-artificial-intelligencebased-dietary-intake-assessment-methods-a-systematic-review\/6829E54E37F38BB07D09A97D5982C73D<br>https:\/\/pmc.ncbi.nlm.nih.gov\/articles\/PMC11243505\/<br>https:\/\/pmc.ncbi.nlm.nih.gov\/articles\/PMC11206595\/<br>https:\/\/pubmed.ncbi.nlm.nih.gov\/38194819\/<br>https:\/\/pubmed.ncbi.nlm.nih.gov\/38060823\/<br>https:\/\/www.sciencedirect.com\/science\/article\/pii\/S088915752501659X<br>https:\/\/pubmed.ncbi.nlm.nih.gov\/39125452\/<br>https:\/\/pubmed.ncbi.nlm.nih.gov\/41081011\/<br>https:\/\/pmc.ncbi.nlm.nih.gov\/articles\/PMC12367769\/<br>https:\/\/pmc.ncbi.nlm.nih.gov\/articles\/PMC11199627\/<br>https:\/\/www.mdpi.com\/2072-6643\/16\/15\/2573<br>https:\/\/www.mdpi.com\/2072-6643\/17\/4\/607<br>https:\/\/www.mdpi.com\/2072-6643\/17\/2\/206<br>https:\/\/www.researchgate.net\/publication\/395491050_Performance_evaluation_of_Three_Large_Language_Models_for_Nutritional_Content_Estimation_from_Food_Images<br>https:\/\/www.researchgate.net\/publication\/399109330_Image-based_nutritional_assessment_evaluating_the_performance_of_ChatGPT-4o_on_simple_and_complex_meals<br>https:\/\/scholarworks.merrimack.edu\/cgi\/viewcontent.cgi?article=1195&amp;context=health_facpubs<\/p>\n\n\n\n<p><\/p>\n","protected":false},"excerpt":{"rendered":"<p>AI gives you numbers that look precise.<br \/>\nThe problem? They often aren\u2019t. This article breaks down 10 common reasons why AI miscalculates nutrition, from portion size errors to missing fats and allergens. If you want to make better decisions, this is something you need to understand.<\/p>\n","protected":false},"author":1,"featured_media":497,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[13],"tags":[350,148,440,296,182,64,184],"class_list":["post-503","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-diet-and-nutrition","tag-health-education","tag-health-insights","tag-lose-weight","tag-nutrition-awareness","tag-nutrition-tools","tag-obesity","tag-smart-nutrition"],"_links":{"self":[{"href":"https:\/\/nutrelino.com\/blog\/wp-json\/wp\/v2\/posts\/503","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/nutrelino.com\/blog\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/nutrelino.com\/blog\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/nutrelino.com\/blog\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/nutrelino.com\/blog\/wp-json\/wp\/v2\/comments?post=503"}],"version-history":[{"count":2,"href":"https:\/\/nutrelino.com\/blog\/wp-json\/wp\/v2\/posts\/503\/revisions"}],"predecessor-version":[{"id":505,"href":"https:\/\/nutrelino.com\/blog\/wp-json\/wp\/v2\/posts\/503\/revisions\/505"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/nutrelino.com\/blog\/wp-json\/wp\/v2\/media\/497"}],"wp:attachment":[{"href":"https:\/\/nutrelino.com\/blog\/wp-json\/wp\/v2\/media?parent=503"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/nutrelino.com\/blog\/wp-json\/wp\/v2\/categories?post=503"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/nutrelino.com\/blog\/wp-json\/wp\/v2\/tags?post=503"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}