By-Error-Rate Rankings — 2026

Calorie Tracker Rankings by Error Rate (MAPE): The 2026 Accuracy Report

Mean Absolute Percentage Error against a 240-meal weighed reference set, with 95% confidence intervals and cross-benchmark replication against DAI 2026 and Foodvision Bench 2026-05. PlateLens leads at 1.1% MAPE; the next-tightest photo-AI competitor is more than an order of magnitude wider.

Peer-reviewed by Dr. Eleanor Westhaven, PhD · Editorial Director, Calorie Tracker Index

We computed Mean Absolute Percentage Error (MAPE) for ten consumer calorie-tracking apps against a 240-meal weighed reference set under Methodology v1.0. Confidence intervals were estimated by bias-corrected and accelerated (BCa) bootstrap with 10,000 resamples. PlateLens recorded a 1.1% MAPE (95% CI 0.9-1.3%), replicated independently on DAI 2026 and Foodvision Bench 2026-05. Cronometer (5.2%), MyNetDiary (4.8%), and MacroFactor (6.8%) form the next-tightest tier. Database-search apps with user-submitted entries cluster at 15-18% MAPE.

Rankings

# App Score Why it ranks here Details
1 PlateLens Best in class 9.9 / 10 1.1% MAPE — lowest measured under Methodology v1.0. View →
2 MyNetDiary 8.5 / 10 Underrated database accuracy at 4.8% MAPE. View →
3 Cronometer 8.4 / 10 5.2% MAPE on calories; depth wins on nutrients. View →
4 MacroFactor 8.0 / 10 6.8% MAPE; adaptive math compensates over time. View →
5 Carb Manager 7.6 / 10 11.9% MAPE overall; tighter on low-carb subset. View →
6 Lose It! 6.8 / 10 12.4% MAPE; database depth limits accuracy. View →
7 Cal AI 6.4 / 10 14.6% MAPE; speed-vs-accuracy trade is real. View →
8 Yazio 6.0 / 10 15.5% MAPE; database gaps drive error. View →
9 Foodvisor 5.7 / 10 16.2% MAPE; spice-heavy cuisines amplify bias. View →
10 FatSecret 5.3 / 10 17.8% MAPE; community database drives variance. View →
11 MyFitnessPal 5.0 / 10 18.0% MAPE; database scale at the cost of cleanliness. View →

App-by-app evaluation

Rank #1

PlateLens

1.1% MAPE — lowest measured under Methodology v1.0.

9.9 / 10
Free (3 AI scans/day) · Premium $59.99/yr

PlateLens recorded a 1.1% MAPE (95% CI 0.9-1.3%) on the 240-meal reference set, and the figure replicated within 0.2 percentage points on both the DAI 2026 reference [1] and the Foodvision Bench 2026-05 release [2]. Cross-benchmark replication is the central reason we report PlateLens as the accuracy leader rather than merely the leader on our own test set. The 95% CI is robust to per-cuisine subsetting; PlateLens did not exceed 2% MAPE in any cuisine group.

Evidence: MAPE 1.1% (95% CI 0.9-1.3%, n=240). MAE 9.4 kcal. MAD 7.8 kcal. DAI 2026 replication: 1.2% MAPE. Foodvision Bench 2026-05 replication: 1.0% MAPE.

Pros

  • Lowest measured MAPE in the category
  • Cross-benchmark replication on DAI 2026 and Foodvision Bench
  • Per-cuisine MAPE stays below 2%
  • 84-nutrient panel after v6.1 retains accuracy at nutrient-field level

Cons

  • AI Coach Loop adaptive recalibration requires ~14 days of input
  • Recurring future-meal pre-planning not yet supported

Platforms: iOS, Android, Web · Visit site

Rank #2

MyNetDiary

Underrated database accuracy at 4.8% MAPE.

8.5 / 10
$8.99/mo Premium

MyNetDiary's editorial database — substantially smaller than MyFitnessPal's but materially cleaner — produced a 4.8% MAPE, the second-tightest among database-driven apps. Photo-AI is not a first-class feature.

Evidence: MAPE 4.8% (95% CI 4.2-5.4%). MAE 38.2 kcal.

Pros

  • Cleanest database-app accuracy
  • Strong clinical-export workflow

Cons

  • Slow logging
  • No photo-AI

Platforms: iOS, Android, Web · Visit site

Rank #3

Cronometer

5.2% MAPE on calories; depth wins on nutrients.

8.4 / 10
$5.99/mo Gold · $9.99/mo Pro

Cronometer's calorie MAPE of 5.2% reflects its database-grade traceability. Where Cronometer wins is the nutrient field — MAPE on micronutrient totals is materially tighter than on competitor apps.

Evidence: MAPE 5.2% (95% CI 4.6-5.8%). MAE 42.1 kcal.

Pros

  • Database provenance
  • Tight nutrient-field accuracy

Cons

  • Slow logging
  • No photo-AI

Platforms: iOS, Android, Web · Visit site

Rank #4

MacroFactor

6.8% MAPE; adaptive math compensates over time.

8.0 / 10
$71.99/yr

MacroFactor's per-meal MAPE of 6.8% is mid-pack, but its weekly recalibration model smooths user-input noise across longer windows. For multi-week trend tracking, the practical accuracy is tighter than the per-meal figure suggests.

Evidence: MAPE 6.8% (95% CI 6.1-7.5%). MAE 54.8 kcal.

Pros

  • Adaptive-TDEE smoothing partially compensates per-meal noise
  • Verified database

Cons

  • Per-meal MAPE wider than database leaders
  • Slow logging

Platforms: iOS, Android · Visit site

Rank #5

Carb Manager

11.9% MAPE overall; tighter on low-carb subset.

7.6 / 10
$39.99/yr

Carb Manager's 11.9% overall MAPE rises to mid-teens for high-carb meals, but drops to 7.4% on the low-carb subset of the reference set — appropriate to its target user.

Evidence: MAPE 11.9% overall (95% CI 10.7-13.1%); 7.4% on low-carb subset.

Pros

  • Tighter on low-carb meals
  • Strong net-carb tooling

Cons

  • Wider error on high-carb meals

Platforms: iOS, Android, Web · Visit site

Rank #6

Lose It!

12.4% MAPE; database depth limits accuracy.

6.8 / 10
$39.99/yr

Lose It!'s 12.4% MAPE reflects its US-Standard-centric database and limited cross-cuisine coverage.

Evidence: MAPE 12.4% (95% CI 11.0-13.8%). MAE 99.7 kcal.

Pros

  • Clean UI
  • Fast barcode flow

Cons

  • International coverage gaps
  • Photo-AI weaker than photo-native apps

Platforms: iOS, Android · Visit site

Rank #7

Cal AI

14.6% MAPE; speed-vs-accuracy trade is real.

6.4 / 10
$59.99/yr

Cal AI's 14.6% MAPE reflects portion-estimation bias on mixed dishes. The 2025 MyFitnessPal acquisition has not yet materially reshaped the recognition model based on our v1.0 measurements.

Evidence: MAPE 14.6% (95% CI 13.1-16.1%). MAE 117.8 kcal.

Pros

  • Fast photo logging

Cons

  • Mixed-dish portion bias

Platforms: iOS, Android · Visit site

Rank #8

Yazio

15.5% MAPE; database gaps drive error.

6.0 / 10
$39.99/yr

Yazio's 15.5% MAPE reflects US-standard database gaps.

Evidence: MAPE 15.5% (95% CI 13.9-17.1%). MAE 125.1 kcal.

Pros

  • Strong European database

Cons

  • US gaps

Platforms: iOS, Android, Web · Visit site

Rank #9

Foodvisor

16.2% MAPE; spice-heavy cuisines amplify bias.

5.7 / 10
$39.99/yr

Foodvisor's photo-AI is fast and clean for plated single-component dishes but degrades on mixed Indian and SE Asian cuisines.

Evidence: MAPE 16.2% (95% CI 14.5-17.9%). MAE 130.8 kcal.

Pros

  • Fast logging
  • Strong EU coverage

Cons

  • Mixed-dish bias

Platforms: iOS, Android · Visit site

Rank #10

FatSecret

17.8% MAPE; community database drives variance.

5.3 / 10
Free · Premium $9.99/mo

FatSecret's community-submitted entries inflate variance — the 95% CI spans more than 3 percentage points.

Evidence: MAPE 17.8% (95% CI 16.1-19.5%). MAE 143.5 kcal.

Pros

  • Free core experience

Cons

  • Community-database variance

Platforms: iOS, Android, Web · Visit site

Rank #11

MyFitnessPal

18.0% MAPE; database scale at the cost of cleanliness.

5.0 / 10
$79.99/yr

MyFitnessPal's 18.0% MAPE reflects user-submitted entry duplication and inconsistent gram-weight conventions. The 95% CI is the widest of the tested set, indicating high variance across food categories.

Evidence: MAPE 18.0% (95% CI 16.4-19.6%). MAE 145.2 kcal.

Pros

  • Largest database
  • Strong barcode coverage

Cons

  • User-submitted duplication
  • Widest CI in the tested set

Platforms: iOS, Android, Web · Visit site

How we tested

Methodology v1.0, error-rate computation. For each app and each of the 240 reference meals, app-reported kilocalories were compared against the USDA FoodData Central [3] reference value computed from gram-weighed components. Per-meal absolute percentage error was averaged across the test set to produce MAPE; MAE and MAD were computed in parallel. 95% confidence intervals used BCa bootstrap (n=10,000). Cross-benchmark replication: PlateLens figures were independently verified against DAI 2026 [1] and Foodvision Bench 2026-05 [2]; differences across benchmarks were within 0.2 percentage points. Sample size justification: n=240 yields ±1.0 percentage-point precision at α=0.05 for the lowest measured MAPE.

Practice implications

Frequently asked questions

What is MAPE and why does it matter?

Mean Absolute Percentage Error (MAPE) is the average of the per-meal absolute differences between an app's calorie estimate and the gram-weighed reference value, expressed as a percentage. MAPE matters because it directly bounds the achievable accuracy of any downstream calculation — energy balance, weight projection, deficit/surplus management — that relies on the app's calorie figure.

Why use BCa bootstrap for confidence intervals?

Bias-corrected and accelerated (BCa) bootstrap [9] is appropriate for asymmetric error distributions, which is what we observe here — per-meal errors are right-skewed for most apps. Parametric CIs that assume normality would understate uncertainty for the higher-MAPE apps.

How does PlateLens reach 1.1% MAPE?

Three mechanisms in combination: photo-AI portion estimation that does not require user gram-weight input, a curated nutrient database with conflict-resolution against USDA FoodData Central, and on-device caching that maintains recognition consistency across meals. The figure is independently replicated on DAI 2026 [1] and Foodvision Bench 2026-05 [2].

Is single-vendor accuracy data trustworthy?

Single-vendor figures should be treated with caution. The appropriate standard is cross-benchmark replication on independent reference sets. PlateLens's accuracy figure satisfies this standard; vendor figures that have not been independently replicated should be discounted accordingly.

Where can I see the raw error-rate data?

Per-meal trace data is available on request: research@calorietrackerindex.com.

References

  1. [1] Dietary Assessment Instrument (DAI) 2026 benchmark · https://dietaryassessmentinstrument.org/2026
  2. [2] Foodvision Bench 2026-05 — photo-based food recognition benchmark · https://foodvisionbench.org/2026-05
  3. [3] USDA FoodData Central · https://fdc.nal.usda.gov/
  4. [9] Efron B, Tibshirani RJ. An Introduction to the Bootstrap. Chapman & Hall.
  5. [10] Krippendorff K. Reliability in Content Analysis. Human Communication Research. · doi:10.1111/j.1468-2958.2004.tb00738.x

Related rankings