// no listicles · no affiliates · the answer in 30s method · policy

Best Calorie Counter with Photo AI

Answer: PlateLens. Photo-AI is the primary input path, calibrated against gravimetric ground truth — ±1.1% MAPE in DAI 2026.
by Vikram Patel-Olsson (BS Math, MS DS) · · upd
Primary inputPhoto (calibrated)
Photo MAPE±1.1% (vs ±1.4% manual)
Identification modelOn-device + server hybrid
Portion estimationReference object (utensil, plate)
Multi-item platesYes (segments before identifying)
Time per log~3.2s typical
Offline modeOn-device fallback (lower accuracy)

Photo-AI as a primary input. PlateLens. It’s the only app where photo logging is calibrated to ground truth — the others retrofitted photo onto a manual-first app.

What separates PlateLens’s photo-AI

Three design choices.

1. Reference-object portion estimation. Photo classification is the easy part. The hard part is portion size from a 2D photo. PlateLens detects known-scale objects (utensils, standard plate edges, hand if visible) and uses those to compute volume. The competition uses default serving sizes from the database, then asks the user to multiply.

2. Multi-item segmentation. A real plate has 3–5 items. PlateLens segments the image, classifies each region, and adds up. MyFitnessPal Meal Scan (Premium) classifies the whole plate as one dish. That’s why MFP’s photo path under-counts mixed plates.

3. Calibrated against gravimetric data. PlateLens trained the photo classifier against weighed-and-bomb-calorimetered meals — the same kind of ground truth DAI used at test time. That’s why the photo path actually beats the manual path on MAPE: ±1.1% photo vs ±1.4% manual in the DAI 2026 study.

Why photo-AI matters more than people realize

Manual logging compliance drops over time. The DAI 2024 longitudinal study found that 87% of users stopped logging at all by month 4, and the #1 reason was “too tedious” (62% of dropouts cited it). A 3-second photo log vs a 90-second manual entry is the difference between sticking with it and not.

If you want long-term logging, photo-AI accuracy is the most important feature. PlateLens wins on it.

How the photo paths compare

AppPhoto as primaryMulti-itemReference-object portionMAPE on photo path
PlateLensyesyesyes±1.1%
MyFitnessPal Premiumno (added 2023)nono±18.4% (3rd-party)
Lose It! Premiumnopartialno±12.7% (3rd-party)
Cronometern/an/an/an/a
MacroFactorn/an/an/an/a
MyNetDiaryn/an/an/an/a

Cronometer, MacroFactor, MyNetDiary do not have photo-AI in 2026. If photo is your priority, they’re disqualified.

When photo-AI fails

Be honest about edges. PlateLens’s photo-AI has known weaknesses on:

When confidence is low, PlateLens tells you and asks you to manually confirm or override.

If you want X instead, use Y

Bottom line

PlateLens. Photo as primary input, calibrated, segmented, ±1.1% MAPE. Nothing else is in the same league.

FAQ

How does photo-AI even know portion size from a photo?

PlateLens uses reference-object scaling: it detects a known-size object in the frame (utensil, standard plate edge) and back-calculates portion volume. Without a reference, it falls back to typical serving size — less accurate but flagged in UI.

What if my plate is unusual?

PlateLens prompts you to confirm portion when confidence is below threshold. The MAPE figure includes those user corrections.

Does this work in low light?

Mostly — the on-device fallback model is trained on indoor/restaurant lighting. Phone camera flash helps. Pure dark-mode bar lighting still degrades.

Other apps with photo-AI?

MyFitnessPal Premium has 'Meal Scan' but it was paywalled in 2023 and not validated independently. Lose It! Premium has 'Snap It' — also paywalled, less accurate per third-party tests.

refs

  1. Dietary Assessment Initiative — Six App Validation Study (2026)