Personalized Nutrition-Aware Dietary Recommendation With Multimodal Large Language Models

  • Xiwen Deng
  • , Wenzhe Tu*
  • *Corresponding author for this work

Research output: Contribution to journalJournal articlepeer-review

Abstract

Poor diet quality is a major modifiable risk factor for obesity, cardiovascular disease, and metabolic disorders, motivating a growing ecosystem of diet-tracking apps and recipe platforms that promise personalized guidance. However, most of these systems still rely on simple scoring rules or ID-based collaborative filtering that ignore rich multimodal food information (ingredients, cooking methods, images, nutrient profiles) and lack explicit nutritional reasoning. To address this gap, we propose NutriRec, a multimodal LLM-based framework for personalized dietary recommendation that links nutritional reasoning with collaborative dietary patterns. We uses a multimodal LLM to generate unified representations of meals and food items from textual descriptions, ingredient lists, nutrient profiles, and images, capturing both sensory and nutritional attributes. On top of these content-based embeddings, NutriRec incorporates collaborative filtering signals through a post-alignment mechanism that aligns user representations from an ID-based sequential model with those from the multimodal LLM. This design enables explainable dietary reasoning by grounding each recommendation in concrete nutritional evidence (e.g., lower sodium, higher fiber, heart-healthy tags) while still reflecting behavioral patterns learned from large-scale consumption logs. We evaluate NutriRec on realistic dietary recommendation scenarios with long-term, multimodal food interaction histories. Across these settings, NutriRec not only yields stronger recommendations than sequential and LLM-based baselines, but also provides more explainable dietary reasoning by explicitly linking suggestions to nutrient profiles, meal composition, and simple guideline-based constraints.

Original languageEnglish
Pages (from-to)21791-21804
Number of pages14
JournalIEEE Access
Volume14
DOIs
Publication statusPublished - 3 Feb 2026

UN SDGs

This output contributes to the following UN Sustainable Development Goals (SDGs)

  1. SDG 3 - Good Health and Well-being
    SDG 3 Good Health and Well-being

User-Defined Keywords

  • dietary recommendation
  • large language models
  • multi-modal recommendation
  • Personalized nutrition

Fingerprint

Dive into the research topics of 'Personalized Nutrition-Aware Dietary Recommendation With Multimodal Large Language Models'. Together they form a unique fingerprint.

Cite this