Skip to main content

Thank you for visiting nature.com. You are using a browser version with limited support for CSS. To obtain the best experience, we recommend you use a more up to date browser (or turn off compatibility mode in Internet Explorer). In the meantime, to ensure continued support, we are displaying the site without styles and JavaScript.

Automated food intake tracking requires depth-refined semantic segmentation to rectify visual-volume discordance in long-term care homes

Abstract

Malnutrition is a multidomain problem affecting 54% of older adults in long-term care (LTC). Monitoring nutritional intake in LTC is laborious and subjective, limiting clinical inference capabilities. Recent advances in automatic image-based food estimation have not yet been evaluated in LTC settings. Here, we describe a fully automatic imaging system for quantifying food intake. We propose a novel deep convolutional encoder-decoder food network with depth-refinement (EDFN-D) using an RGB-D camera for quantifying a plate’s remaining food volume relative to reference portions in whole and modified texture foods. We trained and validated the network on the pre-labelled UNIMIB2016 food dataset and tested on our two novel LTC-inspired plate datasets (689 plate images, 36 unique foods). EDFN-D performed comparably to depth-refined graph cut on IOU (0.879 vs. 0.887), with intake errors well below typical 50% (mean percent intake error: \(-4.2\)%). We identify how standard segmentation metrics are insufficient due to visual-volume discordance, and include volume disparity analysis to facilitate system trust. This system provides improved transparency, approximates human assessors with enhanced objectivity, accuracy, and precision while avoiding hefty semi-automatic method time requirements. This may help address short-comings currently limiting utility of automated early malnutrition detection in resource-constrained LTC and hospital settings.

Introduction

Malnutrition has multidomain effects and should be monitored especially in older adults as evidenced by the clinical ramifications (morbidity1, decreased quality of life2), annual economic impact (USA $15.5 billion3, UK £7.3 billion4), and high prevalence (23% malnourished5). In long-term care (LTC) homes, the prevalence is greater with malnutrition or risk for malnutrition affecting 54% of residents6. Malnutrition and risk for malnutrition is primarily due to low food intake of residents7. Thus, tracking and preventing poor food intake is paramount. However, we lack quality tracking methods for food and fluid intake, especially needed with multiple staff involved in the care of residents over the course of a day or week. While methods for measuring food and fluid intake exist, these methods are subject to self-reporting bias, negatively affecting both validity and accuracy8,9 (e.g., errors up to: 400%, 24-hr recall; 50%, portion size10). In the LTC sector, personnel are mandated to report at-risk resident’s food and fluid intake; however, correct estimation of intake occurs only 44% of the time under routine conditions and as low as 38% of the time with delayed recording11. As a result, trust in these measurements is low, with limited utility in practice but care providers would like to utilize this information if measurement reliability and trust in these measurements could be ensured12.

Automated tools may provide a time efficient and cost effective, and objective alternative. More generally, progress in this field of end-to-end systems for nutrition monitoring has been outside the context of LTC with an emphasis of an individual tracking and managing their personal weight loss or health tracking using mobile devices13,14,15,16,17,18,19. While these approaches could be modified for use in LTC, in their current form, they target a different purpose (e.g., calorie tracking), still rely on self-monitoring, and do not consider the LTC context for food and fluid intake tracking best practices. Two reviews provide an overview of technology-driven methodologies including a summary of work exploring wearable devices and sensor20,21. However, these wearable sensor approaches have typically been developed for individual use and require individuals to wear sensors like microphones or strain sensors. While outside of LTC these may show promise, within the LTC setting, the wearables approach is inappropriate from a privacy perspective, when considering the degree of assistance needed during mealtime, the requirements of proper sterilization practices, as well as the financial implications of the number of devices required to track intake. Additional considerations for wearables are prudent in the LTC sector as there is evidence suggesting it modifies (reduces) food intake behaviour. Two papers included in21 showed how wearable micro-cameras can improve intake estimates, but that participants (both adults and children) were self conscious, would not wear it in public, and their eating behaviour changed (reduced intake)22,23. Given that LTC residents are already consuming too few bites as evidenced by the high degree of malnutrition primarily due to low intake7, solutions with the potential to further reduce intake are inappropriate in this setting. As such, current approaches are infeasible for large-scale monitoring, especially in these time-constrained and financially constrained environments such as LTC or hospital settings.

An image-based solution, where food is monitored by a proxy (e.g., personal support worker) may overcome some of these clinically relevant limitations. Within the LTC context specifically, a comparison to estimate food waste of regular- and modified-texture diets either with the visual estimation method or by using digital photographs for retrospective analysis24 was conducted. However, both methods required significant operator time as it was a manual process. More broadly, further work is needed for developing an accurate, objective and cost-effective automated system20,25. Perhaps, most relevant is the work of Astell and colleagues26, who developed an effective electronic food record system for nutrient tracking system for community dwelling older adults. While promising (approximately 97% agreement for energy intake compared to food diaries), the comparison method of food diaries is similar to the monitoring already in place within LTC so true accuracy of this method remains unclear.

We previously established that the LTC sector requires a system that is reliable, accurate, cost effective and time efficient for measuring food and resulting energy, macro and micronutrient intake12. However, before assessing food and fluid intake, three main questions must first be answered: where is there food (segmentation), what foods are present (classification), and how much food remains relative to the initial amount (volume estimation)? For the purpose of this paper, we focus on the where (segmentation), and the how much (volume estimation), as types of food items in LTC are well constrained through monthly menu-planning. Additionally, food classification has received the most attention; reviews of the literature can be found elsewhere (e.g.,16,20,25,27,28).

Food segmentation is a domain which is relatively unexplored. Existing food intake tracking systems rely on images from multiple perspectives18,29, require a single image with a fiducial marker (i.e., reference object15,30), or may not be suitable for real-time monitoring14. Others are limited to predicting food areas with a bounding box31, require manual labelling for each food item14, or require manual selection of bounding boxes32. These methods involve operator time and may impact accuracy. For example, two operators may segment food differently, foods may be incorrectly labelled, or labels may be missed in some cases. One semi-automatic method, interactive graph cut segmentation, has yielded strong accuracy in food segmentation32,33,34. It does not impart the same degree of burden as manual segmentation and we consider this as an “applied ground truth”. However, interactive annotation graph cut35 requires user input to initialize the segmentation process (e.g., drawing areas to keep or discard). Adding a few seconds per image within the LTC environment makes it prohibitive within this context.

While food image segmentation progress has been made, error assessment in these systems tends not to be reported, or segmentation is coupled with either classification14,31,32,33,36,37,38 or volume34, making sources of error difficult to disentangle. This has practical implications as there is generally no way to systematically assess error propagation as part of the pipeline for predicting nutritional outcomes. This results in the system operating as a “black-box”, which may limit the uptake of these approaches in practice due to low perceived trust-worthiness. Beyond the user and ethical perspectives, several researchers also describe the need for accurate segmentation methods for accurately predicting nutritional information (e.g., energy, macro-/micronutrient content) down-stream in the pipeline14,17,30,37. Intersection over union (IOU) has been the most consistently reported accuracy metric17,19,39,40. IOU has several advantages over more traditional precision/recall metrics as it considers the proportion of properly assigned pixels but also penalizes false positive predictions; however, it may not necessarily capture an assessor’s perspective on what is relevant food (e.g., to include or not include crumbs).

Food volume estimation systems are also relevant for estimating portion sizes of consumed food by subtracting the remainder of food from the original portion size. Several attempts have been made using template shape matching36,41,42,43,44,45,46. Three main drawbacks of these methods are the requirement of a shape library, difficulty with template matching with occlusions, and varying preparation methods of the same food. For example, if a food is prepared differently it may not map onto the appropriate shape model (i.e., 3D banana in peel may be in the library but sliced or diced banana may not). Similar to the segmentation problem, others have applied a multi-image perspective or stereo reconstruction for volume estimation47,48 or building a 3D representation through point-cloud representation49. The main drawback of these approaches is the time required to take the photos from different perspectives or gather enough sample points scanned for an accurate 3D representation; lack of time is a main concern when considering the LTC context. Another challenge of these approaches is accurate modelling/measuring of highly textured foods. This is a particularly salient issue for LTC where modified textures are often prescribed as part of a therapeutic diet so very different foods can appear similar (e.g., minced or pureed foods). Others have employed depth cameras and structured lighting to map the topology of the foods14,50,51,52,53. One drawback of depth- and structured-light-only methods is highly reflective foods (e.g., gelatin, soup) which can throw off the readings but they have an advantage at being more robust against illumination variations53. Despite the deep learning boom, acquiring adequately large and complete food datasets for training and testing has limited progress54. The few forays have been fairly early-stage as they contain either large mean volume estimation errors (up to 400 mL error14), measure in terms of niche units (bread units tailored to diabetes55), or are limited to a small number of or synthetic food items in a highly controlled environment (e.g.,56,57).

Recent reviews corroborate that further work is needed for developing an accurate, objective and cost-effective automated system20,25 and describe the need for more complex meal scenarios (i.e., beyond solid, separated foods, or synthetic foods)16,25 and adequate statistical analyses of methods20. In line with these opportunity areas and current human computer interaction trajectories58, we seek to improve trust and transparency by focusing on developing an explainable system which approximates human assessors but with enhanced objectivity, accuracy, and precision. Specifically in this paper, we describe and evaluate a novel fully automated depth-refined multisensor food intake tracking system. Here the depth-refined segmentation and volume estimation have been decoupled to disentangle and assess potential sources of error. Through this decoupling, we seek to enhance reliability for eventual integration with nutritional intake estimation, and to reduce potential barrier to uptake in practice. We designed the segmentation system to be used in clinical settings (such as LTC or hospitals) with acquisition consistent with LTC food and fluid intake visual assessment procedures. Our system is comprised of an RGB-D camera, a novel deep convolutional neural network encoder-decoder food network (EDFN), fuses output from the EDFN with superpixel processing, and incorporates depth information for enhanced segmentation and volume estimation. The use of a single RGB-D camera brings simplicity over a multi-camera or multi-perspective set-up, reducing processing and acquisition time while removing subjectivity in the assessment. We trained our EDFN on the UNIMIB2016 dataset59 and test it on two novel datasets to reduce bias and enhance generalizability. The two novel datasets are (1) a regular texture foods dataset and, (2) a modified texture foods dataset (e.g., pureed, minced). To the best of our knowledge, this is the first modified texture foods dataset used for segmentation or volume estimation. We conducted comprehensive analyses including IOU, 2D- and 3D- percent intake error, absolute intake error, and mean intake error bias. We use ground-truth hand segmentation and comparison against an “applied ground truth” through the graph cut semi-automated method. We supplement these analyses with volume disparities to illuminate how segmentation strategies impact accuracy and under what conditions IOU may be insufficient to assess true accuracy. Using this more holistic construct for assessment, we aim to enable trust in the system and document potential circumstances and limitations of the system relevant to the LTC domain for early malnutrition detection via plate-by-plate food consumption tracking.

Results

Summary of results

Overview of the food detection system and new long-term care food dataset

Figure 1 illustrates our proposed solution for automatic semantic segmentation with depth refinement and Fig. 2 provides a provides visual examples of our custom long-term care food dataset. Data collection for our custom dataset is described in Methods. This 1,039-image dataset is a fully labelled, high resolution dataset consisting of 47 unique foods representative of LTC and representing a variety of fruits, vegetables, pastas, soup, and meat dishes. A summary of the 47 food items imaged can be found in Table 1. This dataset is comprised of two subsets: a “regular texture” dataset, and a “modified texture foods dataset”. The regular dataset is comprised of 9 regular texture foods across three meals each containing up to three meal items and imaged at every 25% incremental amount relative to the full portion. It yielded 125 unique plates per meal and 375 unique plates across meals. The modified texture foods dataset is comprised 664 images across 93 classes of modified texture food samples representing 47 unique foods prepared by a LTC kitchen. All samples include hand-segmented and hand-labelled pixel-level segmentations. A 314-image subset of the modified texture foods dataset representing 63 food samples (56 unique) and 27 unique foods additionally includes full nutritional information provided by the LTC home. These foods were imaged one per plate and samples were imaged at different simulated intake levels by progressively removing some of the sample. Foods were imaged using the Intel RealSense to obtain RGB-D images.

Figure 1
figure 1

System diagram of the proposed deep food segmentation network comprised of: (1) encoder-decoder food network (EDFN) consisting of a residual encoder microarchitecture60 and a pyramid scene parsing61 decoder microarchitecture which outputs a segmented food mask; (2) pixel height map estimation for assessing food depth; (3) depth refinement which outputs a depth-refined food mask (EDFN-D); and (4) food intake report summarizing the volume of food on the plate, the intake amount (in mL) and the intake percent.Our network was tested on a custom dataset of foods offered in long-term care representing both regular texture and modified texture foods (e.g., purées).

Figure 2
figure 2

Example images from our custom LTC simulated intake dataset consisting of colour (RGB), depth and mass amounts (colour examples shown here) in (a) the modified texture foods dataset, and (b) the regular texture foods dataset. The modified texture foods dataset comprises 5 reference portion images whereas images of the regular texture dataset represent three meals at every iteration of 25% simulated intake (a subset shown here).

Table 1 Imaged foods represented in each of the regular texture foods dataset and the modified texture foods dataset. Regular texture foods were derived from LTC menus and imaged with three foods per meal (breakfast: oatmeal, toast, scrambled eggs; lunch: tortellini, salad, cookie; dinner: meatloaf, potatoes, corn) at every 25% incremental amounts of each food item relative to the full portion. Modified texture foods were imaged separately with at five incrementally smaller portions.

“Visual-volume discordance”: segmentation in the context of food intake

When solely considering intake using segmentation, the form factor of food inherently assumes depth is uniform across the segmented portion. However, this approach fails to fully capture the context of food; we refer to this as the “visual-volume discordance”. For example, consider one tablespoon (15 mL) of tomato sauce, this sauce could be piled relatively high into a mound (i.e., representing relatively few plate pixels), or could be very thinly spread across the majority of the plate (i.e., representing many plate pixels). From a computer vision approach to segmentation, these two plates would yield extremely different segmented areas while the absolute volume of the sauce would be the same. A human assessor is able to note there is little sauce on the plate in either configuration. Depth-refinement, either as part of the segmentation pipeline or conducted through relative changes in volume for food volume intake assessment, circumvents this issue by providing context beyond the pixel count of a segment and brings assessment closer (but with higher precision and accuracy) to a human assessor. Now consider what is deemed “ground truth” from hand segmentation of an image. Here, human assessors indicate where on the plate there was food to generate the ground truth. However, considering segmentation accuracy in isolation misses important context about how much food is present, which is the more pertinent question for assessing food intake. As such, we cannot rely fully on classical segmentation accuracy for evaluating system performance since there can be a strong discordance between visual (RGB) and volume (RGB-D) assessments. Volume consideration is particularly essential for estimating food intake when accounting for the high prevalence of modified texture foods in LTC. While metrics pertaining to visual accuracy are most synonymous with traditional assessments of segmentation accuracy (e.g., IOU), metrics pertaining to volume accuracy may be more representative of true intake, especially for modified texture foods which have higher fluidity. Figure 3 provides a visual analyses of the results taking into account these considerations while Table 2 provides a numerical summary of results. Data are reported as (mean ± SD).

Figure 3
figure 3

Visual comparison of proposed method (EDFN) and our “applied ground truth” (GC) both without and with depth-refinement (EDFN-D, GC-D). Examples span both regular and modified textures and illustrate (a) plates with high visual-volume discordance, as well as (b) examples where depth-refinement rectified oppositional segmentation compared to human assessor. IOU is shown in black bold at the top of the frame while volume error (mL) is indicated in white text at the bottom of each frame; negative volume error implies under-segmentation (i.e., over-estimation of intake).

Table 2 Comparative analyses of system performance within and across LTC datasets between our proposed method (EDFN, EDFN-D) and the “applied ground truth” graph cuts (GC, GC-D). % error intake refers to the proportion of segmented pixels calculated using the predicted estimate minus the target ground-truth hand segmented regions; 2D: no-depth, 3D: with depth-refinement.
Table 3 Summary of volume estimation accuracy across portion sizes for the modified texture foods dataset across our proposed EDFN and EDFN-D.

Regular texture foods

Food segmentation accuracy

Regarding the regular texture foods dataset, food segmentation accuracy for our proposed system without and with depth-refinement was commensurate with semi-automatic graph cuts (EDFN: 0.943 ± 0.047, EDFN-D: 0.918 ± 0.048, GC: 0.955 ± 0.055, GC-D: 0.933\(~\pm \) 0.060). Variance was similar but slightly lower for our proposed EDFN and EDFN-D than GC and GC-D.

Segmentation agreement

IOU was improved through depth refinement (0.885 ± 0.068 vs 0.927 ± 0.029). The graph cut IOU both without and with depth-refinement outperformed our proposed methods owing to user-specified seed points (GC: 0.938 ± 0.036, GC-D: 0.941 ± 0.029). We attribute the improvements of both our proposed system and our applied ground truth with depth-refinement to reduction in the visual-volume discordance due to low-profile yet visually distinct foods on a plate, such as pasta sauce.

Percent intake error (2D and 3D)

Regarding 2D percent intake error (i.e., using segmentation alone), our non-depth-refined and depth-refined proposed systems were outperformed by the depth-refined graph cut implementation (EDFN\(_{\text {2D}}\): –23.4% ± 21.2, EDFN-D\(_{\text {2D}}\): –16.5% ± 17.6, GC\(_{\text {2D}}\): –6.8% ± 13.5, GC-D\(_{\text {2D}}\): –5.3% ± 12.7). These negative values, which were improved with depth-refinement, imply a bias towards under-segmenting an image. Refer to the discussion for clinical implications of this bias towards under-segmentation. Regarding the 3D percent intake error, the graph cut implementations outperformed our proposed system (EDFN\(_{\text {3D}}\): –9.1% ± 8.8, EDFN-D\(_{\text {3D}}\): –9.0% ± 8.9, GC\(_{\text {3D}}\): 0.4% ± 1.3, GC-D\(_{\text {3D}}\): 0.4% ± 1.3).

Volume estimation accuracy

Regarding the volume error (mL) on the regular texture dataset, while initially it appeared depth-refinement worsened performance across methods (EDFN: –14.7 mL ± 50.0, EDFN-D: –17.2 mL ± 50.3, GC: 0.0 mL ± 7.1, GC-D: –1.8 mL ± 6.9), the plate-level absolute volume error was improved with depth refinement and variance was much smaller (e.g., EDFN-D volume intake error –130.2 mL ± 154.8; EDFN-D mean absolute error 18.0 mL ± 50.0). However, intake error from volume was high and both our proposed system and graph cut implementations (EDFN: –129.2 mL ± 154.3, EDFN-D: –130.2 mL ± 154.8, GC: 1.8 mL ± 6.6, GC-D: 0.2 mL ± 6.5). Corroborated by the negative mean error bias for EDFN and EDFN-D, we empirically attribute this wide variance and high intake error paired with low plate-level absolute volume error to salad. As intake error is calculated with a plate relative to the full portion, variability in a food’s appearance could be high leading to high intake error, with accurate absolute volume. Salad, as part of the lunch plates had low-density with widely varying degrees of air pockets between leaves of lettuce at each plating. Depending on how “fluffy” the salad was put into position, it could take on differing volumes. This is discussed in detail later.

Modified texture foods

Food segmentation accuracy

Regarding the modified texture foods dataset, food segmentation accuracy was high for our proposed system, however depth-refinement reduced the accuracy based on classical segmentation accuracy calculations (EDFN: 0.922 ± 0.165, EDFN-D: 0.697 ± 0.348). In both cases however, our proposed system outperformed the graph cut analogs (GC: 0.834 ± 0.157, GC-D: 0.656 ± 0.328. Depth-refinement brings the context of volume, whereas the initial ground-truth hand segmentation used for comparison was based solely on visual appearance of food on the plate. Modified texture foods by nature tend to spread out more due to increased fluidity and as such, are at greater risk for the visual-volume discordance.

Segmentation agreement (IOU)

Compared to the regular texture food dataset, segmentation agreement (IOU) was adequate (over 0.80) and relatively unaffected by depth-refinement. Our proposed system was slightly outperformed by the graph cut implementation on IOU, with the graph cut variance smaller in non-depth-refined segmentation (EDFN: 0.846 ± 0.114, GC: 0.898 ± 0.082). Our proposed method was comparable to graph cut for the depth-refined counterparts (EDFN-D: 0.819 ± 0.166, GC-D: 0.819 ± 0.165).

Percent intake error (2D and 3D)

Regarding the 2D percent error of intake, non-depth-refined implementations had unacceptably high percent error of intakes (EDFN\(_{\text {2D}}\): –44.0% ± 38.6, GC\(_{\text {2D}}\): –27.4% ± 25.5) with unacceptably wide variances still present in depth-refined implementations (EDFN-D\(_{\text {2D}}\): –16.5% ± 20.0, GC-D\(_{\text {2D}}\): –14.5% ± 18.1). This provides additional evidence to support the need to rectify the visual-volume discordance, especially in modified texture foods where visually salient food remnants are more likely to remain on the plate after consumption. Depth-refinement had a minimal impact on percent intake error for all implementations yielding very low error in estimating volume intake error (EDFN\(_{\text {3D}}\): –0.8% ± 5.2, EDFN-D\(_{\text {3D}}\): 1.4% ± 5.9, GC\(_{\text {3D}}\): 0.7% ± 3.1, GC-D\(_{\text {3D}}\): 2.1% ± 4.6). Again, negative intake implications are addressed in the discussion.

Figure 4 depicts these visual-volume discordance errors and illustrates how depth-refinement (black lines) both reduced the variance of the percent intake errors as well as reduced the relative intake error compared to the non-depth-refined counterparts (red) and using P1 as the reference “full-portion”. Across methods, error tended to increase as the remaining portion size diminished with the exception of EDFN-D (2D) with error peaking at the third portion and receding across portions P4 and P5. We attribute this trend to depth-refinement compensating for higher degrees of visual-volume discordance on plates more likely to have smearing (i.e., plates with less on the plate relative to the initial portion size). This highlights why visual representation context of where food resides is inadequate and how additional depth context pertinent to how much food is present is required particularly for modified texture foods.

Figure 4
figure 4

Intake accuracy of our proposed method without (EDFN, red) and with (EDFN-D, black) depth-refinement on the modified texture foods dataset. 2D refers to two-dimensional % intake error in terms of the proportion of pixels estimated compared to the ground-truth hand segmentation of food areas. 3D refers to three-dimensional % intake error in terms of the relative volume across the estimated food segment compared to the volume across the ground-truth segmentation of food areas. Here, P1 is the reference plate with P2-P5 representing simulated degrees of increasing intake (i.e., P5 most eaten).

Volume estimation accuracy across portion sizes

Typically, volume errors, as well as volume intake errors, were low (less than 4.0 mL) across EDFN and GC implementations as shown in Table 3. To supplement this using our proposed EDFN and EDFN-D, Fig. 5 shows volume accuracy across each of the 5 portions (P1-P5) relative to the volume across the ground-truth hand segmented food regions. While our EDFN-D implementation was more accurate, it had similar precision to EDFN. A similar trend in reduced error in mL across smaller portions was observed for both EDFN and EDFN-D albeit with EDFN-D error shifted downwards. With EDFN-D, the three smallest portions (P3, P4, P5) had negative values indicating the depth-refinement omitted depth values across additional pixels relative to the initial segmentation and yielded a slight under-segmentation of food.

Figure 5
figure 5

Volume error of the modified texture dataset, in millilitres, of our proposed method without (EDFN) and with (EDFN-D) depth-refinement compared to the volume across the ground-truth hand segmented food area. P1 is the reference plate with P2-P5 representing simulated degrees of increasing intake (i.e., P5 most eaten).

Discussion

Our proposed fully automatic system imparts a reduced processing burden compared to semi-automatic segmentation. Using graph cuts as our “applied ground-truth”, task completion time represents an additional point for consideration, and our automated approach provides a key advantage in the food and nutrition tracking context. Empirically, user-defined seed initialization for graph cut implementation incurred approximately 5 s of manual annotation time per image. Assuming 192 residents across 6 neighbourhoods (units) in LTC, this implies 48 additional minutes during a meal-service simply to annotate the images. The average time for charting residents’ food intake for a day is already at least 270 minutes12, which implies that annotation could impart an 18% time increase to complete food intake charting. This approach is infeasible and prohibitive within this context. Instead, compared to the graph cuts method which had a tendency to under-segment food, our proposed automatic segmentation method requires minimal additional time commitment from the user enhancing its potential for uptake in the field.

Segmentation errors (under- or over-segmentation) have clinical significance for identification of inadequate intake before low intake progresses to eventual malnutrition. Estimated intake could be incorrect due to either under-estimating food consumption (i.e., less food reported than actually consumed) or over-estimating food consumption (i.e., more food reported than actually consumed). Over-segmenting food areas (i.e., under-estimating intake) implies the predicted food area is over-estimated and translates to reporting that there is more food present (less food intake) than what is true. Conversely, under-segmenting food areas (i.e., over-estimating intake) implies the predicted food area is under-estimated and translates to reporting there is less food present (more food intake) than there is. Over-segmentation (under-estimating food consumption) may be the lesser of two evils for the majority of LTC residents; if food intake is better than what is reported, residents at-risk for malnutrition may be less likely to be missed. Clinically, residents receiving modified texture foods consume significantly fewer calories, have higher cognitive impairment, and require more assistance with activities of daily living than residents on regular texture diets62 making these even higher-risk residents. However, under-segmentation (over-estimating food consumption), increases risk of missing residents with nutrient inadequacies through introducing false negatives and potentially missing residents with poor food intake that could result in malnutrition. While under-reporting may lead to increased referrals for malnutrition screening, most at-risk residents who eat very little would still be identified which might help to identify residents who could benefit from a dietary intervention.

It appears segmentation is not enough and there is a need for additional context from volume. We observed that under-segmentation was an issue mostly with 2D % volume intake across implementations and datasets. We evaluated 2D % volume intake for consistency with existing methods, however, the system’s 3D intake estimation is preferred whenever possible. Additionally, returning to Fig. 4, this plot showcases the need for methods beyond segmentation for food intake tracking. While we observed a decrease in 2D % intake error with depth-refinement across the remaining plate portions, errors for P2, P3, and P4 were more than 20%. A system based on segmentation methods alone may be accurate when very little food is consumed or nearly all the food is consumed. Certainly, there is value in improved accuracy for these edge-cases24. However, were we to rely on a system without-depth refinement, this degree of error may be deemed inappropriately high and be a barrier to uptake. Further refinements from depth-refined volume estimations yielded additional improvement in 3D percent error intake and may provide a palatable alternative within an acceptable error margin of less than 10% (where current practice is up to 62% error11; 50% for portion size10) with the added benefit of more fine-grained assessment (continuous measurement versus 25% incremental bins). While we also observed negative values for volume intake error for our proposed system, EDFN-D, this error was largely due to salad. It appeared as though GC seemed to understand the visual representation of salad better than our proposed method. We suspect this could be improved with additional instances of salad included in the training and validation set. Green vegetables in general were under-represented in UNIMIB2016 and provides an opportunity area for additional comprehensive food intake tracking databases. Whereas the volume intake error for the regular texture foods which included salad was on the order of 130 mL, the modified texture foods circumvented the low density foods problem with volume intake error of 0.8 mL. Arguably, accuracy on these modified texture foods is more clinically relevant for at-risk residents. Additionally, building in some depth redundancy may better approximate a human assessor and improve initial acceptability.

Perhaps more pertinent to uptake of the system and recurring use of the system is how the computer “sees” food on a plate and how this compares to the human experience for supporting trust in the automated system. For example, from the human assessor perspective, a plate containing sauce remnants would be ignored and treated as completely consumed food while the computer vision approach would observe each of the pixels containing sauce remnants and mark it as still containing food as in the case of Fig. 3. This may lead to distrust in the system because the system makes decisions differently and in opposition to how a human assessor would classify the presence of food. Borrowing from clinicians perceptions of artificial intelligence tools, the alignment between the system output and what would be expected from a human’s interpretation is essential for continued use63. By incorporating depth-refinement, even though the IOU for modified texture foods decreased, it brings the assessment closer to how a human would interpret a plate with the added benefits of greater precision and objectivity. This area has been largely unexplored because, with one exception, available food datasets only include full plates (see Supplementary Materials for an overview of available food datasets). For measuring food intake, additional application appropriate metrics such as intake error and volume estimation accuracy must be considered since assessing system performance solely from a segmentation accuracy perspective can be misleading.

While our proposed system is not error-free, it is significantly more accurate than current LTC methods. As aforementioned, current LTC home food intake accuracy shows correct estimation of intake occurring as low as 38% of the time11 and when portion size is mis-estimated, has error up to 50%10. Part of the issue, in practice, may be that the granularity of these estimates is also wide since estimates are recorded as 25% incremental food intake bins64,65. This may introduce further subjectivity between assessors. Hypothetically, one human assessor may estimate a plate to be 30% eaten so a value of 25% would be recorded; another assessor may estimate intake at 45% reflecting a record of 50% consumed. With depth-refinement, our proposed EDFN-D removes this subjectivity, operates on a continuous scale, and has a mean 3D % intake of estimation error of -4.2% across both datasets and a mean volume intake error of 0.8 mL on the modified texture foods dataset.

To further improve reliability and consistency of the system, negative volume errors and the issue of low-density foods should be considered. For the proposed EDFN-D, the volume errors for plates after the highest intake (P4 and P5) were negative because the depth-refinement omitted volumes across pixels that were initially segmented as food. As part of future work, it would be interesting to explore how necessary initial colour-based segmentation is for establishing how much food is present or whether placing greater weight on depth maps could improve accuracy for volume estimation accuracy. Given the issue of low-density foods impacting volume regardless of segmentation method, this must be considered a limitation of over-head food intake systems and these types of foods (e.g., potato chips, salad) may need to be treated differently and separately to other foods. Perhaps repeat imaging of these foods separately, flattening the food before imaging, or applying a general food density score to estimate the range of volume values in these foods could address this limitation.

In summary, we proposed an application-driven design for a novel fully automatic multisensor segmentation system which leverages depth-refinement for improved accuracy. We assessed our system on two representative LTC food intake datasets which included simulated intake plates since current datasets contain only full-plate portions or do not contain pixel-level segmentations. For further advancing the field, additional food intake datasets with pixel-level segmentation are needed. A system such as the one presented here which approximates a human assessor but is objective, faster, more consistent, and can more accurately quantify food intake measurements may provide a valuable step towards automated tracking of food and fluid intake within the LTC sector.

Methods

Data collection

Data were collected in an industrial research kitchen which conforms to LTC kitchen standards. We constructed an image acquisition system that enabled top-down image capture. We imaged 36 foods representative of LTC where 9 were regular texture foods listed as options on a LTC menu comprising our novel “regular foods” dataset. A set of 63 modified texture food samples representing 27 unique foods were prepared by a LTC kitchen (The University Gates, Schlegel Villages) and either minced or pureed comprising our “modified texture foods” dataset. During image acquisition, the room temperature varied from 20.6\(^{\circ }\)C to 22.5\(^{\circ }\)C. Images were saved to a computer for model training and evaluation. For a summary of the types and representation of foods imaged, see Table 1.

Regular texture foods acquisition

Three representative meals each consisting of three food items (breakfast: oatmeal, toast, eggs; lunch: pasta, salad, cookie; and dinner: meatloaf, mashed potatoes, corn) were selected from an LTC menu and imaged as part of this data collection series. Each plate was assembled with up to three food items. One full serving of each food item was defined by the nutritional label serving size.

Plates were imaged at every permutation of 0%, 25%, 50%, 75%, 100% of each food item consumed. Here, 0% corresponds to the initial, largest mass portion (P1), and 100% corresponds to no amount of that food component remaining (P5). The largest mass portion, P1 was deemed a “full” portion with P2-P5 representing smaller and smaller masses. These 25% incremental bins were selected based on standard dietary intake record forms used in LTC66,67. This yielded 125 unique plates per meal (375 unique plates). Foods were selected to be representative based on a LTC menu.

Modified texture foods acquisition

We imaged 63 food samples (56 unique + 7 duplicates) representing 27 unique food items. Each set of samples for a given unique food contained at least one example of a modified texture (i.e., minced, pureed or both) either imaged fresh, after being held at serving temperature, or both. Holding food at serving temperature is standard practice in LTC serveries as meal items are prepared in advance of meal service. Each food sample was imaged at 5 different portions (P1-P5), with one exception containing 4 portions, by progressively removing some of the sample with a spoon to simulate varying degrees of leftovers for a total of 314 images. The largest mass portion, P1, was deemed a “full” portion for intake purposes, with P2-P5 representing smaller and smaller masses. Foods were representative of a typical LTC menu as they were prepared by the LTC kitchen and represented a variety of fruits, vegetables, pastas, soup, and meat dishes.

Imaging system for food volume estimation system

The goal was to estimate volumetric food intake from RGB-D images of food on a plate. We developed a deep convolutional neural network (DCNN) for generating food segmentation maps, which was refined using depth heuristics and combined with calibrated pixel-wise food heights to estimate food consumption (in mL). Figure 1 shows a visual representation of the system diagram. The following subsections describe the primary subsystems in more detail.

Encoder decoder food network (EDFN) architecture

The proposed EDFN system comprises a semantic food segmentation backbone and a per-pixel volumetric analysis pipeline for assessing food volume and computing food intake. The semantic segmentation module was inspired by the success of encoder-decoder networks for semantic image segmentation61,68. In addition to the unique local colour and texture properties of food, inter- and intra-food placement on a plate exhibits strong global spatial properties when considering a top-down view. Thus, we chose PSPNet61 as the segmentation backbone due to its ability to encode multiscale global information through pyramid pooling. More specifically, we wanted multiple scales to be represented in learned features as the texture of a single food is important as well how foods clump together and are presented across a plate. PSPNet accounts for both the macro and micro scales. Thus, we designed the macroarchitecture of the proposed food segmentation DCNN as a multi-scale encoder-decoder network architecture tailored for downsampled, pixel-level semantic segmentation of food images. Figure 1 shows the network architecture, which consists of a residual encoder microarchitecture, a multi-scale hierarchical decoder microarchitecture, and a final high-resolution, per-pixel classification layer for producing a food segmentation map. The residual encoder microarchitecture is responsible for encoding RGB images into a set of feature maps describing the objects in the image. The encoder feature map outputs are then processed through the decoder microarchitecture which parses the scene at multiple spatial scales. These multi-scale representations were concatenated to the feature map outputs, and a 1x1 convolutional layer was trained to output a two-class per-pixel segmentation map.

For the residual encoder microarchitecture, we leveraged a spliced ResNet101 architecture with pre-activation60. The ResNet101 architecture was chosen because of its powerful representational capability for learning discriminative feature representations from complex scenes. We leveraged the notion of transfer learning by beginning with a ResNet101 network architecture designed for classification, trained on the ImageNet dataset of natural scenes69, and splicing off the deeper ResNet101 layers to create the final encoder microarchitecture. More specifically, we splice at the third unit of the first residual block61, leading to the proposed residual encoding microarchitecture, which encodes 120\(\times \)160 RGB images into 256 15\(\times \)20 feature maps. As such, the image was fed through a 7\(\times \)7 convolutional layer with 64 kernels and a stride of 2. Then, a 3\(\times \)3 max pool with stride of 2 was performed to downsample the image. These representations were fed through the first ResNet101 block, consisting of 64 1\(\times \)1 convolution, 64 3\(\times \)3 convolution, and 256 1\(\times \)1 convolution layers three times, with skip connections after every set of 3 layers. The last 3\(\times \)3 layer was downsampled using a stride of 2. Thus, the encoder microarchitecture outputs 256 feature maps at 1/8 the input image size.

The decoder microarchitecture of the proposed food segmentation network was designed to decode the feature maps from the encoder microarchitecture into hierarchical global priors using a region binning scene parsing network architecture design. It is well known that multi-scale context aids pixel segmentation70 which is particularly relevant within the context of food. As humans observing food, there are two main components: the colour and the texture of the food. Texture also varies across scales (i.e., food has a hierarchical visual nature to it). To account for the multi-scale context of food, we leveraged a pyramid scene parsing network (PSPNet)61 which was connected to the feature outputs from the encoder microarchitecture. As such, the PSPNet decoder microarchitecture performs analysis across four spatial scales, which adds information representing the underlying feature representation and provides local-to-global context of the plate of food. The feature maps were fed into four parallel max-pool layers, with bin sizes of 1\(\times \)1, 2\(\times \)2, 3\(\times \)3, and 6\(\times \)6. The upscaled hierarchical global prior outputs were concatenated to the encoder feature maps and two class (food or not food) pixel-level segmentation was performed using a 1\(\times \)1 convolution layer. A circle Hough transform71 was used to mask the plate from the table, eliminating detection of food outside the plate boundaries (e.g., on tables with complex patterns).

Training and validation dataset selection

Evaluating the appropriateness of training and validation dataset selection for application to food intake assessment in LTC requires several considerations:

  • Colour Food comes in many colours. Part of supporting a healthy diet includes the mentality of “eat a rainbow” to ensure various micronutrient needs in additional to macronutrient needs are met72. As such, there should be a wide distribution of colours naturally found in foods captured in the training and validation datasets.

  • Texture Similar to colour, foods also inherently come in a variety of textures. This aspect is particularly salient when considering the LTC population where 47% of residents receive modified texture diets (e.g., minced, pureed) as part of a strategy to address the high prevalence of swallowing difficulties pervasive in LTC73.

  • Portion Given the application to food intake assessment, the ideal training dataset would have representative intake images that include both before and after images of meals (i.e., not just full portions of served food). The rationale is by having more representative examples of what foods can look like partially, or fully disassembled in the case of food mixing, in the training dataset, the network will be more likely to learn representative examples of foods apart from the original served context.

  • Orientation The occlusion conundrum where one food occludes another is very difficult to circumvent especially in the LTC environment when taking multiple images from many perspectives is infeasible due to time constraints. For the purpose of acquisition “in the wild” within LTC, images taken from the above configuration is preferred to facilitate volume estimation and down-stream nutritional intake estimation. The issue of occlusion (e.g., seeing only the bun as part of an assembled hamburger) remains an issue, however is reasonable when accepting the assumption that complex foods (e.g., foods with multiple components like a hamburger) are eaten in similar proportions. While this is an undoubtedly fallible assumption, errors in down-stream nutritional estimation are constrained to a specific food as opposed to influencing the entire plate. As such, the ideal training and validation database would be acquired in the top-view configuration.

  • Label level To facilitate volume estimation and down-stream nutrient estimation, image segmentation must be conducted pixel-wise as opposed to using bounding boxes. As such, the ideal training and validation database would be labelled at the pixel-level.

  • Style Multiple food items on a plate is common in LTC as the “family style” approach to eating has been shown to enhance food intake74. While multiple plates may also be used in LTC (e.g., soup, side salad, dessert), the training and validation dataset would ideally show representation consistent with “family style” as opposed to solely “single item” plates.

  • Accessibility The ideal dataset needs to be readily available (i.e., non-proprietary).

Based on the above considerations, the most imperative are orientation and label level and the most suitable food database available for training and validation for LTC (i.e., top-view with pixel-wise labelling) is therefore UNIMIB201659. Refer to Supplementary Materials for a compilation of popular food databases with a summary of rationale for selecting UNIMIB2016.

Training

We trained the proposed encoder-decoder food network (EDFN) on the UNIMIB2016 food dataset (1027 tray images, 73 categories), which contains per-pixel ground-truth segmentation59. The encoder weights were frozen to conserve deep computational feature extraction from large robust datasets, and only the decoder weights were optimized. The UNIMIB2016 dataset was chosen due to its food variety, overhead view, and pixel-level segmentation annotation. Additionally, since our method was driven by LTC application requirements with data collection in a specific manner, we needed a dataset that was similarly acquired (e.g., pixel-wise annotation, not bounding boxes) so training/fine-tuning could be accomplished without bias by our novel LTC test datasets. While our LTC test datasets comprise a representative sample albeit with relatively few food groupings, training on the UNIMIB2016 dataset with 76 food categories enhanced generalizability. Downsampling was conducted to align the spatial feature sizes of the encoder and decoder microarchitectures with the UNIMIB2016 dataset59. The UNIMIB2016 data were resized to match our image height/width which were at the same aspect ratio (4:3). This resizing to 120x160 images provided two key advantages: (1) computation reduction, (2) better scaled kernels for the image size. We empirically observed that there was not enough global context at the original resolution, resulting in the middle of foods getting misclassified. By downsampling our image, the network was able to identify primary low-level features instead of getting stuck in the texture of the food and could be successfully decoded by the pyramid scene parsing decoder microarchitecture. The UNIMIB2016 data were randomly split into training and validation subsets (80%/20%). Since all UNIMIB2016 plates were placed on the same tan coloured tray, we found that the machine learning model inappropriately learned that the tray colour was always indicative of non-food. Thus, we performed data augmentation on the UNIMIB2016 data by randomly rotating the hue channel of each image’s background (non-food) pixels and adding it to the dataset, thus effectively doubling the training and validation datasets. The network was trained using batch size 32 using RMSProp optimizer with softmax cross-entropy loss, a learning rate of 0.0001 and a decay of 0.995. The network was trained over 200 epochs, and the best model according to the validation loss was kept.

Segmentation depth-refinement

The generated food segmentation map from EDFN is based solely on visual information, and is thus privy to visual-volume discordance. We therefore developed a heuristic for excluding labeled food areas that are irrelevant to food consumption (e.g., pasta sauce remnants). To do this, co-aligned depth maps were acquired synchronously with the RGB images for the plate under analysis as well as an empty calibration plate.

Ten depth maps were averaged for each acquisition to account for measurement noise. The calibration depth map \(d_C\) was registered to the plate depth map \(d_P\) to account for any changes in camera-plate orientation between calibration and plate acquisitions. To accomplish this, the plate edges were identified from the RGB plate images using a Canny edge detector with a Gaussian filter of \(\sigma \)=3 and hysteresis threshold values of 10 and 50. A circle Hough transform, using the range of expected plate radii, was performed on these edge maps to determine the plate center and radius. Denoting the plate centers of the calibration and food plate as \((x_C, y_C)\) and \((x_P, y_P)\), a translation transformation \(T=(x_P-x_C, y_P-y_C)\) was applied to \(d_C\), thus aligning the two depth maps. Pixel-wise food height was then computed:

$$\begin{aligned} h_i = d_{C,i}-d_{P,i} \end{aligned}$$
(1)

where \(h_i\) is the food height relative to the plate for pixel i in mm. The transformation T accounts for planar translation between acquisitions, but changes in tilt may also affect the measurement. Thus, a full-field correction was performed by constraining the left and right limits of \(h_i\) (table regions) to 0 mm. Specifically, for each row, we subtracted the weighted average of the left and right table pixel heights (which should theoretically be 0 mm), weighted to the pixel’s distance from the left and right boundaries. A \(5\times 5\) median filter was applied to the food height map to correct spurious measurements at plate boundaries.

Food height was used to refine the segmentation mask (“depth-refinement”) based on a priori knowledge that visual-volume discordance is observed when very shallow and inconsequential foods are visually apparent, but are irrelevant to volumetric analysis. Specifically, the image was decomposed into 250 perceptually meaningful superpixels using simple linear iterative clustering75 (compactness=20, \(\sigma \)=2). For each superpixel \(\mathcal {S}_i\), the constituent pixels were removed from the food map using statistical thresholding on the pixel height distribution:

$$\begin{aligned} Q_{h_{\mathcal {S}_i}}(p) < \tau \end{aligned}$$
(2)

where \(Q_{h_{\mathcal {S}_i}}\) is the quantile function of the distribution of pixel heights in \(\mathcal {S}_i\). We set \(p=0.75\) and \(\tau =2\) mm based on measurement error along a flat table.

Food volume calculation

Food height was determined in mm units, but to calculate food volume, pixel spacing needed to be calibrated to mm (a pixel-to-mm conversion). Using the known diameter of the plate d (259 mm), we used the detected plate radius \(\hat{r}\) from the circle Hough transform to compute the conversion:

$$\begin{aligned} \Delta x = \frac{d}{2\hat{r}} \end{aligned}$$
(3)

Food volume in mL could then be computed by summing the per-pixel differential volumes within the food mask:

$$\begin{aligned} V = \sum _{i \in F} (\Delta x)^2 h_i \end{aligned}$$
(4)

where F is the set of segmented food pixels. Volumetric food intake was computed by subtracting the plate volume from the full portion volume. Similarly, percent intake was calculated relative to the full portion volume.

Testing

We tested the network on our two custom LTC datasets consisting of 689 (375+314) images representing 36 (9+27) different foods (as outlined in Table 1). Original images were downsampled from 480\(\times \)640 to 120\(\times \)160 to decrease the number of network parameters and improve computation time. The images were hand segmented to define ground truth segmentation masks of the food on the plates.

We compared our results to those generated by semi-automatic graph cut segmentation. Since user input is required for initialization, for consistency in the regular texture dataset, one line was used to denote each food item present on the plate and one squiggled background line was indicated around the top and right side of the image as shown in Fig. 6. The modified texture dataset required additional user-defined seeding. The circle Hough transform plate masking used in our proposed system was used here too. The output from this method is a plate-level food segmentation mask.

Data analysis

To compare quantitative performance between methods, we use the common performance measures of global accuracy (Equation 5) to describe the percentage of correctly classified pixels, food segmentation accuracy (Equation 6 to describe the percentage of correctly classified food pixels, as well as the intersection over union (IOU) (Equation 7) both within a meal (i.e., breakfast, lunch, dinner, modified texture single-imaged foods) and across meals. For this application, the IOU provides a more representative metric for how our segmentation system is performing as it captures accuracy within the context of the true bounded food areas since false positive predictions are penalized. The theoretical maximum value of IOU is 1.0 when the intersection maps perfectly over the union without deviation. We define the metrics described above as follows:

$$\begin{aligned} \text {Global Accuracy}&= \frac{TP + TN}{TP + TN + FP + FN} \end{aligned}$$
(5)
$$\begin{aligned} \text {Food Segmentation Accuracy}&= \frac{TP}{TP+FN} \end{aligned}$$
(6)
$$\begin{aligned} IOU&= \frac{target \cap prediction}{target \cup prediction} \end{aligned}$$
(7)

Volume estimation accuracy was assessed by computing mean absolute intake error (mL) as volume calculated using our proposed method or the “applied ground-truth” method with or without depth-refinement relative to the volume across the ground-truth hand segmented areas. Error (mL) was calculated similarly but preserves the direction of error. Volume intake error (mL) is the difference between the current portion relative to the full portion. Intake error was calculated for both segmentation (2D) and volume (3D) data relative to the full portion. All values are reported as mean ± SD.

Figure 6
figure 6

Sample graph cuts annotation with one line per food item (red) and one background line (blue) and resulting segmentation mask.

Data availability

Data are available by contacting the corresponding author on reasonable request.

References

  1. Pirlich, M. & Lochs, H. Nutrition in the elderly. Best Pract. Res. Clin. Gastroenterol. 15, 869–884 (2001).

    CAS  PubMed  Google Scholar 

  2. Keller, H. H., Østbye, T. & Goy, R. Nutritional risk predicts quality of life in elderly community-living Canadians. J. Gerontol.: Ser. A 59, M68–M74 (2004).

    Google Scholar 

  3. Goates, S., Du, K., Braunschweig, C. A. & Arensberg, M. B. Economic burden of disease-associated malnutrition at the state level. PLoS ONE 11, e0161833 (2016).

  4. Russell, C. A. The impact of malnutrition on healthcare costs and economic considerations for the use of oral nutritional supplements. Clin. Nutr. Suppl. 2, 25–32 (2007).

    Google Scholar 

  5. Kaiser, M. J. et al. Frequency of malnutrition in older adults: a multinational perspective using the mini nutritional assessment. J. Am. Geriatr. Soc. 58, 1734–1738 (2010).

    PubMed  Google Scholar 

  6. Keller, H. et al. Prevalence of malnutrition or risk in residents in long term care: Comparison of four tools. J. Nutrit. Gerontol. Geriatrics 38, 329–344 (2019).

    Google Scholar 

  7. Keller, H. H. et al. Prevalence and determinants of poor food intake of residents living in long-term care. J. Am. Med. Dir. Assoc. 18, 941–947 (2017).

    PubMed  Google Scholar 

  8. Martin, C. K. et al. A novel method to remotely measure food intake of free-living individuals in real time: the remote food photography method. Br. J. Nutr. 101, 446–456 (2008).

    PubMed  PubMed Central  Google Scholar 

  9. Williamson, D. A. et al. Comparison of digital photography to weighed and visual estimation of portion sizes. J. Am. Diet. Assoc. 103, 1139–1145 (2003).

    PubMed  Google Scholar 

  10. Bingham, S. A. Limitations of the various methods for collecting dietary intake data. Ann. Nutr. Metab. 35, 117–127 (1991).

    CAS  PubMed  Google Scholar 

  11. Castellanos, V. H. & Andrews, Y. N. Inherent flaws in a method of estimating meal intake commonly used in long-term-care facilities. J. Am. Diet. Assoc. 102, 826–830 (2002).

    PubMed  Google Scholar 

  12. Pfisterer, K., Boger, J. & Wong, A. Prototyping the automated food imaging and nutrient intake tracking (AFINI-T) system: A modified participatory iterative design sprint. JMIR Hum. Factors 6, e13017 (2019).

  13. Kong, F. Automatic Food Intake Assessment Using Camera Phones. Ph.D. thesis, Michigan Technological University (2012).

  14. Meyers, A. et al. Im2calories: towards an automated mobile vision food diary. In Proceedings of the IEEE International Conference on Computer Vision, 1233–1241 (2015).

  15. Okamoto, K. & Yanai, K. An automatic calorie estimation system of food images on a smartphone. In Proceedings of the 2nd International Workshop on Multimedia Assisted Dietary Management, 63–70 (ACM, 2016).

  16. Pouladzadeh, P., Shirmohammadi, S. & Yassine, A. You are what you eat: So measure what you eat!. IEEE Instrum. Measure. Mag. 19, 9–15 (2016).

    Google Scholar 

  17. Aslan, S., Ciocca, G. & Schettini, R. Semantic food segmentation for automatic dietary monitoring. In Proceedings of the IEEE International Conference on Consumer Electronics-Berlin, 1–6 (2018).

  18. Kong, F., He, H., Raynor, H. A. & Tan, J. DietCam: Multi-view regular shape food recognition with a camera phone. Pervasive Mob. Comput. 19, 108–121 (2015).

    Google Scholar 

  19. Aguilar, E., Remeseiro, B., Bolaños, M. & Radeva, P. Grab, pay, and eat: Semantic food detection for smart restaurants. IEEE Trans. Multimedia 20, 3266–3275 (2018).

    Google Scholar 

  20. Doulah, A., Mccrory, M. A., Higgins, J. A. & Sazonov, E. A systematic review of technology-driven methodologies for estimation of energy intake. IEEE Access 7, 49653–49668 (2019).

    PubMed  PubMed Central  Google Scholar 

  21. Boushey, C. J., Spoden, M., Zhu, F. M., Delp, E. J. & Kerr, D. A. New mobile methods for dietary assessment: Review of image-assisted and image-based dietary assessment methods. Proc. Nutrit. Soc. 76, 283–294. https://doi.org/10.1017/S0029665116002913 (2017).

    CAS  Article  Google Scholar 

  22. Pettitt, C. et al. A pilot study to determine whether using a lightweight, wearable micro-camera improves dietary assessment accuracy and offers information on macronutrients and eating rate. Br. J. Nutr. 115, 160–167 (2016).

    CAS  PubMed  Google Scholar 

  23. Beltran, A. et al. Adapting the ebutton to the abilities of children for diet assessment. In Proceedings of Measuring Behavior 2016: 10th International Conference on Methods and Techniques in Behavioral Research. International Conference on Methods and Techniques in Behavioral Research (10th: 2016: Dublin, Ireland), vol. 2016, 72 (NIH Public Access, 2016).

  24. Parent, M., Niezgoda, H., Keller, H. H., Chambers, L. W. & Daly, S. Comparison of visual estimation methods for regular and modified textures: Real-time vs digital imaging. J. Acad. Nutr. Diet. 112, 1636–1641 (2012).

    PubMed  Google Scholar 

  25. Subhi, M. A., Ali, S. H. & Mohammed, M. A. Vision-based approaches for automatic food recognition and dietary assessment: A survey. IEEE Access 7, 35370–35381 (2019).

    Google Scholar 

  26. Astell, A. J. et al. Validation of the nana (novel assessment of nutrition and ageing) touch screen system for use at home by older adults. Exp. Gerontol. 60, 100–107 (2014).

    CAS  PubMed  Google Scholar 

  27. Lo, F. P. W., Sun, Y., Qiu, J. & Lo, B. Image-based food classification and volume estimation for dietary assessment: A review. IEEE J. Biomed. Health Inform. 24, 1926–1939 (2020).

    PubMed  Google Scholar 

  28. Bruno, V. & Silva Resende, C. J. A survey on automated food monitoring and dietary management systems. J. Health Med. Inform.8 (2017).

  29. Pouladzadeh, P., Shirmohammadi, S. & Al-Maghrabi, R. Measuring calorie and nutrition from food image. IEEE Trans. Instrum. Meas. 63, 1947–1956 (2014).

    Google Scholar 

  30. Zhu, F., Bosch, M., Khanna, N., Boushey, C. J. & Delp, E. J. Multiple hypotheses image segmentation and classification with application to dietary assessment. IEEE J. Biomed. Health Inform. 19, 377–388 (2015).

    PubMed  PubMed Central  Google Scholar 

  31. Shimoda, W. & Yanai, K. CNN-based food image segmentation without pixel-wise annotation. In Proceedings of the International Conference on Image Analysis and Processing, 449–457 (2015).

  32. Kawano, Y. & Yanai, K. Foodcam: A real-time food recognition system on a smartphone. Multimed. Tools Appl. 74, 5263–5287 (2015).

    Google Scholar 

  33. Pouladzadeh, P., Kuhad, P., Peddi, S. V. B., Yassine, A. & Shirmohammadi, S. Food calorie measurement using deep learning neural network. In Proceedings of the IEEE International Instrumentation and Measurement Technology, 1–6 (2016).

  34. Hassannejad, H. et al. A new approach to image-based estimation of food volume. Algorithms 10, 66 (2017).

    MathSciNet  MATH  Google Scholar 

  35. Boykov, Y. Y. & Jolly, M.-P. Interactive graph cuts for optimal boundary & region segmentation of objects in ND images. In Proceedings of the IEEE International Conference on Computer Vision 1, 105–112 (2001).

    Google Scholar 

  36. He, Y., Xu, C., Khanna, N., Boushey, C. J. & Delp, E. J. Food image analysis: segmentation, identification and weight estimation. In Proceedings of the IEEE International Conference on Multimedia and Expo, 1–6 (2013).

  37. Wang, Y., He, Y., Boushey, C. J., Zhu, F. & Delp, E. J. Context based image analysis with application in dietary assessment and evaluation. Multimed. Tools Appl. 77, 19769–19794 (2018).

    PubMed  Google Scholar 

  38. Yunus, R. et al. A framework to estimate the nutritional value of food in real time using deep learning techniques. IEEE Access 7, 2643–2652 (2018).

    Google Scholar 

  39. Zheng, X., Lei, Q., Yao, R., Gong, Y. & Yin, Q. Image segmentation based on adaptive k-means algorithm. EURASIP J. Image Video Process. 2018, 68 (2018).

    Google Scholar 

  40. Ciocca, G., Mazzini, D. & Schettini, R. Evaluating CNN-based semantic food segmentation across illuminants. In Proceedings of the International Workshop on Computational Color Imaging, 247–259 (2019).

  41. Xu, C., He, Y., Khanna, N., Boushey, C. J. & Delp, E. J. Model-based food volume estimation using 3d pose. In Proceedings of the 2013 20th IEEE International Conference on Image Processing (ICIP), 2534–2538 (2013).

  42. Chae, J. et al. Volume estimation using food specific shape templates in mobile image-based dietary assessment. Proc. SPIE 7873, 78730K (2011).

    PubMed Central  Google Scholar 

  43. Jia, W. et al. Accuracy of food portion size estimation from digital pictures acquired by a chest-worn camera. Public Health Nutr. 17, 1671–1681 (2014).

    PubMed  Google Scholar 

  44. Ofei, K. T., Mikkelsen, B. E. & Scheller, R. A. Validation of a novel image-weighed technique for monitoring food intake and estimation of portion size in hospital settings: a pilot study. Public Health Nutr. 22, 1203–1208 (2019).

    PubMed  Google Scholar 

  45. Rachakonda, L., Mohanty, S. P. & Kougianos, E. iLog: An intelligent device for automatic food intake monitoring and stress detection in the IoMT. IEEE Trans. Consum. Electron. 66, 115–124 (2020).

    Google Scholar 

  46. Herzig, D. et al. Volumetric food quantification using computer vision on a depth-sensing smartphone: Preclinical study. JMIR Mhealth Uhealth 8, e15294 (2020).

  47. Dehais, J., Anthimopoulos, M., Shevchik, S. & Mougiakakou, S. Two-view 3d reconstruction for food volume estimation. IEEE Trans. Multimedia 19, 1090–1099 (2017).

    Google Scholar 

  48. Puri, M., Zhu, Z., Yu, Q., Divakaran, A. & Sawhney, H. Recognition and volume estimation of food intake using a mobile device. In 2009 Workshop on Applications of Computer Vision (WACV), 1–8 (2009).

  49. Rahman, M. H. et al. Food volume estimation in a mobile phone based dietary assessment system. In International Conference on Signal Image Technology and Internet Based Systems, 988–995 (2012).

  50. Fang, S. et al. A comparison of food portion size estimation using geometric models and depth images. In Proceedings of the IEEE International Conference on Image Processing (ICIP), 26–30 (2016).

  51. Shang, J. et al. A mobile structured light system for food volume estimation. In IEEE International Conference on Computer Vision Workshops (ICCV Workshops), 100–101 (2011).

  52. Chen, M.-Y. et al. Automatic chinese food identification and quantity estimation. In SIGGRAPH Asia 2012 Technical Briefs, 29 (2012).

  53. Liao, H.-C., Lim, Z.-Y. & Lin, H.-W. Food intake estimation method using short-range depth camera. In Signal and Image Processing (ICSIP), IEEE International Conference on, 198–204 (2016).

  54. Zhou, L., Zhang, C., Liu, F., Qiu, Z. & He, Y. Application of deep learning in food: A review. Comprehens. Rev. Food Sci. Food Saf. 18, 1793–1811 (2019).

    Google Scholar 

  55. Ferdinand Christ, P. et al. Diabetes60-inferring bread units from food images using fully convolutional neural networks. In Proceedings of the IEEE International Conference on Computer Vision Workshops, 1526–1535 (2017).

  56. Lo, F.P.-W., Sun, Y., Qiu, J. & Lo, B. Food volume estimation based on deep learning view synthesis from a single depth map. Nutrients 10, 2005 (2018).

    PubMed Central  Google Scholar 

  57. Lo, F.P.-W., Sun, Y., Qiu, J. & Lo, B. P. Point2volume: A vision-based dietary assessment approach using view synthesis. IEEE Trans. Industr. Inf. 16, 577–586 (2019).

    Google Scholar 

  58. Abdul, A., Vermeulen, J., Wang, D., Lim, B. Y. & Kankanhalli, M. Trends and trajectories for explainable, accountable and intelligible systems: an HCI research agenda. In Proceedings of the CHI Conference on Human Factors in Computing Systems, 582 (2018).

  59. Ciocca, G., Napoletano, P. & Schettini, R. Food recognition: A new dataset, experiments and results. IEEE J. Biomed. Health Inform. 21, 588–598 (2017).

    PubMed  Google Scholar 

  60. He, K., Zhang, X., Ren, S. & Sun, J. Identity mappings in deep residual networks. In Proceedings of the European Conference on Computer Vision, 630–645 (2016).

  61. Zhao, H., Shi, J., Qi, X., Wang, X. & Jia, J. Pyramid scene parsing network. In Proceedings of IEEE Conference on Computer Vision and Pattern Recognition (2017).

  62. Keller, H. H. et al. Prevalence of inadequate micronutrient intakes of Canadian long-term care residents. Br. J. Nutr. 119, 1047–1056 (2018).

    CAS  PubMed  Google Scholar 

  63. Tonekaboni, S., Joshi, S., McCradden, M. D. & Goldenberg, A. What clinicians want: contextualizing explainable machine learning for clinical end use. In Machine Learning for Healthcare Conference, 359–380 (2019).

  64. MED-PASS, I. Dietary intake record - 100/pad (2017).

  65. Healthcare, B. Food intake record form top-punch (2017).

  66. MED-PASS. Dietary intake form (2017).

  67. BRiGGS Healthcare. Dietary intake form (2017).

  68. Badrinarayanan, V., Kendall, A. & Cipolla, R. SegNet: A deep convolutional encoder-decoder architecture for image segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 39, 2481–2495 (2017).

    PubMed  Google Scholar 

  69. Deng, J. et al. ImageNet: A large-scale hierarchical image database. In IEEE Conference on Computer Vision and Pattern Recognition, 248–255 (2009).

  70. Liu, C. et al. Auto-DeepLab: hierarchical neural architecture search for semantic image segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 82–92 (2019).

  71. Atherton, T. J. & Kerbyson, D. J. Size invariant circle detection. Image Vis. Comput. 17, 795–803 (1999).

    Google Scholar 

  72. Minich, D. M. A review of the science of colorful, plant-based food and practical strategies for eating the rainbow. J. Nutr. Metabolism2019 (2019).

  73. Vucea, V. et al. Prevalence and characteristics associated with modified texture food use in long term care: An analysis of making the most of mealtimes (m3) project. Can. J. Diet. Pract. Res. 80, 104–110 (2019).

    PubMed  Google Scholar 

  74. Vucea, V., Keller, H. H. & Ducak, K. Interventions for improving mealtime experiences in long-term care. J. Nutr. Gerontol. Geriatrics 33, 249–324 (2014).

    Google Scholar 

  75. Achanta, R. et al. SLIC superpixels compared to state-of-the-art superpixel methods. IEEE Trans. Pattern Anal. Mach. Intell. 34, 2274–2282 (2012).

    PubMed  Google Scholar 

Download references

Acknowledgements

The authors would like to acknowledge Yitian Wang for her contribution to manual food segmentation. This work was funded by the National Science and Engineering Research Council (NSERC), and the Canada Research Chairs (CRC) program.

Author information

Authors and Affiliations

Authors

Contributions

K.J.P and R.A contributed equally to this work. K.J.P conceptualized the system; R.A and A.W provided additional contributions to system design. K.J.P was the main contributor to experimental design, and contributed to algorithmic design. K.J.P was the main contributor for data acquisition protocols and data collection with contributions from R.A and A.G.C. R.A was the main contributor to algorithmic design with contributions from B.S, A.M, and A.W. K.J.P was the main contributor to data analyses; K.J.P and R.A. conducted data analyses. H.H.K provided the clinical nutrition perspective and direction, and facilitated and oversaw data collection in the test kitchen. K.J.P was the main contributor to writing the manuscript. All authors reviewed the manuscript.

Corresponding author

Correspondence to Kaylen J. Pfisterer.

Ethics declarations

Competing interests

The authors declare no competing interests.

Additional information

Publisher's note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Supplementary Information

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.

Reprints and Permissions

About this article

Verify currency and authenticity via CrossMark

Cite this article

Pfisterer, K.J., Amelard, R., Chung, A.G. et al. Automated food intake tracking requires depth-refined semantic segmentation to rectify visual-volume discordance in long-term care homes. Sci Rep 12, 83 (2022). https://doi.org/10.1038/s41598-021-03972-8

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1038/s41598-021-03972-8

Comments

By submitting a comment you agree to abide by our Terms and Community Guidelines. If you find something abusive or that does not comply with our terms or guidelines please flag it as inappropriate.

Search

Quick links

Nature Briefing

Sign up for the Nature Briefing newsletter — what matters in science, free to your inbox daily.

Get the most important science stories of the day, free in your inbox. Sign up for Nature Briefing