logo
banner

Journals & Publications

Journals Publications Papers

Papers

Being a Supercook: Joint Food Attributes and Multimodal Content Modeling for Recipe Retrieval and Exploration
Jul 13, 2017Author:
PrintText Size A A

 Title: Being a Supercook: Joint Food Attributes and Multimodal Content Modeling for Recipe Retrieval and Exploration

 Authors: Min, WQ; Jiang, SQ; Sang, JT; Wang, HY; Liu, XD; Herranz, L

 Author Full Names: Min, Weiqing; Jiang, Shuqiang; Sang, Jitao; Wang, Huayang; Liu, Xinda; Herranz, Luis

 Source: IEEE TRANSACTIONS ON MULTIMEDIA, 19 (5):1100-1113; 10.1109/TMM.2016.2639382 MAY 2017

 Language: English

 Abstract: This paper considers the problem of recipe-oriented image-ingredient correlation learning with multi-attributes for recipe retrieval and exploration. Existing methods mainly focus on food visual information for recognition while we model visual information, textual content (e.g., ingredients), and attributes (e.g., cuisine and course) together to solve extended recipe-oriented problems, such as multimodal cuisine classification and attributeenhanced food image retrieval. As a solution, we propose a multimodal multitask deep belief network (M3TDBN) to learn joint image-ingredient representation regularized by different attributes. By grouping ingredients into visible ingredients (which are visible in the food image, e.g., "chicken" and "mushroom") and nonvisible ingredients (e. g., "salt" and "oil"), M3TDBN is capable of learning both midlevel visual representation between images and visible ingredients and nonvisual representation. Furthermore, in order to utilize different attributes to improve the intermodality correlation, M3TDBN incorporates multitask learning to make different attributes collaborate each other. Based on the proposed M3TDBN, we exploit the derived deep features and the discovered correlations for three extended novel applications: 1) multimodal cuisine classification; 2) attribute-augmented cross-modal recipe image retrieval; and 3) ingredient and attribute inference fromfood images. The proposed approach is evaluated on the constructed Yummly dataset and the evaluation results have validated the effectiveness of the proposed approach.

 ISSN: 1520-9210

 eISSN: 1941-0077

 IDS Number: EY5XS

 Unique ID: WOS:000404056000017

*Click Here to View Full Record