IDETC-CIE
RECALL-MM
A Multimodal Dataset of Consumer Product Recalls for Risk Analysis using Computational Methods and Large Language Models
Abstract
Product recalls provide valuable insights into potential risks and hazards within the engineering design process, yet their full potential remains underutilized. In this study, we curate data from the United States Consumer Product Safety Commission (CPSC) recalls database to develop a multimodal dataset, RECALL-MM, that informs data-driven risk assessment using historical information, and augment it using generative methods. Patterns in the dataset highlight specific areas where improved safety measures could have significant impact. We extend our analysis by demonstrating interactive clustering maps that embed all recalls into a shared latent space based on recall descriptions and product names. Leveraging these data-driven tools, we explore three case studies to demonstrate the dataset’s utility in identifying product risks and guiding safer design decisions. The first two case studies illustrate how designers can visualize patterns across recalled products and situate new product ideas within the broader recall landscape to proactively anticipate hazards. In the third case study, we extend our approach by employing a large language model (LLM) to predict potential hazards based solely on product images. This demonstrates the model’s ability to leverage visual context to identify risk factors, revealing strong alignment with historical recall data across many hazard categories. However, the analysis also highlights areas where hazard prediction remains challenging, underscoring the importance of risk awareness throughout the design process. Collectively, this work aims to bridge the gap between historical recall data and future product safety, presenting a scalable, data-driven approach to safer engineering design.
Download publicationAssociated Researchers
Diana Bolanos
University of California, Berkeley
Kosa Goucher-Lambert
University of California, Berkeley
Related Publications
2024
Evaluating Large Language Models for Material SelectionThis work evaluates the use of LLMs for predicting materials of…
2024
Elicitron: An LLM Agent-Based Simulation Framework for Design Requirements ElicitationA novel framework that leverages Large Language Models (LLMs) to…
2024
DesignQA: A Multimodal Benchmark for Evaluating Large Language Models’ Understanding of Engineering DocumentationNovel benchmark aimed at evaluating the proficiency of multimodal…
2024
What’s in this LCA Report? A Case Study on Harnessing Large Language Models to Support Designers in Understanding Life Cycle ReportsExploring how large language models like ChatGPT can help designers…
Get in touch
Something pique your interest? Get in touch if you’d like to learn more about Autodesk Research, our projects, people, and potential collaboration opportunities.
Contact us