Large Language Models for Digital Twin Simulation

Best AI papers explained - A podcast by Enoch H. Kang

Categories:

Twin-2K-500: A dataset for building digital twins of over 2,000 people based on their answers to over 500 questionsThis academic paper introduces a novel and extensive dataset of over 2,000 human participants who answered more than 500 questions covering demographics, psychological traits, cognitive abilities, and economic behaviors across four waves of data collection. The dataset is designed to validate and improve the use of Large Language Models (LLMs) in creating "digital twins" that simulate human responses. Initial testing of these digital twins on a subset of behavioral economics experiments shows they can replicate human behavior with an average accuracy of 71.72%, representing a significant improvement over random chance. While the digital twins effectively mirror human responses in many scenarios, the authors highlight areas where the simulations diverge, particularly when humans exhibit suboptimal or non-normative behaviors or in sensitive domains like medicine and politics, suggesting avenues for future research and refinement of LLM-based human simulation.