Sasha: Creative Goal-Oriented Reasoning in Smart Homes with Large Language Models

被引:7
|
作者
King, Evan [1 ]
Yu, Haoxiang [1 ]
Lee, Sangsu [1 ]
Julien, Christine [1 ]
机构
[1] Univ Texas Austin, Austin, TX 78712 USA
来源
PROCEEDINGS OF THE ACM ON INTERACTIVE MOBILE WEARABLE AND UBIQUITOUS TECHNOLOGIES-IMWUT | 2024年 / 8卷 / 01期
基金
美国国家科学基金会;
关键词
smart environments; pervasive computing; ambient intelligence; large language models; USERS;
D O I
10.1145/3643505
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Smart home assistants function best when user commands are direct and well-specified-e.g., "turn on the kitchen light"-or when a hard-coded routine specifies the response. In more natural communication, however, human speech is unconstrained, often describing goals (e.g., "make it cozy in here" or "help me save energy") rather than indicating specific target devices and actions to take on those devices. Current systems fail to understand these under-specified commands since they cannot reason about devices and settings as they relate to human situations. We introduce large language models (LLMs) to this problem space, exploring their use for controlling devices and creating automation routines in response to under-specified user commands in smart homes. We empirically study the baseline quality and failure modes of LLM-created action plans with a survey of age-diverse users. We find that LLMs can reason creatively to achieve challenging goals, but they experience patterns of failure that diminish their usefulness. We address these gaps with Sasha, a smarter smart home assistant. Sasha responds to loosely-constrained commands like "make it cozy" or "help me sleep better" by executing plans to achieve user goals-e.g., setting a mood with available devices, or devising automation routines. We implement and evaluate Sasha in a hands-on user study, showing the capabilities and limitations of LLM-driven smart homes when faced with unconstrained user-generated scenarios.
引用
收藏
页数:38
相关论文
共 43 条
  • [31] Selecting from Multiple Strategies Improves the Foreseeable Reasoning of Tool-Augmented Large Language Models
    Wu, Yongchao
    Henriksson, Aron
    MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES: RESEARCH TRACK, PT III, ECML PKDD 2024, 2024, 14943 : 197 - 212
  • [32] Applications and Challenges of Large Language Models in Smart Government - From technological Advances to Regulated Applications
    Dai, Ziqing
    PROCEEDINGS OF 2024 3RD INTERNATIONAL CONFERENCE ON FRONTIERS OF ARTIFICIAL INTELLIGENCE AND MACHINE LEARNING, FAIML 2024, 2024, : 275 - 280
  • [33] DogChat: A Pet-centered Smart Collar Prototype based on Large Language Models and Wechat
    Xue, Cheng
    Zuo, Zonglin
    Jiang, Xinran
    Fu, Xinyi
    COMPANION OF THE 2024 ACM INTERNATIONAL JOINT CONFERENCE ON PERVASIVE AND UBIQUITOUS COMPUTING, UBICOMP COMPANION 2024, 2024, : 162 - 166
  • [34] Performance evaluation of large language models with chain-of-thought reasoning ability in clinical laboratory case interpretation
    Yang, He S.
    Li, Jieli
    Yi, Xin
    Wang, Fei
    CLINICAL CHEMISTRY AND LABORATORY MEDICINE, 2025,
  • [35] Improving Zero-shot Visual Question Answering via Large Language Models with Reasoning Question Prompts
    Lan, Yunshi
    Li, Xiang
    Liu, Xin
    Li, Yang
    Qin, Wei
    Qian, Weining
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2023, 2023, : 4389 - 4400
  • [36] Smart Drug Delivery Systems using Large Language Models for Real-Time Treatment Personalization
    Balakrishna, Chinnala
    Yadav, Ankit
    Singh, Jagendra
    Saba, Masarath
    Shashikant
    Shrivastava, Vineet
    2024 2ND WORLD CONFERENCE ON COMMUNICATION & COMPUTING, WCONF 2024, 2024,
  • [37] Fluid Transformers and Creative Analogies: Exploring Large Language Models' Capacity for Augmenting Cross-Domain Analogical Creativity
    Ding, Zijian
    Srinivasan, Arvind
    MacNeil, Stephen
    Chan, Joel
    2023 PROCEEDINGS OF THE 15TH CONFERENCE ON CREATIVITY AND COGNITION, C&C 2023, 2023, : 489 - 505
  • [38] Enhancing Chinese comprehension and reasoning for large language models: an efficient LoRA fine-tuning and tree of thoughts framework
    Chen, Songlin
    Wang, Weicheng
    Chen, Xiaoliang
    Zhang, Maolin
    Lu, Peng
    Li, Xianyong
    Du, Yajun
    JOURNAL OF SUPERCOMPUTING, 2025, 81 (01)
  • [39] From Classification to Clinical Insights: Towards Analyzing and Reasoning About Mobile and Behavioral Health Data With Large Language Models
    Englhardt, Zachary
    Ma, Chengqian
    Morris, Margaret E.
    Chang, Chun-Cheng
    Xu, Xuhai Orson
    Qin, Lianhui
    McDduff, Daniel
    Liu, Xin
    Patel, Shwetak
    Iyer, Vikram
    PROCEEDINGS OF THE ACM ON INTERACTIVE MOBILE WEARABLE AND UBIQUITOUS TECHNOLOGIES-IMWUT, 2024, 8 (02):
  • [40] Enhancing In-Context Learning of Large Language Models for Knowledge Graph Reasoning via Rule-and-Reinforce Selected Triples
    Wang, Shaofei
    APPLIED SCIENCES-BASEL, 2025, 15 (03):