top of page

DECISION INTELLIGENCE | CONVERSATIONAL AGENTS | MIMICKING HUMAN SPEECH AND BEHAVIOR

Untitled design (2)_edited.png

DECISION INTELLIGENCE: COMPUTATIONAL LINGUISTIC MODELING AND EVALUATION

[Funded by the National Research Foundation of Korea]

ROLE

Research Assistant

DESCRIPTION

As a Research Assistant of the Human and Artificial Intelligence Research Lab at Yonsei University, I am currently exploring how to assist human collaboration with AI agents for efficient decision-making in corporate settings. With my background in Human-Computer Interaction, I am currently focusing on how to mimic human dialogues for decision-making.

TEAMMATES

PI: Keeheon Lee
Kunhee Ryu, Minje Kim, Arina Svetasheva, Yeajoo Yoo, Hogyun Yoo, Sungjoon Whang

TIMELINE

Sept 2025 - Present

RESEARCH METHODS

Currently replication research
and meta-research

TOTAL RESEARCH FUNDING

Total Funding: 247,180,000 KRW
(approx. $172,986 USD)

Funded by National Research Foundation of Korea

PROJECT INTRODUCTION

This research series aims to explore how decision-making processes could be made more efficient in corporate settings, using a multi-agent system.

​​

In order to advance human-AI collaboration, my research team and I are currently exploring collaborative decision-oriented dialogue, as well as decision-making using Decentralized Partially Observable Markov Decision Process (Dec-POMDP).

INSPIRATIONS

Decision-Oriented Dialogue

for Human-AI Collaboration

Lin, J., Tomlin, N., Andreas, J., & Eisner, J. (2024). Decision-oriented dialogue for human–AI collaboration. Transactions of the Association for Computational Linguistics, 12, 892–911. https://doi.org/10.1162/tacl_a_00679

Introduces decision-oriented dialogues (DOD):
AI assistants collaborate with humans via natural language
to solve complex decision-making problems.

Screenshot 2025-11-15 at 12.04.27 AM.png

Fig 1. An overview of the three featured collaborative dialogue tasks. Tasks include assignment, planning, and mediation (Jessy Lin et al, 2024).

System Features

Involve multiple agents communicating to reach a joint decision.

 

Three daily-life domains studied:

  • Assignment: Conference paper-reviewer matching

  • Planning: City itinerary creation

  • Mediation: Coordinating group travel schedules

Screenshot 2025-11-15 at 12.04.27 AM.png

Assignment Task Observations & Roles

  • Each agent has symmetric roles but different and partial edge information.

  • Scores from a random uniform distribution, scaled positively but unknown between agents to avoid numeric communication.

  • Agents share information via dialogue to infer better matches.

  • Decision proposal, acceptance, and dialogue ends when agreement is reached.

Fig 2. The assignment task. (Jessy Lin et al, 2024).

Planning Task Definition

  • The assistant helps the user build an itinerary across city sites.

  • World state is a fully connected graph where edge weights indicate travel times.

  • User preferences include budget limits and desired amenities, while the assistant knows site features but not the user's private preferences.

  • Goal is to select a travel path balancing these information.

Screenshot 2025-11-15 at 12.04.27 AM.png

Fig 3. The planning task. (Jessy Lin et al, 2024).

Screenshot 2025-11-15 at 12.04.27 AM.png

Mediation Task Definition

  • Only the assistant can propose itineraries, and the user must accept or reject based on their preferences.

  • The iterative nature of dialogue allows refining plans through back and forth conversation.

  • The environment supports programmatic queries over sites to aid assistant decisions.


  • Scoring system normalizes itinerary quality for comparison.

Fig 4. The mediation task. (Jessy Lin et al, 2024).

Key Findings

  • Substantial gaps between AI assistants based on GPT-3 and human assistants.

  • Humans consistently achieved higher reward scores with shorter, more efficient dialogues. GPT-3 assistants tended to engage in longer conversations but often failed to make progress toward optimal decisions.

  • Self-play, where AI agents simulate both roles, resulted in even lower performance, indicating the challenge for models to collaborate without a human in the loop. 

Language models can solve computer tasks 

Kim, G., Baldi, P., & McAleer, S. (2023). Language models can solve computer tasks (arXiv:2303.17491). arXiv. https://doi.org/10.48550/arXiv.2303.17491

Explores how large language models (LLMs) can act as agents to solve computer tasks using natural language instructions.

Screenshot 2025-09-26 at 11.30.34 PM.png

Fig 5. The agent uses RCI prompting to solve a terminal task (deleting a file ending with ".rb") by first generating a step-by-step plan from the natural language instruction (Kim et al, 2023).

Recursive Criticism and Improvement (RCI):

AI assistants collaborate with humans via natural language to solve

complex decision-making problems.

  • technique for making large language models smarter and more reliable when solving computer tasks.

  • RCI helps models learn from their own mistakes, much like how people revise their work.

  •  way for AI models to improve their own answers. Instead of just giving one response, the model reviews its work, finds mistakes, and tries again, just like a human editor.
This process is recursive, meaning it can repeat several times to refine the answer.
RCI helps models become more accurate and thoughtful, especially on complex tasks.

Key Findings

  • RCI enables LLM agents to solve complex computer tasks with minimal data.

  • Limitations include handling long states and limited action spaces; future work may address these.

  • Can be applied to language-based agents in practical settings.

COPYRIGHT @ HEEYOUNG (EMILY) GHANG 2025

bottom of page