15d ago

Researchers release SWE-chat dataset of AI coding interactions

0

Researchers released SWE-chat, first large-scale dataset of real-user interactions with AI coding agents in software engineering sessions. Collected via open-source CLI tool across 200+ GitHub repositories, it contains 2.7 million interaction records. Analysis reveals coding agents wrote nearly all code in 40% of sessions; users intervened or pushed back 39% of time. Resources available via arXiv paper 2604.20779, swe-chat.com, and Hugging Face. Stanford Assistant Professor Diyi Yang among researchers.

Original post

We present SWE-chat: the first large-scale dataset of coding agent interactions from real users in the wild. In 40% of real coding sessions, the agent writes ~all the code. Users push back 39% of the time – agents almost never stop to check. Data, paper, & findings in the 🧵👇

Overview of SWE-chat. Left: a data collection pipeline diagram. Open-source developers install the Entire.io CLI tool, which logs their coding agent sessions and pushes the logs to a dedicated branch on their public GitHub repository. We discover and aggregate these logs into the SWE-chat dataset, with line-level attribution of which lines of code were written by the human versus the agent. Right: a growth chart showing cumulative logged events over time, rising steeply through early 2026. As of April 2026, the dataset contains 2.7 million logged events from over 200 repositories, including 63,000+ user prompts and 355,000+ agent tool calls across nearly 6,000 sessions.
9:44 AM · Apr 27, 2026 View on X
Reposted by

WildChat, OpenAssistant type datasets were very useful for understanding the effect of chat bots at scale.

Now that SWE-agent's are the norm, SWE-chat aims to do the same for AI coding.

Lots of fun findings, great effort led by @joabaum!

Joachim Baumann @ ICLR'26@joabaum

We present SWE-chat: the first large-scale dataset of coding agent interactions from real users in the wild. In 40% of real coding sessions, the agent writes ~all the code. Users push back 39% of the time – agents almost never stop to check. Data, paper, & findings in the 🧵👇

4:44 PM · Apr 27, 2026 · 39.6K Views
4:57 PM · Apr 27, 2026 · 8.2K Views

Really excited to have this dataset released to the community! There's a gap in our understanding of how users interact with coding agents at scale. SWE-chat fills that need to help shape the next generation of human-centered evals and training objectives for coding agents! 🤖🚀

Joachim Baumann @ ICLR'26@joabaum

We present SWE-chat: the first large-scale dataset of coding agent interactions from real users in the wild. In 40% of real coding sessions, the agent writes ~all the code. Users push back 39% of the time – agents almost never stop to check. Data, paper, & findings in the 🧵👇

4:44 PM · Apr 27, 2026 · 39.6K Views
1:24 AM · Apr 28, 2026 · 2.3K Views
Researchers release SWE-chat dataset of AI coding interactions · KRO · Digg