[COLM 2024] OpenAgents: An Open Platform for Language Agents in the Wild
[NeurIPS 2024] OSWorld: Benchmarking Multimodal Agents for Open-Ended Tasks in Real Computer Environments
Paper collection on building and evaluating language model agents via executable language grounding
[ICML 2023] Data and code release for the paper "DS-1000: A Natural and Reliable Benchmark for Data Science Code Generation".
[ICLR 2024 Spotlight] Code for the paper "Text2Reward: Reward Shaping with Language Models for Reinforcement Learning"
BRIGHT: A Realistic and Challenging Benchmark for Reasoning-Intensive Retrieval
[NeurIPS 2024] Spider2-V: How Far Are Multimodal Agents From Automating Data Science and Engineering Workflows?
Scaling Computer-Use Grounding via UI Decomposition and Synthesis