AI Development

AGENTS.md Files Reduce Coding Agent Performance, arXiv Study Reveals

AI Tools Aggregator
#AGENTS.md #coding agents #SWE-bench #arXiv #research #performance optimization

SWE-bench experiments show that repository context files like AGENTS.md decrease task success rates and increase inference costs by over 20%, challenging widely recommended practices by agent developers.

A groundbreaking arXiv paper published on February 12, 2026, titled “Evaluating AGENTS.md: Are Repository-Level Context Files Helpful for Coding Agents?” reveals that AGENTS.md and similar repository context files—widely recommended by coding agent developers—may actually harm agent performance.

Research Background

Many coding agent developers (Claude Code, Cursor, GitHub Copilot, etc.) strongly encourage placing AGENTS.md files in repositories. These files are intended to provide context information to help agents understand repositories and execute appropriate coding tasks.

However, no rigorous investigation has previously examined whether this practice is actually effective for real-world tasks.

Key Findings

Researchers evaluated AGENTS.md effectiveness across multiple coding agents and LLMs in two complementary settings:

  1. SWE-bench Tasks: Established tasks from popular repositories with LLM-generated context files following agent-developer recommendations
  2. Real Repositories: A novel collection of issues from repositories containing developer-committed context files

1. Reduced Task Success Rates

Context files like AGENTS.md tend to reduce task success rates compared to providing no repository context at all. This finding directly contradicts the conventional assumption that context files are beneficial.

2. Significant Inference Cost Increase

Using context files increases inference costs by over 20%, driven by increased token usage.

3. Behavioral Analysis

The presence of context files changes agent behavior:

  • LLM-generated context files: Encourage more thorough testing and file traversal
  • Developer-provided context files: Similarly promote broader exploration
  • Instruction compliance: Agents strongly tend to respect context file instructions

4. Root Cause Identification

The research team concluded:

Unnecessary requirements from context files make tasks harder.

Practical Recommendations

Based on the findings, the paper offers clear guidance:

Human-written context files should describe only minimal requirements.

Specifically:

  • Avoid:

    • Overly detailed guidelines
    • Comprehensive instructions attempting to cover all possible scenarios
    • General development principles not directly related to the task
  • Include:

    • Minimal information directly necessary for the task
    • Repository-specific critical constraints
    • Important context that agents cannot discover independently

Industry Impact

These findings have significant implications for development teams rapidly adopting coding agents.

Impact on Agent Developers

Teams behind Claude Code, Cursor, GitHub Copilot, and similar tools may need to revise their AGENTS.md recommendations. The conventional wisdom that “more detail is better” should be reconsidered.

Impact on Developers

Developers who have already implemented AGENTS.md should consider:

  1. Reviewing existing AGENTS.md files and removing unnecessary requirements
  2. Focusing only on truly essential minimal information
  3. Measuring performance metrics before and after changes

Cost Reduction Potential

The potential to reduce inference costs by over 20% represents significant economic impact for organizations using agents at scale. With API pricing rising, this optimization cannot be ignored.

Research Limitations and Future Directions

The paper identifies several areas for future research:

  • Validation on larger-scale repositories
  • Effect measurement across different task types
  • Determining optimal context file structure
  • Comparing optimal context provision methods per agent

Conclusion

Well-intentioned AGENTS.md files may actually reduce agent performance and increase costs. This research demonstrates the importance of empirically validating practices considered “best practices.”

Coding agent users can potentially achieve both performance improvements and cost reductions by revisiting their AGENTS.md files in light of these findings.

Paper Link: arXiv:2602.11988
Authors: Thibaud Gloaguen et al.
Publication Date: February 12, 2026
Evaluation Environment: SWE-bench, multiple coding agents and LLMs

関連記事

人気記事

Comparison

ChatGPT(OpenAI)とClaude(Anthropic)の機能比較 2026年版。コーディング・長文解析・コスト・API料金の違いを検証

ChatGPT(GPT-4o/o3)とClaude(Sonnet 4.6/Opus 4.5)を2026年時点の最新情報で比較する。コーディング能力、長文処理、日本語品質、API料金、無料プランの違いをSWE-benchなどのベンチマーク結果とともに解説する。

続きを読む →
opinion

【2026年2月20日 所感】「AIがコードを書く」は仮説から現実になった——しかし私たちはその意味をまだ消化できていない

2026年2月20日に観測したコーディングエージェント関連ニュースの総括と所感。Anthropicの自律性研究、cmux、MJ Rathbunのエージェント事故、HN「外骨格 vs チーム」論争、Stripe Minions週1000件PR、Taalas 17k tokens/sec——朝から夜までの流れを通じて見えてきた「AIがコードを書く時代」の実相を考察する。

続きを読む →
tool

868のスキルをnpx 1コマンドで——「Antigravity Awesome Skills」が主要AIコーディングエージェントの共通スキル基盤になりつつある

Claude Code・Gemini CLI・Codex CLI・Cursor・GitHub Copilotなど主要AIコーディングアシスタントを横断する868以上のスキルライブラリ「Antigravity Awesome Skills」(v5.4.0)を詳細分析。Anthropic・Vercel・OpenAI・Supabase・Microsoftの公式スキルを統合した設計思想、ロール別バンドル・ワークフロー機能、SKILL.mdによる相互運用性のアーキテクチャを解説する。

続きを読む →

最新記事

0 tools selected