• Live Crypto Prices
  • Crypto News
    • Worldwide
      • Bitcoin
      • Ethereum
      • Altcoin
      • Blockchain
      • Regulation
    • Australian Crypto News
  • Education
    • Cryptocurrency For Beginners
    • Where to Buy Cryptocurrency
    • Where to Store Cryptos
    • Cryptocurrency Tax in Australia 2021
No Result
View All Result
CryptoABC.net
No Result
View All Result

Anthropic Study Reveals AI Agents Run 45 Minutes Autonomously as Trust Builds

February 18, 2026
in Blockchain
Reading Time: 3min read
0 0
A A
0
Anthropic Launches Claude 3.5 Sonnet Android App with Advanced AI Features
0
SHARES
2
VIEWS
ShareShareShareShareShare


Felix Pinkston
Feb 18, 2026 20:03

New Anthropic research shows Claude Code autonomy nearly doubled in 3 months, with experienced users granting more independence while maintaining oversight.





AI agents are working independently for significantly longer periods as users develop trust in their capabilities, according to new research from Anthropic published February 18, 2026. The study, which analyzed millions of human-agent interactions, found that the longest-running Claude Code sessions nearly doubled from under 25 minutes to over 45 minutes between October 2025 and January 2026.

The findings arrive as Anthropic rides a wave of investor confidence, having just closed a $30 billion Series G round that valued the company at $380 billion. That valuation reflects growing enterprise appetite for AI agents—and this research offers the first large-scale empirical look at how humans actually work with them.

Trust Builds Gradually, Not Through Capability Jumps

Perhaps the most striking finding: the increase in autonomous operation time was smooth across model releases. If autonomy were purely about capability improvements, you’d expect sharp jumps when new models dropped. Instead, the steady climb suggests users are gradually extending trust as they gain experience.

The data backs this up. Among new Claude Code users, roughly 20% of sessions use full auto-approve mode. By the time users hit 750 sessions, that number exceeds 40%. But here’s the counterintuitive part—experienced users also interrupt Claude more frequently, not less. New users interrupt in about 5% of turns; veterans interrupt in roughly 9%.

What’s happening? Users aren’t abandoning oversight. They’re shifting strategy. Rather than approving every action upfront, experienced users let Claude run and step in when something needs correction. It’s the difference between micromanaging and monitoring.

Claude Knows When to Ask

The research revealed something unexpected about Claude’s own behavior. On complex tasks, the AI stops to ask clarifying questions more than twice as often as humans interrupt it. Claude-initiated pauses actually exceed human-initiated interruptions on the most difficult work.

Common reasons Claude stops itself include presenting users with choices between approaches (35% of pauses), gathering diagnostic information (21%), and clarifying vague requests (13%). Meanwhile, humans typically interrupt to provide missing technical context (32%) or because Claude was running slow or excessive (17%).

This suggests Anthropic’s training for uncertainty recognition is working. Claude appears calibrated to its own limitations—though the researchers caution it may not always stop at the right moments.

Software Dominates, But Riskier Domains Emerge

Software engineering accounts for nearly 50% of all agentic tool calls on Anthropic’s public API. That concentration makes sense—code is testable, reviewable, and relatively low-stakes if something breaks.

But the researchers found emerging usage in healthcare, finance, and cybersecurity. Most actions remain low-risk and reversible—only 0.8% of observed actions appeared irreversible, like sending customer emails. Still, the highest-risk clusters involved sensitive security operations, financial transactions, and medical records.

The team acknowledges limitations: many high-risk actions may actually be red-team evaluations rather than production deployments. They can’t always tell the difference from their vantage point.

What This Means for the Industry

Anthropic’s researchers argue against mandating specific oversight patterns like requiring human approval for every action. Their data suggests such requirements would create friction without safety benefits—experienced users naturally develop more efficient monitoring strategies.

Instead, they’re calling for better post-deployment monitoring infrastructure across the industry. Pre-deployment testing can’t capture how humans actually interact with agents in practice. The patterns they observed—trust building over time, shifting oversight strategies, agents limiting their own autonomy—only emerge in real-world usage.

For enterprises evaluating AI agent deployments, the research offers a concrete benchmark: even power users at the extreme end of the distribution are running Claude autonomously for under an hour at a stretch. The gap between what models can theoretically handle (METR estimates five hours for comparable tasks) and what users actually permit suggests significant headroom remains—and that trust, not capability, may be the binding constraint on adoption.

Image source: Shutterstock


Credit: Source link

ShareTweetSendPinShare
Previous Post

Cup And Handle Pattern Puts XRP Price At $60 After Hitting Resistance

Next Post

Harvey AI Launches Global Legal Benchmark for UK, Australia, Spain

Next Post
VeChain Foundation Releases Q1 2024 Treasury Report

Harvey AI Launches Global Legal Benchmark for UK, Australia, Spain

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

You might also like

Crypto Price Prediction Today 6 March – XRP, Bitcoin, Ethereum

Crypto Price Prediction Today 6 March – XRP, Bitcoin, Ethereum

March 6, 2026
Former LAPD Officer Convicted in $350K Crypto ‘Wrench Attack’ Home Invasion

Former LAPD Officer Convicted in $350K Crypto ‘Wrench Attack’ Home Invasion

March 4, 2026
Solana Price Prediction: All Eyes on $95 — Will This Level Launch SOL Toward New Highs?

Solana Price Prediction: All Eyes on $95 — Will This Level Launch SOL Toward New Highs?

March 4, 2026
Bitcoin Hovers Around $70K as Weak Demand and Defensive Positioning Signal Fragile Market, Says Glassnode

Bitcoin Hovers Around $70K as Weak Demand and Defensive Positioning Signal Fragile Market, Says Glassnode

March 6, 2026
South Korea Orders Crypto Asset Review After $4.8M Wallet Blunder

South Korea Orders Crypto Asset Review After $4.8M Wallet Blunder

March 3, 2026
Culper Shorts Ethereum, Says Buterin Selling Signals More Pain

Culper Shorts Ethereum, Says Buterin Selling Signals More Pain

March 6, 2026
CryptoABC.net

This is an Australian online news/education portal that aims to provide the latest crypto news, real-time updates, education and reviews within Australia and around the world. Feel free to get in touch with us!

What's New Here!

Pundit Says XRP Price Could Reach $1,000 By The End Of 2026 If This Happens

Pundit Says XRP Price Could Reach $1,000 By The End Of 2026 If This Happens

March 8, 2026
Bitcoin Price Holds Above $115,000 — Here’s Why This Level Is Significant

Here’s Why Bitcoin Price Must Not Fall To $54K: Analyst

March 7, 2026

Subscribe Now

  • Contact Us
  • Privacy Policy
  • Terms of Use
  • DMCA

© 2021 cryptoabc.net - All rights reserved!

No Result
View All Result
  • Live Crypto Prices
  • Crypto News
    • Worldwide
      • Bitcoin
      • Ethereum
      • Altcoin
      • Blockchain
      • Regulation
    • Australian Crypto News
  • Education
    • Cryptocurrency For Beginners
    • Where to Buy Cryptocurrency
    • Where to Store Cryptos
    • Cryptocurrency Tax in Australia 2021

© 2021 cryptoabc.net - All rights reserved!

Welcome Back!

Login to your account below

Forgotten Password?

Create New Account!

Fill the forms below to register

All fields are required. Log In

Retrieve your password

Please enter your username or email address to reset your password.

Log In
Please enter CoinGecko Free Api Key to get this plugin works.