How are people here giving AI agents reliable access to public web data?
Curious how people here are solving public web data access for AI agent workflows.
A lot of agent demos work fine in controlled examples, but things seem to get much harder once the workflow depends on:
- geo-targeted results
- repeated sessions
- browser-based interaction
- public sites with stronger anti-bot controls
I work on web data and proxy infrastructure, so I’m looking at this from that angle.
What I’m most interested in is where things usually break first in real setups:
- access/routing
- session handling
- retries
- browser execution
- data quality
Are people mostly solving this with simple proxy rotation, browser automation, or something more layered?
Tagged in#Agents
Comments
No comments yet
Be the first to share your thoughts