top | item 45990112

(no title)

irshadnilam | 3 months ago

While I am not familiar with OPs project,

I can somewhat answer this to best of my knowledge.

Right now, businesses communicate with REST Apis.

That is why we have API gateways like AWS Gateway, Apigee, WSO2 (company i used to work in), Kong, etc so businesses can securly deploy and expose APIS.

As LLMS gets better, the idea is we will evenutally move to a world where ai agents do most of business tasks. And businesses will want to expose ai agents instead of APIS.

This is where protocols like a2a comes in. Google partnering with some other giants introduced a2a protocol a while ago, it is now under linux foundation.

It is a standard for one agent to talk to another agent regardless of the framework (langchain, crewai etc) that is used to build the agent.

discuss

order

mac-monet|3 months ago

I see. If iiuc, it's like an extension to an API endpoint. Instead of exposing only endpoints, you can let a user describe an intent and have the agent do the work. Is this not also the goal of an MCP as well?

spwa4|3 months ago

No. What the AI giants want to accomplish is to have agents talk to agents ... so that the business-agent has the choice of what to present, what to do, instead of being forced as an API to list all the options.

Because if they have this, then there will be enormous value in selling smarter agents to businesses (think a server doubling the coffee price when detecting tourists), dollars they're hoping to capture.

magackame|3 months ago

Can't you just put the agent behind a REST API and give the other agents a curl tool + doc?

irshadnilam|3 months ago

You can.

Everyone will have their own versions of the rest endpoints, their own version of input params, and lots and lots of docs scatterd.

A standard, will help the ecosystem grow. Tooling, libraries etc.