-
Notifications
You must be signed in to change notification settings - Fork 0
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Tool learning for LLM #4
Comments
Could you provide a few extra details?
|
Current LLMs are being directed towards building agent systems. An agent is a system based on LLMs by providing tool descriptions in prompts for the LLM My approach filters tools in two steps: Step 1: Use a bi-encoder (quite effective). "query": The query text. |
Thanks for the additional context. If I understand correctly, you want to rank tool descriptions based on a query specifying the need for a tool. I'm surprised that a bi-encoder is more effective than a cross-encoder on this task and would assume that given enough high-quality training data, a cross-encoder will be substantially more effective. That being said, the Set-Encoder most likely will not give you a substantial boost over a standard cross-encoder's effectiveness. The Set-Encoder excels when interactions between the items to be ranked are necessary. In this case, the tools can most likely be ranked independently from one another. |
I think set encoder can still be used, when a query needs the cooperation of many tools to complete. |
Yes, that is a good point. In those cases, the Set-Encoder is likely to be more effective than a standard cross-encoder. |
I am currently working on a problem to rerank tools (retrieving the appropriate tool for LLM), but the cross-encoder models are not converging.
Here is an example:
query: give me btc price
tool: get token price
Is your model feasible for this task?
The text was updated successfully, but these errors were encountered: