Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

evaluate results from elastic search as well #39

Open
SubhadityaMukherjee opened this issue Aug 20, 2024 · 4 comments
Open

evaluate results from elastic search as well #39

SubhadityaMukherjee opened this issue Aug 20, 2024 · 4 comments
Assignees
Labels
evaluation about evaluation model results
Milestone

Comments

@SubhadityaMukherjee
Copy link
Collaborator

No description provided.

@SubhadityaMukherjee SubhadityaMukherjee self-assigned this Aug 20, 2024
@SubhadityaMukherjee SubhadityaMukherjee added the evaluation about evaluation model results label Aug 20, 2024
@SubhadityaMukherjee SubhadityaMukherjee added this to the Search milestone Aug 20, 2024
@SubhadityaMukherjee
Copy link
Collaborator Author

I did evaluate elastic search with the queries that do not rely on filtering but it was not fair because that is not really the strength of the LLM.
Evaluating it with the structured query remains to be done.

@PGijsbers
Copy link
Member

but it was not fair

I don't want to be a hard-ass here, but the user doesn't care what's "fair". If the user isn't happy with the service, we can't say "Oh, but you see, we have someone in the back-end personally reading your query, going to the paper archives to read through the datasets, and then returning a number. It's a very different approach than conventional fast and reliable software, so you shouldn't have the same expectations."

That's a wild exaggeration of course, but it drives the point home. Right now the LLM prototype is a standalone prototype so it's fine. But for normal integration we need to compare to conventional methods so we can make an informed decision about what the user experience will be like.

@SubhadityaMukherjee
Copy link
Collaborator Author

I just meant that we havent done the evaluation yet ): and that the one that we have right now was not fair because we did not compare it to the elastic search yet. So we cannot be certain how much better it is.
Not that it is not fair to compare them.

Just that comparing the keyword search between the RAG and elastic search led to slightly better results for the RAG. But elastic search is not able to say find the appropriate filters and the latter was not evaluated yet.

@PGijsbers
Copy link
Member

okay clear :) just wanted to make sure, sorry if it came off the wrong way

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
evaluation about evaluation model results
Projects
None yet
Development

No branches or pull requests

2 participants