Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

LLM qualitative evaluations and labeling #876

Open
Luca-Blight opened this issue Feb 19, 2025 · 2 comments
Open

LLM qualitative evaluations and labeling #876

Luca-Blight opened this issue Feb 19, 2025 · 2 comments

Comments

@Luca-Blight
Copy link

Description

It would be nice to have a place in the platform for this.
Another option would be to allow for integration with a partner that does provide it.

@samuelcolvin
Copy link
Member

Yup, we're going on this very thing, see pydantic/pydantic-ai#915 and linked pull request.

@Luca-Blight
Copy link
Author

That's awesome to see!

One thing that could be an interesting feature to have, particularly for online performance, is to enable the ability for another model to be set up as evaluator versus using a human.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

2 participants