Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Running RAP with LLMs Missing get_next_token_logits and get_loglikelihood #153

Open
ERoszczyk opened this issue Feb 25, 2025 · 0 comments
Open

Comments

@ERoszczyk
Copy link

I've noticed that many LLMs available through APIs do not support functions like get_next_token_logits or get_loglikelihood. Since RAP relies on these functions, I'm wondering if it's possible to run RAP with an LLM that doesn't provide them.

I'm particularly interested in the prontoqa implementation using RAP. However, I found that both of these functions are used in the fast_reward function within ProntoQAConfig (see this link). Although there’s a comment referring to this as "OLD fast reward code," it appears that this implementation is still in use across several parts of the codebase.

Could someone clarify whether there’s an alternative approach for LLMs that lack these functions, or if this part of the implementation is planned for replacement? Any insights or suggestions would be appreciated.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant