Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add new metric functions: LLMGEvalScore and ChatLLMGEvalScore #125

Draft
wants to merge 9 commits into
base: main
Choose a base branch
from

Conversation

m-ast
Copy link

@m-ast m-ast commented Jan 31, 2025

Summary

  • Add two metric functions: LLMGEvalScore and ChatLLMGEvalScore
  • These two metics calculate a weighted average score for evaluation of lm_output.

Implementation

  • Create flexeval/core/metric/llm_geval_score.py
    • Implement LLMGEvalScore and ChatLLMGEvalScore which inherit Metric
    • Note that these metrics can be used only for HuggingFaceLM and VLLM, implemented batch_compute_log_probs. I plan to implement it for OpenAIChatAPI and OpenAIChatBatchAPI afterwards.
  • Create tests for the script.

@m-ast m-ast self-assigned this Jan 31, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

1 participant