Conversation
johnwalz97
left a comment
There was a problem hiding this comment.
Left a couple comments, but looks good to me so far.
|
Robert McMenamin seems not to be a GitHub user. You need a GitHub account to be able to sign the CLA. If you have already a GitHub account, please add the email address used for this commit to your account. You have signed the CLA already but the status is still pending? Let us recheck it. |
PR SummaryThis PR introduces significant changes to support customizable judge LLM and embeddings across ValidMind's AI and prompt validation tests. The enhancements include:
Overall, these changes allow flexibility in specifying the judge LLM during testing and provide detailed validation for compatibility, ensuring that downstream testing functions receive the correct configurations. Test Suggestions
|
Pull Request Description
What and why?
Currently the RAGAS and Prompt tests only support using OpenAI. This solution adds two ways to point to a new LLM/Embedding model to leverage in those tests.
This includes:
How to test
Notebook used to test can be found in the Solutions Architect repo: https://github.com/validmind/solutions_architects_repo/blob/main/notebooks/byo_llm_for_judge.ipynb
I give an example of 1 RAGAS and 1 prompt test to ensure it works and has sample syntax for three ways Judge LLMs can be set
What needs special review?
Nothing of note
Dependencies, breaking changes, and deployment notes
None. Notebooks using syntax to run tests will continue to work.
Release notes
Adds ability to bring your own LLM. All existing tests using an LLM can now support using a user-defined LLM/Embedding model (assuming it is compatible with Langchain Chat/Embedding framework). This is restricted to RAGAS and Prompt validation tests.
Checklist