notebooks: Quickstart for model validation#376
Conversation
Co-authored-by: Lois Ansah <133300328+LoiAnsah@users.noreply.github.com>
Co-authored-by: Lois Ansah <133300328+LoiAnsah@users.noreply.github.com>
Co-authored-by: Lois Ansah <133300328+LoiAnsah@users.noreply.github.com>
Co-authored-by: Lois Ansah <133300328+LoiAnsah@users.noreply.github.com>
Co-authored-by: Lois Ansah <133300328+LoiAnsah@users.noreply.github.com>
Co-authored-by: Lois Ansah <133300328+LoiAnsah@users.noreply.github.com>
PR SummaryThis pull request introduces several enhancements and additions to the model validation documentation and quickstart guides within the project. Key changes include:
These changes aim to improve the user experience and provide clearer guidance for users working with the ValidMind Library and Platform. Test Suggestions
|
Co-authored-by: Lois Ansah <133300328+LoiAnsah@users.noreply.github.com>
|
@LoiAnsah For the record, this is what happens when the JSON is incorrect and the notebook is corrupted... ;) Can you think about how if this were your notebook, how you would go about fixing it? |
|
|
@LoiAnsah, why did you close this PR? In general, you shouldn't be closing or merging other people's PRs on their behalf, especially without communication. |
|
@validbeck I’d check my logs and switch back to the one just before the current one. |
|
@validbeck Apoligies, I closed it by mistake. I was trying to reply to your comment. I select quote reply. |
In this case, this is not the right approach — you want to apply the suggested changes. The "roll-back" method is only if you don't want to retain the changes and want to revert to known working version and start fresh. You will encounter many situations like this, where you will need to evaluate on a case-by-case basis how to best approach fixing things. Rolling back is not the only answer, especially if you want to retain later work. What I did was:
We can go over this together in a session because I want you to interact with Jupyter Notebooks under the hood and what it looks like. In preparation, please:
|
|
@juanmleng @LoiAnsah I've either committed the suggestions and edited the surrounding context to match the changes, or left explanations via a comment as to why the suggestions weren't applied. Can either of you please take another look, and approve if it looks good enough? 🙏🏻 |
PR SummaryThis pull request introduces several enhancements and bug fixes to the model validation and documentation notebooks within the project. Key changes include:
These changes collectively aim to improve the user experience and effectiveness of the model validation and documentation process using the ValidMind Library. Test Suggestions
|







Pull Request Description
What
There is a net-new notebook under
notebooks/quickstart/: Quickstart for model validationThis notebook is a companion notebook to our existing "Quickstart for model documentation," and goes over the basics of getting started with model validation with the ValidMind Library with the idea that the model you're validating was created using the model documentation quickstart.
Why
Our validator resources are currently very sparse. This is also a step towards retooling our undeveloped "Get Started" section in the documentation.
How to Test
gh pr checkout 376/notebooks/quickstart/quickstart_model_validation.ipynbPull Request Dependencies
Tip
Refer to the deployment notes below.
External Release Notes
Want to get started with validating models with the ValidMind Library? Check out our brand new Quickstart for model validation notebook:
Deployment Notes
Changes to the notebooks will be cherry-picked into the
documentationrepo with this branch when changes are approvedvalidmind-libraryside via this PR: validmind/documentation#731Breaking Changes
n/a
Screenshots/Videos (Frontend Only)
n/a
Checklist
Areas Needing Special Review
Important
I took some creative license with the following sections, so someone should check if the examples are relevant, accurate, and properly described:
Run data comparison tests
Check if the explanatory comments on why we compare the two different sets of paired datasets is accurate, and if these two comparisons are in fact relevant and demonstrative.
Run model performance tests
Check if the lead-in text for why we use the testing dataset for our performance tests is relevant and accurate.
Run diagnostic tests
Check if the lead-in text for why we use the training and testing datasets for our diagnostic tests is relevant and accurate.
Run feature importance tests
Check if the lead-in text for why we use the testing dataset for our feature importance tests is relevant and accurate.
Additional Notes
I also adjusted the following sections in these notebooks as I noticed they were incomplete/out of date:
Validate an application scorecard model
Validate an application scorecard model: Setting up > Assign validator credentials
This was missing the update where you also have to remove yourself as a model owner. Remedied:
Finalize testing and documentation (ValidMind for model development)
Finalize testing and documentation
The next steps section needed some TLC in comparison to the newer validation series, so I spruced it up:
Developing challenger models (ValidMind for model validation)
Developing a potential challenger model: Running model evaluation tests
Since I added more context to why we use certain datasets in the quickstart, I added the same explanations in this introductory notebook as well under the Running model evaluation tests sub-sections:
Note
Refer also to the "Areas Needing Special Review" section above.
FInalize testing and reporting (ValidMind for model validation)
FInalize testing and reporting
Tidied up the Next steps section as well here: