What problem does this solve?
Initial tests on my skills score well apart from low novelty scores, but the validator doesn't seem to provide many clues on what's dragging this value down.
I'd like to fix that, but I'm not sure what the problem is. I've tried asking Claude if the validator provides details on what it considers not-novel, but the answer seems inconclusive:
The validator doesn’t itemize what’s not novel — the feedback is always a holistic qualitative comment.
What would you like to see?
If the validator is aware of what an LLM already knows from training data, it seems like it should be possible to provide a few pointers to specific things to fix rather than just a score.
What problem does this solve?
Initial tests on my skills score well apart from low novelty scores, but the validator doesn't seem to provide many clues on what's dragging this value down.
I'd like to fix that, but I'm not sure what the problem is. I've tried asking Claude if the validator provides details on what it considers not-novel, but the answer seems inconclusive:
What would you like to see?
If the validator is aware of what an LLM already knows from training data, it seems like it should be possible to provide a few pointers to specific things to fix rather than just a score.