Excellent work!
However, it appears that the openclaw-test directory contains only the code for simulating dialogues, while the code for evaluating model performance is missing.
In other words, the evaluation tools corresponding to the scores presented in Table 3 of the paper are currently unavailable.
One indication of this is that I was unable to locate the specific prompts mentioned in Appendix C.3 of the paper—"Personal Agent: Evaluative Prompt from Simulator"—within this repository.
