Hi,
Thank you for your work.
I wanted to ask: are you only using the first question in each QAs.json entry as the instruction during training, as indicated here:
|
question = QAs["questions"][0] |
If so, wouldn’t that underutilize the dataset, since most samples contain 5–6 questions? Have you tried incorporating all questions during training?
Additionally, I noticed that the evaluation metric you used is IoU as mentioned in the paper, which only considers the segmentation masks. In that case, the answers in QAs.json are not directly involved in training. Why, then, are they loaded into the data_dict here:
|
# answer_idx = random.randint(0, answer_num - 1) |
|
answer = QAs["answer"][0] |
Or am I missing something and the answers are also used during training? If so, I noticed you tried both random sampling and using answer[0] as the line above, could you clarify which approach you ultimately used, and whether using different strategies affected performance?
Regards,
Vicky
Hi,
Thank you for your work.
I wanted to ask: are you only using the first question in each QAs.json entry as the instruction during training, as indicated here:
SegEarth-R1/segearth_r1/train/train_dataset.py
Line 321 in f0c382c
If so, wouldn’t that underutilize the dataset, since most samples contain 5–6 questions? Have you tried incorporating all questions during training?
Additionally, I noticed that the evaluation metric you used is IoU as mentioned in the paper, which only considers the segmentation masks. In that case, the answers in QAs.json are not directly involved in training. Why, then, are they loaded into the data_dict here:
SegEarth-R1/segearth_r1/train/train_dataset.py
Lines 327 to 328 in f0c382c
Or am I missing something and the answers are also used during training? If so, I noticed you tried both random sampling and using answer[0] as the line above, could you clarify which approach you ultimately used, and whether using different strategies affected performance?
Regards,
Vicky