Are all the reward models trained with 32 GPUs in total, including models with single machine config, e.g., editscore_7B, editscore_qwen3_vl_4B_instruct, and editscore_qwen3_vl_8B_instruct?
Is all multi-machine reward training done with 2 nodes with 16 GPUs each?
Are all the reward models trained with 32 GPUs in total, including models with single machine config, e.g.,
editscore_7B,editscore_qwen3_vl_4B_instruct, andeditscore_qwen3_vl_8B_instruct?Is all multi-machine reward training done with 2 nodes with 16 GPUs each?