Update README.md
This commit is contained in:
@@ -105,7 +105,7 @@ In this work, we use the Best-of-N evaluation strategy and employ [VisualPRM-8B]
|
||||
|
||||
### Multimodal Reasoning and Mathematics
|
||||
|
||||

|
||||

|
||||
|
||||
### OCR, Chart, and Document Understanding
|
||||
|
||||
@@ -161,7 +161,7 @@ The evaluation results in the Figure below shows that the model with native mult
|
||||
|
||||
As shown in the table below, models fine-tuned with MPO demonstrate superior reasoning performance across seven multimodal reasoning benchmarks compared to their counterparts without MPO. Specifically, InternVL3-78B and InternVL3-38B outperform their counterparts by 4.1 and 4.5 points, respectively. Notably, the training data used for MPO is a subset of that used for SFT, indicating that the performance improvements primarily stem from the training algorithm rather than the training data.
|
||||
|
||||

|
||||

|
||||
|
||||
### Variable Visual Position Encoding
|
||||
|
||||
|
||||
Reference in New Issue
Block a user