Spaces:
Running
Running
JianGuanTHU
commited on
Commit
·
d30eb2d
1
Parent(s):
973737d
initialize
Browse files
.DS_Store
CHANGED
Binary files a/.DS_Store and b/.DS_Store differ
|
|
README.md
CHANGED
@@ -1,4 +1,18 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
# ViLaBench
|
|
|
2 |
This is a web project showcasing a collection of benchmarks for vision-language models.
|
3 |
|
4 |
These benchmark and result data are carefully compiled and merged from technical reports and official blogs of renowned multimodal models, including Google's Gemini series ([Gemini 2.5 Report](https://storage.googleapis.com/deepmind-media/gemini/gemini_v2_5_report.pdf)), OpenAI GPT series and OpenAI o series ([OpenAI o3 and o4-mini](https://openai.com/index/introducing-o3-and-o4-mini/)), [Seed1.5-VL](https://arxiv.org/pdf/2505.07062), [MiMo-VL](https://arxiv.org/pdf/2506.03569), [Kimi-VL](https://huggingface.co/moonshotai/Kimi-VL-A3B-Thinking-2506), [Qwen2.5-VL](https://arxiv.org/pdf/2502.13923), [InternVL3](https://arxiv.org/abs/2504.10479), and other leading models' official technical documentation.
|
@@ -22,15 +36,19 @@ This collection provides researchers and developers with a comprehensive, standa
|
|
22 |
## Local Usage
|
23 |
|
24 |
1. Ensure `vilabench.csv` and `index.html` are in the same directory
|
|
|
25 |
2. Use a local server to open the webpage (to avoid CORS issues):
|
|
|
26 |
```bash
|
27 |
python3 -m http.server 8000
|
28 |
```
|
|
|
29 |
3. Visit `http://localhost:8000` in your browser
|
30 |
|
31 |
## Data Format
|
32 |
|
33 |
The CSV file contains the following columns:
|
|
|
34 |
- Benchmark: Benchmark name
|
35 |
- URL: Paper link
|
36 |
- year: Publication year
|
|
|
1 |
+
---
|
2 |
+
title: ViLaBench
|
3 |
+
emoji: 🧠
|
4 |
+
colorFrom: yellow
|
5 |
+
colorTo: indigo
|
6 |
+
sdk: static
|
7 |
+
pinned: false
|
8 |
+
license: apache-2.0
|
9 |
+
short_description: Benchmark collection for Vision-Language Models (VLMs)
|
10 |
+
---
|
11 |
+
|
12 |
+
|
13 |
+
|
14 |
# ViLaBench
|
15 |
+
|
16 |
This is a web project showcasing a collection of benchmarks for vision-language models.
|
17 |
|
18 |
These benchmark and result data are carefully compiled and merged from technical reports and official blogs of renowned multimodal models, including Google's Gemini series ([Gemini 2.5 Report](https://storage.googleapis.com/deepmind-media/gemini/gemini_v2_5_report.pdf)), OpenAI GPT series and OpenAI o series ([OpenAI o3 and o4-mini](https://openai.com/index/introducing-o3-and-o4-mini/)), [Seed1.5-VL](https://arxiv.org/pdf/2505.07062), [MiMo-VL](https://arxiv.org/pdf/2506.03569), [Kimi-VL](https://huggingface.co/moonshotai/Kimi-VL-A3B-Thinking-2506), [Qwen2.5-VL](https://arxiv.org/pdf/2502.13923), [InternVL3](https://arxiv.org/abs/2504.10479), and other leading models' official technical documentation.
|
|
|
36 |
## Local Usage
|
37 |
|
38 |
1. Ensure `vilabench.csv` and `index.html` are in the same directory
|
39 |
+
|
40 |
2. Use a local server to open the webpage (to avoid CORS issues):
|
41 |
+
|
42 |
```bash
|
43 |
python3 -m http.server 8000
|
44 |
```
|
45 |
+
|
46 |
3. Visit `http://localhost:8000` in your browser
|
47 |
|
48 |
## Data Format
|
49 |
|
50 |
The CSV file contains the following columns:
|
51 |
+
|
52 |
- Benchmark: Benchmark name
|
53 |
- URL: Paper link
|
54 |
- year: Publication year
|