Image-Text-to-Text
Safetensors
English
qwen2_5_vl
agent
conversational

Add library_name and fix paper link

#1
by nielsr HF Staff - opened
Files changed (1) hide show
  1. README.md +10 -7
README.md CHANGED
@@ -1,23 +1,26 @@
1
  ---
2
- license: apache-2.0
 
3
  datasets:
4
  - hitsmy/AdaReasoner-TC-Randomized
5
  - hitsmy/AdaReasoner-TG-Data-Randomized
6
  language:
7
  - en
 
8
  metrics:
9
  - accuracy
10
- base_model:
11
- - Qwen/Qwen2.5-VL-7B-Instruct
12
  pipeline_tag: image-text-to-text
 
13
  tags:
14
  - agent
 
15
  ---
 
16
  <div align="center">
17
  <img src="docs/logo.png" alt="Logo" width="300">
18
  <h1 align="center">Dynamic Tool Orchestration for Iterative Visual Reasoning</h1>
19
 
20
- <a href="#">
21
  <img src="https://img.shields.io/badge/Paper-A42C25?style=for-the-badge&logo=arxiv&logoColor=white" alt="Paper">
22
  </a>
23
  <a href="https://github.com/ssmisya/AdaReasoner/tree/main/docs">
@@ -44,8 +47,7 @@ tags:
44
 
45
  ## πŸ“‹ Model Description
46
 
47
- **AdaReasoner-7B** is a vision-language model trained with dynamic tool orchestration capabilities for iterative visual reasoning. This model is AdaReasoner-7B-Randomized.
48
-
49
 
50
  We provide three variants of AdaReasoner-7B, each optimized for different use cases:
51
 
@@ -64,7 +66,8 @@ We provide three variants of AdaReasoner-7B, each optimized for different use ca
64
 
65
  ## πŸš€ Quick Start
66
 
67
- AdaReasoner-7B can be deployed for single-turn inference using standard inference frameworks such as vLLM.
 
68
  However, AdaReasoner is a tool-planning model whose full capabilities require interaction with an external tool environment.
69
  To fully evaluate or utilize its tool-planning behavior, we recommend using [AdaEval](https://github.com/ssmisya/AdaReasoner/tree/main/tool_server/tf_eval) provided in our repository for batch inference and evaluation, or trying the [Demo](https://github.com/ssmisya/AdaReasoner/tree/main/tool_server/tf_eval/demo) interface for interactive, single-instance GUI-based reasoning.
70
 
 
1
  ---
2
+ base_model:
3
+ - Qwen/Qwen2.5-VL-7B-Instruct
4
  datasets:
5
  - hitsmy/AdaReasoner-TC-Randomized
6
  - hitsmy/AdaReasoner-TG-Data-Randomized
7
  language:
8
  - en
9
+ license: apache-2.0
10
  metrics:
11
  - accuracy
 
 
12
  pipeline_tag: image-text-to-text
13
+ library_name: transformers
14
  tags:
15
  - agent
16
+ arxiv: 2601.18631
17
  ---
18
+
19
  <div align="center">
20
  <img src="docs/logo.png" alt="Logo" width="300">
21
  <h1 align="center">Dynamic Tool Orchestration for Iterative Visual Reasoning</h1>
22
 
23
+ <a href="https://arxiv.org/abs/2601.18631">
24
  <img src="https://img.shields.io/badge/Paper-A42C25?style=for-the-badge&logo=arxiv&logoColor=white" alt="Paper">
25
  </a>
26
  <a href="https://github.com/ssmisya/AdaReasoner/tree/main/docs">
 
47
 
48
  ## πŸ“‹ Model Description
49
 
50
+ **AdaReasoner-7B** is a vision-language model trained with dynamic tool orchestration capabilities for iterative visual reasoning. This model is AdaReasoner-7B-Randomized. It was introduced in the paper [AdaReasoner: Dynamic Tool Orchestration for Iterative Visual Reasoning](https://arxiv.org/abs/2601.18631).
 
51
 
52
  We provide three variants of AdaReasoner-7B, each optimized for different use cases:
53
 
 
66
 
67
  ## πŸš€ Quick Start
68
 
69
+ AdaReasoner-7B can be deployed for single-turn inference using standard inference frameworks such as vLLM or the `transformers` library.
70
+
71
  However, AdaReasoner is a tool-planning model whose full capabilities require interaction with an external tool environment.
72
  To fully evaluate or utilize its tool-planning behavior, we recommend using [AdaEval](https://github.com/ssmisya/AdaReasoner/tree/main/tool_server/tf_eval) provided in our repository for batch inference and evaluation, or trying the [Demo](https://github.com/ssmisya/AdaReasoner/tree/main/tool_server/tf_eval/demo) interface for interactive, single-instance GUI-based reasoning.
73