arashkermani commited on
Commit
0bad3bb
·
verified ·
1 Parent(s): 924a12c

Upload folder using huggingface_hub

Browse files
README.md CHANGED
@@ -21,17 +21,19 @@ This is a **tiny random-initialized version** of the [openbmb/MiniCPM-o-2_6](htt
21
  **⚠️ Important**: This model has randomly initialized weights and is NOT intended for actual inference. It is designed solely for:
22
  - Testing model loading and export functionality
23
  - CI/CD pipeline validation
24
- - OpenVINO conversion testing
25
  - Quantization workflow testing
26
 
27
  ## Model Specifications
28
 
29
  - **Architecture**: MiniCPM-o-2_6 (multimodal: vision + text + audio + TTS)
30
- - **Parameters**: 1,477,376 (~1.48M parameters)
31
- - **Model Binary Size**: 5.64 MB
32
- - **Total Repository Size**: ~21 MB
33
  - **Original Model**: [openbmb/MiniCPM-o-2_6](https://huggingface.co/openbmb/MiniCPM-o-2_6) (~18 GB)
34
- - **Size Reduction**: 853× smaller than the full model
 
 
35
 
36
  ## Architecture Details
37
 
@@ -56,7 +58,7 @@ This is a **tiny random-initialized version** of the [openbmb/MiniCPM-o-2_6](htt
56
  - `hidden_size`: 8
57
  - `num_layers`: 1
58
 
59
- All architectural components are present but miniaturized to ensure API compatibility while drastically reducing compute requirements.
60
 
61
  ## Usage
62
 
@@ -111,7 +113,8 @@ model_id = "arashkermani/tiny-random-MiniCPM-o-2_6"
111
  # Load model for OpenVINO
112
  model = OVModelForVisualCausalLM.from_pretrained(
113
  model_id,
114
- trust_remote_code=True
 
115
  )
116
 
117
  processor = AutoProcessor.from_pretrained(
@@ -148,10 +151,11 @@ This model is intended **exclusively** for:
148
  ## Training Details
149
 
150
  This model was generated by:
151
- 1. Loading the config from `optimum-intel-internal-testing/tiny-random-MiniCPM-o-2_6`
152
  2. Reducing all dimensions to minimal viable values
153
  3. Initializing weights randomly using `AutoModelForCausalLM.from_config()`
154
- 4. Copying all necessary tokenizer, processor, and custom code files
 
155
 
156
  **No training was performed** - all weights are randomly initialized.
157
 
@@ -161,15 +165,27 @@ The model has been validated to ensure:
161
  - ✅ Loads with `trust_remote_code=True`
162
  - ✅ Compatible with transformers AutoModel APIs
163
  - ✅ Supports forward pass with expected input format
164
- - ✅ Compatible with OpenVINO export via optimum-intel
165
  - ✅ Includes all required custom modules and artifacts
 
 
 
 
 
 
 
 
 
 
 
 
166
 
167
- See the [validation report](https://github.com/arashkermani/tiny-minicpm-o) for detailed technical analysis.
168
 
169
  ## Files Included
170
 
171
  - `config.json` - Model configuration
172
- - `pytorch_model.bin` - Model weights (5.64 MB)
173
  - `generation_config.json` - Generation parameters
174
  - `preprocessor_config.json` - Preprocessor configuration
175
  - `processor_config.json` - Processor configuration
 
21
  **⚠️ Important**: This model has randomly initialized weights and is NOT intended for actual inference. It is designed solely for:
22
  - Testing model loading and export functionality
23
  - CI/CD pipeline validation
24
+ - OpenVINO conversion testing
25
  - Quantization workflow testing
26
 
27
  ## Model Specifications
28
 
29
  - **Architecture**: MiniCPM-o-2_6 (multimodal: vision + text + audio + TTS)
30
+ - **Parameters**: 17,390,468 (~17.4M parameters)
31
+ - **Model Binary Size**: 66.45 MB
32
+ - **Total Repository Size**: ~82 MB
33
  - **Original Model**: [openbmb/MiniCPM-o-2_6](https://huggingface.co/openbmb/MiniCPM-o-2_6) (~18 GB)
34
+ - **Size Reduction**: 219× smaller than the full model
35
+ - **OpenVINO Export**: ✅ Fully supported
36
+ - **All Components Enabled**: vision, audio, and TTS modules initialized
37
 
38
  ## Architecture Details
39
 
 
58
  - `hidden_size`: 8
59
  - `num_layers`: 1
60
 
61
+ All architectural components are present and properly initialized, ensuring full compatibility with OpenVINO export and testing workflows.
62
 
63
  ## Usage
64
 
 
113
  # Load model for OpenVINO
114
  model = OVModelForVisualCausalLM.from_pretrained(
115
  model_id,
116
+ trust_remote_code=True,
117
+ export=True
118
  )
119
 
120
  processor = AutoProcessor.from_pretrained(
 
151
  ## Training Details
152
 
153
  This model was generated by:
154
+ 1. Loading the config from `openbmb/MiniCPM-o-2_6`
155
  2. Reducing all dimensions to minimal viable values
156
  3. Initializing weights randomly using `AutoModelForCausalLM.from_config()`
157
+ 4. Keeping all components (vision, audio, TTS) enabled for full compatibility
158
+ 5. Copying all necessary tokenizer, processor, and custom code files
159
 
160
  **No training was performed** - all weights are randomly initialized.
161
 
 
165
  - ✅ Loads with `trust_remote_code=True`
166
  - ✅ Compatible with transformers AutoModel APIs
167
  - ✅ Supports forward pass with expected input format
168
+ - ✅ **Compatible with OpenVINO export via optimum-intel**
169
  - ✅ Includes all required custom modules and artifacts
170
+ - ✅ All multimodal components (vision/audio/TTS) properly initialized
171
+
172
+ ## Comparison with Previous Versions
173
+
174
+ | Metric | v1 (components disabled) | v2 (this version) |
175
+ |--------|-------------------------|-------------------|
176
+ | Parameters | 1.48M | 17.4M |
177
+ | Total Size | 21 MB | 82 MB |
178
+ | OpenVINO Export | ❌ Not supported | ✅ Fully supported |
179
+ | Vision Module | ❌ Disabled | ✅ Enabled |
180
+ | Audio Module | ❌ Disabled | ✅ Enabled |
181
+ | TTS Module | ❌ Disabled | ✅ Enabled |
182
 
183
+ **Recommendation**: Use this version for full test coverage including OpenVINO export tests.
184
 
185
  ## Files Included
186
 
187
  - `config.json` - Model configuration
188
+ - `pytorch_model.bin` - Model weights (66.45 MB)
189
  - `generation_config.json` - Generation parameters
190
  - `preprocessor_config.json` - Preprocessor configuration
191
  - `processor_config.json` - Processor configuration
README.md.backup ADDED
@@ -0,0 +1,215 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ library_name: transformers
4
+ tags:
5
+ - vision
6
+ - image-text-to-text
7
+ - multimodal
8
+ - test-model
9
+ - tiny-model
10
+ - openvino
11
+ - optimum-intel
12
+ pipeline_tag: image-text-to-text
13
+ ---
14
+
15
+ # Tiny Random MiniCPM-o-2_6
16
+
17
+ ## Model Description
18
+
19
+ This is a **tiny random-initialized version** of the [openbmb/MiniCPM-o-2_6](https://huggingface.co/openbmb/MiniCPM-o-2_6) multimodal vision-language model, designed specifically for **testing and CI/CD purposes** in the [optimum-intel](https://github.com/huggingface/optimum-intel) library.
20
+
21
+ **⚠️ Important**: This model has randomly initialized weights and is NOT intended for actual inference. It is designed solely for:
22
+ - Testing model loading and export functionality
23
+ - CI/CD pipeline validation
24
+ - OpenVINO conversion testing
25
+ - Quantization workflow testing
26
+
27
+ ## Model Specifications
28
+
29
+ - **Architecture**: MiniCPM-o-2_6 (multimodal: vision + text + audio + TTS)
30
+ - **Parameters**: 1,477,376 (~1.48M parameters)
31
+ - **Model Binary Size**: 5.64 MB
32
+ - **Total Repository Size**: ~21 MB
33
+ - **Original Model**: [openbmb/MiniCPM-o-2_6](https://huggingface.co/openbmb/MiniCPM-o-2_6) (~18 GB)
34
+ - **Size Reduction**: 853× smaller than the full model
35
+
36
+ ## Architecture Details
37
+
38
+ ### Language Model (LLM) Component
39
+ - `num_hidden_layers`: 2 (reduced from 40)
40
+ - `hidden_size`: 256 (reduced from 2048)
41
+ - `intermediate_size`: 512 (reduced from 8192)
42
+ - `num_attention_heads`: 4 (reduced from 32)
43
+ - `vocab_size`: 320 (reduced from 151,700)
44
+ - `max_position_embeddings`: 128 (reduced from 8192)
45
+
46
+ ### Vision Component (SigLIP-based)
47
+ - `hidden_size`: 8
48
+ - `num_hidden_layers`: 1
49
+
50
+ ### Audio Component (Whisper-based)
51
+ - `d_model`: 64
52
+ - `encoder_layers`: 1
53
+ - `decoder_layers`: 1
54
+
55
+ ### TTS Component
56
+ - `hidden_size`: 8
57
+ - `num_layers`: 1
58
+
59
+ All architectural components are present but miniaturized to ensure API compatibility while drastically reducing compute requirements.
60
+
61
+ ## Usage
62
+
63
+ ### Loading with Transformers
64
+
65
+ ```python
66
+ from transformers import AutoModelForCausalLM, AutoProcessor
67
+ import torch
68
+
69
+ model_id = "arashkermani/tiny-random-MiniCPM-o-2_6"
70
+
71
+ # Load model
72
+ model = AutoModelForCausalLM.from_pretrained(
73
+ model_id,
74
+ trust_remote_code=True,
75
+ torch_dtype=torch.float32,
76
+ device_map="cpu"
77
+ )
78
+
79
+ # Load processor
80
+ processor = AutoProcessor.from_pretrained(
81
+ model_id,
82
+ trust_remote_code=True
83
+ )
84
+
85
+ # Test forward pass
86
+ input_ids = torch.randint(0, 320, (1, 5))
87
+ position_ids = torch.arange(5).unsqueeze(0)
88
+
89
+ data = {
90
+ "input_ids": input_ids,
91
+ "pixel_values": [[]],
92
+ "tgt_sizes": [[]],
93
+ "image_bound": [[]],
94
+ "position_ids": position_ids,
95
+ }
96
+
97
+ with torch.no_grad():
98
+ outputs = model(data=data)
99
+
100
+ print(f"Logits shape: {outputs.logits.shape}") # (1, 5, 320)
101
+ ```
102
+
103
+ ### Using with Optimum-Intel (OpenVINO)
104
+
105
+ ```python
106
+ from optimum.intel.openvino import OVModelForVisualCausalLM
107
+ from transformers import AutoProcessor
108
+
109
+ model_id = "arashkermani/tiny-random-MiniCPM-o-2_6"
110
+
111
+ # Load model for OpenVINO
112
+ model = OVModelForVisualCausalLM.from_pretrained(
113
+ model_id,
114
+ trust_remote_code=True
115
+ )
116
+
117
+ processor = AutoProcessor.from_pretrained(
118
+ model_id,
119
+ trust_remote_code=True
120
+ )
121
+ ```
122
+
123
+ ### Export to OpenVINO
124
+
125
+ ```bash
126
+ optimum-cli export openvino \
127
+ -m arashkermani/tiny-random-MiniCPM-o-2_6 \
128
+ minicpm-o-openvino \
129
+ --task=image-text-to-text \
130
+ --trust-remote-code
131
+ ```
132
+
133
+ ## Intended Use
134
+
135
+ This model is intended **exclusively** for:
136
+ - ✅ Testing optimum-intel OpenVINO export functionality
137
+ - ✅ CI/CD pipeline validation
138
+ - ✅ Model loading and compatibility testing
139
+ - ✅ Quantization workflow testing
140
+ - ✅ Fast prototyping and debugging
141
+
142
+ **Not intended for**:
143
+ - ❌ Production inference
144
+ - ❌ Actual image-text-to-text tasks
145
+ - ❌ Model quality evaluation
146
+ - ❌ Benchmarking performance metrics
147
+
148
+ ## Training Details
149
+
150
+ This model was generated by:
151
+ 1. Loading the config from `optimum-intel-internal-testing/tiny-random-MiniCPM-o-2_6`
152
+ 2. Reducing all dimensions to minimal viable values
153
+ 3. Initializing weights randomly using `AutoModelForCausalLM.from_config()`
154
+ 4. Copying all necessary tokenizer, processor, and custom code files
155
+
156
+ **No training was performed** - all weights are randomly initialized.
157
+
158
+ ## Validation Results
159
+
160
+ The model has been validated to ensure:
161
+ - ✅ Loads with `trust_remote_code=True`
162
+ - ✅ Compatible with transformers AutoModel APIs
163
+ - ✅ Supports forward pass with expected input format
164
+ - ✅ Compatible with OpenVINO export via optimum-intel
165
+ - ✅ Includes all required custom modules and artifacts
166
+
167
+ See the [validation report](https://github.com/arashkermani/tiny-minicpm-o) for detailed technical analysis.
168
+
169
+ ## Files Included
170
+
171
+ - `config.json` - Model configuration
172
+ - `pytorch_model.bin` - Model weights (5.64 MB)
173
+ - `generation_config.json` - Generation parameters
174
+ - `preprocessor_config.json` - Preprocessor configuration
175
+ - `processor_config.json` - Processor configuration
176
+ - `tokenizer_config.json` - Tokenizer configuration
177
+ - `tokenizer.json` - Fast tokenizer
178
+ - `vocab.json` - Vocabulary
179
+ - `merges.txt` - BPE merges
180
+ - Custom Python modules:
181
+ - `modeling_minicpmo.py`
182
+ - `configuration_minicpm.py`
183
+ - `processing_minicpmo.py`
184
+ - `image_processing_minicpmv.py`
185
+ - `tokenization_minicpmo_fast.py`
186
+ - `modeling_navit_siglip.py`
187
+ - `resampler.py`
188
+ - `utils.py`
189
+
190
+ ## Related Models
191
+
192
+ - Original model: [openbmb/MiniCPM-o-2_6](https://huggingface.co/openbmb/MiniCPM-o-2_6)
193
+ - Previous test model: [optimum-intel-internal-testing/tiny-random-MiniCPM-o-2_6](https://huggingface.co/optimum-intel-internal-testing/tiny-random-MiniCPM-o-2_6)
194
+
195
+ ## License
196
+
197
+ This model follows the same license as the original MiniCPM-o-2_6 model (Apache 2.0).
198
+
199
+ ## Citation
200
+
201
+ If you use this test model in your CI/CD or testing infrastructure, please reference:
202
+
203
+ ```bibtex
204
+ @misc{tiny-minicpm-o-2_6,
205
+ author = {Arash Kermani},
206
+ title = {Tiny Random MiniCPM-o-2_6 for Testing},
207
+ year = {2026},
208
+ publisher = {HuggingFace},
209
+ howpublished = {\url{https://huggingface.co/arashkermani/tiny-random-MiniCPM-o-2_6}}
210
+ }
211
+ ```
212
+
213
+ ## Contact
214
+
215
+ For issues or questions about this test model, please open an issue in the [optimum-intel repository](https://github.com/huggingface/optimum-intel/issues).
assets/chattts_tokenizer/special_tokens_map.json ADDED
@@ -0,0 +1,389 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ {
4
+ "content": "[Sasr]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false
9
+ },
10
+ {
11
+ "content": "[Pasr]",
12
+ "lstrip": false,
13
+ "normalized": false,
14
+ "rstrip": false,
15
+ "single_word": false
16
+ },
17
+ {
18
+ "content": "[Easr]",
19
+ "lstrip": false,
20
+ "normalized": false,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ },
24
+ {
25
+ "content": "[Stts]",
26
+ "lstrip": false,
27
+ "normalized": false,
28
+ "rstrip": false,
29
+ "single_word": false
30
+ },
31
+ {
32
+ "content": "[Ptts]",
33
+ "lstrip": false,
34
+ "normalized": false,
35
+ "rstrip": false,
36
+ "single_word": false
37
+ },
38
+ {
39
+ "content": "[Etts]",
40
+ "lstrip": false,
41
+ "normalized": false,
42
+ "rstrip": false,
43
+ "single_word": false
44
+ },
45
+ {
46
+ "content": "[Sbreak]",
47
+ "lstrip": false,
48
+ "normalized": false,
49
+ "rstrip": false,
50
+ "single_word": false
51
+ },
52
+ {
53
+ "content": "[Pbreak]",
54
+ "lstrip": false,
55
+ "normalized": false,
56
+ "rstrip": false,
57
+ "single_word": false
58
+ },
59
+ {
60
+ "content": "[Ebreak]",
61
+ "lstrip": false,
62
+ "normalized": false,
63
+ "rstrip": false,
64
+ "single_word": false
65
+ },
66
+ {
67
+ "content": "[uv_break]",
68
+ "lstrip": false,
69
+ "normalized": false,
70
+ "rstrip": false,
71
+ "single_word": false
72
+ },
73
+ {
74
+ "content": "[v_break]",
75
+ "lstrip": false,
76
+ "normalized": false,
77
+ "rstrip": false,
78
+ "single_word": false
79
+ },
80
+ {
81
+ "content": "[lbreak]",
82
+ "lstrip": false,
83
+ "normalized": false,
84
+ "rstrip": false,
85
+ "single_word": false
86
+ },
87
+ {
88
+ "content": "[llbreak]",
89
+ "lstrip": false,
90
+ "normalized": false,
91
+ "rstrip": false,
92
+ "single_word": false
93
+ },
94
+ {
95
+ "content": "[undefine]",
96
+ "lstrip": false,
97
+ "normalized": false,
98
+ "rstrip": false,
99
+ "single_word": false
100
+ },
101
+ {
102
+ "content": "[laugh]",
103
+ "lstrip": false,
104
+ "normalized": false,
105
+ "rstrip": false,
106
+ "single_word": false
107
+ },
108
+ {
109
+ "content": "[spk_emb]",
110
+ "lstrip": false,
111
+ "normalized": false,
112
+ "rstrip": false,
113
+ "single_word": false
114
+ },
115
+ {
116
+ "content": "[empty_spk]",
117
+ "lstrip": false,
118
+ "normalized": false,
119
+ "rstrip": false,
120
+ "single_word": false
121
+ },
122
+ {
123
+ "content": "[music]",
124
+ "lstrip": false,
125
+ "normalized": false,
126
+ "rstrip": false,
127
+ "single_word": false
128
+ },
129
+ {
130
+ "content": "[pure]",
131
+ "lstrip": false,
132
+ "normalized": false,
133
+ "rstrip": false,
134
+ "single_word": false
135
+ },
136
+ {
137
+ "content": "[break_0]",
138
+ "lstrip": false,
139
+ "normalized": false,
140
+ "rstrip": false,
141
+ "single_word": false
142
+ },
143
+ {
144
+ "content": "[break_1]",
145
+ "lstrip": false,
146
+ "normalized": false,
147
+ "rstrip": false,
148
+ "single_word": false
149
+ },
150
+ {
151
+ "content": "[break_2]",
152
+ "lstrip": false,
153
+ "normalized": false,
154
+ "rstrip": false,
155
+ "single_word": false
156
+ },
157
+ {
158
+ "content": "[break_3]",
159
+ "lstrip": false,
160
+ "normalized": false,
161
+ "rstrip": false,
162
+ "single_word": false
163
+ },
164
+ {
165
+ "content": "[break_4]",
166
+ "lstrip": false,
167
+ "normalized": false,
168
+ "rstrip": false,
169
+ "single_word": false
170
+ },
171
+ {
172
+ "content": "[break_5]",
173
+ "lstrip": false,
174
+ "normalized": false,
175
+ "rstrip": false,
176
+ "single_word": false
177
+ },
178
+ {
179
+ "content": "[break_6]",
180
+ "lstrip": false,
181
+ "normalized": false,
182
+ "rstrip": false,
183
+ "single_word": false
184
+ },
185
+ {
186
+ "content": "[break_7]",
187
+ "lstrip": false,
188
+ "normalized": false,
189
+ "rstrip": false,
190
+ "single_word": false
191
+ },
192
+ {
193
+ "content": "[laugh_0]",
194
+ "lstrip": false,
195
+ "normalized": false,
196
+ "rstrip": false,
197
+ "single_word": false
198
+ },
199
+ {
200
+ "content": "[laugh_1]",
201
+ "lstrip": false,
202
+ "normalized": false,
203
+ "rstrip": false,
204
+ "single_word": false
205
+ },
206
+ {
207
+ "content": "[laugh_2]",
208
+ "lstrip": false,
209
+ "normalized": false,
210
+ "rstrip": false,
211
+ "single_word": false
212
+ },
213
+ {
214
+ "content": "[oral_0]",
215
+ "lstrip": false,
216
+ "normalized": false,
217
+ "rstrip": false,
218
+ "single_word": false
219
+ },
220
+ {
221
+ "content": "[oral_1]",
222
+ "lstrip": false,
223
+ "normalized": false,
224
+ "rstrip": false,
225
+ "single_word": false
226
+ },
227
+ {
228
+ "content": "[oral_2]",
229
+ "lstrip": false,
230
+ "normalized": false,
231
+ "rstrip": false,
232
+ "single_word": false
233
+ },
234
+ {
235
+ "content": "[oral_3]",
236
+ "lstrip": false,
237
+ "normalized": false,
238
+ "rstrip": false,
239
+ "single_word": false
240
+ },
241
+ {
242
+ "content": "[oral_4]",
243
+ "lstrip": false,
244
+ "normalized": false,
245
+ "rstrip": false,
246
+ "single_word": false
247
+ },
248
+ {
249
+ "content": "[oral_5]",
250
+ "lstrip": false,
251
+ "normalized": false,
252
+ "rstrip": false,
253
+ "single_word": false
254
+ },
255
+ {
256
+ "content": "[oral_6]",
257
+ "lstrip": false,
258
+ "normalized": false,
259
+ "rstrip": false,
260
+ "single_word": false
261
+ },
262
+ {
263
+ "content": "[oral_7]",
264
+ "lstrip": false,
265
+ "normalized": false,
266
+ "rstrip": false,
267
+ "single_word": false
268
+ },
269
+ {
270
+ "content": "[oral_8]",
271
+ "lstrip": false,
272
+ "normalized": false,
273
+ "rstrip": false,
274
+ "single_word": false
275
+ },
276
+ {
277
+ "content": "[oral_9]",
278
+ "lstrip": false,
279
+ "normalized": false,
280
+ "rstrip": false,
281
+ "single_word": false
282
+ },
283
+ {
284
+ "content": "[speed_0]",
285
+ "lstrip": false,
286
+ "normalized": false,
287
+ "rstrip": false,
288
+ "single_word": false
289
+ },
290
+ {
291
+ "content": "[speed_1]",
292
+ "lstrip": false,
293
+ "normalized": false,
294
+ "rstrip": false,
295
+ "single_word": false
296
+ },
297
+ {
298
+ "content": "[speed_2]",
299
+ "lstrip": false,
300
+ "normalized": false,
301
+ "rstrip": false,
302
+ "single_word": false
303
+ },
304
+ {
305
+ "content": "[speed_3]",
306
+ "lstrip": false,
307
+ "normalized": false,
308
+ "rstrip": false,
309
+ "single_word": false
310
+ },
311
+ {
312
+ "content": "[speed_4]",
313
+ "lstrip": false,
314
+ "normalized": false,
315
+ "rstrip": false,
316
+ "single_word": false
317
+ },
318
+ {
319
+ "content": "[speed_5]",
320
+ "lstrip": false,
321
+ "normalized": false,
322
+ "rstrip": false,
323
+ "single_word": false
324
+ },
325
+ {
326
+ "content": "[speed_6]",
327
+ "lstrip": false,
328
+ "normalized": false,
329
+ "rstrip": false,
330
+ "single_word": false
331
+ },
332
+ {
333
+ "content": "[speed_7]",
334
+ "lstrip": false,
335
+ "normalized": false,
336
+ "rstrip": false,
337
+ "single_word": false
338
+ },
339
+ {
340
+ "content": "[speed_8]",
341
+ "lstrip": false,
342
+ "normalized": false,
343
+ "rstrip": false,
344
+ "single_word": false
345
+ },
346
+ {
347
+ "content": "[speed_9]",
348
+ "lstrip": false,
349
+ "normalized": false,
350
+ "rstrip": false,
351
+ "single_word": false
352
+ }
353
+ ],
354
+ "cls_token": {
355
+ "content": "[CLS]",
356
+ "lstrip": false,
357
+ "normalized": false,
358
+ "rstrip": false,
359
+ "single_word": false
360
+ },
361
+ "mask_token": {
362
+ "content": "[MASK]",
363
+ "lstrip": false,
364
+ "normalized": false,
365
+ "rstrip": false,
366
+ "single_word": false
367
+ },
368
+ "pad_token": {
369
+ "content": "[PAD]",
370
+ "lstrip": false,
371
+ "normalized": false,
372
+ "rstrip": false,
373
+ "single_word": false
374
+ },
375
+ "sep_token": {
376
+ "content": "[SEP]",
377
+ "lstrip": false,
378
+ "normalized": false,
379
+ "rstrip": false,
380
+ "single_word": false
381
+ },
382
+ "unk_token": {
383
+ "content": "[UNK]",
384
+ "lstrip": false,
385
+ "normalized": false,
386
+ "rstrip": false,
387
+ "single_word": false
388
+ }
389
+ }
assets/chattts_tokenizer/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
assets/chattts_tokenizer/tokenizer_config.json ADDED
@@ -0,0 +1,516 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "100": {
12
+ "content": "[UNK]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "101": {
20
+ "content": "[CLS]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "102": {
28
+ "content": "[SEP]",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "103": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ },
43
+ "21128": {
44
+ "content": "[Sasr]",
45
+ "lstrip": false,
46
+ "normalized": false,
47
+ "rstrip": false,
48
+ "single_word": false,
49
+ "special": true
50
+ },
51
+ "21129": {
52
+ "content": "[Pasr]",
53
+ "lstrip": false,
54
+ "normalized": false,
55
+ "rstrip": false,
56
+ "single_word": false,
57
+ "special": true
58
+ },
59
+ "21130": {
60
+ "content": "[Easr]",
61
+ "lstrip": false,
62
+ "normalized": false,
63
+ "rstrip": false,
64
+ "single_word": false,
65
+ "special": true
66
+ },
67
+ "21131": {
68
+ "content": "[Stts]",
69
+ "lstrip": false,
70
+ "normalized": false,
71
+ "rstrip": false,
72
+ "single_word": false,
73
+ "special": true
74
+ },
75
+ "21132": {
76
+ "content": "[Ptts]",
77
+ "lstrip": false,
78
+ "normalized": false,
79
+ "rstrip": false,
80
+ "single_word": false,
81
+ "special": true
82
+ },
83
+ "21133": {
84
+ "content": "[Etts]",
85
+ "lstrip": false,
86
+ "normalized": false,
87
+ "rstrip": false,
88
+ "single_word": false,
89
+ "special": true
90
+ },
91
+ "21134": {
92
+ "content": "[Sbreak]",
93
+ "lstrip": false,
94
+ "normalized": false,
95
+ "rstrip": false,
96
+ "single_word": false,
97
+ "special": true
98
+ },
99
+ "21135": {
100
+ "content": "[Pbreak]",
101
+ "lstrip": false,
102
+ "normalized": false,
103
+ "rstrip": false,
104
+ "single_word": false,
105
+ "special": true
106
+ },
107
+ "21136": {
108
+ "content": "[Ebreak]",
109
+ "lstrip": false,
110
+ "normalized": false,
111
+ "rstrip": false,
112
+ "single_word": false,
113
+ "special": true
114
+ },
115
+ "21137": {
116
+ "content": "[uv_break]",
117
+ "lstrip": false,
118
+ "normalized": false,
119
+ "rstrip": false,
120
+ "single_word": false,
121
+ "special": true
122
+ },
123
+ "21138": {
124
+ "content": "[v_break]",
125
+ "lstrip": false,
126
+ "normalized": false,
127
+ "rstrip": false,
128
+ "single_word": false,
129
+ "special": true
130
+ },
131
+ "21139": {
132
+ "content": "[lbreak]",
133
+ "lstrip": false,
134
+ "normalized": false,
135
+ "rstrip": false,
136
+ "single_word": false,
137
+ "special": true
138
+ },
139
+ "21140": {
140
+ "content": "[llbreak]",
141
+ "lstrip": false,
142
+ "normalized": false,
143
+ "rstrip": false,
144
+ "single_word": false,
145
+ "special": true
146
+ },
147
+ "21141": {
148
+ "content": "[undefine]",
149
+ "lstrip": false,
150
+ "normalized": false,
151
+ "rstrip": false,
152
+ "single_word": false,
153
+ "special": true
154
+ },
155
+ "21142": {
156
+ "content": "[laugh]",
157
+ "lstrip": false,
158
+ "normalized": false,
159
+ "rstrip": false,
160
+ "single_word": false,
161
+ "special": true
162
+ },
163
+ "21143": {
164
+ "content": "[spk_emb]",
165
+ "lstrip": false,
166
+ "normalized": false,
167
+ "rstrip": false,
168
+ "single_word": false,
169
+ "special": true
170
+ },
171
+ "21144": {
172
+ "content": "[empty_spk]",
173
+ "lstrip": false,
174
+ "normalized": false,
175
+ "rstrip": false,
176
+ "single_word": false,
177
+ "special": true
178
+ },
179
+ "21145": {
180
+ "content": "[music]",
181
+ "lstrip": false,
182
+ "normalized": false,
183
+ "rstrip": false,
184
+ "single_word": false,
185
+ "special": true
186
+ },
187
+ "21146": {
188
+ "content": "[pure]",
189
+ "lstrip": false,
190
+ "normalized": false,
191
+ "rstrip": false,
192
+ "single_word": false,
193
+ "special": true
194
+ },
195
+ "21147": {
196
+ "content": "[break_0]",
197
+ "lstrip": false,
198
+ "normalized": false,
199
+ "rstrip": false,
200
+ "single_word": false,
201
+ "special": true
202
+ },
203
+ "21148": {
204
+ "content": "[break_1]",
205
+ "lstrip": false,
206
+ "normalized": false,
207
+ "rstrip": false,
208
+ "single_word": false,
209
+ "special": true
210
+ },
211
+ "21149": {
212
+ "content": "[break_2]",
213
+ "lstrip": false,
214
+ "normalized": false,
215
+ "rstrip": false,
216
+ "single_word": false,
217
+ "special": true
218
+ },
219
+ "21150": {
220
+ "content": "[break_3]",
221
+ "lstrip": false,
222
+ "normalized": false,
223
+ "rstrip": false,
224
+ "single_word": false,
225
+ "special": true
226
+ },
227
+ "21151": {
228
+ "content": "[break_4]",
229
+ "lstrip": false,
230
+ "normalized": false,
231
+ "rstrip": false,
232
+ "single_word": false,
233
+ "special": true
234
+ },
235
+ "21152": {
236
+ "content": "[break_5]",
237
+ "lstrip": false,
238
+ "normalized": false,
239
+ "rstrip": false,
240
+ "single_word": false,
241
+ "special": true
242
+ },
243
+ "21153": {
244
+ "content": "[break_6]",
245
+ "lstrip": false,
246
+ "normalized": false,
247
+ "rstrip": false,
248
+ "single_word": false,
249
+ "special": true
250
+ },
251
+ "21154": {
252
+ "content": "[break_7]",
253
+ "lstrip": false,
254
+ "normalized": false,
255
+ "rstrip": false,
256
+ "single_word": false,
257
+ "special": true
258
+ },
259
+ "21155": {
260
+ "content": "[laugh_0]",
261
+ "lstrip": false,
262
+ "normalized": false,
263
+ "rstrip": false,
264
+ "single_word": false,
265
+ "special": true
266
+ },
267
+ "21156": {
268
+ "content": "[laugh_1]",
269
+ "lstrip": false,
270
+ "normalized": false,
271
+ "rstrip": false,
272
+ "single_word": false,
273
+ "special": true
274
+ },
275
+ "21157": {
276
+ "content": "[laugh_2]",
277
+ "lstrip": false,
278
+ "normalized": false,
279
+ "rstrip": false,
280
+ "single_word": false,
281
+ "special": true
282
+ },
283
+ "21158": {
284
+ "content": "[oral_0]",
285
+ "lstrip": false,
286
+ "normalized": false,
287
+ "rstrip": false,
288
+ "single_word": false,
289
+ "special": true
290
+ },
291
+ "21159": {
292
+ "content": "[oral_1]",
293
+ "lstrip": false,
294
+ "normalized": false,
295
+ "rstrip": false,
296
+ "single_word": false,
297
+ "special": true
298
+ },
299
+ "21160": {
300
+ "content": "[oral_2]",
301
+ "lstrip": false,
302
+ "normalized": false,
303
+ "rstrip": false,
304
+ "single_word": false,
305
+ "special": true
306
+ },
307
+ "21161": {
308
+ "content": "[oral_3]",
309
+ "lstrip": false,
310
+ "normalized": false,
311
+ "rstrip": false,
312
+ "single_word": false,
313
+ "special": true
314
+ },
315
+ "21162": {
316
+ "content": "[oral_4]",
317
+ "lstrip": false,
318
+ "normalized": false,
319
+ "rstrip": false,
320
+ "single_word": false,
321
+ "special": true
322
+ },
323
+ "21163": {
324
+ "content": "[oral_5]",
325
+ "lstrip": false,
326
+ "normalized": false,
327
+ "rstrip": false,
328
+ "single_word": false,
329
+ "special": true
330
+ },
331
+ "21164": {
332
+ "content": "[oral_6]",
333
+ "lstrip": false,
334
+ "normalized": false,
335
+ "rstrip": false,
336
+ "single_word": false,
337
+ "special": true
338
+ },
339
+ "21165": {
340
+ "content": "[oral_7]",
341
+ "lstrip": false,
342
+ "normalized": false,
343
+ "rstrip": false,
344
+ "single_word": false,
345
+ "special": true
346
+ },
347
+ "21166": {
348
+ "content": "[oral_8]",
349
+ "lstrip": false,
350
+ "normalized": false,
351
+ "rstrip": false,
352
+ "single_word": false,
353
+ "special": true
354
+ },
355
+ "21167": {
356
+ "content": "[oral_9]",
357
+ "lstrip": false,
358
+ "normalized": false,
359
+ "rstrip": false,
360
+ "single_word": false,
361
+ "special": true
362
+ },
363
+ "21168": {
364
+ "content": "[speed_0]",
365
+ "lstrip": false,
366
+ "normalized": false,
367
+ "rstrip": false,
368
+ "single_word": false,
369
+ "special": true
370
+ },
371
+ "21169": {
372
+ "content": "[speed_1]",
373
+ "lstrip": false,
374
+ "normalized": false,
375
+ "rstrip": false,
376
+ "single_word": false,
377
+ "special": true
378
+ },
379
+ "21170": {
380
+ "content": "[speed_2]",
381
+ "lstrip": false,
382
+ "normalized": false,
383
+ "rstrip": false,
384
+ "single_word": false,
385
+ "special": true
386
+ },
387
+ "21171": {
388
+ "content": "[speed_3]",
389
+ "lstrip": false,
390
+ "normalized": false,
391
+ "rstrip": false,
392
+ "single_word": false,
393
+ "special": true
394
+ },
395
+ "21172": {
396
+ "content": "[speed_4]",
397
+ "lstrip": false,
398
+ "normalized": false,
399
+ "rstrip": false,
400
+ "single_word": false,
401
+ "special": true
402
+ },
403
+ "21173": {
404
+ "content": "[speed_5]",
405
+ "lstrip": false,
406
+ "normalized": false,
407
+ "rstrip": false,
408
+ "single_word": false,
409
+ "special": true
410
+ },
411
+ "21174": {
412
+ "content": "[speed_6]",
413
+ "lstrip": false,
414
+ "normalized": false,
415
+ "rstrip": false,
416
+ "single_word": false,
417
+ "special": true
418
+ },
419
+ "21175": {
420
+ "content": "[speed_7]",
421
+ "lstrip": false,
422
+ "normalized": false,
423
+ "rstrip": false,
424
+ "single_word": false,
425
+ "special": true
426
+ },
427
+ "21176": {
428
+ "content": "[speed_8]",
429
+ "lstrip": false,
430
+ "normalized": false,
431
+ "rstrip": false,
432
+ "single_word": false,
433
+ "special": true
434
+ },
435
+ "21177": {
436
+ "content": "[speed_9]",
437
+ "lstrip": false,
438
+ "normalized": false,
439
+ "rstrip": false,
440
+ "single_word": false,
441
+ "special": true
442
+ }
443
+ },
444
+ "additional_special_tokens": [
445
+ "[Sasr]",
446
+ "[Pasr]",
447
+ "[Easr]",
448
+ "[Stts]",
449
+ "[Ptts]",
450
+ "[Etts]",
451
+ "[Sbreak]",
452
+ "[Pbreak]",
453
+ "[Ebreak]",
454
+ "[uv_break]",
455
+ "[v_break]",
456
+ "[lbreak]",
457
+ "[llbreak]",
458
+ "[undefine]",
459
+ "[laugh]",
460
+ "[spk_emb]",
461
+ "[empty_spk]",
462
+ "[music]",
463
+ "[pure]",
464
+ "[break_0]",
465
+ "[break_1]",
466
+ "[break_2]",
467
+ "[break_3]",
468
+ "[break_4]",
469
+ "[break_5]",
470
+ "[break_6]",
471
+ "[break_7]",
472
+ "[laugh_0]",
473
+ "[laugh_1]",
474
+ "[laugh_2]",
475
+ "[oral_0]",
476
+ "[oral_1]",
477
+ "[oral_2]",
478
+ "[oral_3]",
479
+ "[oral_4]",
480
+ "[oral_5]",
481
+ "[oral_6]",
482
+ "[oral_7]",
483
+ "[oral_8]",
484
+ "[oral_9]",
485
+ "[speed_0]",
486
+ "[speed_1]",
487
+ "[speed_2]",
488
+ "[speed_3]",
489
+ "[speed_4]",
490
+ "[speed_5]",
491
+ "[speed_6]",
492
+ "[speed_7]",
493
+ "[speed_8]",
494
+ "[speed_9]"
495
+ ],
496
+ "clean_up_tokenization_spaces": true,
497
+ "cls_token": "[CLS]",
498
+ "do_basic_tokenize": true,
499
+ "do_lower_case": true,
500
+ "mask_token": "[MASK]",
501
+ "max_length": 256,
502
+ "model_max_length": 1000000000000000019884624838656,
503
+ "never_split": null,
504
+ "pad_to_multiple_of": null,
505
+ "pad_token": "[PAD]",
506
+ "pad_token_type_id": 0,
507
+ "padding_side": "right",
508
+ "sep_token": "[SEP]",
509
+ "stride": 0,
510
+ "strip_accents": null,
511
+ "tokenize_chinese_chars": true,
512
+ "tokenizer_class": "BertTokenizer",
513
+ "truncation_side": "right",
514
+ "truncation_strategy": "longest_first",
515
+ "unk_token": "[UNK]"
516
+ }
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "/Users/arashkermanikolankeh/Downloads/Altana/tiny_minicpm_o_2_6",
3
  "architectures": [
4
  "MiniCPMO"
5
  ],
@@ -148,9 +148,9 @@
148
  "hidden_act": "silu",
149
  "hidden_size": 256,
150
  "image_size": 448,
151
- "init_audio": false,
152
- "init_tts": false,
153
- "init_vision": false,
154
  "initializer_range": 0.02,
155
  "intermediate_size": 512,
156
  "listen_speak_type": "asr",
@@ -158,7 +158,6 @@
158
  "max_window_layers": 28,
159
  "model_type": "minicpmo",
160
  "num_attention_heads": 4,
161
- "num_heads": 1,
162
  "num_hidden_layers": 2,
163
  "num_key_value_heads": 4,
164
  "patch_size": 14,
@@ -183,9 +182,7 @@
183
  "model_type": "conditional_chattts",
184
  "num_attention_heads": 1,
185
  "num_audio_tokens": 10,
186
- "num_heads": 1,
187
  "num_hidden_layers": 1,
188
- "num_layers": 1,
189
  "num_mel_bins": 10,
190
  "num_text_tokens": 20
191
  },
 
1
  {
2
+ "_name_or_path": "openbmb/MiniCPM-o-2_6",
3
  "architectures": [
4
  "MiniCPMO"
5
  ],
 
148
  "hidden_act": "silu",
149
  "hidden_size": 256,
150
  "image_size": 448,
151
+ "init_audio": true,
152
+ "init_tts": true,
153
+ "init_vision": true,
154
  "initializer_range": 0.02,
155
  "intermediate_size": 512,
156
  "listen_speak_type": "asr",
 
158
  "max_window_layers": 28,
159
  "model_type": "minicpmo",
160
  "num_attention_heads": 4,
 
161
  "num_hidden_layers": 2,
162
  "num_key_value_heads": 4,
163
  "patch_size": 14,
 
182
  "model_type": "conditional_chattts",
183
  "num_attention_heads": 1,
184
  "num_audio_tokens": 10,
 
185
  "num_hidden_layers": 1,
 
186
  "num_mel_bins": 10,
187
  "num_text_tokens": 20
188
  },
preprocessor_config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "auto_map": {
3
- "AutoImageProcessor": "image_processing_minicpmv.MiniCPMVImageProcessor",
4
- "AutoProcessor": "processing_minicpmo.MiniCPMOProcessor"
5
  },
6
  "chunk_length": 30,
7
  "feature_extractor_type": "WhisperFeatureExtractor",
 
1
  {
2
  "auto_map": {
3
+ "AutoImageProcessor": "openbmb/MiniCPM-o-2_6--image_processing_minicpmv.MiniCPMVImageProcessor",
4
+ "AutoProcessor": "openbmb/MiniCPM-o-2_6--processing_minicpmo.MiniCPMOProcessor"
5
  },
6
  "chunk_length": 30,
7
  "feature_extractor_type": "WhisperFeatureExtractor",
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:979e8692eb0dc05dc0e85077d94d6feaa968f042f906524c0028250d074480eb
3
- size 5919000
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5af168eabd5d0954ff83073fe3b36fc7fd6a8c9e8b6591132313241d8165d91c
3
+ size 69676335
tokenizer_config.json CHANGED
@@ -505,10 +505,9 @@
505
  "<reserved_53>"
506
  ],
507
  "auto_map": {
508
- "AutoProcessor": "processing_minicpmo.MiniCPMOProcessor",
509
  "AutoTokenizer": [
510
- "tokenization_qwen2.Qwen2Tokenizer",
511
- "tokenization_minicpmo_fast.MiniCPMOTokenizerFast"
512
  ]
513
  },
514
  "bos_token": "<|im_start|>",
 
505
  "<reserved_53>"
506
  ],
507
  "auto_map": {
 
508
  "AutoTokenizer": [
509
+ "openbmb/MiniCPM-o-2_6--tokenization_minicpmo_fast.MiniCPMOTokenizerFast",
510
+ null
511
  ]
512
  },
513
  "bos_token": "<|im_start|>",