Update: coreml
Browse files- README.md +5 -4
- coreml_minicpmv40_vit_f16.mlmodelc.zip +3 -0
- coreml_minicpmv40_vit_f16.mlmodelc/analytics/coremldata.bin +3 -0
- coreml_minicpmv40_vit_f16.mlmodelc/coremldata.bin +3 -0
- coreml_minicpmv40_vit_f16.mlmodelc/metadata.json +106 -0
- coreml_minicpmv40_vit_f16.mlmodelc/model.mil +0 -0
- coreml_minicpmv40_vit_f16.mlmodelc/weights/weight.bin +3 -0
- mmproj-model-f16-iOS.gguf +3 -0
README.md
CHANGED
@@ -12,7 +12,6 @@ tags:
|
|
12 |
- multi-image
|
13 |
- video
|
14 |
- custom_code
|
15 |
-
license: apache-2.0
|
16 |
---
|
17 |
|
18 |
<h1>A GPT-4V Level MLLM for Single Image, Multi Image and Video on Your Phone</h1>
|
@@ -619,8 +618,9 @@ print(answer)
|
|
619 |
|
620 |
## License
|
621 |
#### Model License
|
622 |
-
* The
|
623 |
-
*
|
|
|
624 |
|
625 |
|
626 |
#### Statement
|
@@ -644,4 +644,5 @@ If you find our work helpful, please consider citing our papers 📝 and liking
|
|
644 |
journal={Nat Commun 16, 5509 (2025)},
|
645 |
year={2025}
|
646 |
}
|
647 |
-
```
|
|
|
|
12 |
- multi-image
|
13 |
- video
|
14 |
- custom_code
|
|
|
15 |
---
|
16 |
|
17 |
<h1>A GPT-4V Level MLLM for Single Image, Multi Image and Video on Your Phone</h1>
|
|
|
618 |
|
619 |
## License
|
620 |
#### Model License
|
621 |
+
* The code in this repo is released under the [Apache-2.0](https://github.com/OpenBMB/MiniCPM/blob/main/LICENSE) License.
|
622 |
+
* The usage of MiniCPM-V series model weights must strictly follow [MiniCPM Model License.md](https://github.com/OpenBMB/MiniCPM/blob/main/MiniCPM%20Model%20License.md).
|
623 |
+
* The models and weights of MiniCPM are completely free for academic research. After filling out a ["questionnaire"](https://modelbest.feishu.cn/share/base/form/shrcnpV5ZT9EJ6xYjh3Kx0J6v8g) for registration, MiniCPM-V 2.6 weights are also available for free commercial use.
|
624 |
|
625 |
|
626 |
#### Statement
|
|
|
644 |
journal={Nat Commun 16, 5509 (2025)},
|
645 |
year={2025}
|
646 |
}
|
647 |
+
```
|
648 |
+
|
coreml_minicpmv40_vit_f16.mlmodelc.zip
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5255eed60eb864dab617cc3657f9c319ae6fdc2e85febc6d3dfc5c537f5fcd0a
|
3 |
+
size 690694016
|
coreml_minicpmv40_vit_f16.mlmodelc/analytics/coremldata.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:58aebe129df9a4a4ca5aba84e3fa9861308394cbd0cb97fd8cca17f438b0ead4
|
3 |
+
size 243
|
coreml_minicpmv40_vit_f16.mlmodelc/coremldata.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:eb9515d066e82c06b158ff723b8c930a8fd101753b15b219f782d3dba6d23772
|
3 |
+
size 739
|
coreml_minicpmv40_vit_f16.mlmodelc/metadata.json
ADDED
@@ -0,0 +1,106 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
[
|
2 |
+
{
|
3 |
+
"shortDescription" : "MiniCPM-V 4.5 vit on ANE",
|
4 |
+
"metadataOutputVersion" : "3.0",
|
5 |
+
"outputSchema" : [
|
6 |
+
{
|
7 |
+
"hasShapeFlexibility" : "0",
|
8 |
+
"isOptional" : "0",
|
9 |
+
"dataType" : "Float32",
|
10 |
+
"formattedType" : "MultiArray (Float32 1 × 64 × 2560)",
|
11 |
+
"shortDescription" : "",
|
12 |
+
"shape" : "[1, 64, 2560]",
|
13 |
+
"name" : "output",
|
14 |
+
"type" : "MultiArray"
|
15 |
+
}
|
16 |
+
],
|
17 |
+
"version" : "4.5.0",
|
18 |
+
"modelParameters" : [
|
19 |
+
|
20 |
+
],
|
21 |
+
"author" : "tianchi",
|
22 |
+
"specificationVersion" : 6,
|
23 |
+
"storagePrecision" : "Float16",
|
24 |
+
"license" : "Apache 2.0",
|
25 |
+
"mlProgramOperationTypeHistogram" : {
|
26 |
+
"Linear" : 167,
|
27 |
+
"Gelu" : 27,
|
28 |
+
"LayerNorm" : 57,
|
29 |
+
"Transpose" : 114,
|
30 |
+
"Matmul" : 56,
|
31 |
+
"Softmax" : 28,
|
32 |
+
"Mul" : 27,
|
33 |
+
"Cast" : 3,
|
34 |
+
"Reshape" : 113,
|
35 |
+
"Add" : 56,
|
36 |
+
"Gather" : 1,
|
37 |
+
"Conv" : 1
|
38 |
+
},
|
39 |
+
"computePrecision" : "Mixed (Float16, Float32, Int32)",
|
40 |
+
"stateSchema" : [
|
41 |
+
|
42 |
+
],
|
43 |
+
"isUpdatable" : "0",
|
44 |
+
"availability" : {
|
45 |
+
"macOS" : "12.0",
|
46 |
+
"tvOS" : "15.0",
|
47 |
+
"visionOS" : "1.0",
|
48 |
+
"watchOS" : "8.0",
|
49 |
+
"iOS" : "15.0",
|
50 |
+
"macCatalyst" : "15.0"
|
51 |
+
},
|
52 |
+
"modelType" : {
|
53 |
+
"name" : "MLModelType_mlProgram"
|
54 |
+
},
|
55 |
+
"inputSchema" : [
|
56 |
+
{
|
57 |
+
"hasShapeFlexibility" : "0",
|
58 |
+
"isOptional" : "0",
|
59 |
+
"dataType" : "Float32",
|
60 |
+
"formattedType" : "MultiArray (Float32 1 × 3 × 14 × 14336)",
|
61 |
+
"shortDescription" : "",
|
62 |
+
"shape" : "[1, 3, 14, 14336]",
|
63 |
+
"name" : "pixel_values",
|
64 |
+
"type" : "MultiArray"
|
65 |
+
},
|
66 |
+
{
|
67 |
+
"hasShapeFlexibility" : "0",
|
68 |
+
"isOptional" : "0",
|
69 |
+
"dataType" : "Int32",
|
70 |
+
"formattedType" : "MultiArray (Int32 1 × 1024)",
|
71 |
+
"shortDescription" : "",
|
72 |
+
"shape" : "[1, 1024]",
|
73 |
+
"name" : "position_ids",
|
74 |
+
"type" : "MultiArray"
|
75 |
+
},
|
76 |
+
{
|
77 |
+
"hasShapeFlexibility" : "0",
|
78 |
+
"isOptional" : "0",
|
79 |
+
"dataType" : "Float32",
|
80 |
+
"formattedType" : "MultiArray (Float32 1024 × 1 × 2560)",
|
81 |
+
"shortDescription" : "",
|
82 |
+
"shape" : "[1024, 1, 2560]",
|
83 |
+
"name" : "pos_embed",
|
84 |
+
"type" : "MultiArray"
|
85 |
+
}
|
86 |
+
],
|
87 |
+
"userDefinedMetadata" : {
|
88 |
+
"compute_units" : "ALL",
|
89 |
+
"deployment_target" : "iOS15+",
|
90 |
+
"model_id" : "b4a55ee32a106eb7d3c0da90163713c4",
|
91 |
+
"com.github.apple.coremltools.version" : "8.3.0",
|
92 |
+
"batch_size" : "1",
|
93 |
+
"base_model" : "MiniCPM-V 4.5",
|
94 |
+
"owner" : "tianchi",
|
95 |
+
"input_shape" : "torch.Size([1, 3, 14, 14336])",
|
96 |
+
"precision" : "float16",
|
97 |
+
"com.github.apple.coremltools.source_dialect" : "TorchScript",
|
98 |
+
"target_device" : "ANE",
|
99 |
+
"com.github.apple.coremltools.source" : "torch==2.6.0",
|
100 |
+
"converter" : "coremltools",
|
101 |
+
"model_type" : "vit"
|
102 |
+
},
|
103 |
+
"generatedClassName" : "coreml_minicpmv40_vit_f16",
|
104 |
+
"method" : "predict"
|
105 |
+
}
|
106 |
+
]
|
coreml_minicpmv40_vit_f16.mlmodelc/model.mil
ADDED
The diff for this file is too large to render.
See raw diff
|
|
coreml_minicpmv40_vit_f16.mlmodelc/weights/weight.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:81ab9e18839923002b08b49fa2f313bf49db30d5fc6227bbc49f98db15cce494
|
3 |
+
size 894309952
|
mmproj-model-f16-iOS.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8c82c9d2eda698e92f48655ee385bbcd3adb589231784fe4abd5144492630a86
|
3 |
+
size 135019040
|