Update README.md
Browse files
README.md
CHANGED
|
@@ -8,11 +8,10 @@ tags:
|
|
| 8 |
---
|
| 9 |
|
| 10 |
This model provides [HuggingFaceTB/SmolVLM-256M-Instruct](https://huggingface.co/HuggingFaceTB/SmolVLM-256M-Instruct) model in TFLite format.
|
| 11 |
-
You can use this model with [Custom Cpp Pipiline](https://github.com/dragynir/ai-edge-torch-smalvlm/tree/dev/ai_edge_torch/generative/examples/cpp_image)
|
| 12 |
-
or run with python pipeline (see COLAB example below).
|
| 13 |
Please note that, at the moment, [AI Edge Torch](https://github.com/google-ai-edge/ai-edge-torch/tree/main/ai_edge_torch/generative/examples) VLMS not supported
|
| 14 |
on [MediaPipe LLM Inference API](https://ai.google.dev/edge/mediapipe/solutions/genai/llm_inference),
|
| 15 |
-
for example [qwen_vl model](https://github.com/google-ai-edge/ai-edge-torch/tree/main/ai_edge_torch/generative/examples/qwen_vl),
|
| 16 |
that was used as reference to write SmolVLM-256M-Instruct convertation scripts.
|
| 17 |
|
| 18 |
|
|
@@ -23,6 +22,19 @@ that was used as reference to write SmolVLM-256M-Instruct convertation scripts.
|
|
| 23 |
[](https://colab.research.google.com/#fileId=https://huggingface.co/litert-community/SmolVLM-256M-Instruct/blob/main/smalvlm_notebook.ipynb
|
| 24 |
)
|
| 25 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 26 |
|
| 27 |
## TFlite convertation
|
| 28 |
|
|
|
|
| 8 |
---
|
| 9 |
|
| 10 |
This model provides [HuggingFaceTB/SmolVLM-256M-Instruct](https://huggingface.co/HuggingFaceTB/SmolVLM-256M-Instruct) model in TFLite format.
|
| 11 |
+
You can use this model with [Custom Cpp Pipiline](https://github.com/dragynir/ai-edge-torch-smalvlm/tree/dev/ai_edge_torch/generative/examples/cpp_image) or run with python pipeline (see COLAB example below).
|
|
|
|
| 12 |
Please note that, at the moment, [AI Edge Torch](https://github.com/google-ai-edge/ai-edge-torch/tree/main/ai_edge_torch/generative/examples) VLMS not supported
|
| 13 |
on [MediaPipe LLM Inference API](https://ai.google.dev/edge/mediapipe/solutions/genai/llm_inference),
|
| 14 |
+
for example [qwen_vl model](https://github.com/google-ai-edge/ai-edge-torch/tree/main/ai_edge_torch/generative/examples/qwen_vl),
|
| 15 |
that was used as reference to write SmolVLM-256M-Instruct convertation scripts.
|
| 16 |
|
| 17 |
|
|
|
|
| 22 |
[](https://colab.research.google.com/#fileId=https://huggingface.co/litert-community/SmolVLM-256M-Instruct/blob/main/smalvlm_notebook.ipynb
|
| 23 |
)
|
| 24 |
|
| 25 |
+
## Cpp inference
|
| 26 |
+
|
| 27 |
+
```shell
|
| 28 |
+
mkdir cache
|
| 29 |
+
|
| 30 |
+
bazel run --verbose_failures -c opt //ai_edge_torch/generative/examples/cpp_image:text_generator_main -- \
|
| 31 |
+
--tflite_model="/home/dragynir/ai_vlm/ai-edge-torch-smalvlm/ai_edge_torch/generative/examples/smalvlm/models/SmolVLM-256M-Instruct-tflite-single/smalvlm-256m-instruct_q8_ekv2048.tflite" \
|
| 32 |
+
--sentencepiece_model="/home/dragynir/ai_vlm/ai-edge-torch-smalvlm/ai_edge_torch/generative/examples/smalvlm/models/SmolVLM-256M-Instruct-tflite/tokenizer.model" \
|
| 33 |
+
--start_token="<|im_start|>" --stop_token="<end_of_utterance>" --num_threads=16 \
|
| 34 |
+
--prompt="User:<image>What in the image?<end_of_utterance>\nAssistant:" --weight_cache_path="/home/dragynir/llm/ai-edge-torch/ai_edge_torch/generative/examples/cpp/cache/model.xnnpack_cache" \
|
| 35 |
+
--use_single_image=true --image_path="/home/dragynir/ai_vlm/car.jpg" --max_generated_tokens=64
|
| 36 |
+
```
|
| 37 |
+
|
| 38 |
|
| 39 |
## TFlite convertation
|
| 40 |
|