Commit
·
add6575
1
Parent(s):
4762d81
Update README.md
Browse files
README.md
CHANGED
|
@@ -4,4 +4,19 @@ tags:
|
|
| 4 |
library_name: generic
|
| 5 |
---
|
| 6 |
|
| 7 |
-
# Optimized and Quantized DistilBERT with a custom pipeline.py
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 4 |
library_name: generic
|
| 5 |
---
|
| 6 |
|
| 7 |
+
# Optimized and Quantized DistilBERT with a custom pipeline.py
|
| 8 |
+
|
| 9 |
+
> NOTE: Blog post coming soon
|
| 10 |
+
|
| 11 |
+
This is a template repository for Text Classification using Optimum and onnxruntime to support generic inference with Hugging Face Hub generic Inference API. There are two required steps:
|
| 12 |
+
|
| 13 |
+
1. Specify the requirements by defining a `requirements.txt` file.
|
| 14 |
+
2. Implement the `pipeline.py` `__init__` and `__call__` methods. These methods are called by the Inference API. The `__init__` method should load the model and preload the optimum model and tokenizers as well as the `text-classification` pipeline needed for inference. This is only called once. The `__call__` method performs the actual inference. Make sure to follow the same input/output specifications defined in the template for the pipeline to work.
|
| 15 |
+
|
| 16 |
+
add
|
| 17 |
+
```
|
| 18 |
+
library_name: generic
|
| 19 |
+
```
|
| 20 |
+
to the readme.
|
| 21 |
+
|
| 22 |
+
_note: the `generic` community image currently only support `inputs` as parameter and no parameter._
|