DISTILBERT RUNNING ON DEEPSPARSE GOES BRHMMMMMMMM. 🚀🚀🚀

This model is 👇

    ███████╗ ██████╗   █████╗  ██████╗  ███████╗ ███████╗
    ██╔════╝ ██╔══██╗ ██╔══██╗ ██╔══██╗  ██╔════╝ ██╔════╝
    ███████╗ ██████╔╝ ███████║ ██████╔╝  ███████╗ █████╗  
    ╚════██║ ██╔═══╝  ██╔══██║ ██╔══██╗ ╚════██║█ █╔══╝  
    ███████║ ██║      ██║  ██║ ██║  ██ ║███████║ ███████╗
    ╚══════╝ ╚═╝      ╚═╝  ╚═╝ ╚═╝  ╚═ ╝╚══════╝ ╚══════╝
                                                                                                     

Alt Text

LOOKS LIKE THIS 👇

Imgur

Inference endpoints, outside of outliers (4ms) is avg. latency on 2 vCPUs:

Imgur

Handler for access to inference endpoints

class EndpointHandler:

    def __init__(self, path=""):

        self.pipeline = Pipeline.create(task="text-classification", model_path=path)

    def __call__(self, data: Dict[str, Any]) -> Dict[str, str]:
        """
        Args:
            data (:obj:): prediction input text
        """
        inputs = data.pop("inputs", data)

        start = perf_counter()
        prediction = self.pipeline(inputs)
        end = perf_counter()
        latency = end - start

        return {
            "labels": prediction.labels, 
            "scores": prediction.scores,
            "latency (secs.)": latency
        }

̷͈̍ ̵̳͒R̶̙̓i̸̟͘c̴̻̆k̸̑͜ÿ̷̳́ ̸̪̚ ̷͖̀

Downloads last month
-
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support