Issues Deploying Sesame / CSM-1B on Hugging Face Inference and SageMaker
#53
by
Vishal05
- opened
Hi everyone,
We’re trying to deploy the sesame/csm-1b model, but running into multiple issues:
- Hugging Face Inference Deployment:
- The UI displays:
"This model is not part of our Model Catalog, and does not come with verified configuration. It might behave unexpectedly." - Additionally, we get a warning:
"Warning: deploying this model will probably fail because no handler.py file was found in the repository. Try selecting a different model or creating a custom handler." - As expected, the deployment does not proceed correctly.
- SageMaker Deployment:
- We tried deploying using the script provided in the Sesame repo.
- While the deployment starts, the endpoint never responds.
- Given the large size of the model, the nodes often time out or need to be manually destroyed.
Has anyone successfully deployed this model either via Hugging Face Inference Endpoints or on SageMaker?
Specific questions:
- Is there a sample handler.py or recommended configuration for this model?
- Are there any known hardware requirements (e.g. GPU spec, RAM) for this model to run stably?
- Has Hugging Face considered adding this model to the verified model catalog, or is that up to the authors?
Any guidance would be much appreciated!
Vishal
did you get any help on it
Facing the same problem. I think sesame's ML head has been poached by Meta hence the no response :)