id
stringlengths 5
118
| author
stringlengths 2
42
⌀ | lastModified
stringlengths 19
19
⌀ | downloads
float64 0
117M
⌀ | downloadsAllTime
float64 0
2.17B
⌀ | tags
stringlengths 2
28.2k
⌀ | pipeline_tag
stringclasses 52
values | createdAt
stringlengths 19
19
⌀ | dataset
stringlengths 1
5.92k
⌀ | license
stringclasses 104
values | architectures
stringlengths 2
125
⌀ | base_model
stringlengths 10
2.07k
⌀ | base_model_relation
stringclasses 11
values |
---|---|---|---|---|---|---|---|---|---|---|---|---|
nlpaueb/bert-base-uncased-contracts
|
nlpaueb
|
2022-04-28 14:43:56
| 35,114 | 677,936 |
['transformers', 'pytorch', 'tf', 'jax', 'bert', 'legal', 'fill-mask', 'en', 'endpoints_compatible']
|
fill-mask
|
2022-03-02 23:29:05
|
unknown
|
cc-by-sa-4.0
|
unknown
| null |
unknown
|
nlpaueb/bert-base-uncased-echr
|
nlpaueb
|
2022-04-28 14:44:26
| 164 | 13,577 |
['transformers', 'pytorch', 'tf', 'jax', 'bert', 'legal', 'fill-mask', 'en', 'endpoints_compatible']
|
fill-mask
|
2022-03-02 23:29:05
|
unknown
|
cc-by-sa-4.0
|
unknown
| null |
unknown
|
nlpaueb/bert-base-uncased-eurlex
|
nlpaueb
|
2022-04-28 14:44:15
| 360 | 160,645 |
['transformers', 'pytorch', 'tf', 'jax', 'bert', 'legal', 'fill-mask', 'en', 'endpoints_compatible']
|
fill-mask
|
2022-03-02 23:29:05
|
unknown
|
cc-by-sa-4.0
|
unknown
| null |
unknown
|
nlpaueb/legal-bert-small-uncased
|
nlpaueb
|
2022-04-28 14:43:32
| 27,390 | 7,938,613 |
['transformers', 'pytorch', 'tf', 'jax', 'bert', 'legal', 'fill-mask', 'en', 'endpoints_compatible']
|
fill-mask
|
2022-03-02 23:29:05
|
unknown
|
cc-by-sa-4.0
|
unknown
| null |
unknown
|
nlpaueb/sec-bert-base
|
nlpaueb
|
2022-04-28 14:46:31
| 757 | 86,109 |
['transformers', 'pytorch', 'tf', 'bert', 'pretraining', 'finance', 'financial', 'fill-mask', 'en', 'endpoints_compatible']
|
fill-mask
|
2022-03-02 23:29:05
|
unknown
|
cc-by-sa-4.0
|
BertForPreTraining
| null |
unknown
|
nlpaueb/sec-bert-num
|
nlpaueb
|
2022-04-28 14:46:16
| 171 | 2,249 |
['transformers', 'pytorch', 'tf', 'bert', 'pretraining', 'finance', 'financial', 'fill-mask', 'en', 'endpoints_compatible']
|
fill-mask
|
2022-03-02 23:29:05
|
unknown
|
cc-by-sa-4.0
|
BertForPreTraining
| null |
unknown
|
nlpaueb/sec-bert-shape
|
nlpaueb
|
2022-04-28 14:46:51
| 207 | 5,601 |
['transformers', 'pytorch', 'tf', 'bert', 'pretraining', 'finance', 'financial', 'fill-mask', 'en', 'endpoints_compatible']
|
fill-mask
|
2022-03-02 23:29:05
|
unknown
|
cc-by-sa-4.0
|
BertForPreTraining
| null |
unknown
|
nlpconnect/vit-gpt2-image-captioning
|
nlpconnect
|
2023-02-27 15:00:09
| 1,714,464 | 47,602,343 |
['transformers', 'pytorch', 'vision-encoder-decoder', 'image-text-to-text', 'image-to-text', 'image-captioning', 'endpoints_compatible']
|
image-to-text
|
2022-03-02 23:29:05
|
unknown
|
apache-2.0
|
VisionEncoderDecoderModel
| null |
unknown
|
novakat/nerkor-cars-onpp-hubert
|
novakat
|
2023-09-14 19:28:41
| 4,909 | 260,657 |
['transformers', 'pytorch', 'safetensors', 'bert', 'token-classification', 'hu', 'autotrain_compatible', 'endpoints_compatible']
|
token-classification
|
2022-03-02 23:29:05
|
unknown
|
gpl
|
BertForTokenClassification
| null |
unknown
|
nreimers/MiniLM-L6-H384-uncased
|
nreimers
|
2021-08-30 20:05:29
| 1,325 | 523,806 |
['transformers', 'pytorch', 'jax', 'bert', 'feature-extraction', 'endpoints_compatible']
|
feature-extraction
|
2022-03-02 23:29:05
|
unknown
|
mit
|
BertModel
|
['microsoft/MiniLM-L12-H384-uncased']
|
unknown_annotated
|
nreimers/mMiniLMv2-L6-H384-distilled-from-XLMR-Large
|
nreimers
|
2021-06-20 19:03:02
| 437 | 155,052 |
['transformers', 'pytorch', 'xlm-roberta', 'fill-mask', 'autotrain_compatible', 'endpoints_compatible']
|
fill-mask
|
2022-03-02 23:29:05
|
unknown
|
unknown
|
XLMRobertaForMaskedLM
| null |
unknown
|
nsi319/legal-led-base-16384
|
nsi319
|
2021-03-01 12:33:48
| 723 | 34,858 |
['transformers', 'pytorch', 'led', 'text2text-generation', 'summarization', 'en', 'autotrain_compatible']
|
summarization
|
2022-03-02 23:29:05
|
unknown
|
mit
|
LEDForConditionalGeneration
| null |
unknown
|
ntu-spml/distilhubert
|
ntu-spml
|
2023-07-24 18:30:45
| 5,498 | 7,155,028 |
['transformers', 'pytorch', 'safetensors', 'hubert', 'feature-extraction', 'speech', 'en', 'endpoints_compatible']
|
feature-extraction
|
2022-03-02 23:29:05
|
librispeech_asr
|
apache-2.0
|
HubertModel
| null |
source
|
nvidia/mit-b0
|
nvidia
|
2023-11-15 07:49:03
| 84,346 | 2,181,371 |
['transformers', 'pytorch', 'tf', 'segformer', 'image-classification', 'vision', 'autotrain_compatible', 'endpoints_compatible']
|
image-classification
|
2022-03-02 23:29:05
|
imagenet_1k
|
other
|
SegformerForImageClassification
| null |
unknown
|
nvidia/segformer-b0-finetuned-cityscapes-1024-1024
|
nvidia
|
2022-08-08 13:43:30
| 6,285 | 69,414 |
['transformers', 'pytorch', 'tf', 'segformer', 'vision', 'image-segmentation', 'endpoints_compatible']
|
image-segmentation
|
2022-03-02 23:29:05
|
cityscapes
|
other
|
SegformerForSemanticSegmentation
| null |
unknown
|
nvidia/segformer-b1-finetuned-cityscapes-1024-1024
|
nvidia
|
2022-08-09 11:33:04
| 10,922 | 230,456 |
['transformers', 'pytorch', 'tf', 'segformer', 'vision', 'image-segmentation', 'endpoints_compatible']
|
image-segmentation
|
2022-03-02 23:29:05
|
cityscapes
|
other
|
SegformerForSemanticSegmentation
| null |
unknown
|
obi/deid_roberta_i2b2
|
obi
|
2025-02-23 03:36:26
| 588,074 | 12,084,051 |
['transformers', 'pytorch', 'safetensors', 'roberta', 'token-classification', 'deidentification', 'medical notes', 'ehr', 'phi', 'en', 'autotrain_compatible', 'endpoints_compatible']
|
token-classification
|
2022-03-02 23:29:05
|
I2B2
|
mit
|
RobertaForTokenClassification
| null |
unknown
|
oliverguhr/fullstop-punctuation-multilang-large
|
oliverguhr
|
2023-11-16 09:35:35
| 317,211 | 10,032,982 |
['transformers', 'pytorch', 'tf', 'onnx', 'safetensors', 'xlm-roberta', 'token-classification', 'punctuation prediction', 'punctuation', 'en', 'de', 'fr', 'it', 'multilingual', 'autotrain_compatible', 'endpoints_compatible']
|
token-classification
|
2022-03-02 23:29:05
|
wmt/europarl
|
mit
|
XLMRobertaForTokenClassification
| null |
unknown
|
openai/clip-vit-base-patch16
|
openai
|
2022-10-04 09:42:28
| 4,594,375 | 187,773,162 |
['transformers', 'pytorch', 'jax', 'clip', 'zero-shot-image-classification', 'vision', 'endpoints_compatible']
|
zero-shot-image-classification
|
2022-03-02 23:29:05
|
unknown
|
unknown
|
CLIPModel
| null |
source
|
opensource/extract_names
|
opensource
|
2021-01-19 04:59:04
| 86 | 4,178 |
['transformers', 'tf', 'xlm-roberta', 'token-classification', 'Extract Names', 'multilingual', 'autotrain_compatible', 'endpoints_compatible']
|
token-classification
|
2022-03-02 23:29:05
|
unknown
|
apache-2.0
|
XLMRobertaForTokenClassification
| null |
unknown
|
ozcangundes/mt5-small-turkish-summarization
|
ozcangundes
|
2021-09-22 09:31:27
| 862 | 16,007 |
['transformers', 'pytorch', 'jax', 'mt5', 'text2text-generation', 'summarization', 'tr', 'autotrain_compatible', 'endpoints_compatible']
|
summarization
|
2022-03-02 23:29:05
|
MLSUM
|
mit
|
MT5ForConditionalGeneration
| null |
unknown
|
patrickvonplaten/bert2bert_cnn_daily_mail
|
patrickvonplaten
|
2022-06-25 17:06:49
| 1,239 | 39,262 |
['transformers', 'pytorch', 'encoder-decoder', 'text2text-generation', 'summarization', 'en', 'model-index', 'autotrain_compatible', 'endpoints_compatible']
|
summarization
|
2022-03-02 23:29:05
|
cnn_dailymail
|
apache-2.0
|
EncoderDecoderModel
| null |
unknown
|
pdelobelle/robbert-v2-dutch-base
|
pdelobelle
|
2023-12-04 15:14:12
| 16,267 | 1,692,365 |
['transformers', 'pytorch', 'tf', 'jax', 'safetensors', 'roberta', 'fill-mask', 'Dutch', 'Flemish', 'RoBERTa', 'RobBERT', 'BERT', 'nl', 'autotrain_compatible', 'endpoints_compatible']
|
fill-mask
|
2022-03-02 23:29:05
|
oscar_dbrd_lassy-ud_europarl-mono_conll2002
|
mit
|
RobertaForMaskedLM
| null |
unknown
|
peterhsu/marian-finetuned-kde4-en-to-zh_TW
|
peterhsu
|
2022-02-28 11:26:43
| 347 | 1,023 |
['transformers', 'pytorch', 'tensorboard', 'marian', 'text2text-generation', 'translation', 'generated_from_trainer', 'model-index', 'autotrain_compatible', 'endpoints_compatible']
|
translation
|
2022-03-02 23:29:05
|
kde4
|
apache-2.0
|
MarianMTModel
| null |
unknown
|
philschmid/bart-large-cnn-samsum
|
philschmid
|
2022-12-23 19:48:57
| 197,506 | 20,759,846 |
['transformers', 'pytorch', 'bart', 'text2text-generation', 'sagemaker', 'summarization', 'en', 'model-index', 'autotrain_compatible', 'endpoints_compatible']
|
summarization
|
2022-03-02 23:29:05
|
samsum
|
mit
|
BartForConditionalGeneration
| null |
unknown
|
phueb/BabyBERTa-1
|
phueb
|
2022-01-18 14:44:02
| 202 | 14,805 |
['transformers', 'pytorch', 'roberta', 'fill-mask', 'BabyBERTa', 'en', 'autotrain_compatible', 'endpoints_compatible']
|
fill-mask
|
2022-03-02 23:29:05
|
CHILDES
|
unknown
|
RobertaForMaskedLM
| null |
unknown
|
pierreguillou/bert-base-cased-squad-v1.1-portuguese
|
pierreguillou
|
2022-01-04 09:57:53
| 4,463 | 224,328 |
['transformers', 'pytorch', 'tf', 'jax', 'bert', 'question-answering', 'bert-base', 'pt', 'endpoints_compatible']
|
question-answering
|
2022-03-02 23:29:05
|
brWaC_squad_squad_v1_pt
|
mit
|
BertForQuestionAnswering
| null |
unknown
|
pierreguillou/bert-large-cased-squad-v1.1-portuguese
|
pierreguillou
|
2022-01-04 09:57:00
| 775 | 156,566 |
['transformers', 'pytorch', 'tf', 'bert', 'question-answering', 'bert-large', 'pt', 'endpoints_compatible']
|
question-answering
|
2022-03-02 23:29:05
|
brWaC_squad_squad_v1_pt
|
mit
|
BertForQuestionAnswering
| null |
unknown
|
pierreguillou/gpt2-small-portuguese
|
pierreguillou
|
2021-05-23 10:59:56
| 7,787 | 174,123 |
['transformers', 'pytorch', 'tf', 'jax', 'gpt2', 'text-generation', 'pt', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible']
|
text-generation
|
2022-03-02 23:29:05
|
wikipedia
|
mit
|
GPT2LMHeadModel
| null |
unknown
|
pierreguillou/ner-bert-base-cased-pt-lenerbr
|
pierreguillou
|
2021-12-29 19:32:39
| 25,728 | 1,479,840 |
['transformers', 'pytorch', 'bert', 'token-classification', 'generated_from_trainer', 'pt', 'model-index', 'autotrain_compatible', 'endpoints_compatible']
|
token-classification
|
2022-03-02 23:29:05
|
lener_br
|
unknown
|
BertForTokenClassification
| null |
unknown
|
ponteineptique/latin-classical-small
|
ponteineptique
|
2020-04-24 16:05:14
| 142 | 660 |
['transformers', 'pytorch', 'xlm', 'feature-extraction', 'endpoints_compatible']
|
feature-extraction
|
2022-03-02 23:29:05
|
unknown
|
unknown
|
XLMModel
| null |
unknown
|
prajjwal1/bert-medium
|
prajjwal1
|
2021-10-27 18:30:16
| 92,533 | 1,417,851 |
['transformers', 'pytorch', 'BERT', 'MNLI', 'NLI', 'transformer', 'pre-training', 'en', 'endpoints_compatible']
|
unknown
|
2022-03-02 23:29:05
|
unknown
|
mit
|
unknown
| null |
unknown
|
princeton-nlp/sup-simcse-bert-base-uncased
|
princeton-nlp
|
2021-05-20 02:54:31
| 468,423 | 7,086,093 |
['transformers', 'pytorch', 'jax', 'bert', 'feature-extraction', 'endpoints_compatible']
|
feature-extraction
|
2022-03-02 23:29:05
|
unknown
|
unknown
|
BertModel
| null |
unknown
|
princeton-nlp/sup-simcse-roberta-large
|
princeton-nlp
|
2022-11-11 20:04:02
| 88,758 | 24,470,221 |
['transformers', 'pytorch', 'jax', 'roberta', 'feature-extraction', 'endpoints_compatible']
|
feature-extraction
|
2022-03-02 23:29:05
|
unknown
|
unknown
|
RobertaModel
| null |
unknown
|
prithivida/parrot_paraphraser_on_T5
|
prithivida
|
2021-05-18 07:53:27
| 944,963 | 28,181,828 |
['transformers', 'pytorch', 't5', 'text2text-generation', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible']
|
text2text-generation
|
2022-03-02 23:29:05
|
unknown
|
unknown
|
T5ForConditionalGeneration
| null |
unknown
|
pszemraj/led-base-book-summary
|
pszemraj
|
2023-11-28 19:11:49
| 3,605 | 233,478 |
['transformers', 'pytorch', 'safetensors', 'led', 'text2text-generation', 'summarization', 'summary', 'longformer', 'booksum', 'long-document', 'long-form', 'model-index', 'autotrain_compatible', 'endpoints_compatible']
|
summarization
|
2022-03-02 23:29:05
|
kmfoda/booksum
|
apache-2.0_bsd-3-clause
|
LEDForConditionalGeneration
| null |
unknown
|
pszemraj/led-large-book-summary
|
pszemraj
|
2023-11-28 19:16:42
| 4,355 | 395,612 |
['transformers', 'pytorch', 'safetensors', 'led', 'text2text-generation', 'summarization', 'summary', 'longformer', 'booksum', 'long-document', 'long-form', 'en', 'model-index', 'autotrain_compatible', 'endpoints_compatible']
|
summarization
|
2022-03-02 23:29:05
|
kmfoda/booksum
|
apache-2.0_bsd-3-clause
|
LEDForConditionalGeneration
| null |
unknown
|
pucpr/biobertpt-all
|
pucpr
|
2022-11-27 16:54:34
| 2,564 | 245,407 |
['transformers', 'pytorch', 'tf', 'jax', 'bert', 'fill-mask', 'pt', 'autotrain_compatible', 'endpoints_compatible']
|
fill-mask
|
2022-03-02 23:29:05
|
unknown
|
unknown
|
BertForMaskedLM
| null |
unknown
|
pucpr/biobertpt-clin
|
pucpr
|
2022-11-27 16:55:14
| 120 | 15,061 |
['transformers', 'pytorch', 'tf', 'jax', 'bert', 'fill-mask', 'pt', 'autotrain_compatible', 'endpoints_compatible']
|
fill-mask
|
2022-03-02 23:29:05
|
unknown
|
unknown
|
BertForMaskedLM
| null |
unknown
|
pucpr/clinicalnerpt-medical
|
pucpr
|
2021-10-13 09:28:28
| 162 | 11,281 |
['transformers', 'pytorch', 'bert', 'token-classification', 'pt', 'autotrain_compatible', 'endpoints_compatible']
|
token-classification
|
2022-03-02 23:29:05
|
SemClinBr
|
unknown
|
BertForTokenClassification
| null |
unknown
|
pucpr/gpt2-bio-pt
|
pucpr
|
2022-11-23 12:33:37
| 160 | 8,691 |
['transformers', 'pytorch', 'tf', 'jax', 'gpt2', 'text-generation', 'pt', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible']
|
text-generation
|
2022-03-02 23:29:05
|
unknown
|
unknown
|
GPT2LMHeadModel
| null |
unknown
|
pyannote/embedding
|
pyannote
|
2024-05-10 19:36:51
| 377,351 | 15,124,561 |
['pyannote-audio', 'pytorch', 'tensorboard', 'pyannote', 'pyannote-audio-model', 'audio', 'voice', 'speech', 'speaker', 'speaker-recognition', 'speaker-verification', 'speaker-identification', 'speaker-embedding']
|
unknown
|
2022-03-02 23:29:05
|
voxceleb
|
mit
|
unknown
| null |
unknown
|
pysentimiento/robertuito-base-uncased
|
pysentimiento
|
2023-05-22 14:06:59
| 454 | 35,428 |
['transformers', 'pytorch', 'safetensors', 'roberta', 'fill-mask', 'twitter', 'masked-lm', 'es', 'autotrain_compatible', 'endpoints_compatible']
|
fill-mask
|
2022-03-02 23:29:05
|
unknown
|
unknown
|
RobertaForMaskedLM
| null |
unknown
|
pysentimiento/robertuito-hate-speech
|
pysentimiento
|
2023-02-20 19:04:44
| 5,693 | 886,958 |
['pysentimiento', 'pytorch', 'roberta', 'twitter', 'hate-speech', 'es']
|
unknown
|
2022-03-02 23:29:05
|
unknown
|
unknown
|
RobertaForSequenceClassification
| null |
unknown
|
qanastek/pos-french
|
qanastek
|
2024-04-09 15:42:08
| 179 | 95,198 |
['flair', 'pytorch', 'token-classification', 'sequence-tagger-model', 'fr']
|
token-classification
|
2022-03-02 23:29:05
|
qanastek/ANTILLES
|
unknown
|
unknown
| null |
unknown
|
raynardj/wenyanwen-ancient-translate-to-modern
|
raynardj
|
2022-01-08 04:22:30
| 514 | 5,293 |
['transformers', 'pytorch', 'encoder-decoder', 'text2text-generation', 'translation', '古文', '文言文', 'ancient', 'classical', 'zh', 'autotrain_compatible', 'endpoints_compatible']
|
translation
|
2022-03-02 23:29:05
|
unknown
|
unknown
|
EncoderDecoderModel
| null |
unknown
|
raynardj/wenyanwen-chinese-translate-to-ancient
|
raynardj
|
2021-11-29 14:42:25
| 2,473 | 17,082 |
['transformers', 'pytorch', 'encoder-decoder', 'text2text-generation', 'translation', '文言文', 'ancient', 'zh', 'autotrain_compatible', 'endpoints_compatible']
|
translation
|
2022-03-02 23:29:05
|
unknown
|
apache-2.0
|
EncoderDecoderModel
| null |
unknown
|
recobo/agri-sentence-transformer
|
recobo
|
2022-10-15 21:04:01
| 179 | 1,684 |
['sentence-transformers', 'pytorch', 'bert', 'feature-extraction', 'sentence-similarity', 'transformers', 'en', 'autotrain_compatible', 'text-embeddings-inference', 'endpoints_compatible']
|
sentence-similarity
|
2022-03-02 23:29:05
|
unknown
|
unknown
|
BertModel
| null |
unknown
|
rinna/japanese-gpt-1b
|
rinna
|
2024-07-20 07:52:31
| 2,272 | 732,621 |
['transformers', 'pytorch', 'safetensors', 'gpt2', 'text-generation', 'gpt', 'lm', 'nlp', 'ja', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible']
|
text-generation
|
2022-03-02 23:29:05
|
cc100_wikipedia_c4
|
mit
|
GPT2LMHeadModel
| null |
unknown
|
rinna/japanese-gpt2-medium
|
rinna
|
2024-07-20 07:50:47
| 7,762 | 602,756 |
['transformers', 'pytorch', 'tf', 'jax', 'safetensors', 'gpt2', 'text-generation', 'lm', 'nlp', 'ja', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible']
|
text-generation
|
2022-03-02 23:29:05
|
cc100_wikipedia
|
mit
|
GPT2LMHeadModel
| null |
unknown
|
saattrupdan/nbailab-base-ner-scandi
|
saattrupdan
|
2024-12-16 16:21:04
| 73,213 | 440,778 |
['transformers', 'pytorch', 'safetensors', 'bert', 'token-classification', 'da', 'no', 'nb', 'nn', 'sv', 'fo', 'is', 'autotrain_compatible', 'endpoints_compatible']
|
token-classification
|
2022-03-02 23:29:05
|
dane_norne_wikiann_suc3.0
|
mit
|
BertForTokenClassification
| null |
unknown
|
Saibo-creator/legal-roberta-base
|
Saibo-creator
|
2021-08-31 15:36:35
| 243 | 120,022 |
['transformers', 'pytorch', 'tf', 'jax', 'roberta', 'fill-mask', 'legal', 'en', 'autotrain_compatible', 'endpoints_compatible']
|
fill-mask
|
2022-03-02 23:29:05
|
unknown
|
apache-2.0
|
RobertaForMaskedLM
| null |
unknown
|
sakares/wav2vec2-large-xlsr-thai-demo
|
sakares
|
2021-03-22 07:15:18
| 952 | 23,628 |
['transformers', 'pytorch', 'wav2vec2', 'automatic-speech-recognition', 'audio', 'speech', 'xlsr-fine-tuning-week', 'th', 'model-index', 'endpoints_compatible']
|
automatic-speech-recognition
|
2022-03-02 23:29:05
|
common_voice
|
apache-2.0
|
Wav2Vec2ForCTC
| null |
unknown
|
Ashishkr/query_wellformedness_score
|
Ashishkr
|
2024-03-30 11:51:12
| 979 | 73,426,480 |
['transformers', 'pytorch', 'jax', 'safetensors', 'roberta', 'text-classification', 'autotrain_compatible']
|
text-classification
|
2022-03-02 23:29:05
|
google_wellformed_query
|
apache-2.0
|
RobertaForSequenceClassification
| null |
unknown
|
savasy/bert-base-turkish-ner-cased
|
savasy
|
2024-02-01 09:21:04
| 98,434 | 284,627 |
['transformers', 'pytorch', 'jax', 'safetensors', 'bert', 'token-classification', 'tr', 'autotrain_compatible', 'endpoints_compatible']
|
token-classification
|
2022-03-02 23:29:05
|
unknown
|
unknown
|
BertForTokenClassification
| null |
unknown
|
savasy/bert-base-turkish-sentiment-cased
|
savasy
|
2024-02-01 09:18:53
| 6,640 | 1,128,334 |
['transformers', 'pytorch', 'jax', 'safetensors', 'bert', 'text-classification', 'tr', 'autotrain_compatible', 'endpoints_compatible']
|
text-classification
|
2022-03-02 23:29:05
|
unknown
|
unknown
|
BertForSequenceClassification
| null |
unknown
|
savasy/bert-base-turkish-squad
|
savasy
|
2024-02-01 09:17:38
| 1,318 | 117,651 |
['transformers', 'pytorch', 'jax', 'safetensors', 'bert', 'question-answering', 'tr', 'endpoints_compatible']
|
question-answering
|
2022-03-02 23:29:05
|
unknown
|
unknown
|
BertForQuestionAnswering
| null |
unknown
|
ai-forever/Real-ESRGAN
|
ai-forever
|
2022-09-25 13:17:44
| 0 | 0 |
['PyTorch', 'ru', 'en']
|
unknown
|
2022-03-02 23:29:05
|
unknown
|
unknown
|
unknown
| null |
unknown
|
ai-forever/ruT5-base
|
ai-forever
|
2023-12-11 17:45:55
| 5,939 | 70,625 |
['transformers', 'pytorch', 't5', 'text2text-generation', 'PyTorch', 'Transformers', 'ru', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible']
|
text2text-generation
|
2022-03-02 23:29:05
|
unknown
|
unknown
|
T5ForConditionalGeneration
| null |
unknown
|
ai-forever/ruT5-large
|
ai-forever
|
2023-12-28 09:01:22
| 3,025 | 31,382 |
['transformers', 'pytorch', 't5', 'text2text-generation', 'PyTorch', 'Transformers', 'ru', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible']
|
text2text-generation
|
2022-03-02 23:29:05
|
unknown
|
unknown
|
T5ForConditionalGeneration
| null |
unknown
|
ai-forever/rugpt3large_based_on_gpt2
|
ai-forever
|
2023-12-04 14:43:51
| 16,893 | 406,205 |
['transformers', 'pytorch', 'jax', 'gpt2', 'text-generation', 'PyTorch', 'Transformers', 'ru', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible']
|
text-generation
|
2022-03-02 23:29:05
|
unknown
|
unknown
|
GPT2LMHeadModel
| null |
unknown
|
ai-forever/rugpt3medium_based_on_gpt2
|
ai-forever
|
2023-12-05 08:43:34
| 5,798 | 233,529 |
['transformers', 'pytorch', 'gpt2', 'text-generation', 'PyTorch', 'Transformers', 'ru', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible']
|
text-generation
|
2022-03-02 23:29:05
|
unknown
|
unknown
|
GPT2LMHeadModel
| null |
unknown
|
ai-forever/rugpt3small_based_on_gpt2
|
ai-forever
|
2023-12-05 09:19:39
| 29,624 | 346,849 |
['transformers', 'pytorch', 'jax', 'gpt2', 'text-generation', 'PyTorch', 'Transformers', 'ru', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible']
|
text-generation
|
2022-03-02 23:29:05
|
unknown
|
unknown
|
GPT2LMHeadModel
| null |
unknown
|
ai-forever/sbert_large_mt_nlu_ru
|
ai-forever
|
2024-06-13 07:29:25
| 3,309 | 402,631 |
['transformers', 'safetensors', 'bert', 'feature-extraction', 'PyTorch', 'Transformers', 'ru', 'text-embeddings-inference', 'endpoints_compatible']
|
feature-extraction
|
2022-03-02 23:29:05
|
unknown
|
unknown
|
BertModel
| null |
unknown
|
ai-forever/sbert_large_nlu_ru
|
ai-forever
|
2024-10-07 11:09:02
| 870,047 | 12,808,600 |
['transformers', 'pytorch', 'safetensors', 'bert', 'feature-extraction', 'PyTorch', 'Transformers', 'ru', 'text-embeddings-inference', 'endpoints_compatible']
|
feature-extraction
|
2022-03-02 23:29:05
|
unknown
|
unknown
|
BertModel
| null |
unknown
|
sdadas/polish-roberta-large-v2
|
sdadas
|
2024-04-23 15:32:52
| 1,250 | 57,057 |
['transformers', 'pytorch', 'safetensors', 'roberta', 'fill-mask', 'pl', 'autotrain_compatible', 'endpoints_compatible']
|
fill-mask
|
2022-03-02 23:29:05
|
unknown
|
lgpl-3.0
|
RobertaForMaskedLM
| null |
unknown
|
seidel/plsum-base-ptt5
|
seidel
|
2022-04-28 16:59:49
| 133 | 1,345 |
['transformers', 'pytorch', 't5', 'text2text-generation', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible']
|
text2text-generation
|
2022-03-02 23:29:05
|
unknown
|
unknown
|
T5ForConditionalGeneration
| null |
unknown
|
sentence-transformers/LaBSE
|
sentence-transformers
|
2025-03-06 13:40:25
| 588,072 | 9,223,941 |
['sentence-transformers', 'pytorch', 'tf', 'jax', 'onnx', 'safetensors', 'bert', 'feature-extraction', 'sentence-similarity', 'multilingual', 'af', 'sq', 'am', 'ar', 'hy', 'as', 'az', 'eu', 'be', 'bn', 'bs', 'bg', 'my', 'ca', 'ceb', 'zh', 'co', 'hr', 'cs', 'da', 'nl', 'en', 'eo', 'et', 'fi', 'fr', 'fy', 'gl', 'ka', 'de', 'el', 'gu', 'ht', 'ha', 'haw', 'he', 'hi', 'hmn', 'hu', 'is', 'ig', 'id', 'ga', 'it', 'ja', 'jv', 'kn', 'kk', 'km', 'rw', 'ko', 'ku', 'ky', 'lo', 'la', 'lv', 'lt', 'lb', 'mk', 'mg', 'ms', 'ml', 'mt', 'mi', 'mr', 'mn', 'ne', 'no', 'ny', 'or', 'fa', 'pl', 'pt', 'pa', 'ro', 'ru', 'sm', 'gd', 'sr', 'st', 'sn', 'si', 'sk', 'sl', 'so', 'es', 'su', 'sw', 'sv', 'tl', 'tg', 'ta', 'tt', 'te', 'th', 'bo', 'tr', 'tk', 'ug', 'uk', 'ur', 'uz', 'vi', 'cy', 'wo', 'xh', 'yi', 'yo', 'zu', 'autotrain_compatible', 'text-embeddings-inference', 'endpoints_compatible']
|
sentence-similarity
|
2022-03-02 23:29:05
|
unknown
|
apache-2.0
|
BertModel
| null |
unknown
|
sentence-transformers/all-mpnet-base-v1
|
sentence-transformers
|
2025-03-06 13:19:48
| 14,790 | 1,238,396 |
['sentence-transformers', 'pytorch', 'onnx', 'safetensors', 'openvino', 'mpnet', 'fill-mask', 'feature-extraction', 'sentence-similarity', 'transformers', 'en', 'autotrain_compatible', 'endpoints_compatible']
|
sentence-similarity
|
2022-03-02 23:29:05
|
unknown
|
apache-2.0
|
MPNetForMaskedLM
| null |
unknown
|
sentence-transformers/average_word_embeddings_glove.6B.300d
|
sentence-transformers
|
2025-03-06 13:40:43
| 0 | 0 |
['sentence-transformers', 'feature-extraction', 'sentence-similarity', 'autotrain_compatible', 'endpoints_compatible']
|
sentence-similarity
|
2022-03-02 23:29:05
|
unknown
|
apache-2.0
|
unknown
| null |
unknown
|
sentence-transformers/clip-ViT-B-32-multilingual-v1
|
sentence-transformers
|
2024-11-05 16:39:09
| 123,936 | 3,135,782 |
['sentence-transformers', 'pytorch', 'tf', 'onnx', 'safetensors', 'openvino', 'distilbert', 'feature-extraction', 'sentence-similarity', 'multilingual', 'autotrain_compatible', 'text-embeddings-inference', 'endpoints_compatible']
|
sentence-similarity
|
2022-03-02 23:29:05
|
unknown
|
apache-2.0
|
DistilBertModel
| null |
unknown
|
sentence-transformers/clip-ViT-B-32
|
sentence-transformers
|
2024-02-12 10:14:11
| 0 | 0 |
['sentence-transformers', 'feature-extraction', 'sentence-similarity', 'autotrain_compatible', 'endpoints_compatible']
|
sentence-similarity
|
2022-03-02 23:29:05
|
unknown
|
unknown
|
unknown
| null |
unknown
|
sentence-transformers/gtr-t5-base
|
sentence-transformers
|
2025-03-06 13:39:57
| 158,312 | 1,045,731 |
['sentence-transformers', 'pytorch', 'safetensors', 't5', 'feature-extraction', 'sentence-similarity', 'en', 'autotrain_compatible', 'endpoints_compatible']
|
sentence-similarity
|
2022-03-02 23:29:05
|
unknown
|
apache-2.0
|
T5EncoderModel
| null |
unknown
|
sentence-transformers/gtr-t5-xxl
|
sentence-transformers
|
2025-03-06 13:38:08
| 979 | 35,509 |
['sentence-transformers', 'pytorch', 'safetensors', 't5', 'feature-extraction', 'sentence-similarity', 'en', 'autotrain_compatible', 'endpoints_compatible']
|
sentence-similarity
|
2022-03-02 23:29:05
|
unknown
|
apache-2.0
|
T5EncoderModel
| null |
unknown
|
sentence-transformers/msmarco-MiniLM-L6-v3
|
sentence-transformers
|
2025-03-06 13:17:32
| 23,725 | 2,036,563 |
['sentence-transformers', 'pytorch', 'tf', 'jax', 'onnx', 'safetensors', 'openvino', 'bert', 'feature-extraction', 'sentence-similarity', 'transformers', 'autotrain_compatible', 'text-embeddings-inference', 'endpoints_compatible']
|
sentence-similarity
|
2022-03-02 23:29:05
|
unknown
|
apache-2.0
|
BertModel
| null |
unknown
|
sentence-transformers/multi-qa-MiniLM-L6-cos-v1
|
sentence-transformers
|
2024-11-05 17:17:16
| 17,035,956 | 79,722,672 |
['sentence-transformers', 'pytorch', 'tf', 'onnx', 'safetensors', 'openvino', 'bert', 'feature-extraction', 'sentence-similarity', 'transformers', 'en', 'autotrain_compatible', 'text-embeddings-inference', 'endpoints_compatible']
|
sentence-similarity
|
2022-03-02 23:29:05
|
flax-sentence-embeddings/stackexchange_xml_ms_marco_gooaq_yahoo_answers_topics_search_qa_eli5_natural_questions_trivia_qa_embedding-data/QQP_embedding-data/PAQ_pairs_embedding-data/Amazon-QA_embedding-data/WikiAnswers
|
unknown
|
BertModel
|
['nreimers/MiniLM-L6-H384-uncased']
|
unknown_annotated
|
sentence-transformers/multi-qa-distilbert-cos-v1
|
sentence-transformers
|
2024-11-05 17:18:43
| 107,566 | 3,781,387 |
['sentence-transformers', 'pytorch', 'onnx', 'safetensors', 'openvino', 'distilbert', 'fill-mask', 'feature-extraction', 'sentence-similarity', 'transformers', 'en', 'autotrain_compatible', 'text-embeddings-inference', 'endpoints_compatible']
|
sentence-similarity
|
2022-03-02 23:29:05
|
flax-sentence-embeddings/stackexchange_xml_ms_marco_gooaq_yahoo_answers_topics_search_qa_eli5_natural_questions_trivia_qa_embedding-data/QQP_embedding-data/PAQ_pairs_embedding-data/Amazon-QA_embedding-data/WikiAnswers
|
unknown
|
DistilBertForMaskedLM
| null |
unknown
|
sentence-transformers/multi-qa-mpnet-base-cos-v1
|
sentence-transformers
|
2024-11-05 17:21:14
| 192,406 | 4,242,743 |
['sentence-transformers', 'pytorch', 'onnx', 'safetensors', 'openvino', 'mpnet', 'fill-mask', 'feature-extraction', 'sentence-similarity', 'transformers', 'en', 'autotrain_compatible', 'endpoints_compatible']
|
sentence-similarity
|
2022-03-02 23:29:05
|
unknown
|
unknown
|
MPNetForMaskedLM
| null |
unknown
|
sentence-transformers/multi-qa-mpnet-base-dot-v1
|
sentence-transformers
|
2024-11-05 15:15:42
| 1,693,788 | 28,435,893 |
['sentence-transformers', 'pytorch', 'onnx', 'safetensors', 'openvino', 'mpnet', 'fill-mask', 'feature-extraction', 'sentence-similarity', 'transformers', 'en', 'autotrain_compatible', 'endpoints_compatible']
|
sentence-similarity
|
2022-03-02 23:29:05
|
flax-sentence-embeddings/stackexchange_xml_ms_marco_gooaq_yahoo_answers_topics_search_qa_eli5_natural_questions_trivia_qa_embedding-data/QQP_embedding-data/PAQ_pairs_embedding-data/Amazon-QA_embedding-data/WikiAnswers
|
unknown
|
MPNetForMaskedLM
| null |
unknown
|
sentence-transformers/paraphrase-MiniLM-L6-v2
|
sentence-transformers
|
2025-03-06 13:26:35
| 7,458,416 | 92,938,663 |
['sentence-transformers', 'pytorch', 'tf', 'onnx', 'safetensors', 'openvino', 'bert', 'feature-extraction', 'sentence-similarity', 'transformers', 'autotrain_compatible', 'text-embeddings-inference', 'endpoints_compatible']
|
sentence-similarity
|
2022-03-02 23:29:05
|
unknown
|
apache-2.0
|
BertModel
| null |
unknown
|
sentence-transformers/sentence-t5-base
|
sentence-transformers
|
2025-03-06 13:39:02
| 485,213 | 3,160,724 |
['sentence-transformers', 'pytorch', 'rust', 'safetensors', 't5', 'feature-extraction', 'sentence-similarity', 'en', 'autotrain_compatible', 'endpoints_compatible']
|
sentence-similarity
|
2022-03-02 23:29:05
|
unknown
|
apache-2.0
|
T5EncoderModel
| null |
unknown
|
sentence-transformers/sentence-t5-large
|
sentence-transformers
|
2025-03-06 13:37:56
| 19,555 | 1,009,256 |
['sentence-transformers', 'pytorch', 'onnx', 'safetensors', 't5', 'feature-extraction', 'sentence-similarity', 'en', 'autotrain_compatible', 'endpoints_compatible']
|
sentence-similarity
|
2022-03-02 23:29:05
|
unknown
|
apache-2.0
|
T5EncoderModel
| null |
unknown
|
sentence-transformers/stsb-roberta-base-v2
|
sentence-transformers
|
2025-03-06 13:36:30
| 14,415 | 691,785 |
['sentence-transformers', 'pytorch', 'tf', 'jax', 'onnx', 'safetensors', 'openvino', 'roberta', 'feature-extraction', 'sentence-similarity', 'transformers', 'autotrain_compatible', 'text-embeddings-inference', 'endpoints_compatible']
|
sentence-similarity
|
2022-03-02 23:29:05
|
unknown
|
apache-2.0
|
RobertaModel
| null |
unknown
|
seokho/gpt2-emotion
|
seokho
|
2021-07-06 06:07:33
| 172 | 2,144 |
['transformers', 'pytorch', 'gpt2', 'text-generation', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible']
|
text-generation
|
2022-03-02 23:29:05
|
unknown
|
unknown
|
GPT2LMHeadModel
| null |
unknown
|
serdarakyol/interpress-turkish-news-classification
|
serdarakyol
|
2022-03-10 16:06:33
| 217 | 3,262 |
['transformers', 'pytorch', 'tf', 'jax', 'bert', 'text-classification', 'tr', 'autotrain_compatible', 'endpoints_compatible']
|
text-classification
|
2022-03-02 23:29:05
|
unknown
|
unknown
|
BertForSequenceClassification
| null |
unknown
|
seyonec/ChemBERTa-zinc-base-v1
|
seyonec
|
2021-05-20 20:55:33
| 255,206 | 3,686,376 |
['transformers', 'pytorch', 'jax', 'roberta', 'fill-mask', 'chemistry', 'autotrain_compatible', 'endpoints_compatible']
|
fill-mask
|
2022-03-02 23:29:05
|
unknown
|
unknown
|
RobertaForMaskedLM
| null |
unknown
|
seyonec/ChemBERTa_zinc250k_v2_40k
|
seyonec
|
2021-05-20 20:57:42
| 772 | 183,895 |
['transformers', 'pytorch', 'jax', 'roberta', 'fill-mask', 'autotrain_compatible', 'endpoints_compatible']
|
fill-mask
|
2022-03-02 23:29:05
|
unknown
|
unknown
|
RobertaForMaskedLM
| null |
unknown
|
shahp7575/gpt2-horoscopes
|
shahp7575
|
2021-08-24 02:34:10
| 643 | 28,330 |
['transformers', 'pytorch', 'gpt2', 'text-generation', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible']
|
text-generation
|
2022-03-02 23:29:05
|
unknown
|
unknown
|
GPT2LMHeadModel
| null |
unknown
|
shahrukhx01/bert-mini-finetune-question-detection
|
shahrukhx01
|
2023-03-29 22:00:48
| 7,686 | 1,227,014 |
['transformers', 'pytorch', 'safetensors', 'bert', 'text-classification', 'neural-search-query-classification', 'neural-search', 'en', 'autotrain_compatible', 'endpoints_compatible']
|
text-classification
|
2022-03-02 23:29:05
|
unknown
|
unknown
|
BertForSequenceClassification
| null |
unknown
|
shahrukhx01/question-vs-statement-classifier
|
shahrukhx01
|
2023-03-29 22:01:12
| 65,309 | 1,976,779 |
['transformers', 'pytorch', 'safetensors', 'bert', 'text-classification', 'neural-search-query-classification', 'neural-search', 'en', 'autotrain_compatible', 'endpoints_compatible']
|
text-classification
|
2022-03-02 23:29:05
|
unknown
|
unknown
|
BertForSequenceClassification
| null |
unknown
|
shibing624/code-autocomplete-distilgpt2-python
|
shibing624
|
2024-02-19 09:34:30
| 346 | 376,473 |
['transformers', 'pytorch', 'safetensors', 'gpt2', 'text-generation', 'code', 'autocomplete', 'en', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible']
|
text-generation
|
2022-03-02 23:29:05
|
unknown
|
apache-2.0
|
GPT2LMHeadModel
| null |
unknown
|
shibing624/macbert4csc-base-chinese
|
shibing624
|
2024-09-27 06:55:18
| 32,677 | 1,507,210 |
['transformers', 'pytorch', 'onnx', 'safetensors', 'bert', 'fill-mask', 'zh', 'pycorrector', 'text2text-generation', 'autotrain_compatible', 'endpoints_compatible']
|
text2text-generation
|
2022-03-02 23:29:05
|
shibing624/CSC
|
apache-2.0
|
BertForMaskedLM
| null |
unknown
|
siebert/sentiment-roberta-large-english
|
siebert
|
2024-06-11 16:40:11
| 178,937 | 6,704,022 |
['transformers', 'pytorch', 'tf', 'jax', 'roberta', 'text-classification', 'sentiment', 'twitter', 'reviews', 'siebert', 'en', 'autotrain_compatible', 'endpoints_compatible']
|
text-classification
|
2022-03-02 23:29:05
|
unknown
|
unknown
|
RobertaForSequenceClassification
| null |
unknown
|
skt/ko-gpt-trinity-1.2B-v0.5
|
skt
|
2021-09-23 16:29:25
| 1,841 | 169,114 |
['transformers', 'pytorch', 'gpt2', 'text-generation', 'gpt3', 'ko', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible']
|
text-generation
|
2022-03-02 23:29:05
|
unknown
|
cc-by-nc-sa-4.0
|
GPT2LMHeadModel
| null |
unknown
|
skt/kobert-base-v1
|
skt
|
2021-07-01 07:16:05
| 80,878 | 5,202,484 |
['transformers', 'pytorch', 'bert', 'feature-extraction', 'endpoints_compatible']
|
feature-extraction
|
2022-03-02 23:29:05
|
unknown
|
unknown
|
BertModel
| null |
unknown
|
skt/kogpt2-base-v2
|
skt
|
2021-09-23 16:29:28
| 27,956 | 2,219,083 |
['transformers', 'pytorch', 'jax', 'gpt2', 'text-generation', 'ko', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible']
|
text-generation
|
2022-03-02 23:29:05
|
unknown
|
cc-by-nc-sa-4.0
|
GPT2LMHeadModel
| null |
unknown
|
snunlp/KR-FinBert-SC
|
snunlp
|
2022-04-28 05:07:18
| 16,011 | 970,724 |
['transformers', 'pytorch', 'bert', 'text-classification', 'ko', 'autotrain_compatible', 'endpoints_compatible']
|
text-classification
|
2022-03-02 23:29:05
|
unknown
|
unknown
|
BertForSequenceClassification
| null |
unknown
|
sonoisa/sentence-bert-base-ja-mean-tokens-v2
|
sonoisa
|
2024-04-17 11:39:38
| 253,911 | 8,266,890 |
['sentence-transformers', 'pytorch', 'safetensors', 'bert', 'sentence-bert', 'feature-extraction', 'sentence-similarity', 'ja', 'autotrain_compatible', 'endpoints_compatible']
|
feature-extraction
|
2022-03-02 23:29:05
|
unknown
|
cc-by-sa-4.0
|
BertModel
| null |
unknown
|
sonoisa/t5-base-japanese
|
sonoisa
|
2024-12-12 13:36:45
| 7,665 | 615,096 |
['transformers', 'pytorch', 'jax', 'safetensors', 't5', 'feature-extraction', 'text2text-generation', 'seq2seq', 'ja', 'text-generation-inference', 'endpoints_compatible']
|
text2text-generation
|
2022-03-02 23:29:05
|
wikipedia_oscar_cc100
|
cc-by-sa-4.0
|
T5Model
| null |
unknown
|
spacy/en_core_web_lg
|
spacy
|
2023-11-21 08:13:57
| 265 | 17,215 |
['spacy', 'token-classification', 'en', 'model-index']
|
token-classification
|
2022-03-02 23:29:05
|
unknown
|
mit
|
unknown
| null |
unknown
|
Subsets and Splits
FLUX.1-dev Adapter Models
The query performs basic filtering to retrieve specific entries related to a particular base model, which provides limited analytical value.
FLUX.1-dev Adapter Models
This query retrieves a limited number of entries where the base model is from a specific vendor and relation is 'adapter', which provides basic filtering but limited analytical value.