All Models
sentence-transformers/all-MiniLM-L6-v2
sentence-transformers
This model is designed for sentence similarity and feature extraction tasks. It supports various datasets including s2orc, flax-sentence-embeddings/stackexchange_xml, ms_marco, gooaq, yahoo_answers_topics, code_search_net, search_qa, eli5, snli, multi_nli, wikihow, natural_questions, trivia_qa, and several embedding data sources such as sentence-compression, flickr30k-captions, altlex, simple-wiki.
Falconsai/nsfw_image_detection
Falconsai
Image classification model licensed under Apache 2.0.
google/electra-base-discriminator
Electra is a model designed for natural language understanding tasks. It is trained to distinguish between real and fake text, making it useful for various applications in NLP.
dima806/fairface_age_image_detection
dima806
Detects age group with about 59% accuracy based on an image. License: Apache 2.0. Metrics include accuracy and F1 score. Base model used is Google ViT Base Patch 16 224 in 21k. The model is part of the image classification pipeline and utilizes the dataset nateraw/fairface.
google-bert/bert-base-uncased
google-bert
Bert is a language model trained on the BookCorpus and Wikipedia datasets. It is licensed under Apache 2.0 and is designed for various natural language processing tasks.
openai/clip-vit-base-patch32
openai
This model is designed for vision tasks. It includes candidate labels such as playing music and playing sports. An example title is Cat & Dog. You can find a sample image at https://huggingface.co/datasets/mishig/sample_images/resolve/main/cat-dog-music.png.
timm/mobilenetv3_small_100.lamb_in1k
timm
Image classification model using the timm library. It is licensed under Apache 2.0 and trained on the ImageNet-1K dataset.
FacebookAI/roberta-large
FacebookAI
Roberta Large is a language model trained on datasets including BookCorpus and Wikipedia. It is licensed under MIT and is designed for various natural language processing tasks.
sentence-transformers/all-mpnet-base-v2
sentence-transformers
This model is designed for sentence similarity and feature extraction using sentence-transformers. It supports various tasks such as text embeddings inference and is compatible with datasets including s2orc, flax-sentence-embeddings/stackexchange_xml, ms_marco, gooaq, yahoo_answers_topics, code_search_net, search_qa, eli5, snli, multi_nli, wikihow, natural_questions, trivia_qa, embedding-data/sentence-compression, embedding-data/flickr30k-captions, and embedding-data/altlex.
pyannote/segmentation-3.0
pyannote
Segmentation is a model for speaker diarization, speaker change detection, speaker segmentation, voice activity detection, and overlapped speech detection. It is part of the pyannote audio framework, which aims to improve the understanding of audio and voice data. The model is open-source under the MIT license, and the collected information will help acquire a better knowledge of the pyannote.audio user base and assist its maintainers in further improvements.
FacebookAI/roberta-base
FacebookAI
Roberta Base is a language model trained on datasets including BookCorpus and Wikipedia. It is licensed under MIT.
Bingsu/adetailer
Bingsu
Adetailer is a model that utilizes the Apache 2.0 license and is built on the Ultralytics library. It is designed to work with datasets such as WIDER FACE and skytnt/anime-segmentation, and it is tagged with PyTorch.
distilbert/distilbert-base-uncased
distilbert
DistilBERT is a smaller, faster, cheaper, and lighter version of BERT. It is designed for natural language processing tasks and is trained on the BookCorpus and Wikipedia datasets. The model is licensed under Apache 2.0 and is tagged with exbert.
sentence-transformers/paraphrase-multilingual-MiniLM-L12-v2
sentence-transformers
This model is designed for multilingual sentence similarity tasks, supporting a wide range of languages including Arabic, Bulgarian, Catalan, Czech, Danish, German, Greek, English, Spanish, Estonian, Persian, Finnish, French, Galician, Gujarati, Hebrew, Hindi, Croatian, Hungarian, Armenian, Indonesian, Italian, Japanese, Georgian, Korean, Kurdish, Lithuanian, Latvian, Macedonian, Mongolian, Marathi, Malay, Burmese, Norwegian Bokmål, Dutch, Polish, Portuguese, Romanian, Russian, Slovak, Slovenian, Albanian, Serbian, Swedish, Thai, Turkish, Ukrainian, Urdu, and Vietnamese. It is licensed under Apache 2.0 and is part of the sentence-transformers library, suitable for feature extraction and sentence similarity tasks.
facebook/esm2_t33_650M_UR50D
The model is licensed under MIT and is designed for text processing tasks. It includes a sample input sequence: MQIFVKTLTGKTITLEVEPS<mask>TIENVKAKIQDKEGIPPDQQRLIFAGKQLEDGRTLSDYNIQKESTLHLVLRLRGG.
pyannote/wespeaker-voxceleb-resnet34-LM
pyannote
Wespeaker is a model designed for audio processing, specifically for voice and speech tasks such as speaker recognition, speaker verification, speaker identification, and speaker embedding. It utilizes the Voxceleb dataset and is licensed under CC BY 4.0. The model is part of the pyannote audio framework.
pyannote/speaker-diarization-3.1
pyannote
Speaker diarization model for audio processing, capable of speaker change detection, voice activity detection, overlapped speech detection, and automatic speech recognition. Tags include pyannote, pyannote-audio, and pyannote-audio-pipeline. The model is licensed under MIT.
openai-community/gpt2
openai-community
Gpt2 is a language model designed for natural language processing tasks.
openai/clip-vit-large-patch14
openai
This model is designed for vision tasks. It includes candidate labels such as playing music and playing sports. An example title is Cat & Dog. You can find a sample image at the following source: https://huggingface.co/datasets/mishig/sample_images/resolve/main/cat-dog-music.png.
omni-research/Tarsier2-Recap-7b
omni-research
A video LLM with an Apache 2.0 license.
kyutai/moshiko-pytorch-bf16
kyutai
Moshiko is a PyTorch model designed for efficient training and inference using BF16 precision.
FacebookAI/xlm-roberta-base
FacebookAI
XLM Roberta Base is a multilingual model supporting a wide range of languages including Afrikaans, Amharic, Arabic, Assamese, Azerbaijani, Belarusian, Bulgarian, Bengali, Brazilian Portuguese, Bosnian, Catalan, Czech, Welsh, Danish, German, Greek, English, Esperanto, Spanish, Estonian, Basque, Persian, Finnish, French, Frisian, Irish, Scottish Gaelic, Galician, Gujarati, Hausa, Hebrew, Hindi, Croatian, Hungarian, Armenian, Indonesian, Icelandic, Italian, Japanese, Javanese, Georgian, Kazakh, Khmer, Kannada, Korean, Kurdish, Kyrgyz, Latin, Lao, Lithuanian, Latvian, Malagasy, Macedonian, Malayalam, Mongolian, Marathi, Malay, Nepali, Dutch, Norwegian, Oromo, Oriya, Punjabi, Polish, Pashto, Portuguese, Romanian, Russian, Arabic, Sindhi, Sinhala, Slovak, Slovenian, Somali, Albanian, Serbian, Sundanese, Swedish, Swahili, Tamil, Telugu, Thai, Tagalog, Turkish, Uighur, Ukrainian, Urdu, Uzbek, Vietnamese, and Xhosa.
Qwen/Qwen2.5-7B-Instruct
Qwen
This model is designed for text generation. It is licensed under the Apache 2.0 license, and you can find the license details at https://huggingface.co/Qwen/Qwen2.5-7B-Instruct/blob/main/LICENSE. The model supports the English language and is tagged for chat applications, utilizing the transformers library.
facebook/contriever
Trained without supervision following the approach described in Towards Unsupervised Dense Information Retrieval with Contrastive Learning. The associated GitHub repository is available here https://github.com/facebookresearch/contriever.
Qwen/Qwen2.5-VL-3B-Instruct
Qwen
This model is designed for image-text-to-text tasks. It is licensed under the qwen-research license, which can be found at https://huggingface.co/Qwen/Qwen2.5-VL-3B-Instruct/blob/main/LICENSE. The model supports the English language and is part of the multimodal category, utilizing the transformers library.
Qwen/Qwen3-0.6B
Qwen
Text generation model based on Qwen/Qwen3-0.6B. It is licensed under the Apache 2.0 license, and the license can be found at https://huggingface.co/Qwen/Qwen3-0.6B/blob/main/LICENSE. The base model is Qwen/Qwen3-0.6B-Base.
BAAI/bge-m3
BAAI
Pipeline tag: sentence-similarity. Tags: sentence-transformers, feature-extraction, sentence-similarity. License: MIT.
colbert-ir/colbertv2.0
colbert-ir
ColBERT is a model designed for efficient retrieval in information retrieval tasks. It is licensed under MIT and supports the English language. The model is associated with the ColBERT project.
Isotonic/distilbert_finetuned_ai4privacy_v2
Isotonic
This model is a token classification model based on DistilBERT, fine-tuned for privacy-related tasks. It is licensed under CC BY-NC 4.0 and is built on the distilbert-base-uncased architecture. The model was generated from a trainer and has been evaluated using the seqeval metric. It is trained on datasets including ai4privacy/pii-masking-200k and Isotonic/pii-masking-200k.
Gensyn/Qwen2.5-0.5B-Instruct
Gensyn
This model is designed for text generation. It is licensed under the Apache 2.0 license, and you can find the license details at https://huggingface.co/Gensyn/Qwen2.5-0.5B-Instruct/blob/main/LICENSE. The model supports English language and is tagged with chat, rl-swarm, and gensyn. It utilizes the transformers library.
Comfy-Org/Wan_2.2_ComfyUI_Repackaged
Comfy-Org
This model is designed for diffusion in a single file format, optimized for use with ComfyUI.
autogluon/chronos-bolt-small
autogluon
Time series forecasting model licensed under Apache 2.0. It is a pretrained foundation model designed for time series analysis and forecasting.
distilbert/distilbert-base-uncased-finetuned-sst-2-english
distilbert
This model is designed for text classification tasks, specifically fine-tuned on the SST-2 dataset from the GLUE benchmark. It achieves an accuracy of 0.9106 on the validation split of the SST-2 dataset. The model is licensed under Apache 2.0.
BAAI/bge-small-en-v1.5
BAAI
The model is designed for tasks such as sentence transformation, feature extraction, and sentence similarity using transformers. It has been evaluated on the MTEB Amazon Counterfactual Classification dataset with the following results: accuracy of 73.79, average precision of 37.22, and F1 score of 68.09.
Kijai/WanVideo_comfy
Kijai
Combined and quantized models for WanVideo, originating from here. Tags include diffusion single file and comfyui base model with versions Wan-AI/Wan2.1-VACE-14B and Wan-AI/Wan2.1-VACE-1.3B.
google-bert/bert-base-multilingual-cased
google-bert
BERT is a multilingual model that supports a wide range of languages including Afrikaans, Albanian, Arabic, Armenian, Asturian, Azerbaijani, Basque, Bavarian, Belarusian, Bengali, Bihari, Bosnian, Breton, Bulgarian, Burmese, Catalan, Cebuano, Chechen, Chinese, Chuvash, Croatian, Czech, Danish, Dutch, English, Estonian, Finnish, French, Galician, Georgian, German, Greek, Gujarati, Haitian Creole, Hebrew, Hindi, Hungarian, Icelandic, Interlingua, Indonesian, Irish, Italian, Japanese, Javanese, Kannada, Kazakh, Kyrgyz, Korean, Latin, Latvian, Lithuanian, Lombard, Macedonian, Malagasy, Malay, Malayalam, Marathi, Mongolian, Minangkabau, Nepali, Newar, Norwegian Bokmål, Norwegian Nynorsk, Occitan, Persian, Piedmontese, Polish, Portuguese, Punjabi, Romanian, Russian, Scots, Serbian, Serbian (Cyrillic), Slovak, Slovenian, Spanish, Sundanese, Swahili, Swedish, Tagalog, Tajik, Thai, Tamil, Tatar, Telugu.
jinaai/jina-embeddings-v3
jinaai
Feature extraction model for sentence similarity, supporting multilingual capabilities across various languages including Afrikaans, Amharic, Arabic, Assamese, Azerbaijani, Belarusian, Bulgarian, Bengali, Brazilian Portuguese, Bosnian, Catalan, Czech, Welsh, Danish, German, Greek, English, Esperanto, Spanish, Estonian, Basque, Persian, Finnish, French, Frisian, Irish, Scottish Gaelic, Galician, Gujarati, Hausa, Hebrew, Hindi, Croatian, Hungarian, Armenian, Indonesian, Icelandic, Italian, Japanese, Javanese, Georgian, Kazakh, Khmer, Kannada, Korean, Kurdish, Kyrgyz, Latin, Lao, Lithuanian, Latvian, Malagasy, Macedonian, Malayalam, Mongolian, Marathi, Malay, Burmese, Nepali, Dutch, Norwegian, Oromo, Oriya, Punjabi, Polish, Pashto, Portuguese, Romanian, Russian, Arabic, Sindhi, Sinhala, Slovak, Slovenian.
facebook/wav2vec2-base-960h
Automatic Speech Recognition model trained on the LibriSpeech dataset. It supports English language audio processing and is part of the Hugging Face ASR leaderboard. The model is licensed under Apache 2.0 and includes example audio samples for demonstration.
nlpaueb/legal-bert-base-uncased
nlpaueb
This model is designed for fill-mask tasks in the legal domain. It is trained on English language data and is licensed under CC BY-SA 4.0. The model can be used to predict masked words in legal texts, such as in the example: "The applicant submitted that her husband was subjected to treatment amounting to [MASK] whilst in the custody of police." A thumbnail image is available at https://i.ibb.co/p3kQ7Rw/Screenshot-2020-10-06-at-12-16-36-PM.png.
jonatasgrosman/wav2vec2-large-xlsr-53-russian
jonatasgrosman
Automatic Speech Recognition model for the Russian language. It is licensed under Apache 2.0 and trained on datasets including Common Voice. The model is evaluated using metrics such as Word Error Rate (WER) and Character Error Rate (CER). It is tagged with audio, automatic-speech-recognition, and robust-speech-event, among others.
coqui/XTTS-v2
coqui
Text-to-speech model with a coqui public model license. License link: https://coqui.ai/cpml. Library name: coqui. Pipeline tag: text-to-speech. Example text: Once when I was six years old I saw a magnificent picture.
sentence-transformers/paraphrase-multilingual-mpnet-base-v2
sentence-transformers
This model is designed for multilingual sentence similarity and feature extraction. It supports a wide range of languages including Arabic, Bulgarian, Catalan, Czech, Danish, German, Greek, English, Spanish, Estonian, Persian, Finnish, French, Galician, Gujarati, Hebrew, Hindi, Croatian, Hungarian, Armenian, Indonesian, Italian, Japanese, Georgian, Korean, Kurdish, Lithuanian, Latvian, Macedonian, Mongolian, Marathi, Malay, Burmese, Norwegian Bokmål, Dutch, Polish, Portuguese, Romanian, Russian, Slovak, Slovenian, Albanian, Serbian, Swedish, Thai, Turkish, Ukrainian, Urdu, and Vietnamese. The model is licensed under Apache 2.0 and is part of the sentence-transformers library.
cross-encoder/ms-marco-MiniLM-L6-v2
cross-encoder
This model is designed for text ranking tasks. It is licensed under Apache 2.0 and is based on the cross-encoder/ms-marco-MiniLM-L12-v2 model. It utilizes the sentence-transformers library and is trained on the ms-marco dataset, specifically for the English language.
meta-llama/Llama-3.1-8B-Instruct
meta-llama
Llama 3.1 is a text generation model supporting multiple languages including English, German, French, Italian, Portuguese, Hindi, Spanish, and Thai. It is based on the Meta-Llama 3.1 8B architecture and is licensed under the Llama 3.1 Community License Agreement. The release date is July 23, 2024. The model is tagged with facebook, meta, pytorch, llama, and llama-3.
sentence-transformers/all-MiniLM-L12-v2
sentence-transformers
This model is designed for sentence similarity and feature extraction using transformers. It supports various datasets including s2orc, flax-sentence-embeddings/stackexchange_xml, ms_marco, gooaq, yahoo_answers_topics, code_search_net, search_qa, eli5, snli, multi_nli, wikihow, natural_questions, trivia_qa, and several embedding-data sources.
BAAI/bge-large-en-v1.5
BAAI
The model is designed for tasks such as sentence transformation, feature extraction, and sentence similarity using transformers. It is part of the MTEB model index, specifically for the Amazon Counterfactual Classification dataset in English. The model's configuration is set to English with a test split. The revision identifier is e8379541af4e31359cca9fbcf4b00f2671dba205. The performance metrics include an accuracy of 75.85%, an average precision of 38.57, and an F1 score of 69.69.
Comfy-Org/Wan_2.1_ComfyUI_repackaged
Comfy-Org
A model designed for diffusion in a single file format, optimized for use with ComfyUI.
dphn/dolphin-2.9.1-yi-1.5-34b
dphn
This model is based on the Yi 1.5 34B architecture and is licensed under Apache 2.0. It was generated from a trainer and utilizes various datasets including cognitivecomputations/Dolphin-2.9, teknium/OpenHermes-2.5, m-a-p/CodeFeedback-Filtered-Instruction, cognitivecomputations/dolphin-coder, cognitivecomputations/samantha-data, microsoft/orca-math-word-problems-200k, Locutusque/function-calling-chatml, and internlm/Agent-FLAN.
Qwen/Qwen2.5-VL-7B-Instruct
Qwen
This model is designed for image-text-to-text tasks. It supports the English language and is built using the transformers library under the Apache 2.0 license.
cardiffnlp/twitter-roberta-base-sentiment-latest
cardiffnlp
This model is designed for sentiment analysis on tweets. It is trained on the TweetEval dataset and is available under the CC BY 4.0 license.
google-bert/bert-base-multilingual-uncased
google-bert
BERT is a multilingual model that supports a wide range of languages including Afrikaans, Albanian, Arabic, Armenian, Asturian, Azerbaijani, Basque, Belarusian, Bengali, Bihari, Bosnian, Breton, Bulgarian, Burmese, Catalan, Cebuano, Chechen, Chinese, Chuvash, Croatian, Czech, Danish, Dutch, English, Estonian, Finnish, French, Galician, Georgian, German, Greek, Gujarati, Haitian Creole, Hebrew, Hindi, Hungarian, Icelandic, Interlingua, Indonesian, Irish, Italian, Japanese, Javanese, Kannada, Kazakh, Kyrgyz, Korean, Latin, Latvian, Lithuanian, Luxembourgish, Macedonian, Malagasy, Malay, Malayalam, Marathi, Minangkabau, Nepali, Newar, Norwegian Bokmål, Norwegian Nynorsk, Occitan, Persian, Piedmontese, Polish, Portuguese, Punjabi, Romanian, Russian, Scots, Serbian, Slovak, Slovenian, Spanish, Sundanese, Swahili, Swedish, Tagalog, Tajik, Tamil, Tatar, Telugu, Turkish, Ukrainian.
openai/gpt-oss-20b
openai
Text generation model licensed under Apache 2.0, utilizing the transformers library and tagged with vllm.
sentence-transformers/paraphrase-MiniLM-L6-v2
sentence-transformers
This model is designed for sentence similarity tasks and feature extraction using the sentence-transformers library. It is licensed under Apache 2.0.
BAAI/bge-base-en-v1.5
BAAI
The model is designed for sentence transformation, feature extraction, and sentence similarity tasks. It is part of the MTEB model index, specifically for the Amazon Counterfactual Classification dataset in English. The test split has a revision identifier of e8379541af4e31359cca9fbcf4b00f2671dba205. The model achieved an accuracy of 76.15%, an average precision of 39.32%, and an F1 score of 70.17%.
google/gemma-3-1b-it
Gemma is a text generation model that requires users to review and agree to Google’s usage license. To access Gemma on Hugging Face, ensure you’re logged in and click the provided button. Requests are processed immediately.
dandelin/vilt-b32-mlm
dandelin
This model is licensed under the Apache 2.0 license.
facebook/dinov2-base
Dino is a vision model licensed under Apache 2.0.
Qwen/Qwen3-Embedding-0.6B
Qwen
This model is based on Qwen/Qwen3-0.6B-Base and is designed for tasks such as sentence similarity, feature extraction, and text embeddings inference. It is licensed under Apache 2.0 and utilizes transformers and sentence-transformers.
hexgrad/Kokoro-82M
hexgrad
Kokoro is an open-weight TTS model with 82 million parameters. Despite its lightweight architecture, it delivers comparable quality to larger models while being significantly faster and more cost-efficient. With Apache-licensed weights, Kokoro can be deployed anywhere from production environments to personal projects.
Qwen/Qwen2.5-1.5B-Instruct
Qwen
This model is licensed under the Apache 2.0 license. It supports the English language and is designed for text generation tasks. The base model is Qwen/Qwen2.5-1.5B and it is compatible with the transformers library.
autogluon/chronos-bolt-base
autogluon
Time series forecasting model licensed under Apache 2.0. It is designed for time series analysis and includes pretrained models, foundation models, and time series foundation models.
MahmoudAshraf/mms-300m-1130-forced-aligner
MahmoudAshraf
The model supports a wide range of languages including but not limited to: Abkhazian, Afrikaans, Akan, Amharic, Arabic, Assamese, Avar, Aymara, Azerbaijani, Bashkir, Bambara, Belarusian, Bengali, Bislama, Tibetan, Bosnian, Bulgarian, Catalan, Czech, Chechen, Chuvash, Kurdish, Welsh, Danish, German, Divehi, Dzongkha, Greek, English, Esperanto, Estonian, Basque, Finnish, French, Western Frisian, Fula, Irish, Galician, Guarani, Gujarati, Chinese, Haitian Creole, Hausa, Hebrew, Hindi, Hungarian, Armenian, Igbo, Interlingua, Malay, Icelandic, Italian, Javanese, Japanese, Kannada, Georgian, Kazakh, Korean, Khmer, Kikuyu, Kinyarwanda, Kyrgyz, Korean, Komi, Lao, Latin, Latvian, Lingala, Lithuanian, Luxembourgish, Ganda, Marshallese, Malayalam, Marathi, Malay, Macedonian, Malagasy, Maltese, Mongolian, Maori, Burmese, Chinese, Dutch, Norwegian, Nepali, Nyanja, Occitan, Oromo, and others.
openai/whisper-large-v3
openai
This model supports multiple languages including English, Chinese, German, Spanish, Russian, Korean, French, Japanese, Portuguese, Turkish, Polish, Catalan, Dutch, Arabic, Swedish, Italian, Indonesian, Hindi, Finnish, Vietnamese, Hebrew, Ukrainian, Greek, Malay, Czech, Romanian, Danish, Hungarian, Tamil, Norwegian, Thai, Urdu, Croatian, Bulgarian, Lithuanian, Latin, Maori, Malayalam, Welsh, Slovak, Telugu, Persian, Latvian, Bengali, Serbian, Azerbaijani, Slovenian, Kannada, Estonian, Macedonian, Breton, Basque, Icelandic, Armenian, Nepali, Mongolian, Bosnian, Kazakh, Albanian, Swahili, Galician, Marathi, Punjabi, Sinhala, Khmer, Shona, Yoruba, Somali, Afrikaans, Occitan, Georgian, Tajik, Sudanese, Gujarati, Amharic, Yiddish, Lao, Uzbek, Faroese, Haitian Creole, Pashto, Tatar, Norwegian Nynorsk, Maltese, Arabic (Saudi), Luxembourgish, Malay (Malaysian), Tibetan, Tagalog, Malagasy, Assamese, Tatar, Hawaiian, Lingala, Hausa, Bashkir.
facebook/opt-125m
Text generation model designed for English language inference.
openai/whisper-large-v3-turbo
openai
This model supports multiple languages including English, Chinese, German, Spanish, Russian, Korean, French, Japanese, Portuguese, Turkish, Polish, Catalan, Dutch, Arabic, Swedish, Italian, Indonesian, Hindi, Finnish, Vietnamese, Hebrew, Ukrainian, Greek, Malay, Czech, Romanian, Danish, Hungarian, Tamil, Norwegian, Thai, Urdu, Croatian, Bulgarian, Lithuanian, Latin, Maori, Malayalam, Welsh, Slovak, Telugu, Persian, Latvian, Bengali, Serbian, Azerbaijani, Slovenian, Kannada, Estonian, Macedonian, Brazilian Portuguese, Basque, Icelandic, Armenian, Nepali, Mongolian, Bosnian, Kazakh, Albanian, Swahili, Galician, Marathi, Punjabi, Sinhala, Khmer, Shona, Yoruba, Somali, Afrikaans, Occitan, Georgian, Tajik, Sudanese, Gujarati, Amharic, Yiddish, Lao, Uzbek, Faroese, Haitian Creole, Pashto, Tatar, Nynorsk, Maltese, Arabic, and more.
TinyLlama/TinyLlama-1.1B-Chat-v1.0
TinyLlama
A chatbot that can help code. License: Apache-2.0. Datasets include Cerebras SlimPajama-627B, BigCode Starcoderdata, HuggingFaceH4 Ultrachat 200k, and HuggingFaceH4 Ultrafeedback Binarized. The example widget demonstrates a Python function to calculate the first 10 digits of the Fibonacci sequence and print it to the CLI.
trl-internal-testing/tiny-Qwen2ForCausalLM-2.5
trl-internal-testing
Tiny Qwen 2 For Causal LM is a model that utilizes the transformers library and is tagged with trl.
Qwen/Qwen3-4B-Instruct-2507
Qwen
This model is designed for text generation. It is part of the Qwen library and is licensed under the Apache 2.0 license. The license can be found at https://huggingface.co/Qwen/Qwen3-4B-Instruct-2507/blob/main/LICENSE.
hustvl/vitmatte-small-composition-1k
hustvl
A model designed for vision tasks, licensed under Apache 2.0.
jonatasgrosman/wav2vec2-large-xlsr-53-portuguese
jonatasgrosman
Automatic Speech Recognition model trained on Portuguese language using datasets such as Common Voice. It is licensed under Apache 2.0 and includes metrics like WER and CER. The model is tagged with audio, automatic-speech-recognition, and robust-speech-event, among others.
jonatasgrosman/wav2vec2-large-xlsr-53-japanese
jonatasgrosman
This model is designed for automatic speech recognition in Japanese, utilizing the Wav2Vec2 architecture. It is trained on the Common Voice dataset and evaluates performance using Word Error Rate (WER) and Character Error Rate (CER) metrics. The model is tagged for audio processing and fine-tuning in speech recognition tasks.
meta-llama/Llama-3.2-1B-Instruct
meta-llama
This model is designed for text generation and supports multiple languages including English, German, French, Italian, Portuguese, Hindi, Spanish, and Thai. It utilizes the transformers library and is built on PyTorch. The model is tagged with keywords such as facebook, meta, pytorch, llama, and llama-3. It is licensed under llama3.2 and includes an extra gated prompt.
sentence-transformers/multi-qa-mpnet-base-dot-v1
sentence-transformers
This model is designed for sentence similarity tasks, providing feature extraction and text embeddings inference. It supports various datasets including Stack Exchange, MS MARCO, and Yahoo Answers Topics, among others.
Alibaba-NLP/gte-large-en-v1.5
Alibaba-NLP
This model is designed for sentence similarity tasks and is based on the MTEB Amazon Counterfactual Classification dataset. It utilizes the transformers library and is tagged with sentence-transformers, gte, mteb, and transformers.js. The model is licensed under Apache 2.0 and supports the English language. The evaluation results show an accuracy of 73.01% and an average precision of 35.05% on the test split.
openai/gpt-oss-120b
openai
Text generation model licensed under Apache 2.0, utilizing the transformers library and tagged with vllm.
Qwen/Qwen2.5-3B-Instruct
Qwen
Text generation model with a focus on chat applications. It is licensed under an other license. The model is based on Qwen/Qwen2.5-3B and is compatible with the transformers library. The primary language supported is English.
bigscience/bloomz-560m
bigscience
BloomZ 560M is a text generation model trained on the bigscience/xP3 dataset. It supports multiple languages including ak, ar, as, bm, bn, ca, code, en, es, eu, fon, fr, gu, hi, id, ig, ki, kn, lg, ln, ml, mr, ne, nso, ny, or, pa, pt, rn, rw, sn, st, sw, ta, te, tn, ts, tum, tw, ur, vi, wo, xh, yo, zh, and zu. The model is compatible with various programming languages such as C, C++, C#, Go, Java, JavaScript, Lua, PHP, Python, Ruby, Rust, Scala, and TypeScript. It is licensed under bigscience-bloom-rail-1.0.
timm/convnextv2_nano.fcmae_ft_in22k_in1k
timm
A model designed for image classification, utilizing the timm library. It is trained on the ImageNet-1K dataset and is suitable for various image classification tasks.
stabilityai/sd-turbo
stabilityai
Pipeline tag: text-to-image inference: false.
facebook/esmfold_v1
Esmfold V1 is a model designed for protein structure prediction, leveraging advanced machine learning techniques to provide accurate and efficient results.
emilyalsentzer/Bio_ClinicalBERT
emilyalsentzer
Fill-mask model for English language processing with MIT license.
tech4humans/yolov8s-signature-detector
tech4humans
Object detection model for signature detection using the Ultralytics YOLOv8 framework. It is based on the AGPL-3.0 license and utilizes the tech4humans/signature-detection dataset. The model evaluates performance using metrics such as F1 score, precision, and recall. It is implemented in PyTorch with library version 8.0.239.
timm/resnet50.a1_in1k
timm
Image classification model licensed under Apache 2.0, utilizing the timm library and supporting transformers.
google-bert/bert-base-cased
google-bert
Bert is a language model trained on the BookCorpus and Wikipedia datasets. It is licensed under Apache 2.0 and is tagged with exbert.
BAAI/bge-reranker-v2-m3
BAAI
This model is designed for text classification and supports multilingual capabilities. It is licensed under Apache 2.0 and utilizes transformers, sentence-transformers, and text embeddings for inference.
Salesforce/moirai-2.0-R-small
Salesforce
This model is designed for time series forecasting. It is licensed under CC BY-NC 4.0 and is categorized as a pretrained foundation model for time series applications.
neulab/codebert-python
neulab
Trained for 1,000,000 steps with a batch size of 32 on Python code from the codeparrot/github-code-clean dataset, on the masked-language-modeling task.
facebook/bart-large-mnli
Zero-shot classification model trained on the MultiNLI dataset. It is licensed under MIT. For more information, visit the thumbnail at https://huggingface.co/front/thumbnails/facebook.png.
patrickjohncyh/fashion-clip
patrickjohncyh
A model designed for vision and language tasks in the fashion and ecommerce domain. It supports English and utilizes the transformers library. The model can classify candidate labels such as black shoe and red shoe, and includes an example of a black shoe. A widget is available for integration.
w11wo/indonesian-roberta-base-posp-tagger
w11wo
Token classification model for part-of-speech tagging using the Indonesian RoBERTa base architecture. It is based on the flax-community/indonesian-roberta-base model and has been trained on the indonlu dataset. The model evaluates performance using metrics such as precision, recall, F1 score, and accuracy. The results for the test split of the indonlu dataset show a precision of 0.9625 and a recall of 0.9625.
context-labs/meta-llama-Llama-3.2-3B-Instruct-FP16
context-labs
Language support includes English, German, French, Italian, Portuguese, Hindi, Spanish, and Thai. The model utilizes the transformers library and is designed for text generation. It is associated with tags such as facebook, meta, pytorch, llama, and llama-3. The license for this model is llama3.2 and it includes an extra gated prompt.
ggml-org/models-moved
ggml-org
Various models to be used in llama.cpp CI workflow.
google/vit-base-patch16-224-in21k
Vision model trained on the ImageNet-21k dataset. License: Apache-2.0. Inference capability: false.
openai/clip-vit-base-patch16
openai
A vision model designed for image classification tasks. It can categorize images into candidate labels such as playing music and playing sports. An example includes a cat and dog image.
zhihan1996/DNABERT-S
zhihan1996
DNABERT S is designed for biological and genomic applications, providing advanced capabilities for DNA sequence analysis.
openai/clip-vit-large-patch14-336
openai
This model is designed for image classification tasks using candidate labels such as playing music and playing sports. It includes a sample image of a cat and dog. The model index contains the name clip-vit-large-patch14-336 with no results listed.
facebook/bart-large-cnn
This model is designed for summarization tasks, specifically trained on the CNN/DailyMail dataset. It achieves high performance with metrics such as ROUGE-1 at 42.95, ROUGE-2 at 20.81, and ROUGE-L at 30.62. The model is licensed under MIT and is suitable for generating concise summaries from longer texts.
Qwen/Qwen3-VL-30B-A3B-Instruct
Qwen
This model is designed for image-text-to-text tasks and is licensed under the Apache 2.0 license.
mistralai/Mistral-7B-Instruct-v0.2
mistralai
This model is a fine-tuned version of Mistral designed for instruction-based tasks. It is built using the transformers library and is licensed under Apache 2.0. The model includes tags such as finetuned and mistral-common. The current version is 0.2, with a new version available as 0.3. Inference is not enabled. For more information on how personal data is processed, please refer to the Privacy Policy.
datasocietyco/bge-base-en-v1.5-course-recommender-v5
datasocietyco
This model is designed for sentence similarity and feature extraction using the sentence-transformers library. It was generated from a trainer with a dataset size of 45 and utilizes the Multiple Negatives Ranking Loss. The base model is BAAI/bge-base-en-v1.5 and is tagged for sentence-transformers.
OpenGVLab/InternVL3_5-241B-A28B-Instruct
OpenGVLab
Image-text-to-text model licensed under Apache 2.0. It is based on the pretrained model OpenGVLab/InternVL3_5-241B-Pretrained and is fine-tuned on datasets including OpenGVLab/MMPR-v1.2 and OpenGVLab/MMPR-Tiny. The model supports multilingual capabilities and is tagged with internvl and custom_code.
google/vit-base-patch16-224
This model is designed for vision tasks, specifically image classification. It is trained on datasets including ImageNet-1K and ImageNet-21K. Example images include a tiger, a teapot, and a palace.
sentence-transformers/distiluse-base-multilingual-cased-v2
sentence-transformers
This model is designed for sentence similarity tasks and supports multiple languages including Arabic, Bulgarian, Catalan, Czech, Danish, German, Greek, English, Spanish, Estonian, Persian, Finnish, French, Galician, Gujarati, Hebrew, Hindi, Croatian, Hungarian, Armenian, Indonesian, Italian, Japanese, Georgian, Korean, Kurdish, Lithuanian, Latvian, Macedonian, Mongolian, Marathi, Malay, Burmese, Norwegian Bokmål, Dutch, Polish, Portuguese, Romanian, Russian, Slovak, Slovenian, Albanian, Serbian, Swedish, Thai, Turkish, Ukrainian, Urdu, and Vietnamese. It is licensed under Apache 2.0 and is part of the sentence-transformers library, focusing on feature extraction and sentence similarity.
timm/resnet18.a1_in1k
timm
ResNet-18 model for image classification, pretrained on ImageNet-1k dataset.
laion/CLIP-ViT-H-14-laion2B-s32B-b79K
laion
This model is designed for zero-shot image classification. It supports candidate labels such as playing music and playing sports. The model is available under the MIT license and includes a widget for displaying sample images.
Qwen/Qwen3-8B
Qwen
Text generation model based on Qwen/Qwen3-8B-Base. Licensed under Apache 2.0. License details can be found at https://huggingface.co/Qwen/Qwen3-8B/blob/main/LICENSE.
trpakov/vit-face-expression
trpakov
This model is designed for recognizing and interpreting facial expressions.
usyd-community/vitpose-plus-base
usyd-community
A keypoint detection model designed for human pose estimation, built using the transformers library. It is licensed under Apache 2.0 and supports the English language.
mixedbread-ai/mxbai-embed-large-v1
mixedbread-ai
This model is designed for embedding tasks using the MTEB framework. It includes various classification tasks with datasets such as MTEB Amazon Counterfactual Classification and MTEB Amazon Polarity. The model has achieved an accuracy of 75.04%, an average precision of 37.74%, and an F1 score of 68.93%. The configurations are set for the English language and the test split.
google-t5/t5-small
google-t5
T5 Small is a multilingual model that supports English, French, Romanian, and German. It is designed for tasks such as summarization and translation, and is licensed under Apache 2.0. The model is trained on the C4 dataset.
cardiffnlp/twitter-xlm-roberta-base-sentiment
cardiffnlp
This model is designed for multilingual sentiment analysis on Twitter data. It can understand and analyze sentiments expressed in various languages, making it useful for applications that require insights from diverse linguistic sources.
microsoft/deberta-v3-base
microsoft
DeBERTa V3 Base is a language model designed for fill-mask tasks. It utilizes advanced techniques to improve performance on various natural language processing tasks.
ProsusAI/finbert
ProsusAI
Finbert is designed for financial sentiment analysis, providing insights into market trends and sentiments. It analyzes text data to determine the sentiment surrounding financial topics, such as stock movements and currency fluctuations.
microsoft/TRELLIS-image-large
microsoft
Trellis is a library designed for image-to-3D processing. It provides tools and pipelines to convert 2D images into 3D representations, enabling various applications in computer vision and graphics.
FacebookAI/xlm-roberta-large
FacebookAI
XLM-RoBERTa is a multilingual model designed for various language tasks. It supports a wide range of languages including Afrikaans, Amharic, Arabic, Assamese, Azerbaijani, Belarusian, Bulgarian, Bengali, Brazilian Portuguese, Bosnian, Catalan, Czech, Welsh, Danish, German, Greek, English, Esperanto, Spanish, Estonian, Basque, Persian, Finnish, French, Frisian, Irish, Scottish Gaelic, Galician, Gujarati, Hausa, Hebrew, Hindi, Croatian, Hungarian, Armenian, Indonesian, Icelandic, Italian, Japanese, Javanese, Georgian, Kazakh, Khmer, Kannada, Korean, Kurdish, Kyrgyz, Latin, Lao, Lithuanian, Latvian, Malagasy, Macedonian, Malayalam, Mongolian, Marathi, Malay, Nepali, Dutch, Norwegian, Oromo, Odia, Punjabi, Polish, Pashto, Portuguese, Romanian, Russian, Arabic, Sindhi, Sinhala, Slovak, Slovenian, Somali, Albanian, Serbian, Sundanese, Swedish, Swahili, Tamil, Telugu, Thai, Tagalog, Turkish, Uyghur, Ukrainian, Urdu, Uzbek, Vietnamese, and Xhosa.
openai/whisper-small
openai
Whisper Small is a multilingual speech recognition model that supports a wide range of languages including English, Chinese, German, Spanish, Russian, Korean, French, Japanese, Portuguese, Turkish, Polish, and many more. It is designed to transcribe audio into text accurately across various languages, making it a versatile tool for applications in transcription, translation, and accessibility.
stabilityai/stable-diffusion-xl-base-1.0
stabilityai
OpenRail license. Tags: text-to-image, stable-diffusion.
stable-diffusion-v1-5/stable-diffusion-v1-5
stable-diffusion-v1-5
Stable Diffusion is a text-to-image model that allows for image generation based on textual descriptions. It is licensed under CreativeML OpenRail and includes tags such as stable-diffusion, stable-diffusion-diffusers.
nlpconnect/vit-gpt2-image-captioning
nlpconnect
Vision Transformer with GPT-2 for automatic image captioning, generating natural language descriptions of images.
Salesforce/blip-image-captioning-base
Salesforce
Image captioning model that generates descriptive text for images. It supports English language and is licensed under the BSD-3-Clause.
jonatasgrosman/wav2vec2-large-xlsr-53-dutch
jonatasgrosman
Wav2Vec2 model for automatic speech recognition in Dutch, fine-tuned on Common Voice dataset.
google/siglip-so400m-patch14-384
This model is designed for vision tasks. It is licensed under Apache 2.0. The candidate labels include playing music and playing sports. An example title is Cat & Dog. You can find sample images at the following source: https://huggingface.co/datasets/mishig/sample_images/resolve/main/cat-dog-music.png.
inference-net/Schematron-3B
inference-net
3B parameter model based on Llama 3.2 3B Instruct.
distilbert/distilgpt2
distilbert
Language model for English with tags including exbert.
intfloat/multilingual-e5-large
intfloat
Multilingual E5 Large is a model designed for sentence similarity and feature extraction tasks. It is part of the MTEB benchmark, specifically for the Amazon Counterfactual Classification dataset in English. The model has been evaluated on the test split with a revision identifier e8379541af4e31359cca9fbcf4b00f2671dba205. The performance metrics include an accuracy of 79.06%, an average precision of 43.49%, and an F1 score of 73.33%.
deepseek-ai/DeepSeek-R1-Distill-Qwen-32B
deepseek-ai
DeepSeek R1 Distill Qwen 32B is a model designed for advanced AI tasks, utilizing the transformers library under the MIT license.
Qwen/Qwen2-VL-2B-Instruct
Qwen
This model is designed for image-text-to-text tasks. It supports the English language and is based on the Qwen/Qwen2-VL-2B architecture. It is licensed under Apache 2.0 and utilizes the transformers library.
papluca/xlm-roberta-base-language-detection
papluca
Detects multiple languages including Arabic, Bulgarian, German, Greek, English, Spanish, French, Hindi, Italian, Japanese, Dutch, Polish, Portuguese, Russian, Swahili, Thai, Turkish, Urdu, Vietnamese, and Chinese. The model is based on XLM-Roberta and is licensed under MIT. It has been trained on the papluca/language-identification dataset and evaluated using metrics such as accuracy and F1 score.
CIDAS/clipseg-rd64-refined
CIDAS
A model for image segmentation licensed under Apache 2.0.
opensearch-project/opensearch-neural-sparse-encoding-doc-v2-distill
opensearch-project
Neural sparse encoding model for OpenSearch, optimized for document retrieval and passage ranking using learned sparse representations.
CompVis/stable-diffusion-safety-checker
CompVis
CLIP-based safety checker for Stable Diffusion, designed to filter inappropriate content.
deepseek-ai/DeepSeek-OCR
deepseek-ai
DeepSeek OCR is a multilingual image-to-text model designed for optical character recognition. It supports various languages and is equipped with advanced vision-language capabilities. The model is built using the Transformers library and is available under the MIT license.
Qwen/Qwen2.5-VL-32B-Instruct
Qwen
This model is designed for image-text-to-text tasks, supporting the English language. It utilizes the transformers library and is tagged as multimodal.
jonatasgrosman/wav2vec2-large-xlsr-53-arabic
jonatasgrosman
Wav2Vec2 model for automatic speech recognition in Arabic, trained on Common Voice and Arabic Speech Corpus datasets.
E-MIMIC/inclusively-reformulation-it5
E-MIMIC
--- license: cc-by-nc-sa-4.0 ---
jonatasgrosman/wav2vec2-large-xlsr-53-chinese-zh-cn
jonatasgrosman
Wav2Vec2 model for automatic speech recognition in Chinese (Mandarin), trained on Common Voice dataset.
roadz/qwen3-1.7b-oxxroad
roadz
Qwen3-style causal language model with 1.7B parameters.
Qwen/Qwen2-VL-7B-Instruct
Qwen
This model is designed for image-text-to-text tasks. It supports the English language and is built on the Qwen/Qwen2-VL-7B architecture. The model is licensed under Apache 2.0 and is compatible with the transformers library.
nvidia/parakeet-tdt-0.6b-v2
nvidia
Automatic speech recognition model using the Transducer architecture. It is built with the NeMo library and supports English language. The model is trained on datasets such as nvidia/Granary and nvidia/nemo-asr-set-3.0. It features tags like FastConformer, Conformer, and is part of the hf-asr-leaderboard. Example audio samples include Librispeech sample 1 and Librispeech sample 2.
meta-llama/Llama-3.2-3B-Instruct
meta-llama
This model is designed for text generation and supports multiple languages including English, German, French, Italian, Portuguese, Hindi, Spanish, and Thai. It utilizes the transformers library and is built on the PyTorch framework. The model is tagged with keywords such as Facebook, Meta, and Llama, indicating its origin and functionality.
nvidia/parakeet-rnnt-0.6b
nvidia
600M parameter RNN-T speech recognition model trained on multiple English speech datasets including LibriSpeech, Fisher, Switchboard, and Common Voice.
depth-anything/Depth-Anything-V2-Base-hf
depth-anything
Depth Anything V2 Base model for monocular depth estimation.
microsoft/table-transformer-structure-recognition
microsoft
Transformer model for recognizing and analyzing table structures in documents.
vikhyatk/moondream2
vikhyatk
Image-text-to-text model licensed under Apache 2.0. The new version is Moondream 3 Preview.
petals-team/StableBeluga2
petals-team
--- datasets: - conceptofmind/cot_submix_original - conceptofmind/flan2021_submix_original - conceptofmind/t0_submix_original - conceptofmind/niv2_submix_original language: - en pipeline_tag: text-generation ---
pyannote/segmentation
pyannote
Segmentation is a model designed for audio processing tasks such as speaker segmentation, voice activity detection, and overlapped speech detection. It is part of the pyannote audio framework, which aims to enhance the understanding of audio data. The model is licensed under MIT and is intended for research purposes. Users are encouraged to cite relevant papers in their publications to support the development and improvement of the pyannote.audio user base.
meta-llama/Llama-3.2-1B
meta-llama
Llama 3.2 is a text generation model that supports multiple languages including English, German, French, Italian, Portuguese, Hindi, Spanish, and Thai. It is built on the PyTorch library and is part of the Llama series developed by Meta. The model is designed for various text generation tasks and is tagged for easy discovery.
dslim/bert-base-NER
dslim
This model is designed for token classification tasks using the CoNLL2003 dataset. It achieves an accuracy of 0.9118, with a precision of 0.9212, recall of 0.9306, and an F1 score of 0.9259. The model is licensed under MIT.
sentence-transformers/multi-qa-MiniLM-L6-cos-v1
sentence-transformers
This model is designed for sentence similarity tasks. It supports English language and is part of the sentence-transformers library. Key features include feature extraction and sentence similarity. It has been trained on various datasets including flax-sentence-embeddings/stackexchange_xml, ms_marco, gooaq, yahoo_answers_topics, search_qa, eli5, natural_questions, trivia_qa, embedding-data/QQP, embedding-data/PAQ_pairs, embedding-data/Amazon-QA, and embedding-data/WikiAnswers. The model is tagged for sentence similarity.
facebook/bart-base
--- license: apache-2.0 language: en ---
meta-llama/Meta-Llama-3-8B
meta-llama
Language model for text generation. Tags include facebook, meta, pytorch, llama, and llama-3. License is llama3. New version available: Llama 3.1 8B. Extra gated prompt is included.
Qwen/Qwen2.5-7B
Qwen
This model is designed for text generation tasks. It is licensed under the Apache 2.0 license, and the license details can be found at the provided link. The model supports the English language and is compatible with the transformers library.
nomic-ai/nomic-embed-text-v1.5
nomic-ai
The model is designed for sentence similarity tasks using the sentence-transformers library. It supports feature extraction and is part of the MTEB benchmark, specifically for the Amazon Counterfactual Classification dataset in English. The model's performance metrics include an accuracy of 75.21%, an average precision of 38.58%, and an F1 score.
unslothai/1
unslothai
--- library_name: transformers tags: [] ---
speechbrain/spkrec-ecapa-voxceleb
speechbrain
This model is designed for speaker verification and identification using embeddings. It utilizes the ECAPA-TDNN architecture and is implemented in PyTorch. The model is trained on the VoxCeleb dataset and is licensed under Apache 2.0. It supports various applications in speech processing.
allenai/longformer-base-4096
allenai
Language model licensed under Apache 2.0.
intfloat/multilingual-e5-base
intfloat
Multilingual E5 Base is designed for sentence similarity tasks and is part of the MTEB benchmark. It has been evaluated on the Amazon Counterfactual Classification dataset, achieving an accuracy of 78.97%, an average precision of 43.69%, and an F1 score of 73.38%. This model is suitable for various multilingual applications in natural language processing.
Comfy-Org/Qwen-Image_ComfyUI
Comfy-Org
A diffusion model designed for use with ComfyUI, offering a streamlined interface for image generation.
jonatasgrosman/wav2vec2-large-xlsr-53-persian
jonatasgrosman
--- language: fa datasets: - common_voice metrics: - wer - cer tags: - audio - automatic-speech-recognition - speech - xlsr-fine-tuning-week license: apache-2.0 model-index: - name: XLSR Wav2Vec2 Persian by Jonatas Grosman results: - task: name: Speech Recognition type: automatic-speech-recognition dataset: name: Common Voice fa type: common_voice args: fa metrics: - name: Test WER type: wer value: 30.12 - name: Test CER type: cer value: 7.37 ---
distilbert/distilroberta-base
distilbert
--- language: en tags: - exbert
Qwen/Qwen2.5-0.5B-Instruct
Qwen
This model is designed for text generation tasks in English. It is based on the Qwen 2.5 architecture and is compatible with the transformers library. The model is licensed under the Apache 2.0 license.
pysentimiento/robertuito-sentiment-analysis
pysentimiento
--- language: - es library_name: pysentimiento pipeline_tag: text-classification tags: - twitter - sentiment-analysis
cross-encoder/ms-marco-TinyBERT-L2-v2
cross-encoder
--- license: apache-2.0 datasets: - sentence-transformers/msmarco language: - en base_model: - nreimers/BERT-Tiny_L-2_H-128_A-2 pipeline_tag: text-ranking library_name: sentence-transformers tags: - transformers ---
Qwen/Qwen3-32B
Qwen
This model is designed for text generation. It is part of the transformers library and is licensed under the Apache 2.0 license. For more information, you can visit the license link at https://huggingface.co/Qwen/Qwen3-32B/blob/main/LICENSE.
facebook/dinov2-small
--- license: apache-2.0 tags: - dino - vision ---
allenai/OLMo-2-0425-1B
allenai
--- license: apache-2.0 language: - en library_name: transformers ---
intfloat/multilingual-e5-small
intfloat
This model supports multiple languages, including Afrikaans, Amharic, Arabic, Assamese, Azerbaijani, Belarusian, Bulgarian, Bengali, Breton, Bosnian, Catalan, Czech, Welsh, Danish, German, Greek, English, Esperanto, Spanish, Estonian, Basque, Persian, Finnish, French, Frisian, Irish, Scottish Gaelic, Galician, Gujarati, Hausa, Hebrew, Hindi, Croatian, Hungarian, Armenian, Indonesian, Icelandic, Italian, Japanese, Javanese, Georgian, Kazakh, Khmer, Kannada, Korean, Kurdish, Kyrgyz, Latin, Lao, Lithuanian, Latvian, Malagasy, Macedonian, Malayalam, Mongolian, Marathi, Malay, Nepali, Dutch, Norwegian, Oromo, Punjabi, Polish, Pashto, Portuguese, Romanian, Russian, Arabic, Sindhi, Sinhala, Slovak, Slovenian, Somali, Albanian, Serbian, Sundanese, Swedish, Swahili, Tamil, Telugu, Thai, Tagalog, Turkish, Uyghur, Ukrainian, Urdu, Uzbek, Vietnamese, Xhosa, Yiddish, and Chinese.
facebook/dino-vits16
--- license: apache-2.0 tags: - dino - vision datasets: - imagenet-1k ---
black-forest-labs/FLUX.1-dev
black-forest-labs
Flux is a text-to-image model designed for image generation. It is licensed under the Flux-1-Dev Non-Commercial License. By clicking 'Agree', you agree to the FluxDev Non-Commercial License Agreement and acknowledge the Acceptable Use Policy.
microsoft/Phi-3-mini-4k-instruct
microsoft
This model is designed for text generation and supports English and French languages. It is tagged for natural language processing and code inference. The model has parameters such as temperature set to 0. It can assist users with various tasks, including providing ways to eat combinations of bananas and dragonfruits.
Xenova/speecht5_tts
Xenova
Base model for text-to-speech using the SpeechT5 architecture from Microsoft, implemented in the transformers.js library.
cambridgeltl/SapBERT-from-PubMedBERT-fulltext
cambridgeltl
--- license: apache-2.0 language: - en tags: - biomedical - lexical semantics - bionlp - biology - science - embedding - entity linking --- ---
google-bert/bert-base-chinese
google-bert
Bert Base Chinese is a language model designed for understanding and processing Chinese text. It is licensed under the Apache 2.0 license.
google-bert/bert-large-uncased
google-bert
--- language: en license: apache-2.0 datasets: - bookcorpus - wikipedia ---
cross-encoder/ms-marco-MiniLM-L4-v2
cross-encoder
--- license: apache-2.0 datasets: - sentence-transformers/msmarco language: - en base_model: - cross-encoder/ms-marco-MiniLM-L12-v2 pipeline_tag: text-ranking library_name: sentence-transformers tags: - transformers ---
facebook/musicgen-medium
--- inference: true tags: - musicgen license: cc-by-nc-4.0 pipeline_tag: text-to-audio widget: - text: a funky house with 80s hip hop vibes example_title: Prompt 1 - text: a chill song with influences from lofi, chillstep and downtempo example_title: Prompt 2 - text: a catchy beat for a podcast intro example_title: Prompt 3 ---
apple/mobilevit-small
apple
MobileVit Small is a vision model designed for image classification. It is trained on the ImageNet-1K dataset. Example images include a tiger, a teapot, and a palace.
intfloat/e5-base-v2
intfloat
E5 Base V2 is a sentence-transformers model designed for sentence similarity tasks. It has been evaluated on the MTEB Amazon Counterfactual Classification dataset with a test split in English. The model achieved an accuracy of 77.78%, an average precision of 42.05%, and an F1 score of 72.12%. The revision used for this evaluation is e8379541af4e31359cca9fbcf4b00f2671dba205.
IDEA-Research/grounding-dino-base
IDEA-Research
Grounding Dino Base is a model designed for zero-shot object detection. It operates under the Apache 2.0 license and is focused on vision tasks.
microsoft/table-transformer-detection
microsoft
--- license: mit widget: - src: https://www.invoicesimple.com/wp-content/uploads/2018/06/Sample-Invoice-printable.png example_title: Invoice ---
google/gemma-3-12b-it
Gemma is a model designed for image-to-text tasks, utilizing the transformers library. To access Gemma on Hugging Face, you must review and agree to Google's usage license. Ensure you are logged in to Hugging Face to proceed with the request.
Alibaba-NLP/gte-multilingual-base
Alibaba-NLP
A multilingual model designed for sentence similarity and text embeddings inference. It supports a wide range of languages including Afrikaans, Arabic, Azerbaijani, Belarusian, Bulgarian, Bengali, Catalan, Cebuano, Czech, Welsh, Danish, German, Greek, English, Spanish, Estonian, Basque, Persian, Finnish, French, Galician, Gujarati, Hebrew, Hindi, Croatian, Haitian Creole, Hungarian, Armenian, Indonesian, Icelandic, Italian, Japanese, Javanese, Georgian, Kazakh, Khmer, Kannada, Korean, Kyrgyz, Lao, Lithuanian, Latvian, Macedonian, Malayalam, Mongolian, Marathi, Malay, Burmese, Nepali, Dutch, Norwegian, Punjabi, Polish, Portuguese, Quechua, Romanian, Russian, Sinhala, Slovak, Slovenian, Somali, Albanian, Serbian, Swedish, Swahili, Tamil, Telugu, Thai, Tagalog.
amazon/chronos-bolt-base
amazon
Time series forecasting model licensed under Apache 2.0. It is a pretrained foundation model designed for time series analysis. The latest version is amazon/chronos-2.
pyannote/voice-activity-detection
pyannote
Voice activity detection identifies segments of audio that contain speech. It is useful for various applications in audio processing, including automatic speech recognition and speaker identification. The model is associated with tags such as pyannote, pyannote-audio, and voice-activity-detection. It has been trained on datasets like AMI, DiHard, and VoxConverse. The model is licensed under MIT. The collected information will help acquire a better knowledge of the pyannote.audio user base and assist maintainers in applying for grants to improve it further. Academic researchers are encouraged to cite relevant papers in their publications.
answerdotai/answerai-colbert-small-v1
answerdotai
--- license: apache-2.0 language: - en tags: - ColBERT - RAGatouille - passage-retrieval ---
google-t5/t5-base
google-t5
The model is designed for translation tasks, supporting languages such as English, French, Romanian, and German. It is trained on the C4 dataset and is tagged for summarization and translation. The model is licensed under Apache 2.0.
dbmdz/bert-large-cased-finetuned-conll03-english
dbmdz
openai-community/gpt2-large
openai-community
--- language: en license: mit ---
lpiccinelli/unidepth-v2-vitl14
lpiccinelli
--- library_name: UniDepth tags: - model_hub_mixin - monocular-metric-depth-estimation - pytorch_model_hub_mixin ---
distil-whisper/distil-large-v3
distil-whisper
Automatic speech recognition model supporting English language. Licensed under MIT. Utilizes the transformers library. Tags include audio, automatic-speech-recognition, and transformers.js. Example audio samples include LibriSpeech sample 1 and LibriSpeech sample 2, available at https://cdn-media.huggingface.co/speech_samples/sample1.flac and https://cdn-media.huggingface.co/speech_samples/sample2.flac respectively. The model is designed for efficient and accurate transcription of spoken language.