_id stringlengths 24 24 | id stringlengths 7 122 | author stringlengths 2 41 | base_models dict | downloads int64 0 206M | downloads_all_time int64 0 2.81B | gated stringclasses 3
values | created_at timestamp[us, tz=UTC]date 2022-03-02 23:29:04 2025-11-03 14:41:54 | last_modified timestamp[us, tz=UTC]date 2020-12-11 21:34:15 2026-03-29 10:14:05 | library_name stringclasses 36
values | likes int64 0 13.1k | trending_score float64 0 47 | model_index stringlengths 30 911k ⌀ | pipeline_tag stringclasses 46
values | safetensors stringlengths 30 122 ⌀ | siblings listlengths 0 10k | sha stringlengths 40 40 | tags listlengths 2 1.82k | gguf stringclasses 789
values | config stringlengths 2 53.7k ⌀ | transformers_info dict | card_data stringlengths 234 905k ⌀ | card stringlengths 0 638k ⌀ | spaces null | licenses listlengths 1 3 ⌀ | datasets listlengths 1 289 ⌀ | languages listlengths 1 1.81k ⌀ | safetensors_params float64 0 1,019B ⌀ | gguf_params float64 0 122B ⌀ | metrics listlengths 1 15 ⌀ | architectures listlengths 1 6 ⌀ | tasks listlengths 1 6 ⌀ | modalities listlengths 1 4 ⌀ | input_modalities listlengths 1 3 ⌀ | output_modalities listlengths 1 4 ⌀ |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
6698d8a0653e4babe21e1e7d | meta-llama/Llama-3.1-8B-Instruct | meta-llama | {
"models": [
{
"_id": "66944f1fe0c5c2e493a804f5",
"id": "meta-llama/Llama-3.1-8B"
}
],
"relation": "finetune"
} | 8,384,864 | 133,735,004 | manual | 2024-07-18T08:56:00 | 2024-09-25T17:00:57 | transformers | 5,631 | 47 | null | text-generation | {"parameters": {"BF16": 8030261248}, "total": 8030261248} | [
".gitattributes",
"LICENSE",
"README.md",
"USE_POLICY.md",
"config.json",
"generation_config.json",
"model-00001-of-00004.safetensors",
"model-00002-of-00004.safetensors",
"model-00003-of-00004.safetensors",
"model-00004-of-00004.safetensors",
"model.safetensors.index.json",
"original/consolid... | 0e9e39f249a16976918f6564b8830bc894c89659 | [
"transformers",
"safetensors",
"llama",
"text-generation",
"facebook",
"meta",
"pytorch",
"llama-3",
"conversational",
"en",
"de",
"fr",
"it",
"pt",
"hi",
"es",
"th",
"arxiv:2204.05149",
"base_model:meta-llama/Llama-3.1-8B",
"base_model:finetune:meta-llama/Llama-3.1-8B",
"lic... | null | {"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{{- bos_token }}\n{%- if custom_tools is defined %}\n {%- set tools = custom_tools %}\n{%- endif %}\n{%- if not tools_in_user_message is defined %}\n {%- set tools_in_user_messag... | {
"auto_model": "AutoModelForCausalLM",
"custom_class": null,
"pipeline_tag": "text-generation",
"processor": "AutoTokenizer"
} | {"base_model": "meta-llama/Meta-Llama-3.1-8B", "datasets": null, "eval_results": null, "language": ["en", "de", "fr", "it", "pt", "hi", "es", "th"], "library_name": null, "license": "llama3.1", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": "text-generation", "tags": ["... | null | null | [
"llama3.1"
] | null | [
"en",
"de",
"fr",
"it",
"pt",
"hi",
"es",
"th"
] | 8,030,261,248 | null | null | [
"AutoModelForCausalLM",
"llama",
"LlamaForCausalLM"
] | [
"text-generation"
] | [
"text"
] | [
"text"
] | [
"text"
] |
654a84cadff2f49007ce6c37 | openai/whisper-large-v3 | openai | null | 4,832,162 | 115,357,760 | False | 2023-11-07T18:41:14 | 2024-08-12T10:20:10 | transformers | 5,522 | 30 | null | automatic-speech-recognition | null | [
".gitattributes",
"README.md",
"added_tokens.json",
"config.json",
"flax_model.msgpack",
"generation_config.json",
"merges.txt",
"model.fp32-00001-of-00002.safetensors",
"model.fp32-00002-of-00002.safetensors",
"model.safetensors",
"model.safetensors.index.fp32.json",
"normalizer.json",
"pre... | 06f233fe06e710322aca913c1bc4249a0d71fce1 | [
"transformers",
"pytorch",
"jax",
"safetensors",
"whisper",
"automatic-speech-recognition",
"audio",
"hf-asr-leaderboard",
"en",
"zh",
"de",
"es",
"ru",
"ko",
"fr",
"ja",
"pt",
"tr",
"pl",
"ca",
"nl",
"ar",
"sv",
"it",
"id",
"hi",
"fi",
"vi",
"he",
"uk",
"... | null | {"architectures": ["WhisperForConditionalGeneration"], "model_type": "whisper", "tokenizer_config": {"bos_token": "<|endoftext|>", "eos_token": "<|endoftext|>", "pad_token": "<|endoftext|>", "unk_token": "<|endoftext|>"}} | {
"auto_model": "AutoModelForSpeechSeq2Seq",
"custom_class": null,
"pipeline_tag": "automatic-speech-recognition",
"processor": "AutoProcessor"
} | {"base_model": null, "datasets": null, "eval_results": null, "language": ["en", "zh", "de", "es", "ru", "ko", "fr", "ja", "pt", "tr", "pl", "ca", "nl", "ar", "sv", "it", "id", "hi", "fi", "vi", "he", "uk", "el", "ms", "cs", "ro", "da", "hu", "ta", "no", "th", "ur", "hr", "bg", "lt", "la", "mi", "ml", "cy", "sk", "te", ... | # Whisper
Whisper is a state-of-the-art model for automatic speech recognition (ASR) and speech translation, proposed in the paper
[Robust Speech Recognition via Large-Scale Weak Supervision](https://huggingface.co/papers/2212.04356) by Alec Radford
et al. from OpenAI. Trained on >5M hours of labeled data, Whisper d... | null | [
"apache-2.0"
] | null | [
"en",
"zh",
"de",
"es",
"ru",
"ko",
"fr",
"ja",
"pt",
"tr",
"pl",
"ca",
"nl",
"ar",
"sv",
"it",
"id",
"hi",
"fi",
"vi",
"he",
"uk",
"el",
"ms",
"cs",
"ro",
"da",
"hu",
"ta",
"no",
"th",
"ur",
"hr",
"bg",
"lt",
"la",
"mi",
"ml",
"cy",
"sk"... | null | null | null | [
"WhisperForConditionalGeneration",
"whisper",
"AutoModelForSpeechSeq2Seq"
] | [
"automatic-speech-recognition"
] | [
"multimodal"
] | [
"audio"
] | [
"text"
] |
68913522f16f3c8aaffccf1f | openai/gpt-oss-120b | openai | null | 4,363,728 | 29,644,344 | False | 2025-08-04T22:33:06 | 2025-08-26T17:25:03 | transformers | 4,621 | 23 | null | text-generation | {"parameters": {"BF16": 2167371072, "U8": 118244966400}, "total": 120412337472} | [
".gitattributes",
"LICENSE",
"README.md",
"USAGE_POLICY",
"chat_template.jinja",
"config.json",
"generation_config.json",
"metal/model.bin",
"model-00000-of-00014.safetensors",
"model-00001-of-00014.safetensors",
"model-00002-of-00014.safetensors",
"model-00003-of-00014.safetensors",
"model-... | b5c939de8f754692c1647ca79fbf85e8c1e70f8a | [
"transformers",
"safetensors",
"gpt_oss",
"text-generation",
"vllm",
"conversational",
"arxiv:2508.10925",
"license:apache-2.0",
"eval-results",
"endpoints_compatible",
"8-bit",
"mxfp4",
"deploy:azure",
"region:us"
] | null | {"architectures": ["GptOssForCausalLM"], "model_type": "gpt_oss", "quantization_config": {"quant_method": "mxfp4"}, "tokenizer_config": {"bos_token": "<|startoftext|>", "eos_token": "<|return|>", "pad_token": "<|endoftext|>"}, "chat_template_jinja": "{#-\n In addition to the normal inputs of `messages` and `tools`, th... | {
"auto_model": "AutoModelForCausalLM",
"custom_class": null,
"pipeline_tag": "text-generation",
"processor": "AutoTokenizer"
} | {"base_model": null, "datasets": null, "eval_results": null, "language": null, "library_name": "transformers", "license": "apache-2.0", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": "text-generation", "tags": ["vllm"]} | <p align="center">
<img alt="gpt-oss-120b" src="https://raw.githubusercontent.com/openai/gpt-oss/main/docs/gpt-oss-120b.svg">
</p>
<p align="center">
<a href="https://gpt-oss.com"><strong>Try gpt-oss</strong></a> ·
<a href="https://cookbook.openai.com/topic/gpt-oss"><strong>Guides</strong></a> ·
<a href="https... | null | [
"apache-2.0"
] | null | null | 120,412,337,472 | null | null | [
"AutoModelForCausalLM",
"GptOssForCausalLM",
"gpt_oss"
] | [
"text-generation"
] | [
"text"
] | [
"text"
] | [
"text"
] |
621ffdc036468d709f17434d | openai-community/gpt2 | openai-community | null | 11,636,493 | 836,599,686 | False | 2022-03-02T23:29:04 | 2024-02-19T10:57:45 | transformers | 3,153 | 22 | null | text-generation | {"parameters": {"F32": 137022720}, "total": 137022720} | [
".gitattributes",
"64-8bits.tflite",
"64-fp16.tflite",
"64.tflite",
"README.md",
"config.json",
"flax_model.msgpack",
"generation_config.json",
"merges.txt",
"model.safetensors",
"onnx/config.json",
"onnx/decoder_model.onnx",
"onnx/decoder_model_merged.onnx",
"onnx/decoder_with_past_model.... | 607a30d783dfa663caf39e06633721c8d4cfcd7e | [
"transformers",
"pytorch",
"tf",
"jax",
"tflite",
"rust",
"onnx",
"safetensors",
"gpt2",
"text-generation",
"exbert",
"en",
"doi:10.57967/hf/0039",
"license:mit",
"text-generation-inference",
"endpoints_compatible",
"deploy:azure",
"region:us"
] | null | {"architectures": ["GPT2LMHeadModel"], "model_type": "gpt2", "tokenizer_config": {}} | {
"auto_model": "AutoModelForCausalLM",
"custom_class": null,
"pipeline_tag": "text-generation",
"processor": "AutoTokenizer"
} | {"base_model": null, "datasets": null, "eval_results": null, "language": "en", "library_name": null, "license": "mit", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": null, "tags": ["exbert"]} | # GPT-2
Test the whole generation capabilities here: https://transformer.huggingface.co/doc/gpt2-large
Pretrained model on English language using a causal language modeling (CLM) objective. It was introduced in
[this paper](https://d4mucfpksywv.cloudfront.net/better-language-models/language_models_are_unsupervised_mu... | null | [
"mit"
] | null | [
"en"
] | 137,022,720 | null | null | [
"GPT2LMHeadModel",
"AutoModelForCausalLM",
"gpt2"
] | [
"text-generation"
] | [
"text"
] | [
"text"
] | [
"text"
] |
66e81cbcd683a3f4e5291bbf | Qwen/Qwen2.5-7B-Instruct | Qwen | {
"models": [
{
"_id": "66e6d06401e2adb00b0be996",
"id": "Qwen/Qwen2.5-7B"
}
],
"relation": "finetune"
} | 17,917,551 | 109,933,464 | False | 2024-09-16T11:55:40 | 2025-01-12T02:10:10 | transformers | 1,165 | 22 | null | text-generation | {"parameters": {"BF16": 7615616512}, "total": 7615616512} | [
".gitattributes",
"LICENSE",
"README.md",
"config.json",
"generation_config.json",
"merges.txt",
"model-00001-of-00004.safetensors",
"model-00002-of-00004.safetensors",
"model-00003-of-00004.safetensors",
"model-00004-of-00004.safetensors",
"model.safetensors.index.json",
"tokenizer.json",
"... | a09a35458c702b33eeacc393d103063234e8bc28 | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"chat",
"conversational",
"en",
"arxiv:2309.00071",
"arxiv:2407.10671",
"base_model:Qwen/Qwen2.5-7B",
"base_model:finetune:Qwen/Qwen2.5-7B",
"license:apache-2.0",
"text-generation-inference",
"endpoints_compatible",
"deploy:azure... | null | {"architectures": ["Qwen2ForCausalLM"], "model_type": "qwen2", "tokenizer_config": {"bos_token": null, "chat_template": "{%- if tools %}\n {{- '<|im_start|>system\\n' }}\n {%- if messages[0]['role'] == 'system' %}\n {{- messages[0]['content'] }}\n {%- else %}\n {{- 'You are Qwen, created by Aliba... | {
"auto_model": "AutoModelForCausalLM",
"custom_class": null,
"pipeline_tag": "text-generation",
"processor": "AutoTokenizer"
} | {"base_model": "Qwen/Qwen2.5-7B", "datasets": null, "eval_results": null, "language": ["en"], "library_name": "transformers", "license": "apache-2.0", "license_name": null, "license_link": "https://huggingface.co/Qwen/Qwen2.5-7B-Instruct/blob/main/LICENSE", "metrics": null, "model_name": null, "pipeline_tag": "text-gen... | # Qwen2.5-7B-Instruct
<a href="https://chat.qwenlm.ai/" target="_blank" style="margin: 2px;">
<img alt="Chat" src="https://img.shields.io/badge/%F0%9F%92%9C%EF%B8%8F%20Qwen%20Chat%20-536af5" style="display: inline-block; vertical-align: middle;"/>
</a>
## Introduction
Qwen2.5 is the latest series of Qwen large la... | null | [
"apache-2.0",
"https://huggingface.co/Qwen/Qwen2.5-7B-Instruct/blob/main/LICENSE"
] | null | [
"en"
] | 7,615,616,512 | null | null | [
"AutoModelForCausalLM",
"Qwen2ForCausalLM",
"qwen2"
] | [
"text-generation"
] | [
"text"
] | [
"text"
] | [
"text"
] |
67b79c8700245b72c5706777 | google/gemma-3-4b-it | google | {
"models": [
{
"_id": "67b79c6c01ad68cfed14677a",
"id": "google/gemma-3-4b-pt"
}
],
"relation": "finetune"
} | 1,702,746 | 15,276,165 | manual | 2025-02-20T21:20:07 | 2025-03-21T20:20:53 | transformers | 1,265 | 22 | null | image-text-to-text | null | [
".gitattributes",
"README.md",
"added_tokens.json",
"chat_template.json",
"config.json",
"generation_config.json",
"model-00001-of-00002.safetensors",
"model-00002-of-00002.safetensors",
"model.safetensors.index.json",
"preprocessor_config.json",
"processor_config.json",
"special_tokens_map.js... | 093f9f388b31de276ce2de164bdc2081324b9767 | [
"transformers",
"safetensors",
"gemma3",
"image-text-to-text",
"conversational",
"arxiv:1905.07830",
"arxiv:1905.10044",
"arxiv:1911.11641",
"arxiv:1904.09728",
"arxiv:1705.03551",
"arxiv:1911.01547",
"arxiv:1907.10641",
"arxiv:1903.00161",
"arxiv:2009.03300",
"arxiv:2304.06364",
"arxi... | null | {"architectures": ["Gemma3ForConditionalGeneration"], "model_type": "gemma3", "processor_config": {"chat_template": "{{ bos_token }}\n{%- if messages[0]['role'] == 'system' -%}\n {%- if messages[0]['content'] is string -%}\n {%- set first_user_prefix = messages[0]['content'] + '\n\n' -%}\n {%- else -%}\n ... | {
"auto_model": "AutoModelForImageTextToText",
"custom_class": null,
"pipeline_tag": "image-text-to-text",
"processor": "AutoProcessor"
} | {"base_model": "google/gemma-3-4b-pt", "datasets": null, "eval_results": null, "language": null, "library_name": "transformers", "license": "gemma", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": "image-text-to-text", "tags": null, "extra_gated_heading": "Access Gemma o... | null | null | [
"gemma"
] | null | null | null | null | null | [
"AutoModelForImageTextToText",
"Gemma3ForConditionalGeneration",
"gemma3"
] | [
"image-text-to-text"
] | [
"multimodal"
] | [
"text",
"image"
] | [
"text"
] |
621ffdc136468d709f180294 | sentence-transformers/all-MiniLM-L6-v2 | sentence-transformers | null | 206,073,068 | 2,406,656,367 | False | 2022-03-02T23:29:05 | 2025-03-06T13:37:44 | sentence-transformers | 4,620 | 21 | null | sentence-similarity | {"parameters": {"I64": 512, "F32": 22713216}, "total": 22713728} | [
".gitattributes",
"1_Pooling/config.json",
"README.md",
"config.json",
"config_sentence_transformers.json",
"data_config.json",
"model.safetensors",
"modules.json",
"onnx/model.onnx",
"onnx/model_O1.onnx",
"onnx/model_O2.onnx",
"onnx/model_O3.onnx",
"onnx/model_O4.onnx",
"onnx/model_qint8_... | c9745ed1d9f207416be6d2e6f8de32d1f16199bf | [
"sentence-transformers",
"pytorch",
"tf",
"rust",
"onnx",
"safetensors",
"openvino",
"bert",
"feature-extraction",
"sentence-similarity",
"transformers",
"en",
"dataset:s2orc",
"dataset:flax-sentence-embeddings/stackexchange_xml",
"dataset:ms_marco",
"dataset:gooaq",
"dataset:yahoo_a... | null | {"architectures": ["BertModel"], "model_type": "bert", "tokenizer_config": {"unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]"}} | {
"auto_model": "AutoModel",
"custom_class": null,
"pipeline_tag": "feature-extraction",
"processor": "AutoTokenizer"
} | {"base_model": null, "datasets": ["s2orc", "flax-sentence-embeddings/stackexchange_xml", "ms_marco", "gooaq", "yahoo_answers_topics", "code_search_net", "search_qa", "eli5", "snli", "multi_nli", "wikihow", "natural_questions", "trivia_qa", "embedding-data/sentence-compression", "embedding-data/flickr30k-captions", "emb... | # all-MiniLM-L6-v2
This is a [sentence-transformers](https://www.SBERT.net) model: It maps sentences & paragraphs to a 384 dimensional dense vector space and can be used for tasks like clustering or semantic search.
## Usage (Sentence-Transformers)
Using this model becomes easy when you have [sentence-transformers](ht... | null | [
"apache-2.0"
] | [
"s2orc",
"flax-sentence-embeddings/stackexchange_xml",
"ms_marco",
"gooaq",
"yahoo_answers_topics",
"code_search_net",
"search_qa",
"eli5",
"snli",
"multi_nli",
"wikihow",
"natural_questions",
"trivia_qa",
"embedding-data/sentence-compression",
"embedding-data/flickr30k-captions",
"emb... | [
"en"
] | 22,713,728 | null | null | [
"BertModel",
"AutoModel",
"bert"
] | [
"sentence-similarity",
"feature-extraction"
] | [
"text",
"multimodal"
] | [
"text"
] | [
"logits",
"embeddings"
] |
66944f1fe0c5c2e493a804f5 | meta-llama/Llama-3.1-8B | meta-llama | null | 1,387,690 | 20,616,230 | manual | 2024-07-14T22:20:15 | 2024-10-16T22:00:37 | transformers | 2,128 | 16 | null | text-generation | {"parameters": {"BF16": 8030261248}, "total": 8030261248} | [
".gitattributes",
"LICENSE",
"README.md",
"USE_POLICY.md",
"config.json",
"generation_config.json",
"model-00001-of-00004.safetensors",
"model-00002-of-00004.safetensors",
"model-00003-of-00004.safetensors",
"model-00004-of-00004.safetensors",
"model.safetensors.index.json",
"original/consolid... | d04e592bb4f6aa9cfee91e2e20afa771667e1d4b | [
"transformers",
"safetensors",
"llama",
"text-generation",
"facebook",
"meta",
"pytorch",
"llama-3",
"en",
"de",
"fr",
"it",
"pt",
"hi",
"es",
"th",
"arxiv:2204.05149",
"license:llama3.1",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | null | {"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "eos_token": "<|end_of_text|>"}} | {
"auto_model": "AutoModelForCausalLM",
"custom_class": null,
"pipeline_tag": "text-generation",
"processor": "AutoTokenizer"
} | {"base_model": null, "datasets": null, "eval_results": null, "language": ["en", "de", "fr", "it", "pt", "hi", "es", "th"], "library_name": "transformers", "license": "llama3.1", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": "text-generation", "tags": ["facebook", "meta... | null | null | [
"llama3.1"
] | null | [
"en",
"de",
"fr",
"it",
"pt",
"hi",
"es",
"th"
] | 8,030,261,248 | null | null | [
"AutoModelForCausalLM",
"llama",
"LlamaForCausalLM"
] | [
"text-generation"
] | [
"text"
] | [
"text"
] | [
"text"
] |
67fddfa9a7fe1f21ec1d3026 | microsoft/bitnet-b1.58-2B-4T | microsoft | null | 16,753 | 190,249 | False | 2025-04-15T04:25:13 | 2025-12-17T18:13:05 | transformers | 1,404 | 15 | null | text-generation | {"parameters": {"BF16": 328775890, "U8": 521011200}, "total": 849787090} | [
".gitattributes",
"LICENSE",
"README.md",
"config.json",
"data_summary_card.md",
"generation_config.json",
"model.safetensors",
"special_tokens_map.json",
"tokenizer.json",
"tokenizer_config.json"
] | 04c3b9ad9361b824064a1f25ea60a8be9599b127 | [
"transformers",
"safetensors",
"bitnet",
"text-generation",
"chat",
"large-language-model",
"conversational",
"custom_code",
"en",
"arxiv:2504.12285",
"license:mit",
"endpoints_compatible",
"8-bit",
"deploy:azure",
"region:us"
] | null | {"architectures": ["BitNetForCausalLM"], "auto_map": {"AutoConfig": "configuration_bitnet.BitNetConfig", "AutoModelForCausalLM": "modeling_bitnet.BitNetForCausalLM"}, "model_type": "bitnet", "quantization_config": {"quant_method": "bitnet"}, "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{% se... | {
"auto_model": "AutoModelForCausalLM",
"custom_class": null,
"pipeline_tag": "text-generation",
"processor": "AutoTokenizer"
} | {"base_model": null, "datasets": null, "eval_results": null, "language": ["en"], "library_name": "transformers", "license": "mit", "license_name": null, "license_link": "https://huggingface.co/microsoft/bitnet-b1.58-2B-4T/blob/main/LICENSE", "metrics": null, "model_name": null, "pipeline_tag": "text-generation", "tags"... | # BitNet b1.58 2B4T - Scaling Native 1-bit LLM
This repository contains the weights for **BitNet b1.58 2B4T**, the first open-source, native 1-bit Large Language Model (LLM) at the 2-billion parameter scale, developed by Microsoft Research.
Trained on a corpus of 4 trillion tokens, this model demonstrates that native... | null | [
"mit",
"https://huggingface.co/microsoft/bitnet-b1.58-2B-4T/blob/main/LICENSE"
] | null | [
"en"
] | 849,787,090 | null | null | [
"AutoModelForCausalLM",
"BitNetForCausalLM",
"bitnet"
] | [
"text-generation"
] | [
"text"
] | [
"text"
] | [
"text"
] |
680da79d48c19d19cf82ab5e | Qwen/Qwen3-8B | Qwen | {
"models": [
{
"_id": "680f0c3e6eacf22a04ee1219",
"id": "Qwen/Qwen3-8B-Base"
}
],
"relation": "finetune"
} | 9,481,229 | 43,490,001 | False | 2025-04-27T03:42:21 | 2025-07-26T03:49:13 | transformers | 1,013 | 15 | null | text-generation | {"parameters": {"BF16": 8190735360}, "total": 8190735360} | [
".gitattributes",
"LICENSE",
"README.md",
"config.json",
"generation_config.json",
"merges.txt",
"model-00001-of-00005.safetensors",
"model-00002-of-00005.safetensors",
"model-00003-of-00005.safetensors",
"model-00004-of-00005.safetensors",
"model-00005-of-00005.safetensors",
"model.safetensor... | b968826d9c46dd6066d109eabc6255188de91218 | [
"transformers",
"safetensors",
"qwen3",
"text-generation",
"conversational",
"arxiv:2309.00071",
"arxiv:2505.09388",
"base_model:Qwen/Qwen3-8B-Base",
"base_model:finetune:Qwen/Qwen3-8B-Base",
"license:apache-2.0",
"text-generation-inference",
"endpoints_compatible",
"deploy:azure",
"region... | null | {"architectures": ["Qwen3ForCausalLM"], "model_type": "qwen3", "tokenizer_config": {"bos_token": null, "chat_template": "{%- if tools %}\n {{- '<|im_start|>system\\n' }}\n {%- if messages[0].role == 'system' %}\n {{- messages[0].content + '\\n\\n' }}\n {%- endif %}\n {{- \"# Tools\\n\\nYou may call o... | {
"auto_model": "AutoModelForCausalLM",
"custom_class": null,
"pipeline_tag": "text-generation",
"processor": "AutoTokenizer"
} | {"base_model": ["Qwen/Qwen3-8B-Base"], "datasets": null, "eval_results": null, "language": null, "library_name": "transformers", "license": "apache-2.0", "license_name": null, "license_link": "https://huggingface.co/Qwen/Qwen3-8B/blob/main/LICENSE", "metrics": null, "model_name": null, "pipeline_tag": "text-generation"... | null | null | [
"apache-2.0",
"https://huggingface.co/Qwen/Qwen3-8B/blob/main/LICENSE"
] | null | null | 8,190,735,360 | null | null | [
"AutoModelForCausalLM",
"Qwen3ForCausalLM",
"qwen3"
] | [
"text-generation"
] | [
"text"
] | [
"text"
] | [
"text"
] |
683f05dc98de733bf4c59f3d | Qwen/Qwen3-Embedding-0.6B | Qwen | {
"models": [
{
"_id": "680f0da2f99a20ddfc9a9f26",
"id": "Qwen/Qwen3-0.6B-Base"
}
],
"relation": "finetune"
} | 5,507,791 | 36,810,259 | False | 2025-06-03T14:25:32 | 2025-06-20T09:31:05 | sentence-transformers | 950 | 15 | null | feature-extraction | {"parameters": {"BF16": 595776512}, "total": 595776512} | [
".gitattributes",
"1_Pooling/config.json",
"README.md",
"config.json",
"config_sentence_transformers.json",
"generation_config.json",
"merges.txt",
"model.safetensors",
"modules.json",
"tokenizer.json",
"tokenizer_config.json",
"vocab.json"
] | c54f2e6e80b2d7b7de06f51cec4959f6b3e03418 | [
"sentence-transformers",
"safetensors",
"qwen3",
"text-generation",
"transformers",
"sentence-similarity",
"feature-extraction",
"text-embeddings-inference",
"arxiv:2506.05176",
"base_model:Qwen/Qwen3-0.6B-Base",
"base_model:finetune:Qwen/Qwen3-0.6B-Base",
"license:apache-2.0",
"endpoints_co... | null | {"architectures": ["Qwen3ForCausalLM"], "model_type": "qwen3", "tokenizer_config": {"bos_token": null, "chat_template": "{%- if tools %}\n {{- '<|im_start|>system\\n' }}\n {%- if messages[0].role == 'system' %}\n {{- messages[0].content + '\\n\\n' }}\n {%- endif %}\n {{- \"# Tools\\n\\nYou may call o... | {
"auto_model": "AutoModelForCausalLM",
"custom_class": null,
"pipeline_tag": "text-generation",
"processor": "AutoTokenizer"
} | {"base_model": ["Qwen/Qwen3-0.6B-Base"], "datasets": null, "eval_results": null, "language": null, "library_name": null, "license": "apache-2.0", "license_name": null, "license_link": null, "metrics": null, "model_name": null, "pipeline_tag": null, "tags": ["transformers", "sentence-transformers", "sentence-similarity"... | # Qwen3-Embedding-0.6B
<p align="center">
<img src="https://qianwen-res.oss-accelerate-overseas.aliyuncs.com/logo_qwen3.png" width="400"/>
<p>
## Highlights
The Qwen3 Embedding model series is the latest proprietary model of the Qwen family, specifically designed for text embedding and ranking tasks. Building up... | null | [
"apache-2.0"
] | null | null | 595,776,512 | null | null | [
"AutoModelForCausalLM",
"Qwen3ForCausalLM",
"qwen3"
] | [
"sentence-similarity",
"feature-extraction",
"text-generation"
] | [
"text",
"multimodal"
] | [
"text"
] | [
"logits",
"text",
"embeddings"
] |
End of preview. Expand in Data Studio
README.md exists but content is empty.
- Downloads last month
- 58