Merge pull request #555 from exo-explore/modelvariations

add llama-3.2-1b-8bit, llama-3.2-3b-8bit, llama-3.2-3b-bf16
This commit is contained in:
Alex Cheema
2024-12-12 17:14:28 +00:00
committed by GitHub

View File

@@ -17,7 +17,28 @@ model_cards = {
"TinygradDynamicShardInferenceEngine": "unsloth/Llama-3.2-1B-Instruct",
},
},
"llama-3.2-1b-8bit": {
"layers": 16,
"repo": {
"MLXDynamicShardInferenceEngine": "mlx-community/Llama-3.2-1B-Instruct-8bit",
"TinygradDynamicShardInferenceEngine": "unsloth/Llama-3.2-1B-Instruct",
},
},
"llama-3.2-3b": {
"layers": 28,
"repo": {
"MLXDynamicShardInferenceEngine": "mlx-community/Llama-3.2-3B-Instruct-4bit",
"TinygradDynamicShardInferenceEngine": "unsloth/Llama-3.2-3B-Instruct",
},
},
"llama-3.2-3b-8bit": {
"layers": 28,
"repo": {
"MLXDynamicShardInferenceEngine": "mlx-community/Llama-3.2-3B-Instruct-8bit",
"TinygradDynamicShardInferenceEngine": "unsloth/Llama-3.2-3B-Instruct",
},
},
"llama-3.2-3b-bf16": {
"layers": 28,
"repo": {
"MLXDynamicShardInferenceEngine": "mlx-community/Llama-3.2-3B-Instruct",
@@ -94,7 +115,10 @@ model_cards = {
pretty_name = {
"llama-3.3-70b": "Llama 3.3 70B",
"llama-3.2-1b": "Llama 3.2 1B",
"llama-3.2-1b-8bit": "Llama 3.2 1B (8-bit)",
"llama-3.2-3b": "Llama 3.2 3B",
"llama-3.2-3b-8bit": "Llama 3.2 3B (8-bit)",
"llama-3.2-3b-bf16": "Llama 3.2 3B (BF16)",
"llama-3.1-8b": "Llama 3.1 8B",
"llama-3.1-70b": "Llama 3.1 70B",
"llama-3.1-70b-bf16": "Llama 3.1 70B (BF16)",