File size: 2,457 Bytes
172db2d
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
e32e654
172db2d
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
---
language:
- vi
- en
license: apache-2.0
library_name: transformers
tags:
- moe
- mixture-of-experts
- text-generation
- decode-series
- llm
- vietnamese-llm
datasets:
- markov-ai/computer-use-large
metrics:
- loss
- perplexity
model-index:
- name: Decode-12B-MoE
  results: []
---

# ๐Ÿš€ Decode-12B-MoE: High-Performance Mixture of Experts Model

**Decode-12B-MoE** is a Large Language Model (LLM) utilizing a **Sparse Mixture of Experts (MoE)** architecture with a total of **12.5 billion parameters**. This model is engineered to bridge the gap between massive parameter counts and computational efficiency, activating only a fraction of its weights (~2.5B) during inference.
** Untrained model! **
## ๐Ÿ“Œ Technical Specifications

| Attribute | Value |
| :--- | :--- |
| **Total Parameters** | 12,500,340,736 (12.5B) |
| **Active Parameters** | ~2.5B per token |
| **Architecture** | Sparse MoE (Decoder-only) |
| **Context Window** | 4096 tokens |
| **Format** | Bfloat16 / Float16 |
| **Training Hardware** | NVIDIA Tesla T4 (Prototyping) / [Your_Main_GPU] |

## ๐Ÿ›  Training Methodology

The model was trained with advanced memory optimization techniques to ensure stability on consumer and enterprise-grade hardware:
- **8-bit Optimizer:** Utilized `bitsandbytes` AdamW to reduce optimizer state memory footprint by 75%.
- **Gradient Checkpointing:** Enabled to manage activation memory for deep MoE layers.
- **Dataset:** Fine-tuned on a diverse corpus of Vietnamese and English text, focusing on reasoning, logic, and natural conversation.

## ๐Ÿ’ป Quick Start (Usage)

To use this model, ensure you have `transformers` and `accelerate` installed.

```python
from transformers import AutoModelForCausalLM, AutoTokenizer
import torch

# Replace with your actual Hugging Face repo ID
model_id = "your-username/decode-12b-moe"

tokenizer = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForCausalLM.from_pretrained(
    model_id,
    torch_dtype=torch.bfloat16,
    device_map="auto",
    trust_remote_code=True # Required for custom MoE architectures
)

# Test Prompt
prompt = "Explain the concept of Quantum Computing in simple terms."
inputs = tokenizer(prompt, return_tensors="pt").to("cuda")

with torch.no_grad():
    outputs = model.generate(
        **inputs, 
        max_new_tokens=512, 
        temperature=0.7, 
        top_p=0.9,
        do_sample=True
    )

print(tokenizer.decode(outputs[0], skip_special_tokens=True))