sr5434/CodegebraGPT_data
Viewer • Updated • 1.25M • 123 • 1
How to use sr5434/CodegebraGPT-10b with PEFT:
from peft import PeftModel
from transformers import AutoModelForCausalLM
base_model = AutoModelForCausalLM.from_pretrained("upstage/SOLAR-10.7B-v1.0")
model = PeftModel.from_pretrained(base_model, "sr5434/CodegebraGPT-10b")This model is a fine-tuned version of upstage/SOLAR-10.7B-v1.0 on the text only 100k samples subset of sr5434/CodegebraGPT_Data dataset. It stopped at 37k steps(for an unknown reason) instead of at 100k steps.
It can chat with you about science, engineering, math, or coding.
This is not finetuned with RLHF and is not intended to be used in production.
CodegebraGPT 100k text dataset
The following hyperparameters were used during training:
Detailed results can be found here
| Metric | Value |
|---|---|
| Avg. | 62.68 |
| AI2 Reasoning Challenge (25-Shot) | 59.81 |
| HellaSwag (10-Shot) | 83.42 |
| MMLU (5-Shot) | 60.20 |
| TruthfulQA (0-shot) | 46.57 |
| Winogrande (5-shot) | 80.98 |
| GSM8k (5-shot) | 45.11 |
Base model
upstage/SOLAR-10.7B-v1.0