Update README.md
Browse files
README.md
CHANGED
|
@@ -271,13 +271,20 @@ extra_gated_fields:
|
|
| 271 |
|
| 272 |
These files are GPTQ 4bit model files for [Bigcode's Starcoder](https://huggingface.co/bigcode/starcoder).
|
| 273 |
|
| 274 |
-
It is the result of quantising to 4bit using [
|
| 275 |
|
| 276 |
## Repositories available
|
| 277 |
|
| 278 |
* [4-bit GPTQ models for GPU inference](https://huggingface.co/TheBloke/starcoder-GPTQ)
|
| 279 |
-
* [
|
| 280 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 281 |
|
| 282 |
## How to easily download and use this model in text-generation-webui
|
| 283 |
|
|
@@ -308,7 +315,6 @@ from auto_gptq import AutoGPTQForCausalLM, BaseQuantizeConfig
|
|
| 308 |
import argparse
|
| 309 |
|
| 310 |
model_name_or_path = "TheBloke/starcoder-GPTQ"
|
| 311 |
-
model_basename = "gptq_model-4bit--1g"
|
| 312 |
|
| 313 |
use_triton = False
|
| 314 |
|
|
@@ -322,33 +328,9 @@ model = AutoGPTQForCausalLM.from_quantized(model_name_or_path,
|
|
| 322 |
use_triton=use_triton,
|
| 323 |
quantize_config=None)
|
| 324 |
|
| 325 |
-
|
| 326 |
-
|
| 327 |
-
|
| 328 |
-
output = model.generate(inputs=input_ids, temperature=0.7, max_new_tokens=512)
|
| 329 |
-
print(tokenizer.decode(output[0]))
|
| 330 |
-
|
| 331 |
-
# Inference can also be done using transformers' pipeline
|
| 332 |
-
|
| 333 |
-
# Prevent printing spurious transformers error when using pipeline with AutoGPTQ
|
| 334 |
-
logging.set_verbosity(logging.CRITICAL)
|
| 335 |
-
|
| 336 |
-
prompt = "Tell me about AI"
|
| 337 |
-
prompt_template=f'''### Human: {prompt}
|
| 338 |
-
### Assistant:'''
|
| 339 |
-
|
| 340 |
-
print("*** Pipeline:")
|
| 341 |
-
pipe = pipeline(
|
| 342 |
-
"text-generation",
|
| 343 |
-
model=model,
|
| 344 |
-
tokenizer=tokenizer,
|
| 345 |
-
max_new_tokens=512,
|
| 346 |
-
temperature=0.7,
|
| 347 |
-
top_p=0.95,
|
| 348 |
-
repetition_penalty=1.15
|
| 349 |
-
)
|
| 350 |
-
|
| 351 |
-
print(pipe(prompt_template)[0]['generated_text'])
|
| 352 |
```
|
| 353 |
|
| 354 |
## Provided files
|
|
@@ -361,7 +343,7 @@ It was created without group_size to lower VRAM requirements, and with --act-ord
|
|
| 361 |
|
| 362 |
* `gptq_model-4bit--1g.safetensors`
|
| 363 |
* Works with AutoGPTQ in CUDA or Triton modes.
|
| 364 |
-
*
|
| 365 |
* Works with text-generation-webui, including one-click-installers.
|
| 366 |
* Parameters: Groupsize = -1. Act Order / desc_act = True.
|
| 367 |
|
|
|
|
| 271 |
|
| 272 |
These files are GPTQ 4bit model files for [Bigcode's Starcoder](https://huggingface.co/bigcode/starcoder).
|
| 273 |
|
| 274 |
+
It is the result of quantising to 4bit using [AutoGPTQ](https://github.com/PanQiWei/AutoGPTQ).
|
| 275 |
|
| 276 |
## Repositories available
|
| 277 |
|
| 278 |
* [4-bit GPTQ models for GPU inference](https://huggingface.co/TheBloke/starcoder-GPTQ)
|
| 279 |
+
* [Bigcoder's unquantised fp16 model in pytorch format, for GPU inference and for further conversions](https://huggingface.co/bigcode/starcoder)
|
| 280 |
+
|
| 281 |
+
## Prompting
|
| 282 |
+
|
| 283 |
+
The model was trained on GitHub code.
|
| 284 |
+
|
| 285 |
+
As such it is _not_ an instruction model and commands like "Write a function that computes the square root." do not work well.
|
| 286 |
+
|
| 287 |
+
However, by using the [Tech Assistant prompt](https://huggingface.co/datasets/bigcode/ta-prompt) you can turn it into a capable technical assistant.
|
| 288 |
|
| 289 |
## How to easily download and use this model in text-generation-webui
|
| 290 |
|
|
|
|
| 315 |
import argparse
|
| 316 |
|
| 317 |
model_name_or_path = "TheBloke/starcoder-GPTQ"
|
|
|
|
| 318 |
|
| 319 |
use_triton = False
|
| 320 |
|
|
|
|
| 328 |
use_triton=use_triton,
|
| 329 |
quantize_config=None)
|
| 330 |
|
| 331 |
+
inputs = tokenizer.encode("def print_hello_world():", return_tensors="pt").to(device)
|
| 332 |
+
outputs = model.generate(inputs)
|
| 333 |
+
print(tokenizer.decode(outputs[0]))
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 334 |
```
|
| 335 |
|
| 336 |
## Provided files
|
|
|
|
| 343 |
|
| 344 |
* `gptq_model-4bit--1g.safetensors`
|
| 345 |
* Works with AutoGPTQ in CUDA or Triton modes.
|
| 346 |
+
* Does not work with GPTQ-for-LLaMa.
|
| 347 |
* Works with text-generation-webui, including one-click-installers.
|
| 348 |
* Parameters: Groupsize = -1. Act Order / desc_act = True.
|
| 349 |
|