Wie benutze ich das quantitative AWQ -Klassifizierungsmodell?Python

Python-Programme
Guest
 Wie benutze ich das quantitative AWQ -Klassifizierungsmodell?

Post by Guest »

Wenn ich ein Klassifizierungsmodell habe, das auf QWEN2.5-0.5B-Training basiert: < /p>

Code: Select all

from transformers import AutoTokenizer, AutoModelForSequenceClassification

tokenizer = AutoTokenizer.from_pretrained("Qwen/Qwen2.5-0.5B")
model = AutoModelForSequenceClassification.from_pretrained(
"Qwen/Qwen2.5-0.5B",
device_map="auto",
num_labels=2,
torch_dtype=torch.bfloat16,
)
Wie quantifiziere ich es für AWQ und kalibrieren Sie es? p>

Code: Select all

from transformers import AutoTokenizer, AutoModelForSequenceClassification

tokenizer = AutoTokenizer.from_pretrained("Qwen/Qwen2.5-0.5B")
model = AutoModelForSequenceClassification.from_pretrained(
"Qwen/Qwen2.5-0.5B",
device_map="auto",
num_labels=2,
torch_dtype=torch.bfloat16,
)
model.save_pretrained(model_path)
tokenizer.save_pretrained(model_path)
```python

The conversion was then performed using AutoAwq, but it was found that the head layer changed from score to lm_head after quantisation (the model architecture was changed).

```python
from awq import AutoAWQForCausalLM
from transformers import AutoTokenizer, AutoModelForSequenceClassification
from transformers import AwqConfig, AutoConfig
import torch
quant_config = {"zero_point": True, "q_group_size": 128, "w_bit": 4, "version": "GEMM" }

tokenizer = AutoTokenizer.from_pretrained(model_path, trust_remote_code=True)
model = AutoAWQForCausalLM.from_pretrained(model_path, device_map="auto", safetensors=True)

model.quantize(tokenizer, quant_config=quant_config, calib_data=data)
model.save_quantized(quant_path, safetensors=True, shard_size="4GB")
tokenizer.save_pretrained(quant_path)

Quick Reply

Change Text Case: 
   
  • Similar Topics
    Replies
    Views
    Last post