|
|
|
|
|
"""Packed unary loader. Loads weights, passes pointers to C engine.""" |
|
|
import ctypes, os, sys, time, json |
|
|
import numpy as np |
|
|
from ctypes import c_int, c_float, c_void_p, POINTER, c_uint8, c_uint64 |
|
|
|
|
|
class PackedEngine: |
|
|
def __init__(self, model_dir, engine_path="./packed_engine.so"): |
|
|
self.lib = ctypes.CDLL(engine_path) |
|
|
self.lib.model_alloc.restype = c_void_p |
|
|
self.lib.forward_token.restype = POINTER(c_float) |
|
|
self.model_dir = model_dir |
|
|
|
|
|
with open(os.path.join(model_dir, "manifest.json")) as f: |
|
|
self.manifest = json.load(f) |
|
|
with open(os.path.join(model_dir, "config.json")) as f: |
|
|
self.config = json.load(f) |
|
|
|
|
|
self.arrays = [] |
|
|
self.model = self.lib.model_alloc() |
|
|
self._load_weights() |
|
|
|
|
|
def _keep(self, arr): |
|
|
self.arrays.append(arr) |
|
|
return arr.ctypes.data |
|
|
|
|
|
def _load_file(self, key, ext, dtype): |
|
|
path = os.path.join(self.model_dir, key.replace(".", "_") + ext) |
|
|
return np.fromfile(path, dtype=dtype) |
|
|
|
|
|
def _load_weights(self): |
|
|
t0 = time.time() |
|
|
fp16_keys = self.manifest["fp16"] |
|
|
packed_keys = self.manifest["packed"] |
|
|
|
|
|
|
|
|
emb = self._load_file("model.embed_tokens.weight", ".fp16", np.uint16) |
|
|
self.lib.model_set_embed(self.model, self._keep(emb)) |
|
|
print(f" Embeddings: {emb.nbytes/1e6:.1f} MB") |
|
|
|
|
|
|
|
|
lm = self._load_file("lm_head.weight", ".fp16", np.uint16) |
|
|
od, id_ = fp16_keys["lm_head.weight"] |
|
|
self.lib.model_set_lm_head(self.model, self._keep(lm), od, id_) |
|
|
print(f" LM head: {lm.nbytes/1e6:.1f} MB") |
|
|
|
|
|
|
|
|
fn = self._load_file("model.norm.weight", ".fp16", np.uint16).astype(np.float32) |
|
|
|
|
|
fn_f16 = self._load_file("model.norm.weight", ".fp16", np.float16) |
|
|
fn = fn_f16.astype(np.float32) |
|
|
self.lib.model_set_final_norm(self.model, self._keep(fn)) |
|
|
|
|
|
n_layers = self.config["num_hidden_layers"] |
|
|
for l in range(n_layers): |
|
|
pfx = f"model.layers.{l}" |
|
|
|
|
|
|
|
|
in_f16 = self._load_file(f"{pfx}.input_layernorm.weight", ".fp16", np.float16) |
|
|
pn_f16 = self._load_file(f"{pfx}.post_attention_layernorm.weight", ".fp16", np.float16) |
|
|
in_f = in_f16.astype(np.float32) |
|
|
pn_f = pn_f16.astype(np.float32) |
|
|
self.lib.layer_set_norms(self.model, l, self._keep(in_f), self._keep(pn_f)) |
|
|
|
|
|
|
|
|
qb = kb = vb = None |
|
|
qb_key = f"{pfx}.self_attn.q_proj.bias" |
|
|
if qb_key in fp16_keys: |
|
|
qb_f16 = self._load_file(qb_key, ".fp16", np.float16) |
|
|
qb = qb_f16.astype(np.float32) |
|
|
kb_f16 = self._load_file(f"{pfx}.self_attn.k_proj.bias", ".fp16", np.float16) |
|
|
kb = kb_f16.astype(np.float32) |
|
|
vb_f16 = self._load_file(f"{pfx}.self_attn.v_proj.bias", ".fp16", np.float16) |
|
|
vb = vb_f16.astype(np.float32) |
|
|
self.lib.layer_set_bias(self.model, l, |
|
|
self._keep(qb), self._keep(kb), self._keep(vb)) |
|
|
else: |
|
|
self.lib.layer_set_bias(self.model, l, None, None, None) |
|
|
|
|
|
|
|
|
args = [] |
|
|
for name in ['self_attn.q_proj','self_attn.k_proj','self_attn.v_proj', |
|
|
'self_attn.o_proj','mlp.gate_proj','mlp.up_proj','mlp.down_proj']: |
|
|
key = f"{pfx}.{name}.weight" |
|
|
shape = packed_keys[key] |
|
|
od, id_ = shape |
|
|
mags = self._load_file(key, ".mags", np.uint8) |
|
|
signs = self._load_file(key, ".signs", np.uint64) |
|
|
scales = self._load_file(key, ".scales", np.float32) |
|
|
rmm = self._load_file(key, ".rmm", np.uint8) |
|
|
args.extend([self._keep(mags), self._keep(signs), |
|
|
self._keep(scales), self._keep(rmm), od, id_]) |
|
|
|
|
|
self.lib.layer_set_linears(self.model, l, *args) |
|
|
|
|
|
if (l+1) % 7 == 0 or l == n_layers-1: |
|
|
print(f" Loaded {l+1}/{n_layers} layers") |
|
|
|
|
|
dt = time.time() - t0 |
|
|
total = sum(a.nbytes for a in self.arrays) |
|
|
print(f"\nModel loaded in {dt:.1f}s, {total/1e6:.0f} MB in Python arrays") |
|
|
|
|
|
def generate(self, token_ids, max_new_tokens=100, temperature=0.6, top_p=0.9, eos_id=151643): |
|
|
prompt = (c_int * len(token_ids))(*token_ids) |
|
|
output = (c_int * max_new_tokens)() |
|
|
self.lib.model_reset_cache(self.model) |
|
|
t0 = time.time() |
|
|
n = self.lib.generate(self.model, prompt, len(token_ids), |
|
|
output, max_new_tokens, c_float(temperature), |
|
|
c_float(top_p), eos_id) |
|
|
dt = time.time() - t0 |
|
|
tokens = [output[i] for i in range(n)] |
|
|
return tokens, n, dt |
|
|
|
|
|
|
|
|
if __name__ == "__main__": |
|
|
from transformers import AutoTokenizer |
|
|
model_dir = sys.argv[1] if len(sys.argv) > 1 else "deepseek-r1-1.5b-packed" |
|
|
tok_dir = sys.argv[2] if len(sys.argv) > 2 else "deepseek-r1-1.5b-hf" |
|
|
|
|
|
print("Loading tokenizer...") |
|
|
tok = AutoTokenizer.from_pretrained(tok_dir, trust_remote_code=True) |
|
|
print("Loading packed unary engine...") |
|
|
engine = PackedEngine(model_dir, "./packed_engine.so") |
|
|
|
|
|
prompts = ["What is 2+2?", "Explain gravity in one sentence.", "Write a haiku about snow."] |
|
|
for prompt in prompts: |
|
|
msgs = [{"role": "user", "content": prompt}] |
|
|
ids = tok.apply_chat_template(msgs, add_generation_prompt=True) |
|
|
tokens, n, dt = engine.generate(ids, max_new_tokens=100, temperature=0.6) |
|
|
text = tok.decode(tokens, skip_special_tokens=False) |
|
|
print(f"\n[{prompt}] ({n} tok, {dt:.1f}s, {n/dt:.1f} tok/s)") |
|
|
print(text[:300]) |
|
|
print("---") |
|
|
|