VaibhavHD commited on
Commit
5352ede
·
verified ·
1 Parent(s): 14aa92c

Update inference.py

Browse files
Files changed (1) hide show
  1. inference.py +6 -2
inference.py CHANGED
@@ -1,15 +1,19 @@
1
  from transformers import AutoTokenizer, AutoModelForCausalLM
2
  from peft import PeftModel
 
 
 
 
 
3
 
4
  BASE_MODEL = "deepseek-ai/deepseek-coder-1.3b-base"
5
  LORA_REPO = "VaibhavHD/deepseek-lora-monthly"
6
 
7
-
8
  tokenizer = AutoTokenizer.from_pretrained(BASE_MODEL, trust_remote_code=True)
9
  base = AutoModelForCausalLM.from_pretrained(BASE_MODEL, trust_remote_code=True)
10
  model = PeftModel.from_pretrained(base, LORA_REPO)
11
 
12
- def generate_response(prompt:str)->str:
13
  inputs = tokenizer(prompt, return_tensors="pt")
14
  out = model.generate(**inputs, max_new_tokens=200)
15
  return tokenizer.decode(out[0], skip_special_tokens=True)
 
1
  from transformers import AutoTokenizer, AutoModelForCausalLM
2
  from peft import PeftModel
3
+ import torch
4
+ import transformers.training_args
5
+
6
+ # ✅ Fix: allow this class for safe loading in PyTorch 2.6+
7
+ torch.serialization.add_safe_globals([transformers.training_args.TrainingArguments])
8
 
9
  BASE_MODEL = "deepseek-ai/deepseek-coder-1.3b-base"
10
  LORA_REPO = "VaibhavHD/deepseek-lora-monthly"
11
 
 
12
  tokenizer = AutoTokenizer.from_pretrained(BASE_MODEL, trust_remote_code=True)
13
  base = AutoModelForCausalLM.from_pretrained(BASE_MODEL, trust_remote_code=True)
14
  model = PeftModel.from_pretrained(base, LORA_REPO)
15
 
16
+ def generate_response(prompt: str) -> str:
17
  inputs = tokenizer(prompt, return_tensors="pt")
18
  out = model.generate(**inputs, max_new_tokens=200)
19
  return tokenizer.decode(out[0], skip_special_tokens=True)