Mohaddz commited on
Commit
264168f
·
verified ·
1 Parent(s): 47f7d3b

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +5 -2
app.py CHANGED
@@ -5,8 +5,11 @@ import gradio as gr
5
  from transformers import AutoModelForCausalLM, AutoTokenizer
6
  from huggingface_hub import snapshot_download
7
  from dotenv import load_dotenv
 
 
8
  load_dotenv()
9
 
 
10
  # Check if CUDA is available
11
  device = "cuda" if torch.cuda.is_available() else "cpu"
12
 
@@ -38,9 +41,9 @@ snapshot_download(
38
  ]
39
  )
40
 
41
- model = AutoModelForCausalLM.from_pretrained(model_name, torch_dtype=torch.bfloat16)
42
  model.to(device)
43
- tokenizer = AutoTokenizer.from_pretrained(model_name)
44
  print(f"Orpheus model loaded to {device}")
45
 
46
  # Process text prompt
 
5
  from transformers import AutoModelForCausalLM, AutoTokenizer
6
  from huggingface_hub import snapshot_download
7
  from dotenv import load_dotenv
8
+ import os
9
+ token_hf = os.getenv('HF_TOKEN')
10
  load_dotenv()
11
 
12
+
13
  # Check if CUDA is available
14
  device = "cuda" if torch.cuda.is_available() else "cpu"
15
 
 
41
  ]
42
  )
43
 
44
+ model = AutoModelForCausalLM.from_pretrained(model_name, torch_dtype=torch.bfloat16, token=token_hf)
45
  model.to(device)
46
+ tokenizer = AutoTokenizer.from_pretrained(model_name, token_hf)
47
  print(f"Orpheus model loaded to {device}")
48
 
49
  # Process text prompt