rahul7star commited on
Commit
551f60a
·
verified ·
1 Parent(s): 8dcb874

Update app1.py

Browse files
Files changed (1) hide show
  1. app1.py +16 -9
app1.py CHANGED
@@ -33,16 +33,23 @@ pipe.to("cuda")
33
  # Load FP8 Text Encoder
34
  # ------------------------
35
  log("Loading FP8 text encoder...")
 
 
 
 
 
 
36
  config = AutoConfig.from_pretrained("jiangchengchengNLP/qwen3-4b-fp8-scaled")
37
- fp8_encoder = AutoModel.from_pretrained(
38
- "jiangchengchengNLP/qwen3-4b-fp8-scaled",
39
- config=config,
40
- device_map="auto",
41
- torch_dtype=torch.bfloat16,
42
- )
43
- pipe.text_encoder = fp8_encoder
44
- log("FP8 text encoder loaded successfully.")
45
- log(f"Tokenizer remains: {pipe.tokenizer.__class__.__name__}")
 
46
 
47
  # ------------------------
48
  # PIPELINE DEBUG INFO
 
33
  # Load FP8 Text Encoder
34
  # ------------------------
35
  log("Loading FP8 text encoder...")
36
+ from safetensors.torch import load_file
37
+ import torch
38
+ from transformers import AutoConfig
39
+ from qwen.modeling_qwen import QwenForCausalLM # ensure Qwen3 model class is imported
40
+
41
+ # Load config
42
  config = AutoConfig.from_pretrained("jiangchengchengNLP/qwen3-4b-fp8-scaled")
43
+
44
+ # Initialize model
45
+ fp8_encoder = QwenForCausalLM(config)
46
+
47
+ # Load weights manually
48
+ weights = load_file("qwen3_4b_fp8_scaled.safetensors")
49
+ fp8_encoder.load_state_dict(weights, strict=False)
50
+
51
+ fp8_encoder.to("cuda")
52
+
53
 
54
  # ------------------------
55
  # PIPELINE DEBUG INFO