Spaces:
Running
on
Zero
Running
on
Zero
Update app1.py
Browse files
app1.py
CHANGED
|
@@ -33,16 +33,23 @@ pipe.to("cuda")
|
|
| 33 |
# Load FP8 Text Encoder
|
| 34 |
# ------------------------
|
| 35 |
log("Loading FP8 text encoder...")
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 36 |
config = AutoConfig.from_pretrained("jiangchengchengNLP/qwen3-4b-fp8-scaled")
|
| 37 |
-
|
| 38 |
-
|
| 39 |
-
|
| 40 |
-
|
| 41 |
-
|
| 42 |
-
)
|
| 43 |
-
|
| 44 |
-
|
| 45 |
-
|
|
|
|
| 46 |
|
| 47 |
# ------------------------
|
| 48 |
# PIPELINE DEBUG INFO
|
|
|
|
| 33 |
# Load FP8 Text Encoder
|
| 34 |
# ------------------------
|
| 35 |
log("Loading FP8 text encoder...")
|
| 36 |
+
from safetensors.torch import load_file
|
| 37 |
+
import torch
|
| 38 |
+
from transformers import AutoConfig
|
| 39 |
+
from qwen.modeling_qwen import QwenForCausalLM # ensure Qwen3 model class is imported
|
| 40 |
+
|
| 41 |
+
# Load config
|
| 42 |
config = AutoConfig.from_pretrained("jiangchengchengNLP/qwen3-4b-fp8-scaled")
|
| 43 |
+
|
| 44 |
+
# Initialize model
|
| 45 |
+
fp8_encoder = QwenForCausalLM(config)
|
| 46 |
+
|
| 47 |
+
# Load weights manually
|
| 48 |
+
weights = load_file("qwen3_4b_fp8_scaled.safetensors")
|
| 49 |
+
fp8_encoder.load_state_dict(weights, strict=False)
|
| 50 |
+
|
| 51 |
+
fp8_encoder.to("cuda")
|
| 52 |
+
|
| 53 |
|
| 54 |
# ------------------------
|
| 55 |
# PIPELINE DEBUG INFO
|