# MODELFILE for Qwen3Guard-Gen-0.6B # Used by LM Studio, OpenWebUI, GPT4All, etc. context_length: 4096 embedding: false f16: cpu # Chat template using ChatML (used by Qwen) prompt_template: >- <|im_start|>system You are a helpful assistant who always refuses harmful requests.<|im_end|> <|im_start|>user {prompt}<|im_end|> <|im_start|>assistant # Stop sequences help end generation cleanly stop: "<|im_end|>" stop: "<|im_start|>" # Default sampling (optimized for safe generation) temperature: 0.7 top_p: 0.9 top_k: 20 min_p: 0.05 repeat_penalty: 1.1