yushaohan commited on
Commit
556c2ea
·
verified ·
1 Parent(s): ce35302

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +24 -6
README.md CHANGED
@@ -1,21 +1,33 @@
1
  ---
 
 
2
  datasets:
3
  - yushaohan/ProGuard-data
4
  language:
5
  - en
6
- base_model:
7
- - Qwen/Qwen2.5-VL-3B-Instruct
8
  tags:
9
  - vlm
10
  - safety
11
  - guard
 
 
12
  ---
13
 
14
- This model is the official open-source implementation of **[ProGuard: Towards Proactive Multimodal Safeguard](https://arxiv.org/abs/2512.23573)**. For deployment instructions, please refer to **[this link](https://github.com/yushaohan/ProGuard/tree/master/deploy)**.
 
 
15
 
16
- If you find this model helpful, please cite our paper 🤗
 
 
17
 
18
- ```
 
 
 
 
 
 
19
  @article{yu2025proguard,
20
  title={ProGuard: Towards Proactive Multimodal Safeguard},
21
  author={Yu, Shaohan and Li, Lijun and Si, Chenyang and Sheng, Lu and Shao, Jing},
@@ -23,5 +35,11 @@ If you find this model helpful, please cite our paper 🤗
23
  year={2025},
24
  url={https://yushaohan.github.io/ProGuard/}
25
  }
26
- ```
27
 
 
 
 
 
 
 
 
 
1
  ---
2
+ base_model:
3
+ - Qwen/Qwen2.5-VL-3B-Instruct
4
  datasets:
5
  - yushaohan/ProGuard-data
6
  language:
7
  - en
 
 
8
  tags:
9
  - vlm
10
  - safety
11
  - guard
12
+ library_name: transformers
13
+ pipeline_tag: image-text-to-text
14
  ---
15
 
16
+ # ProGuard-3B
17
+
18
+ ProGuard is a proactive multimodal safeguard model. It is designed to identify and reason about unknown risks across both text and visual modalities, moving beyond rigid predefined classification systems.
19
 
20
+ - **Arxiv Paper:** [ProGuard: Towards Proactive Multimodal Safeguard](https://arxiv.org/abs/2512.23573)
21
+ - **Project Page:** [ProGuard Homepage](https://yushaohan.github.io/ProGuard/)
22
+ - **GitHub Repository:** [ProGuard Implementation](https://github.com/yushaohan/ProGuard), [DeepSafe Implementation](https://github.com/AI45Lab/DeepSafe)
23
 
24
+ This model is the official open-source implementation of **ProGuard**. For deployment instructions, please refer to **[this link](https://github.com/yushaohan/ProGuard/tree/master/deploy)**.
25
+
26
+ ## Citation
27
+
28
+ If you find this model helpful, please cite our research:
29
+
30
+ ```bibtex
31
  @article{yu2025proguard,
32
  title={ProGuard: Towards Proactive Multimodal Safeguard},
33
  author={Yu, Shaohan and Li, Lijun and Si, Chenyang and Sheng, Lu and Shao, Jing},
 
35
  year={2025},
36
  url={https://yushaohan.github.io/ProGuard/}
37
  }
 
38
 
39
+ @article{zhang2026deepsight,
40
+ title={DeepSight: An All-in-One LM Safety Toolkit},
41
+ author={Zhang, Bo and Guo, Jiaxuan and Li, Lijun and Liu, Dongrui and Chen, Sujin and Chen, Guanxu and Zheng, Zhijie and Lin, Qihao and Yan, Lewen and Qian, Chen and others},
42
+ journal={arXiv preprint arXiv:2602.12092},
43
+ year={2026}
44
+ }
45
+ ```