Jyhan003 commited on
Commit
65a00e3
·
1 Parent(s): 0eeef3f

Updated README to clarify: code under MIT, weights under CC BY-NC 4.0

Browse files
Files changed (1) hide show
  1. README.md +6 -2
README.md CHANGED
@@ -1,5 +1,5 @@
1
  ---
2
- license: mit
3
  library_name: transformers
4
  pipeline_tag: voice-activity-detection
5
  tags:
@@ -16,7 +16,7 @@ tags:
16
  ## Overview
17
  This hub features the pre-trained model by [DiariZen](https://github.com/BUTSpeechFIT/DiariZen). The EEND component is built upon WavLM Large and Conformer layers. The model was trained on far-field, single-channel audio from a diverse set of public datasets, including AMI, AISHELL-4, AliMeeting, NOTSOFAR-1, MSDWild, DIHARD3, RAMC, and VoxConverse.
18
 
19
- Then structured pruning at 80% sparsity is applied. After pruning, the number of parameters in WavLM Large is reduced from **316.6M to 63.3M**, and the computational cost (MACs) decreases from **17.8G to 3.8G** per second.
20
 
21
 
22
 
@@ -79,3 +79,7 @@ If you found this work helpful, please consider citing:
79
  }
80
  ```
81
 
 
 
 
 
 
1
  ---
2
+ license: cc-by-nc-4.0
3
  library_name: transformers
4
  pipeline_tag: voice-activity-detection
5
  tags:
 
16
  ## Overview
17
  This hub features the pre-trained model by [DiariZen](https://github.com/BUTSpeechFIT/DiariZen). The EEND component is built upon WavLM Large and Conformer layers. The model was trained on far-field, single-channel audio from a diverse set of public datasets, including AMI, AISHELL-4, AliMeeting, NOTSOFAR-1, MSDWild, DIHARD3, RAMC, and VoxConverse.
18
 
19
+ Then structured pruning at 80% sparsity is applied. After pruning, the number of parameters in WavLM Large is reduced from **316.6M to 63.3M**, and the computational cost (MACs) decreases from **17.8G to 3.8G** per second. When loading this model, please ensure **non-commercial** usage, in accordance with the CC BY-NC 4.0 license.
20
 
21
 
22
 
 
79
  }
80
  ```
81
 
82
+ ## License
83
+ - **Source code**: MIT (see the [project’s GitHub repository](https://github.com/BUTSpeechFIT/DiariZen)).
84
+ - **Model weights**: CC BY-NC 4.0 (non-commercial).
85
+ - Rationale: some training datasets are research-only or non-commercial, so the released weights cannot be used commercially.