| | --- |
| | license: apache-2.0 |
| | library_name: pytorch |
| | tags: |
| | - materials-science |
| | - crystal-structures |
| | - solid-state-batteries |
| | - representation-learning |
| | - screening |
| | model-index: |
| | - name: SSB Screening Model (RTX6000x2) |
| | results: |
| | - task: |
| | type: text-classification |
| | name: Screening Proxy (3-class) |
| | metrics: |
| | - type: accuracy |
| | value: 0.8118937 |
| | - type: f1 |
| | value: 0.8060277 |
| | - type: precision |
| | value: 0.7671543 |
| | - type: recall |
| | value: 0.8694215 |
| | - type: val_loss |
| | value: 0.2856999 |
| | --- |
| | |
| | # SSB Screening Model (RTX6000x2) |
| |
|
| | ## Model Summary |
| | This model is a lightweight MLP classifier trained on NPZ-encoded inorganic crystal structure features for solid-state battery (SSB) screening proxies. It is intended to prioritize candidate structures, not to replace DFT or experimental validation. |
| |
|
| | - **Architecture**: MLP (input_dim=144, hidden_dims=[512, 256, 128], dropout variable by sweep) |
| | - **Output**: 3-class classification proxy for screening tasks |
| | - **Training Regime**: supervised training on curated NPZ dataset with class-weighted loss |
| | - **Best checkpoint**: `checkpoint_epoch45.pt` (lowest observed val_loss in logs) |
| | |
| | ## Intended Use |
| | - **Primary**: ranking/prioritization of SSB electrolyte candidates |
| | - **Not intended**: absolute property prediction or experimental ground truth replacement |
| | |
| | ## Training Data |
| | - **Dataset**: `ssb_npz_v1` (curated NPZ features) |
| | - **Split**: 80/10/10 (train/val/test) |
| | - **Features**: composition + lattice + derived scalar statistics (144-dim) |
| | |
| | ## Evaluation |
| | Metrics from the latest run summary: |
| | - **Val loss**: 0.2857 |
| | - **Val accuracy**: 0.8119 |
| | - **Holdout accuracy**: 0.8096 |
| | - **F1**: 0.8060 |
| | - **Precision**: 0.7672 |
| | - **Recall**: 0.8694 |
| | |
| | ## Limitations |
| | - The model is a proxy classifier; it does not predict ground-truth physical properties. |
| | - Performance is tied to the training distribution of `ssb_npz_v1`. |
| | - Chemical regimes underrepresented in the training set may be poorly ranked. |
| | |
| | ## Training Configuration (abridged) |
| | - Optimizer: AdamW |
| | - LR: sweep (best around ~3e-4) |
| | - Weight decay: sweep (0.005–0.02) |
| | - Scheduler: cosine |
| | - Batch size: sweep (128–512) |
| | - Epochs: sweep (20–60) |
| | - Gradient accumulation: sweep (1–4) |
| | |
| | ## Citation |
| | If you use this model, please cite the dataset and training pipeline from the Nexa_compute repository. |
| |
|