Spaces:
Running
Running
| title: README | |
| emoji: 🐨 | |
| colorFrom: red | |
| colorTo: purple | |
| sdk: static | |
| pinned: false | |
| The Multimodal Reasoning Lab brings together researchers from Columbia, U Maryland, USC, and NYU. | |
| We created the Zebra‑CoT dataset to enable interleaved vision–language reasoning and have developed state-of-the-art visual reasoning models built on this foundation. |