Add task category, paper & code links, full abstract and tags

#2
by nielsr HF Staff - opened
Files changed (1) hide show
  1. README.md +11 -3
README.md CHANGED
@@ -1,5 +1,12 @@
1
  ---
2
  license: mit
 
 
 
 
 
 
 
3
  dataset_info:
4
  features:
5
  - name: edge_index
@@ -53,14 +60,15 @@ configs:
53
 
54
  # MuMo Pretraining Dataset
55
 
56
- - 📄 Paper: [NeurIPS 2025 Poster](https://neurips.cc/virtual/2025/poster/119127)
 
57
  - 📬 Contact:
58
  - Zihao Jing: [email protected] | Wechat: A2016A315214 | Instagram: nobeljing25
59
  - Pingzhao Hu: [email protected]
60
 
61
  ## Abstract
62
 
63
- Multimodal molecular models often suffer from 3D conformer unreliability and modality collapse, limiting robustness and generalization. MuMo addresses these challenges with a Structured Fusion Pipeline (SFP) that combines 2D topology and 3D geometry into a stable structural prior, and a Progressive Injection (PI) mechanism that asymmetrically integrates this prior into the sequence stream. Built on a state space backbone, MuMo supports long-range dependency modeling. Across 21+ benchmarks, MuMo achieves strong improvements and robustness to 3D conformer noise. See paper for details: [NeurIPS 2025](https://neurips.cc/virtual/2025/poster/119127).
64
 
65
  ## Dataset Overview
66
 
@@ -127,7 +135,7 @@ We use `preprocess/mol3d_processor.py` to derive graph and geometry features fro
127
 
128
  If you find this work useful, please cite:
129
 
130
- Zihao Jing, Yan Sun, Yanyi Li, Sugitha Janarthanan, Alana Deng, and Pingzhao Hu. "MuMo: Multimodal Molecular Representation Learning via Structural Fusion and Progressive Injection." In Advances in Neural Information Processing Systems (NeurIPS), 2025. ([paper](https://neurips.cc/virtual/2025/poster/119127))
131
 
132
  ```bibtex
133
  @inproceedings{jing2025mumo,
 
1
  ---
2
  license: mit
3
+ task_categories:
4
+ - graph-ml
5
+ tags:
6
+ - chemistry
7
+ - molecular-biology
8
+ - drug-discovery
9
+ - multi-modal
10
  dataset_info:
11
  features:
12
  - name: edge_index
 
60
 
61
  # MuMo Pretraining Dataset
62
 
63
+ - 📄 Paper: [Structure-Aware Fusion with Progressive Injection for Multimodal Molecular Representation Learning](https://huggingface.co/papers/2510.23640)
64
+ - 💻 Code: [https://github.com/selmiss/MuMo](https://github.com/selmiss/MuMo)
65
  - 📬 Contact:
66
  - Zihao Jing: [email protected] | Wechat: A2016A315214 | Instagram: nobeljing25
67
  - Pingzhao Hu: [email protected]
68
 
69
  ## Abstract
70
 
71
+ Multimodal molecular models often suffer from 3D conformer unreliability and modality collapse, limiting their robustness and generalization. We propose MuMo, a structured multimodal fusion framework that addresses these challenges in molecular representation through two key strategies. To reduce the instability of conformer-dependent fusion, we design a Structured Fusion Pipeline (SFP) that combines 2D topology and 3D geometry into a unified and stable structural prior. To mitigate modality collapse caused by naive fusion, we introduce a Progressive Injection (PI) mechanism that asymmetrically integrates this prior into the sequence stream, preserving modality-specific modeling while enabling cross-modal enrichment. Built on a state space backbone, MuMo supports long-range dependency modeling and robust information propagation. Across 29 benchmark tasks from Therapeutics Data Commons (TDC) and MoleculeNet, MuMo achieves an average improvement of 2.7% over the best-performing baseline on each task, ranking first on 22 of them, including a 27% improvement on the LD50 task. These results validate its robustness to 3D conformer noise and the effectiveness of multimodal fusion in molecular representation. The code is available at: this http URL .
72
 
73
  ## Dataset Overview
74
 
 
135
 
136
  If you find this work useful, please cite:
137
 
138
+ Zihao Jing, Yan Sun, Yanyi Li, Sugitha Janarthanan, Alana Deng, and Pingzhao Hu. "MuMo: Multimodal Molecular Representation Learning via Structural Fusion and Progressive Injection." In Advances in Neural Information Processing Systems (NeurIPS), 2025. ([paper](https://huggingface.co/papers/2510.23640))
139
 
140
  ```bibtex
141
  @inproceedings{jing2025mumo,