Update README.md
Browse files
README.md
CHANGED
|
@@ -3,10 +3,12 @@ library_name: transformers
|
|
| 3 |
tags: []
|
| 4 |
---
|
| 5 |
|
| 6 |
-
# ***Mol-MoE***:
|
| 7 |
*Diego Calanzone (1, 2), Pierluca D'Oro (2), Pierre-Luc Bacon (1, 2)* <br>
|
| 8 |
*(1) Universite de Montreal, (2) Mila Quebec AI Institute* <br>
|
| 9 |
|
|
|
|
|
|
|
| 10 |
|
| 11 |
## How to use this model
|
| 12 |
This LM is fine-tuned to generate molecules in the SMILES format wrt. desired properties.
|
|
|
|
| 3 |
tags: []
|
| 4 |
---
|
| 5 |
|
| 6 |
+
# ***Mol-MoE***: Training Preference-Guided Routers for Molecule Generation
|
| 7 |
*Diego Calanzone (1, 2), Pierluca D'Oro (2), Pierre-Luc Bacon (1, 2)* <br>
|
| 8 |
*(1) Universite de Montreal, (2) Mila Quebec AI Institute* <br>
|
| 9 |
|
| 10 |
+
**Abstract**: Recent advances in language models have enabled framing molecule generation as sequence modeling. However, existing approaches often rely on single-objective reinforcement learning, limiting their applicability to real-world drug design, where multiple competing properties must be optimized. Traditional multi-objective reinforcement learning (MORL) methods require costly retraining for each new objective combination, making rapid exploration of trade-offs impractical. To overcome these limitations, we introduce Mol-MoE, a mixture-of-experts (MoE) architecture that enables efficient test-time steering of molecule generation without retraining. Central to our approach is a preference-based router training objective that incentivizes the router to combine experts in a way that aligns with user-specified trade-offs. This provides improved flexibility in exploring the chemical property space at test time, facilitating rapid trade-off exploration. Benchmarking against state-of-the-art methods, we show that Mol-MoE achieves superior sample quality and steerability.
|
| 11 |
+
|
| 12 |
|
| 13 |
## How to use this model
|
| 14 |
This LM is fine-tuned to generate molecules in the SMILES format wrt. desired properties.
|