How do we achieve interpretability in generative models for drug design? Disentangle!

Generative models for molecules are hard to interpret – their latent spaces are typically overly convoluted. Learning disentangled representations is important for unraveling the underlying

complex interactions between latent generative factors pertaining to different molecular properties of interest (e.g., ADMET properties). For example, we might want to improve the metabolism property (‘M’ in ADMET) of a molecule without significantly altering its other properties. 

However, not all properties can be disentangled – e.g., usually it’s hard to optimize or improve bioactivity of a candidate molecule without enhancing its toxicity as well. In a NeurIPS 2022 paper, we address this issue with a novel concept of conditional disentanglement: our method  segregates the latent space into uncoupled and entangled parts. This allows us to exercise finer control over molecule generation and optimization. Experimental results on molecular data strongly corroborate the interpretability of our method. Interestingly, this interpretability is accompanied with improved molecular generation performance, as evaluated across multiple standard metrics.

Innovations

Successful Business Evolution through AI & Quantum Tech

Discovering new promising molecule candidates that could translate into effective...

State-of-the-art approaches for structure-based drug design (SBDD) use extremely complex...

Unlocking novel generative capabilities with diffusion via “loop guidance”! In...