Starkly Speaking: Leveraging Discrete Function Decomposability for Scientific Design
23 views
Skip to first unread message
Hannes Stärk
unread,
May 3, 2026, 7:14:05 PM (8 days ago) May 3
Reply to author
Sign in to reply to author
Forward
Sign in to forward
Delete
You do not have permission to delete messages in this group
Copy link
Report message
Show original message
Either email addresses are anonymous for this group or you need the view member email addresses permission to view the original message
to stark...@googlegroups.com
Hi together,
Tomorrow we will talk about:
Speaker: James Bowden who is a PhD student at Berkeley and NSF GRFP fellow, working at the intersection of ML and scientific design (proteins, materials). He is advised by Jennifer Listgarten and Sergey Levine. Lately he has been interested in understanding how domain-specific aspects of the scientific design problems he works on interact with various models and design algorithms. For instance, how might we enable experimentalists to more precisely specify the details of their problem settings and design desiderata?
Paper: Leveraging Discrete Function Decomposability for Scientific Design https://arxiv.org/abs/2511.03032 (James C. Bowden, Sergey Levine, Jennifer Listgarten) In the era of AI-driven science and engineering, we often want to design discrete objects in silico according to user-specified properties. For example, we may wish to design a protein to bind its target, arrange components within a circuit to minimize latency, or find materials with certain properties. Given a property predictive model, in silico design typically involves training a generative model over the design space (e.g., protein sequence space) to concentrate on designs with the desired properties. Distributional optimization–which can be formalized as an estimation of distribution algorithm or as reinforcement learning policy optimization–finds the generative model that maximizes an objective function in expectation. Optimizing a distribution over discrete-valued designs is in general challenging because of the combinatorial nature of the design space. However, many property predictors in scientific applications are decomposable in the sense that they can be factorized over design variables in a way that could in principle enable more effective optimization. For example, amino acids at a catalytic site of a protein may only loosely interact with amino acids of the rest of the protein to achieve maximal catalytic activity. Current distributional optimization algorithms are unable to make use of such decomposability structure. Herein, we propose and demonstrate use of a new distributional optimization algorithm, Decomposition-Aware Distributional Optimization (DADO), that can leverage any decomposability defined by a junction tree on the design variables, to make optimization more efficient. At its core, DADO employs a soft-factorized "search distribution"–a learned generative model–for efficient navigation of the search space, invoking graph message-passing to coordinate optimization across linked factors.