Could Leo use hypergraphs?

64 views
Skip to first unread message

Edward K. Ream

unread,
Jun 10, 2025, 8:32:32 AM6/10/25
to leo-editor

Could Leo use hypergraphs? What do you think?

Edward

Edward K. Ream

unread,
Jun 10, 2025, 11:01:38 AM6/10/25
to leo-editor
On Tuesday, June 10, 2025 at 7:32:32 AM UTC-5 Edward K. Ream wrote:


I should acknowledge the source of the links above. They come from this YouTube video from Anastasiia Nosova.  Her YouTube channel is here.

The 5-minute mark of the video starts her interview with Pushmeet Kohli, along with overlaid video of the Gemini LLM in action evolving the initial template. It's worth a look, imo. It's brilliantly clear explanation. Notice the references (in playback!) to the python tools jax and jaxline. Wow.

The references to Dr. Kohli are also the highlight of my first AI conversation.

Edward

Edward K. Ream

unread,
Jun 10, 2025, 11:33:52 AM6/10/25
to leo-editor
On Tuesday, June 10, 2025 at 10:01:38 AM UTC-5 Edward K. Ream wrote:

Some highlights of the AlphaChip paper:

Training placement policies that generalize across chips is extremely challenging, because it requires learning to optimize the placement of all possible chip netlists onto all possible canvases. Chip floorplanning is analogous to a game with [a state space] of 1000 [factorial] (greater than 10**2,500), whereas Go has a state space of 10**360.

The underlying problem is a high-dimensional contextual bandits problem [not to be confused with a multi-armed bandit problem] but, as in prior work, ... we have chosen to reformulate it as a sequential Markov decision process (MDP), because this allows us to more easily incorporate the problem constraints as described below.

To address [the challenge of an enormous state space] we first focused on learning rich representations of the state space. Our intuition was that a policy capable of the general task of chip placement should also be able to encode the state associated with a new unseen chip into a meaningful signal at inference time. We therefore trained a neural network architecture capable of predicting reward on placements of new netlists, with the ultimate goal of using this architecture as the encoder layer of our policy.

Seems like brilliant science and mathematics to me :-)

Edward
Reply all
Reply to author
Forward
0 new messages