π Research-in-Progress (RiP): Monthly Data Science Seminar - Dr. Yanfu Zhang (William & Mary College) Notes
π Started at 2:56PM on 13 Nov, lasted 1h 9m
Key Points
- Avinash Sahu introduced Dr. Yanfu Zhang as an assistant professor in Computer Science at William & Mary College who received his Ph.D. in Electrical and Computer Engineering from UTD under Dr. Huang's supervision.
- Collaboration between the teams has produced several published papers focusing on heterogeneous graph neural networks and EHR genomic information using large language models.
- Graph representation learning addresses the challenge of transforming irregular graph data into readable formats for predictors, unlike regular structures in images and text that use convolution and attention mechanisms.
- Over-smoothing problem in graph convolutional networks occurs when multiple layers cause all nodes to have indistinguishable representations due to repeated low-pass filtering effects.
- Continuous graph neural networks using second-order dynamics were proposed to solve the over-smoothing limitation by incorporating velocity and acceleration terms similar to a spring-mass system.
- Second-order differential equations prevent nodes from overlapping in representation space, ensuring different final representations and avoiding the over-smoothing problem that affects first-order diffusion processes.
- Self-supervised learning approach using subgraph-level proximity was developed to handle unlabeled big data by comparing node similarities within sampled subgraphs rather than relying on manual first-order or second-order proximity definitions.
- Wasserstein distance computation enables end-to-end trainable optimization for comparing subgraph representations through differentiable convex optimization and KKT conditions.
- Brain connectome analysis leverages small-world property of neural networks to predict depression scores, with random graph generators serving as auxiliary tasks for graph neural network training.
- Graph-level representations for brain connectivity data showed improved performance when preserving small-world properties compared to fully connected graph approaches in depression prediction tasks.
- Deep metric learning for image retrieval was reformulated as a graph construction problem where images become nodes and similarity relationships form edges between same-category clusters.
- Unified loss function combining pairwise and proxy-based methods achieved superior performance across CUB, SOP, and In-Shop datasets by approximating pairwise loss while maintaining computational efficiency.
- Transformer architectures avoid over-smoothing in language processing because each input creates a different fully connected graph context, unlike fixed graph structures in traditional graph neural networks.
- Future research directions include integrating graph neural network principles with large language models and developing graph foundation models for single-cell data analysis applications.
Summary
Research-in-Progress Seminar Introduction and Background
AI Research Programs and Collaboration Opportunities
Representation Learning Fundamentals and Applications
Machine learning has achieved remarkable success across various domains including AlphaGo beating human champions, AlphaFold predicting protein structures, and ChatGPT demonstrating powerful language capabilities
- Common feature of successful approaches involves task-specific architectures where simple predictors work with complex representation learning components
- Benefits include allowing researchers to focus on specific representation learning problems rather than exploiting theoretical predictor structures
- Examples demonstrate how humans process visual information by creating representations that enable easy question answering about images.
Graph Neural Networks and Their Challenges
Graphs serve as powerful tools for describing diverse data forms including social networks, brain connections, protein interactions, and even images and text as special graph types
- Major applications include node classification, link prediction, and community detection for understanding complex relationships
- Irregularity presents the primary challenge as graphs lack the regular structure that enables convolution and attention mechanisms in images and text
- Combining neighbor information with node features while handling different numbers and types of edges creates significant algorithmic design difficulties.
Graph Convolution Networks and Over-smoothing Problem
Continuous Graph Neural Networks and Second-Order Dynamics
Self-Supervised Learning and Graph Contrastive Methods
Scalability Solutions and Wasserstein Distance Integration
Graph-Level Representation Learning for Brain Connectomes
Deep Metric Learning and Graph Construction
Integration with Large Language Models and Future Directions
Next Steps