Daily TMLR digest for Aug 09, 2025

0 views
Skip to first unread message

TMLR

unread,
Aug 9, 2025, 12:06:08 AMAug 9
to tmlr-anno...@googlegroups.com

Accepted papers
===============


Title: From Spikes to Heavy Tails: Unveiling the Spectral Evolution of Neural Networks

Authors: Vignesh Kothapalli, Tianyu Pang, Shenyang Deng, Zongmin Liu, Yaoqing Yang

Abstract: Training strategies for modern deep neural networks (NNs) tend to induce a heavy-tailed (HT) empirical spectral density (ESD) in the layer weights. While previous efforts have shown that the HT phenomenon correlates with good generalization in large NNs, a theoretical explanation of its occurrence is still lacking. Especially, understanding the conditions which lead to this phenomenon can shed light on the interplay between generalization and weight spectra. Our work aims to bridge this gap by presenting a simple, rich setting to model the emergence of HT ESD. In particular, we present a theory-informed setup for ‘crafting’ heavy tails in the ESD of two-layer NNs and present a systematic analysis of the HT ESD emergence without any gradient noise. This is the first work to analyze a noise-free setting, and we also incorporate optimizer (GD/Adam) dependent (large) learning rates into the HT ESD analysis. Our results highlight the role of learning rates on the Bulk+Spike and HT shape of the ESDs in the early phase of training, which can facilitate generalization in the two-layer NN. These observations shed light on the behavior of large-scale NNs, albeit in a much simpler setting.

URL: https://openreview.net/forum?id=DJHB8eBUnt

---


New submissions
===============


Title: Multimodal Deception in Explainable AI: Concept-Level Backdoor Attacks on Concept Bottleneck Models

Abstract: Deep learning has demonstrated transformative potential across domains, yet its inherent opacity has driven the development of Explainable Artificial Intelligence (XAI). Concept Bottleneck Models (CBMs), which enforce interpretability through human-understandable concepts, represent a prominent advancement in XAI. However, despite their semantic transparency, CBMs remain vulnerable to security threats such as backdoor attacks—malicious manipulations that induce controlled misbehaviors during inference. While CBMs leverage multimodal representations (visual inputs and textual concepts) to enhance interpretability, heir dual-modality structure introduces new attack surfaces. To address the unexplored risk of concept-level backdoor attacks in multimodal XAI systems, we propose CAT (Concept-level Backdoor ATtacks), a methodology that injects triggers into conceptual representations during training, enabling precise prediction manipulation without compromising clean-data performance. An enhanced variant, CAT+, incorporates a concept correlation function to systematically optimize trigger-concept associations, thereby improving attack effectiveness and stealthiness. Through a comprehensive evaluation framework assessing attack success rate, stealth metrics, and model utility preservation, we demonstrate that CAT and CAT+ maintain high performance on clean data while achieving significant targeted effects on backdoored datasets. This work highlights critical security risks in interpretable AI systems and provides a robust methodology for future security assessments of CBMs.

URL: https://openreview.net/forum?id=N8CTyY5FbR

---

Title: MTMT: Tiered Treatment Effect Decomposition for Multi-Task Uplift Modeling

Abstract: As a key component in boosting online user growth, uplift modeling aims to measure individual user responses (e.g., whether to play the game) to various treatments, such as gaming bonuses, thereby enhancing business outcomes. However, previous research typically considers a single-task, single-treatment setting, where only one treatment exists and the overall treatment effect is measured by a single type of user response. In this paper, we propose a Multi-Treatment Multi-Task (MTMT) uplift network to estimate treatment effects in a multi-task scenario. We identify the multi-treatment problem as a causal inference problem with a tiered response, comprising a base effect (from offering a treatment) and an incremental effect (from offering a specific type of treatment), where the base effect can be numerically much larger than the incremental effect. Specifically, MTMT separately encodes user features and treatments. The user feature encoder uses a multi-gate mixture of experts (MMOE) network to encode relevant user features, explicitly learning inter-task relations. The resultant embeddings are used to measure natural responses per task. Furthermore, we introduce a user-treatment feature interaction module to model correlations between each treatment and user feature. Consequently, we separately measure the base and incremental treatment effect for each task based on the produced treatment-aware representations. Experimental results based on an offline public dataset and an online proprietary dataset demonstrate the effectiveness of MTMT in single/multi-treatment and single/multi-task settings. Additionally, MTMT has been deployed in our gaming platform to improve user experience.

URL: https://openreview.net/forum?id=yo9NTZsGMa

---

Title: Unveiling Transfer Learning Effectiveness Through Latent Feature Distributions

Abstract: Transfer learning leverages large-scale pretraining to adapt models to specific downstream
tasks. It has emerged as a powerful and widely adopted training strategy in deep learning
frameworks. So, what makes it effective? Prior research has attributed its success to feature
reuse, pretrained weights reuse, domain alignment, and the transfer of low-level data statistics.
This study goes beyond these perspectives and focuses on a more fundamental factor:
the evolution of logits distribution within the latent feature space of pretrained models. We
introduce a novel approach using the Wasserstein distance to track distributional changes in
the latent features. We find that pretraining not only learns the input distributions but also
transforms them into generalizable internal representations in a consistent manner across
all frozen layers. This finding underpins the effectiveness of transfer learning and provides
a unifying explanation for these established theoretical perspectives.

URL: https://openreview.net/forum?id=2oqUNlfRmB

---

Reply all
Reply to author
Forward
0 new messages