| Day-1 Saturday - 22nd November 2025 | ||
|---|---|---|
| 08:00 - 09:00 - Breakfast | ||
| Morning Session: 09:30 - 13:00, Venue: CSA 104 | ||
| Session Chair: Sumit Kumar Mandal (IISc, Bangalore) | ||
| Time Slot | Title of the Talk | Speaker (Affiliation) |
| 09:30 - 10:30 |
Federated Learning, Robustness and Incentives
Federated Learning (FL) enables collaborative model training without centralizing data, but its distributed nature introduces new challenges. In this talk, I will discuss two key aspects: robustness and incentives. Robustness focuses on ensuring reliable learning under data heterogeneity and adversarial updates. Incentive design addresses how to motivate self-interested or strategic clients to contribute high-quality data and computation truthfully. I will outline recent approaches to each and reflect on how integrating the two can lead to more stable, fair, and trustworthy federated systems.
|
Manisha Padala (IIT Gandhinagar) |
| 10:30 - 11:00 | Coffee Break | |
| 11:00 - 12:00 |
Old dog, Old tricks, New show: Fast 1st order methods for training Kernel Machines
Kernel Machines are a classical family of models in Machine Learning that overcome several limitations of Neural Networks. These models have regained popularity following some landmark results showing their equivalence to Neural Networks. We propose a state of the art algorithm - EigenPro - based on gradient descent in the RKHS. This algorithm is much faster and requires less memory compared to previous attempts, and enables training large scale Kernel Machines over large datasets.
|
Parthe Pandit (IIT Bombay) |
| 12:00 - 13:00 |
Statistical Inference for Stochastic Gradient Descent
The stochastic gradient descent (SGD) algorithm is used for parameter estimation, particularly for massive datasets and online learning. Inference in SGD has been a generally neglected problem and has only recently started to get some attention. I will first introduce SGD for relatively simple statistical models and explain the limiting behavior of Averaged SGD. Then, I will present a memory-reduced batch-means estimator of the limiting covariance matrix that is both consistent and amenable to finite-sample corrections. Further, I will discuss the practical usability of error covariance matrices for problems where SGD is relevant, and present ongoing challenges in this area.
|
Dootika Vats (IIT Kanpur) |
| 13:00 - 14:00 | Lunch Break | |
| Afternoon Session: 14:00 - 16:30, Venue: CSA 104 | ||
| Session Chair: Dootika Vats (IIT Kanpur) | ||
| 14:00 - 14:50 |
Towards practical and provable privacy preservation in the age of foundation models
While foundation models and LLMs unlock new and unprecedented AI capabilities, they come with a substantially increased risk of memorising, regurgitating, and leaking privacy-sensitive data. Differential privacy, now a well-established standard for privacy protection, provides a principled solution to prevent such leakage, but is often computationally expensive (for good performance). I'll present some of our work on developing efficient and scalable algorithms AI inference and fine-tuning to make differential privacy practical in the era of foundation models. First, I will describe our method for generating artificial text with LLMs that is statistically similar to real data while preserving privacy. Our algorithm reduces the computational overhead of differential privacy from roughly 100-1000x in prior work to about 4x, making deployment feasible at scale. Next, I will discuss fine-tuning with differential privacy, where we build on a recent approach that injects correlated Gaussian noise across stochastic gradient steps. Our variant reduces the time complexity from quadratic to nearly linear, while maintaining comparable accuracy and privacy guarantees. I will conclude with a brief outlook on ongoing and future directions.
|
Krishna Pillutla (IIT Madras) |
| 14:50 - 15:40 |
Towards safety-aware autonomous navigation
Safe autonomous navigation lies at the heart of deploying robots in the real world. Whether guiding self-driving cars through traffic or mobile robots through dynamic indoor spaces, the challenge is the same: perceive the world, predict how it might evolve, and choose actions that avoid harm while still making progress toward a goal. This demands robustness to uncertainty, an understanding of environmental affordances, and principled ways to reason about risk. In this talk, I will present some of our ongoing work on safe reinforcement learning and eliciting safety guidance from foundation models for safe autonomous navigation.
|
Raunak Bhattacharyya (IIT Delhi) |
| 15:40 - 16:30 |
Exploration via Optimal Transport
Computation of optimal exploration strategies for sequential learning problems such as bandits and RL with large action and policy spaces poses significant challenges. While there exist approaches for optimal exploration with finite action sets, much less is known about techniques to tackle the continuous case. In this paper, we take up the goal of algorithms for pure exploration in linear bandits with continuous arm sets. A key challenge in this setting is to determine an asymptotically optimal sampling distribution over the arm space, which requires solving a minimax problem arising from information-theoretic considerations. This problem is a non-smooth objective over the (infinite dimensional) space of probability measures on the continuous action space. Using the framework of optimal transport over the Wasserstein space of probability measures over arms, we introduce an idealized subgradient descent method operating directly in this space. Under mild conditions, we prove a local convergence result, demonstrating that our method converges to an asymptotically optimal sampling distribution when initialized appropriately. The Wasserstein geometry allows us to naturally develop a computationally feasible mean-field particle approximation of this new exploration method. Empirical evaluation on synthetic datasets shows promising preliminary results.
|
Aditya Gopalan (IISc, Bangalore) |
| 16:30 - Onwards | High Tea | |
| Day-2 Sunday - 23rd November 2025 | ||
|---|---|---|
| 08:00 - 09:00 - Breakfast | ||
| Morning Session: 09:00 - 13:00, Venue: CSA 104 | ||
| Session Chair: Anant Raj (IISc, Bangalore) | ||
| Time Slot | Title of the Talk | Speaker (Affiliation) |
| 09:00 - 09:55 |
Reimagining Computing through Systems that Drive AI and AI that Shapes Systems
Artificial intelligence has not only transformed applications but has also redefined the design principles of modern computing systems. This presentation explores the emerging co-evolution of systems that drive AI and AI that shapes systems. It first discusses system-level innovations in AI accelerators, including dataflow architectures, memory hierarchies, and interconnect optimizations that enable scalable and energy-efficient model execution. In parallel, it examines the use of AI techniques to enhance system intelligence, with particular emphasis on learning-guided cache management and adaptive memory control. By integrating predictive models into runtime and architectural decisions, computing platforms can dynamically adapt to diverse workloads and access patterns, leading to more efficient and intelligent system behavior.
|
Palash Das (IIT Jodhpur) |
| 09:55 - 10:50 |
Machine Learning for Electronic Design Automation
The increasing complexity of integrated circuit (IC) design has made traditional Electronic Design Automation (EDA) approaches computationally intensive and time-consuming. Recent advances in Machine Learning (ML) are transforming the EDA landscape by enabling data-driven optimization, faster design closure, and improved accuracy in prediction-based tasks. This talk will provide an overview of how ML techniques are being integrated into various stages of the VLSI design flow-such as synthesis, placement, routing, and verification. Emphasis will be placed on key challenges, successful use cases, and future research directions where ML can significantly enhance design productivity and innovation in EDA.
|
Chandan Karfa (IIT Guwahati) |
| 10:50 - 11:10 | Coffee Break | |
| 11:10 - 12:05 |
From Static to Strategic: Rethinking Testing and Benchmarking in Data Systems
Data systems are the backbone of modern data-driven computing, powering everything from decision-making processes to critical enterprise applications. However, the current landscape of testing and benchmarking these systems is plagued by a lack of automation and an inability to effectively evaluate systems in real-world customer deployments. This talk focuses on bridging the gap between synthetic benchmarks and real-world performance evaluations, moving beyond static methods toward adaptive and strategic approaches. Specifically, we will discuss a dynamic data generation approach that leverages query execution plans from customer deployments to synthesize data. This enables the creation of synthetic data that replicates customer query processing environments, allowing for more realistic system evaluations. Additionally, we will present our evaluation study examining how effectively language models-a cornerstone of the impending AI-driven data systems - understand and process enterprise data. The study highlights the challenges these models face when transitioning from general-purpose public datasets to the complexity of enterprise data. Together, these efforts contribute to the development of a smarter, automated testing and benchmarking paradigm, essential for ensuring the reliability and robustness of data systems.
|
Anupam Sanghi (IIT Hyderabad) |
| 12:05 - 13:00 |
Learning from Human Feedback
Learning from human feedback is an essential component of many AI systems. For example, recommendation systems rely on human feedback for curating high-quality personalized content. Large language models (LLMs) rely on human feedback for fine-tuning models to better align with human intent and values. In this talk, I will provide a brief introduction to the field of learning from human feedback and discuss my work on efficient learning from pairwise feedback.
|
Arpit Agarwal (IIT Bombay) |
| 13:00 - 14:00 | Lunch Break | |
| Afternoon Session: 14:00 - 16:30, Venue: CSA 104 | ||
| Session Chair: Chandan Karfa (IIT Guwahati) | ||
| 14:00 - 15:00 |
Harnessing Generative Intelligence for Healthcare: Models, Methods and Evaluations
Generative AI, especially Large Language Models (LLMs) and Multimodal Language Models (MLMs), is creating exciting opportunities in healthcare. However, real-world use is still challenging due to the need for models that are compact, personalized, safe, and capable of handling multiple languages and data types. This work tackles these challenges in three main directions: building specialized models, developing advanced methods for key healthcare tasks, and creating strong evaluation benchmarks. First, we build a small, domain-specific language model for veterinary medicine, a field that is often overlooked. This model is trained from scratch with proper pretraining, fine-tuning, and safety alignment. Second, we design models for summarizing medical inputs that include both text and images, focusing on low-resource and code-mixed languages to help healthcare professionals better understand complex patient data. Finally, we introduce new benchmarks to evaluate model performance in medical settings, including a) M3Retrieve, a large multimodal retrieval benchmark across 5 domains and 16 medical fields, and b) a multilingual trust benchmark that covers 15 languages and 18 detailed tasks. Together, these efforts aim to make generative AI more practical, reliable, and inclusive for healthcare use.
|
Sriparna Saha (IIT Patna) |
| 15:00 - 16:00 |
On Stopping Times of Power-one Sequential Tests: Tight Lower and Upper Bounds
Sequential hypothesis testing is a fundamental tool in statistics and machine learning. It enables decisions to be made from streaming data while controlling errors. But what is the minimum number of samples that you need to make such decisions with high confidence? In this talk, we will see two tight lower bounds on the stopping times of power-one sequential tests that guarantee small false positive rates and eventual detection under the alternative. These results extend the classical works (Wald, Farrell) to modern, fully nonparametric composite settings, using an information-theoretic quantity, which we call KL-inf, the minimal KL divergence between the null and alternative sets. We will also see sufficient conditions for designing tests that match these lower bounds. Given past work, these upper and lower bounds are unsurprising in their form; our main contribution is the generality in which they hold, for example, not requiring reference measures or compactness of the classes.
|
Shubhada Agrawal (IISc, Bangalore) |
| 16:00 - 17:00 | High Tea | |
| 17:00 - 19:00 | Campus Tour | |
| 19:00 - Onwards | Banquet Dinner (Venue: CSA Garden) | |
| Day-3 Monday - 24th November 2025 | ||
|---|---|---|
| 08:00 - 09:00 - Breakfast | ||
| Morning Session: 09:00 - 13:00, Venue: CSA 104 | ||
| Session Chair: Neha Karanjkar (IIT Goa) | ||
| Time Slot | Title of the Talk | Speaker (Affiliation) |
| 09:00 - 09:55 |
Principled approaches to Model Editing
Model Editing-modifying well-trained models for tasks such as compression or classwise unlearning—has become increasingly important, especially when training data or original training details are unavailable. In this talk, we present recent results showing how structural components of a network can be identified using only their distributional responses. We introduce Discriminative Components, filters whose responses differ sharply across classes (in TV distance), and demonstrate their utility for pruning. We also identify HiFi Components, subsets of filters that can faithfully reconstruct layer-wise activations. Together, these insights enable efficient, principled, and fully data-free model editing across both vision and language tasks.
|
Chiranjib Bhattacharyya (IISc, Bangalore) |
| 09:55 - 10:50 |
On Device Agents with Small Action Models (SAMs)
The term Large Action Model (LAM) is typically used in the context of specific agentic workflows leveraging Large Language Models (LLMs) designed to accomplish tasks by interacting with tools/APIs, device controls, web pages or user Interfaces. Since such actions are often local and frequent, there is a compelling case to consider Small Action Models (SAMs) to unlock on-device, private, low-power, low-latency agents. However, small models pose significant challenges since they are often less reliable. This talk discusses the challenges in building high-quality, reliable on-device action models and presents an approach to building SAMs that are capable of complex query orchestration
|
Niranjan Damera (Hewlett Packards) |
| 10:50 - 11:10 | Coffee Break | |
| 11:10 - 12:05 |
Efficient Stochastic Machine Learning at the Edge
In this talk, I will talk about some hardware/software work my group has done in the area of stochastic computing based machine learning acceleration. I will talk about suitability of the SC to this workload, how to deal with its inherent approximate nature and briefly discuss few chip prototypes which we leverage both logic and in-memory implementations of SC-based accelerators for dense as well as a sparse compute.
|
Puneet Gupta (UCLA) |
| 12:05 - 13:00 |
Adaptive and Efficient Deep Neural Network Inference on Heterogeneous Edge Platforms
Deploying deep neural networks (DNNs) on edge devices presents unique challenges due to limited computational resources, diverse hardware architectures, and dynamic runtime conditions. This talk presents a comprehensive overview of our research efforts aimed at making DNN inference on heterogeneous edge platforms more efficient, adaptive, and responsive to real-world constraints.
|
Gayathri Ananthanarayanan (IIT Dharwad) |
| 13:00 - 14:00 | Lunch Break | |
| Afternoon Session: 14:00 - 16:30, Venue: CSA 104 | ||
| Session Chair: Gayathri Ananthanarayanan (IIT Dharwad) | ||
| 14:00 - 14:50 |
From Guess to Guarantee: Fusing AI and AR for Automated Synthesis
We increasingly entrust large parts of our daily lives to computer systems that are growing ever more complex. Developing scalable and trustworthy methods for designing, building, and verifying these systems is therefore crucial. In this talk, I will focus on automated synthesis a technique that uses formal specifications to automatically generate systems such as functions, programs, or circuits that provably satisfy their requirements. Can we leverage AI-particularly machine learning and large language models-to propose candidate programs or functions? Yes. But can these candidates be guaranteed correct? Can we verify them, or even repair them so that they provably meet the intended specifications? This talk will revolve around these questions-how AI and Automated Reasoning can be effectively fused to synthesize reliable systems.
|
Priyanka Golia (IIT Delhi) |
| 14:50 - 15:40 |
Toward Faithful and Human-Interpretable Explanations for Graph Neural Networks
Understanding why Graph Neural Networks (GNNs) make certain predictions remains a central challenge. Existing explainability methods, though insightful, often produce complex or large explanations that are difficult for humans to interpret, and they primarily focus on local reasoning around individual predictions. Yet, a GNN learns global reasoning patterns that govern its behavior across data. This motivates our broader vision-to design explainability algorithms that are both faithful to the model's reasoning and interpretable to humans. We first addressed this through GraphTrail, the first global, post-hoc GNN explainer that represents model behavior as Boolean formulae over subgraph-level concepts discovered using Shapley values, offering a symbolic understanding of GNN reasoning. However, graph datasets are inherently multi-modal, combining topology with rich node and edge attributes, and a graph with n nodes admits up to 2n subgraphs, making the isolation of neural reasoning patterns combinatorially prohibitive. Consequently, GraphTrail is limited to small, labeled graphs. Our latest work, GNNXemplar, overcomes these challenges. Inspired by Exemplar Theory in cognitive science, it identifies representative nodes-exemplars-in the embedding space and derives interpretable natural language rules for their neighborhoods using large language models.
|
Sayan Ranu (IIT Delhi) |
| 15:40 - 16:30 |
Efficient Solutions for Machine Learning at the Edge
The rapid growth of edge devices has created a dynamic, AI-powered data ecosystem with significant potential for societal advancement. However, privacy concerns restrict data sharing across multiple owners, hindering the full potential of AI. Furthermore, edge devices often have substantial resource constraints and heterogeneity, severely restricting their ability to handle large models. This talk will provide innovative solutions to overcome these challenges and enable efficient and privacy-preserving ML in diverse edge settings. As a key highlight, we will address the following question: How to enable federated learning of a large global model, when each edge device can only train a small local model?
|
Saurav Prakash (IIT Madras) |
| 16:30 - Onwards | High Tea | |
| Day-4 Tuesday - 25th November 2025 | ||
|---|---|---|
| 08:00 - 09:00 - Breakfast | ||
| Morning Session: 09:30 - 13:00, Venue: CSA 104 | ||
| Session Chair: Anant Raj (IISc, Bangalore) | ||
| Time Slot | Title of the Talk | Speaker (Affiliation) |
| 09:30 - 10:30 |
Scalable and Interactive 3D Ocean Data Visualization
Oceanographers struggle with the scalability that is required for visual analysis of massive and multivariate ocean model output for tasks like event tracking and phenomenon identification. Our research addresses this challenge by introducing a novel methodology centered on two key innovations: integrating specialized domain-specific analysis modules directly as efficient ParaView filters, and developing parallel solutions that leverage the use of computing resources available to the analyst. This approach culminated in the development of pyParaOcean, an extendible and easily deployable visualization system that leverages ParaView's parallel processing capabilities. We demonstrate the utility of this research with a Bay of Bengal case study and present scaling studies that confirm the high efficiency of the system.
|
Vijay Natarajan (IISc, Bangalore) |
| 10:30 - 11:30 |
Theory and Practice: Performance Engineering Beyond Steady State
Performance models assume steady state and infinite resources. Real systems don't. Learn how performance engineers diagnose where theory breaks, design experiments to expose the gaps, and use that understanding to solve hard problems. We'll explore the principles through real debugging examples.
|
Shishir Kumar (AMD) |
| 11:30 - 12:00 | Coffee Break | |
| 12:00 - 13:00 |
Improving the Quality of GPU-accelerated Software
An ever-growing number of applications today rely on Graphics Processing Units (GPUs) for their computation. Consequently, the quality of GPU programs has a significant impact on the reliability and efficiency of a substantial portion of today's software ecosystem. In this talk, we will discuss two key aspects of GPU program quality -- software reliability and efficiency. We will demonstrate how writing correct and efficient GPU programs can be challenging, even for experienced programmers. In the first part of the talk, we will discuss how subtle synchronization bugs in GPU programs can lead to intermittent failures, threatening the reliability of GPU-accelerated software. We will then discuss our tool, named iGUARD, which can pinpoint such synchronization bugs for programmers. In the second part of the talk, we will discuss how sub-optimal programming can lead to performance bugs in GPU programs, causing underutilization of costly and power-hungry GPUs. We will then discuss our second tool, ScopeAdvice, which helps programmers identify performance bugs caused by over- and redundant synchronization in GPU programs. Together, these tools can help improve the quality of GPU programs.
|
Arkaprava Basu (IISc, Bangalore) |
| 13:00 - 14:00 | Lunch Break | |
| Afternoon Session: 14:30 - 16:30, Venue: CSA 104 | ||
| Session Chair: Sumit Kumar Mandal (IISc, Bangalore) | ||
| 14:00 - 14:50 |
Scalable Simulation for Performance Modeling and Design Exploration
Simulation plays a key role in systems research. This session will begin with a gentle introduction to Discrete-Event Simulation (DES), covering event-driven and cycle-based approaches and explain why parallelizing DES is essential yet non-trivial. I will then introduce SiTAR, a parallel simulation framework that we have developed over several years, outline its key ideas, modeling language, and runtime, and share result on modeling and parallel simulation of multicore and memory subsystem models for design exploration. This talk will conclude with an overview of our ongoing work on automated model generation and hybrid (discrete-continuous) simulation.
| Neha Karanjkar (IIT Goa) |
| 14:50 - 15:40 |
From Radiance Fields to Gaussian Splatting: Learning 3D Scenes from Sparse Inputs
Neural Radiance Fields (NeRFs) have offer remarkable scene reconstructions through differentiable volumetric rendering. Yet, their reliance on dense image capture and heavy optimization limits practical deployment. This talk will first introduce the foundations of NeRFs — their scene representation, rendering pipeline, and learning dynamics — and then discuss advances that extend NeRFs to sparse input settings. I will also introduce 3D Gaussian Splatting, a recent point-based alternative enabling real-time rendering, and highlight our recent work on sparse-input 3D Gaussian Splatting.
|
Rajiv Soundararajan (IISc, Bangalore) |
| 15:40 - 16:30 |
Pathway to PhD
TBD
|
PhD Scholar |
| 16:30 - Onwards | High Tea | |