Portrait of Sharut Gupta is unavailable

Sharut Gupta

Alumni

Publications

FL Games: A federated learning framework for distribution shifts
Federated learning aims to train predictive models for data that is distributed across clients, under the orchestration of a server. However… (see more), participating clients typically each hold data from a different distribution, whereby predictive models with strong in-distribution generalization can fail catastrophically on unseen domains. In this work, we argue that in order to generalize better across non-i.i.d. clients, it is imperative to only learn correlations that are stable and invariant across domains. We propose FL Games, a game-theoretic framework for federated learning for learning causal features that are invariant across clients. While training to achieve the Nash equilibrium, the traditional best response strategy suffers from high-frequency oscillations. We demonstrate that FL Games effectively resolves this challenge and exhibits smooth performance curves. Further, FL Games scales well in the number of clients, requires significantly fewer communication rounds, and is agnostic to device heterogeneity. Through empirical evaluation, we demonstrate that FL Games achieves high out-of-distribution performance on various benchmarks.
QU-BraTS: MICCAI BraTS 2020 Challenge on Quantifying Uncertainty in Brain Tumor Segmentation - Analysis of Ranking Scores and Benchmarking Results
Raghav Mehta
Angelos Filos
Ujjwal Baid
Chiharu Sako
Richard McKinley
Michael Rebsamen
Katrin Datwyler
Raphaël Meier
Piotr Radojewski
Gowtham Krishnan Murugesan
Sahil Nalawade
Chandan Ganesh
Ben Wagner
Fang Yu
Baowei Fei
Ananth J. Madhuranthakam
Joseph A. Maldjian
Laura Daza
Catalina Gómez
Pablo Arbeláez … (see 72 more)
Chengliang Dai
Shuo Wang
Hadrien Reynaud
Yuan-han Mo
Elsa D. Angelini
Yike Guo
Wenjia Bai
Subhashis Banerjee
Lin-min Pei
Murat AK
Sarahi Rosas-González
Ilyess Zemmoura
Clovis Tauber
Minh H. Vu
Tufve Nyholm
Tommy Löfstedt
Laura Mora Ballestar
Verónica Vilaplana
Hugh McHugh
Gonzalo D. Maso Talou
Alan Wang
Jay Patel
Ken Chang
Katharina Hoebel
Mishka Gidwani
Nishanth Arun
Mehak Aggarwal
Praveer Singh
Elizabeth R. Gerstner
Jayashree Kalpathy-Cramer
Nicolas Boutry
Alexis Huard
Lasitha Vidyaratne
Md Monibor Rahman
Khan M. Iftekharuddin
Joseph Chazalon
Élodie Puybareau
Guillaume Tochon
Jun Ma
Mariano Cabezas
Xavier Lladó
Arnau Oliver
Liliana Patricia Marlés Valencia
Sergi Valverde
Mehdi Amian
Mohammadreza Soltaninejad
Andriy Myronenko
Ali Hatamizadeh
Xue Feng
Fang Yu
Nicholas Tustison
Craig H. Meyer
Nisarg A. Shah
Sanjay N. Talbar
Marc‐André Weber
Abhishek Mahajan
Andras Jakab
Roland Wiest
Hassan M. Fathallah‐Shaykh
Arash Nazeri
Mikhail Milchenko
Daniel C. Marcus
Aikaterini Kotrotsou
Rivka R. Colen
John Freymann
Justin Kirby
Christos Davatzikos
Bjoern Menze
Spyridon Bakas
Yarin Gal
Deep learning (DL) models have provided state-of-the-art performance in various medical imaging benchmarking challenges, including the Brain… (see more) Tumor Segmentation (BraTS) challenges. However, the task of focal pathology multi-compartment segmentation (e.g., tumor and lesion sub-regions) is particularly challenging, and potential errors hinder translating DL models into clinical workflows. Quantifying the reliability of DL model predictions in the form of uncertainties could enable clinical review of the most uncertain regions, thereby building trust and paving the way toward clinical translation. Several uncertainty estimation methods have recently been introduced for DL medical image segmentation tasks. Developing scores to evaluate and compare the performance of uncertainty measures will assist the end-user in making more informed decisions. In this study, we explore and evaluate a score developed during the BraTS 2019 and BraTS 2020 task on uncertainty quantification (QU-BraTS) and designed to assess and rank uncertainty estimates for brain tumor multi-compartment segmentation. This score (1) rewards uncertainty estimates that produce high confidence in correct assertions and those that assign low confidence levels at incorrect assertions, and (2) penalizes uncertainty measures that lead to a higher percentage of under-confident correct assertions. We further benchmark the segmentation uncertainties generated by 14 independent participating teams of QU-BraTS 2020, all of which also participated in the main BraTS segmentation task. Overall, our findings confirm the importance and complementary value that uncertainty estimates provide to segmentation algorithms, highlighting the need for uncertainty quantification in medical image analyses. Finally, in favor of transparency and reproducibility, our evaluation code is made publicly available at https://github.com/RagMeh11/QU-BraTS.