Portrait de Prakash Panangaden

Prakash Panangaden

Membre académique principal
Sujets de recherche
Apprentissage par renforcement
Modèles probabilistes
Raisonnement
Théorie de l'apprentissage automatique
Théorie de l'information quantique

Biographie

Prakash Panangaden a étudié la physique à l'Indian Institute of Technology de Kanpur, en Inde. Il a obtenu une maîtrise en physique de l'Université de Chicago, où il a étudié l'émission stimulée des trous noirs. Il a ensuite obtenu un doctorat en physique de l'Université du Wisconsin-Milwaukee, dans lequel il s’est penché sur la théorie quantique des champs dans un espace-temps courbe.

Il a été professeur adjoint d'informatique à l'Université Cornell, où il a principalement travaillé sur la sémantique des langages de programmation concurrents. Depuis 1990, il travaillait à l'Université McGill. Au cours des 25 dernières années, il s'est intéressé à de nombreux aspects des processus de Markov : équivalence des processus, caractérisation logique, approximation et métrique.

Récemment, il a travaillé sur l'utilisation des métriques pour améliorer l'apprentissage des représentations. Il a également publié des articles sur la physique, l'information quantique et les mathématiques pures. Il est membre de la Société royale du Canada et de l'Association for Computing Machinery (ACM).

Étudiants actuels

Maîtrise recherche - McGill
Co-superviseur⋅e :

Publications

Studying the Interplay Between the Actor and Critic Representations in Reinforcement Learning
Trevor McInroe
Christopher G. Lucas
David Abel
Stefano V Albrecht
Conditions on Preference Relations that Guarantee the Existence of Optimal Policies
Learning from Preferential Feedback (LfPF) plays an essential role in training Large Language Models, as well as certain types of interactiv… (voir plus)e learning agents. However, a substantial gap exists between the theory and application of LfPF algorithms. Current results guaranteeing the existence of optimal policies in LfPF problems assume that both the preferences and transition dynamics are determined by a Markov Decision Process. We introduce the Direct Preference Process, a new framework for analyzing LfPF problems in partially-observable, non-Markovian environments. Within this framework, we establish conditions that guarantee the existence of optimal policies by considering the ordinal structure of the preferences. We show that a decision-making problem can have optimal policies -- that are characterized by recursive optimality equations -- even when no reward function can express the learning goal. These findings underline the need to explore preference-based learning strategies which do not assume that preferences are generated by reward.
Polynomial Lawvere Logic
Giorgio Bacci
Radu Mardare
Gordon D. Plotkin
Optimal Approximate Minimization of One-Letter Weighted Finite Automata
In this paper, we study the approximate minimization problem of weighted finite automata (WFAs): to compute the best possible approximation … (voir plus)of a WFA given a bound on the number of states. By reformulating the problem in terms of Hankel matrices, we leverage classical results on the approximation of Hankel operators, namely the celebrated Adamyan-Arov-Krein (AAK) theory. We solve the optimal spectral-norm approximate minimization problem for irredundant WFAs with real weights, defined over a one-letter alphabet. We present a theoretical analysis based on AAK theory, and bounds on the quality of the approximation in the spectral norm and
Policy Gradient Methods in the Presence of Symmetries and State Abstractions
Reinforcement learning (RL) on high-dimensional and complex problems relies on abstraction for improved efficiency and generalization. In th… (voir plus)is paper, we study abstraction in the continuous-control setting, and extend the definition of Markov decision process (MDP) homomorphisms to the setting of continuous state and action spaces. We derive a policy gradient theorem on the abstract MDP for both stochastic and deterministic policies. Our policy gradient results allow for leveraging approximate symmetries of the environment for policy optimization. Based on these theorems, we propose a family of actor-critic algorithms that are able to learn the policy and the MDP homomorphism map simultaneously, using the lax bisimulation metric. Finally, we introduce a series of environments with continuous symmetries to further demonstrate the ability of our algorithm for action abstraction in the presence of such symmetries. We demonstrate the effectiveness of our method on our environments, as well as on challenging visual control tasks from the DeepMind Control Suite. Our method's ability to utilize MDP homomorphisms for representation learning leads to improved performance, and the visualizations of the latent space clearly demonstrate the structure of the learned abstraction.
Sum and Tensor of Quantitative Effects
Giorgio Bacci
Radu Mardare
Gordon Plotkin
Behavioural pseudometrics for continuous-time diffusions
Propositional Logics for the Lawvere Quantale
Giorgio Bacci
Radu Mardare
Gordon Plotkin
Behavioural equivalences for continuous-time Markov processes
A Kernel Perspective on Behavioural Metrics for Markov Decision Processes
We present a novel perspective on behavioural metrics for Markov decision processes via the use of positive definite kernels. We define a ne… (voir plus)w metric under this lens that is provably equivalent to the recently introduced MICo distance (Castro et al., 2021). The kernel perspective enables us to provide new theoretical results, including value-function bounds and low-distortion finite-dimensional Euclidean embeddings, which are crucial when using behavioural metrics for reinforcement learning representations. We complement our theory with strong empirical results that demonstrate the effectiveness of these methods in practice.
Towards an AAK Theory Approach to Approximate Minimization in the Multi-Letter Case
We study the approximate minimization problem of weighted finite automata (WFAs): given a WFA, we want to compute its optimal approximation … (voir plus)when restricted to a given size. We reformulate the problem as a rank-minimization task in the spectral norm, and propose a framework to apply Adamyan-Arov-Krein (AAK) theory to the approximation problem. This approach has already been successfully applied to the case of WFAs and language modelling black boxes over one-letter alphabets \citep{AAK-WFA,AAK-RNN}. Extending the result to multi-letter alphabets requires solving the following two steps. First, we need to reformulate the approximation problem in terms of noncommutative Hankel operators and noncommutative functions, in order to apply results from multivariable operator theory. Secondly, to obtain the optimal approximation we need a version of noncommutative AAK theory that is constructive. In this paper, we successfully tackle the first step, while the second challenge remains open.
Augmenting Human Selves Through Artificial Agents – Lessons From the Brain
Georg Northoff
Maia Fraser
John Griffiths
Dimitris A. Pinotsis
Rosalyn Moran
Karl Friston
Much of current artificial intelligence (AI) and the drive toward artificial general intelligence (AGI) focuses on developing machines for f… (voir plus)unctional tasks that humans accomplish. These may be narrowly specified tasks as in AI, or more general tasks as in AGI – but typically these tasks do not target higher-level human cognitive abilities, such as consciousness or morality; these are left to the realm of so-called “strong AI” or “artificial consciousness.” In this paper, we focus on how a machine can augment humans rather than do what they do, and we extend this beyond AGI-style tasks to augmenting peculiarly personal human capacities, such as wellbeing and morality. We base this proposal on associating such capacities with the “self,” which we define as the “environment-agent nexus”; namely, a fine-tuned interaction of brain with environment in all its relevant variables. We consider richly adaptive architectures that have the potential to implement this interaction by taking lessons from the brain. In particular, we suggest conjoining the free energy principle (FEP) with the dynamic temporo-spatial (TSD) view of neuro-mental processes. Our proposed integration of FEP and TSD – in the implementation of artificial agents – offers a novel, expressive, and explainable way for artificial agents to adapt to different environmental contexts. The targeted applications are broad: from adaptive intelligence augmenting agents (IA’s) that assist psychiatric self-regulation to environmental disaster prediction and personal assistants. This reflects the central role of the mind and moral decision-making in most of what we do as humans.