Kiran Chhatre

I am an MSCA ITN PhD fellow at the KTH Royal Institute of Technology (2021.01 – Present), where I am advised by Christopher Peters. During my PhD, I have had the opportunity to work closely with Michael J. Black, Timo Bolkart, and Srikrishna Karanam. I have also gained valuable experience as a PhD intern at Adobe Research, the Max Planck Institute for Intelligent Systems (Perceiving Systems), Electronic Arts (SEED), and Ubisoft La Forge.

Prior to my PhD, I worked as a Research Affiliate with the BEAM team at Lawrence Berkeley National Laboratory. I earned my Master’s degree in Mechanical Engineering from RWTH Aachen University (2017.09 – 2020.10), where I was advised by Mikhail Itskov. During my master’s studies, I interned with IBM Research, Dassault Systèmes (SIMULIA), and the Innovation & Learning Center Aachen. Before that, I worked at Dassault Systèmes Solutions Lab and Autocop Telematics, and I completed my Bachelor’s degree in Mechanical Engineering at COEP Technological University (2010.07 – 2014.05).

 /  CV  /  Scholar  /  LinkedIn  /  GitHub  /  X  /  Bluesky

profile photo

Research

My research lies at the intersection of computer vision and machine learning. I am particularly focused on the controllability of generative models, with an emphasis on generating realistic 3D human face and body motions. My work also includes the low-rank (LoRA) adaptation of vision-language models for visual grounding. Prior to specializing in computer vision, I explored multi-fidelity parallel Bayesian optimization for multi-agent systems, robotics (vision + control), and computer-aided engineering (CAE).

project image

Audiopedia: Audio QA with Knowledge


Abhirama S. Penamakuri*, Kiran Chhatre*, Akshat Jain (* denotes equal contribution)
IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 2025
arxiv / code / website /

Audiopedia introduces a novel, knowledge-intensive audio question answering task and proposes a framework to enhance audio language models by integrating external knowledge.

project image

AMUSE: Emotional Speech-driven 3D Body Animation via Disentangled Latent Diffusion


Kiran Chhatre, Radek Daněček, Nikos Athanasiou, Giorgio Becherini, Christopher Peters, Michael J. Black, Timo Bolkart
IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2024
arxiv / code / poster / website / youtube / x thread /

AMUSE generates realistic emotional 3D body gestures directly from a speech sequence. It provides user control over the generated emotion by combining the driving speech with a different emotional audio.

project image

EMOTE: Emotional Speech-Driven Animation with Content-Emotion Disentanglement


Radek Daněček, Kiran Chhatre, Shashank Tripathi, Yandong Wen, Michael J. Black, Timo Bolkart
ACM SIGGRAPH Asia Conference Papers, 2023
arxiv / video / code / website / x thread /

Given audio input and an emotion label, EMOTE generates an animated 3D head that has state-of-the-art lip synchronization while expressing the emotion. The method is trained from 2D video sequences using a novel video emotion loss and a mechanism to disentangle emotion from speech.

project image

Spatio-temporal priors in 3D human motion


Anna Deichler*, Kiran Chhatre*, Christopher Peters, Jonas Beskow (* denotes equal contribution)
IEEE International Conference on Development and Learning (StEPP) workshop, 2021
arxiv / website /

This workshop paper investigates spatial-temporal priors for 3D human motion synthesis by comparing graph convolutional networks and transformer architectures to capture dynamic joint dependencies.




Other Projects

project image

BEAMBayesOpt: Parallel Bayesian Optimization of Agent-Based Transportation Simulation


Kiran Chhatre, Sidney Feygin, Colin Sheppard, Rashid Waraich
Springer Nature International Conference on Machine Learning, Optimization, and Data Science (LOD), 2022
paper / code / code #2 / documentation / award /

BEAMBayesOpt introduces a parallel Bayesian optimization approach with early stopping that autonomously calibrates hyperparameters in BEAM’s large-scale multi-agent transportation simulations and enables efficient surrogate modeling of complex scenarios.

project image

Rethinking Computer-Aided Architectural Design (CAAD) – From Generative Algorithms and Architectural Intelligence to Environmental Design and Ambient Intelligence


Todor Stojanovski, Hui Zhang, Emma Frid, Kiran Chhatre, Christopher Peters, Ivor Samuels, Paul Sanders, Jenni Partanen, Deborah Lefosse
Springer Nature International Conference on Computer-Aided Architectural Design Futures (CAAD Futures), 2021
paper /

This paper reviews the evolution of CAAD—from generative algorithms and BIM to current AI developments—and argues that integrating AI-driven ambient intelligence into digital design tools can transform architectural and urban design for smarter, more sustainable cities.




Academic Services

Conference Reviewer: ICLR, SIGGRAPH, SIGGRAPH Asia, CVPRW, ISMAR, AAMAS, LOD, CoG, IVA
Journal Reviewer: Pattern Recognition
Program Committee: CLIPE Workshop at Eurographics 2024

Miscellaneous

EU Reports: MSCA ITN Clipe project
CAE Projects: Rear View Camera System (IJSRD 2016), Glass Fiber Reinforced Polymer (IRF 2014)
Co-curricular: IAESTE Intern at Qatar University 2019, RoboCup Logistics League (Team Carologistics) 2018
Extra-curricular: PhD Chapter at KTH, KTH AI Society, KTH Ethics Committee, Swedish Red Cross
Other links: KTH website, MPI-IS website, Old website (Robotics demos)


Updated on: 2025-03-19


Thanks, Jon Barron!