Kiran Chhatre

I am an MSCA ITN PhD fellow at the KTH Royal Institute of Technology (2021.01 – Present), where I am advised by Christopher Peters. During my PhD, I have had the opportunity to work closely with Michael J. Black, Timo Bolkart, Srikrishna Karanam, and Jun Rekimoto. I have also gained valuable experience as a PhD intern at Adobe Research, Sony Computer Science Laboratories, the Max Planck Institute for Intelligent Systems (Perceiving Systems), Electronic Arts (SEED), and Ubisoft La Forge.

Prior to my PhD, I worked as a Research Affiliate with the BEAM team at Lawrence Berkeley National Laboratory. I earned my Master’s degree in Mechanical Engineering from RWTH Aachen University (2017.09 – 2020.10), where I was advised by Mikhail Itskov. During my master’s studies, I interned with IBM Research, Dassault Systèmes (SIMULIA), and the Innovation & Learning Center Aachen. Before that, I worked at Dassault Systèmes Solutions Lab and Autocop Telematics, and I completed my Bachelor’s degree in Mechanical Engineering at COEP Technological University (2010.07 – 2014.05).

 /  CV  /  Scholar  /  LinkedIn  /  GitHub  /  X  /  Bluesky

profile photo

Research

My research lies at the intersection of computer vision and machine learning. I am particularly focused on the controllability of generative models across various modalities (image, video, 3D, 4D). I have broadly worked on visual grounding and audio reasoning of Multimodal Large Language Models (MLLMs). I have also worked on emotional 3D animation of virtual humans, with applications in VR-based human–computer interaction (HCI). Previously, I explored Bayesian optimization for multi-agent systems, robotic systems, and computer-aided engineering (CAE).

▸ Currently at Adobe Research London with Paul, Chun-Hao, Yulia, and Hyeonho, working on video foundation models.

/ / / / /
project image

Spectrum: Learning 3D Texture-Aware Representations for Parsing Diverse Human Clothing and Body Parts


Kiran Chhatre, Christopher Peters, Srikrishna Karanam
arXiv Preprint, 2025
arxiv / website / patent /

Spectrum introduces a novel repurposing of an Image-to-Texture diffusion model for improved alignment with body parts and clothing, enabling detailed human parsing that handles diverse clothing types and complex poses across any number of humans in the scene.

project image

Synthetically Expressive: Evaluating gesture and voice for emotion and empathy in VR and 2D scenarios


Haoyang Du, Kiran Chhatre, Christopher Peters, Brian Keegan, Rachel McDonnell, Cathy Ennis
International Conference on Intelligent Virtual Agents (IVA), 2025
arxiv / website / youtube /

This work evaluates gesture and voice synthesis for conveying emotion and empathy in both VR and 2D scenarios, providing insights into the effectiveness of synthetic emotional expressions across different interaction modalities.

project image

Evaluation of Generative Models for Emotional 3D Animation Generation in VRjournal


Kiran Chhatre, Renan Guarese, Andrii Matviienko, Christopher Peters
Frontiers in Computer Science (Human-Media Interaction) & ACM SIGGRAPH I3D, 2025
frontiers paper / i3d workshop / website / video / supp. material /

This work evaluates emotional 3D animation generative models within an immersive Virtual Reality environment, emphasizing user-centric metrics including emotional arousal realism, naturalness, enjoyment, diversity, face-body congruence, and interaction quality in real-time human-agent interaction scenarios.

project image

Audiopedia: Audio QA with Knowledgeoral


Abhirama S. Penamakuri*, Kiran Chhatre*, Akshat Jain (* denotes equal contribution)
IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 2025
arxiv / website / code /

Audiopedia introduces a novel, knowledge-intensive audio question answering task and proposes a framework to enhance audio language models by integrating external knowledge.

project image

AMUSE: Emotional Speech-driven 3D Body Animation via Disentangled Latent Diffusion


Kiran Chhatre, Radek Daněček, Nikos Athanasiou, Giorgio Becherini, Christopher Peters, Michael J. Black, Timo Bolkart
IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2024
arxiv / website / youtube / code / poster / x thread /

AMUSE generates realistic emotional 3D body gestures directly from a speech sequence. It provides user control over the generated emotion by combining the driving speech with a different emotional audio.

project image

EMOTE: Emotional Speech-Driven Animation with Content-Emotion Disentanglement


Radek Daněček, Kiran Chhatre, Shashank Tripathi, Yandong Wen, Michael J. Black, Timo Bolkart
ACM SIGGRAPH Asia Conference Papers, 2023
arxiv / website / video / code / x thread /

Given audio input and an emotion label, EMOTE generates an animated 3D head that has state-of-the-art lip synchronization while expressing the emotion. The method is trained from 2D video sequences using a novel video emotion loss and a mechanism to disentangle emotion from speech.

project image

BEAMBayesOpt: Parallel Bayesian Optimization of Agent-Based Transportation Simulationspecial mentions


Kiran Chhatre, Sidney Feygin, Colin Sheppard, Rashid Waraich
Springer Nature International Conference on Machine Learning, Optimization, and Data Science (LOD), 2022
paper / code / BEAM-integration /

BEAMBayesOpt introduces a parallel Bayesian optimization approach with early stopping that autonomously calibrates hyperparameters in BEAM’s large-scale multi-agent transportation simulations and enables efficient surrogate modeling of complex scenarios.

project image

Spatio-temporal priors in 3D human motion


Anna Deichler*, Kiran Chhatre*, Christopher Peters, Jonas Beskow (* denotes equal contribution)
IEEE International Conference on Development and Learning (StEPP) workshop, 2021
arxiv / website /

This workshop paper investigates spatial-temporal priors for 3D human motion synthesis by comparing graph convolutional networks and transformer architectures to capture dynamic joint dependencies.

project image

Rethinking Computer-Aided Architectural Design (CAAD) – From Generative Algorithms and Architectural Intelligence to Environmental Design and Ambient Intelligence


Todor Stojanovski, Hui Zhang, Emma Frid, Kiran Chhatre, Christopher Peters, Ivor Samuels, Paul Sanders, Jenni Partanen, Deborah Lefosse
Springer Nature International Conference on Computer-Aided Architectural Design Futures (CAAD Futures), 2021
paper /

This paper reviews the evolution of CAAD—from generative algorithms and BIM to current AI developments—and argues that integrating AI-driven ambient intelligence into digital design tools can transform architectural and urban design for smarter, more sustainable cities.




Academic Services

Conference Reviewer: NeurIPS, ICLR, AAAI, ICCV, SIGGRAPH, SIGGRAPH Asia, ISMAR, CoG, IVA
Journal Reviewer: Pattern Recognition, IEEE Transactions on Affective Computing
Program Committee: CLIPE Workshop at Eurographics 2024

Miscellaneous

EU Reports: MSCA ITN Clipe project
CAE Projects: Rear View Camera System (IJSRD 2016), Glass Fiber Reinforced Polymer (IRF 2014)
Co-curricular: IAESTE Intern at Qatar University 2019, RoboCup Logistics League (Team Carologistics) 2018
Extra-curricular: PhD Chapter at KTH, KTH AI Society, KTH Ethics Committee, Swedish Red Cross
Other links: KTH website, MPI-IS website, Old website (Robotics demos)


Updated on: 2025-08-25


Thanks, Jon Barron!