Hi :)

I am a PhD student at Imperial College London. My research interests include controllable generation, interpretability, and trustworthy AI.

My research focuses on understanding and mitigating risks in generative AI systems with the help of localisation methods in generative models. Previously, I have worked towards identifying attribute leakage in text-to-image diffusion models and building methods for stopping this leakage. I have also worked on building controllable generation methods for text-to-image diffusion models.

Previously, I completed my MRes in Artificial Intelligence and Machine Learning at Imperial College London, and my B.Tech in Computer Science and Engineering from IIT Guwahati. I also gained valuable industry experience as a Software Engineer at Microsoft.

I'm deeply interested in Mechanistic Interpretability, Explainable AI, and Reinforcement Learning. When I am not experimenting or judging papers 🤓, I enjoy mentoring junior students, playing chess, running, and reading books.

Education

🎓
Imperial College London
PhD in Computing
Oct 2025 - Present
🎓
Imperial College London
MRes in Artificial Intelligence and Machine Learning
Sept 2024 - Sept 2025
🎓
Indian Institute of Technology, Guwahati
B.Tech in Computer Science and Engineering
Nov 2020 - May 2024

Selected Publications

Understanding Dementia Speech Alignment with Diffusion-Based Image Generation

Mansi, Anastasios Lepipas, Soteris Demetriou, Dominika C Woszczyk, Yiying Guan
INTERSPEECH 2025 (Oral)
We investigate attribute alignment signals between text and images in diffusion-based generation for dementia speech contexts. Our work explores how generative models can inadvertently reveal sensitive cognitive information through text-to-image generation, with implications for privacy and healthcare applications. We develop novel methods to detect and analyze these alignment patterns, contributing to the understanding of information leakage in generative AI systems.

Leaky Diffusion: Attribute Leakage in Text-Guided Image Generation

Anastasios Lepipas, Mansi, Marios Charalambides, Jiani Liu, Yiying Guan, Dominika C Woszczyk, Thanh Hai Le, Soteris Demetriou
PoPETS 2025
We show and analyze attribute leakage paths in text-guided diffusion models, demonstrating how sensitive information can inadvertently leak through generative processes. Our research reveals novel attack vectors for authorship identification using text-to-image diffusion models, highlighting significant privacy concerns in current generative AI systems. We propose comprehensive analysis frameworks and mitigation strategies for these emerging security vulnerabilities.

AmalREC: A Dataset for Relation Extraction and Classification Leveraging Amalgamation of Large Language Models

Mansi, Pranshu Pandya, Mahek Bhavesh Vora, Soumya Bharadwaj, Ashish Anand
Preprint 2024
We present AmalREC, a comprehensive relation extraction dataset created by combining LLM-based approaches with template-based methods. Our dataset leverages a 6-level relation hierarchy to generate diverse and high-quality relation extraction examples. We provide detailed analysis of bias and noise across different relation buckets, comparing the effectiveness of various generation strategies and their impact on downstream performance.

On the Impact of Sparsification on Quantitative Argumentative Explanations in Neural Networks

Daniel Peacock, Mansi, Nico Potyka, Francesca Toni and Xiang Yin
3rd International Workshop on Argumentation for eXplainable AI
We investigate the impact of sparsification techniques on the quality and interpretability of quantitative argumentative explanations in neural networks. Our research explores how different sparsification methods affect the ability of neural networks to generate coherent, logical explanations for their predictions. We develop novel metrics to evaluate the argumentative quality of explanations and demonstrate that strategic sparsification can enhance both model efficiency and explanation interpretability without compromising predictive performance.

Contact

Email: m24@ic.ac.uk · Phone: +44 7818915997

LinkedIn: mansi-973736204 · GitHub: AnMaJ

Experience

Research Experience

  1. Imperial College London

    Sept 2024 – Present · London, UK
    • Leakage of neuro-cognitive decline markers in T2I diffusion; accepted at INTERSPEECH 2025.
    • Authorship leakage and a new attack for authorship identification using T2I diffusion; accepted at PETS 2025.
    • Developing adversarially robust T2I framework with policy-driven safety controls via concept localization.
    • Building argumentative LLMs for faithful, self-debated explanations.
  2. University of Pennsylvania (Research Intern)

    Nov 2022 – Sept 2024 · Remote · Mentor: Dr. Vivek Gupta
    • Diverse, coherent QA from tables using SQL-based pipeline and LLM few-shot prompting.
    • Graph-to-QA generation using DPlot and automatic SQL template generation.
  3. IIT Guwahati (B.Tech Thesis)

    May 2023 – Sept 2024 · Mentor: Prof. Ashish Anand
    • Generated a diverse relation extraction dataset using a 6-level relation hierarchy.
    • Compared LLM-based and template-based methods; analyzed bias and noise across relation buckets.
  4. DDoS in D2D Communications

    Aug 2023 – May 2024 · IIT Guwahati
    • Detection for seven DDoS attack types using Random Forest.
    • Prevention via relay deployment in ns-3 with CIC Flowmeter for packet processing.

Industry Experience

  1. Software Engineer, Microsoft

    May 2024 – Sept 2024 · Bangalore, India · Cloud & AI
    • AI-powered customer care for Dynamics 365 Omnichannel.
    • Scalability migration across communication channels.
    • Stack: C#, Kusto, REST APIs, Docker, Azure.
  2. Software Engineer Intern, Microsoft

    May 2023 – July 2023 · Bangalore, India · Cloud & AI
    • ICM auto-routing with BERT; efficiency improved from 53% → 93% across 98 teams.
    • Marker model, feedback pipeline, and RL for self-learning.
    • Stack: PyTorch, TensorFlow, Kusto, REST APIs, Docker, Azure.

Things I've worked on.

Understanding Dementia Speech Alignment with Diffusion-Based Image Generation

Mansi, Anastasios Lepipas, Soteris Demetriou, Dominika C Woszczyk, Yiying Guan
INTERSPEECH 2025

We investigate attribute alignment signals between text and images in diffusion-based generation for dementia speech contexts. Our work explores how generative models can inadvertently reveal sensitive cognitive information through text-to-image generation, with implications for privacy and healthcare applications. We develop novel methods to detect and analyze these alignment patterns, contributing to the understanding of information leakage in generative AI systems.

Understanding Dementia Speech Alignment with Diffusion-Based Image Generation

Leaky Diffusion: Attribute Leakage in Text-Guided Image Generation

Anastasios Lepipas, Mansi, Marios Charalambides, Jiani Liu, Yiying Guan, Dominika C Woszczyk, Thanh Hai Le, Soteris Demetriou
PoPETS 2025

We show and analyze attribute leakage paths in text-guided diffusion models, demonstrating how sensitive information can inadvertently leak through generative processes. Our research reveals novel attack vectors for authorship identification using text-to-image diffusion models, highlighting significant privacy concerns in current generative AI systems. We propose comprehensive analysis frameworks and mitigation strategies for these emerging security vulnerabilities.

Leaky Diffusion: Attribute Leakage in Text-Guided Image Generation

AmalREC: A Dataset for Relation Extraction and Classification Leveraging Amalgamation of Large Language Models

Mansi, Pranshu Pandya, Mahek Bhavesh Vora, Soumya Bharadwaj, Ashish Anand
Preprint 2024

We present AmalREC, a comprehensive relation extraction dataset created by combining LLM-based approaches with template-based methods. Our dataset leverages a 6-level relation hierarchy to generate diverse and high-quality relation extraction examples. We provide detailed analysis of bias and noise across different relation buckets, comparing the effectiveness of various generation strategies and their impact on downstream performance.

AmalREC: A Dataset for Relation Extraction and Classification

On the Impact of Sparsification on Quantitative Argumentative Explanations in Neural Networks

Daniel Peacock, Mansi, Nico Potyka, Francesca Toni and Xiang Yin
3rd International Workshop on Argumentation for eXplainable AI

We investigate the impact of sparsification techniques on the quality and interpretability of quantitative argumentative explanations in neural networks. Our research explores how different sparsification methods affect the ability of neural networks to generate coherent, logical explanations for their predictions. We develop novel metrics to evaluate the argumentative quality of explanations and demonstrate that strategic sparsification can enhance both model efficiency and explanation interpretability without compromising predictive performance.

On the Impact of Sparsification on Quantitative Argumentative Explanations in Neural Networks

Key Projects

Lost Child Found Application

May 2022

Flutter Android app using a face recognition REST API to help find lost children.

Play Store · GitHub

Using AED to estimate Annotator Count

Sept 2022 – Nov 2022

Pipeline to estimate annotator counts for classification tasks using AED with DistilBERT.

Report · GitHub

Live Location Tracking App

Jul 2021 – Aug 2021

Flutter app to share and monitor live locations with Google Maps API and safety alerts.

GitHub

Raman — The Vision Bot

Apr 2022 – May 2022

CNN-based hand-gesture recognition integrated with OpenCV and Arduino for RPS game.

GitHub

Other Projects

Achievements

  • 3rd prize at Imperial Computing Conference for work on attribute leakage.
  • India Future Leaders and Anne Seagrim Scholarships.
  • Microsoft Engage mentee; received full-time role offer.
  • Barclays Global Student Discovery Program: top 10 sophomores.
  • Grace Hopper Celebration Scholarship 2024 & 2025.
  • KVPY National Research Fellowship (AIR 219).
  • JEE Main: AIR 1 in Physics; AIR 439 overall.
  • JEE Advanced: AIR 1450.
  • ICPC: AIR 93 (certificate linked).
  • Flipkart Grid 4.0: Top 3% among 50k+ teams.
  • National Standard Examination in Physics: top 1% nationwide.
  • Regional Mathematics Olympiad qualifier (top 5%).
  • Technothlon Silver Medalist (AIR 239).

Extracurricular

  • Poster designer — examples.
  • Football — Silver medal in Spardha inter-hostel championship.