About Me
My research interests lie in inference and post-training techniques for generative models. I focus on multimodal models, in particular diffusion and flow-based architectures, and work on improving their capabilities and controllability.
Beyond generation research, I am deeply interested in the safety of multimodal models. As these systems become more powerful and widely deployed, ensuring they behave responsibly and do not produce harmful or biased outputs is critical. My work addresses both the advancement and the responsible deployment of generative AI.
Research Topics: Multimodal AI · Diffusion/Flow Models · Inference & Post-Training · AI Safety · AI Alignment
Selected Publications
For a full list, see my Google Scholar profile.
Inference-time Physics Alignment of Video Generative Models with Latent World Models
J Yuan, X Zhang*, F Friedrich*, N Beltran-Velez*, M Hall, R Askari-Hemmat, ...
CVPR 2026
Measuring and Guiding Monosemanticity
R Härle*, F Friedrich*, M Brack, S Wäldchen, B Deiseroth, P Schramowski, ...
NeurIPS 2025 (Spotlight)
Multilingual Text-to-Image Generation Magnifies Gender Stereotypes and Prompt Engineering May Not Help You
F Friedrich, K Hämmerl, P Schramowski, M Brack, J Libovicky, K Kersting, ...
ACL 2025
Evaluating the Social Impact of Generative AI Systems in Systems and Society
I Solaiman, Z Talat, W Agnew, L Ahmad, D Baker, SL Blodgett, C Chen, ...
Oxford Handbook on the Foundations and Regulation of Generative AI 2025
LlavaGuard: An Open VLM-based Framework for Safeguarding Vision Datasets and Models
L Helff*, F Friedrich*, M Brack*, K Kersting, P Schramowski
ICML 2025
Aurora-M: Open Source Continual Pre-training for Multilingual Language and Code
T Nakamura, M Mishra, S Tedeschi, Y Chai, JT Stillerman, F Friedrich, ...
COLING 2025
FairDiffusion: Auditing and Instructing Text-to-Image Generation Models on Fairness
F Friedrich, M Brack, L Struppek, D Hintersdorf, P Schramowski, ...
AI and Ethics 2024
LEdits++: Limitless Image Editing using Text-to-Image Models
M Brack*, F Friedrich*, K Kornmeier*, L Tsaban, P Schramowski, ...
CVPR 2024
ALERT: A Comprehensive Benchmark for Assessing Large Language Models' Safety through Red Teaming
S Tedeschi, F Friedrich, P Schramowski, K Kersting, R Navigli, H Nguyen, ...
Online Workshop on Red Teaming Generative AI Models 2024
Learning by Self-Explaining
W Stammer, F Friedrich, D Steinmann, M Brack, H Shindo, K Kersting
TMLR 2024
SEGA: Instructing Text-to-Image Models using Semantic Guidance
M Brack, F Friedrich, D Hintersdorf, L Struppek, P Schramowski, ...
NeurIPS 2023
MultiFusion: Fusing Pre-trained Models for Multi-lingual, Multi-modal Image Generation
M Bellagente, M Brack, H Teufel, F Friedrich, B Deiseroth, C Eichenberg, ...
NeurIPS 2023
Exploiting Cultural Biases via Homoglyphs in Text-to-Image Synthesis
L Struppek, D Hintersdorf, F Friedrich, P Schramowski, K Kersting
JAIR 2023
A Typology for Exploring the Mitigation of Shortcut Behaviour
F Friedrich, W Stammer, P Schramowski, K Kersting
Nature Machine Intelligence 2023