Safe Latent Diffusion: Mitigating Inappropriate Degeneration in Diffusion Models
- P. Schramowski, Manuel Brack, Björn Deiseroth, K. Kersting
- Computer ScienceArXiv
- 9 November 2022
This work establishes a novel image generation test bed—inappropriate image prompts (I2P)—containing dedicated, real-world image-to-text prompts covering concepts such as nudity and violence, and introduces safe latent diffusion (SLD), a method to combat degenerated and biased human behavior.
How Long Do Vulnerabilities Live in the Code? A Large-Scale Empirical Measurement Study on FOSS Vulnerability Lifetimes
- Nikolaos Alexopoulos, Manuel Brack, Jan Philipp Wagner, Tim Grube
- Computer ScienceUSENIX Security Symposium
- 2022
The first large-scale measurement of Free and Open Source Software vulnerability lifetimes is performed, and it is found that the average lifetime of a vulnerability is around 4 years, varying significantly between projects (~2 years for Chromium, ~7 years for OpenSSL).
Does CLIP Know My Face?
- Dominik Hintersdorf, Lukas Struppek, Manuel Brack, Felix Friedrich, P. Schramowski, K. Kersting
- Computer Science
- 15 September 2022
This work introduces a novel method to assess privacy for multi-modal models, specifically vision-language models like CLIP, and suggests that IDIAs can be used to prove the unauthorized use of data for training and to enforce privacy laws.
Image Classifiers Leak Sensitive Attributes About Their Classes
- Lukas Struppek, Dominik Hintersdorf, Felix Friedrich, Manuel Brack, P. Schramowski, K. Kersting
- Computer Science
- 16 March 2023
This work introduces the first Class Attribute Inference Attack (Caia), which leverages recent advances in text-to-image synthesis to infer sensitive attributes of individual classes in a black-box setting, while remaining competitive with related white-box attacks.
The Stable Artist: Steering Semantics in Diffusion Latent Space
- Manuel Brack, P. Schramowski, Felix Friedrich, Dominik Hintersdorf, K. Kersting
- Computer ScienceArXiv
- 12 December 2022
The Stable Artist is presented, to enable control by allow-ing the artist to steer the diffusion process along a variable number of semantic directions, which allows for subtle edits to images, changes in composition and style, as well as optimization of the overall artistic conception.
Fair Diffusion: Instructing Text-to-Image Generation Models on Fairness
- Felix Friedrich, P. Schramowski, K. Kersting
- Computer ScienceArXiv
- 7 February 2023
This work presents a novel strategy, called Fair Diffusion, to attenuate biases after the deployment of generative text-to-image models, and demonstrates shifting a bias, based on human instructions, in any direction yielding arbitrarily new proportions for, e.g., identity groups.
SEGA: Instructing Diffusion using Semantic Dimensions
- Manuel Brack, Felix Friedrich, Dominik Hintersdorf, Lukas Struppek, P. Schramowski, K. Kersting
- Computer ScienceArXiv
- 28 January 2023
Text-to-image diffusion models have recently re-ceived a lot of interest for their astonishing ability to produce high-fidelity images from text only. However, achieving one-shot generation that…
ILLUME: Rationalizing Vision-Language Models by Interacting with their Jabber
- Manuel Brack, P. Schramowski, Björn Deiseroth, K. Kersting
- Computer ScienceArXiv
- 2022
This work proposes an iterative sampling and tuning paradigm, called I LLUME, that executes the following loop: Given an image-question-answer prompt, the VLM samples multiple candidate rationales, and a human critic provides minimal feedback via preference selec- tion, used for fine-tuning.
Exploiting Cultural Biases via Homoglyphs in Text-to-Image Synthesis
- Lukas Struppek, Dominik Hintersdorf, Felix Friedrich, Manuel Brack, P. Schramowski, K. Kersting
- Computer Science
- 19 September 2022
It is shown that by simply inserting single non-Latin characters in a textual description, common models reflect cultural stereotypes and biases in their generated images, and a novel homoglyph unlearning method is proposed to fine-tune a text encoder, making it robust againsthomoglyph manipulations.
ILLUME: Rationalizing Vision-Language Models through Human Interactions
- Manuel Brack, P. Schramowski, Björn Deiseroth, K. Kersting
- Computer Science
- 17 August 2022
This work proposes a tuning paradigm based on human interactions with machine generated data that is competitive with standard supervised tuning while using significantly fewer training data and only requiring minimal feedback.