Universal and Transferable Adversarial Attacks on Aligned Language Models
Large language models (LLMs) are typically trained on massive text corpora scraped from the
internet, which are known to contain a substantial amount of objectionable content. In an attempt to make AI systems better aligned with human values. Read more
You might also like
-
New guide: key questions to ask before using GenAI for research on violence against women
-
RightsCon Recap – Assessing humanitarian AI: what M&E frameworks do humanitarians need in the face of emerging AI?
-
Event recap: Humans in the machine – the impact of AI on workers
-
Evidence and Learning in the Context of Climate Change: Invitation to Action