AI you can trust
What we build now will decide what AI becomes
We bring together AI researchers, builders, and legal minds to ensure AI is developed with accountability and governance at its core.
Join professionals from leading organizations
What we build now will decide what AI becomes
We bring together AI researchers, builders, and legal minds to ensure AI is developed with accountability and governance at its core.
Join professionals from leading organizations
Open Source
We believe AI debates and infrastructure must be widely available.
This opens the field to many more voices and organizations.
Looking for open‑source contributors and projects to collaborate on.
Community & Research
Events, papers, and collaborations that push the boundaries of AI safety research and practice.

Gemma is a family of open models built by Google DeepMind. Gemma 4 models are multimodal, handling text and image input (with audio supported on small models) and generating text output. Google is opening a Hackathon for the Gemma 4 Good models. The goal is to create projects that use the models in domains like Safety & Trust, Health & Sciences or Future of Education. In their own words: "This is more than just a competition; it is a chance to show the world that when the right tools are accessible to everyone, the possibilities for positive change are truly endless". Prizes are up to $200,000 USD in cash and awards. Join us in person this weekend to work on our projects and get feedback before the Kaggle submission.

Open-weight LLMs are increasingly deployed in agentic workflows where they can invoke external tools, creating new attack vectors beyond traditional text generation. We present an evaluation framework that measures how model tampering...
No One Thanks You for Disasters That Never Happened
Pricing AI risk while making AI safety investable.
DetectPromptInjection
The go-to API for detecting and preventing prompt injection attacks.
What exactly is an AI system under the EU AI Act?
The starting point for anyone working on AI compliance in Europe — defining whether your software falls under the scope of the regulation.
What is shallow safety alignment and why it matters?
Practical approaches to quantifying AI risks and key metrics for evaluating AI system safety and reliability.
PyCon APAC 2025
Presenting LLM and vector database use case for scenario-based narrative generation.