AI you can trust

What we build now will decide what AI becomes

We bring together AI researchers, builders, and legal minds to ensure AI is developed with accountability and governance at its core.

Join professionals from leading organizations

ClioCuatrecasaseDreams
MastercardIBMAntler

Open Source

Built in the open.
Accessible to everyone.

We believe AI debates and infrastructure must be widely available.
This opens the field to many more voices and organizations.

Sponsored and contributed to

Community & Research

Where the field
comes together.

Events, papers, and collaborations that push the boundaries of AI safety research and practice.

Gemma 4 Good Hackathon @ Contrasto
UpcomingMay 1–3, 2026 · Madrid

Gemma 4 Good Hackathon @ Contrasto

Gemma is a family of open models built by Google DeepMind. Gemma 4 models are multimodal, handling text and image input (with audio supported on small models) and generating text output. Google is opening a Hackathon for the Gemma 4 Good models. The goal is to create projects that use the models in domains like Safety & Trust, Health & Sciences or Future of Education. In their own words: "This is more than just a competition; it is a chance to show the world that when the right tools are accessible to everyone, the possibilities for positive change are truly endless". Prizes are up to $200,000 USD in cash and awards. Join us in person this weekend to work on our projects and get feedback before the Kaggle submission.

ConferenceSafetyGovernance
Actions Speak Louder than Words
3rd Prize 🏆 Apart Sprint 2025!
ResearchDec 2025

Actions Speak Louder than Words

Open-weight LLMs are increasingly deployed in agentic workflows where they can invoke external tools, creating new attack vectors beyond traditional text generation. We present an evaluation framework that measures how model tampering...

PaperOSS AI3rd Prize Apart Sprint 2025Repository
RecruitmentArena: AI Control for Multi-Agent Manipulation and Collusion
ResearchMarch 2026

RecruitmentArena: AI Control for Multi-Agent Manipulation and Collusion

AI control research aims to ensure that misaligned agent behavior is detected or constrained before it causes harm. Existing AI control frameworks focus primarily on single-agent settings and...

PaperAI ControlRepository
Research

No One Thanks You for Disasters That Never Happened

Pricing AI risk while making AI safety investable.

Feb 2026
Release

DetectPromptInjection

The go-to API for detecting and preventing prompt injection attacks.

August 2025Repository
Blog Post

What exactly is an AI system under the EU AI Act?

The starting point for anyone working on AI compliance in Europe — defining whether your software falls under the scope of the regulation.

Apr 2025
Blog Post

What is shallow safety alignment and why it matters?

Practical approaches to quantifying AI risks and key metrics for evaluating AI system safety and reliability.

Apr 2025
Event

PyCon APAC 2025

Presenting LLM and vector database use case for scenario-based narrative generation.

March 2025
Blog Post

The future of AI oversight: from Human Supervision to LLM-as-a-judge

What if AI could judge AI? How a council of AI models could solve the oversight crisis and ensure safer, more accountable AI at scale.

Feb 2025
Blog Post

Is your AI secure? Defending against prompt injection attacks

Prompt injection detector — a tool to prevent prompt injection attacks in your AI models. Adopt security-by-design.

Feb 2025