# Enterprise-Grade Safety and Transparency by Design

The Bria platform is built for zero-risk AI implementation, with enterprise-grade safety, compliance, and content integrity embedded across every layer. Our proprietary foundation models are trained exclusively on licensed, safe-for-commercial-use data—avoiding scraped, harmful, or infringing content by design. While safety starts at the model training stage, Bria’s APIs offer developers built-in controls in-generation and post-generation through a multi-layered architecture. These include prompt and content moderation, flagging of IP-related prompts, and post-generation compliance features—helping teams meet legal, brand, and platform requirements by default.

# Safety Architecture

Bria’s enterprise-grade safety framework spans three layers:

## 1. Pre-Training Layer – Data Integrity

- Models are trained on **100% licensed data**
- **No scraped internet content o**r unauthorized material
- **No public figures, fictional characters, biometric data, NSFW, or violent content**
- Balanced, diverse, and **inclusive dataset representation**


## 2. In-Generation Layer – Real-Time Controls

Bria provides two configurable, opt-in runtime safety features: prompt content moderation and visual content moderation.

### Prompt Moderation

- Enabled via the `prompt_content_moderation` parameter
- Scans textual prompts for NSFW or restricted concepts before generation
- Uses non-AI-based blocklist filtering


#### Handling IP-Related Prompts

Prompts that reference public figures, brands, or other protected content are not blocked, but the models are not trained on this type of data.
As a result, the output may be unrelated or differ significantly from what you intended.
When an IP-related reference is detected in the prompt, the following warning will appear in the API response:


```text
  This prompt may contain intellectual property (IP)-protected content.
  To ensure compliance and safety, certain elements may be omitted or altered.
  As a result, the output may not fully meet your request.
```

**Example:**

- **Prompt:** "a Nike sneaker on a reflective white surface"


**Bria Output:**


**Outputs from Other Providers:**
The following images show how different visual generation providers handled the same prompt.

img
img
img
img
img
### Input & Output Visual Moderation

- Enabled via the `content_moderation` parameter
- Scans both input and output visuals


#### Moderated Categories

- **Explicit Content**: Nudity, sexual activity, sex toys
- **Violence**: self-harm, gore
- **Visually Disturbing Content**: Crashes, corpses, emaciated bodies
- **Hate Symbols**


## 3. Post-Generation Layer – Data Traceability & Compliance

- **C2PA Image Marking** adds metadata for content authenticity and traceability
- **Attribution Engine Layer** enables revenue-sharing with original data owners and provides transparency to Bria customers into the data  used to train the models


## Indemnity Guarantee

Bria provides **full indemnity against copyright infringement** for all outputs generated by its models, **for enterprise customers only**. This assurance is made possible through our use of 100% licensed, safe-for-commercial-use training data — ensuring that every visual generated with Bria’s platform is compliant by design.