Last updated

Enterprise-Grade Safety and Transparency by Design

The Bria platform is built for zero-risk AI implementation, with enterprise-grade safety, compliance, and content integrity embedded across every layer. Our proprietary foundation models are trained exclusively on licensed, safe-for-commercial-use data—avoiding scraped, harmful, or infringing content by design. While safety starts at the model training stage, Bria’s APIs offer developers built-in controls in-generation and post-generation through a multi-layered architecture. These include prompt and content moderation, flagging of IP-related prompts, and post-generation compliance features—helping teams meet legal, brand, and platform requirements by default.

Safety Architecture

Bria’s enterprise-grade safety framework spans three layers:

1. Pre-Training Layer – Data Integrity

  • Models are trained on 100% licensed data
  • No scraped internet content or unauthorized material
  • No public figures, fictional characters, biometric data, NSFW, or violent content
  • Balanced, diverse, and inclusive dataset representation

2. In-Generation Layer – Real-Time Controls

Bria provides two configurable, opt-in runtime safety features: prompt content moderation and visual content moderation.

Prompt Moderation

  • Enabled via the prompt_content_moderation parameter
  • Scans textual prompts for NSFW or restricted concepts before generation
  • Uses non-AI-based blocklist filtering

Prompts that reference public figures, brands, or other protected content are not blocked, but the models are not trained on this type of data. As a result, the output may be unrelated or differ significantly from what you intended. When an IP-related reference is detected in the prompt, the following warning will appear in the API response:

  This prompt may contain intellectual property (IP)-protected content.
  To ensure compliance and safety, certain elements may be omitted or altered.
  As a result, the output may not fully meet your request.

Example:

  • Prompt: "a Nike sneaker on a reflective white surface"

Bria Output:
Bria Output Image

Outputs from Other Providers:
The following images show how different visual generation providers handled the same prompt.

Other Output 1Other Output 2Other Output 3Other Output 4Other Output 5

Input & Output Visual Moderation

  • Enabled via the content_moderation parameter
  • Scans both input and output visuals

Moderated Categories

  • Explicit Content: Nudity, sexual activity, sex toys
  • Non-Explicit Nudity: Implied nudity, kissing
  • Swimwear/Underwear
  • Violence: Weapons, blood, self-harm, gore
  • Visually Disturbing Content: Crashes, corpses, emaciated bodies
  • Substances: Alcohol, pills, smoking
  • Offensive Gestures
  • Gambling
  • Hate Symbols

3. Post-Generation Layer – Data Traceability & Compliance

  • C2PA Image Marking adds metadata for content authenticity and traceability
  • Attribution Engine Layer enables revenue-sharing with original data owners and provides transparency to Bria customers into the data used to train the models

Indemnity Guarantee

Bria provides full indemnity against copyright infringement for all outputs generated by its models. This assurance is made possible through our use of 100% licensed, safe-for-commercial-use training data — ensuring that every visual generated with Bria’s platform is compliant by design.