The Vision Language Model (VLM)
Analyzes visual intent and outputs a structured JSON representation—geometry, lighting, composition, attributes—all explicit, all inspectable.
No Black Box. No Monolith. No Lock-In. Full visibility into every model. Open weights you can run anywhere. A modular architecture that lets you use exactly what you need.
Most visual AI platforms are opaque bundles. You get everything or nothing-with no visibility into what's happening inside.
This creates risk:
Can't explain, audit, or reproduce results
Pay for capabilities you don't need
Switching costs grow with every integration
Their cloud, their rules
Fibo isn't one model—it's two, working in sequence.

Analyzes visual intent and outputs a structured JSON representation—geometry, lighting, composition, attributes—all explicit, all inspectable.
Renders the final image from that structured state. No ambiguity. No interpolation from vague prompts.
You can inspect, modify, and version the intermediate state- before a single pixel is rendered.
No guessing. No "prompt and pray." Full observability.
Prompt in → image out (black box)
Can't see why you got what you got
Reproduce results? Good luck
Reproduce results? Good luck
Prompt in → structured state → image out
Full visibility into every parameter
Deterministic: same state = same output
Deterministic: same state = same output
The Fibo family shares a common control language. Only execution characteristics change. Same paradigm. Different problems solved. No paradigm shifts when requirements change.
Full image creation from structured state
Optimized for: Quality, prompt adherence
Targeted mutations to existing visual state
Optimized for: Precision, consistency
Production-grade generation
Optimized for: Latency, throughput
Lightweight on-device generation
Optimized for: Size, offline capability
Custom-trained on your data
Optimized for: Brand fidelity
Style-locked generation
Optimized for: Visual consistency
RMBG-2.0 is Bria's segmentation model, designed to preserve information rather than discard it. Instead of binary masks, it outputs continuous alpha mattes—retaining fine-grained boundary detail.
Clean compositing. Reliable background replacement. Seamless integration into structured pipelines.
Part of the modular toolkit—use it standalone or as a building block in larger workflows.
Bria is committed to open, responsible AI development. Our foundation models are available on Hugging Face with full transparency—no black boxes, no hidden constraints.
What's available:
Model weights - Download and run Fibo models, RMBG-2.0, and more.
Source code - Full implementation, not just inference.
Training recipes - Reproduce, extend, customize
Documentation - Comprehensive guides and examples
Open source builds trust. Commercial support builds production systems. With Bria, you get both:
Download models, run locally, test against your requirements
For startups under $1M ARR or $25M funding
Enterprise pricing and dedicated solutions
1M+ downloads on Hugging Face
Active contributor community
Peer-reviewed research (NeurIPS, CVPR)