
facetron
FaceTron is a high-performance face embedding server using ONNX Runtime, supporting dynamic multi-model loading, offline deployment, and scalable environments. It exposes an OpenAPI endpoint with MCP-compatible metadata and integrates with OpenTelemetry for observability.
Repository Info
About This Server
FaceTron is a high-performance face embedding server using ONNX Runtime, supporting dynamic multi-model loading, offline deployment, and scalable environments. It exposes an OpenAPI endpoint with MCP-compatible metadata and integrates with OpenTelemetry for observability.
Model Context Protocol (MCP) - This server can be integrated with AI applications to provide additional context and capabilities, enabling enhanced AI interactions and functionality.
Documentation
FaceTron MCP Server
FaceTron is a high-performance, modular face(recognition) embedding server with ONNX Runtime support, featuring dynamic multi-model loading, take your models offline, built to run anywhere, from local machines to scalable cloud environments. It also provides an OpenAPI Specification (OAS) endpoint with MCP-compatible metadata, and with OpenTelemetry integration.
Added lightweight ONNX models under the models/ directory for testing purposes. Replace or mount this directory with actual high-resolution face models such as 1k3d68, 2d106det, arcface, buffalo, genderage, glintr100, and scrfd_10g_bnkps for full functionality.
🚀 Features
-
🔗 APIs:
- GET /models: Get all loaded models from registry.
- POST /infer: Get embeddings from an image.
- POST /infer_visualize: Get boxed image with face detection markings.
- GET /download: Download boxed image with face detection markings.
- GET /openapi.json: Get OpenAPI v3.1.0 JSON spec for MCP metadata.
-
🧠 Face Embedding Inference
Accepts face images, detects faces, and returns clean, standardized vector embeddings. -
📦 Dynamic Multi-Model Support
Load multiple ONNX models (e.g., ArcFace, SCRFD, Glint360K) via/modelswith plug-and-play architecture. -
🔍 Face Detection & Visualization
Returns original images with bounding boxes and aligned face crops for inference validation. -
📚 OpenAPI 3.1 & Swagger UI
Easy integration via built-in interactive API documentation. -
🏷️ Out-of-the-Box OpenTelemetry Tracing Integrated support for OpenTelemetry with OTLP exporter (Jaeger compatible). Use setup_telemetry() in main.py to enable tracing automatically, or disable via the DISABLE_OTEL=true environment variable
-
🧩 Future-Ready Modular Design
Clean architecture withservices/,routes/, andutils/layers—ready for:- TensorFlow, PyTorch runtime extension
- Deployment into model orchestration pipelines
-
🧠 MCP Metadata Support
Exposes model info in OpenAPI spec for compliance with Model Connector Protocol (MCP) agents.
🧱 Tech Stack
- FastAPI (async web server)
- ONNX Runtime (model execution)
- Python 3.9+
- Dockerized & DevOps-ready
🗂 Project Structure
facetron/
├── main.py # FastAPI app entrypoint
├── routes.py # API route definitions
├── services/ # Core face detection & embedding logic
│ ├── face_detection_service.py
│ ├── face_embedding_service.py
│ └── model_registry_service.py
├── utils/ # Utility helpers
│ └── image_utils.py
├── models/ # Pretrained ONNX models (test models included)
│ ├── 1k3d68.onnx
│ ├── 2d106det.onnx
│ ├── arcface.onnx
│ ├── buffalo.onnx
│ ├── genderage.onnx
│ ├── glintr100.onnx
│ └── scrfd_10g_bnkps.onnx
├── tester/ # Client script for local testing
│ └── local_tester_client.py
├── temp/ # Temp storage for incoming image files
├── resources/ # Visual assets for API & architecture
│ └── *.png, *.jpg
├── requirements.txt
├── docker-compose.yml
├── Dockerfile
├── .env
├── readme.md
├── telemetry_utils.py # setup telemetry configs
└── license
⚙️ Setup
1. Clone and navigate
git clone https://github.com/13shivam/facetron.git
cd facetron
2. Add your ONNX models to /models
Make sure your models implement a .get_embedding(np.ndarray) -> np.ndarray interface via a wrapper.
3. Add/Remove OpenTelemetry Configuration
The server includes built-in OpenTelemetry, Telemetry is enabled on startup if setup_telemetry() is called in main.py.
Disable Telemetry: Set the environment variable DISABLE_OTEL=true before starting the server and skip invoking setup_telemetry().
Run with docker compose/pip
Create .env
MODEL_DIR=models
LOG_LEVEL=info
OTEL_EXPORTER_OTLP_ENDPOINT=http://host.docker.internal:4317
DISABLE_OTEL=false
Option 1: Start server using docker
docker-compose up -d
Access APIs via Swagger
http://127.0.0.1:8000/docs#/
Access Jaeger APIs explore trace
http://localhost:16686/search
Option 2: Start server using pip
pip install -r requirements.txt
python main.py
...
INFO: Will watch for changes in these directories: ['/facetron']
INFO: Uvicorn running on http://0.0.0.0:8000 (Press CTRL+C to quit)
INFO: Started reloader process [11385] using StatReload
INFO: Started server process [11388]
INFO: Waiting for application startup.
INFO: Application startup complete.
(logs contd...)
#run tester script
python local_tester_client.py
...
200
{'message': 'Image saved successfully', 'saved_image_path': '/facetron/temp/temp_276ce9b8.jpg'}
200
{'message': 'Image saved successfully', 'saved_image_path': '/facetron/temp/temp_47179b1b.jpg'}
200
{'message': 'Image saved successfully', 'saved_image_path': '/facetron/temp/temp_feb98d6b.jpg'}
200
{'message': 'Image saved successfully', 'saved_image_path': '/facetron/temp/temp_5e4a77e5.jpg'}
...
Sequence Diagram All APIs in test order
!My Local Image
Swagger view All APIs
!My Local Image
Get API: Get All loaded Models
!get_all_models_configured.png
Infer API: returns map of face_ids with embeddings detected
!infer_api.png
Infer API Visualize: returns boxed image
!infer_visualize.png
Jaeger Trace Visualisation
!opentelemetry_jaegar.jpg
Sample Images
!temp_1cc1baa9.jpg !temp_f16f0348.jpg
🧪 Example Inference
curl --location 'http://localhost:8000/infer' --header 'Content-Type: application/json' --form 'model_name="buffalo"' --form 'image_format="jpg"' --form 'file=@"/path/to/your/image.jpg"'
Example response:
{
"faces": [
{
"face_id": "15f9b282-ea6f-43f5-a60f-0e1a3b422fcd",
"bbox": [403, 46, 465, 108],
"embedding": [0, 0, 3434, -343...]
}
]
}
Performance Report
!console_log.jpg
📝 License
This project is licensed under the MIT License - see the license file for details.
Important Notice Regarding Open Source Dependencies:
This project relies on various open-source models and libraries, each with its own licensing terms. It is the user's responsibility to understand and adhere to the specific licenses of all the open-source components they choose to use. Consult the individual licenses provided by the respective model and library providers.
Quick Start
Clone the repository
git clone https://github.com/13shivam/facetronInstall dependencies
cd facetron
npm installFollow the documentation
Check the repository's README.md file for specific installation and usage instructions.
Repository Details
Recommended MCP Servers
Discord MCP
Enable AI assistants to seamlessly interact with Discord servers, channels, and messages.
Knit MCP
Connect AI agents to 200+ SaaS applications and automate workflows.
Apify MCP Server
Deploy and interact with Apify actors for web scraping and data extraction.
BrowserStack MCP
BrowserStack MCP Server for automated testing across multiple browsers.
Zapier MCP
A Zapier server that provides automation capabilities for various apps.