Why?

Is it not?
The Idea
A journalist knows how to identify “security framing” in news coverage. A policy analyst knows what counts as “meaningful stakeholder engagement” in legislative proposals. A bureaucrat knows whether a grant application is properly filled out. That expertise lives in their heads, maybe in spreadsheets and notes. This works great for tens of documents. At hundreds or thousands, you’re either stuck or you need to hire engineers. Meanwhile, sophisticated analysis infrastructure, the kind that lets you systematically apply analytical frameworks at scale, has only been available to well-funded institutions. This is where HQ shines: define your analytical questions in plain language. Apply them at scale. Your schemas are your “lenses” you use to see the data. They are shareable, transparent, and improvable. Other researchers can see exactly how you defined your framework, critique it, refine it, or apply it to their own data. For example: Imagine as a journalist, you are analysing 200 news articles. You create a schema:




How It Works
1
Ingest Content
Upload files, URLs, search results, RSS feeds
2
Define Schemas
Describe what information to extract in plain language (see above)
3
Run Analysis
Use AI to apply your schema at scale across documents
4
Explore Results
View through tables, visualizations, maps, or export the data
Supported Formats
PDFs, web articles, text, CSV, RSS feeds
Coming Soon
Images, audio, email inbox ingestion

Chat & MCP
The Chat & MCP (Model Context Protocol) centralizes all platform tools (asset management, schema based analysis, vector search, content ingestion) into a unified analysis access point that lets you work with structured analytical methods through conversational AI. This combination allows for various workflows for researching & synthesizing data. Editing MCP configs (upcoming on a per user level) furthermore allows you to grant your LLM access to almost arbitrary outside data and use it for your analysis.Infospaces & Vector Storage
Your assets, schemas and analysis results are scoped to “Information Spaces” that you can use to curate information. Each Information Space is a dedicated vector space. Use vector embeddings from local or cloud models to search through your data, cluster it (upcoming) and find duplicates.Getting Started
Option 1: Use the Hosted Instance
The easiest way to start. We host the infrastructure, you bring your own LLM API keys.1
Register
Sign up at open-politics.org/accounts/register
2
Add API Keys
Configure your LLM API keys on the home page
3
Start Analysing
Upload content and create your first schema
Your account also works on the forum for community support.
Option 2: Self-Host with Docker
For privacy, customization, or institutional requirements. Run everything on your own infrastructure.Deployment Flexibility
Fully Local
Run everything on your own hardware. Good for air gapped environments or complete data control.
Hybrid
Run the application locally but use managed services (AWS RDS, Upstash Redis, S3) to reduce operational burden.
Kubernetes
Architecture
The platform is built from several independent services that work together. You can run them all locally or mix local and managed services.Core Components
Component | What It Does | Technology |
---|---|---|
Backend | API, analysis jobs, MCP server | FastAPI + Python |
Frontend | Web interface | Next.js + React |
Worker | Background processing for large jobs | Celery |
Database | Data storage with vector search | PostgreSQL + PGVector |
Object Storage | File storage for uploads | MinIO (S3-compatible) |
Cache/Queue | Session management, job queues | Redis |
Geocoding | Location extraction and mapping | Pelias |
LLM (optional) | Local AI inference | Ollama |
LLM Support
Connect any of these AI providers:Anthropic
Claude Sonnet 4.5, etc. (Best Experience with tools)
OpenAI
GPT-5 etc. (Not sure)
Gemini models (Large context, speed and modality capabilities)
Ollama
Run models locally: Llama, OAI OSS, Qwen, etc. (Best Privacy)
Contributing
We’re building this in the open. The codebase, analytical methods, and documentation are all public and improvable.Code Contributions
Report bugs, suggest features, contribute code
Documentation
Improve docs, add examples, share schemas
Community
Join forum discussions, share knowledge
Research
Build analytical frameworks, publish findings
Contact & Community
License
AGPLv3, see LICENSE This means you can use, modify, and distribute this software, but any modifications or services built on it must also be open source. You can get an enterprise license for private use modifications which are not publicly deployed for one year at a time under strict ethical guidelines.Project Origins & Story
Talk: Open Source Political Intelligence
CCCB Datengarten Presentation