Our Competition Data Observatory is a fully automated, open source, open data observatory that produces new indicators from open data sources and experimental big data sources, with authoritative copies and a modern API.
Our observatory is monitoring the certain segments of the European economy, and develops tools for computational antitrust in Europe. We take a critical SME-, intellectual property policy and competition policy point of view automation, robotization, and the AI revolution on the service-oriented European social market economy.
We would like to create early-warning, risk, economic effect, and impact indicators that can be used in scientific, business and policy contexts for professionals who are working on re-setting the European economy after a devastating pandemic and in the age of AI. We would like to map data between economic activities (NACE), antitrust markets, and sub-national, regional, metropolitian area data.
Get involved in services: our ongoing projects, team of contributors, open-source libraries and use our data for publications. See some use cases.
Follow news about us or the more comprehensive Data & Lyrics blog.
Contact us .
Download our competition presentation
Our Product/Market Fit was validated in the world’s 2nd ranked university-backed incubator program, the Yes!Delft AI Validation Lab.
Automated Data Services
Check out our more general Data & Lyrics blog.
See what we are doing, why we are doing it, and how you can contribute.
API to our data, our sources, how we process and validate it, how we increase its quality.
International Open Data Day 2021:
Data, Maps Produced & Their Uses
Our first use case is about identifying potential political roadblocks for climate policies. We are combining survey data about attitudes to climate change policies with socio-economic coal mining and and voting data. We examine the relationship between voter attitudes and economic dependency on coal mining.
Our peer-reviewed, open source statistical software packages
We believe that transparency is the key to the highest data quality. We use only open source software. We open up the critical elements of our software for peer-review.
We use open-source software, there is no vendor lock-in.
Our data products go through many, automated (unit) tests, replacing countless error-prone human validation working hours.
The critical elements of our code go through external validation and peer-review by computational statisticians and data scientists.
The goal of retroharmonize is to facilitate retrospective (ex-post) harmonization of data, particularly survey data, in a reproducible manner.
Our peer-reviewed, open source statistical software packages
Conversations with users, contributors and volunteers.