Data Sharing Community
Welcome to the Portal of the CDQ Data Sharing CommunityThe CDQ Data Sharing Community is a trusted network of user companies to manage business partner data collaboratively. | |||||||||||||||||||||||||||||||||||||||||||||||||||||
What's new? (RSS)Integration of California Business Register (13 March 2025)SummaryWe are excited to announce the integration of the California Business Register into CDQ's Business Partner Lookup service. This new data source enhances our coverage of businesses registered in California, providing valuable insights to support compliance, data enrichment, and risk management efforts. DetailsThe California Business Register data is sourced from the California Secretary of State's BizFile Online database and the publicly available business entity dataset. It contains over 8 million records and is updated weekly to ensure the most current information is available. Each record may include the following details:
Augmentation Configuration Support in Generate Golden Record Endpoint (11 March 2025)OverviewWe have enhanced the Generate Golden Record endpoint by introducing support for Augmentation Configuration. This improvement offers greater flexibility and control over the Golden Record generation process by allowing users to specify a configuration that defines how data should be enriched and prioritized New functionalityThe Generate Golden Record endpoint now accepts an optional parameter called configurationId in the request payload.
Payload example:
Benefits
This enhancement is now available for use in the API. For additional details or support, please refer to the API documentation. Seed Updates for Customer-Modified Business Partners (7 March 2025)We are introducing a new feature that automatically generates a seed update whenever a business partner (BP) is modified within the customer mirror. This update allows you to quickly compare the latest BP details against those in the reference data source.
Data modelAn important prerequisite for collaborative data management is a common understanding of the shared data. For the CDQ Data Sharing Community, this common understanding is specified by the CDQ Data Model. The concepts of this model are defined and documented in this wiki which can be used as a business vocabulary. Moreover, the wiki provides a machine-readable interface to reuse this metadata by using semantic annotations. ![]() Data maintenance proceduresA procedure is a common standard or "how-to" for a specific data management task. Within the CDQ Data Sharing Community, companies agree on such procedures to ensure similar rules and guidelines for similar tasks. For several countries, the CDQ Wiki provides such information, e.g. data quality rules, trusted information sources, legal forms, or tax numbers. Try
or select another country from the list. |
Data sources
Metadata and Standards: Metadata-driven Data QualityData quality plays a pivotal role in ensuring compliance with legal, regulatory, and industry standards. One of the core challenges in achieving high data quality is adhering to dynamic data requirements that evolve due to changes in national regulations. These requirements vary by country, making it essential for businesses to track and update compliance criteria continuously. In many countries, official company information is available as Open Data, but the lack of a standardized data model or provision method complicates the process of integrating this data. The Data Sharing Community actively collaborates to identify global data requirements and reference data sources, whether Open Data or commercial.
Data Quality RulesTransformation of human-documented data requirements into executable data quality rules is mostly a manual IT effort. Changing requirements cause IT efforts again and again. Some checks, e.g. tax number validity (not just format!), require external services. Other checks, e.g. validity of legal forms, require managed reference data (e.g. legal forms by country, plus abbreviations). Continuous data quality assurance (i.e. batch analyses) and real-time checks in workflows often use different rule sets. Data requirements and related reference data are collected and updated collaboratively by the Data Sharing Community. Data quality rules are derived from these requirements automatically. All data quality rules are executed behind 1 interface, in real-time. Batch jobs and single-record checks use the same rule set and can be integrated by APIs. For proving that a data quality rule is content-wise correct we maintain supporting document(s) per data quality rule which share the rule's source. This could be:
We manage the URL (if any), a screenshot of the relevant parts (if any) and the source's name (e.g. Community member data standard, European Commission, National ....) See Identifier format invalid (SIREN (France)) as an exemplary rule that was specified and implemented based on information provided by the OECD. |