Company
-
September 27, 2022

Announcing Bigconfig and developer tools for enterprise data monitoring as code

We’re thrilled to announce the general availability of Bigconfig, a powerful new way for Bigeye customers to define data monitoring as code and automatically apply it across their enterprise data warehouse.

Kendall Lovett
Get Data Insights Delivered
Join hundreds of data professionals who subscribe to the Data Leaders Digest for actionable insights and expert advice.
Stay Informed
Sign up for the Data Leaders Digest and get the latest trends, insights, and strategies in data management delivered straight to your inbox.
Get the Best of Data Leadership
Subscribe to the Data Leaders Digest for exclusive content on data reliability, observability, and leadership from top industry experts.

Get the Best of Data Leadership

Subscribe to the Data Leaders Digest for exclusive content on data reliability, observability, and leadership from top industry experts.

Stay Informed

Sign up for the Data Leaders Digest and get the latest trends, insights, and strategies in data management delivered straight to your inbox.

Get Data Insights Delivered

Join hundreds of data professionals who subscribe to the Data Leaders Digest for actionable insights and expert advice.

We’re thrilled to announce the general availability of Bigconfig, a powerful new way for Bigeye customers to define data monitoring as code and automatically apply it across their enterprise data warehouse.

Bigconfig is the first monitoring-as-code solution to support enterprise-scale data observability.  It was designed in partnership with data engineering teams who use Bigeye on hundreds to thousands of tables and need to integrate data observability into their existing version control and code review processes. Bigconfig includes declarative formatting, dynamic data tagging, and reusable monitoring definitions so customers can define what they want to monitor and how they want to monitor it with just a few lines of human-readable YAML.

As a team of engineers, we love that Bigeye gives us the option to create version-controlled data monitoring as code with an elegant, ‘Terraform-like’ solution. With Bigconfig, we use a  simple YAML file to define data monitoring rules and then let Bigeye automatically apply them across our entire data warehouse, including new tables that come online.

Simon Dong, Udacity, Senior manager, Data engineering

With Bigconfig, data observability is:

  • Dynamic and automated: Deploy metrics automatically on new data that matches your specifications with dynamic tagging
  • Flexible: Specify and customize your monitoring with full granular control or apply any of Bigeye’s 60+ metrics out of the box
  • Version controlled: Manage your data observability from a central location and track changes with version-controlled audit logs
  • Fast and easy: Use optimized, human-readable YAML so there’s less code to write and no new language to learn
  • Integrated: Manage your entire Bigeye operation without leaving the command line

☝️ See a Bigconfig YAML template ☝️

More monitoring, less configuration code

As we developed Bigconfig, we strived to ensure “as code” never became “too much code.” Other monitoring as code solutions require the user to specify each table to monitor and which checks to apply for each, making them impractical and unsuitable for enterprise applications. Bigconfig uses human-readable YAML so there’s no domain-specific language to learn, intuitive tagging with simple wildcards so you can specify and group tables to monitor across your warehouse, and the ability to customize and save metrics so you can apply them globally or directly to specific groups of tables.

Deeply customizable data observability with developer tools

In addition to Bigconfig, Bigeye customers now have access to Bigeye-CLI—an ergonomic command-line interface—and a robust set of developer tools so they can integrate data observability into any stack or workflow, no matter how custom. Bigeye-CLI offers access to Bigeye’s full automated monitoring, anomaly detection, and data diff’ing capabilities without leaving the command line. Adding outlier detection to thousands of tables, integrating blue-green testing into your data model’s CI/CD pipeline, or validating the migration of thousands of tables into a new data warehouse can all be automated in a few simple commands.

Complete control in an ergonomic API

Data engineers can also choose to interact with Bigeye directly through a friendly and familiar REST API, making it easy to extend Bigeye into a custom workflow. Are you using a custom ETL orchestrator and want to build a circuit-breaker pattern? No problem. Want to pull Bigeye monitoring data into a third-party tool like Datadog for additional analysis? The Bigeye REST API makes it fast and easy.

Deploy data observability right from your Airflow DAGs

For data teams orchestrating dbt jobs with Airflow, Bigeye offers several simple yet powerful Airflow operators to configure and deploy Bigeye quality checks right from an Airflow DAG.  

Bigeye Airflow Operator

Python SDK for data science notebooks

Data Science teams are responsible for the performance of their models but they are often blind to the quality of the data coming into them. Bigeye provides a Python SDK that makes it easy to deploy Bigeye monitoring directly from a data science notebook like Jupyter or Hex for integrated data validation and confidence.

Ready to get started? Register for a live overview or request a demo to see Bigconfig in action.

share this episode
Resource
Monthly cost ($)
Number of resources
Time (months)
Total cost ($)
Software/Data engineer
$15,000
3
12
$540,000
Data analyst
$12,000
2
6
$144,000
Business analyst
$10,000
1
3
$30,000
Data/product manager
$20,000
2
6
$240,000
Total cost
$954,000
Role
Goals
Common needs
Data engineers
Overall data flow. Data is fresh and operating at full volume. Jobs are always running, so data outages don't impact downstream systems.
Freshness + volume
Monitoring
Schema change detection
Lineage monitoring
Data scientists
Specific datasets in great detail. Looking for outliers, duplication, and other—sometimes subtle—issues that could affect their analysis or machine learning models.
Freshness monitoringCompleteness monitoringDuplicate detectionOutlier detectionDistribution shift detectionDimensional slicing and dicing
Analytics engineers
Rapidly testing the changes they’re making within the data model. Move fast and not break things—without spending hours writing tons of pipeline tests.
Lineage monitoringETL blue/green testing
Business intelligence analysts
The business impact of data. Understand where they should spend their time digging in, and when they have a red herring caused by a data pipeline problem.
Integration with analytics toolsAnomaly detectionCustom business metricsDimensional slicing and dicing
Other stakeholders
Data reliability. Customers and stakeholders don’t want data issues to bog them down, delay deadlines, or provide inaccurate information.
Integration with analytics toolsReporting and insights

Get the Best of Data Leadership

Subscribe to the Data Leaders Digest for exclusive content on data reliability, observability, and leadership from top industry experts.

Stay Informed

Sign up for the Data Leaders Digest and get the latest trends, insights, and strategies in data management delivered straight to your inbox.

Get Data Insights Delivered

Join hundreds of data professionals who subscribe to the Data Leaders Digest for actionable insights and expert advice.

Join the Bigeye Newsletter

1x per month. Get the latest in data observability right in your inbox.