Skip to main content

Core Concepts

Full Stack Data Observability

Full Stack Data observability is a holistic approach for maintaining complex data ecosystems. While classic data observability systems focus on the data quality aspects, full stack systems monitor the software & code that generates & consumes the data.

Once monitoring is shifted from the data to the software the user gains visibility to:

  1. The Data application health & stability
  2. The performance, resource efficiency & cost of the data applications
  3. The quality of the processed data tested for the specific requirements of each application accessing it.

AI for data platform insights, optimizations & data incidents

Modern data platforms & pipelines are complex and the teams that run them prefer to focus on the business logic & value they are expected to deliver and not on quality checks and optimizations.

definity monitors hundreds of metrics covering all aspects: Execution health, Data health & resource utilization, learns the behavior of your pipelines and uses AI-ML to generate tests and optimizations accordingly:

  1. Data Quality Tests
  2. Execution Health Tests
  3. Health insights
  4. Resource utilization recommendations

Real-time proactive observability

Unlike traditional data quality tools definity does not run on the data after the pipeline has ended. Instead, it is injected to run as part of the pipeline itself with negligible footprint. This allows real-time analysis & proactive response to incidents. Below are some examples:

Real time analysisActive interventionImpact
Identify stale input data before the application started to transform itPreempt the pipeline from runningResources saved & no downstream impact of stale data
Identify stuck jobs earlyStop the stuck jobs automaticallyMeet SLA, allow rerun & save wasted resources
Identify faulty data output automatically in runtimeDivert final output to an alternate locationAvoid downstream contamination while allowing debug & approval