By using this site, you agree to the Privacy Policy and Terms of Use.
Accept
World of SoftwareWorld of SoftwareWorld of Software
  • News
  • Software
  • Mobile
  • Computing
  • Gaming
  • Videos
  • More
    • Gadget
    • Web Stories
    • Trending
    • Press Release
Search
  • Privacy
  • Terms
  • Advertise
  • Contact
Copyright © All Rights Reserved. World of Software.
Reading: How Agoda Unified Multiple Data Pipelines Into a Single Source of Truth
Share
Sign In
Notification Show More
Font ResizerAa
World of SoftwareWorld of Software
Font ResizerAa
  • Software
  • Mobile
  • Computing
  • Gadget
  • Gaming
  • Videos
Search
  • News
  • Software
  • Mobile
  • Computing
  • Gaming
  • Videos
  • More
    • Gadget
    • Web Stories
    • Trending
    • Press Release
Have an existing account? Sign In
Follow US
  • Privacy
  • Terms
  • Advertise
  • Contact
Copyright © All Rights Reserved. World of Software.
World of Software > News > How Agoda Unified Multiple Data Pipelines Into a Single Source of Truth
News

How Agoda Unified Multiple Data Pipelines Into a Single Source of Truth

News Room
Last updated: 2026/01/14 at 8:29 AM
News Room Published 14 January 2026
Share
How Agoda Unified Multiple Data Pipelines Into a Single Source of Truth
SHARE

Agoda recently described how it consolidated multiple independent data pipelines into a centralized Apache Spark-based platform to eliminate inconsistencies in financial data. The company implemented a multi-layered quality framework that combines automated validations, machine-learning-based anomaly detection, and data contracts with upstream teams to ensure the accuracy of financial metrics used in statements and strategic planning, while processing millions of daily booking transactions.

The problem emerged from a typical enterprise pattern: Agoda’s Data Engineering, Business Intelligence, and Data Analysis teams had each developed separate financial data pipelines with independent logic and definitions. While this offered simplicity and clear ownership, it created duplicate processing and inconsistent metrics across the organization. As Warot Jongboondee from Agoda’s engineering team explains, these discrepancies “could potentially impact Agoda’s financial statements.”


Separate financial data pipelines (source)

The solution, called Financial Unified Data Pipeline (FINUDP), establishes a single source of truth for financial data, including sales, cost, revenue, and margin calculations. Built on Apache Spark, the system delivers hourly updates to downstream teams for reconciliation and financial planning. The consolidation required significant effort: aligning stakeholders across product, finance, and engineering on shared data definitions was time-intensive, and the initial runtime of five hours required optimization through query tuning and infrastructure adjustments to reach approximately 30 minutes.



Unified Financial Data Pipeline (FINUDP) architecture (source)

Agoda’s quality framework implements multiple defensive layers. Automated validations check data tables for null values, range constraints, and data integrity. When business-critical rules fail, the pipeline automatically halts rather than processing potentially incorrect data. The team uses Quilliup to compare target and source tables. Data contracts with upstream teams define required rules; violations trigger immediate alerts. Machine learning models monitor patterns to identify anomalies. A three-tier alerting system ensures rapid response via email, Slack notifications, and an internal tool that escalates to Agoda’s 24/7 Network Operations Center when updates lag.

The approach aligns with broader industry trends. According to recent industry research, 64% of organizations cite poor data quality as their biggest challenge, with data contracts emerging as what Gartner calls an “increasingly popular way to manage, deliver, and govern data products.” These formal agreements between producers and consumers define expectations for schemas and quality requirements.

The consolidation came with explicit trade-offs. Development velocity decreased because changes now require testing the entire pipeline. Data dependencies mean the full pipeline waits for all upstream datasets before proceeding. Thorough documentation and stakeholder consensus slowed implementation but built trust. Jongboondee noted that centralization “demands tighter coordination and careful change management at every step.”

The system currently achieves 95.6% uptime and targets 99.5% availability. All changes undergo shadow testing where queries run on both proposed and previous versions, with results compared within merge requests. A dedicated staging environment mirrors production, allowing teams to test before release.

The FINUDP initiative demonstrates how organizations handling critical business data at scale are moving beyond ad-hoc quality checks toward comprehensive, architecturally enforced reliability systems that prioritize consistency and auditability over development speed, characteristics increasingly essential as financial data feeds reporting, machine learning models, and regulatory compliance processes.

Sign Up For Daily Newsletter

Be keep up! Get the latest breaking news delivered straight to your inbox.
By signing up, you agree to our Terms of Use and acknowledge the data practices in our Privacy Policy. You may unsubscribe at any time.
Share This Article
Facebook Twitter Email Print
Share
What do you think?
Love0
Sad0
Happy0
Sleepy0
Angry0
Dead0
Wink0
Previous Article TikTok Shop Showed Me Search Suggestions for Products With Nazi Symbolism TikTok Shop Showed Me Search Suggestions for Products With Nazi Symbolism
Next Article UK police blame Microsoft Copilot for intelligence mistake UK police blame Microsoft Copilot for intelligence mistake
Leave a comment

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

Stay Connected

248.1k Like
69.1k Follow
134k Pin
54.3k Follow

Latest News

Linux 7.0 To Focus Just On Full & Lazy Preemption Models For Up-To-Date CPU Archs
Linux 7.0 To Focus Just On Full & Lazy Preemption Models For Up-To-Date CPU Archs
Computing
I Used Smart Glasses to Cover CES. Here’s What Actually Worked
I Used Smart Glasses to Cover CES. Here’s What Actually Worked
News
London’s Tube network extends 4G/5G connectivity | Computer Weekly
London’s Tube network extends 4G/5G connectivity | Computer Weekly
News
Reverse Engineering the AI Supply Chain: Why Regex Won’t Save Your PyTorch Models | HackerNoon
Reverse Engineering the AI Supply Chain: Why Regex Won’t Save Your PyTorch Models | HackerNoon
Computing

You Might also Like

I Used Smart Glasses to Cover CES. Here’s What Actually Worked
News

I Used Smart Glasses to Cover CES. Here’s What Actually Worked

16 Min Read
London’s Tube network extends 4G/5G connectivity | Computer Weekly
News

London’s Tube network extends 4G/5G connectivity | Computer Weekly

7 Min Read
The year of ultra-ultraportable Windows laptops is here
News

The year of ultra-ultraportable Windows laptops is here

11 Min Read
Tesla to stop selling FSD as a standalone package and switch to subscription only
News

Tesla to stop selling FSD as a standalone package and switch to subscription only

3 Min Read
//

World of Software is your one-stop website for the latest tech news and updates, follow us now to get the news that matters to you.

Quick Link

  • Privacy Policy
  • Terms of use
  • Advertise
  • Contact

Topics

  • Computing
  • Software
  • Press Release
  • Trending

Sign Up for Our Newsletter

Subscribe to our newsletter to get our newest articles instantly!

World of SoftwareWorld of Software
Follow US
Copyright © All Rights Reserved. World of Software.
Welcome Back!

Sign in to your account

Lost your password?