Custom Styles

Dark Clouds: Why Enterprises are re-evaluating multi-cloud architecture

This October, the digital world learned that even clouds can fail. For more than a decade, a handful of technical giants have been the invisible gravity that holds the digital world together.

November 13, 2025
Why Enterprises Are Re-Evaluating Multi-Cloud Architecture

For more than a decade, a handful of technical giants have been the invisible gravity that holds the digital world together. Together, they power over half of the world’s cloud workloads with Amazon S3 alone peaking at nearly 1 petabyte per second in bandwidth. With average uptimes measured at 99.999% and data centers spanning every continent, these clouds have made reliability feel almost ordinary.

When you order a meal, book a flight, or send a message, you don’t wonder where the data lives. You expect it to work – instantly, everywhere, all the time. That’s the brilliance, and the paradox, of hyperscale computing: the better it gets, the less we remember it’s there.

So, when two giants falter, the world didn’t just face downtime – it felt disconnected from its digital heartbeat. Snapchat went silent. Coinbase froze. Heathrow check-ins halted. Webflow  blinked out.

And Meredith Whittaker, the President of Signal, reminded the internet in a now-viral post, “There are only a handful of entities on Earth capable of offering the kind of global infrastructure a service like Signal requires.”  

She’s right, and that’s precisely the problem. If so much of the world runs on so few providers, what happens when the sky blinks?  

In this piece, we’ll explore what the recent AWS and Azure outages teach us about dependency, and why multi-cloud resilience may be the only way forward. And how doing it right requires re-thinking how enterprises design for continuity itself.

Why even the most resilient systems go down

For global enterprises, only three cloud providers on the planet – AWS, Azure, and Google Cloud – offer true global reach with the compliance, scale, and performance demanded by millions of concurrent users and devices.

Their dominance wasn’t luck; it was engineered. Over the past decade, these hyperscalers built astonishingly resilient systems with unmatched global reach, distributing workloads across regions, synchronizing backups between data centers, and making downtime feel mythical.

As these three providers grew, they didn’t just sell compute – they sold confidence. The pitch to enterprises was simple: stay within our ecosystem, and you’ll never go down. To prove it, they built seamless multi-region replication, allowing workloads and databases to mirror across geographies in real time. A failover in Oregon could instantly shift to Virginia; a backup in Singapore could keep services running if Tokyo stumbled. Multi-region became both a technological marvel and a marketing assurance – proof that a single-cloud strategy could deliver global continuity without the complexity of managing multiple vendors.  

That’s why multi-region architecture became the de facto safety net. Within a single cloud system, creating secondary zones and failover systems was a simple, cost-effective, and largely automated process. For most organizations, it was the rational resilient architecture. For a decade, it worked beautifully.

Until this October.

The AWS and Azure outages didn’t start in a data center or a regional cluster. They began in the global orchestration layers – the digital data traffic control systems that manage routing, authentication, and coordination across every region. When those systems blinked, every dependent region blinked with them.

Essentially, the same architecture that made cloud redundancy easy also created a dependency that no customer of these three service providers can escape. As Meredith Whittaker added in her post, “Cloud infrastructure is a choke point for the entire digital ecosystem.

Her words capture the uncomfortable truth that the strength of cloud infrastructure – its globe-straddling, unifying scale – has become its vulnerability. Control-plane failures have happened before, but they were rare enough and systems recovered fast enough that single-vendor, multi-region strategies felt sufficient. The events of October changed that calculus. Even the global scaffolding of these global cloud providers can falter – and when it does, no amount of intra-cloud redundancy can substitute for independence.

If multi-region resilience can no longer guarantee uptime, the next evolution isn’t redundancy; it is reinvention. Multi-cloud resilience – not as a buzzword, but as a design discipline that treats portability, data liquidity, and provider-agnostic uptime as first-class principles of modern architecture.

Multi-cloud is the answer – and why it’s hard

For years, multi-cloud has been the white whale of IT strategy – admired from afar, rarely captured. The premise was simple: distribute workloads across providers to minimize risk, prevent downtime, and avoid over-reliance on a single vendor.

The challenge was never conviction – it was complexity. Because true multi-cloud isn’t just about having backups elsewhere – it’s about keeping two living systems in sync.

Every transaction, every log, every user action must decide: Do I replicate this now or later? To which system? In what format? When one cloud slows or fails, automation must not only redirect workloads but also determine what state of data to recover, when to switch back, and how to avoid conflicts when both sides come online again.

The system needs to determine which version of a record is authoritative, how to maintain integrity during mid-flight transactions, and how to ensure compliance when one region’s laws differ from those of another. Testing these scenarios is notoriously difficult. Simulating a global outage can disrupt production; not testing leaves blind spots.

This is why multi-cloud used to be a luxury reserved for a few technology giants with large engineering teams. For everyone else, the math – and the risk – didn’t work.

Cloud’s rise, after all, was powered by convenience. AWS, Azure, and Google Cloud offered a unified ecosystem where scale, replication, and resilience were built in. They let engineering teams move faster by outsourcing undifferentiated heavy lifting – from storage and security to global networking. Within those walls, resilience felt like a solved problem.

Due to this complexity and convenience, single-vendor multi-region architectures have become the gold standard. They were cost-effective, automated, and easy to manage. The architecture made sense – until it didn’t.

The October outages revealed the blind spot. And that is where the conversation shifts.
This isn’t about distrust in cloud vendors – their reliability remains extraordinary. It’s about responsible risk management in a world where that reliability can no longer be assumed as absolute.

Forward-looking leaders are now asking a new question:
Can emerging technologies finally make multi-cloud feasible – not as a hedge, but as a new standard for resilience?

That’s the opportunity. To transform what was once an engineering burden into a business imperative – to use automation, data fabrics, and AI-assisted operations to not just distribute workloads, but to create enterprise-grade confidence.

The Five Principles of true multi-cloud resilience

Modern enterprises don’t just run on data: they run on uninterrupted access to it.
In a world where customers expect every transaction, login, and workflow to be instantaneous, resilience has become the most accurate measure of trust.

That’s why multi-cloud matters. It’s the only architectural model that promises “always-up” systems – platforms capable of staying operational even when a primary cloud provider experiences disruption. By distributing workloads, data, and control across multiple providers, enterprises can insulate their business from global outages and deliver the reliability their customers already expect to be guaranteed. It would put enterprises back in the driver’s seat on their systems, rather than leaving them vulnerable to provider failures.

The question is no longer whether multi-cloud is desirable, but how it can be achieved without increasing complexity to the extent of making it unfeasible. Enterprises that succeed tend to follow five foundational principles – pragmatic guardrails for transforming resilience into a lasting architecture.

  1. Start at the Edge: Independent Traffic Control
    Resilience begins with control over routing. In most single-cloud designs, DNS, load balancing, and traffic steering live inside the provider’s control plane – the very layer that failed in October. A neutral, provider-independent edge – using external DNS and traffic managers – creates a first line of defense. When one cloud falters, requests can automatically shift to another entry point in seconds.
  1. Dual-Home Identity and Access
    Authentication outages often outlast infrastructure ones. Enterprises should maintain a secondary identity and secrets system – an auxiliary OIDC or SAML provider, or escrowed credentials – that can mint and validate tokens even if a cloud’s native IAM or Entra service goes dark.
  1. Make Data Liquid
    Data is the most complex system to move and the easiest to lose. True multi-cloud architecture treats data as a flowing asset, not as a static store. This means continuous replication across providers, standardized schemas, and automated reconciliation to keep operational data within defined RPO/RTO windows. Modern data fabrics and object storage replication make this feasible without doubling costs. AI-powered data pipelines can also provide schema standardization, indexing, and tagging at the point of ingesting, and prioritizing, routing, duplicating, and routing data with granular policy implementation with edge governance.
  1. Build Cloud-agnostic Application Layers
    Every dependency on proprietary PaaS services – queues, functions, monitoring agents – ties resilience to a single vendor. Abstracting the application tier with containers, service meshes, and portable frameworks ensures that workloads can be deployed or recovered anywhere, providing flexibility and scalability. Kubernetes, Kafka, and open telemetry stacks are not silver bullets, but they serve as the connective tissue of mobility.  
  1. Govern for Autonomy, not Abandonment
    Multi-cloud isn’t about rejecting providers; it is about de-risking dependence. That requires unified governance – visibility, cost control, compliance, and observability – that transcends vendor dashboards. Modern automation and AI-assisted orchestration can maintain policy consistency across environments, ensuring resilience without becoming operational debt.  

When these five principles converge, resilience stops being reactive and becomes a design property of the enterprise itself. It turns multi-cloud from an engineering aspiration into a business continuity strategy – one that keeps critical services available, customer trust intact, and the brand’s promise uninterrupted.

From pioneers to the possible

Not long ago, multi-cloud resilience was a privilege reserved for the few – projects measured in years, not quarters.

Coca-Cola began its multi-cloud transformation around 2017, building a governance and management system that could span AWS, Azure, and Google Cloud. It took years of integration and cost optimization for the company to achieve unified visibility across its environments.

Goldman Sachs followed, extending its cloud footprint from AWS into Google Cloud by 2019, balancing trading workloads on one platform with data analytics and machine learning on another. Their multi-cloud evolution unfolded gradually through 2023, aligning high-performance finance systems with specialized AI infrastructure.

In Japan, Mizuho Financial Group launched its multi-cloud modernization initiative in 2020, achieving strict financial-sector compliance while reducing server build time by nearly 80 percent by 2022.

Each of these enterprises demonstrated the principle: true continuity and flexibility are possible, but historically only through multi-year engineering programs, deep vendor collaboration, and substantial internal bandwidth.

That equation is evolving. Advances in AI, automation, and unified data fabrics now make the kind of resilience these pioneers sought achievable in a fraction of the time – without rebuilding every system from scratch.

Modern platforms like Databahn represent this shift, enabling enterprises to seamlessly orchestrate, move, and analyze data across clouds. They transform multi-cloud from merely an infrastructure concern into an intelligence layer – one that detects disruptions, adapts automatically, and keeps the enterprise operational even when the clouds above encounter issues.

Owning the future: building resilience on liquid data

Every outage leaves a lesson in its wake. The October disruptions made one thing unmistakably clear: even the best-engineered clouds are not immune to failure.
For enterprises that live and breathe digital uptime, resilience can no longer be delegated — it must be designed.

And at the heart of that design lies data. Not just stored or secured, but liquid – continuously available, intelligently replicated, and ready to flow wherever it’s needed.
Liquid data powers cross-cloud recovery, real-time visibility, and adaptive systems that think and react faster than disruptions.

That’s the future of enterprise architecture: always-on systems built not around a single provider, but around intelligent fabrics that keep operations alive through uncertainty.
It’s how responsible leaders will measure resilience in the next decade – not by the cloud they choose, but by the continuity they guarantee.

At Databahn, we believe that liquid data is the defining resource of the 21st century –  both the foundation of AI and the reporting layer that drives the world’s most critical business decisions. We help enterprises control and own their data in the most resilient and fault-tolerant way possible.

Did the recent outages impact you? Are you looking to make your systems multi-cloud, resilient, and future-proof? Get in touch and let’s see if a multi-cloud system is worthwhile for you.

Ready to unlock full potential of your data?
Share

See related articles

What is a SIEM?

A Security Information and Event Management (SIEM) system aggregates logs and security events from across an organization’s IT infrastructure. It correlates and analyzes data in real time, using built-in rules, analytics, and threat intelligence to identify anomalies and attacks as they happen. SIEMs provide dashboards, alerts, and reports that help security teams respond quickly to incidents and satisfy compliance requirements. In essence, a SIEM acts as a central security dashboard, giving analysts a unified view of events and threats across their environment.

Pros and Cons of SIEM

Pros of SIEM:

  • Real-time monitoring and alerting for known threats via continuous data collection
  • Centralized log management provides a unified view of security events
  • Built-in compliance reporting and audit trails simplify regulatory obligations
  • Extensive integration ecosystem with standard enterprise tools
  • Automated playbooks and correlation rules accelerate incident triage and response

Cons of SIEM:

  • High costs for licensing, storage, and processing at large data volumes
  • Scalability issues often require filtering or short retention windows
  • May struggle with cloud-native environments or unstructured data without heavy customization
  • Requires ongoing tuning and maintenance to reduce false positives
  • Vendor lock-in due to proprietary data formats and closed architectures

What is a Security Data Lake?

A security data lake is a centralized big-data repository (often cloud-based) designed to store and analyze vast amounts of security-related data in its raw form. It collects logs, network traffic captures, alerts, endpoint telemetry, threat intelligence feeds, and more, without enforcing a strict schema on ingestion. Using schema-on-read, analysts can run SQL queries, full-text searches, machine learning, and AI algorithms on this raw data. Data lakes can scale to petabytes, making it possible to retain years of data for forensic analysis.

Pros and Cons of Security Data Lakes

Pros of Data Lakes:

  • Massive scalability and lower storage costs, especially with cloud-based storage
  • Flexible ingestion: accepts any data type without predefined schema
  • Enables advanced analytics and threat hunting via machine learning and historical querying
  • Breaks down data silos and supports collaboration across security, IT, and compliance
  • Long-term data retention supports regulatory and forensic needs

Cons of Data Lakes:

  • Requires significant data engineering effort and strong data governance
  • Lacks native real-time detection—requires custom detections and tooling
  • Centralized sensitive data increases security and compliance challenges
  • Integration with legacy workflows and analytics tools can be complex
  • Without proper structure and tooling, can become an unmanageable “data swamp”  

A Hybrid Approach: Security Data Fabric

Rather than choosing one side, many security teams adopt a hybrid architecture that uses both SIEM and data lake capabilities. Often called a “security data fabric,” this strategy decouples data collection, storage, and analysis into flexible layers. For example:

  • Data Filtering and Routing: Ingest all security logs through a centralized pipeline that tags and routes data. Send only relevant events and alerts to the SIEM (to reduce noise and license costs), while streaming raw logs and enriched telemetry to the data lake for deep analysis.
  • Normalized Data Model: Preprocess and normalize data on the way into the lake so that fields (timestamps, IP addresses, user IDs, etc.) are consistent. This makes it easier for analysts to query and correlate data across sources.
  • Tiered Storage Strategy: Keep recent or critical logs indexed in the SIEM for fast, interactive queries. Offload bulk data to the data lake’s cheaper storage tiers (including cold storage) for long-term retention. Compliance logs can be archived in the lake where they can be replayed if needed.
  • Unified Analytics: Let the SIEM focus on real-time monitoring and alerting. Use the data lake for ad-hoc investigations and machine-learning-driven threat hunting. Security analysts can run complex queries on the full dataset in the lake, while SIEM alerts feed into a coordinated response plan.
  • Integration with Automation: Connect the SIEM and data lake to orchestration/SOAR platforms. This ensures that alerts or insights from either system trigger a unified incident response workflow.

This modular security data fabric is an emerging industry best practice. It helps organizations avoid vendor lock-in and balance cost with capability. For instance, by filtering out irrelevant data, the SIEM can operate leaner and more accurately. Meanwhile, threat hunters gain access to the complete historical dataset in the lake.

Choosing the Right Strategy

Every organization’s needs differ. A full-featured SIEM might be sufficient for smaller environments or for teams that prioritize quick alerting and compliance out-of-the-box. Large enterprises or those with very high data volumes often need data lake capabilities to scale analytics and run advanced machine learning. In practice, many CISOs opt for a combined approach: maintain a core SIEM for active monitoring and use a security data lake for additional storage and insights.

Key factors include data volume, regulatory requirements, budget, and team expertise. Data lakes can dramatically reduce storage costs and enable new analytics, but they require dedicated data engineering and governance. SIEMs provide mature detection features and reporting, but can become costly and complex at scale. A hybrid “data fabric” lets you balance these trade-offs and future-proof the security stack.

At the end of the day, rethinking SIEM doesn’t necessarily mean replacing it. It means integrating SIEM tools with big-data analytics in a unified way. By leveraging both technologies — the immediate threat detection of SIEM and the scalable depth of data lakes — security teams can build a more flexible, robust analytics platform.

Ready to modernize your security analytics? Book a demo with Databahn to see how a unified security data fabric can streamline threat detection and response across your organization.

The Old Guard of Data Governance: Access and Static Rules

For years, data governance has been synonymous with gatekeeping. Enterprises set up permissions, role-based access controls, and policy checklists to ensure the right people had the right access to the right data. Compliance meant defining who could see customer records, how long logs were retained, and what data could leave the premises. This access-centric model worked in a simpler era – it put up fences and locks around data. But it did little to improve the quality, context, or agility of data itself. Governance in this traditional sense was about restriction more than optimization. As long as data was stored and accessed properly, the governance box was checked.

However, simply controlling access doesn’t guarantee that data is usable, accurate, or safe in practice. Issues like data quality, schema changes, or hidden sensitive information often went undetected until after the fact. A user might have permission to access a dataset, but if that dataset is full of errors or policy violations (e.g. unmasked personal data), traditional governance frameworks offer no immediate remedy. The cracks in the old model are growing more visible as organizations deal with modern data challenges.

Why Traditional Data Governance Is Buckling  

Today’s data environment is defined by velocity, variety, and volume. Rigid governance frameworks are struggling to keep up. Several pain points illustrate why the old access-based model is reaching a breaking point:

Unmanageable Scale: Data growth has outpaced human capacity. Firehoses of telemetry, transactions, and events are pouring in from cloud apps, IoT devices, and more. Manually reviewing and updating rules for every new source or change is untenable. In fact, every new log source or data format adds more drag to the system – analysts end up chasing false positives from mis-parsed fields, compliance teams wrestle with unmasked sensitive data, and engineers spend hours firefighting schema drift. Scaling governance by simply throwing more people at the problem no longer works.

Constant Change (Schema Drift): Data is not static. Formats evolve, new fields appear, APIs change, and schemas drift over time. Traditional pipelines operating on “do exactly what you’re told” logic will quietly fail when an expected field is missing or a new log format arrives. By the time humans notice the broken schema, hours or days of bad data may have accumulated. Governance based on static rules can’t react to these fast-moving changes.

Reactive Compliance: In many organizations, compliance checks happen after data is already collected and stored. Without enforcement woven into the pipeline, sensitive data can slip into the wrong systems or go unmasked in transit. Teams are then stuck auditing and cleaning up after the fact instead of controlling exposure at the source. This reactive posture not only increases legal risk but also means governance is always a step behind the data. As one industry leader put it, “moving too fast without solid data governance is exactly why many AI and analytics initiatives ultimately fail”.

Operational Overhead: Legacy governance often relies on manual effort and constant oversight. Someone has to update access lists, write new parser scripts, patch broken ETL jobs, and double-check compliance on each dataset. These manual processes introduce latency at every step. Each time a format changes or a quality issue arises, downstream analytics suffer delays as humans scramble to patch pipelines. It’s no surprise that analysts and engineers end up spending over 50% of their time fighting data issues instead of delivering insights. This drag on productivity is unsustainable.

Rising Costs & Noise: When governance doesn’t intelligently filter or prioritize data, everything gets collected “just in case.” The result is mountains of low-value logs stored in expensive platforms, driving up SIEM licensing and cloud storage costs. Security teams drown in noisy alerts because the pipeline isn’t smart enough to distinguish signal from noise. For example, trivial heartbeat logs or duplicates continue flowing into analytics tools, adding cost without adding value. Traditional governance has no mechanism to optimize data volumes – it was never designed for cost-efficiency, only control.

The old model of governance is cracking under the pressure. Access controls and check-the-box policies can’t cope with dynamic, high-volume data. The status quo leaves organizations with blind spots and reactive fixes: false alerts from bad data, sensitive fields slipping through unmasked, and engineers in a constant firefight to patch leaks. These issues demand excessive manual effort and leave little time for innovation. Clearly, a new approach is needed – one that doesn’t just control data access, but actively manages data quality, compliance, and context at scale.

From Access Control to Autonomous Agents: A New Paradigm

What would it look like if data governance were proactive and intelligent instead of reactive and manual? Enter the world of agentic data governance – where intelligent agents imbued in the data pipeline itself take on the tasks of enforcing policies, correcting errors, and optimizing data flow autonomously. This shift is as radical as it sounds: moving from static rules to living, learning systems that govern data in real time.

Instead of simply access management, the focus shifts to agency – giving the data pipeline itself the ability to act. Traditional automation can execute predefined steps, but it “waits” for something to break or for a human to trigger a script. In contrast, an agentic system learns from patterns, anticipates issues, and makes informed decisions on the fly. It’s the difference between a security guard who follows a checklist and an analyst who can think and adapt. With intelligent agents, data governance becomes an active process: the system doesn’t need to wait for a human to notice a compliance violation or a broken schema – it handles those situations in real time.

Consider a simple example of this autonomy in action. In a legacy pipeline, if a data source adds a new field or changes its format, the downstream process would typically fail silently – dropping the field or halting ingestion – until an engineer debugs it hours later. During that window, you’d have missing or malformed data and maybe missed alerts. Now imagine an intelligent agent in that pipeline: it recognizes the schema change before it breaks anything, maps the new field against known patterns, and automatically updates the parsing logic to accommodate it. No manual intervention, no lost data, no blind spots. That is the leap from automation to true autonomy – predicting and preventing failures rather than merely reacting to them.

This new paradigm doesn’t just prevent errors; it builds trust. When your governance processes can monitor themselves, fix issues, and log every decision along the way, you gain confidence that your data is complete, consistent, and compliant. For security teams, it means the data feeding their alerts and reports is reliable, not full of unseen gaps. For compliance officers, it means controls are enforced continuously, not just at periodic checkpoints. And for data engineers, it means a lot less 3 AM pager calls and tedious patching – the boring stuff is handled by the system. Organizations need more than an AI co-pilot; they need “a complementary data engineer that takes over all the exhausting work,” freeing up humans for strategic tasks. In other words, they need agentic AI working for them.

How Databahn’s Cruz Delivers Agentic Governance

At DataBahn, we’ve turned this vision of autonomous data governance into reality. It’s embodied in Cruz, our agentic AI-powered data engineer that works within DataBahn’s security data fabric. Cruz is not just another monitoring tool or script library – as we often describe it, Cruz is “an autonomous AI data engineer that monitors, detects, adapts, and actively resolves issues with minimal human intervention.” In practice, that means Cruz and the surrounding platform components (from smart edge collectors to our central data fabric) handle the heavy lifting of governance automatically. Instead of static pipelines with bolt-on rules, DataBahn provides a self-healing, policy-aware pipeline that governs itself in real time.

With these agentic capabilities, DataBahn’s platform transforms data governance from a static, after-the-fact function into a dynamic, self-healing workflow. Instead of asking “Who should access this data?” you can start trusting the system to ask “Is this data correct, compliant, and useful – and if not, how do we fix it right now?”. Governance becomes an active verb, not just a set of nouns (policies, roles, classifications) sitting on a shelf. By moving governance into the fabric of data operations, DataBahn ensures your pipelines are not only efficient, but defensible and trustworthy by default.

Embracing Autonomous Data Governance

The shift from access to agency means your governance framework can finally scale with your data and complexity. Instead of a gatekeeper saying “no,” you get a guardian angel for your data: one that tirelessly cleans, repairs, and protects your information assets across the entire journey from collection to storage. For CISOs and compliance leaders, this translates to unprecedented confidence – policies are enforced continuously and audit trails are built into every transaction. For data engineers and analysts, it means freedom from the drudgery of pipeline maintenance and an end to the 3 AM pager calls; they gain an automated colleague who has their back in maintaining data integrity.

The era of autonomous, agentic governance is here, and it’s changing data management forever. Organizations that embrace this model will see their data pipelines become strategic assets rather than liabilities. They’ll spend less time worrying about broken feeds or inadvertent exposure, and more time extracting value and insights from a trusted data foundation. In a world of exploding data volumes and accelerating compliance demands, intelligent agents aren’t a luxury – they’re the new necessity for staying ahead.

If you’re ready to move from static control to proactive intelligence in your data strategy, it’s time to explore what agentic AI can do for you. Contact DataBahn or book a demo to see how Cruz and our security data fabric can transform your governance approach.

Every second, billions of connected devices quietly monitor the pulse of the physical world: measuring pressure in refineries, tracking vibrations on turbine blades, adjusting the temperature of precision manufacturing lines, counting cars at intersections, and watching valves that regulate clean water. This is the telemetry that keeps our world running. It is also increasingly what’s putting the world at risk.

Why is OT telemetry becoming a cybersecurity priority?

In 2021, attackers tried to poison a water plant in Oldsmar, Florida, by changing chemical levels. In 2022, ransomware actors breached Tata Power in India, exfiltrating operational data and disrupting key functions. These weren’t IT breaches – they targeted operational technology (OT): the systems where the digital meets the physical. When compromised, they can halt production, damage equipment, or endanger lives.

Despite this growing risk, the telemetry from these systems – the rich, continuous streams of data describing what’s happening in the real world – aren’t entering enterprise-grade security and analytics tools such as SIEMs.

What makes OT telemetry data so hard to integrate into security tools?

For decades, OT telemetry was designed for control, not correlation. Its data is continuous, dense, and expensive to store – the exact opposite of the discrete, event-based logs that SIEMs and observability tools were built for. This mismatch created an architectural blind spot: the systems that track our physical world can’t speak the same language as the systems that secure our digital one. Today, as plants and utilities connect to the cloud, that divide has become a liability.  

OT Telemetry is Different by Design

Security teams managed discrete events – a log, an edit to a file, an alert. OT telemetry reflects continuous signals – temperature, torque, flow, vibrations, cycles. Traditional security logs are timestamped records of what happened. OT data describes what’s happening, sampled dozens or even thousands of times per minute. This creates three critical mismatches in OT and IT telemetry data:

  • Format: Continuous numeric data doesn’t fit text-based log schemas
  • Purpose: OT telemetry optimizes continuing performance while security telemetry is used to flag anomalies and detect threats
  • Economics: SIEMs and analytics tools charge on the basis of ingestion. Continuous data floods these models, turning visibility into runaway cost.

This is why most enterprises either down-sample OT data or skip it entirely; and why most SIEMs don’t have the capacity to ingest OT data out of the box.

Why does this increase risk?

Without unified telemetry, security teams only see fragments of their operational truth. Silent sources or anomalous readings might seem harmless to OT engineers but might signal malicious interference; but that clue needs to be seen and investigated with SOCs to uncover the truth. Each uncollected and unanalyzed bit of data widens the gap between what has happened, what is happening, and what could happen in the future. In our increasingly connected and networked enterprises, that’s where risk lies.

From isolation to integration: bridging the gap

For decades, OT systems operated in isolated environments – air-gapped networks, proprietary closed-loop control systems, and field devices that only speak to their own kind. However, as enterprises sought real-time visibility and data-driven optimization, operational systems started getting linked to enterprise networks and cloud platforms. Plants started streaming production metrics to dashboards; energy firms connected sensors to predictive maintenance systems, and industrial vendors began managing equipment remotely.  

The result: enormous gains in efficiency – and a sudden explosion of exposure.

Attackers can now reach into building control systems inside manufacturing facilities, power plants, and supply chain networks to reach what was once unreachable. Suddenly, a misconfigured VPN or a vulnerability in middleware systems that connect OT to IT systems (current consensus suggests this is what exposed the JLR systems in the recent hack) could become an attacker’s entry point into core operations.

Why is telemetry still a cost center and not a value stream?

For many CISOs, CIOs, and CTOs, OT telemetry remains a budget line item – something to collect sparingly because of the cost of ingesting and storing it, especially in their favorite security tools and systems built over years of operations. But this misses the larger shift underway.

This data is no longer about just monitoring machines – it’s about protecting business continuity and understanding operational risk. The same telemetry that can predict a failing compressor can also help security teams catch and track a cyber intrusion.  

Organizations that treat this data and its security management purely as a compliance expense will always be reactive; those that see this as a strategic dataset – feeding security, reliability, and AI-driven optimization – will turn it into a competitive advantage.

AI as a catalyst: turning telemetry into value

AI has always been most effective when it’s fed by diverse, high-quality data. This is the mindset with which the modern security team treated data, but ingestion-based pricing made them allergic to collecting OT telemetry at scale. But this same mindset is now reaching operational systems, and leading organizations around the world are treating IoT and OT telemetry as strategic data sources for AI-driven security, optimization, and resilience.

AI thrives on context, and no data source offers more context than telemetry that connects the digital and physical worlds. Patterns in OT data can reveal early indications of faltering equipment, sub-optimal logistical choices, and resource allocation signals that can help the enterprise save. It can also provide early indication of attack and defray significant business continuity and operational safety risk.

But for most enterprises, this value is still locked behind scale, complexity, and gaps in their existing systems and tools. Collecting, normalizing, and routing billions of telemetry signals from globally distributed sites is challenging to build manually. Existing tools to solve these problems (SIEM collectors, log forwarders) aren’t built for these data types and still require extensive effort to repurpose.  

This is where Agentic AI can become transformative. Rather than analyzing data downstream after extensive tooling to manage data, AI can be harnessed to manage and govern telemetry from the point of ingestion.

  • Automatically detect new data formats or schema drifts, and generate parsers in minutes on the fly
  • Recognize patterns of redundancy and noise and recommend filtering or forking of data by security relevance to store everything while analyzing only that data which matters
  • Enforce data governance policies in real time – routing sensitive telemetry to compliant destinations
  • Learn from historical behavior to predict which signals are security-relevant versus purely operational

The result is a system that scales not by collecting less, but by collecting everything and routing intelligently. AI is not just the reason to collect more telemetry – it is also the means to make that data valuable and sustainable at scale.

Case Study: Turning 80 sites of OT chaos into connected intelligence

A global energy producer operating more than 80 distributed industrial sites faced the same challenge shared by many manufacturers: limited bandwidth, siloed OT networks, and inconsistent data formats. Each site generates between a few gigabytes to hundreds of gigabytes of log data daily – a mix of access control logs, process telemetry, and infrastructure events. Only a fraction of this data reached their security operations center. The rest stayed on-premise, trapped in local systems that couldn’t easily integrate with their SIEM or data lake. This created blind spots and with recent compliance developments in their region, they needed to integrate this into their security architecture.

The organization decided to re-architect their telemetry layer around a modular, pipeline-first approach. After an evaluation process, they chose Databahn as their vendor to accomplish this. They deployed Databahn’s collectors at the edge, capable of compressing and filtering data locally before securely transmitting it to centralized storage and security tools.

With bandwidth and network availability varying dramatically across sites, edge intelligence became critical. The collectors automatically prioritized security-relevant data for streaming, compressing non-relevant telemetry for slower transmission to conserve network capacity when needed. When a new physical security system needed to be onboarded – one with no existing connectors – an AI-assisted parser system was built in a few days, not months. This agility helped the team reduce their backlog of pending log sources and immediately increase their visibility across their OT environments.

In parallel, they used policy-driven routing to send filtered telemetry not only to their security tools, but also to the organization’s data lake – enabling business and engineering teams to analyze the same data for operational insights.

The outcome?

  • Improved visibility across all their sites in a few weeks
  • Data volume to their SIEM dropped to 60% despite increased coverage, due to intelligent reduction and compression
  • New source of centralized and continuous intelligence established for multiple functional teams to analyze and understand

This is the power of treating telemetry as a strategic asset: and of using the pipeline as the control plane to ensure that the increased coverage and visibility don’t come at the cost of security posture or by destroying the IT/Security budget.

Continuous data, continuous resilience, continuous value

The convergence of IT and OT has and will continue to represent an increase in the attack surface and the vulnerability of digital systems deeply connected to physical reality. For factories and manufacturers like Jaguar Land Rover, this is about protecting their systems from ransomware actors. For power manufacturers and utilities distributors, it could mean the difference between life and death for their business, employees, and citizens with major national security implications.  

To meet this increased risk threshold, telemetry must become the connective tissue of resilience. It must be more closely watched, more deeply understood, and more intelligently managed. Its value must be gauged as early as possible, and its volume must be routed intelligently to sanctify detection and analytics equipment while retaining the underlying data for bulk analysis.

The next decade of enterprise security and AI will depend upon how effectively organizations bridge this divide from the present into the ideal future. The systems that today are being kept out of SIEMs to stop them from flooding will need to fuel your AI. The telemetry from isolated networks will have to be connected to power real-time visibility across your enterprise.

The world will run on this data – and so should the security of your organization.

Hi 👋 Let’s schedule your demo

Thank you! Your submission has been received!
Oops! Something went wrong while submitting the form.

Trusted by leading brands and partners

optiv
mobia
la esfera
inspira
evanssion
KPMG
Guidepoint Security
EY
ESI