Scaling Security Operations using Data Orchestration

Learn how decoupling data ingestion and collection from your SIEM can unlock exceptional scalability and value for your security and IT teams

February 28, 2024

Scaling Security Operations using Data Orchestration

Lately, there has been a surge in discussions through numerous articles and blogs emphasizing the importance of disentangling the processes of data collection and ingestion from the conventional SIEM (Security Information and Event Management) systems. Leading detection engineering teams within the industry are already adapting to this transformation. They are moving away from the conventional approach of considering security data ingestion, analytics (detection), and storage as a single, monolithic task.

Instead, they have opted to separate the facets of data collection and ingestion from the SIEM, granting them the freedom to expand their detection and threat-hunting capabilities within the platforms of their choice. This approach not only enhances flexibility to bring the best-of-breed technologies but also proves to be cost-effective, as it empowers them to bring in the most pertinent data for their security operations.

Staying ahead of threats requires innovative solutions. One such advancement is the emergence of next-generation data-focused orchestration platforms.

So, what is Security Data Orchestration?

Security data orchestration is a process or technology that involves the collection, normalization, and organization of data related to cybersecurity and information security. It aims to streamline the handling of security data from various sources, making it more accessible in destinations where the data is actionable for security professionals.

 

Why is Security Data Orchestration becoming a big deal now?

Not too long ago, security teams adhered to a philosophy of sending every bit of data everywhere. During that era, the allure of extensive on-premise infrastructure was irresistible, and organizations justified the sustained costs over time. However, in the subsequent years, a paradigm shift occurred as the entire industry began to shift its gaze towards the cloud.

This transformative shift meant that all the entities downstream from data sources—such as SIEM (Security Information and Event Management) systems, UEBA (User and Entity Behavior Analytics), and Data Warehouses—all made their migration to the cloud. This marked the inception of a new era defined by subscription and licensing models that held data as a paramount factor in their quest to maximize profit margins.

In the contemporary landscape, most downstream products, without exception, revolve around the notion of data as a pivotal element. It's all about the data you ingest, the data you process, the data you store, and, not to be overlooked, the data you search in your quest for security and insights.

This paradigm shift has left many security teams grappling to extract the full value they deserve from these downstream systems. They frequently find themselves constrained by the limitations of their SIEMs, struggling to accommodate additional valuable data. Moreover, they often face challenges related to storage capacity and data retention, hindering their ability to run complex hunting scenarios or retrospectively delve deeper into their data for enhanced visibility and insights.

It's quite amusing, but also concerning, to note the significant volume of redundant data that accumulates when companies simply opt for vendor default audit configurations. Take a moment to examine your data for outbound traffic to Office 365 applications, corporate intranets, or routine process executions like Teams.exe or Zoom.exe.


Sample data redundancy illustration with logs collected by these product types in your SIEM Upon inspection, you'll likely discover that within your SIEM, at least three distinct sources are capturing identical information within their respective logs. This level of data redundancy often flies under the radar, and it's a noteworthy issue that warrants attention. And quite simply, this hinders the value that your teams expect to see from the investments made in your SIEM and data warehouse.

Conversely, many security teams amass extensive datasets, but only a fraction of this data finds utility in the realms of threat detection, hunting, and investigations. Here's a snapshot of Active Directory (AD) events, categorized by their event IDs and the daily volume within SIEMs across four distinct organizations.

It is evident that, despite AD audit logs being a staple in SIEM implementations, no two organizations exhibit identical log profiles or event volume trends.

 

Adhering solely to vendor default audit configurations often leads to several noteworthy issues:

  1. Overwhelming Log Collection: In certain cases, such as Org 3, organizations end up amassing an astronomical number of logs from event IDs like EID 4658 or 4690, despite their detection teams rarely leveraging these logs for meaningful analysis.
  2. Redundant Event Collection: Org 4, for example, inadvertently collects redundant events, such as EID 5156, which are also gathered by their firewalls and endpoint systems. This redundancy complicates data management and adds little value.
  3. Blind spots: Standard vendor configurations may result in the omission of critical events, thereby creating security blind spots. These unmonitored areas leave organizations vulnerable to potential threats

On the other hand, it's vital to recognize that in today's multifaceted landscape, no single platform can serve as the definitive, all-encompassing detection system. Although there are numerous purpose-built detection systems painstakingly crafted for specific log types, customers often find themselves grappling with the harsh reality that they can't readily incorporate a multitude of best-of-breed platforms.

The formidable challenges emerge from the intricate intricacies of data acquisition, system management, and the prevalent issue of the ingestion layer being tightly coupled with their SIEMs. Frequently, data cascades into various systems from the SIEM, further compounding the complexity of the situation. The overwhelming burden, both in terms of cost and operational intricacies, can make the pursuit of best-of-breed solutions an impractical endeavor for many organizations.

Today’s SOC teams do not have the strength or capacity to look at each source that is logging to weed out these redundancies or address blind spots or take only the right and relevant data to expensive downstream systems like the SIEM or analytics platforms or even manage multiple data pipelines for multiple platforms.

This underscores the growing necessity for Security Data Orchestration, with an even more vital emphasis on Context-Aware Security Data Orchestration. The rationale is clear: we want the Security Engineering team to focus on security, not get bogged down in data operations.

So, how do you go about Security Data Orchestration?

In its simplest form, envision this layer as a sandwich, positioned neatly between your data sources and their respective destinations.

 

The foundational principles of a Security Data Orchestration platform are -

Centralize your log collection:-  Gather all your security-related logs and data from various sources through a centralized collection layer. This consolidation simplifies data management and analysis, making it easier for downstream platforms to consume the data effectively.

Decouple data ingestion:- Separate the processes of data collection and data ingestion from the downstream systems like SIEMs. This decoupling provides flexibility and scalability, allowing you to fine-tune data ingestion without disrupting your entire security infrastructure.

Filter to send only what is relevant to your downstream system:- Implement intelligent data orchestration to filter and direct only the most pertinent and actionable data to your downstream systems. This not only streamlines cost management but also optimizes the performance of your downstream systems with remarkable efficiency.

Enter DataBahn

At databahn.ai, our mission is clear: to forge the path toward the next-generation Data Orchestration platform. We're dedicated to empowering our customers to seize control of their data but without the burden of relying on communities or embarking on the arduous journey of constructing complex Kafka clusters and writing intricate code to track data changes.

We are purpose-built for Security, our platform captures telemetry once, improves its quality and usability, and then distributes it to multiple destinations - streamlining cybersecurity operations and data analytics.

DataBahn seamlessly ingests data from multiple feeds, aggregates compresses, reduces, and intelligently routes it. With advanced capabilities, it standardizes, enriches, correlates, and normalizes the data before transferring a comprehensive time-series dataset to your data lake, SIEM, UEBA, AI/ML, or any downstream platform.


DataBahn offers continuous ML and AI-powered insights and recommendations on the data collected to unlock maximum visibility and ROI. Our platform natively comes with

  • Out-of-the-box connectors and integrations:- DataBahn offers effortless integration and plug-and-play connectivity with a wide array of products and devices, allowing SOCs to swiftly adapt to new data sources.
  • Threat Research Enabled Filtering Rules:- Pre-configured filtering rules, underpinned by comprehensive threat research, guarantee a minimum volume reduction of 35%, enhancing data relevance for analysis.
  • Enrichment support against Multiple Contexts:- DataBahn enriches data against various contexts including Threat Intelligence, User, Asset, and Geo-location, providing a contextualized view of the data for precise threat identification.
  • Format Conversion and Schema Monitoring:- The platform supports seamless conversion into popular data formats like CIM, OCSF, CEF, and others, facilitating faster downstream onboarding. It intelligently monitors log schema changes for proactive adaptability.
  • Schema Drift Detection:- Detect changes to log schema intelligently for proactive adaptability.
  • Sensitive data detection:- Identify, isolate, and mask sensitive data ensuring data security and compliance.
  • Continuous Support for New Event Types:- DataBahn provides continuous support for new and unparsed event types, ensuring consistent data processing and adaptability to evolving data sources.

Data orchestration revolutionizes the traditional cybersecurity data architecture by efficiently collecting, normalizing, and enriching data from diverse sources, ensuring that only relevant and purposeful data reaches detection and hunting platforms. Data Orchestration is the next big evolution in cybersecurity, that gives Security teams both control and flexibility simultaneously, with agility and cost-efficiency.

Uncover hidden visitor insights to improve their website journey
Share

See related articles

In their article about how banks can extract value from a new generation of AI technology, notable strategy and management consulting firm McKinsey identified AI-enabled data pipelines as an essential part of the ‘Core Technology and Data Layer’. They found this infrastructure to be necessary for AI transformation, as an important intermediary step in the evolution banks and financial institutions will have to make for them to see tangible results from their investments in AI.

The technology stack for the AI-powered banking of the future relies greatly on an increased focus on managing enterprise data better. McKinsey’s Financial Services Practice forecasts that with these tools, banks will have the capacity to harness AI and “… become more intelligent, efficient, and better able to achieve stronger financial performance.

What McKinsey says

The promise of AI in banking

The authors point to increased adoption of AI across industries and organizations, but the depth of the adoption remains low and experimental. They express their vision of an AI-first bank, which -

  1. Reimagines the customer experience through personalization and streamlined, frictionless use across devices, for bank-owned platforms and partner ecosystems
  2. Leverages AI for decision-making, by building the architecture to generate real-time insights and translating them into output which addresses precise customer needs. (They could be talking about Reef)
  3. Modernizes core technology with automation and streamlined architecture to enable continuous, secure data exchange (and now, Cruz)

They recommend that banks and financial service enterprises set a bold vision for AI-powered transformation, and root the transformation in business value.

AI stack powered by multiagent systems

The true potential of AI will require banks of the future to tread beyond just AI models, the authors claim. With embedding AI into four capability layers as the goal, they identify ‘data and core tech’ as one of those four critical components. They have augmented an earlier AI capability stack, specifically adding data preprocessing, vector databases, and data post-processing to create an ‘enterprise data’ part of the ‘core technology and data layer’. They indicate that this layer would build a data-driven foundation for multiple AI agents to deliver customer engagement and enable AI-powered decision-making across various facets of a bank’s functioning.

Our perspective

Data quality is the single greatest predictor of LLM effectiveness today, and our current generation of AI tools are fundamentally wired to convert large volumes of data into patterns, insights, and intelligence. We believe the true value of enterprise AI lies in depth, where Agentic AI modules can speak and interact with each other while automating repetitive tasks and completing specific and niche workstreams and workflows. This is only possible when the AI modules have access to purposeful, meaningful, and contextual data to rely on.

We are already working with multiple banks and financial services institutions to enable data processing (pre and post), and our Cruz and Reef products are deployed in many financial institutions to become the backbone of their transformation into AI-first organizations.

Are you curious to see how you can come closer to building the data infrastructure of the future? Set up a call with our experts to see what’s possible when data is managed with intelligence.

Two years ago, our DataBahn journey began with a simple yet urgent realization: security data management is fundamentally flawed. Enterprises are overwhelmed by security and telemetry, struggling to collect, store, and process it, while finding it harder and harder to gain timely insights from it. As leaders and practitioners in cybersecurity, data engineering, and data infrastructure, we saw this pattern everywhere: spiraling SIEM costs, tool sprawl, noisy data, tech debt, brittle pipelines, and AI initiatives blocked by legacy systems and architectures.

We founded DataBahn to break this cycle. Our platform is specifically designed to help enterprises regain control: disconnecting data pipelines from outdated tools, applying AI to automate data engineering, and constructing systems that empower security, data, and IT teams. We believe data infrastructure should be dynamic, resilient, and scalable, and we are creating systems that leverage these core principles to enhance efficiency, insight, and reliability.

Today, we’re announcing a significant milestone in this journey: a $17M Series A funding round led by Forgepoint Capital, with participation from S3 Ventures and returning investor GTM Capital. Since coming out of stealth, our trajectory has been remarkable – we’ve secured a Fortune 10 customer and have already helped several Fortune 500 and Global 200 companies cut over 50% of their telemetry processing costs and automate most of their data engineering workloads. We're excited by this opportunity to partner with these incredible customers and investors to reimagine how telemetry data is managed.

Tackling an industry problem

As operators, consultants, and builders, we worked with and interacted with CISOs across continents who complained about how they had gone from managing gigabytes of data every month to being drowned by terabytes of data daily, while using the same pipelines as before. Layers and levels of complexity were added by proprietary formats, growing disparity in sources and devices, and an evolving threat landscape. With the advent of Generative AI, CISOs and CIOs found themselves facing an incredible opportunity wrapped in an existential threat, and without the right tools to prepare for it.

DataBahn is setting a new benchmark for how modern enterprises and their CISO/CIOs can manage and operationalize their telemetry across security, observability, and IOT/OT systems and AI ecosystems. Built on a revolutionary AI-driven architecture, DataBahn parses, enriches, and suppresses noise at scale, all while minimizing egress costs. This is the approach our current customers are excited about, because it addresses key pain points they have been unable to solve with other solutions.

Our two new Agentic AI products are solving problems for enterprise data engineering and analytics teams. Cruz automates complex data engineering tasks from log discovery, pipeline creation, tracking and maintaining telemetry health, to providing insights on data quality. Reef surfaces context-aware and enriched insights from streaming telemetry data, turning hours of complex querying across silos into seconds of natural-language queries.

The Right People

We’re incredibly grateful to our early customers; their trust, feedback, and high expectations have shaped who we are. Their belief drives us every day to deliver meaningful outcomes. We’re not just solving problems with them, we’re building long-term partnerships to help enterprise security and IT teams take control of their data, and design systems that are flexible, resilient, and built to last. There’s more to do, and we’re excited to keep building together.

We’re also deeply thankful for the guidance and belief of our advisors, and now our investors. Their support has not only helped us get here but also sharpened our understanding of the opportunity ahead. Ernie, Aaron, and Saqib’s support has made this moment more meaningful than the funding; it’s the shared conviction that the way enterprises manage and use data must fundamentally change. Their backing gives us the momentum tomove faster, and the guidance to keep building towards that mission.

Above all, we want to thank our team. Your passion, resilience, and belief in what we’re building together are what got us here. Every challenge you’ve tackled, every idea you’ve contributed, every late night and early morning has laid the foundation for what we have done so far and for what comes next. We’re excited about this next chapter and are grateful to have been on this journey with all of you.

The Next Chapter

The complexity of enterprise data management is growing exponentially. But we believe that with the right foundation, enterprises can turn that complexity into clarity, efficiency, and competitive advantage.

If you’re facing challenges with your security or observability data, and you’re ready to make your data work smarter for AI, we’d love to show you what DataBahn can do. Request a demo and see how we can help.

Onwards and upwards!

Nanda and Nithya
Cofounders, DataBahn

In September 2022, cybercriminals accessed, encrypted, and stole a substantial amount of data from Suffolk County’s IT systems, which included personally identifiable information (PII) of county residents, employees, and retirees. Although Suffolk County did not pay the ransom demand of $2.5 million, it ultimately spent $25 million to address and remediate the impact of the attack.

Members of the county’s IT team reported receiving hundreds of alerts every day in the weeks leading up to the attack. Several months earlier, frustrated by the excessive number of unnecessary alerts, the team redirected the notifications from their tools to a Slack channel. Although the frequency and severity of the alerts increased leading up to the September breach, the constant stream of alerts wore the small team down, leaving them too exhausted to respond and distinguish false positives from relevant alerts. This situation created an opportunity for malicious actors to successfully circumvent security systems.

The alert fatigue problem

Today, cybersecurity teams are continually bombarded by alerts from security tools throughout the data lifecycle. Firewalls, XDRs/EDRs, and SIEMs are among the common tools that trigger these alerts. In 2020, Forrester reported that SOC teams received 11,000 alerts daily, and 55% of cloud security professionals admitted to missing critical alerts. Organizations cannot afford to ignore a single alert, yet alert fatigue (and an overwhelming number of unnecessary alerts) causes SOCs to miss up to 30% of security alerts that go uninvestigated or are completely overlooked.

While this creates a clear cybersecurity and business continuity problem, it also presents a pressing human issue. Alert fatigue leads to cognitive overload, emotional exhaustion, and disengagement, resulting in stress, mental health concerns, and attrition. More than half of cybersecurity professionals cite their workload as the primary source of stress, two-thirds reported experiencing burnout, and over 60% of cybersecurity professionals surveyed stated it contributed to staff turnover and talent loss.

Alert fatigue poses operational challenges, represents a critical security risk, and truly becomes an adversary of the most vital resource that enterprises rely on for their security — SOC professionals doing their utmost to combat cybercriminals. SOCs are spending so much time and effort triaging alerts and filtering false positives that there’s little room for creative threat hunting.

Data is the problem – and the solution

Alert fatigue is a result, not a root cause. When these security tools were initially developed, cybersecurity teams managed gigabytes of data each month from a limited number of computers on physically connected sites. Today, Security Operations Centers (SOCs) are tasked with handling security data from thousands of sources and devices worldwide, which arrive through numerous distinct devices in various formats. The developers of these devices did not intend to simplify the lives of security teams, and the tools they designed to identify patterns often resemble a fire alarm in a volcano. The more data that is sent as an input to these machines, the more likely they are to malfunction – further exhausting and overwhelming already stretched security teams.

Well-intentioned leaders advocate for improved triaging, the use of automation, refined rules to reduce false-positive rates, and the application of popular technologies like AI and ML. Until we can stop security tools from being overwhelmed by large volumes of unstructured, unrefined, and chaotic data from diverse sources and formats, these fixes will be band aids on a gaping wound.

The best way to address alert fatigue is to filter out the data being ingested into downstream security tools. Consolidate, correlate, parse, and normalize data before it enters your SIEM or UEBA. If it isn’t necessary, store it in blob storage. If it’s duplicated or irrelevant, discard it. Don’t clutter your SIEM with poor data so it doesn’t overwhelm your SOC with alerts no one requested.

How Databahn helps

At DataBahn, we help enterprises cut through cybersecurity noise with our security data pipeline solution, which works around the clock to:

1. Aggregates and normalizes data across tools and environments automatically

2. Uses AI-driven correlation and prioritization

3. Denoises the data going into the SIEM, ensuring more actionable alerts with full context

SOCs using DataBahn aren’t overwhelmed with alerts; they only see what’s relevant, allowing them to respond more quickly and effectively to threats. They are empowered to take a more strategic approach in managing operations, as their time isn’t wasted triaging and filtering out unnecessary alerts.

Organizations looking to safeguard their systems – and protect their SOC members – should shift from raw alert processing to smarter alert management, driven by an intelligent pipeline which combines automation, correlation, and transformation that filters out the noise and combats alert fatigue.

Interested in saving your SOC from alert fatigue? Contact DataBahn
In the past, we've written about how we solve this problem for Sentinel. You can read more here: 
AI-powered Sentinel Log Optimization