All Resources
In this article:
minus iconplus icon
Share the Blog

Your AI Is Only as Secure as Your Data: Celebrating a $100M Milestone

April 23, 2025
3
Min Read
Data Security

Over the past year, we’ve seen an incredible surge in enterprise AI adoption. Companies across industries are integrating AI agents and generative AI into their operations to move faster, work smarter, and unlock innovation. But behind every AI breakthrough lies a foundational truth: AI is only as secure as the data behind it.

At Sentra, securing that data has always been our mission, not just to prevent breaches and data leaks, but to empower prosperity and innovation with confidence and control.

Data Security: The Heartbeat of Your Organization

As organizations push forward with AI, massive volumes of data, often sensitive, regulated, or business-critical are being used to train models or power AI agents. Too often, this happens without full visibility or governance. 


The explosion of the data security market reflects how critical this challenge has become. At Sentra, we’ve long believed that a Data Security Platform (DSP) must be cloud-native, scalable, and adaptable to real-world enterprise environments. We’ve been proud to lead the way, and our continued growth, especially among Fortune 500 customers, is a testament to the urgency and relevance of our approach.

Scaling for What's Next

With the announcement of our $50 million Series B funding round, bringing our total funding to over $100 million, we’re scaling Sentra to meet the moment. We're counting on strong customer momentum and more than tripling revenue year-over-year, and we’re using this investment to grow our team, strengthen our platform, and continue defining what modern data security looks like.

We’ve always said security shouldn’t slow innovation - it should fuel it. And that’s exactly what we’re enabling.

It's All About the People


At the end of the day, it’s people who build it, scale it, and believe in it. I want to extend a heartfelt thank you to our investors, customers, and, most importantly, our team. It’s all about you! Your belief in Sentra and your relentless execution make everything possible. We couldn’t make it without each and every one of you.

We’re not just building a product, we’re setting the gold standard for data security, because securing your data is the heartbeat of your organization!

Innovation without security isn’t progress. Let’s shape a future where both go together!

<blogcta-big>

Yoav Regev has over two decades of experience in the world of cybersecurity, cloud, big data, and machine learning. He was the Head of Cyber Department (Colonel) in the Israeli Military Intelligence (Unit 8200) for nearly 25 years. Reflecting on this experience, it was clear to him that sensitive data had become the most important asset in the world. In the private sector, enterprises that were leveraging data to generate new insights, develop new products, and provide better experiences, were separating themselves from the competition. As data becomes more valuable, it becomes a bigger target, and as the amount of sensitive data grows, so does the importance of finding the most effective way to secure it. That’s why he co-founded Sentra, together with accomplished co-founders, Asaf Kochan, Ron Reiter, and Yair Cohen.

Subscribe

Latest Blog Posts

Ariel Rimon
Ariel Rimon
Daniel Suissa
Daniel Suissa
February 16, 2026
4
Min Read

How Modern Data Security Discovers Sensitive Data at Cloud Scale

How Modern Data Security Discovers Sensitive Data at Cloud Scale

Modern cloud environments contain vast amounts of data stored in object storage services such as Amazon S3, Google Cloud Storage, and Azure Blob Storage. In large organizations, a single data store can contain billions (or even tens of billions) of objects. In this reality, traditional approaches that rely on scanning every file to detect sensitive data quickly become impractical.

Full object-level inspection is expensive, slow, and difficult to sustain over time. It increases cloud costs, extends onboarding timelines, and often fails to keep pace with continuously changing data. As a result, modern data security platforms must adopt more intelligent techniques to build accurate data inventories and sensitivity models without scanning every object.

Why Object-Level Scanning Fails at Scale

Object storage systems expose data as individual objects, but treating each object as an independent unit of analysis does not reflect how data is actually created, stored, or used.

In large environments, scanning every object introduces several challenges:

  • Cost amplification from repeated content inspection at massive scale
  • Long time to actionable insights during the first scan
  • Operational bottlenecks that prevent continuous scanning
  • Diminishing returns, as many objects contain redundant or structurally identical data

The goal of data discovery is not exhaustive inspection, but rather accurate understanding of where sensitive data exists and how it is organized.

The Dataset as the Correct Unit of Analysis

Although cloud storage presents data as individual objects, most data is logically organized into datasets. These datasets often follow consistent structural patterns such as:

  • Time-based partitions
  • Application or service-specific logs
  • Data lake tables and exports
  • Periodic reports or snapshots

For example, the following objects are separate files but collectively represent a single dataset:

logs/2026/01/01/app_events_001.json

logs/2026/01/02/app_events_002.json

logs/2026/01/03/app_events_003.json

While these objects differ by date, their structure, schema, and sensitivity characteristics are typically consistent. Treating them as a single dataset enables more accurate and scalable analysis.

Analyzing Storage Structure Without Reading Every File

Modern data discovery platforms begin by analyzing storage metadata and object structure, rather than file contents.

This includes examining:

  • Object paths and prefixes
  • Naming conventions and partition keys
  • Repeating directory patterns
  • Object counts and distribution

By identifying recurring patterns and natural boundaries in storage layouts, platforms can infer how objects relate to one another and where dataset boundaries exist. This analysis does not require reading object contents and can be performed efficiently at cloud scale.

Configurable by Design

Sampling can be disabled for specific data sources, and the dataset grouping algorithm can be adjusted by the user. This allows teams to tailor the discovery process to their environment and needs.


Automatic Grouping into Dataset-Level Assets

Using structural analysis, objects are automatically grouped into dataset-level assets. Clustering algorithms identify related objects based on path similarity, partitioning schemes, and organizational patterns. This process requires no manual configuration and adapts as new objects are added. Once grouped, these datasets become the primary unit for further analysis, replacing object-by-object inspection with a more meaningful abstraction.

Representative Sampling for Sensitivity Inference

After grouping, sensitivity analysis is performed using representative sampling. Instead of inspecting every object, the platform selects a small, statistically meaningful subset of files from each dataset.

Sampling strategies account for factors such as:

  • Partition structure
  • File size and format
  • Schema variation within the dataset

By analyzing these samples, the platform can accurately infer the presence of sensitive data across the entire dataset. This approach preserves accuracy while dramatically reducing the amount of data that must be scanned.

Handling Non-Standard Storage Layouts

In some environments, storage layouts may follow unconventional or highly customized naming schemes that automated grouping cannot fully interpret. In these cases, manual grouping provides additional precision. Security analysts can define logical dataset boundaries, often supported by LLM-assisted analysis to better understand complex or ambiguous structures. Once defined, the same sampling and inference mechanisms are applied, ensuring consistent sensitivity assessment even in edge cases.

Scalability, Cost, and Operational Impact

By combining structural analysis, grouping, and representative sampling, this approach enables:

  • Scalable data discovery across millions or billions of objects
  • Predictable and significantly reduced cloud scanning costs
  • Faster onboarding and continuous visibility as data changes
  • High confidence sensitivity models without exhaustive inspection

This model aligns with the realities of modern cloud environments, where data volume and velocity continue to increase.

From Discovery to Classification and Continuous Risk Management

Dataset-level asset discovery forms the foundation for scalable classification, access governance, and risk detection. Once assets are defined at the dataset level, classification becomes more accurate and easier to maintain over time. This enables downstream use cases such as identifying over-permissioned access, detecting risky data exposure, and managing AI-driven data access patterns.

Applying These Principles in Practice

Platforms like Sentra apply these principles to help organizations discover, classify, and govern sensitive data at cloud scale - without relying on full object-level scans. By focusing on dataset-level discovery and intelligent sampling, Sentra enables continuous visibility into sensitive data while keeping costs and operational overhead under control.

<blogcta-big>

Read More
Elie Perelman
Elie Perelman
February 12, 2026
3
Min Read

Best Data Access Governance Tools

Best Data Access Governance Tools

Managing access to sensitive information is becoming one of the most critical challenges for organizations in 2026. As data sprawls across cloud platforms, SaaS applications, and on-premises systems, enterprises face compliance violations, security breaches, and operational inefficiencies. Data Access Governance Tools provide automated discovery, classification, and access control capabilities that ensure only authorized users interact with sensitive data. This article examines the leading platforms, essential features, and implementation strategies for effective data access governance.

Best Data Access Governance Tools

The market offers several categories of solutions, each addressing different aspects of data access governance. Enterprise platforms like Collibra, Informatica Cloud Data Governance, and Atlan deliver comprehensive metadata management, automated workflows, and detailed data lineage tracking across complex data estates.

Specialized Data Access Governance (DAG) platforms focus on permissions and entitlements. Varonis, Immuta, and Securiti provide continuous permission mapping, risk analytics, and automated access reviews. Varonis identifies toxic combinations by discovering and classifying sensitive data, then correlating classifications with access controls to flag scenarios where high-sensitivity files have overly broad permissions.

User Reviews and Feedback

Varonis

  • Detailed file access analysis and real-time protection capabilities
  • Excellent at identifying toxic permission combinations
  • Learning curve during initial implementation

BigID

  • AI-powered classification with over 95% accuracy
  • Handles both structured and unstructured data effectively
  • Strong privacy automation features
  • Technical support response times could be improved

OneTrust

  • User-friendly interface and comprehensive privacy management
  • Deep integration into compliance frameworks
  • Robust feature set requires organizational support to fully leverage

Sentra

  • Effective data discovery and automation capabilities (January 2026 reviews)
  • Significantly enhances security posture and streamlines audit processes
  • Reduces cloud storage costs by approximately 20%

Critical Capabilities for Modern Data Access Governance

Effective platforms must deliver several core capabilities to address today's challenges:

Unified Visibility

Tools need comprehensive visibility across IaaS, PaaS, SaaS, and on-premises environments without moving data from its original location. This "in-environment" architecture ensures data never leaves organizational control while enabling complete governance.

Dynamic Data Movement Tracking

Advanced platforms monitor when sensitive assets flow between regions, migrate from production to development, or enter AI pipelines. This goes beyond static location mapping to provide real-time visibility into data transformations and transfers.

Automated Classification

Modern tools leverage AI and machine learning to identify sensitive data with high accuracy, then apply appropriate tags that drive downstream policy enforcement. Deep integration with native cloud security tools, particularly Microsoft Purview, enables seamless policy enforcement.

Toxic Combination Detection

Platforms must correlate data sensitivity with access permissions to identify scenarios where highly sensitive information has broad or misconfigured controls. Once detected, systems should provide remediation guidance or trigger automated actions.

Infrastructure and Integration Considerations

Deployment architecture significantly impacts governance effectiveness. Agentless solutions connecting via cloud provider APIs offer zero impact on production latency and simplified deployment. Some platforms use hybrid approaches combining agentless scanning with lightweight collectors when additional visibility is required.

Integration Area Key Considerations Example Capabilities
Microsoft Ecosystem Native integration with Microsoft Purview, Microsoft 365, and Azure Varonis monitors Copilot AI prompts and enforces consistent policies
Data Platforms Direct remediation within platforms such as Snowflake BigID automatically enforces dynamic data masking and tagging
Cloud Providers API-based scanning without performance overhead Sentra’s agentless architecture scans environments without deploying agents

Open Source Data Governance Tools

Organizations seeking cost-effective or customizable solutions can leverage open source tools. Apache Atlas, originally designed for Hadoop environments, provides mature governance capabilities that, when integrated with Apache Ranger, support tag-based policy management for flexible access control.

DataHub, developed at LinkedIn, features AI-powered metadata ingestion and role-based access control. OpenMetadata offers a unified metadata platform consolidating information across data sources with data lineage tracking and customized workflows.

While open source tools provide foundational capabilities, metadata cataloging, data lineage tracking, and basic access controls, achieving enterprise-grade governance typically requires additional customization, integration work, and infrastructure investment. The software is free, but self-hosting means accounting for operational costs and expertise needed to maintain these platforms.

Understanding the Gartner Magic Quadrant for Data Governance Tools

Gartner's Magic Quadrant assesses vendors on ability to execute and completeness of vision. For data access governance, Gartner examines how effectively platforms define, automate, and enforce policies controlling user access to data.

<blogcta-big>

Read More
Yair Cohen
Yair Cohen
February 11, 2026
4
Min Read

DSPM vs DLP vs DDR: How to Architect a Data‑First Stack That Actually Stops Exfiltration

DSPM vs DLP vs DDR: How to Architect a Data‑First Stack That Actually Stops Exfiltration

Many security stacks look impressive at first glance. There is a DLP agent on every endpoint, a CASB or SSE proxy watching SaaS traffic, EDR and SIEM for hosts and logs, and perhaps a handful of identity and access governance tools. Yet when a serious incident is investigated, it often turns out that sensitive data moved through a path nobody was really watching, or that multiple tools saw fragments of the story but never connected them.

The common thread is that most stacks were built around infrastructure, not data. They understand networks, workloads, and log lines, but they don’t share a single, consistent understanding of:

  • What your sensitive data is
  • Where it actually lives
  • Who and what can access it
  • How it moves across cloud, SaaS, and AI systems

To move beyond that, security leaders are converging on a data‑first architecture that brings together four capabilities: DSPM (Data Security Posture Management), DLP (Data Loss Prevention), DAG (Data Access Governance), and DDR (Data Detection & Response) in a unified model.

Clarifying the Roles

At the heart of this architecture is DSPM. DSPM is your data‑at‑rest intelligence layer. It continuously discovers data across clouds, SaaS, on‑prem, and AI pipelines, classifies it, and maps its posture; configurations, locations, access paths, and regulatory obligations. Instead of a static inventory, you get a living view of where sensitive data resides and how risky it is.

DLP sits at the edges of the system. Its job is to enforce policy on data in motion and in use: emails leaving the organization, files uploaded to the web, documents synced to endpoints, content copied into SaaS apps, or responses generated by AI tools. DLP decides whether to block, encrypt, quarantine, or simply log based on policies and the context it receives.

DAG bridges the gap between “what” and “who.” It’s responsible for least‑privilege access; understanding which human and machine identities can access which datasets, whether they really need that access, and what toxic combinations exist when sensitive data is exposed to broad groups or powerful service accounts.

DDR closes the loop. It monitors access to and movement of sensitive data in real time, looking for unusual or risky behavior: anomalous downloads, mass exports, unusual cross‑region copies, suspicious AI usage. When something looks wrong, DDR triggers detections, enriches them with data context, and kicks off remediation workflows.

When these four functions work together, you get a stack that doesn’t just warn you about potential issues; it actively reduces your exposure and stops exfiltration in motion.

Why “DSPM vs DLP” Is the Wrong Framing

It’s tempting to think of DSPM and DLP as competing answers to the same problem. In reality, they address different parts of the lifecycle. DSPM shows you what’s at risk and where; DLP controls how that risk can materialize as data moves.

Trying to use DLP as a discovery and classification engine is what leads to the noise and blind spots described in the previous section. Conversely, running DSPM without any enforcement at the edges leaves you with excellent visibility but too little control over where data can go.

DSPM and DAG reduce your attack surface; DLP and DDR reduce your blast radius. DSPM and DAG shrink the pool of exposed data and over‑privileged identities. DLP and DDR watch the edges and intervene when data starts to move in risky ways.

A Unified, Data‑First Reference Architecture

In a data‑first architecture, DSPM sits at the center, connected API‑first into cloud accounts, SaaS platforms, data warehouses, on‑prem file systems, and AI infrastructure. It continuously updates an inventory of data assets, understands which are sensitive or regulated, and applies labels and context that other tools can use.

On top of that, DAG analyzes which users, groups, service principals, and AI agents can access each dataset. Over‑privileged access is identified and remediated, sometimes automatically: by tightening IAM roles, restricting sharing, or revoking legacy permissions. The result is a significant reduction in the number of places where a single identity can cause significant damage.

DLP then reads the labels and access context from DSPM and DAG instead of inferring everything from scratch. Email and endpoint DLP, cloud DLP via SSE/CASB, and even platform‑native solutions like Purview DLP all begin enforcing on the same sensitivity definitions and labels. Policies become more straightforward: “Block Highly Confidential outside the tenant,” “Encrypt PHI sent to external partners,” “Require justification for Customer‑Identifiable data leaving a certain region.”

DDR runs alongside this, monitoring how labeled data actually moves. It can see when a typically quiet user suddenly downloads thousands of PHI records, when a service account starts copying IP into a new data store, or when an AI tool begins interacting with a dataset marked off‑limits. Because DDR is fed by DSPM’s inventory and DAG’s access graph, detections are both higher fidelity and easier to interpret.

From there, integration points into SIEM, SOAR, IAM/CIEM, ITSM, and AI gateways allow you to orchestrate end‑to‑end responses: open tickets, notify owners, roll back risky changes, block certain actions, or update policies.

Where Sentra Fits

Sentra’s product vision aligns directly with this data‑first model. Rather than treating DSPM, DAG, DDR, and DLP intelligence as separate products, Sentra brings them together into a single, cloud‑native data security platform.

That means you get:

  • DSPM that discovers and classifies data across cloud, SaaS, on‑prem, and AI
  • DAG that maps and rationalizes access to that data
  • DDR that monitors sensitive data in motion and detects threats
  • Integrations that feed this intelligence into DLP, SSE/CASB, Purview, EDR, and other controls

In other words, Sentra is positioned as the brain of the data‑first stack, giving DLP and the rest of your security stack the insight they need to actually stop exfiltration, not just report on it afterward.

<blogcta-big>

Read More
Expert Data Security Insights Straight to Your Inbox
What Should I Do Now:
1

Get the latest GigaOm DSPM Radar report - see why Sentra was named a Leader and Fast Mover in data security. Download now and stay ahead on securing sensitive data.

2

Sign up for a demo and learn how Sentra’s data security platform can uncover hidden risks, simplify compliance, and safeguard your sensitive data.

3

Follow us on LinkedIn, X (Twitter), and YouTube for actionable expert insights on how to strengthen your data security, build a successful DSPM program, and more!

Before you go...

Get the Gartner Customers' Choice for DSPM Report

Read why 98% of users recommend Sentra.

White Gartner Peer Insights Customers' Choice 2025 badge with laurel leaves inside a speech bubble.