All Resources
In this article:
minus iconplus icon
Share the Blog

Minimizing your Data Attack Surface in the Cloud

November 8, 2022
4
Min Read

The cloud is one of the most important developments in the history of information technology. It drives innovation and speed for companies, giving engineers instant access to virtually any type of workload with unlimited scale.

But with opportunity comes a price - moving at these speeds increases the risk that data ends up in places that are not monitored for governance, risk and compliance issues. Of course, this increases the risk of a data breach, but it’s not the only reason we’re seeing so many breaches in the cloud era. Other reasons include: 

  • Systems are being built quickly for business units without adequate regard for security
  • More data is moving through the company as teams use and mine data more efficiently using tools such as cloud data warehouses, BI, and big data analytics
  • New roles are being created constantly for people who need to gain access to organizational data
  • New technologies are being adopted for business growth which require access to vast amounts of data - such as deep learning, novel language models, and new processors in the cloud
  • Anonymous cryptocurrencies have made data leaks lucrative.
  • Nation state powers are increasing cyber attacks due to new conflicts

Ultimately, there are only two methods which can mitigate the risk of cloud data leaks - better protecting your cloud infrastructure, and minimizing your data attack surface.

Protecting Cloud Infrastructure

Companies such as Wiz, Orca Security and Palo Alto provide great cloud security solutions, the most important of which is a Cloud Security Posture Management tool. CSPM tools help security teams to understand and remediate infrastructure related cloud security risks which are mostly related to misconfigurations, lateral movements of attackers, and vulnerable software that needs to be patched.

However, these tools cannot mitigate all attacks. Insider threats, careless handling of data, and malicious attackers will always find ways to get a hold of organizational data, whether it is in the cloud, in different SaaS services, or on employee workstations. Even the most protected infrastructure cannot withstand social engineering attacks or accidental mishandling of sensitive data. The best way to mitigate the risk for sensitive data leaks is by minimizing the “data attack surface” of the cloud.

What is the "Data Attack Surface"?

Data attack surface is a term that describes the potential exposure of an organization’s sensitive data in the event of a data breach. If a traditional attack surface is the sum of all an organization’s vulnerabilities, a data attack surface is the sum of all sensitive data that isn’t secured properly. 

The larger the data attack surface - the more sensitive data you have - the higher the chances are that a data breach will occur.

There are several ways to reduce the chances of a data breach:

  • Reduce access to sensitive data
  • Reduce the number of systems that process sensitive data
  • Reduce the number of outputs that data processing systems write
  • Address misconfigurations of the infrastructure which holds sensitive data
  • Isolate infrastructure which holds sensitive data
  • Tokenize data
  • Encrypt data at rest
  • Encrypt data in transit
  • Use proxies which limit and govern access to sensitive data of engineers

Reduce Your Data Attack Surface by using a Least Privilege Approach

The less people and systems have access to sensitive data, the less chances a misconfiguration or an insider will cause a data breach. 

The most optimal method of reducing access to data is by using the least privilege approach  of only granting access to entities that need the data.  The type of access is also important  - if read-only access is enough, then it’s important to make sure that write access or administrative access is not accidentally granted. 

To know which entities need what access, engineering teams need to be responsible for mapping all systems in the organization and ensuring that no data stores are accessible to entities which do not need access.

Engineers can get started by analyzing the actual use of the data using cloud tools such as Cloudtrail.  Once there’s an understanding of which users and services access infrastructure with sensitive data, the actual permissions to the data stores should be reviewed and matched against usage data. If partial permissions are adequate to keep operations running, then it’s possible to reduce the existing permissions within existing roles. 

Reducing Your Data Attack Surface by Tokenizing Your Sensitive Data

Tokenization is a great tool which can protect your data - however it’s hard to deploy and requires a lot of effort from engineers. 

Tokenization is the act of replacing sensitive data such as email addresses and credit card information with tokens, which correspond to the actual data. These tokens can reside in databases and logs throughout your cloud environment without any concern, since exposing them does not reveal the actual data but only a reference to the data.

When the data actually needs to be used (e.g. when emailing the customer or making a transaction with their credit card) the token can be used to access a vault which holds the sensitive information. This vault is highly secured using throttling limits, strong encryption, very strict access limits, and even hardware-based methods to provide adequate protection.

This method also provides a simple way to purge sensitive customer data, since the tokens that represent the sensitive data are meaningless if the data was purged from the sensitive data vault.

Reducing Your Data Attack Surface by Encrypting Your Sensitive Data

Encryption is an important technique which should almost always be used to protect sensitive data. There are two methods of encryption: using the infrastructure or platform you are using to encrypt and decrypt the data, or encrypting it on your own. In most cases, it’s more convenient to encrypt your data using the platform because it is simply a configuration change. This will allow you to ensure that only the people who need access to data will have access via encryption keys. In Amazon Web Services for example, only principals with access to the KMS vault will be able to decrypt information in an S3 bucket with KMS encryption enabled.

It is also possible to encrypt the data by using a customer-managed key, which has its advantages and disadvantages. The advantage is that it’s harder for a misconfiguration to accidentally allow access to the encryption keys, and that you don’t have to rely on the platform you are using to store them. However, using customer-managed keys means you need to send the keys over more frequently to the systems which encrypt and decrypt it, which increases the chance of the key being exposed.

Reducing Your Data Attack Surface by using Privileged Access Management Solutions

There are many tools that centrally manage access to databases. In general, they are divided into two categories: Zero-Trust Privilege Access Management solutions, and Database Governance proxies. Both provide protection against data leaks in different ways.

Zero-Trust Privilege Access Management solutions replace traditional database connectivity with stronger authentication methods combined with network access. Tools such as StrongDM and Teleport (open-source) allow developers to connect to production databases by using authentication with the corporate identity provider.

Database Governance proxies such as Satori and Immuta control how developers interact with sensitive data in production databases. These proxies control not only who can access sensitive data, but how they access the data. By proxying the requests, sensitive data can be tracked and these proxies guarantee that no sensitive information is being queried by developers. When sensitive data is queried, these proxies can either mask the sensitive information, or simply omit or disallow the requests ensuring that sensitive data doesn’t leave the database.

Reducing the data attack surface reflects the reality of the attackers mindset. They’re not trying to get into your infrastructure to breach the network. They’re doing it to find the sensitive data. By ensuring that sensitive data always is secured, tokenized, encrypted, and  with least privilege access, they’ll be nothing valuable for an attacker to find - even in the event of a breach. 

 

Discover Ron’s expertise, shaped by over 20 years of hands-on tech and leadership experience in cybersecurity, cloud, big data, and machine learning. As a serial entrepreneur and seed investor, Ron has contributed to the success of several startups, including Axonius, Firefly, Guardio, Talon Cyber Security, and Lightricks, after founding a company acquired by Oracle.

Subscribe

Latest Blog Posts

David Stuart
David Stuart
January 28, 2026
3
Min Read

Data Privacy Day: Why Discovery Isn’t Enough

Data Privacy Day: Why Discovery Isn’t Enough

Data Privacy Day is a good reminder for all of us in the tech world: finding sensitive data is only the first step. But in today’s environment, data is constantly moving -across cloud platforms, SaaS applications, and AI workflows. The challenge isn’t just knowing where your sensitive data lives; it’s also understanding who or what can touch it, whether that access is still appropriate, and how it changes as systems evolve.

I’ve seen firsthand that privacy breaks down not because organizations don’t care, but because access decisions are often disconnected from how data is actually being used. You can have the best policies on paper, but if they aren’t continuously enforced, they quickly become irrelevant.

Discovery is Just the Beginning

Most organizations start with data discovery. They run scans, identify sensitive files, and map out where data lives. That’s an important first step, and it’s necessary, but it’s far from sufficient. Data is not static. It moves, it gets copied, it’s accessed by humans and machines alike. Without continuously governing that access, all the discovery work in the world won’t stop privacy incidents from happening.

The next step, and the one that matters most today, is real-time governance. That means understanding and controlling access as it happens. 

Who can touch this data? Why do they have access? Is it still needed? And crucially, how do these permissions evolve as your environment changes?

Take, for example, a contractor who needs temporary access to sensitive customer data. Or an AI workflow that processes internal HR information. If those access rights aren’t continuously reviewed and enforced, a small oversight can quickly become a significant privacy risk.

Privacy in an AI and Automation Era

AI and automation are changing the way we work with data, but they also change the privacy equation. Automated processes can move and use data in ways that are difficult to monitor manually. AI models can generate insights using sensitive information without us even realizing it. This isn’t a hypothetical scenario, it’s happening right now in organizations of all sizes.

That’s why privacy cannot be treated as a once-a-year exercise or a checkbox in an audit report. It has to be embedded into daily operations, into the way data is accessed, used, and monitored. Organizations that get this right build systems that automatically enforce policies and flag unusual access - before it becomes a problem.

Beyond Compliance: Continuous Responsibility

The companies that succeed in protecting sensitive data are those that treat privacy as a continuous responsibility, not a regulatory obligation. They don’t wait for audits or compliance reviews to take action. Instead, they embed privacy into how data is accessed, shared, and used across the organization.

This approach delivers real results. It reduces risk by catching misconfigurations before they escalate. It allows teams to work confidently with data, knowing that sensitive information is protected. And it builds trust - both internally and with customers because people know their data is being handled responsibly.

A New Mindset for Data Privacy Day

So this Data Privacy Day, I challenge organizations to think differently. The question is no longer “Do we know where our sensitive data is?” Instead, ask:

“Are we actively governing who can touch our data, every moment, everywhere it goes?”

In a world where cloud platforms, AI systems, and automated workflows touch nearly every piece of data, privacy isn’t a one-time project. It’s a continuous practice, a mindset, and a responsibility that needs to be enforced in real time.

Organizations that adopt this mindset don’t just meet compliance requirements, they gain a competitive advantage. They earn trust, strengthen security, and maintain a dynamic posture that adapts as systems and access needs evolve.

Because at the end of the day, true privacy isn’t something you achieve once a year. It’s something you maintain every day, in every process, with every decision. This Data Privacy Day, let’s commit to moving beyond discovery and audits, and make continuous data privacy the standard.

<blogcta-big>

Read More
David Stuart
David Stuart
January 27, 2026
4
Min Read

DSPM for Modern Fintech: From Masking to AI-Aware Data Protection

DSPM for Modern Fintech: From Masking to AI-Aware Data Protection

Fintech leaders, from digital-first banks to API-driven investment platforms, face a major data dilemma today. With cloud-native architectures, real-time analytics, and the rapid integration of AI, the scale, speed, and complexity of sensitive data have skyrocketed. Fintech platforms are quickly surpassing what legacy Data Loss Prevention (DLP) and Data Security Posture Management (DSPM) tools can handle.

Why? Fintech companies now need more than surface-level safeguards. They require true depth: AI-driven data classification, dynamic masking, and fluid integrations across a massive tech stack that includes Snowflake, AWS Bedrock, and Microsoft 365. Below, we look at why DSPM in financial services is at a defining moment, what recurring pain points exist with traditional, and even many emerging, tools, and how Sentra is reimagining what the modern data protection stack should deliver.

The Pitfalls of Legacy DLP and Early DSPM in Fintech

Legacy DLP wasn’t built for fintech’s speed or expanding data footprint. These tools focus on rigid rules and tight boundaries, which aren’t equipped to handle petabyte-scale, multi-cloud, or AI-powered environments. Early DSPM tools brought some improvements in visibility, but problems persisted: incomplete data discovery, basic classification, lots of manual steps, and limited support for dynamic masking.

For fintech companies, this creates mounting regulatory risk as compliance pressures rise, and slow, manual processes lead to both security and operational headaches. Teams waste hours juggling alerts and trying to piece together patchwork fixes, often resorting to clunky add-on masking tools. The cost is obvious: a scattered protection strategy, long breach response times, and constant exposure to regulatory issues - especially as environments get more distributed and complex.

Why "Good Enough" DSPM Isn’t Enough Anymore

Change in fintech moves faster than ever. The DSPM for the financial services sector is growing at breakneck speed. But as financial applications get more sophisticated, and with cloud and AI adoption soaring, the old "good enough" DSPM falls short. Sensitive data is everywhere now. 82% percent of breaches happen in the cloud, with 39% stretching across multi-cloud or hybrid setups according to The Future of Data Security: Why DSPM is Here to Stay. Enterprise data is set to exceed 181 zettabytes by 2025, raising the stakes for automation, real-time classification, and tight integration with core infrastructure.

AI and automation are no longer optional. To effectively reduce risk and keep compliance manageable and truly auditable, DSPM systems need to automate classification, masking, remediation, and reporting as a central part of operations, not as last-minute additions.

Where Most DSPM Solutions Fall Short

Fintech organizations often struggle to scale legacy or early DSPM and DLP products, especially those similar to emerging DSPM or large CNAPP vendors. These tools might offer broad control and AI-powered classification, but they usually require too much manual orchestration to achieve full remediation, only automate certain pieces of the workflow, and rely on separate masking add-ons.

That leads to gaps in AI and multi-cloud data context, choppy visibility, and much of the workflow stuck in manual gear, a recipe for persistent exposure of sensitive data, especially in fast-moving fintech environments.

Fintech buyers, especially those scaling quickly, also point to a crucial need: ensuring DSPM tools natively and deeply support platforms like Snowflake, AWS Bedrock, and Macie. They want automated, business-driven policy enforcement without constantly babysitting the system.

Sentra’s Next-Gen DSPM: AI-Native, Masking-Aware, and Stack-Integrated for Fintech

Sentra was created with these modern fintech challenges in mind. It offers real-time, continuous, agentless classification and deep context for cloud, SaaS, and AI-powered environments.

What makes Sentra different?

  • Petabyte-scale agentless discovery: Always-on, friction-free classification, with no heavy infrastructure or manual tweaks.
  • AI-native contextualization: Pinpoints sensitive data at a business level and connects instantly with masking policies across Snowflake, Microsoft Purview, and more inferred masking synergy.
  • Automation-driven compliance: Handles everything from discovery to masking to changing permissions, with clear, auditable reporting automated masking/remediation.
  • Integrated for modern stacks: Ready-made, with out-of-the-box connections for Snowflake, Bedrock, Microsoft 365, and the wider AWS/fintech ecosystem.

More and more fintech companies are switching to Sentra DSPM to achieve true cross-cloud visibility and meet regulations without slowing down. By plugging into fintech data flows and covering AI model pipelines, Sentra lets organizations use DSPM with the same speed as their business.

Building a Future-Ready DSPM Strategy in Financial Services

Managing and protecting sensitive data is a competitive edge for fintech, not just a security concern. With compliance rising up the agenda - 84% of IT and security leaders now list it as a top driver - your DSPM investments need to focus on automation, consistent visibility, and enforceable policies throughout your architecture.

Next-gen DSPM means: less busywork, no more juggling between masking and classification tools, and instant, actionable insight into data risk, wherever your information lives. In other words, you spend less time firefighting, move faster, and can assure partners and customers that their data is in good hands.

See How SoFi

Request a demo and technical assessment to discover how Sentra’s AI-aware DSPM can speed up both your compliance and your innovation.

Conclusion

Legacy data protection simply can’t keep up with the size, complexity, and regulatory demands of financial data today. DSPM is now table stakes - as long as it’s automated, built with AI at its core, and actively reduces risk in real time, not just points it out.

Sentra helps you move forward confidently: always-on, agentless classification, automated fixes and masking, and deep stack integration designed for the most complex fintech systems. As you build the future of financial services, your DSPM should make it easier to stay compliant, agile, and protected - no matter how quickly your technology changes.

<blogcta-big>

Read More
Romi Minin
Romi Minin
Nikki Ralston
Nikki Ralston
January 26, 2026
4
Min Read

How to Choose a Data Access Governance Tool

How to Choose a Data Access Governance Tool

Introduction: Why Data Access Governance Is Harder Than It Should Be

Data access governance should be simple: know where your sensitive data lives, understand who has access to it, and reduce risk without breaking business workflows. In practice, it’s rarely that straightforward. Modern organizations operate across cloud data stores, SaaS applications, AI pipelines, and hybrid environments. Data moves constantly, permissions accumulate over time, and visibility quickly degrades. Many teams turn to data access governance tools expecting clarity, only to find legacy platforms that are difficult to deploy, noisy, or poorly suited for dynamic, fast-proliferating cloud environments.

A modern data access governance tool should provide continuous visibility into who and what can access sensitive data across cloud and SaaS environments, and help teams reduce overexposure safely and incrementally.

What Organizations Actually Need from Data Access Governance

Before evaluating vendors, it’s important to align on outcomes, just not features. Most teams are trying to solve the same core problems:

  • Unified visibility across cloud data stores, SaaS platforms, and hybrid environments
  • Clear answers to “which identities have access to what, and why?”
  • Risk-based prioritization instead of long, unmanageable lists of permissions
  • Safe remediation that tightens access without disrupting workflows

Tools that focus only on periodic access reviews or static policies often fall short in dynamic environments where data and permissions change constantly.

Why Legacy and Over-Engineered Tools Fall Short

Many traditional data governance and IGA tools were designed for on-prem environments and slower change cycles. In cloud and SaaS environments, these tools often struggle with:

  • Long deployment timelines and heavy professional services requirements
  • Excessive alert noise without clear guidance on what to fix first
  • Manual access certifications that don’t scale
  • Limited visibility into modern SaaS and cloud-native data stores

Overly complex platforms can leave teams spending more time managing the tool than reducing actual data risk.

Key Capabilities to Look for in a Modern Data Access Governance Tool

1. Continuous Data Discovery and Classification

A strong foundation starts with knowing where sensitive data lives. Modern tools should continuously discover and classify data across cloud, SaaS, and hybrid environments using automated techniques, not one-time scans.

2. Access Mapping and Exposure Analysis

Understanding data sensitivity alone isn’t enough. Tools should map access across users, roles, applications, and service accounts to show how sensitive data is actually exposed.

3. Risk-Based Prioritization

Not all exposure is equal. Effective platforms correlate data sensitivity with access scope and usage patterns to surface the highest-risk scenarios first, helping teams focus remediation where it matters most.

4. Low-Friction Deployment

Look for platforms that minimize operational overhead:

  • Agentless or lightweight deployment models
  • Fast time-to-value
  • Minimal disruption to existing workflows

5. Actionable Remediation Workflows

Visibility without action creates frustration. The right tool should support guided remediation, tightening access incrementally and safely rather than enforcing broad, disruptive changes.

How Teams Are Solving This Today

Security teams that succeed tend to adopt platforms that combine data discovery, access analysis, and real-time risk detection in a single workflow rather than stitching together multiple legacy tools. For example, platforms like Sentra focus on correlating data sensitivity with who or what can actually access it, making it easier to identify over-permissioned data, toxic access combinations, and risky data flows, without breaking existing workflows or requiring intrusive agents.

The common thread isn’t the tool itself, but the ability to answer one question continuously:

“Who can access our most sensitive data right now, and should they?”

Teams using these approaches often see faster time-to-value and more actionable insights compared to legacy systems.

Common Gotchas to Watch Out For

When evaluating tools, buyers often overlook a few critical issues:

  • Hidden costs for deployment, tuning, or ongoing services
  • Tools that surface risk but don’t help remediate it
  • Point-in-time scans that miss rapidly changing environments
  • Weak integration with identity systems, cloud platforms, and SaaS apps

Asking vendors how they handle these scenarios during a pilot can prevent surprises later.
Download The Dirt on DSPM POVs: What Vendors Don’t Want You to Know

How to Run a Successful Pilot

A focused pilot is the best way to evaluate real-world effectiveness:

  1. Start with one or two high-risk data stores
  2. Measure signal-to-noise, not alert volume
  3. Validate that remediation steps work with real teams and workflows
  4. Assess how quickly the tool delivers actionable insights

The goal is to prove reduced risk, not just improved reporting.

Final Takeaway: Visibility First, Enforcement Second

Effective data access governance starts with visibility. Organizations that succeed focus first on understanding where sensitive data lives and how it’s exposed, then apply controls gradually and intelligently. Combining DAG with DSPM is an effective way to achieve this.

In 2026, the most effective data access governance tools are continuous, risk-driven, and cloud-native, helping security teams reduce exposure without slowing the business down.

Frequently Asked Questions (FAQs)

What is data access governance?

Data access governance is the practice of managing and monitoring who can access sensitive data, ensuring access aligns with business needs and security requirements.

How is data access governance different from IAM?

IAM focuses on identities and permissions. Data access governance connects those permissions to actual data sensitivity and exposure, and alerts when violations occur.

How do organizations reduce over-permissioned access safely?

By using risk-based prioritization and incremental remediation instead of broad access revocations.

What should teams look for in a modern data access governance tool?

This question comes up frequently in real-world evaluations, including Reddit discussions where teams share what’s worked and what hasn’t. Teams should prioritize tools that give fast visibility into who can access sensitive data, provide context-aware insights, and allow incremental, safe remediation - all without breaking workflows or adding heavy operational overhead. Cloud- and SaaS-aware platforms tend to outperform legacy or overly complex solutions.

<blogcta-big>

Read More
Expert Data Security Insights Straight to Your Inbox
What Should I Do Now:
1

Get the latest GigaOm DSPM Radar report - see why Sentra was named a Leader and Fast Mover in data security. Download now and stay ahead on securing sensitive data.

2

Sign up for a demo and learn how Sentra’s data security platform can uncover hidden risks, simplify compliance, and safeguard your sensitive data.

3

Follow us on LinkedIn, X (Twitter), and YouTube for actionable expert insights on how to strengthen your data security, build a successful DSPM program, and more!

Before you go...

Get the Gartner Customers' Choice for DSPM Report

Read why 98% of users recommend Sentra.

White Gartner Peer Insights Customers' Choice 2025 badge with laurel leaves inside a speech bubble.