Skip to content

Privacy by design in a data-hungry app

Affinity Atlas needs rich behavioural data to match well. It also believes your data is yours. Here is how we handle that tension - and why most dating apps do not.

Deep Dive   By 0xBrewEntropy - 30 March 2026 · 12 min read

The state of dating app privacy

Dating apps collect more intimate data than almost any other category of software. Your photos, your location, your sexual orientation, your relationship history, your messaging patterns, your purchase behaviour - and increasingly, your biometric data. This is not an exaggeration. It is the baseline.

In 2024, the Mozilla Foundation reviewed 25 major dating apps for their *Privacy Not Included* report and concluded that dating apps have "gotten even worse for your privacy" since their previous review in 2021. 22 out of 25 apps received a privacy warning. Over half had experienced a data breach, leak, or hack in the preceding three years. And the data being collected went far beyond what any reasonable person would expect.

A Norwegian Consumer Council study found that popular dating apps - including Tinder, OkCupid, and Grindr - were sharing sensitive personal data with dozens of advertising companies. This included GPS coordinates, sexual orientation, drug use information, and political opinions. Users had no meaningful way to prevent this sharing.

The Electronic Frontier Foundation put it bluntly in a 2025 report: "Dating apps need to learn how consent works." The EFF highlighted cases where dating platforms shared user data with AI companies without explicit consent - including Bumble sending personal information to OpenAI to power AI-generated icebreaker messages, a practice that the European privacy nonprofit noyb has filed a GDPR complaint over.

โš ๏ธ The uncomfortable truth: When you create a profile on a mainstream dating app, you are not just sharing data with potential matches. You are feeding a data pipeline that extends to advertisers, analytics firms, AI training sets, and in some cases, data brokers who sell your information to anyone willing to pay - including, in at least one documented case, a Catholic organisation that purchased Grindr location data to monitor clergy members.

The breach hall of shame

Privacy failures in dating apps are not hypothetical risks. They are documented history, with real consequences for real people.

2015
Ashley Madison

32 million user records exposed, including real names, home addresses, and credit card transactions. The breach led to documented cases of blackmail, divorce, job losses, and reported suicides. The site had charged users $19 to "permanently delete" their data - but never actually deleted it.

2016
AdultFriendFinder

Over 400 million accounts leaked - one of the largest data breaches in history. Exposed data included sexual preference information, making it among the most sensitive breaches ever recorded.

2018
Tinder

Security researchers demonstrated that Tinder's lack of encryption made it possible for anyone on the same Wi-Fi network to reconstruct a user's entire app experience in real time - including which profiles they viewed and which direction they swiped.

2019
Coffee Meets Bagel

6 million users' data exposed on Valentine's Day. Names, email addresses, and profile information were stolen as part of a larger supply-chain attack affecting multiple services.

2025
M.A.D Mobile

Over 1.5 million private and explicit images from five dating apps exposed via unprotected cloud storage. No passwords, no encryption - just open buckets. Up to 900,000 users put at risk of blackmail and extortion.

2026
Match Group

A breach at Tinder, Hinge, and OkCupid's parent company exposed user data. The incident underscored that even the largest players in the industry - with the most resources to invest in security - remain vulnerable.

These are not fringe apps run by amateurs. Ashley Madison, Tinder, and Match Group are (or were) industry leaders. The pattern is clear: if your data exists in a centralised database, it is a target. And dating data - with its combination of identity, location, sexual preference, and intimate imagery - is among the most valuable and most dangerous data to lose.

The false trade-off: data vs privacy

The dating app industry presents a false binary: share everything or get bad matches. The more data you give us, the better we can match you. Your privacy is the price of finding love.

This framing is convenient for companies whose business model depends on data collection. But it is not technically accurate.

Affinity Atlas needs more data than a typical dating app - not less. The whole point is to match across multiple dimensions of who you are: your music, your games, your books, your beer, your code, your fitness. That requires pulling data from Spotify, Steam, Untappd, Goodreads, GitHub, Strava, and more.

So how does a data-intensive matching system also claim to be privacy-first? The answer lies in a distinction that most dating apps ignore entirely:

๐Ÿ”‘ The key distinction: There is a fundamental difference between collecting data and storing data. And there is a fundamental difference between storing raw data and storing derived signals. Affinity Atlas is designed around these distinctions at every layer.

The system needs to read your Spotify listening history to calculate compatibility. It does not need to keep your full listening history. It needs to know that you and another user both love an artist with a popularity of 14/100. It does not need to store every track you have ever played.

This is the core principle: process rich data, store minimal signals.

Every platform integration in Affinity Atlas follows the same consent model. No exceptions, no defaults-to-on, no dark patterns.

This is not a novel legal requirement. GDPR already mandates clear consent, data minimisation, and the right to deletion. The difference is that most dating apps treat these as compliance checkboxes - buried in 8,000-word privacy policies that nobody reads. Affinity Atlas treats them as product features that are visible, accessible, and central to the user experience.

Derived signals: the privacy layer

The concept of derived signals is the technical foundation of Affinity Atlas's privacy model. Here is how it works in practice:

What gets pulled (temporarily)

When you connect Spotify, the system reads your top artists, top tracks, recently played tracks, and saved library via Spotify's API. This data is processed in memory to extract compatibility-relevant features.

What gets stored (permanently)

The raw Spotify data is not stored. Instead, the system derives and stores:

What gets shown to other users

When you match with someone, the match card shows:

It does not show your full library, your listening frequency, your recently played tracks, or any data from integrations the other person has not also connected. The principle is simple: show the overlap, not the individual.

๐Ÿ›ก๏ธ The practical test: If Affinity Atlas's entire database were breached tomorrow, an attacker would find derived taste signals and anonymised compatibility scores - not raw Spotify libraries, not Untappd check-in locations, not GitHub repository lists. The raw data never persists beyond the processing window.

Technical safeguards

Beyond the consent model and derived signals, Affinity Atlas implements several technical safeguards:

Data minimisation by default

Every API call requests the minimum scope necessary. The Spotify integration does not request access to your email, your payment method, or your social graph - even though Spotify's API would allow it. The system asks for exactly what it needs for matching and nothing more. This is not just good practice - it is the GDPR principle of data minimisation applied as an engineering constraint, not a policy afterthought.

No third-party data sharing

Your data - raw or derived - is never shared with advertisers, analytics companies, AI training pipelines, or any third party. Period. This is possible because Affinity Atlas is a passion project with no advertising revenue model. There is no structural incentive to monetise user data, because there are no investors demanding growth metrics that require it.

Compare this to the Norwegian Consumer Council's findings that Grindr alone was sharing data with over 50 advertising partners. Or the EFF's documentation of Bumble sharing user data with OpenAI without explicit consent.

OAuth tokens, not passwords

Affinity Atlas never sees your password for any connected platform. All integrations use OAuth 2.0, the industry standard for delegated authorisation. You authenticate directly with Spotify, Steam, or whichever platform you are connecting - and grant Affinity Atlas a scoped, revocable token. Revoking access from either the Affinity Atlas settings or the connected platform's own settings immediately invalidates the token.

Encryption at rest and in transit

All stored data is encrypted at rest. All API communications use TLS. Derived signals are stored in encrypted database fields with access controls that prevent even database administrators from reading individual user profiles without an audit trail.

Transparent data export

Users can export everything Affinity Atlas stores about them - every derived signal, every match score, every consent record - in a machine-readable format. This is the GDPR right of access implemented as a one-click feature, not a 30-day support ticket.

How this compares

Here is how Affinity Atlas's privacy model compares to mainstream dating apps across the dimensions that matter most:

Dimension Affinity Atlas Mainstream Apps
Default data collection โœ“ Opt-in only, per integration โœ— Collects by default; opt-out buried in settings
What is stored โœ“ Derived signals only; raw data discarded after processing โœ— Raw data retained indefinitely; often unclear what is kept
Third-party sharing โœ“ None. No advertisers, no analytics, no AI training โœ— Shared with dozens of ad partners
Consent granularity โœ“ Per-integration, per-category toggles โ—‹ All-or-nothing; take it or leave it
Revocation โœ“ Instant disconnect; derived signals deleted โœ— Data retained even after deletion requests
Breach exposure risk โœ“ Derived signals only - no raw libraries, no locations, no messages โœ— Full profiles, messages, photos, GPS coordinates
Data export โœ“ One-click, machine-readable export โ—‹ Available but often requires manual request; slow turnaround
Transparency โœ“ Users see exactly what is stored, scored, and shown โœ— Privacy policies average 8,000+ words; real practices opaque

Why the structural difference matters

The most important privacy feature of Affinity Atlas is not a technical safeguard. It is the absence of a business model that requires data exploitation.

Match Group, Bumble, and other publicly traded dating companies are structurally incentivised to collect, retain, and monetise user data. Their advertising revenue, investor reporting, and growth metrics all depend on having large, rich datasets. Privacy protections in this context are always in tension with the core business model.

Affinity Atlas, as an independent passion project, does not have this tension. There are no investors demanding user data for ad targeting. There is no advertising revenue stream to protect. There is no incentive to retain data beyond what is needed for matching. The privacy architecture is not a compromise - it is the design.

This structural independence means privacy-first decisions are baked into the foundation rather than bolted on later. No investor pressure, no advertising revenue to protect, no incentive to compromise.


Your data, your call. Always.

Affinity Atlas treats privacy as a feature, not a footnote. Explore the demo to see how transparent, consent-first matching works.

Try the demo