
In this piece
Qualitative research collects far more personal data than most research teams realize, and the compliance gap usually lives between what teams think they're collecting and what they're actually collecting. Whether you're an agency running studies on behalf of clients or an in-house insights team conducting your own research, that gap is your problem to close.
Key Takeaways
- Qualitative research routinely produces biometric data (voice, video) and sensitive disclosures that most privacy regimes treat as high-risk categories
- The PII surface area of a qual study expands the moment a participant mentions health, finances, or political views, regardless of what the screener asked
- GDPR, PIPL, and CCPA all impose specific obligations on research vendors handling participant recordings and transcripts
- A signed Data Processing Agreement (DPA) is a baseline requirement before processing EU participant data, not an optional addendum
- Compliance posture is a design decision, not a legal review at the end of a project
The PII Surface Area Nobody Maps
Most research teams think of PII as names, email addresses, and maybe IP addresses. The real inventory is wider. A standard qualitative interview produces voice recordings, which qualify as biometric data under GDPR and most similar regimes. If you're running video interviews, you're adding facial imagery. The transcript, which looks like text, contains free-form disclosures: a participant mentions their mother's cancer diagnosis, their household income, their political frustration, their undocumented family member. None of that was in your screener. All of it is now in your data.
Sensitive personal information, as defined under GDPR Article 9, includes health data, racial or ethnic origin, political opinions, religious beliefs, and sexual orientation. Under China's PIPL, the list is similarly broad. The fact that you didn't ask for it doesn't reduce your obligation to handle it appropriately. For in-house teams, the obligation sits with your organization directly. For agencies, it's split: you control the data operationally, but your client may share controller status and carries reputational exposure if something goes wrong. Neither side can outsource accountability to the other.
What "Handling" Actually Requires
Handling PII data in qualitative research has four concrete dimensions that both agencies and end-clients frequently underestimate.
Lawful basis and consent. Under GDPR, you need a documented lawful basis for processing. For research, legitimate interest or explicit consent are the most defensible bases. Consent language must describe what data is collected, how it's stored, whether it's transcribed, who can access it, and how long it's retained. Blanket "your responses may be used for research" language won't hold up. Agencies should not assume the client has handled this; in-house teams should not assume the agency has.
Data processing agreements. If you're passing participant data to a research platform, a transcription service, or an analysis tool, you are sharing personal data with a processor. A DPA is legally required before that transfer occurs if EU residents are involved. For agencies, this means a DPA with your platform and a separate data sharing agreement with your client. For in-house teams running their own studies, it means a DPA with every tool in the stack.
Cross-border transfers. If your platform stores data outside the EU, you need Standard Contractual Clauses or another approved transfer mechanism. Post-Schrems II, SCCs alone require a Transfer Impact Assessment for certain destination countries. Agencies running global studies for enterprise clients face this on every project involving EU participants and non-EU infrastructure. In-house teams working with multinational respondent pools face the same obligation.
Retention and deletion. Participant data should have a defined retention period, enforced programmatically. If a participant exercises their right to erasure, your platform needs to honor it, including deletion from transcripts, analysis outputs, and downstream reports. Agencies need to confirm their clients' retention policies before archiving deliverables that contain participant data.
Run your next study on Enumerate.
See how Enumerate works on a study like yours. Book a 30-minute demo and we'll walk you through it.
Book a demoTailored to your use case
Redaction, Format Choice, and Sharing Deliverables
One of the most underused tools in qual compliance is active redaction before data is shared internally or stored long-term. On the text side, transcript redaction tools can automatically flag and strip direct identifiers: names, locations, employer mentions, and account numbers. The residual transcript is still analytically useful; the identifiable layer is gone. Redacting at the transcript stage is faster and more auditable than trying to scrub downstream reports after the fact.
Format choice matters too. Audio-only interviews generate voice biometrics but not facial imagery. Where video adds little analytical value, audio is the lower-risk default. When video is genuinely necessary, facial masking applied to the recording before it enters shared storage substantially reduces biometric exposure.
The same principles apply when agencies share highlight reels with clients, or when in-house teams share clips with internal stakeholders. Short video clips pulled from interviews to illustrate a key theme are a common deliverable, but they're also a PII transfer. Before a reel goes anywhere outside the core research team, faces should be masked and any incidental identifiers visible on screen should be blurred or cropped. A masked reel preserves the emotional texture and verbal nuance that makes qual compelling, while keeping the participant's identity out of a deck that will circulate well beyond anyone who signed the DPA.
The Research Platform Is in Scope
The research platform you use to conduct interviews, store recordings, and generate transcripts is a data processor in the GDPR sense. Its compliance posture is your exposure, whether you're the agency that procured it or the client whose participant data lives inside it. A platform without SOC 2 Type II certification has not demonstrated that its security controls operate reliably over time. A platform without a clear sub-processor list cannot tell you where your participant data is actually going.
Enumerate's built-in compliance architecture, including SOC 2 Type II and ISO 27001 certification, is designed for exactly this: research teams and the clients they serve shouldn't have to engineer privacy compliance on top of a generic tool. The platform handles participant data with the controls that enterprise procurement and legal teams require, so the conversation with your CISO is a short one, regardless of which side of the agency relationship you're on.
Compliance isn't a legal review at the end of a project. It's a design decision at the beginning. Book a demo with Enumerate to see how compliant qual at scale actually works in practice.
Related Reading

What Are the Most Important Factors in a Successful DIY Research Study?
The most important factors in a successful DIY research study: clear objectives, sound recruitment, incentives, probing depth, and rigorous analysis. A practical guide for in-house teams.
Read more
GDPR and AI Research: What You Actually Need to Know
GDPR compliance for AI-moderated research isn't optional. Here's what research teams and agencies need to know about consent, transfers, data residency, and data handling.
Read more
HIPAA Compliance for Medical Research: What You Actually Need to Know
HIPAA compliance for medical research is trickier than most research teams expect. Here's what qualitative research platforms actually need to get right.
Read more
Run your next study on Enumerate.
See how Enumerate works on a study like yours. Book a 30-minute demo and we'll walk you through it.
Book a demoTailored to your use case