Strip every SSN, PHI, secret, and identifier before any prompt reaches the model.
PortEden's redaction engine inspects every field bound for Claude, ChatGPT, Copilot, or Gemini — masking the 18 HIPAA Safe Harbor identifiers, payment data, API keys, and your custom rules at the egress boundary. The AI gets useful structure. Your data never leaves your perimeter.
Free tier · No credit card · Works with any AI client
Every prompt is an unmonitored data export.
The moment your team types into ChatGPT, Claude, Copilot, or Gemini, your data leaves your perimeter. Once it lands at OpenAI, Anthropic, Google, or Microsoft, you have no control over how long it lives, who reads it, or whether it shows up in someone else's training set.
Your team pastes customer emails into ChatGPT
One copy-paste sends names, addresses, account numbers, and full message bodies to OpenAI's servers in plain text. By the time procurement finds out, it's been logged, possibly used for abuse review, and you have no record of what went where.
Claude summarizes a contract and ingests every clause
Counterparty names, settlement figures, NDAs, and proprietary terms are now sitting in Anthropic's context window — protected by their privacy policy, not your control. If a client or auditor asks what data was disclosed, you can't answer.
A developer pastes a stack trace with API keys into Copilot
Error logs and config dumps are full of secrets — DB connection strings, JWTs, AWS access keys, OAuth tokens. Your secrets scanning catches Git pushes; it doesn't catch AI chat.
Sensitive data, redacted before it reaches the model.
PortEden inspects every field. SSNs, PHI, payment data, secrets, and your custom identifiers are masked at the boundary — never sent to OpenAI, Anthropic, Google, or Microsoft.
200+ patterns, four pillars of protection.
PortEden ships with deterministic regex, transformer NER, and Luhn-validated payment detectors. Add your own rules for matter numbers, claim IDs, internal SKUs — anything with a pattern.
PHI · 18 HIPAA Safe Harbor identifiers
- Patient names
- Geographic subdivisions smaller than a state
- All elements of dates (DOB, admission, discharge)
- Phone & fax numbers
- Email addresses
- Social Security Numbers
- Medical Record Numbers (MRN)
- Health plan beneficiary numbers
- Account numbers
- Certificate / license numbers
- Vehicle identifiers (VIN, plate)
- Device identifiers and serial numbers
- URLs
- IP addresses
- Biometric identifiers
- Full-face photographs
- Any other unique identifying number / characteristic
- Diagnoses & procedure codes (ICD-10, CPT)
PCI · Payment Card Industry data
- Primary Account Numbers (PAN, with Luhn check)
- Card expiration dates
- CVV / CVC / CID
- Cardholder names tied to PAN
- IBAN & SWIFT codes
- ACH routing & account numbers
- Stripe / Plaid / payment processor IDs
- Tax ID / EIN / VAT numbers
PII · GDPR Article 4 personal data
- Full names & aliases
- Government IDs (passport, driver's license)
- Date of birth & age combined with location
- Home & work addresses
- Geolocation coordinates
- Online identifiers (cookie IDs, device fingerprints)
- IP addresses & MAC addresses
- Biometric & genetic data
Secrets · keys, tokens, credentials
- API keys (OpenAI, Stripe, Twilio, SendGrid)
- AWS access keys & session tokens
- GCP service-account JSON
- Azure connection strings
- GitHub & GitLab PATs
- JWTs & OAuth bearer tokens
- Private keys (RSA, EC, SSH, PGP)
- Database connection strings & passwords
Inspect. Detect. Redact. Re-hydrate.
1. Inspect
Every payload bound for an AI client is parsed in your perimeter. Email bodies, attachments, calendar events, chat messages, document text, and code snippets are tokenized and scanned by both regex and an NER model.
2. Detect
Layered detection: ~120 deterministic patterns (SSN, PCI, IBAN, JWT, AWS keys, GitHub PATs), a transformer NER for context-dependent entities (names, MRNs, locations, organizations), and your custom rules. Detections are confidence-scored.
3. Redact
Sensitive values are replaced with structure-preserving placeholders ([PERSON_1], [SSN_a4f2], [DATE_2024-03-15]) so the LLM still has enough signal to be useful. The original values go into a short-lived encrypted token vault.
4. Re-hydrate
When the model's response comes back referencing the placeholders, PortEden swaps them for real values in the user's browser. The model never sees the originals; the user never sees the placeholders. Every event is logged.
Same email, two very different exposures.
Here's a real-looking message from your inbox. On the left, what you typed. On the right, exactly what reaches the AI. The model still has enough context to draft a useful reply — your patient's details never leave your network.
Hi team — quick update on Jane Doe (DOB 03/15/1985).
Her MRN is 0048-2231 and her insurance member ID is BC889922.
She called from (555) 123-4567 about her prescription.
Her card on file (4532-1234-5678-9012) was declined yesterday.
Can you draft a reply asking her to confirm the new copay amount?
Hi team — quick update on [PERSON_1] (DOB [DATE_1]).
Her MRN is [MRN_1] and her insurance member ID is [MEMBER_1].
She called from [PHONE_1] about her prescription.
Her card on file ([CARD_1]) was declined yesterday.
Can you draft a reply asking her to confirm the new copay amount?
Of course — here's a friendly draft asking Jane Doe to confirm the updated copay for her prescription, with a link to update her card on file. Tone is reassuring and avoids referencing the decline directly. Want me to adjust the formality?
The model wrote [PERSON_1] in its response. PortEden swapped it back to Jane Doe in your browser before you saw the reply — so you get the useful answer, and her name never reaches OpenAI or Anthropic.
The same workflow, two very different audit trails.
Citations, not vague reassurances.
Each redaction control maps to a specific clause in the framework your auditor is reading. Evidence is exportable from the audit trail.
Every source your AI tries to read from.
One redaction engine, six regulated workflows.
Egress redaction, not endpoint trust.
Browser extensions and per-app DLP miss as much as they catch — every new AI tool is another rollout. PortEden sits at the integration boundary instead, so the rules apply uniformly whether the prompt comes from Claude desktop, ChatGPT web, a Cursor agent, or an MCP server.
Integration-side, not endpoint-side
Redaction happens once at the data source — Gmail, Drive, Slack, etc. — not on every laptop. Add a new AI client and rules apply automatically.
Tokens stay in your tenant
Original values are stored encrypted in a short-lived per-tenant vault. Default 5-minute TTL, configurable. Vault is isolated; the model never reaches it.
Every event is an audit record
Rule fired, count, category, integration, user, AI client, timestamp, payload hash. Stream to Splunk, Datadog, Elastic, or download a signed CSV.
Pairs well with
Data redaction questions
What is AI data redaction and why do I need it?
Which identifiers does PortEden's redaction engine cover by default?
How does redaction work technically — regex, ML, or both?
Will redaction break the AI's ability to be useful?
What evidence does this produce for HIPAA, GDPR, PCI-DSS, and SOC 2 auditors?
Can I see exactly what was redacted from each prompt?
Does redaction work with Claude, ChatGPT, Copilot, and Gemini?
How fast is the redaction engine? Will it slow down my AI workflows?
Can I customize redaction rules for my own internal identifiers?
What happens to data we've already sent to OpenAI or Anthropic before installing PortEden?
Is the redaction reversible? How do I see the original values?
What pricing tier includes data redaction?
Stop your data from leaking into someone else's AI.
Set up redaction in under 5 minutes. Free tier covers solo users; Enterprise adds SSO/SAML, SCIM, and SIEM export.