Text Translation
Detect and replace PII in medical documents using WebLLM + WebGPU — entirely in your browser.
⚠️ Disclaimer: Anonymization is not guaranteed to detect all PII. Always manually review the output before sharing documents. This tool is an aid, not a replacement for human review.
Pipeline details & hardware
LLM only (recommended): Qwen3 handles all PII detection in a single pass. Best accuracy, especially with the 4B or 8B model.
NER + LLM: Run a NER model first, then let the LLM verify and catch remaining PII. Useful for additional coverage or comparison.
NER only: Run just the selected NER model. Fastest option, no GPU required, but lower accuracy.
- Qwen3 0.6B — ~1.4 GB VRAM. Smallest & fastest. Good for quick scans.
- Qwen3 1.7B — ~2 GB VRAM. Good balance of speed and quality.
- Qwen3 4B (default) — ~3.4 GB VRAM. Best quality for most devices.
- Qwen3 8B — ~5.7 GB VRAM. Highest quality, needs a powerful GPU.
- Multilingual PII NER — ~280 MB. XLM-RoBERTa, no GPU needed. Names, addresses, dates, IDs in 8+ languages.
- GLiNER PII Edge — ~46 MB. Zero-shot, no GPU needed. Best for English.
- Multilingual BERT NER — ~100 MB. Lighter general-purpose NER.
Always manually review the output. First load downloads the selected model; subsequent runs use the browser cache.
Medical Document
Drop document here
.xlsx, .docx, .txt
Mapping File Optional
Drop mapping file
.xlsx or .json
Generate structured reports from medical documents using WebLLM + WebGPU — entirely in your browser.
Or Paste Text
Transcribe audio recordings or live microphone input using OpenAI Whisper via Transformers.js — entirely offline.
Model details
- Whisper Tiny — ~150 MB. Fastest, basic quality.
- Whisper Base — ~300 MB. Fair quality.
- Whisper Small (recommended) — ~500 MB. Good quality, best for clinical use.
All models run via WebAssembly (ONNX Runtime). First use downloads the model; subsequent runs use browser cache.
Record Audio
Or Upload Audio
Drop audio file here
.mp3, .wav, .ogg, .webm, .m4a, .flac
Index and sort unsorted DICOM data from any directory or network path. Smart scanning uses file size, naming, and folder structure heuristics to avoid opening every file.
Source Directory
Scan Settings
Download models in advance so you can work offline when patients arrive. Inspect and manage all browser-stored data below.
Prepare for Offline Use
Download all models now while you have internet. Once cached, everything works in airplane mode.
Personal Data — Where It Goes
Your documents, text and patient data are never written to disk or browser storage. They exist only in temporary JavaScript memory and are automatically removed.
Currently in memory
No personal data in memory
Cache API (Model Files)
Translation and NER models are cached here. These are safe AI model weights — no personal data.
Scanning...
IndexedDB (LLM Model Cache)
WebLLM/MLC caches Qwen model weights here. These are safe AI model data — no personal data.
Scanning...
This will remove all cached models. You will need to re-download them on next use.