๐Ÿฅ Open-Source Clinical Documentation

Your Conversations.
Your Notes. Your Machine.

FerriScribe listens to your doctor-patient conversations and generates structured SOAP notes using a local AI model of your choice that runs entirely on your computer. No cloud APIs. No data transmission. No subscriptions. HIPAA-compliant by design โ€” because nothing ever leaves your device.

Why FerriScribe?
๐ŸŽฏ

Clinically Accurate

System prompt systematically optimized and benchmarked across 21 real clinical cases. Achieves 90%+ schema compliance with zero fabricated patient data โ€” every field defaults to "Not discussed" unless explicitly stated in the transcript.

๐Ÿ”’

100% Local & Private

All audio processing and note generation happen on your machine. No cloud APIs, no data transmission, no third-party access. HIPAA-compliant by architecture, not by policy.

โšก

Minutes, Not Hours

Records the consultation, transcribes speech to text, and generates a structured SOAP note โ€” ready for your review before the next patient sits down.

๐Ÿ› ๏ธ

Adapts to Your Practice

Fine-tune the model to your specialty, your templates, and your documentation style. Support for context templates, custom vocabulary, and patient history injection. Your notes, your way.

๐ŸŒ

Fully Open Source

Audit every line of code. Fork it, contribute, or customize it for your clinic. No black boxes, no vendor lock-in, no future price hikes.

๐Ÿ–ฅ๏ธ

Cross-Platform

macOS, Linux, and Windows. Runs on any capable machine with 16GB+ RAM โ€” from a MacBook Air to a clinic workstation.

๐Ÿ’ฐ

Zero Recurring Costs

No per-seat licenses. No API fees. No cloud subscriptions. Download it, run it, own it. One less line item on your practice budget.

๐Ÿ”—

Multi-Device Sharing

Pair multiple devices to a central office server over your local network. Record on a laptop in one room, generate notes on the desktop in another. Connected clients auto-resume across restarts.

๐Ÿ’พ

Runs on Consumer Hardware

Proven on Apple Silicon Mac Studio with 192GB unified memory. Runs quantized models (Qwen 3.6 35B A3B) via LM Studio โ€” no GPU server, no special equipment needed.

Three Steps to Done
1

Record

Hit record before the consultation. FerriScribe captures audio and transcribes it locally with high-quality speech-to-text via Whisper.

2

Process

The local AI model analyzes the transcript on-device โ€” extracting clinical findings, ICD codes, differential diagnoses, assessments, and plans.

3

Document

A structured SOAP note appears in seconds. Review, edit, and copy it into your EMR. Done before your next patient.

๐Ÿ›ก๏ธ

HIPAA-Compliant by Design

FerriScribe runs entirely on your local machine โ€” no cloud APIs, no data transmission, no third-party servers. Patient conversations never leave your device. This isn't a privacy policy you have to trust; it's an architecture you can verify. Every line of code is open source and auditable.

โœ“ Tested in Production

Battle-Tested in a Live Clinic

FerriScribe isn't a prototype โ€” it's running daily in a real family practice, generating SOAP notes from actual patient encounters. Here's the reference deployment:

๐Ÿ–ฅ๏ธ Hardware

  • Apple Mac Studio
  • M-series chip with 96GB unified memory
  • All processing on a single silent desktop
  • No external GPU or server hardware

๐Ÿง  AI & Speech Models

  • LM Studio โ€” local model serving (localhost:1234)
  • Qwen 3.6 35B A3B (quantized, MLX-accelerated)
  • Whisper โ€” local speech-to-text transcription
  • Tauri v2 โ€” native desktop framework

๐ŸŒ Networking

  • Tailscale โ€” secure mesh VPN between devices
  • Multi-device pairing over local network
  • Office server mode with auto-resume
  • Zero exposed ports to the public internet
Runs on What You Have
๐ŸŽ

macOS

Native Apple Silicon support with MLX optimization. Runs great on MacBook Air, MacBook Pro, and Mac Studio with M-series chips.

๐Ÿง

Linux

Full NVIDIA CUDA and AMD ROCm support. Run it on your workstation, a home server, or a dedicated clinic machine.

๐ŸชŸ

Windows

DirectML and CUDA support. Works on any capable Windows laptop or desktop with sufficient RAM for local inference.

The only requirement: a computer with enough RAM to run local models (16GB+ recommended).

Tech Stack
Qwen 3.6 35B
A3B quantized model
LM Studio
Local model serving
Whisper
Speech-to-text
Tauri v2
Desktop framework
Rust
Core engine (12 crates)
Tailscale
Secure mesh VPN
BERTScore
0.837 semantic accuracy
90%+ Schema
SOAP compliance rate