⚠
High Risk:This skill has significant security concerns. Review the findings below before installing.
voice-note-to-midi
⚠Caution·Scanned 2/17/2026
This skill converts voice notes, humming, and melodic recordings into quantized MIDI using a Python pipeline and Spotify's Basic Pitch. The package includes a shell setup.sh that installs packages, downloads hum2midi from https://raw.githubusercontent.com/basic-pitch/basic-pitch/main/hum2midi, and appends export PATH to ~/.bashrc, enabling network downloads and local execution.
from clawhub.ai·v522c387·19.8 KB·0 installs
Scanned from 0.1.0 at 522c387 · Transparency log ↗
$ vett add clawhub.ai/danbennettuk/voice-note-to-midiReview security findings before installing
🎵 Voice Note to MIDI
Convert voice memos, humming, and melodic recordings to clean, quantized MIDI files.
Transform your 3 AM melody ideas into MIDI notes for your DAW — no keyboard required.
What It Does
Voice Note (WAV/M4A/MP3)
↓
┌─────────────────────────────────────┐
│ 1. Stem Separation (HPSS) │
│ Isolate melody from noise │
└─────────────────────────────────────┘
↓
┌─────────────────────────────────────┐
│ 2. Basic Pitch ML (Spotify) │
│ Detect fundamental frequencies │
└─────────────────────────────────────┘
↓
┌─────────────────────────────────────┐
│ 3. Key Detection │
│ Identify musical key │
└─────────────────────────────────────┘
↓
┌─────────────────────────────────────┐
│ 4. Quantization & Cleanup │
│ • Snap to timing grid │
│ • Key-aware pitch correction │
│ • Harmonic pruning (octave/overlap)│
│ • Note merging (legato) │
└─────────────────────────────────────┘
↓
MIDI File → Your DAW
Quick Start
git clone https://github.com/DanBennettUK/voice-note-to-midi.git
cd voice-note-to-midi
./setup.sh
Then convert a voice memo:
hum2midi my_humming.wav # Creates my_humming.mid
hum2midi voice.wav song.mid # Custom output name
hum2midi hum.wav --key-aware # Auto-detect key & quantize
Features
- ML-Powered Pitch Detection — Spotify's Basic Pitch model
- Key Detection — Automatic musical key identification
- Key-Aware Quantization — Snap notes to detected scale
- Harmonic Pruning — Remove overtones, keep fundamentals
- Legato Merging — Combine note chunks into sustained tones
- Configurable Grid — 1/4, 1/8, 1/16, or 1/32 note quantization
Usage
hum2midi <input.wav> [output.mid] [options]
Options:
--grid <value> Quantization: 1/4, 1/8, 1/16, 1/32 (default: 1/16)
--min-note <ms> Minimum note duration (default: 50ms)
--key-aware Enable key-aware pitch correction
--no-quantize Skip quantization (raw Basic Pitch output)
--no-analysis Skip pitch analysis
Requirements
- Python 3.11+
- FFmpeg (optional but recommended)
- See
setup.shfor full dependency installation
Limitations
- Monophonic only — one note at a time (voice can't do chords)
- Pitched audio required — humming/singing works, whisper/"air notes" don't
- Quality matters — loud, clear melody = better results
- Background noise — can confuse pitch detection
How It Works
The pipeline uses a multi-stage approach:
- HPSS separates harmonic (melodic) content from percussive sounds
- Basic Pitch neural network extracts pitch information
- Music21 analyzes pitch classes to detect the musical key
- Post-processing cleans up harmonics and merges legato notes
Documentation
- Full Guide: See SKILL.md
- Quick Reference: See QUICKSTART.md
- Setup Script: setup.sh
License
MIT — see LICENSE
Uses Basic Pitch by Spotify, librosa, and music21.
Made with 🦊 for Clawdbot