From 09561c565d02239dce7e8de82ccb5c6798a0a27b Mon Sep 17 00:00:00 2001 From: Andrew Duberstein Date: Thu, 11 Sep 2014 17:40:03 -0700 Subject: [PATCH] Update README.md Corrected a typo --- README.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/README.md b/README.md index ee46dbb..64e2fc7 100755 --- a/README.md +++ b/README.md @@ -179,7 +179,7 @@ The algorithm works off a fingerprint based system, much like: * [MusicRetrieval](http://www.cs.cmu.edu/~yke/musicretrieval/) * [Chromaprint](https://oxygene.sk/2011/01/how-does-chromaprint-work/) -The "fingerprints" are locality sensitve hashes that are computed from the spectrogram of the audio. This is done by taking the FFT of the signal over overlapping windows of the song and identifying peaks. A very robust peak finding algorithm is needed, otherwise you'll have a terrible signal to noise ratio. +The "fingerprints" are locality sensitive hashes that are computed from the spectrogram of the audio. This is done by taking the FFT of the signal over overlapping windows of the song and identifying peaks. A very robust peak finding algorithm is needed, otherwise you'll have a terrible signal to noise ratio. Here I've taken the spectrogram over the first few seconds of "Blurred Lines". The spectrogram is a 2D plot and shows amplitude as a function of time (a particular window, actually) and frequency, binned logrithmically, just as the human ear percieves it. In the plot below you can see where local maxima occur in the amplitude space: