dpwe / audfprint

Landmark-based audio fingerprinting
MIT License
544 stars 122 forks source link

How to avoid big % Dropped #94

Open lightcomc opened 7 months ago

lightcomc commented 7 months ago

Hello i want to scan all collection I read that to improve accuracy, you need to increase the density. I increased the Desity parameter from 20 to 100 and the result is a higher % Dropped. (for example YOUTUBE 01 D2.pklz (79.70% dropped)) Some of the records became less recognizable than with standard Density = 20 How can I increase the output file and decrease the % Dropped to increase the overall recognition accuracy. I take as a basis the material I want to recognize low quality audio from tapes from the early 90's, so accuracy is important. It doesn't matter how much space the database of fingerprints will take up.

dpwe commented 7 months ago

“% dropped” comes from buckets overflowing in the database. You can increase the database size with the bucketsize flag; the default is 100. The database in memory will grow in proportion, and at some point you’ll run out of RAM.

DAn.

On Wed, Mar 6, 2024 at 07:27 lightcomc @.***> wrote:

Hello i want to scan all collection I read that to improve accuracy, you need to increase the density. I increased the Desity parameter from 20 to 100 and the result is a higher % Dropped. (for example YOUTUBE 01 D2.pklz (79.70% dropped)) Some of the records became less recognizable than with standard Density = 20 How can I increase the output file and decrease the % Dropped to increase the overall recognition accuracy. I take as a basis the material I want to recognize low quality audio from tapes from the early 90's, so accuracy is important. It doesn't matter how much space the database of fingerprints will take up.

— Reply to this email directly, view it on GitHub https://github.com/dpwe/audfprint/issues/94, or unsubscribe https://github.com/notifications/unsubscribe-auth/AAEGZUKBROI5XN6CAC73KYLYW4DRZAVCNFSM6AAAAABEI6VCB6VHI2DSMVQWIX3LMV43ASLTON2WKOZSGE3TCMZYHE3TKMI . You are receiving this because you are subscribed to this thread.Message ID: @.***>

lightcomc commented 7 months ago

Thank you i'll check thatt parameter to find optimal size and memory comsumption in process, speed is not critical. Started from 4x of current size (400 mb to 1200 mb database) i have about 12 000 tracks (64 gb of music) in each scan

ZhymabekRoman commented 7 months ago

You can try to use my fork with some minor changes. Recognition speed is increased and memory consumption is reduced. - https://github.com/ZhymabekRoman/audfprint-enhanced

godzfire commented 6 months ago

You can try to use my fork with some minor changes. Recognition speed is increased and memory consumption is reduced. - https://github.com/ZhymabekRoman/audfprint-enhanced

Are you doing any active commits anymore?

ZhymabekRoman commented 6 months ago

You can try to use my fork with some minor changes. Recognition speed is increased and memory consumption is reduced. - ZhymabekRoman/audfprint-enhanced

Are you doing any active commits anymore?

Yeap