Is segmentation of sounds works well ?
I`ve listened to TalkingMachinesPodcast.wav, watched waves spectre in Camtasia Studio and compared with output of
segmentation.py:
('0.0m 0.0s', 0)
('0.0m 2.Ss', 1)
('0.0m 5.0s', 0)
( '0.0m 7 .Ss ', 1)
('0.0m 10.0s', 1)
('0.0m 12.Ss', 1)
('0.0m 15.0s', 1)
('0.0m 17.Ss', 0)
('0.0m 20.0s', 1)
('0.0m 22.Ss', 1)
('0.0m 25.0s', 0)
('0.0m 27.Ss', 0)
And i dont think that it works correct as it makes segment no exectly close where they should be.
Moreover if i put k=3 i see:
('0.0m 0.0s', 0)
('0.0m 5.0s', 1)
('0.0m 10.0s', 2)
('0.0m 15.0s', 2)
('0.0m 20.0s', 1)
('0.0m 25.0s', 2)
('0.0m 30.0s', 1)
('0.0m 35.0s', 2)
('0.0m 40.0s', 1)
('0.0m 45.0s', 2)
('0.0m 50.0s', 2)
('0.0m 55.0s', 2)
Why timeline in code run away ? And we have much more clusters ?
Is segmentation of sounds works well ? I`ve listened to TalkingMachinesPodcast.wav, watched waves spectre in Camtasia Studio and compared with output of segmentation.py: ('0.0m 0.0s', 0) ('0.0m 2.Ss', 1) ('0.0m 5.0s', 0) ( '0.0m 7 .Ss ', 1) ('0.0m 10.0s', 1) ('0.0m 12.Ss', 1) ('0.0m 15.0s', 1) ('0.0m 17.Ss', 0) ('0.0m 20.0s', 1) ('0.0m 22.Ss', 1) ('0.0m 25.0s', 0) ('0.0m 27.Ss', 0)
And i dont think that it works correct as it makes segment no exectly close where they should be. Moreover if i put k=3 i see: ('0.0m 0.0s', 0) ('0.0m 5.0s', 1) ('0.0m 10.0s', 2) ('0.0m 15.0s', 2) ('0.0m 20.0s', 1) ('0.0m 25.0s', 2) ('0.0m 30.0s', 1) ('0.0m 35.0s', 2) ('0.0m 40.0s', 1) ('0.0m 45.0s', 2) ('0.0m 50.0s', 2) ('0.0m 55.0s', 2) Why timeline in code run away ? And we have much more clusters ?