Skip to content

Fine-Tuning Hybrid Demucs in Cross-Talk scenario with Self-Knowledge Distillation with Progressive Refinement of Targets

Notifications You must be signed in to change notification settings

spear011/mss-pskd

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

4 Commits
 
 
 
 
 
 
 
 

Repository files navigation

Hybrid Demucs with Self-Knowledge Distillation (PS-KD)

This repository introduces the Self-Knowledge Distillation with Progressive Refinement of Targets (PS-KD) approach to training Hybrid Demucs. The model training incorporates a curriculum learning approach. Additionally, it includes the codebase for the Cadenza Challenge 2024.

Music Source Separation Training Code

Training code for the mss models based on [here]

Results

  • The results indicate the performance of the validation set for the cadenza challenge.
Title PS-KD Curriculum-learning Augmentation SDR avg Haaqi Score
no fine-tune - - - 3.701 0.6677
w/o aug - - - 4.1838 0.6776
w/ aug - - O 4.0762 0.6733
PS-KD w/o aug O - - 4.2505 0.6764
PS-KD w/ aug O - O 4.4060 0.6818
PS-KD Curri w/o aug O O - 4.2002 0.6772
PS-KD Curri w/ aug O O O 4.5481 0.6836

PS-KD

  • The PS-KD method uses targets by creating soft targets, and we used an alpha of 0.8.

Curriculum-learning

  • We took the SDR score and used the top 75% of the scoring datasets as the EASY dataset.

Augmentation

  • We used a random augmentation method on pitch {-2, -1, 0, 1, 2} tempo {-20, -10, 0, 10, 20} with random augmentation. We also applied a random source mix, channel shuffle, and random audio effects (reverb, phaser, distortion) with a 0.05% probability. Soundstretch was used to augment pitch and tempo, and other techniques are available in code.

About

Fine-Tuning Hybrid Demucs in Cross-Talk scenario with Self-Knowledge Distillation with Progressive Refinement of Targets

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages