Skip to content

Commit

Permalink
Update README.md
Browse files Browse the repository at this point in the history
  • Loading branch information
jordipons authored Apr 16, 2024
1 parent 4a5f329 commit 8c1ec49
Showing 1 changed file with 1 addition and 1 deletion.
2 changes: 1 addition & 1 deletion README.md
Original file line number Diff line number Diff line change
Expand Up @@ -2,7 +2,7 @@

Audio-based generative models for music have seen great strides recently, but so far have not managed to produce full-length music tracks with coherent musical structure. We show that by training a generative model on long temporal contexts it is possible to produce long-form music of up to 4m\,45s. Our model consists of a diffusion-transformer operating on a highly downsampled continuous latent representation {(latent rate of 21.5\,Hz)}. It obtains state-of-the-art generations according to metrics on audio quality and prompt alignment, and subjective tests reveal that it produces full-length music with coherent structure.

## Comparison with state-of-the-art (w/ song describer dataset prompts)
## Comparison with state-of-the-art (song describer dataset prompts)

**Prompt**: An uplifting jazz song that makes your head shake.

Expand Down

0 comments on commit 8c1ec49

Please sign in to comment.