Endpoint for requesting text2music inference
reference_audio file for style transfer — the model will use it as a stylistic reference for the generated track. The endpoint returns a task ID to track processing status. Ideal for apps needing automated music creation — background tracks, jingles, or full songs with vocals.
slug and check specific limits.Bearer authentication header of the form Bearer <token>, where <token> is your auth token.
application/json Music generation parameters
Text description of the music to generate
"upbeat electronic dance music with energetic synths"
The model to use for music generation. Available models can be retrieved via the GET /api/v1/client/models endpoint.
"ACE-Step-v1.5-turbo"
Lyrics for the music. Use "[Instrumental]" for instrumental tracks without vocals.
"[Instrumental]"
Duration in seconds (10-600)
30
Number of diffusion inference steps (1-100). Use 8 for turbo models, 32+ for base models.
8
Classifier-free guidance scale (0-20)
7
Random seed. Use -1 for random.
-1
Audio output format
"flac"
Beats per minute (30-300)
120
Musical key/scale (e.g. "C major", "F# minor")
"C major"
Time signature. Must be 2, 3, 4, or 6.
4
Language code for vocals (e.g. "en", "es", "fr")
"en"
Optional reference audio file for style transfer. Supported formats: mp3, wav, flac, ogg, m4a. Max size configurable (default 10MB). Duration must be within model-specific limits.
Optional HTTPS URL to receive webhook notifications for job status changes (processing, completed, failed). Must be HTTPS. Max 2048 characters.
2048"https://your-server.com/webhooks/deapi"
ID of the inference request.
Information from success endpoint