Hey r/bard! Been diving deep into the possibilities of direct AI music creation and wanted to share a concept I've been tinkering on that leverages the power of LLMs like Gemini 2.5. The idea is to create a pipeline where an AI could take a simple musical idea and generate an entire 5-minute track, complete with development and structure (think A-B-C-D-E sections!). Here's the envisioned workflow: 1. Start with a small text description of the music. 2. A Gemini LLM enriches this with detailed musical elements. 3. The LLM then converts this detailed description into a compact symbolic representation of the entire track, broken down into musical sections to allow for development over time. 4. This symbolic data is algorithmically translated into a structured JSON format that's specifically designed for easy conversion to MIDI. 5. Finally, this JSON is used to generate a standard MIDI file. What do you all think about this kind of approach? Could we see LLMs like this becoming serious tools for composers in the future, maybe even assisting with complex arrangements or offering completely new creative avenues? Edit: I actually asked Pro 2.5 to implelement all this and got my first piano piece, very sad but cool [https://g.co/gemini/share/2460a7f4f617](https://g.co/gemini/share/2460a7f4f617)