2 min read
MGU-V

Published: IEEE Access

MGU-V: Lo-Fi Music Generation Using Variational Autoencoders

Overview

MGU-V is a deep learning-based framework for generating Lo-Fi music using a hybrid approach combining Long Short-Term Memory (LSTM) networks and Variational Autoencoders (VAEs). This model is capable of producing high-quality music by learning from a curated MIDI dataset and generating realistic-sounding Lo-Fi music, which is perfect for use in real-time applications like background scores, video games, or music streaming.

Features

  • Hybrid Architecture: Combines the temporal modeling power of LSTMs with the generative capabilities of VAEs.
  • Custom Datasets: Trained on a merged dataset of over 2300 MIDI files, including datasets like Nottingham Music Database, Maestro Piano Midi, and others.
  • High Performance: Achieves state-of-the-art performance with an accuracy of 96.2% and minimal loss of 0.19.
  • Lo-Fi Music Generation: Specifically designed to generate continuous, high-fidelity Lo-Fi music suitable for listening during concentration or background environments.