A deep learning library for audio and music analysis, providing time-frequency transforms and feature extraction for tasks like classification and MIR.
audioFlux is a deep learning tool library for audio and music analysis, specializing in time-frequency transformation and feature extraction. It provides dozens of transform methods and hundreds of feature combinations to support tasks like audio classification, music information retrieval (MIR), source separation, and ASR. The library is designed for high performance with a core C implementation and cross-platform support, including mobile devices.
Audio researchers, machine learning engineers, and developers working on audio deep learning tasks such as music analysis, speech processing, and sound classification. It's also suitable for those needing real-time audio feature extraction on mobile platforms.
Developers choose audioFlux for its comprehensive set of audio transforms and features, high performance due to C-core optimization, and flexibility in combining multi-dimensional features for various deep learning models. Its cross-platform nature and mobile support make it unique for both research and production environments.
A library for audio and music analysis, feature extraction.
Open-Awesome is built by the community, for the community. Submit a project, suggest an awesome list, or help improve the catalog on GitHub.
Core implemented in C with FFT hardware acceleration, enabling efficient processing for large-scale data and real-time mobile audio streams, as highlighted in the benchmark results.
Supports dozens of time-frequency analysis methods including BFT, NSGT, CWT, and reassignment techniques like synchrosqueezing, providing sharp representations for diverse audio tasks.
Offers spectral features, cepstral coefficients (e.g., MFCC), deconvolution, and chroma features for various spectrum types, allowing multi-dimensional combinations tailored to deep learning models.
Runs on Linux, macOS, Windows, iOS, and Android, ensuring broad compatibility for research and production environments across devices.
Requires building from source for iOS and Android, as noted in the 'Other Build' section, which adds overhead compared to plug-and-play Python packages.
Primarily provides low-level feature extraction tools; lacks pre-built models or pipelines, forcing users to manually integrate with deep learning frameworks for end-to-end tasks.
With numerous transforms and features, the library assumes familiarity with audio signal processing concepts, which can be overwhelming for developers new to the domain.