Multimodal Art Projection
Multimodal Art Projection (M-A-P) is an opensource research community. The coummnity members are working on Artificial Intelligence-Generated Content (AIGC) topics, including text, audio, and vision modalities. We do large language models (LMMs) training, data collection, and development of fun applications. Welcome to join us~ See our released models at our huggingface organization page: https://huggingface.co/m-a-p.
Projects
-
MERT
MERT is a series of large-scale acoustic music understanding pre-trained models.
-
MARBLE
MARBLE is a benchmark proposed to help the academic & industrial to study, compare, and select pre-trained models according to comprehensive evaluation.
-
MuPT
MuPT is a series of pre-trained models for symbolic music generation. It was trained on a large-scale dataset of symbolic music, including millions of monophonic and polyphonic pieces from different genres and styles.
-
COIG Series
Chinese Open Instruction Generalist is a series of large-scale Chinese textual datasets for supervised fine-tuning.
-
SciMMIR
The scientific multimodal information retrieval (SciMMIR) is a image-text retrieval benchmark with 500K pairs extracted from the scholarly papers.
News
- 2022-07-20» The Launch Of Multimodal Art Projection