a43992899 commited on
Commit
f7159a3
β€’
1 Parent(s): adf53b5

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -2
README.md CHANGED
@@ -16,8 +16,10 @@ Welcome to join us!
16
  Organization page: https://m-a-p.ai
17
 
18
  The development log of our Multimodal Art Projection (m-a-p) model family:
19
- - πŸ”₯2024-2-28: The release of [**ChatMusician**](https://huggingface.co/collections/m-a-p/chatmusician-65de07b3b87b189c2a588329)'s demo, code, model, data, and benchmark. πŸ˜†
20
- - πŸ”₯2024-2-23: The release of [**OpenCodeInterpreter**](https://huggingface.co/collections/m-a-p/opencodeinterpreter-65d312f6f88da990a64da456), beats GPT-4 code interpreter on HumanEval.
 
 
21
  - 23/01/2024: we release [**CMMMU**](https://huggingface.co/datasets/m-a-p/CMMMU) for better Chinese LMMs' Evaluation.
22
  - 13/01/2024: we release a series of **Music Pretrained Transformer (MuPT)** checkpoints, with [**size up to 1.3B and 8192 context length**](https://huggingface.co/m-a-p/MuPT_v0_8192_1.3B). Our models are **LLAMA2**-based, pre-trained on **world's largest 10B tokens symbolic music dataset** ([ABC notation format](https://en.wikipedia.org/wiki/ABC_notation)). We currently support Megatron-LM format and will release huggingface checkpoints soon.
23
  - 02/06/2023: officially release the [MERT pre-print paper](https://arxiv.org/abs/2306.00107) and training [codes](https://github.com/yizhilll/MERT).
 
16
  Organization page: https://m-a-p.ai
17
 
18
  The development log of our Multimodal Art Projection (m-a-p) model family:
19
+ - πŸ”₯11/04/2024: MuPT [**paper**](https://arxiv.org/abs/2404.06393) and [**demo**](https://twitter.com/GeZhang86038849/status/1778089599035941279) is out.
20
+ - πŸ”₯08/04/2024: Chinese Tiny LLM is out. [**HF collection**](https://huggingface.co/collections/m-a-p/chinese-tiny-llm-660d0133dff6856f94ce0fc6).
21
+ - πŸ”₯28/02/2024: The release of [**ChatMusician**](https://huggingface.co/collections/m-a-p/chatmusician-65de07b3b87b189c2a588329)'s demo, code, model, data, and benchmark. πŸ˜†
22
+ - πŸ”₯23/02/2024: The release of [**OpenCodeInterpreter**](https://huggingface.co/collections/m-a-p/opencodeinterpreter-65d312f6f88da990a64da456), beats GPT-4 code interpreter on HumanEval.
23
  - 23/01/2024: we release [**CMMMU**](https://huggingface.co/datasets/m-a-p/CMMMU) for better Chinese LMMs' Evaluation.
24
  - 13/01/2024: we release a series of **Music Pretrained Transformer (MuPT)** checkpoints, with [**size up to 1.3B and 8192 context length**](https://huggingface.co/m-a-p/MuPT_v0_8192_1.3B). Our models are **LLAMA2**-based, pre-trained on **world's largest 10B tokens symbolic music dataset** ([ABC notation format](https://en.wikipedia.org/wiki/ABC_notation)). We currently support Megatron-LM format and will release huggingface checkpoints soon.
25
  - 02/06/2023: officially release the [MERT pre-print paper](https://arxiv.org/abs/2306.00107) and training [codes](https://github.com/yizhilll/MERT).