xurantju commited on
Commit
c24836a
1 Parent(s): 32753c9

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -4
README.md CHANGED
@@ -12,10 +12,10 @@ We are excited to announce the continuation and rebranding of our **BLIP series*
12
  `XGen-MM` is a series of the latest foundational Large Multimodal Models (LMMs) developed by Salesforce AI Research. This series advances upon the successful designs of the `BLIP` series, incorporating fundamental enhancements that ensure a more robust and superior foundation. These models have been trained at scale on high-quality image caption datasets and interleaved image-text data.
13
 
14
  In the v1.1 (08/2024) release, we present a series of XGen-MM models including:
15
- - Base model `xgen-mm-phi3-mini-base-r-v1.1`
16
- - Single-image instruct model `xgen-mm-phi3-mini-instruct-r-v1.1`
17
- - Multi-image instruct model `xgen-mm-phi3-mini-instruct-multi-r-v1.1`
18
- - DPO instruct model `xgen-mm-phi3-mini-instruct-dpo-r-v1.1`
19
 
20
  In addition to the models, we are also releasing a series of datasets for multi-modal pre-training, including:
21
  - [MINT-1T: Scaling Open-Source Multimodal Data by 10x: A Multimodal Dataset with One Trillion Tokens](https://arxiv.org/abs/2406.11271)
 
12
  `XGen-MM` is a series of the latest foundational Large Multimodal Models (LMMs) developed by Salesforce AI Research. This series advances upon the successful designs of the `BLIP` series, incorporating fundamental enhancements that ensure a more robust and superior foundation. These models have been trained at scale on high-quality image caption datasets and interleaved image-text data.
13
 
14
  In the v1.1 (08/2024) release, we present a series of XGen-MM models including:
15
+ - Base model `xgen-mm-phi3-mini-base-r-v1.5`
16
+ - Single-image instruct model `xgen-mm-phi3-mini-instruct-r-v1.5`
17
+ - Multi-image instruct model `xgen-mm-phi3-mini-instruct-multi-r-v1.5`
18
+ - DPO instruct model `xgen-mm-phi3-mini-instruct-dpo-r-v1.5`
19
 
20
  In addition to the models, we are also releasing a series of datasets for multi-modal pre-training, including:
21
  - [MINT-1T: Scaling Open-Source Multimodal Data by 10x: A Multimodal Dataset with One Trillion Tokens](https://arxiv.org/abs/2406.11271)