Skip to content

v2.2.1

Compare
Choose a tag to compare
@tastelikefeet tastelikefeet released this 08 Jul 07:08
· 82 commits to main since this release

English Version

New Features

  1. Multimodal: Supported a large number of multimodal datasets and restructured the multimodal architecture. Some models now support grounding tasks.
  2. Web-ui: Added support for RLHF, evaluation, and quantization.
  3. Evaluation Functionality: Refactored the evaluation functionality, now using OpenCompass internally, supporting over 50 evaluation datasets.
  4. Deployment Functionality: VLLM infer_backend now supports multimodal models.
  5. Agent Training: Refactored the construction, training, and deployment of agent datasets, making the agent pipeline more complete.
  6. Human Alignment: Added alignment algorithms such as KTO and CPO, and refactored the human alignment code.

New Models

  1. openbuddy-llama3-70b
  2. Deepseek-coder-v2
  3. llava1.5, llava1.6, llava-next-video
  4. gemma2
  5. Florence
  6. phi3-4k
  7. internlm2.5, xcomposer2.5
  8. internvl2
  9. codegeex4
  10. mistral-7b-instruct-v0.3

New Datasets

  1. Over 30 foundational multimodal datasets, including GQA, RefCOCO, and Llava-pretrain.
  2. Swift-mix general mixed dataset.
  3. Video-chatgpt video dataset.

中文版本

新功能

  1. 多模态:支持了非常多的多模态数据集,并重构了多模态架构,部分模型开始支持grounding任务
  2. Web-ui:支持了RLHF、评测和量化
  3. 评测功能:进行了重构,内部使用了OpenCompass,支持50+评测集
  4. 部署功能:VLLM infer_backend支持多模态模型
  5. Agent训练:重构了Agent数据集构造、训练、部署,Agent链路更加完整
  6. 人类对齐:增加了KTO、CPO等对齐算法,并重构了人类对齐的代码

新模型

  1. openbuddy-llama3-70b
  2. Deepseek-coder-v2
  3. llava1.5, llava1.6, llava-next-video
  4. gemma2
  5. Florence
  6. phi3-4k
  7. internlm2.5, xcomposer2.5
  8. internvl2
  9. codegeex4
  10. mistral-7b-instruct-v0.3

新数据集

  1. GQA、RefCOCO、Llava-pretrain等30+多模态基础数据集
  2. swift-mix通用混合数据集
  3. video-chatgpt视频数据集

What's Changed

New Contributors

Full Changelog: v2.1.1...v2.2.1