Skip to content

Releases: modelscope/ms-swift

v1.5.0

01 Jan 06:09
Compare
Choose a tag to compare

English Version

New features:

  1. Support multi-line inference
  2. Support multi node training
  3. Add benchmarks
  4. Support UI training, start by swift web-ui
  5. Support VLLM inference
  6. Support RLHF(DPO) training

New tuners:

  1. SCEdit, adopted by TongYi Lab, uses lesser memory but produces better performance than LoRA, and can be used to replace ControlNet in a series of scenarios like Pose control/In-Painting/Out-Paining/Label-removing, etc.

New models:

  1. SUS series models
  2. Mixtral-MoE series models
  3. deepseek series models
  4. phi2-3b
  5. cogagent-chat/cogagent-vqa
  6. codegeex2-6b

New datasets:

Datasets used in RLHF:

  1. hh-rlhf
  2. stack-exchange-paired

中文版

SWIFT本月新版本已经发布!

新特性:

  1. 支持多行输入推理
  2. 支持多卡训练
  3. 添加了模型训练的benchmarks
  4. 支持界面训练和推理,通过swift web-ui开启
  5. 支持VLLM推理
  6. 支持RLHF(DPO)训练

新tuners:

SCEdit: 通义实验室自研的优秀U-Net微调框架,显存占用远小于LoRA,效果较LoRA更好,且可以替代实现ControlNet的效果,实现In-Painting/Out-Paining/标签去除/Pose控制等能力

新模型:

SUS系列模型
Mixtral-MoE系列模型
deepseek系列模型
phi2-3b
cogagent-chat/cogagent-vqa
codegeex2-6b

新数据集:

用于RLHF的数据集:
hh-rlhf
stack-exchange-paired

What's Changed

Full Changelog: v1.4.0...v1.5.0

v1.4.0

08 Dec 03:09
Compare
Choose a tag to compare

English Version

New features:

  1. Support for self-awareness fine-tuning.
  2. Support for fine-tuning and inference of the AnimateDiff model in the AIGC direction.
  3. Support for flash attention in more models: qwen series, qwen-vl series, llama series, openbuddy series, mistral series, yi series, ziya series, using the use_flash_attn parameter.
  4. Support for multiple loras to take effect simultaneously.

New tuners:

  1. NEFTune
  2. ROME supports more models: chatglm

New models:

  1. AnimateDiff
  2. zephyr-7b-beta-chat, openbuddy-zephyr-7b-chat
  3. qwen-1_8b, qwen-1_8b-chat, qwen-1_8b-chat-int4, qwen-1_8b-chat-int8
  4. qwen-72b, qwen-72b-chat, qwen-72b-chat-int4, qwen-72b-chat-int8
  5. qwen-audio, qwen-audio-chat
  6. yi-34b-chat, codefuse-codellama-34b-chat
  7. tongyi-finance-14b, tongyi-finance-14b-chat, tongyi-finance-14b-chat-int4
  8. bluelm-7b, bluelm-7b-chat, bluelm-7b-32k, bluelm-7b-chat-32k

New datasets:

  1. hc3-zh, hc3-en
  2. codefuse-python-en, codefuse-eval-instruction-zh
  3. aishell1-zh, aishell1-mini-zh

中文版

新特性:

  1. 支持自我认知微调.
  2. 支持AIGC方向的AnimateDiff模型的微调与推理.
  3. 支持更多模型的flash attention: qwen series, qwen-vl series, llama series, openbuddy series, mistral series, yi series, ziya series. 使用use_flash_attn参数.
  4. 支持多个lora同时生效

新tuners:

  1. NEFTune
  2. ROME支持更多模型: chatglm

新模型:

  1. AnimateDiff
  2. zephyr-7b-beta-chat, openbuddy-zephyr-7b-chat
  3. qwen-1_8b, qwen-1_8b-chat, qwen-1_8b-chat-int4, qwen-1_8b-chat-int8
  4. qwen-72b, qwen-72b-chat, qwen-72b-chat-int4, qwen-72b-chat-int8
  5. qwen-audio, qwen-audio-chat
  6. yi-34b-chat, codefuse-codellama-34b-chat
  7. tongyi-finance-14b, tongyi-finance-14b-chat, tongyi-finance-14b-chat-int4
  8. bluelm-7b, bluelm-7b-chat, bluelm-7b-32k, bluelm-7b-chat-32k

新数据集:

  1. hc3-zh, hc3-en
  2. codefuse-python-en, codefuse-eval-instruction-zh
  3. aishell1-zh, aishell1-mini-zh

What's Changed

Full Changelog: v1.3.0...v1.4.0

v1.3.0 Release

09 Nov 03:55
Compare
Choose a tag to compare

English Version

New Features:

  1. Serving supported: LoRA and full-parameter training models are supported in vllm/chatglm.cpp/xinference deployment, check the documentation for details by make docs or docs/source/GetStarted/Deployment.md file.
  2. Support training and inference with CLI and inference with Web-UI.

New Adapters:

  • QALoRA
  • Long-LoRA
  • ROME

New Models:

  1. xverse-65b
  2. yi-6b
  3. ziya2-13b
  4. ziya2-13b-chat
  5. mistral-7b
  6. openbuddy-mistral-7b-chat
  7. mistral-7b-chat
  8. chatglm3-6b-base
  9. chatglm3-6b
  10. chatglm3-6b-32k

New Quantized Models:

  1. qwen-7b-chat-int4
  2. qwen-14b-chat-int4
  3. qwen-vl-chat-int4
  4. baichuan2-7b-chat-int4
  5. baichuan2-13b-chat-int4
  6. qwen-7b-chat-int8
  7. qwen-14b-chat-int8

中文版
新功能:

  1. 支持部署:全参数训练及LoRA训练支持以vllm/chatglm.cpp/xinference方式进行部署,可以通过make docs生成官方文档或查看docs/source/GetStarted/Deployment.md文件
  2. 支持CLI方式运行训练和推理,同时支持以Web-UI方式运行推理

新的Adapters:

  • QALoRA
  • Long-LoRA
  • ROME

支持训练和推理的新模型:

  1. xverse-65b
  2. yi-6b
  3. ziya2-13b
  4. ziya2-13b-chat
  5. mistral-7b
  6. openbuddy-mistral-7b-chat
  7. mistral-7b-chat
  8. chatglm3-6b-base
  9. chatglm3-6b
  10. chatglm3-6b-32k

支持训练和推理的新量化模型:

  1. qwen-7b-chat-int4
  2. qwen-14b-chat-int4
  3. qwen-vl-chat-int4
  4. baichuan2-7b-chat-int4
  5. baichuan2-13b-chat-int4
  6. qwen-7b-chat-int8
  7. qwen-14b-chat-int8

Feature Commits

Bug Fix:

New Contributors

  • @zzclynn made their first contribution in #105

Full Changelog: v1.2.0...v1.3.0

v1.1.1 release

22 Sep 06:06
Compare
Choose a tag to compare

Features:

  1. Add make docs command to build docs
  2. Add notebook examples for stable diffusion model
  3. Fix some bugs

v1.1.0

15 Sep 06:58
Compare
Choose a tag to compare