Skip to content

Conversation

@jxt1234
Copy link
Collaborator

@jxt1234 jxt1234 commented Jul 23, 2025

  1. CPU 增加 SME 支持(FP32 / FP16 / Int8)
  2. LSTM / GRU 算子支持量化(通过转换成 while + Convolution 来实现)
  3. LLM 增加 HQQ 量化方案
  4. LLM 增加 MTP 解码加速支持
  5. 支持 QNN 运行 LLM 的 Vision 部分
  6. CUDA 后端支持 LLM 运行及 int4 / int8 反量化计算

@CLAassistant
Copy link

CLA assistant check
Thank you for your submission! We really appreciate it. Like many open source projects, we ask that you sign our Contributor License Agreement before we can accept your contribution.


xiaying seems not to be a GitHub user. You need a GitHub account to be able to sign the CLA. If you have already a GitHub account, please add the email address used for this commit to your account.
You have signed the CLA already but the status is still pending? Let us recheck it.

@jxt1234 jxt1234 merged commit a739ea5 into master Jul 23, 2025
23 of 24 checks passed
@jxt1234 jxt1234 deleted the feature/sync branch July 23, 2025 07:45
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants