Skip to content

Conversation

@lizhenyun01
Copy link
Collaborator

@lizhenyun01 lizhenyun01 commented Dec 25, 2025

Motivation

  • attention优化及重构第一部分:
    • attention重构,合并投机解码/非投机解码分支,消除冗余逻辑
    • 拆分decoder_write_cache_with_rope为单独算子,便于维护
    • 新增decode attention backend,当前只支持PD分离下D节点
    • 优化decode attention C8kernel性能,优化后在group_size=14下单步投机场景性能提升5%-113%
image


  • TODO:
    • ROPE,write_cache重构及投机解码等分支融合
    • C16 C4支持
    • 单测完善
    • backend逐步替换append_attention

Checklist

  • Add at least a tag in the PR title.
    • Tag list: [[FDConfig],[APIServer],[Engine], [Scheduler], [PD Disaggregation], [Executor], [Graph Optimization], [Speculative Decoding], [RL], [Models], [Quantization], [Loader], [OP], [KVCache], [DataProcessor], [BugFix], [Docs], [CI], [Optimization], [Feature], [Benchmark], [Others], [XPU], [HPU], [GCU], [DCU], [Iluvatar], [Metax]]
    • You can add new tags based on the PR content, but the semantics must be clear.
  • Format your code, run pre-commit before commit.
  • Add unit tests. Please write the reason in this PR if no unit tests.
  • Provide accuracy results.
  • If the current PR is submitting to the release branch, make sure the PR has been submitted to the develop branch, then cherry-pick it to the release branch with the [Cherry-Pick] PR tag.

@paddle-bot
Copy link

paddle-bot bot commented Dec 25, 2025

Thanks for your contribution!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant