Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Feature] adapt fused sigmoid gate for MoE model #2739

Open
2 tasks
zhyncs opened this issue Jan 5, 2025 · 2 comments
Open
2 tasks

[Feature] adapt fused sigmoid gate for MoE model #2739

zhyncs opened this issue Jan 5, 2025 · 2 comments
Assignees
Labels

Comments

@zhyncs
Copy link
Member

zhyncs commented Jan 5, 2025

Checklist

Motivation

ref https://github.com/NVIDIA/TensorRT-LLM/blob/be1788106245496872d18e702978e59b6bfd50e0/cpp/tensorrt_llm/kernels/mixtureOfExperts/moe_kernels.cu#L232

Related resources

No response

@zhaochenyang20
Copy link
Collaborator

@NovTi

@sitabulaixizawaluduo
Copy link

Checklist

Motivation

ref https://github.com/NVIDIA/TensorRT-LLM/blob/be1788106245496872d18e702978e59b6bfd50e0/cpp/tensorrt_llm/kernels/mixtureOfExperts/moe_kernels.cu#L232

Related resources

No response

LMDeploy should also do the same optimization, which should greatly improve the performance of MoE

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
Projects
None yet
Development

No branches or pull requests

3 participants