Skip to content

Add Qwen 2.5 using llama_transformer.py #8471

@jackzhxng

Description

@jackzhxng

🚀 The feature, motivation and pitch

Implement Qwen 2.5 by fitting into llama_transformer.py and export_llama, instead of manual tracing and lowering. Part of an exploration into how we can more quickly add new transformer decoder-based models.

Alternatives

No response

Additional context

No response

RFC (Optional)

No response

cc @mergennachin @iseeyuan @lucylq @helunwencser @tarun292 @kimishpatel @cccclai

Metadata

Metadata

Assignees

Labels

module: examplesIssues related to demos under examples/module: llmIssues related to LLM examples and apps, and to the extensions/llm/ codetriagedThis issue has been looked at a team member, and triaged and prioritized into an appropriate module

Projects

Status

Done

Milestone

No milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions