Skip to content

cuda : fix rope pos data #7452

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Merged
merged 4 commits into from
May 22, 2024
Merged

cuda : fix rope pos data #7452

merged 4 commits into from
May 22, 2024

Conversation

ggerganov
Copy link
Member

ref #7225
fix #7451

Accidentally broke CUDA rope in #7225 for all models that do not use NeoX-style RoPE

@github-actions github-actions bot added Nvidia GPU Issues specific to Nvidia GPUs ggml changes relating to the ggml tensor library for machine learning labels May 22, 2024
@ggerganov ggerganov merged commit 3e5faa8 into master May 22, 2024
24 of 79 checks passed
@ggerganov ggerganov deleted the gg/cuda-fix-rope-pos branch May 22, 2024 08:01
@github-actions github-actions bot added the testing Everything test related label May 22, 2024
teleprint-me pushed a commit to teleprint-me/llama.cpp that referenced this pull request May 23, 2024
* cuda : fix rope pos data

ggml-ci

* ggml : drop mode & 1 == 1 support for ggml_rope

ggml-ci

* ggml : support freq_factors for f16 rope (CPU)

ggml-ci

* tests : add rope tests using frequency factors

ggml-ci
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
ggml changes relating to the ggml tensor library for machine learning Nvidia GPU Issues specific to Nvidia GPUs testing Everything test related
Projects
None yet
Development

Successfully merging this pull request may close these issues.

regression: output is nonsense with latest commit and CUDA support enabled
2 participants