Skip to content

GPU Memory Leak #5873

Closed
Closed
@zsogitbe

Description

@zsogitbe

After disposing everything the GPU memory is still not freed. When running the same code for the second time it crashes with a message that it cannot allocate enough memory. There are some things which are not freed from GPU memory.

You can test the issue by using ggml_backend_cuda_get_device_memory(0, out freemem, out totalmem); before and after using a llama model. freemem will show how much memory is still allocated.

See this also: SciSharp/LLamaSharp#575

Metadata

Metadata

Assignees

No one assigned

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions