Files
llama.cpp/ggml-cuda.h
slaren 295f85654a allocators wip
renamed ggml_backend functions
changed ggml_buffer and ggml_backend to always be used as pointers
rename ggml_tensor::params -> op_params
2023-07-19 02:43:44 +02:00

401 B