mirror of
https://github.com/ollama/ollama.git
synced 2026-04-27 19:25:55 +02:00
Revert "Update vendored llama.cpp to b7847" (#14061)
This commit is contained in:
4
llama/llama.cpp/src/llama-adapter.h
vendored
4
llama/llama.cpp/src/llama-adapter.h
vendored
@@ -77,10 +77,6 @@ struct llama_adapter_lora {
|
||||
~llama_adapter_lora() = default;
|
||||
|
||||
llama_adapter_lora_weight * get_weight(ggml_tensor * w);
|
||||
|
||||
uint32_t get_n_nodes() const {
|
||||
return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat
|
||||
}
|
||||
};
|
||||
|
||||
using llama_adapter_loras = std::unordered_map<llama_adapter_lora *, float>;
|
||||
|
||||
Reference in New Issue
Block a user