mirror of
https://github.com/ggerganov/llama.cpp.git
synced 2024-12-25 05:48:47 +01:00
fc8ef549e5
* metal: enable ggml-alloc Make ggml-alloc work with concurrently dispatch. * style-fix Co-authored-by: slaren <slarengh@gmail.com> --------- Co-authored-by: slaren <slarengh@gmail.com> Co-authored-by: Georgi Gerganov <ggerganov@gmail.com>
27 lines
899 B
C
27 lines
899 B
C
#pragma once
|
|
|
|
#include "ggml.h"
|
|
|
|
#ifdef __cplusplus
|
|
extern "C" {
|
|
#endif
|
|
|
|
|
|
GGML_API struct ggml_allocr * ggml_allocr_new(void * data, size_t size, size_t alignment);
|
|
GGML_API struct ggml_allocr * ggml_allocr_new_measure(size_t alignment);
|
|
|
|
// tell the allocator to parse nodes following the order described in the list
|
|
// you should call this if your graph are optimized to execute out-of-order
|
|
GGML_API void ggml_allocr_set_parse_seq(struct ggml_allocr * alloc, int * list, int n);
|
|
|
|
GGML_API void ggml_allocr_free(struct ggml_allocr * alloc);
|
|
GGML_API bool ggml_allocr_is_measure(struct ggml_allocr * alloc);
|
|
GGML_API void ggml_allocr_reset(struct ggml_allocr * alloc);
|
|
GGML_API void ggml_allocr_alloc(struct ggml_allocr * alloc, struct ggml_tensor * tensor);
|
|
GGML_API size_t ggml_allocr_alloc_graph(struct ggml_allocr * alloc, struct ggml_cgraph * graph);
|
|
|
|
|
|
#ifdef __cplusplus
|
|
}
|
|
#endif
|