text-generation-inference/backends/llamacpp
2024-11-28 23:56:57 +01:00
..
cmake feat(backend): bind thread and memory affinity for thread 2024-11-21 13:52:38 +01:00
csrc feat(backend): add missing temperature parameter 2024-11-28 16:55:17 +01:00
offline misc(offline): update model creation as std::shared_ptr 2024-11-28 17:45:22 +01:00
src feat(backend): add missing temperature parameter 2024-11-28 16:55:17 +01:00
build.rs feat(backend): add some test to the backend for core allocation 2024-11-28 14:53:46 +01:00
Cargo.toml feat(backend): rely on multi consumer queue to scheduler workers 2024-11-22 13:32:56 +01:00
CMakeLists.txt feat(backend): update llama.cpp to 4215 2024-11-28 23:56:57 +01:00