2024-10-04 08:42:31 +00:00
|
|
|
//
|
|
|
|
// Created by mfuntowicz on 10/3/24.
|
|
|
|
//
|
|
|
|
|
2024-10-22 22:09:10 +00:00
|
|
|
#include <fmt/color.h>
|
2024-10-04 08:42:31 +00:00
|
|
|
#include <fmt/format.h>
|
2024-10-22 13:22:56 +00:00
|
|
|
#include <fmt/std.h>
|
2024-10-22 22:09:10 +00:00
|
|
|
#include <fmt/ranges.h>
|
2024-10-04 08:42:31 +00:00
|
|
|
#include <spdlog/spdlog.h>
|
|
|
|
#include "../csrc/backend.hpp"
|
|
|
|
|
2024-10-24 14:42:50 +00:00
|
|
|
using namespace huggingface::tgi::backends::llamacpp;
|
|
|
|
|
2024-10-26 20:24:05 +00:00
|
|
|
int main(int argc, char **argv) {
|
2024-10-22 22:09:10 +00:00
|
|
|
if (argc < 2) {
|
2024-10-04 08:42:31 +00:00
|
|
|
fmt::print("No model folder provider");
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
|
|
|
|
spdlog::set_level(spdlog::level::debug);
|
|
|
|
|
2024-10-22 22:09:10 +00:00
|
|
|
const auto prompt = "My name is Morgan";
|
|
|
|
|
2024-10-22 13:22:56 +00:00
|
|
|
const auto modelPath = absolute(std::filesystem::path(argv[1]));
|
2024-10-30 21:40:49 +00:00
|
|
|
const auto params = llama_model_default_params();
|
|
|
|
auto *model = llama_load_model_from_file(modelPath.c_str(), params);
|
|
|
|
|
|
|
|
auto backend = single_worker_backend_t(model, {});
|
|
|
|
|
|
|
|
// generate
|
|
|
|
const auto promptTokens = {128000, 9906, 856, 836, 374, 23809, 128001};
|
|
|
|
const auto out = backend.generate(promptTokens, {.max_new_tokens = 32}, {.top_k = 40});
|
|
|
|
|
|
|
|
if (out.has_value())
|
|
|
|
fmt::print(FMT_STRING("Generated: {}"), *out);
|
|
|
|
else {
|
|
|
|
const auto err = out.error();
|
|
|
|
fmt::print(fmt::emphasis::bold | fg(fmt::color::red), "Got an error: {:d}", static_cast<uint8_t>(err));
|
2024-10-22 22:09:10 +00:00
|
|
|
}
|
|
|
|
}
|