#include "common.h" #include "llama.h" #include #include #include #include #include #include #include #include #include #if defined (__unix__) || (defined (__APPLE__) && defined (__MACH__)) #include #include #elif defined (_WIN32) #include #endif #if defined (_WIN32) #pragma comment(lib,"kernel32.lib") extern "C" __declspec(dllimport) void* __stdcall GetStdHandle(unsigned long nStdHandle); extern "C" __declspec(dllimport) int __stdcall GetConsoleMode(void* hConsoleHandle, unsigned long* lpMode); extern "C" __declspec(dllimport) int __stdcall SetConsoleMode(void* hConsoleHandle, unsigned long dwMode); #endif #define ANSI_COLOR_RED "\x1b[31m" #define ANSI_COLOR_GREEN "\x1b[32m" #define ANSI_COLOR_YELLOW "\x1b[33m" #define ANSI_COLOR_BLUE "\x1b[34m" #define ANSI_COLOR_MAGENTA "\x1b[35m" #define ANSI_COLOR_CYAN "\x1b[36m" #define ANSI_COLOR_RESET "\x1b[0m" #define ANSI_BOLD "\x1b[1m" /* Keep track of current color of output, and emit ANSI code if it changes. */ enum console_state { CONSOLE_STATE_DEFAULT=0, CONSOLE_STATE_PROMPT, CONSOLE_STATE_USER_INPUT }; static console_state con_st = CONSOLE_STATE_DEFAULT; static bool con_use_color = false; void set_console_state(console_state new_st) { if (!con_use_color) return; // only emit color code if state changed if (new_st != con_st) { con_st = new_st; switch(con_st) { case CONSOLE_STATE_DEFAULT: printf(ANSI_COLOR_RESET); return; case CONSOLE_STATE_PROMPT: printf(ANSI_COLOR_YELLOW); return; case CONSOLE_STATE_USER_INPUT: printf(ANSI_BOLD ANSI_COLOR_GREEN); return; } } } static bool is_interacting = false; #if defined (__unix__) || (defined (__APPLE__) && defined (__MACH__)) || defined (_WIN32) void sigint_handler(int signo) { set_console_state(CONSOLE_STATE_DEFAULT); printf("\n"); // this also force flush stdout. if (signo == SIGINT) { if (!is_interacting) { is_interacting=true; } else { _exit(130); } } } #endif int main(int argc, char ** argv) { gpt_params params; params.model = "models/llama-7B/ggml-model.bin"; if (gpt_params_parse(argc, argv, params) == false) { return 1; } if (params.perplexity) { printf("\n************\n"); printf("%s: please use the 'perplexity' tool for perplexity calculations\n", __func__); printf("************\n\n"); return 0; } if (params.n_ctx > 2048) { fprintf(stderr, "%s: warning: model does not support context sizes greater than 2048 tokens (%d specified);" "expect poor results\n", __func__, params.n_ctx); } if (params.seed <= 0) { params.seed = time(NULL); } fprintf(stderr, "%s: seed = %d\n", __func__, params.seed); std::mt19937 rng(params.seed); if (params.random_prompt) { params.prompt = gpt_random_prompt(rng); } // save choice to use color for later // (note for later: this is a slightly awkward choice) con_use_color = params.use_color; // params.prompt = R"(// this function checks if the number n is prime //bool is_prime(int n) {)"; llama_context * ctx; // load the model { auto lparams = llama_context_default_params(); lparams.n_ctx = params.n_ctx; lparams.n_parts = params.n_parts; lparams.seed = params.seed; lparams.f16_kv = params.memory_f16; lparams.use_mlock = params.use_mlock; ctx = llama_init_from_file(params.model.c_str(), lparams); if (ctx == NULL) { fprintf(stderr, "%s: error: failed to load model '%s'\n", __func__, params.model.c_str()); return 1; } } // print system information { fprintf(stderr, "\n"); fprintf(stderr, "system_info: n_threads = %d / %d | %s\n", params.n_threads, std::thread::hardware_concurrency(), llama_print_system_info()); } // determine the maximum memory usage needed to do inference for the given n_batch and n_predict parameters // uncomment the "used_mem" line in llama.cpp to see the results if (params.mem_test) { { const std::vector tmp(params.n_batch, 0); llama_eval(ctx, tmp.data(), tmp.size(), 0, params.n_threads); } { const std::vector tmp = { 0, }; llama_eval(ctx, tmp.data(), tmp.size(), params.n_predict - 1, params.n_threads); } llama_print_timings(ctx); llama_free(ctx); return 0; } int n_past = 0; // Add a space in front of the first character to match OG llama tokenizer behavior params.prompt.insert(0, 1, ' '); // tokenize the prompt auto embd_inp = ::llama_tokenize(ctx, params.prompt, true); const int n_ctx = llama_n_ctx(ctx); params.n_predict = std::min(params.n_predict, n_ctx - (int) embd_inp.size()); // prefix & suffix for instruct mode const auto inp_pfx = ::llama_tokenize(ctx, "\n\n### Instruction:\n\n", true); const auto inp_sfx = ::llama_tokenize(ctx, "\n\n### Response:\n\n", false); // in instruct mode, we inject a prefix and a suffix to each input by the user if (params.instruct) { params.interactive = true; params.antiprompt.push_back("### Instruction:\n\n"); } // enable interactive mode if reverse prompt is specified if (params.antiprompt.size() != 0) { params.interactive = true; } if (params.interactive_start) { params.interactive = true; } // determine newline token auto llama_token_newline = ::llama_tokenize(ctx, "\n", false); if (params.verbose_prompt) { fprintf(stderr, "\n"); fprintf(stderr, "%s: prompt: '%s'\n", __func__, params.prompt.c_str()); fprintf(stderr, "%s: number of tokens in prompt = %zu\n", __func__, embd_inp.size()); for (int i = 0; i < (int) embd_inp.size(); i++) { fprintf(stderr, "%6d -> '%s'\n", embd_inp[i], llama_token_to_str(ctx, embd_inp[i])); } fprintf(stderr, "\n"); } if (params.interactive) { #if defined (__unix__) || (defined (__APPLE__) && defined (__MACH__)) struct sigaction sigint_action; sigint_action.sa_handler = sigint_handler; sigemptyset (&sigint_action.sa_mask); sigint_action.sa_flags = 0; sigaction(SIGINT, &sigint_action, NULL); #elif defined (_WIN32) signal(SIGINT, sigint_handler); #endif fprintf(stderr, "%s: interactive mode on.\n", __func__); if(params.antiprompt.size()) { for (auto antiprompt : params.antiprompt) { fprintf(stderr, "Reverse prompt: '%s'\n", antiprompt.c_str()); } } if (!params.input_prefix.empty()) { fprintf(stderr, "Input prefix: '%s'\n", params.input_prefix.c_str()); } } fprintf(stderr, "sampling parameters: temp = %f, top_k = %d, top_p = %f, repeat_last_n = %i, repeat_penalty = %f\n", params.temp, params.top_k, params.top_p, params.repeat_last_n, params.repeat_penalty); fprintf(stderr, "\n\n"); std::vector embd; int last_n_size = params.repeat_last_n; std::vector last_n_tokens(last_n_size); std::fill(last_n_tokens.begin(), last_n_tokens.end(), 0); if (params.interactive) { fprintf(stderr, "== Running in interactive mode. ==\n" #if defined (__unix__) || (defined (__APPLE__) && defined (__MACH__)) || defined (_WIN32) " - Press Ctrl+C to interject at any time.\n" #endif " - Press Return to return control to LLaMa.\n" " - If you want to submit another line, end your input in '\\'.\n\n"); is_interacting = params.interactive_start || params.instruct; } int input_consumed = 0; bool input_noecho = false; int remaining_tokens = params.n_predict; #if defined (_WIN32) if (params.use_color) { // Enable ANSI colors on Windows 10+ unsigned long dwMode = 0; void* hConOut = GetStdHandle((unsigned long)-11); // STD_OUTPUT_HANDLE (-11) if (hConOut && hConOut != (void*)-1 && GetConsoleMode(hConOut, &dwMode) && !(dwMode & 0x4)) { SetConsoleMode(hConOut, dwMode | 0x4); // ENABLE_VIRTUAL_TERMINAL_PROCESSING (0x4) } } #endif // the first thing we will do is to output the prompt, so set color accordingly set_console_state(CONSOLE_STATE_PROMPT); while (remaining_tokens > 0 || params.interactive) { // predict if (embd.size() > 0) { if (llama_eval(ctx, embd.data(), embd.size(), n_past, params.n_threads)) { fprintf(stderr, "%s : failed to eval\n", __func__); return 1; } } n_past += embd.size(); embd.clear(); if ((int) embd_inp.size() <= input_consumed && !is_interacting) { // out of user input, sample next token const float top_k = params.top_k; const float top_p = params.top_p; const float temp = params.temp; const float repeat_penalty = params.repeat_penalty; llama_token id = 0; { auto logits = llama_get_logits(ctx); if (params.ignore_eos) { logits[llama_token_eos()] = 0; } id = llama_sample_top_p_top_k(ctx, last_n_tokens.data(), last_n_tokens.size(), top_k, top_p, temp, repeat_penalty); last_n_tokens.erase(last_n_tokens.begin()); last_n_tokens.push_back(id); } // replace end of text token with newline token when in interactive mode if (id == llama_token_eos() && params.interactive && !params.instruct) { id = llama_token_newline.front(); if (params.antiprompt.size() != 0) { // tokenize and inject first reverse prompt const auto first_antiprompt = ::llama_tokenize(ctx, params.antiprompt.front(), false); embd_inp.insert(embd_inp.end(), first_antiprompt.begin(), first_antiprompt.end()); } } // add it to the context embd.push_back(id); // echo this to console input_noecho = false; // decrement remaining sampling budget --remaining_tokens; } else { // some user input remains from prompt or interaction, forward it to processing while ((int) embd_inp.size() > input_consumed) { embd.push_back(embd_inp[input_consumed]); last_n_tokens.erase(last_n_tokens.begin()); last_n_tokens.push_back(embd_inp[input_consumed]); ++input_consumed; if ((int) embd.size() >= params.n_batch) { break; } } } // display text if (!input_noecho) { for (auto id : embd) { printf("%s", llama_token_to_str(ctx, id)); } fflush(stdout); } // reset color to default if we there is no pending user input if (!input_noecho && (int)embd_inp.size() == input_consumed) { set_console_state(CONSOLE_STATE_DEFAULT); } // in interactive mode, and not currently processing queued inputs; // check if we should prompt the user for more if (params.interactive && (int) embd_inp.size() <= input_consumed) { // check for reverse prompt std::string last_output; for (auto id : last_n_tokens) { last_output += llama_token_to_str(ctx, id); } // Check if each of the reverse prompts appears at the end of the output. for (std::string & antiprompt : params.antiprompt) { if (last_output.find(antiprompt.c_str(), last_output.length() - antiprompt.length(), antiprompt.length()) != std::string::npos) { is_interacting = true; set_console_state(CONSOLE_STATE_USER_INPUT); fflush(stdout); break; } } if (n_past > 0 && is_interacting) { // potentially set color to indicate we are taking user input set_console_state(CONSOLE_STATE_USER_INPUT); if (params.instruct) { input_consumed = embd_inp.size(); embd_inp.insert(embd_inp.end(), inp_pfx.begin(), inp_pfx.end()); printf("\n> "); } std::string buffer; if (!params.input_prefix.empty()) { buffer += params.input_prefix; printf("%s", buffer.c_str()); } std::string line; bool another_line = true; do { std::getline(std::cin, line); if (line.empty() || line.back() != '\\') { another_line = false; } else { line.pop_back(); // Remove the continue character } buffer += line + '\n'; // Append the line to the result } while (another_line); // done taking input, reset color set_console_state(CONSOLE_STATE_DEFAULT); auto line_inp = ::llama_tokenize(ctx, buffer, false); embd_inp.insert(embd_inp.end(), line_inp.begin(), line_inp.end()); if (params.instruct) { embd_inp.insert(embd_inp.end(), inp_sfx.begin(), inp_sfx.end()); } remaining_tokens -= line_inp.size(); input_noecho = true; // do not echo this again } if (n_past > 0) { is_interacting = false; } } // end of text token if (embd.back() == llama_token_eos()) { if (params.instruct) { is_interacting = true; } else { fprintf(stderr, " [end of text]\n"); break; } } // In interactive mode, respect the maximum number of tokens and drop back to user input when reached. if (params.interactive && remaining_tokens <= 0) { remaining_tokens = params.n_predict; is_interacting = true; } } #if defined (_WIN32) signal(SIGINT, SIG_DFL); #endif llama_print_timings(ctx); llama_free(ctx); set_console_state(CONSOLE_STATE_DEFAULT); return 0; }