minor : fix compile warnings

This commit is contained in:
Georgi Gerganov 2023-05-19 20:14:51 +03:00
parent 5ea4339273
commit 4b7e245adf
No known key found for this signature in database
GPG Key ID: 449E073F9DC10735
4 changed files with 12 additions and 10 deletions

View File

@ -749,7 +749,7 @@ bool console_readline(console_state & con_st, std::string & line) {
break; break;
} }
if (input_char == WEOF || input_char == 0x04 /* Ctrl+D*/) { if (input_char == (char32_t) WEOF || input_char == 0x04 /* Ctrl+D*/) {
end_of_stream = true; end_of_stream = true;
break; break;
} }
@ -764,7 +764,7 @@ bool console_readline(console_state & con_st, std::string & line) {
char32_t code = getchar32(); char32_t code = getchar32();
if (code == '[' || code == 0x1B) { if (code == '[' || code == 0x1B) {
// Discard the rest of the escape sequence // Discard the rest of the escape sequence
while ((code = getchar32()) != WEOF) { while ((code = getchar32()) != (char32_t) WEOF) {
if ((code >= 'A' && code <= 'Z') || (code >= 'a' && code <= 'z') || code == '~') { if ((code >= 'A' && code <= 'Z') || (code >= 'a' && code <= 'z') || code == '~') {
break; break;
} }

View File

@ -44,15 +44,15 @@ struct gpt_params {
float mirostat_tau = 5.00f; // target entropy float mirostat_tau = 5.00f; // target entropy
float mirostat_eta = 0.10f; // learning rate float mirostat_eta = 0.10f; // learning rate
std::string model = "models/7B/ggml-model.bin"; // model path std::string model = "models/7B/ggml-model.bin"; // model path
std::string prompt = ""; std::string prompt = "";
std::string path_prompt_cache = ""; // path to file for saving/loading prompt eval state std::string path_prompt_cache = ""; // path to file for saving/loading prompt eval state
std::string input_prefix = ""; // string to prefix user inputs with std::string input_prefix = ""; // string to prefix user inputs with
std::string input_suffix = ""; // string to suffix user inputs with std::string input_suffix = ""; // string to suffix user inputs with
std::vector<std::string> antiprompt; // string upon seeing which more user input is prompted std::vector<std::string> antiprompt; // string upon seeing which more user input is prompted
std::string lora_adapter = ""; // lora adapter path std::string lora_adapter = ""; // lora adapter path
std::string lora_base = ""; // base model path for the lora adapter std::string lora_base = ""; // base model path for the lora adapter
bool memory_f16 = true; // use f16 instead of f32 for memory kv bool memory_f16 = true; // use f16 instead of f32 for memory kv
bool random_prompt = false; // do not randomize prompt if none provided bool random_prompt = false; // do not randomize prompt if none provided

View File

@ -941,7 +941,7 @@ static void llama_model_load_internal(
size_t ctx_size; size_t ctx_size;
size_t mmapped_size; size_t mmapped_size;
ml->calc_sizes(&ctx_size, &mmapped_size); ml->calc_sizes(&ctx_size, &mmapped_size);
fprintf(stderr, "%s: ggml ctx size = %6.2f KB\n", __func__, ctx_size/1024.0); fprintf(stderr, "%s: ggml ctx size = %6.2f MB\n", __func__, ctx_size/1024.0/1024.0);
// print memory requirements // print memory requirements
{ {

View File

@ -1,14 +1,16 @@
#include "llama.h"
#include "ggml.h" #include "ggml.h"
#include <cassert> #include "llama.h"
#include <cmath>
#ifdef NDEBUG
#undef NDEBUG
#endif
#include <numeric> #include <numeric>
#include <cassert> #include <cassert>
#include <iostream> #include <iostream>
#include <vector> #include <vector>
#include <algorithm> #include <algorithm>
void dump(const llama_token_data_array * candidates) { void dump(const llama_token_data_array * candidates) {
for (size_t i = 0; i < candidates->size; i++) { for (size_t i = 0; i < candidates->size; i++) {
printf("%d: %f (%f)\n", candidates->data[i].id, candidates->data[i].p, candidates->data[i].logit); printf("%d: %f (%f)\n", candidates->data[i].id, candidates->data[i].p, candidates->data[i].logit);