}
void common_init() {
+#if defined(_WIN32)
+ SetConsoleOutputCP(CP_UTF8);
+ SetConsoleCP(CP_UTF8);
+#endif
+
llama_log_set(common_log_default_callback, NULL);
#ifdef NDEBUG
const char * build_type = " (debug)";
#endif
- LOG_INF("build: %d (%s) with %s for %s%s\n", LLAMA_BUILD_NUMBER, LLAMA_COMMIT, LLAMA_COMPILER, LLAMA_BUILD_TARGET, build_type);
+ LOG_DBG("build: %d (%s) with %s for %s%s\n", LLAMA_BUILD_NUMBER, LLAMA_COMMIT, LLAMA_COMPILER, LLAMA_BUILD_TARGET, build_type);
}
std::string common_params_get_system_info(const common_params & params) {
params.prompt = "Hello my name is";
params.n_predict = 32;
+ common_init();
+
if (!common_params_parse(argc, argv, params, LLAMA_EXAMPLE_BATCHED, print_usage)) {
return 1;
}
- common_init();
-
// number of parallel batches
int n_parallel = params.n_parallel;
int main(int argc, char ** argv) {
common_params params;
+ common_init();
+
if (!common_params_parse(argc, argv, params, LLAMA_EXAMPLE_DEBUG, print_usage)) {
return 1;
}
- common_init();
-
llama_backend_init();
llama_numa_init(params.numa);
common_params params;
+ common_init();
+
if (!common_params_parse(argc, argv, params, LLAMA_EXAMPLE_DIFFUSION)) {
return 1;
}
- common_init();
llama_backend_init();
llama_model_params model_params = llama_model_default_params();
common_params params;
+ common_init();
+
if (!common_params_parse(argc, argv, params, LLAMA_EXAMPLE_EMBEDDING)) {
return 1;
}
- common_init();
-
params.embedding = true;
// get max number of sequences per batch
common_params params;
+ common_init();
+
if (!common_params_parse(argc, argv, params, LLAMA_EXAMPLE_COMMON)) {
return 1;
}
- common_init();
-
llama_backend_init();
llama_numa_init(params.numa);
int main(int argc, char ** argv) {
common_params params;
+ common_init();
+
if (!common_params_parse(argc, argv, params, LLAMA_EXAMPLE_COMMON, print_usage)) {
return 1;
}
- common_init();
-
// init LLM
llama_backend_init();
common_params params;
+ common_init();
+
if (!common_params_parse(argc, argv, params, LLAMA_EXAMPLE_COMMON)) {
return 1;
}
- common_init();
-
const int W = 15; // lookahead window
const int N = 5; // n-gram size
const int G = 15; // max verification n-grams
common_params params;
+ common_init();
+
if (!common_params_parse(argc, argv, params, LLAMA_EXAMPLE_LOOKUP)) {
return 1;
}
common_params params;
+ common_init();
+
if (!common_params_parse(argc, argv, params, LLAMA_EXAMPLE_LOOKUP)) {
return 1;
}
- common_init();
-
const int n_draft = params.speculative.n_max;
// init llama.cpp
common_params params;
+ common_init();
+
if (!common_params_parse(argc, argv, params, LLAMA_EXAMPLE_LOOKUP)) {
return 1;
}
- common_init();
-
// max. number of additional tokens to draft if match is found
const int n_draft = params.speculative.n_max;
params.n_predict = 128;
params.n_junk = 1;
+ common_init();
+
if (!common_params_parse(argc, argv, params, LLAMA_EXAMPLE_PARALLEL)) {
return 1;
}
- common_init();
-
// number of simultaneous "clients" to simulate
const int32_t n_clients = params.n_parallel;
params.n_keep = 32;
params.i_pos = -1;
+ common_init();
+
if (!common_params_parse(argc, argv, params, LLAMA_EXAMPLE_PASSKEY, print_usage)) {
return 1;
}
- common_init();
-
int n_junk = params.n_junk;
int n_keep = params.n_keep;
int n_grp = params.grp_attn_n;
common_params params;
+ common_init();
+
if (!common_params_parse(argc, argv, params, LLAMA_EXAMPLE_RETRIEVAL, print_usage)) {
return 1;
}
- common_init();
-
// For BERT models, batch size must be equal to ubatch size
params.n_ubatch = params.n_batch;
params.embedding = true;
const std::string_view state_file = "dump_state.bin";
+ common_init();
+
if (!common_params_parse(argc, argv, params, LLAMA_EXAMPLE_COMMON)) {
return 1;
}
params.kv_unified = true;
}
- common_init();
-
if (params.n_predict < 0) {
params.n_predict = 16;
}
common_params params;
+ common_init();
+
if (!common_params_parse(argc, argv, params, LLAMA_EXAMPLE_SPECULATIVE)) {
return 1;
}
return 1;
}
- common_init();
-
if (params.speculative.mparams_dft.path.empty()) {
LOG_ERR("%s: --model-draft is required\n", __func__);
return 1;
// needed to get candidate probs even for temp <= 0.0
params.sampling.n_probs = 128;
+ common_init();
+
if (!common_params_parse(argc, argv, params, LLAMA_EXAMPLE_SPECULATIVE)) {
return 1;
}
return 1;
}
- common_init();
-
if (params.speculative.mparams_dft.path.empty()) {
LOG_ERR("%s: --model-draft is required\n", __func__);
return 1;
common_params params;
params.escape = false;
+ common_init();
+
if (!common_params_parse(argc, argv, params, LLAMA_EXAMPLE_FINETUNE)) {
return 1;
}
params.cache_type_v = GGML_TYPE_F32;
}
- common_init();
llama_backend_init();
llama_numa_init(params.numa);
// load the model and apply lora adapter, if any
common_params params;
params.out_file = "tests.txt";
+ common_init();
+
if (!common_params_parse(argc, argv, params, LLAMA_EXAMPLE_EXPORT_GRAPH_OPS)) {
return 1;
}
- common_init();
-
// Load CPU-only
ggml_backend_dev_t cpu_device = ggml_backend_dev_by_type(GGML_BACKEND_DEVICE_TYPE_CPU);
params.devices = { cpu_device, nullptr };
params.n_parallel = 3;
params.n_ctx = 256;
+ common_init();
+
if (!common_params_parse(argc, argv, params, LLAMA_EXAMPLE_COMMON)) {
return 1;
}
- common_init();
-
// init
common_init_result_ptr llama_init = common_init_from_params(params);
int main(int argc, char ** argv) {
common_params params;
+ common_init();
+
if (!common_params_parse(argc, argv, params, LLAMA_EXAMPLE_COMMON)) {
return 1;
}
- common_init();
-
llama_backend_init();
llama_numa_init(params.numa);
common_params params;
+ common_init();
+
if (!common_params_parse(argc, argv, params, LLAMA_EXAMPLE_BENCH, print_usage)) {
return 1;
}
- common_init();
-
int is_pp_shared = params.is_pp_shared;
int is_tg_separate = params.is_tg_separate;
params.verbosity = LOG_LEVEL_ERROR; // by default, less verbose logs
+ common_init();
+
if (!common_params_parse(argc, argv, params, LLAMA_EXAMPLE_CLI)) {
return 1;
}
console::error("please use llama-completion instead\n");
}
- common_init();
-
// struct that contains llama context and inference
cli_context ctx_cli(params);
common_params params;
g_params = ¶ms;
+ common_init();
+
if (!common_params_parse(argc, argv, params, LLAMA_EXAMPLE_COMPLETION, print_usage)) {
return 1;
}
- common_init();
-
auto & sparams = params.sampling;
// save choice to use color for later
params.out_file = "control_vector.gguf";
+ common_init();
+
if (!common_params_parse(argc, argv, params, LLAMA_EXAMPLE_CVECTOR_GENERATOR, print_usage)) {
return 1;
}
params.out_file = "ggml-lora-merged-f16.gguf";
+ common_init();
+
if (!common_params_parse(argc, argv, params, LLAMA_EXAMPLE_EXPORT_LORA, print_usage)) {
return 1;
}
int main(int argc, char ** argv) {
common_params params;
+ common_init();
+
if (!common_params_parse(argc, argv, params, LLAMA_EXAMPLE_COMMON)) {
return 1;
}
- common_init();
llama_backend_init();
llama_numa_init(params.numa);
auto mparams = common_model_params_to_llama(params);
params.n_ctx = 512;
params.escape = false;
+ common_init();
+
if (!common_params_parse(argc, argv, params, LLAMA_EXAMPLE_IMATRIX, print_usage)) {
return 1;
}
return 0;
}
- common_init();
-
const int32_t n_ctx = params.n_ctx;
if (n_ctx <= 0) {
common_params params;
+ common_init();
+
if (!common_params_parse(argc, argv, params, LLAMA_EXAMPLE_MTMD, show_additional_info)) {
return 1;
}
- common_init();
mtmd_helper_log_set(common_log_default_callback, nullptr);
if (params.mmproj.path.empty()) {
common_params params;
+ common_init();
+
if (!common_params_parse(argc, argv, params, LLAMA_EXAMPLE_MTMD, show_additional_info)) {
return 1;
}
- common_init();
mtmd_helper_log_set(common_log_default_callback, nullptr);
if (params.mmproj.path.empty()) {
params.n_ctx = 512;
params.escape = false;
+ common_init();
+
if (!common_params_parse(argc, argv, params, LLAMA_EXAMPLE_PERPLEXITY)) {
return 1;
}
- common_init();
-
const int32_t n_ctx = params.n_ctx;
if (n_ctx <= 0) {
int main(int argc, char ** argv) {
common_params params;
params.escape = false;
+
+ common_init();
+
if (!common_params_parse(argc, argv, params, LLAMA_EXAMPLE_RESULTS)) {
return 1;
}
LOG_ERR("%s: an output file must be specified", __func__);
return 1;
}
- common_init();
llama_backend_init();
llama_numa_init(params.numa);
common_init_result_ptr llama_init = common_init_from_params(params);
// own arguments required by this example
common_params params;
+ common_init();
+
if (!common_params_parse(argc, argv, params, LLAMA_EXAMPLE_SERVER)) {
return 1;
}
params.model_alias.insert(params.model.name);
}
- common_init();
-
// struct that contains llama context and inference
server_context ctx_server;
params.sampling.top_k = 4;
params.sampling.samplers = { COMMON_SAMPLER_TYPE_TOP_K, };
+ common_init();
+
if (!common_params_parse(argc, argv, params, LLAMA_EXAMPLE_TTS, print_usage)) {
return 1;
}
const int n_parallel = params.n_parallel;
const int n_predict = params.n_predict;
- common_init();
-
// init LLM
llama_backend_init();