float mirostat_tau = 5.00f; // target entropy
float mirostat_eta = 0.10f; // learning rate
- std::string model = "models/lamma-7B/ggml-model.bin"; // model path
+ std::string model = "models/7B/ggml-model.bin"; // model path
std::string prompt = "";
std::string path_prompt_cache = ""; // path to file for saving/loading prompt eval state
std::string input_prefix = ""; // string to prefix user inputs with
int main(int argc, char ** argv) {
gpt_params params;
- params.model = "models/llama-7B/ggml-model.bin";
if (gpt_params_parse(argc, argv, params) == false) {
return 1;
int main(int argc, char ** argv) {
gpt_params params;
- params.model = "models/llama-7B/ggml-model.bin";
if (gpt_params_parse(argc, argv, params) == false) {
return 1;
int main(int argc, char ** argv) {
gpt_params params;
- params.model = "models/llama-7B/ggml-model.bin";
params.n_batch = 512;
if (gpt_params_parse(argc, argv, params) == false) {
int main(int argc, char ** argv) {
gpt_params params;
- params.model = "models/llama-7B/ggml-model.bin";
params.seed = 42;
params.n_threads = 4;
params.repeat_last_n = 64;