diff --git a/leap/Q4_0.json b/leap/Q4_0.json new file mode 100644 index 0000000..cd412f6 --- /dev/null +++ b/leap/Q4_0.json @@ -0,0 +1,14 @@ +{ + "inference_type": "llama.cpp/text-to-text", + "schema_version": "1.0.0", + "load_time_parameters": { + "model": "../LFM2-700M-Q4_0.gguf" + }, + "generation_time_parameters": { + "sampling_parameters": { + "temperature": 0.3, + "min_p": 0.15, + "repetition_penalty": 1.05 + } + } +} \ No newline at end of file diff --git a/leap/Q4_K_M.json b/leap/Q4_K_M.json new file mode 100644 index 0000000..d13f663 --- /dev/null +++ b/leap/Q4_K_M.json @@ -0,0 +1,14 @@ +{ + "inference_type": "llama.cpp/text-to-text", + "schema_version": "1.0.0", + "load_time_parameters": { + "model": "../LFM2-700M-Q4_K_M.gguf" + }, + "generation_time_parameters": { + "sampling_parameters": { + "temperature": 0.3, + "min_p": 0.15, + "repetition_penalty": 1.05 + } + } +} \ No newline at end of file diff --git a/leap/Q8_0.json b/leap/Q8_0.json new file mode 100644 index 0000000..a67b1d6 --- /dev/null +++ b/leap/Q8_0.json @@ -0,0 +1,14 @@ +{ + "inference_type": "llama.cpp/text-to-text", + "schema_version": "1.0.0", + "load_time_parameters": { + "model": "../LFM2-700M-Q8_0.gguf" + }, + "generation_time_parameters": { + "sampling_parameters": { + "temperature": 0.3, + "min_p": 0.15, + "repetition_penalty": 1.05 + } + } +} \ No newline at end of file