diff --git a/leap/Q4_0.json b/leap/Q4_0.json new file mode 100644 index 0000000..9d79785 --- /dev/null +++ b/leap/Q4_0.json @@ -0,0 +1,14 @@ +{ + "inference_type": "llama.cpp/text-to-text", + "schema_version": "1.0.0", + "load_time_parameters": { + "model": "../LFM2-1.2B-Q4_0.gguf" + }, + "generation_time_parameters": { + "sampling_parameters": { + "temperature": 0.3, + "min_p": 0.15, + "repetition_penalty": 1.05 + } + } +} \ No newline at end of file diff --git a/leap/Q4_K_M.json b/leap/Q4_K_M.json new file mode 100644 index 0000000..05877f4 --- /dev/null +++ b/leap/Q4_K_M.json @@ -0,0 +1,14 @@ +{ + "inference_type": "llama.cpp/text-to-text", + "schema_version": "1.0.0", + "load_time_parameters": { + "model": "../LFM2-1.2B-Q4_K_M.gguf" + }, + "generation_time_parameters": { + "sampling_parameters": { + "temperature": 0.3, + "min_p": 0.15, + "repetition_penalty": 1.05 + } + } +} \ No newline at end of file diff --git a/leap/Q8_0.json b/leap/Q8_0.json new file mode 100644 index 0000000..a1d14b8 --- /dev/null +++ b/leap/Q8_0.json @@ -0,0 +1,14 @@ +{ + "inference_type": "llama.cpp/text-to-text", + "schema_version": "1.0.0", + "load_time_parameters": { + "model": "../LFM2-1.2B-Q8_0.gguf" + }, + "generation_time_parameters": { + "sampling_parameters": { + "temperature": 0.3, + "min_p": 0.15, + "repetition_penalty": 1.05 + } + } +} \ No newline at end of file