初始化项目,由ModelHub XC社区提供模型
Model: mlabonne/NeuralDaredevil-8B-abliterated Source: Original Platform
This commit is contained in:
149
mlabonne_NeuralDaredevil-8B-abliterated.json
Normal file
149
mlabonne_NeuralDaredevil-8B-abliterated.json
Normal file
@@ -0,0 +1,149 @@
|
||||
{
|
||||
"bomFormat": "CycloneDX",
|
||||
"specVersion": "1.6",
|
||||
"serialNumber": "urn:uuid:ec2a096a-6435-4b15-94a2-3cdfe51237d6",
|
||||
"version": 1,
|
||||
"metadata": {
|
||||
"timestamp": "2025-07-10T08:46:54.205541+00:00",
|
||||
"component": {
|
||||
"type": "machine-learning-model",
|
||||
"bom-ref": "mlabonne/NeuralDaredevil-8B-abliterated-137a94eb-5b88-59d1-88c3-66d6367d780d",
|
||||
"name": "mlabonne/NeuralDaredevil-8B-abliterated",
|
||||
"externalReferences": [
|
||||
{
|
||||
"url": "https://huggingface.co/mlabonne/NeuralDaredevil-8B-abliterated",
|
||||
"type": "documentation"
|
||||
}
|
||||
],
|
||||
"modelCard": {
|
||||
"modelParameters": {
|
||||
"task": "text-generation",
|
||||
"architectureFamily": "llama",
|
||||
"modelArchitecture": "LlamaForCausalLM",
|
||||
"datasets": [
|
||||
{
|
||||
"ref": "mlabonne/orpo-dpo-mix-40k-898bbffe-f59a-53fe-a5b7-846dc484fbab"
|
||||
}
|
||||
]
|
||||
},
|
||||
"properties": [
|
||||
{
|
||||
"name": "library_name",
|
||||
"value": "transformers"
|
||||
}
|
||||
],
|
||||
"quantitativeAnalysis": {
|
||||
"performanceMetrics": [
|
||||
{
|
||||
"slice": "dataset: ai2_arc, split: test, config: ARC-Challenge",
|
||||
"type": "acc_norm",
|
||||
"value": 69.28
|
||||
},
|
||||
{
|
||||
"slice": "dataset: hellaswag, split: validation",
|
||||
"type": "acc_norm",
|
||||
"value": 85.05
|
||||
},
|
||||
{
|
||||
"slice": "dataset: cais/mmlu, split: test, config: all",
|
||||
"type": "acc",
|
||||
"value": 69.1
|
||||
},
|
||||
{
|
||||
"slice": "dataset: truthful_qa, split: validation, config: multiple_choice",
|
||||
"type": "mc2",
|
||||
"value": 60
|
||||
},
|
||||
{
|
||||
"slice": "dataset: winogrande, split: validation, config: winogrande_xl",
|
||||
"type": "acc",
|
||||
"value": 78.69
|
||||
},
|
||||
{
|
||||
"slice": "dataset: gsm8k, split: test, config: main",
|
||||
"type": "acc",
|
||||
"value": 71.8
|
||||
}
|
||||
]
|
||||
}
|
||||
},
|
||||
"authors": [
|
||||
{
|
||||
"name": "mlabonne"
|
||||
}
|
||||
],
|
||||
"licenses": [
|
||||
{
|
||||
"license": {
|
||||
"name": "llama3"
|
||||
}
|
||||
}
|
||||
],
|
||||
"tags": [
|
||||
"transformers",
|
||||
"safetensors",
|
||||
"llama",
|
||||
"text-generation",
|
||||
"dpo",
|
||||
"conversational",
|
||||
"dataset:mlabonne/orpo-dpo-mix-40k",
|
||||
"license:llama3",
|
||||
"model-index",
|
||||
"autotrain_compatible",
|
||||
"text-generation-inference",
|
||||
"endpoints_compatible",
|
||||
"region:us"
|
||||
]
|
||||
}
|
||||
},
|
||||
"components": [
|
||||
{
|
||||
"type": "data",
|
||||
"bom-ref": "mlabonne/orpo-dpo-mix-40k-898bbffe-f59a-53fe-a5b7-846dc484fbab",
|
||||
"name": "mlabonne/orpo-dpo-mix-40k",
|
||||
"data": [
|
||||
{
|
||||
"type": "dataset",
|
||||
"bom-ref": "mlabonne/orpo-dpo-mix-40k-898bbffe-f59a-53fe-a5b7-846dc484fbab",
|
||||
"name": "mlabonne/orpo-dpo-mix-40k",
|
||||
"contents": {
|
||||
"url": "https://huggingface.co/datasets/mlabonne/orpo-dpo-mix-40k",
|
||||
"properties": [
|
||||
{
|
||||
"name": "task_categories",
|
||||
"value": "text-generation"
|
||||
},
|
||||
{
|
||||
"name": "language",
|
||||
"value": "en"
|
||||
},
|
||||
{
|
||||
"name": "pretty_name",
|
||||
"value": "ORPO-DPO-mix-40k"
|
||||
},
|
||||
{
|
||||
"name": "configs",
|
||||
"value": "Name of the dataset subset: default {\"split\": \"train\", \"path\": \"data/train-*\"}"
|
||||
},
|
||||
{
|
||||
"name": "license",
|
||||
"value": "apache-2.0"
|
||||
}
|
||||
]
|
||||
},
|
||||
"governance": {
|
||||
"owners": [
|
||||
{
|
||||
"organization": {
|
||||
"name": "mlabonne",
|
||||
"url": "https://huggingface.co/mlabonne"
|
||||
}
|
||||
}
|
||||
]
|
||||
},
|
||||
"description": "\n\t\n\t\t\n\t\tORPO-DPO-mix-40k v1.2\n\t\n\n\nThis dataset is designed for ORPO or DPO training.\nSee Fine-tune Llama 3 with ORPO for more information about how to use it.\nIt is a combination of the following high-quality DPO datasets:\n\nargilla/Capybara-Preferences: highly scored chosen answers >=5 (7,424 samples)argilla/distilabel-intel-orca-dpo-pairs: highly scored chosen answers >=9, not in GSM8K (2,299 samples)\nargilla/ultrafeedback-binarized-preferences-cleaned: highly scored chosen answers >=5 (22\u2026 See the full description on the dataset page: https://huggingface.co/datasets/mlabonne/orpo-dpo-mix-40k."
|
||||
}
|
||||
]
|
||||
}
|
||||
]
|
||||
}
|
||||
Reference in New Issue
Block a user