Update README.md
Browse files
README.md
CHANGED
|
@@ -20,13 +20,6 @@ pipeline_tag: feature-extraction
|
|
| 20 |
|
| 21 |
<!-- Provide a quick summary of what the model is/does. -->
|
| 22 |
|
| 23 |
-
|
| 24 |
-
|
| 25 |
-
## Model Details
|
| 26 |
-
Here’s a more detailed and structured version of the **Model Details** section for your README:
|
| 27 |
-
|
| 28 |
-
---
|
| 29 |
-
|
| 30 |
### Model Details
|
| 31 |
|
| 32 |
#### Model Description
|
|
@@ -78,13 +71,7 @@ The model operates by processing raw code into tokenized representations, then u
|
|
| 78 |
- **Language(s) (NLP):** [(Code-based vulnerabilities in various programming languages like Python, C++, JavaScript, etc.]
|
| 79 |
- **License:** [MIT License]
|
| 80 |
- **Finetuned from model [ meta-llama/Llama-3.1-8B-Instruct]:** [GRPO]
|
| 81 |
-
|
| 82 |
-
|
| 83 |
-
|
| 84 |
-
## Uses
|
| 85 |
-
|
| 86 |
-
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
|
| 87 |
-
|
| 88 |
### Direct Use
|
| 89 |
|
| 90 |
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
|
|
@@ -97,9 +84,7 @@ Analyze code snippets or entire codebases to identify potential security flaws.
|
|
| 97 |
Integrate into CI/CD pipelines for real-time vulnerability detection during development.
|
| 98 |
Provide developers with actionable feedback and recommendations on how to fix the issues identified]
|
| 99 |
|
| 100 |
-
### Downstream Use
|
| 101 |
-
|
| 102 |
-
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
|
| 103 |
|
| 104 |
[This model can be fine-tuned further for specific use cases or integrated into larger security frameworks, such as:
|
| 105 |
|
|
@@ -130,11 +115,8 @@ The model should not be used to mislead or manipulate software security in malic
|
|
| 130 |
[The model was fine-tuned on the Google CodeXGlue Defect Detection dataset, a part of the CodeXGlue benchmark. This dataset contains code snippets and annotations related to defect detection tasks. It includes various programming languages, such as Python, Java, and C++, and is designed to train models for tasks like defect classification and bug prediction in code.]
|
| 131 |
|
| 132 |
|
| 133 |
-
|
| 134 |
#### Training Hyperparameters
|
| 135 |
|
| 136 |
-
- #### Training Hyperparameters
|
| 137 |
-
|
| 138 |
- **Training regime**: fp16 mixed precision
|
| 139 |
- The model was fine-tuned using **16-bit mixed precision (fp16)** training. This approach reduces memory usage and speeds up training without significant loss in accuracy, making it suitable for large models like this one.
|
| 140 |
|
|
@@ -164,8 +146,6 @@ The model should not be used to mislead or manipulate software security in malic
|
|
| 164 |
<!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
|
| 165 |
|
| 166 |
|
| 167 |
-
|
| 168 |
-
|
| 169 |
## Model Card Contact
|
| 170 |
|
| 171 |
For inquiries or more information about this model, please contact:
|
|
|
|
| 20 |
|
| 21 |
<!-- Provide a quick summary of what the model is/does. -->
|
| 22 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 23 |
### Model Details
|
| 24 |
|
| 25 |
#### Model Description
|
|
|
|
| 71 |
- **Language(s) (NLP):** [(Code-based vulnerabilities in various programming languages like Python, C++, JavaScript, etc.]
|
| 72 |
- **License:** [MIT License]
|
| 73 |
- **Finetuned from model [ meta-llama/Llama-3.1-8B-Instruct]:** [GRPO]
|
| 74 |
+
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 75 |
### Direct Use
|
| 76 |
|
| 77 |
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
|
|
|
|
| 84 |
Integrate into CI/CD pipelines for real-time vulnerability detection during development.
|
| 85 |
Provide developers with actionable feedback and recommendations on how to fix the issues identified]
|
| 86 |
|
| 87 |
+
### Downstream Use
|
|
|
|
|
|
|
| 88 |
|
| 89 |
[This model can be fine-tuned further for specific use cases or integrated into larger security frameworks, such as:
|
| 90 |
|
|
|
|
| 115 |
[The model was fine-tuned on the Google CodeXGlue Defect Detection dataset, a part of the CodeXGlue benchmark. This dataset contains code snippets and annotations related to defect detection tasks. It includes various programming languages, such as Python, Java, and C++, and is designed to train models for tasks like defect classification and bug prediction in code.]
|
| 116 |
|
| 117 |
|
|
|
|
| 118 |
#### Training Hyperparameters
|
| 119 |
|
|
|
|
|
|
|
| 120 |
- **Training regime**: fp16 mixed precision
|
| 121 |
- The model was fine-tuned using **16-bit mixed precision (fp16)** training. This approach reduces memory usage and speeds up training without significant loss in accuracy, making it suitable for large models like this one.
|
| 122 |
|
|
|
|
| 146 |
<!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
|
| 147 |
|
| 148 |
|
|
|
|
|
|
|
| 149 |
## Model Card Contact
|
| 150 |
|
| 151 |
For inquiries or more information about this model, please contact:
|