|
|
--- |
|
|
library_name: peft |
|
|
license: apache-2.0 |
|
|
base_model: meta-llama/Llama-2-7b-hf |
|
|
tags: |
|
|
- resume-screening |
|
|
- hr-tech |
|
|
- llama2 |
|
|
- lora |
|
|
- peft |
|
|
- fine-tuned |
|
|
--- |
|
|
|
|
|
# Advanced Resume Screening Model |
|
|
|
|
|
## Model Description |
|
|
|
|
|
This is a LoRA (Low-Rank Adaptation) fine-tuned version of Llama-2-7B specifically optimized for resume screening and candidate evaluation tasks. The model can analyze resumes, extract key information, and provide structured assessments of candidate qualifications. |
|
|
|
|
|
- **Developed by:** kiritps |
|
|
- **Model type:** Causal Language Model (LoRA Fine-tuned) |
|
|
- **Language(s):** English |
|
|
- **License:** Apache 2.0 |
|
|
- **Finetuned from model:** meta-llama/Llama-2-7b-hf |
|
|
|
|
|
## Model Sources |
|
|
|
|
|
- **Repository:** https://huggingface.co/kiritps/Advanced-resume-screening |
|
|
|
|
|
## Uses |
|
|
|
|
|
### Direct Use |
|
|
|
|
|
This model is designed for HR professionals and recruitment systems to: |
|
|
- Analyze and screen resumes automatically |
|
|
- Extract key qualifications and skills |
|
|
- Provide structured candidate assessments |
|
|
- Filter candidates based on specific criteria |
|
|
- Generate summaries of candidate profiles |
|
|
|
|
|
### Downstream Use |
|
|
|
|
|
The model can be integrated into: |
|
|
- Applicant Tracking Systems (ATS) |
|
|
- HR management platforms |
|
|
- Recruitment automation tools |
|
|
- Candidate matching systems |
|
|
|
|
|
### Out-of-Scope Use |
|
|
|
|
|
- Should not be used as the sole decision-maker in hiring processes |
|
|
- Not intended for discriminatory screening based on protected characteristics |
|
|
- Not suitable for general-purpose text generation outside of resume/HR context |
|
|
|
|
|
## How to Get Started with the Model |
|
|
|
|
|
from transformers import AutoTokenizer, AutoModelForCausalLM |
|
|
from peft import PeftModel |
|
|
|
|
|
Load base model and tokenizer |
|
|
base_model = AutoModelForCausalLM.from_pretrained("meta-llama/Llama-2-7b-hf") |
|
|
tokenizer = AutoTokenizer.from_pretrained("meta-llama/Llama-2-7b-hf") |
|
|
|
|
|
Load LoRA adapter |
|
|
model = PeftModel.from_pretrained(base_model, "kiritps/Advanced-resume-screening") |
|
|
|
|
|
Example usage |
|
|
prompt = "Analyze this resume and provide key qualifications: [RESUME TEXT HERE]" |
|
|
inputs = tokenizer(prompt, return_tensors="pt") |
|
|
outputs = model.generate(**inputs, max_length=512, temperature=0.7) |
|
|
response = tokenizer.decode(outputs, skip_special_tokens=True) |
|
|
|
|
|
text |
|
|
|
|
|
## Training Details |
|
|
|
|
|
### Training Data |
|
|
|
|
|
The model was fine-tuned on a curated dataset of resume-response pairs, designed to teach the model how to: |
|
|
- Extract relevant information from resumes |
|
|
- Provide structured analysis of candidate qualifications |
|
|
- Generate appropriate screening responses |
|
|
|
|
|
### Training Procedure |
|
|
|
|
|
#### Training Hyperparameters |
|
|
|
|
|
- **Training regime:** 4-bit quantization with bfloat16 mixed precision |
|
|
- **LoRA rank:** 64 |
|
|
- **LoRA alpha:** 16 |
|
|
- **Learning rate:** 2e-4 |
|
|
- **Batch size:** 4 |
|
|
- **Gradient accumulation steps:** 4 |
|
|
- **Training epochs:** Multiple checkpoints saved (3840, 4320, 4800, 5280, 5760 steps) |
|
|
|
|
|
#### Quantization Configuration |
|
|
|
|
|
- **Quantization method:** bitsandbytes |
|
|
- **Load in 4bit:** True |
|
|
- **Quantization type:** nf4 |
|
|
- **Double quantization:** True |
|
|
- **Compute dtype:** bfloat16 |
|
|
|
|
|
## Bias, Risks, and Limitations |
|
|
|
|
|
### Limitations |
|
|
|
|
|
- Model responses should be reviewed by human recruiters |
|
|
- May exhibit biases present in training data |
|
|
- Performance may vary across different industries or job types |
|
|
- Requires careful prompt engineering for optimal results |
|
|
|
|
|
### Recommendations |
|
|
|
|
|
- Use as a screening aid, not a replacement for human judgment |
|
|
- Regularly audit outputs for potential bias |
|
|
- Combine with diverse evaluation methods |
|
|
- Ensure compliance with local employment laws and regulations |
|
|
|
|
|
## Technical Specifications |
|
|
|
|
|
### Model Architecture |
|
|
|
|
|
- **Parameter Count:** ~7B parameters (base) + LoRA adapters |
|
|
- **Quantization:** 4-bit NF4 quantization |
|
|
|
|
|
### Compute Infrastructure |
|
|
|
|
|
#### Hardware |
|
|
- GPU training environment |
|
|
- Compatible with consumer and enterprise GPUs |
|
|
|
|
|
#### Software |
|
|
- **Framework:** PyTorch |
|
|
- **PEFT Version:** 0.6.2 |
|
|
- **Transformers:** Latest compatible version |
|
|
- **Quantization:** bitsandbytes |
|
|
|
|
|
## Training Procedure |
|
|
|
|
|
The following `bitsandbytes` quantization config was used during training: |
|
|
- quant_method: bitsandbytes |
|
|
- load_in_8bit: False |
|
|
- load_in_4bit: True |
|
|
- llm_int8_threshold: 6.0 |
|
|
- llm_int8_skip_modules: None |
|
|
- llm_int8_enable_fp32_cpu_offload: False |
|
|
- llm_int8_has_fp16_weight: False |
|
|
- bnb_4bit_quant_type: nf4 |
|
|
- bnb_4bit_use_double_quant: True |
|
|
- bnb_4bit_compute_dtype: bfloat16 |
|
|
|
|
|
### Framework Versions |
|
|
- PEFT 0.6.2 |
|
|
- Transformers (compatible version) |
|
|
- PyTorch (latest stable) |
|
|
- bitsandbytes (for quantization) |
|
|
|
|
|
## Model Card Authors |
|
|
|
|
|
kiritps |
|
|
|
|
|
## Model Card Contact |
|
|
|
|
|
For questions or issues regarding this model, please open an issue in the model repository. |