kiritps's picture
Update README.md
3ae40a6 verified
---
library_name: peft
license: apache-2.0
base_model: meta-llama/Llama-2-7b-hf
tags:
- resume-screening
- hr-tech
- llama2
- lora
- peft
- fine-tuned
---
# Advanced Resume Screening Model
## Model Description
This is a LoRA (Low-Rank Adaptation) fine-tuned version of Llama-2-7B specifically optimized for resume screening and candidate evaluation tasks. The model can analyze resumes, extract key information, and provide structured assessments of candidate qualifications.
- **Developed by:** kiritps
- **Model type:** Causal Language Model (LoRA Fine-tuned)
- **Language(s):** English
- **License:** Apache 2.0
- **Finetuned from model:** meta-llama/Llama-2-7b-hf
## Model Sources
- **Repository:** https://huggingface.co/kiritps/Advanced-resume-screening
## Uses
### Direct Use
This model is designed for HR professionals and recruitment systems to:
- Analyze and screen resumes automatically
- Extract key qualifications and skills
- Provide structured candidate assessments
- Filter candidates based on specific criteria
- Generate summaries of candidate profiles
### Downstream Use
The model can be integrated into:
- Applicant Tracking Systems (ATS)
- HR management platforms
- Recruitment automation tools
- Candidate matching systems
### Out-of-Scope Use
- Should not be used as the sole decision-maker in hiring processes
- Not intended for discriminatory screening based on protected characteristics
- Not suitable for general-purpose text generation outside of resume/HR context
## How to Get Started with the Model
from transformers import AutoTokenizer, AutoModelForCausalLM
from peft import PeftModel
Load base model and tokenizer
base_model = AutoModelForCausalLM.from_pretrained("meta-llama/Llama-2-7b-hf")
tokenizer = AutoTokenizer.from_pretrained("meta-llama/Llama-2-7b-hf")
Load LoRA adapter
model = PeftModel.from_pretrained(base_model, "kiritps/Advanced-resume-screening")
Example usage
prompt = "Analyze this resume and provide key qualifications: [RESUME TEXT HERE]"
inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(**inputs, max_length=512, temperature=0.7)
response = tokenizer.decode(outputs, skip_special_tokens=True)
text
## Training Details
### Training Data
The model was fine-tuned on a curated dataset of resume-response pairs, designed to teach the model how to:
- Extract relevant information from resumes
- Provide structured analysis of candidate qualifications
- Generate appropriate screening responses
### Training Procedure
#### Training Hyperparameters
- **Training regime:** 4-bit quantization with bfloat16 mixed precision
- **LoRA rank:** 64
- **LoRA alpha:** 16
- **Learning rate:** 2e-4
- **Batch size:** 4
- **Gradient accumulation steps:** 4
- **Training epochs:** Multiple checkpoints saved (3840, 4320, 4800, 5280, 5760 steps)
#### Quantization Configuration
- **Quantization method:** bitsandbytes
- **Load in 4bit:** True
- **Quantization type:** nf4
- **Double quantization:** True
- **Compute dtype:** bfloat16
## Bias, Risks, and Limitations
### Limitations
- Model responses should be reviewed by human recruiters
- May exhibit biases present in training data
- Performance may vary across different industries or job types
- Requires careful prompt engineering for optimal results
### Recommendations
- Use as a screening aid, not a replacement for human judgment
- Regularly audit outputs for potential bias
- Combine with diverse evaluation methods
- Ensure compliance with local employment laws and regulations
## Technical Specifications
### Model Architecture
- **Parameter Count:** ~7B parameters (base) + LoRA adapters
- **Quantization:** 4-bit NF4 quantization
### Compute Infrastructure
#### Hardware
- GPU training environment
- Compatible with consumer and enterprise GPUs
#### Software
- **Framework:** PyTorch
- **PEFT Version:** 0.6.2
- **Transformers:** Latest compatible version
- **Quantization:** bitsandbytes
## Training Procedure
The following `bitsandbytes` quantization config was used during training:
- quant_method: bitsandbytes
- load_in_8bit: False
- load_in_4bit: True
- llm_int8_threshold: 6.0
- llm_int8_skip_modules: None
- llm_int8_enable_fp32_cpu_offload: False
- llm_int8_has_fp16_weight: False
- bnb_4bit_quant_type: nf4
- bnb_4bit_use_double_quant: True
- bnb_4bit_compute_dtype: bfloat16
### Framework Versions
- PEFT 0.6.2
- Transformers (compatible version)
- PyTorch (latest stable)
- bitsandbytes (for quantization)
## Model Card Authors
kiritps
## Model Card Contact
For questions or issues regarding this model, please open an issue in the model repository.