Rithankoushik commited on
Commit
0fb8f24
·
verified ·
1 Parent(s): a0c8aaa

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +74 -3
README.md CHANGED
@@ -1,3 +1,74 @@
1
- ---
2
- license: apache-2.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ language:
4
+ - en
5
+ base_model:
6
+ - Qwen/Qwen3-0.6B
7
+ tags:
8
+ - job-parsing
9
+ - qwen3
10
+ - lora
11
+ - json-extraction
12
+ ---
13
+
14
+ # 📦 Qwen3-0.6B — Job Description Struct-Extractor
15
+
16
+ A fine-tuned version of **Qwen3-0.6B** designed for accurate extraction of structured job attributes from raw job descriptions.
17
+ This model outputs strict, schema-aligned JSON, making it perfect for downstream applications like search, analytics, and recommendation systems.
18
+
19
+ ---
20
+
21
+ ## 🚀 Model Highlights
22
+
23
+ - **Base Model:** Qwen/Qwen3-0.6B
24
+ - **Architecture:** Decoder-only Transformer (Causal LM)
25
+ - **Tokenizer:** QwenTokenizer (same as base)
26
+
27
+ **Fine-Tuned For:**
28
+ - Zero-hallucination extraction
29
+ - Schema-conformant JSON outputs
30
+
31
+ ---
32
+
33
+ ## 🎯 Task Overview
34
+
35
+ - **Task:** Extract structured fields from job descriptions
36
+ - **Output:** JSON strictly following a predefined schema
37
+
38
+ **Use Cases:**
39
+ - Automated JD parsing into structured fields
40
+ - Talent platform search & recommendation engines
41
+ - HR data cleaning & analytics pipelines
42
+ - Resume ↔ Job matching systems
43
+
44
+ ---
45
+
46
+ ## 🖥️ Inference Example (Python)
47
+
48
+ ```python
49
+ import torch
50
+ import re
51
+ import time
52
+ import json
53
+ import json5
54
+ from transformers import AutoTokenizer, AutoModelForCausalLM
55
+ from peft import PeftModel
56
+
57
+ # Model paths
58
+ base_model_id = "Qwen/Qwen3-0.6B"
59
+ lora_model_id = "Rithankoushik/Qwen-0.6-Job-parser-Model"
60
+
61
+ # Load tokenizer
62
+ tokenizer = AutoTokenizer.from_pretrained(base_model_id, trust_remote_code=True)
63
+ tokenizer.pad_token = tokenizer.eos_token
64
+
65
+ # Load model + LoRA
66
+ base_model = AutoModelForCausalLM.from_pretrained(
67
+ base_model_id,
68
+ trust_remote_code=True,
69
+ torch_dtype=torch.float16,
70
+ device_map="auto"
71
+ )
72
+ model = PeftModel.from_pretrained(base_model, lora_model_id, device_map="auto")
73
+ model = model.merge_and_unload()
74
+ model.eval()