chore(data): update knowledge base docs dataset with 14098 entries
Browse fileschore(data): update knowledge base docs dataset with 14098 entries
| source | count |
|:---------------------------|--------:|
| argo-cd | 131 |
| aws-eks-best-practices | 85 |
| cilium | 254 |
| ckad-exercises | 12 |
| ckad-exercises-2 | 12 |
| consul | 28 |
| docker | 10 |
| elasticsearch-6-2 | 451 |
| external-dns | 39 |
| external-secrets | 109 |
| gcp-professional-services | 225 |
| github-actions | 192 |
| gitlab | 18 |
| google-kubernetes-engine | 68 |
| grafana | 605 |
| helm | 97 |
| ingress-nginx | 61 |
| istio | 273 |
| jenkins | 141 |
| jupyter | 17 |
| kubernetes | 920 |
| kustomize | 36 |
| linux | 61 |
| loki | 197 |
| model-context-protocol-mcp | 3 |
| openai-agents | 125 |
| openai-codex | 5 |
| pandas | 30 |
| prometheus | 55 |
| redis | 87 |
| scikit-learn | 108 |
| smolagents | 21 |
| spacelift | 215 |
| tekton | 63 |
| terraform | 258 |
| vault | 571 |
- README.md +3 -3
- data/train-00000-of-00001.parquet +2 -2
|
@@ -15,10 +15,10 @@ dataset_info:
|
|
| 15 |
dtype: float64
|
| 16 |
splits:
|
| 17 |
- name: train
|
| 18 |
-
num_bytes:
|
| 19 |
num_examples: 14098
|
| 20 |
-
download_size:
|
| 21 |
-
dataset_size:
|
| 22 |
configs:
|
| 23 |
- config_name: default
|
| 24 |
data_files:
|
|
|
|
| 15 |
dtype: float64
|
| 16 |
splits:
|
| 17 |
- name: train
|
| 18 |
+
num_bytes: 96474038
|
| 19 |
num_examples: 14098
|
| 20 |
+
download_size: 53533914
|
| 21 |
+
dataset_size: 96474038
|
| 22 |
configs:
|
| 23 |
- config_name: default
|
| 24 |
data_files:
|
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:002af06fb20ae06fd2179dc2dc164030dee51a42f899ab173fb31aa787e5890e
|
| 3 |
+
size 53533914
|