chore(data): update knowledge base docs dataset with 14046 entries
Browse fileschore(data): update knowledge base docs dataset with 14046 entries
| source | count |
|:---------------------------|--------:|
| argo-cd | 133 |
| aws-eks-best-practices | 85 |
| cilium | 258 |
| ckad-exercises | 12 |
| ckad-exercises-2 | 12 |
| consul | 28 |
| docker | 10 |
| elasticsearch-6-2 | 451 |
| external-dns | 39 |
| external-secrets | 110 |
| gcp-professional-services | 225 |
| github-actions | 193 |
| gitlab | 18 |
| google-kubernetes-engine | 68 |
| grafana | 599 |
| helm | 97 |
| ingress-nginx | 61 |
| istio | 274 |
| jenkins | 141 |
| jupyter | 17 |
| kubernetes | 924 |
| kustomize | 36 |
| linux | 61 |
| loki | 198 |
| model-context-protocol-mcp | 4 |
| openai-agents | 19 |
| openai-codex | 8 |
| pandas | 30 |
| prometheus | 55 |
| redis | 87 |
| scikit-learn | 109 |
| smolagents | 21 |
| spacelift | 216 |
| streamlit | 14 |
| tekton | 63 |
| terraform | 258 |
| vault | 571 |
- README.md +4 -4
- data/train-00000-of-00001.parquet +2 -2
|
@@ -15,10 +15,10 @@ dataset_info:
|
|
| 15 |
dtype: float64
|
| 16 |
splits:
|
| 17 |
- name: train
|
| 18 |
-
num_bytes:
|
| 19 |
-
num_examples:
|
| 20 |
-
download_size:
|
| 21 |
-
dataset_size:
|
| 22 |
configs:
|
| 23 |
- config_name: default
|
| 24 |
data_files:
|
|
|
|
| 15 |
dtype: float64
|
| 16 |
splits:
|
| 17 |
- name: train
|
| 18 |
+
num_bytes: 95999084
|
| 19 |
+
num_examples: 14046
|
| 20 |
+
download_size: 53275956
|
| 21 |
+
dataset_size: 95999084
|
| 22 |
configs:
|
| 23 |
- config_name: default
|
| 24 |
data_files:
|
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4aa0b68525005d8563daeaa3619148f4049b6b9167c3ad7a678afec6dcb4e9c5
|
| 3 |
+
size 53275956
|