omim / README.md
Bgoood's picture
upload
857fc7a
---
annotations_creators:
- found
language:
- en
license: mit
multilinguality: monolingual
pretty_name: vep_mendelian_traits_chr11_split
size_categories:
- 10K<n<100K
source_datasets:
- original
task_categories:
- sequence-modeling
task_ids:
- sequence-classification
---
# vep_clinvar_chr1_split
- 字段: ref, alt, label, chromosome, position
- 划分: chromosome=1为test,其余为train
- 支持自动生成ref/alt序列
## 用法
```python
from datasets import load_dataset
ds = load_dataset(
"Bgoood/vep_mendelian_traits_chr11_split",
sequence_length=2048,
fasta_path="/path/to/hg38.fa.gz",
data_dir="."
)
```
```
---
## 5. 上传到 HuggingFace
1. **初始化git repo(如果还没有)**
```bash
git lfs install
git clone https://huggingface.co/datasets/Bgoood/vep_mendelian_traits_chr11_split
cd vep_mendelian_traits_chr11_split
# 把 train.csv, test.csv, vep_mendelian_traits_chr11_split.py, README.md 放到这个目录
git add .
git commit -m "init dataset with script"
git push
```
2. **或者直接网页上传**
在你的数据集页面,点击“Add file”,上传上述文件。
---
## 6. 用户使用方式
用户只需这样调用即可自动生成ref/alt序列:
```python
from datasets import load_dataset
ds = load_dataset(
"Bgoood/vep_mendelian_traits_chr11_split",
sequence_length=2048,
fasta_path="/path/to/hg38.fa.gz",
data_dir="."
)
```
---
## 7. 依赖
确保用户环境已安装:
```bash
pip install datasets pyfaidx pandas
```
---
## 8. 注意事项
- `fasta_path` 必须是本地可访问的 hg38.fa.gz 路径。
- 你上传到HF的数据集只需包含原始csv和脚本,不需要包含fasta文件。
---
如需自动化脚本生成csv、或有其他定制需求,请随时告知!