wiki-sample / README.md
sebawita's picture
Update README.md
456237f verified
|
raw
history blame
1.31 kB
metadata
license: bsd-3-clause
configs:
  - config_name: openai-text-embedding-3-large
    data_files:
      - split: train
        path: openai/text-embedding-3-large/*.parquet
  - config_name: snowflake-arctic-embed
    data_files:
      - split: train
        path: ollama/snowflake-arctic/*.parquet
size_categories:
  - 100K<n<1M

Loading dataset with vector embeddings

You can load the dataset like this:

from datasets import load_dataset
dataset = load_dataset("weaviate/wiki-sample", "openai-text-embedding-3-large", split="train", streaming=True)
# dataset = load_dataset("weaviate/wiki-sample", "snowflake-arctic-embed", split="train", streaming=True)

for item in dataset:
    print(item["text"])
    print(item["title"])
    print(item["url"])
    print(item["wiki_id"])
    print(item["vector"])
    print()

Supported Datasets

OpenAI

text-embedding-3-large - 3072d vectors - generated with OpenAI

from datasets import load_dataset
dataset = load_dataset("weaviate/wiki-sample", "openai-text-embedding-3-large", split="train", streaming=True)

Snowflake

snowflake-arctic-embed - 1024 vectors - generated with Ollama

from datasets import load_dataset
dataset = load_dataset("weaviate/wiki-sample", "snowflake-arctic-embed", split="train", streaming=True)