hc99's picture
Add files using upload-large-folder tool
362a075 verified
# SPDX-FileCopyrightText: 2022-present deepset GmbH <info@deepset.ai>
#
# SPDX-License-Identifier: Apache-2.0
import json
from pathlib import Path
from typing import Any, Dict, List, Literal, Optional, Set, Tuple, Union
from haystack import component, default_from_dict, default_to_dict, logging
from haystack.components.converters.utils import get_bytestream_from_source, normalize_metadata
from haystack.dataclasses import ByteStream, Document
from haystack.lazy_imports import LazyImport
logger = logging.getLogger(__name__)
with LazyImport("Run 'pip install jq'") as jq_import:
import jq
@component
class JSONConverter:
"""
Converts one or more JSON files into a text document.
### Usage examples
```python
import json
from haystack.components.converters import JSONConverter
from haystack.dataclasses import ByteStream
source = ByteStream.from_string(json.dumps({"text": "This is the content of my document"}))
converter = JSONConverter(content_key="text")
results = converter.run(sources=[source])
documents = results["documents"]
print(documents[0].content)
# 'This is the content of my document'
```
Optionally, you can also provide a `jq_schema` string to filter the JSON source files and `extra_meta_fields`
to extract from the filtered data:
```python
import json
from haystack.components.converters import JSONConverter
from haystack.dataclasses import ByteStream
data = {
"laureates": [
{
"firstname": "Enrico",
"surname": "Fermi",
"motivation": "for his demonstrations of the existence of new radioactive elements produced "
"by neutron irradiation, and for his related discovery of nuclear reactions brought about by"
" slow neutrons",
},
{
"firstname": "Rita",
"surname": "Levi-Montalcini",
"motivation": "for their discoveries of growth factors",
},
],
}
source = ByteStream.from_string(json.dumps(data))
converter = JSONConverter(
jq_schema=".laureates[]", content_key="motivation", extra_meta_fields={"firstname", "surname"}
)
results = converter.run(sources=[source])
documents = results["documents"]
print(documents[0].content)
# 'for his demonstrations of the existence of new radioactive elements produced by
# neutron irradiation, and for his related discovery of nuclear reactions brought
# about by slow neutrons'
print(documents[0].meta)
# {'firstname': 'Enrico', 'surname': 'Fermi'}
print(documents[1].content)
# 'for their discoveries of growth factors'
print(documents[1].meta)
# {'firstname': 'Rita', 'surname': 'Levi-Montalcini'}
```
"""
def __init__(
self,
jq_schema: Optional[str] = None,
content_key: Optional[str] = None,
extra_meta_fields: Optional[Union[Set[str], Literal["*"]]] = None,
):
"""
Creates a JSONConverter component.
An optional `jq_schema` can be provided to extract nested data in the JSON source files.
See the [official jq documentation](https://jqlang.github.io/jq/) for more info on the filters syntax.
If `jq_schema` is not set, whole JSON source files will be used to extract content.
Optionally, you can provide a `content_key` to specify which key in the extracted object must
be set as the document's content.
If both `jq_schema` and `content_key` are set, the component will search for the `content_key` in
the JSON object extracted by `jq_schema`. If the extracted data is not a JSON object, it will be skipped.
If only `jq_schema` is set, the extracted data must be a scalar value. If it's a JSON object or array,
it will be skipped.
If only `content_key` is set, the source JSON file must be a JSON object, else it will be skipped.
`extra_meta_fields` can either be set to a set of strings or a literal `"*"` string.
If it's a set of strings, it must specify fields in the extracted objects that must be set in
the extracted documents. If a field is not found, the meta value will be `None`.
If set to `"*"`, all fields that are not `content_key` found in the filtered JSON object will
be saved as metadata.
Initialization will fail if neither `jq_schema` nor `content_key` are set.
:param jq_schema:
Optional jq filter string to extract content.
If not specified, whole JSON object will be used to extract information.
:param content_key:
Optional key to extract document content.
If `jq_schema` is specified, the `content_key` will be extracted from that object.
:param extra_meta_fields:
An optional set of meta keys to extract from the content.
If `jq_schema` is specified, all keys will be extracted from that object.
"""
self._compiled_filter = None
if jq_schema:
jq_import.check()
self._compiled_filter = jq.compile(jq_schema)
self._jq_schema = jq_schema
self._content_key = content_key
self._meta_fields = extra_meta_fields
if self._compiled_filter is None and self._content_key is None:
msg = "No `jq_schema` nor `content_key` specified. Set either or both to extract data."
raise ValueError(msg)
def to_dict(self) -> Dict[str, Any]:
"""
Serializes the component to a dictionary.
:returns:
Dictionary with serialized data.
"""
return default_to_dict(
self, jq_schema=self._jq_schema, content_key=self._content_key, extra_meta_fields=self._meta_fields
)
@classmethod
def from_dict(cls, data: Dict[str, Any]) -> "JSONConverter":
"""
Deserializes the component from a dictionary.
:param data:
Dictionary to deserialize from.
:returns:
Deserialized component.
"""
return default_from_dict(cls, data)
def _get_content_and_meta(self, source: ByteStream) -> List[Tuple[str, Dict[str, Any]]]:
"""
Utility function to extract text and metadata from a JSON file.
:param source:
UTF-8 byte stream.
:returns:
Collection of text and metadata dict tuples, each corresponding
to a different document.
"""
try:
file_content = source.data.decode("utf-8")
except UnicodeError as exc:
logger.warning(
"Failed to extract text from {source}. Skipping it. Error: {error}",
source=source.meta["file_path"],
error=exc,
)
meta_fields = self._meta_fields or set()
if self._compiled_filter is not None:
try:
objects = list(self._compiled_filter.input_text(file_content))
except Exception as exc:
logger.warning(
"Failed to extract text from {source}. Skipping it. Error: {error}",
source=source.meta["file_path"],
error=exc,
)
return []
else:
# We just load the whole file as JSON if the user didn't provide a jq filter.
# We put it in a list even if it's not to ease handling it later on.
objects = [json.loads(file_content)]
result = []
if self._content_key is not None:
for obj in objects:
if not isinstance(obj, dict):
logger.warning("Expected a dictionary but got {obj}. Skipping it.", obj=obj)
continue
if self._content_key not in obj:
logger.warning(
"'{content_key}' not found in {obj}. Skipping it.", content_key=self._content_key, obj=obj
)
continue
text = obj[self._content_key]
if isinstance(text, (dict, list)):
logger.warning("Expected a scalar value but got {obj}. Skipping it.", obj=obj)
continue
meta = {}
if meta_fields == "*":
meta = {k: v for k, v in obj.items() if k != self._content_key}
else:
for field in meta_fields:
meta[field] = obj.get(field, None)
result.append((text, meta))
else:
for obj in objects:
if isinstance(obj, (dict, list)):
logger.warning("Expected a scalar value but got {obj}. Skipping it.", obj=obj)
continue
result.append((str(obj), {}))
return result
@component.output_types(documents=List[Document])
def run(
self,
sources: List[Union[str, Path, ByteStream]],
meta: Optional[Union[Dict[str, Any], List[Dict[str, Any]]]] = None,
):
"""
Converts a list of JSON files to documents.
:param sources:
A list of file paths or ByteStream objects.
:param meta:
Optional metadata to attach to the documents.
This value can be either a list of dictionaries or a single dictionary.
If it's a single dictionary, its content is added to the metadata of all produced documents.
If it's a list, the length of the list must match the number of sources.
If `sources` contain ByteStream objects, their `meta` will be added to the output documents.
:returns:
A dictionary with the following keys:
- `documents`: A list of created documents.
"""
documents = []
meta_list = normalize_metadata(meta=meta, sources_count=len(sources))
for source, metadata in zip(sources, meta_list):
try:
bytestream = get_bytestream_from_source(source)
except Exception as exc:
logger.warning("Could not read {source}. Skipping it. Error: {error}", source=source, error=exc)
continue
data = self._get_content_and_meta(bytestream)
for text, extra_meta in data:
merged_metadata = {**bytestream.meta, **metadata, **extra_meta}
document = Document(content=text, meta=merged_metadata)
documents.append(document)
return {"documents": documents}