inputs
stringlengths
312
52k
targets
stringlengths
1
3.1k
block_type
stringclasses
11 values
scenario
stringclasses
7 values
<filename>ollama-python/ollama/_client.py<fim_prefix>import os import io import json import httpx import binascii import platform import urllib.parse from os import PathLike from pathlib import Path from hashlib import sha256 from base64 import b64encode, b64decode from typing import Any, AnyStr, Union, Optional, Sequ...
except ResponseError as e: if e.status_code != 404: raise with open(path, 'rb') as r: self._request('POST', f'/api/blobs/{digest}', content=r)
CATCH
prefix_suffix_full_complete_current_block_no_evidence
<filename>ollama-python/ollama/_client.py<fim_prefix>import os import io import json import httpx import binascii import platform import urllib.parse from os import PathLike from pathlib import Path from hashlib import sha256 from base64 import b64encode, b64decode from typing import Any, AnyStr, Union, Optional, Sequ...
except httpx.HTTPStatusError as e: e.response.read() raise ResponseError(e.response.text, e.response.status_code) from None
CATCH
prefix_suffix_full_complete_current_block_no_evidence
<filename>ollama-python/ollama/_client.py<fim_prefix>import os import io import json import httpx import binascii import platform import urllib.parse from os import PathLike from pathlib import Path from hashlib import sha256 from base64 import b64encode, b64decode from typing import Any, AnyStr, Union, Optional, Sequ...
except httpx.HTTPStatusError as e: e.response.read() raise ResponseError(e.response.text, e.response.status_code) from None
CATCH
prefix_suffix_full_complete_current_block_no_evidence
<filename>ollama-python/ollama/_client.py<fim_prefix>import os import io import json import httpx import binascii import platform import urllib.parse from os import PathLike from pathlib import Path from hashlib import sha256 from base64 import b64encode, b64decode from typing import Any, AnyStr, Union, Optional, Sequ...
except ResponseError as e: if e.status_code != 404: raise async def upload_bytes(): with open(path, 'rb') as r: while True: chunk = r.read(32 * 1024) if not chunk: break yield chunk await self._request('POST', f'/api/blobs/{...
CATCH
prefix_suffix_full_complete_current_block_no_evidence
<filename>ollama-python/ollama/_types.py<fim_prefix>import json from typing import Any, TypedDict, Sequence, Literal import sys if sys.version_info < (3, 11): from typing_extensions import NotRequired else: from typing import NotRequired class BaseGenerateResponse(TypedDict): model: str 'Model used to gener...
except json.JSONDecodeError: ...
CATCH
prefix_suffix_full_complete_current_block_no_evidence
<filename>ollama-python/ollama/_client.py<fim_prefix>import os import io import json import httpx import binascii import platform import urllib.parse from os import PathLike from pathlib import Path from hashlib import sha256 from base64 import b64encode, b64decode from typing import Any, AnyStr, Union, Optional, Sequ...
except httpx.HTTPStatusError as e: e.response.read() raise ResponseError(e.response.text, e.response.status_code) from None
CATCH
prefix_suffix_full_complete_current_block_no_evidence
<filename>ollama-python/ollama/_client.py<fim_prefix>import os import io import json import httpx import binascii import platform import urllib.parse from os import PathLike from pathlib import Path from hashlib import sha256 from base64 import b64encode, b64decode from typing import Any, AnyStr, Union, Optional, Sequ...
for line in io.StringIO(modelfile): command, _, args = line.partition(' ') if command.upper() not in ['FROM', 'ADAPTER']: print(line, end='', file=out) continue path = Path(args.strip()).expanduser() path = path if path.is_absolute() else base / path if path.exists(): ...
FOR
prefix_suffix_full_complete_current_block_no_evidence
<filename>ollama-python/ollama/_client.py<fim_prefix>import os import io import json import httpx import binascii import platform import urllib.parse from os import PathLike from pathlib import Path from hashlib import sha256 from base64 import b64encode, b64decode from typing import Any, AnyStr, Union, Optional, Sequ...
for message in messages or []: if not isinstance(message, dict): raise TypeError('messages must be a list of strings') if not (role := message.get('role')) or role not in ['system', 'user', 'assistant']: raise RequestError('messages must contain a role and it must be one of "system", "user",...
FOR
prefix_suffix_full_complete_current_block_no_evidence
<filename>ollama-python/ollama/_client.py<fim_prefix>import os import io import json import httpx import binascii import platform import urllib.parse from os import PathLike from pathlib import Path from hashlib import sha256 from base64 import b64encode, b64decode from typing import Any, AnyStr, Union, Optional, Sequ...
for line in r.iter_lines(): partial = json.loads(line) if e := partial.get('error'): raise ResponseError(e) yield partial
FOR
prefix_suffix_full_complete_current_block_no_evidence
<filename>ollama-python/ollama/_client.py<fim_prefix>import os import io import json import httpx import binascii import platform import urllib.parse from os import PathLike from pathlib import Path from hashlib import sha256 from base64 import b64encode, b64decode from typing import Any, AnyStr, Union, Optional, Sequ...
for line in io.StringIO(modelfile): command, _, args = line.partition(' ') if command.upper() not in ['FROM', 'ADAPTER']: print(line, end='', file=out) continue path = Path(args.strip()).expanduser() path = path if path.is_absolute() else base / path if path.exists(): ...
FOR
prefix_suffix_full_complete_current_block_no_evidence
<filename>ollama-python/ollama/_client.py<fim_prefix>import os import io import json import httpx import binascii import platform import urllib.parse from os import PathLike from pathlib import Path from hashlib import sha256 from base64 import b64encode, b64decode from typing import Any, AnyStr, Union, Optional, Sequ...
for message in messages or []: if not isinstance(message, dict): raise TypeError('messages must be a list of Message or dict-like objects') if not (role := message.get('role')) or role not in ['system', 'user', 'assistant']: raise RequestError('messages must contain a role and it must be one...
FOR
prefix_suffix_full_complete_current_block_no_evidence
<filename>ollama-python/ollama/_client.py<fim_prefix>import os import io import json import httpx import binascii import platform import urllib.parse from os import PathLike from pathlib import Path from hashlib import sha256 from base64 import b64encode, b64decode from typing import Any, AnyStr, Union, Optional, Sequ...
async for line in r.aiter_lines(): partial = json.loads(line) if e := partial.get('error'): raise ResponseError(e) yield partial
FOR
prefix_suffix_full_complete_current_block_no_evidence
<filename>ollama-python/ollama/_client.py<fim_prefix>import os import io import json import httpx import binascii import platform import urllib.parse from os import PathLike from pathlib import Path from hashlib import sha256 from base64 import b64encode, b64decode from typing import Any, AnyStr, Union, Optional, Sequ...
async for line in r.aiter_lines(): partial = json.loads(line) if e := partial.get('error'): raise ResponseError(e) yield partial
FOR
prefix_suffix_full_complete_current_block_no_evidence
<filename>ollama-python/ollama/_client.py<fim_prefix>import os import io import json import httpx import binascii import platform import urllib.parse from os import PathLike from pathlib import Path from hashlib import sha256 from base64 import b64encode, b64decode from typing import Any, AnyStr, Union, Optional, Sequ...
async def inner(): async with self._client.stream(method, url, **kwargs) as r: try: r.raise_for_status() except httpx.HTTPStatusError as e: e.response.read() raise ResponseError(e.response.text, e.response.status_code) from None async for line in r.aiter_line...
METHOD
prefix_suffix_full_complete_current_block_no_evidence
<filename>ollama-python/ollama/_client.py<fim_prefix>import os import io import json import httpx import binascii import platform import urllib.parse from os import PathLike from pathlib import Path from hashlib import sha256 from base64 import b64encode, b64decode from typing import Any, AnyStr, Union, Optional, Sequ...
async def upload_bytes(): with open(path, 'rb') as r: while True: chunk = r.read(32 * 1024) if not chunk: break yield chunk
METHOD
prefix_suffix_full_complete_current_block_no_evidence
<filename>ollama-python/ollama/_client.py<fim_prefix>import os import io import json import httpx import binascii import platform import urllib.parse from os import PathLike from pathlib import Path from hashlib import sha256 from base64 import b64encode, b64decode from typing import Any, AnyStr, Union, Optional, Sequ...
while True: chunk = r.read(32 * 1024) if not chunk: break yield chunk
WHILE
prefix_suffix_full_complete_current_block_no_evidence
<filename>ollama-python/ollama/_client.py<fim_prefix>import os import io import json import httpx import binascii import platform import urllib.parse from os import PathLike from pathlib import Path from hashlib import sha256 from base64 import b64encode, b64decode from typing import Any, AnyStr, Union, Optional, Sequ...
while True: chunk = r.read(32 * 1024) if not chunk: break sha256sum.update(chunk)
WHILE
prefix_suffix_full_complete_current_block_no_evidence
<filename>ollama-python/ollama/_client.py<fim_prefix>import os import io import json import httpx import binascii import platform import urllib.parse from os import PathLike from pathlib import Path from hashlib import sha256 from base64 import b64encode, b64decode from typing import Any, AnyStr, Union, Optional, Sequ...
while True: chunk = r.read(32 * 1024) if not chunk: break sha256sum.update(chunk)
WHILE
prefix_suffix_full_complete_current_block_no_evidence
<filename>open-iris/src/iris/nodes/iris_response/conv_filter_bank.py<fim_prefix>from typing import List, Tuple import numpy as np from pydantic import root_validator, validator from iris.io.class_configs import Algorithm from iris.io.dataclasses import IrisFilterResponse, NormalizedIris from iris.io.validators import...
"""Apply convolution to a given normalized iris image with the filter and probe schema. Args: img_filter (ImageFilter): filter used for convolution. probe_schema (ProbeSchema): probe schema used for convolution. normalization_output (NormalizedIris): Output of the normalizat...
BLOCK_COMMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>open-iris/src/iris/nodes/geometry_refinement/contour_points_filter.py<fim_prefix>import cv2 import numpy as np from pydantic import Field from iris.io.class_configs import Algorithm from iris.io.dataclasses import GeometryPolygons, NoiseMask class ContourPointNoiseEyeballDistanceFilter(Algorithm): """I...
"""Perform polygon refinement by filtering out those iris/pupil polygons points which are to close to eyeball or noise. Args: polygons (GeometryPolygons): Polygons to refine. geometry_mask (NoiseMask): Geometry noise mask. Returns: GeometryPolygons: Refined geometry...
BLOCK_COMMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>open-iris/src/iris/nodes/eye_properties_estimation/iris_bbox_calculator.py<fim_prefix>from numbers import Number from typing import Tuple, Union import numpy as np from pydantic import validator from iris.io.class_configs import Algorithm from iris.io.dataclasses import BoundingBox, GeometryPolygons, IRImag...
"""Assign parameters. WARNING: Depending on wether `buffer` is a float or an int, behaviour differs. `buffer=2.0` => the iris size will be multiplied by 2., `buffer=2` => 2 pixels padding will be added. Args: buffer (Union[int, float, Tuple[Number, Number]], optional): Iris buffer ...
BLOCK_COMMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>open-iris/src/iris/nodes/normalization/common.py<fim_prefix>from typing import Tuple import numpy as np from pydantic import NonNegativeInt from iris.io.dataclasses import GeometryPolygons from iris.utils import common def generate_iris_mask(extrapolated_contours: GeometryPolygons, noise_mask: np.ndarray)...
"""Extract interpolation points coordinates. Args: image (np.ndarray): Original, not normalized image. pixel_x (float): Pixel x coordinate. pixel_y (float): Pixel y coordinate. Returns: Tuple[float, float, float, float]: Tuple with interpolation points c...
BLOCK_COMMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>open-iris/src/iris/callbacks/pipeline_trace.py<fim_prefix>from __future__ import annotations from typing import Any, Dict, Iterable, List, Optional from iris.callbacks.callback_interface import Callback from iris.io.class_configs import Algorithm from iris.orchestration.pipeline_dataclasses import PipelineN...
"""Return stored error. Returns: Optional[Exception]: error. """
BLOCK_COMMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>open-iris/src/iris/nodes/validators/cross_object_validators.py<fim_prefix>import numpy as np from pydantic import Field from iris.io.class_configs import Algorithm from iris.io.dataclasses import EyeCenters, GeometryPolygons, IRImage from iris.io.errors import ExtrapolatedPolygonsInsideImageValidatorError, E...
"""Assign parameters. Args: min_distance_to_border (float, optional): Minimum allowed distance to image boundary. Defaults to 0.0 (Eye centers can be at the image border). """
BLOCK_COMMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>open-iris/src/iris/callbacks/pipeline_trace.py<fim_prefix>from __future__ import annotations from typing import Any, Dict, Iterable, List, Optional from iris.callbacks.callback_interface import Callback from iris.io.class_configs import Algorithm from iris.orchestration.pipeline_dataclasses import PipelineN...
"""Get result_name result. Args: result_name (str): Result name. Raises: PipelineCallTraceStorageError: Raised if result_name is not found. Returns: Any: Result object. """
BLOCK_COMMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>open-iris/src/iris/nodes/geometry_refinement/smoothing.py<fim_prefix>from typing import List, Tuple import numpy as np from pydantic import Field from iris.io.class_configs import Algorithm from iris.io.dataclasses import EyeCenters, GeometryPolygons from iris.io.errors import GeometryRefinementError from i...
"""Cut contour into arcs. Args: polygon (np.ndarray): Contour polygon. center_xy (Tuple[float, float]): Polygon's center. Returns: Tuple[List[np.ndarray], int]: Tuple with: (list of list of vertices, number of gaps detected in a contour). """
BLOCK_COMMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>open-iris/src/iris/nodes/eye_properties_estimation/pupil_iris_property_calculator.py<fim_prefix>from typing import List from pydantic import Field from iris.callbacks.callback_interface import Callback from iris.io.class_configs import Algorithm from iris.io.dataclasses import EyeCenters, GeometryPolygons, ...
"""Assign parameters. Args: min_pupil_diameter (float): minimum pupil diameter. Defaults to 1.0. min_iris_diameter (float): minimum iris diameter. Defaults to 150.0. callbacks (List[Callback]): callbacks list. Defaults to []. """
BLOCK_COMMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>open-iris/src/iris/nodes/vectorization/contouring.py<fim_prefix>from typing import Callable, List import cv2 import numpy as np from pydantic import NonNegativeFloat from iris.io.class_configs import Algorithm from iris.io.dataclasses import GeometryMask, GeometryPolygons from iris.io.errors import Vectoriz...
"""Assign parameters. Args: contour_filters (List[Callable[[List[np.ndarray]], List[np.ndarray]]], optional): List of filter functions used to filter out noise in polygons. Defaults to [ContouringAlgorithm.filter_polygon_areas]. """
BLOCK_COMMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>open-iris/src/iris/nodes/iris_response/probe_schemas/regular_probe_schema.py<fim_prefix>from typing import List, Literal, Optional, Tuple, Union import numpy as np from pydantic import Field, PositiveInt, confloat, fields, validator from iris.io.errors import ProbeSchemaError from iris.nodes.iris_response.p...
phis = phis.flatten()
STATEMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>open-iris/src/iris/nodes/eye_properties_estimation/occlusion_calculator.py<fim_prefix>from typing import List, Tuple import numpy as np from pydantic import Field from iris.callbacks.callback_interface import Callback from iris.io.class_configs import Algorithm from iris.io.dataclasses import EyeCenters, Ey...
pupil_mask = common.contour_to_mask(extrapolated_polygons.pupil_array, mask_shape=(img_w, img_h))
STATEMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>open-iris/src/iris/nodes/normalization/nonlinear_normalization.py<fim_prefix>from typing import Collection, Tuple import numpy as np from pydantic import Field from iris.io.class_configs import Algorithm from iris.io.dataclasses import EyeOrientation, GeometryPolygons, IRImage, NoiseMask, NormalizedIris fro...
src_shape = src_points.shape[0:2]
STATEMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>open-iris/src/iris/utils/common.py<fim_prefix>from typing import Tuple import cv2 import numpy as np def contour_to_mask(vertices: np.ndarray, mask_shape: Tuple[int, int]) -> np.ndarray: """Generate binary mask based on polygon's vertices. Args: vertices (np.ndarray): Vertices points array...
width, height = mask_shape
STATEMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>open-iris/src/iris/nodes/eye_properties_estimation/bisectors_method.py<fim_prefix>from typing import Tuple import numpy as np from pydantic import Field from iris.io.class_configs import Algorithm from iris.io.dataclasses import EyeCenters, GeometryPolygons from iris.io.errors import EyeCentersEstimationErr...
mask = norms > min_distance_between_sector_points_in_px
STATEMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>open-iris/src/iris/nodes/iris_response/probe_schemas/regular_probe_schema.py<fim_prefix>from typing import List, Literal, Optional, Tuple, Union import numpy as np from pydantic import Field, PositiveInt, confloat, fields, validator from iris.io.errors import ProbeSchemaError from iris.nodes.iris_response.p...
phi = np.linspace(0, 1, self.params.n_cols, endpoint=False)
STATEMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>open-iris/src/iris/nodes/normalization/common.py<fim_prefix>from typing import Tuple import numpy as np from pydantic import NonNegativeInt from iris.io.dataclasses import GeometryPolygons from iris.utils import common def generate_iris_mask(extrapolated_contours: GeometryPolygons, noise_mask: np.ndarray)...
return xmin, ymin, xmax, ymax
STATEMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>open-iris/src/iris/nodes/iris_response/image_filters/gabor_filters.py<fim_prefix>from typing import Any, Dict, Tuple import numpy as np from pydantic import Field, conint, root_validator, validator import iris.io.validators as pydantic_v from iris.io.errors import ImageFilterError from iris.nodes.iris_respo...
correction_term_mean = np.mean(envelope, axis=-1)
STATEMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>open-iris/src/iris/io/validators.py<fim_prefix>from typing import Any, Callable, Dict, Iterable, List import numpy as np from pydantic import fields # ----- validators ----- def is_odd(cls: type, v: int, field: fields.ModelField) -> int: """Check that kernel size are odd numbers. Args: cl...
return values
STATEMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>open-iris/src/iris/nodes/vectorization/contouring.py<fim_prefix>from typing import Callable, List import cv2 import numpy as np from pydantic import NonNegativeFloat from iris.io.class_configs import Algorithm from iris.io.dataclasses import GeometryMask, GeometryPolygons from iris.io.errors import Vectoriz...
return filtered_polygons
STATEMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>open-iris/src/iris/nodes/iris_response_refinement/fragile_bits_refinement.py<fim_prefix>from typing import Literal, Tuple import numpy as np from pydantic import confloat from iris.io.class_configs import Algorithm from iris.io.dataclasses import IrisFilterResponse class FragileBitRefinement(Algorithm): ...
for iris_response, iris_mask in zip(iris_filter_response.iris_responses, iris_filter_response.mask_responses): if self.params.fragile_type == "cartesian": mask_value_real = np.abs(np.real(iris_response)) >= self.params.value_threshold[0] mask_value_imaginary = np.abs(np.imag(...
FOR
prefix_suffix_full_complete_current_block_no_evidence
<filename>open-iris/src/iris/nodes/geometry_refinement/smoothing.py<fim_prefix>from typing import List, Tuple import numpy as np from pydantic import Field from iris.io.class_configs import Algorithm from iris.io.dataclasses import EyeCenters, GeometryPolygons from iris.io.errors import GeometryRefinementError from i...
for i in range(-kernel_offset, kernel_offset + 1): stacked_signals.append(np.roll(signal, i))
FOR
prefix_suffix_full_complete_current_block_no_evidence
<filename>open-iris/src/iris/nodes/eye_properties_estimation/bisectors_method.py<fim_prefix>from typing import Tuple import numpy as np from pydantic import Field from iris.io.class_configs import Algorithm from iris.io.dataclasses import EyeCenters, GeometryPolygons from iris.io.errors import EyeCentersEstimationErr...
for _ in range(self.params.max_iterations): random_indices = np.random.choice(len(polygon), size=(self.params.num_bisectors, 2)) first_drawn_points = polygon[random_indices[:, 0]] second_drawn_points = polygon[random_indices[:, 1]] norms = np.linalg.norm(first_drawn_poi...
FOR
prefix_suffix_full_complete_current_block_no_evidence
<filename>open-iris/src/iris/io/dataclasses.py<fim_prefix>from __future__ import annotations from typing import Any, Dict, List, Literal, Tuple import numpy as np from pydantic import Field, NonNegativeInt, root_validator, validator from iris.io import validators as v from iris.io.class_configs import ImmutableModel...
for index, name in self.index2class.items(): if name == class_name: return index
FOR
prefix_suffix_full_complete_current_block_no_evidence
<filename>open-iris/src/iris/nodes/iris_response/conv_filter_bank.py<fim_prefix>from typing import List, Tuple import numpy as np from pydantic import root_validator, validator from iris.io.class_configs import Algorithm from iris.io.dataclasses import IrisFilterResponse, NormalizedIris from iris.io.validators import...
for i_filter, i_schema in zip(self.params.filters, self.params.probe_schemas): iris_response, mask_response = self._convolve(i_filter, i_schema, normalization_output) iris_responses.append(iris_response) mask_responses.append(mask_response)
FOR
prefix_suffix_full_complete_current_block_no_evidence
<filename>open-iris/src/iris/callbacks/pipeline_trace.py<fim_prefix>from __future__ import annotations from typing import Any, Dict, Iterable, List, Optional from iris.callbacks.callback_interface import Callback from iris.io.class_configs import Algorithm from iris.orchestration.pipeline_dataclasses import PipelineN...
for result_name in self._storage.keys(): self._storage[result_name] = None
FOR
prefix_suffix_full_complete_current_block_no_evidence
<filename>open-iris/src/iris/nodes/geometry_refinement/contour_interpolation.py<fim_prefix>from typing import List import numpy as np from pydantic import Field from iris.io.class_configs import Algorithm from iris.io.dataclasses import GeometryPolygons class ContourInterpolation(Algorithm): """Implementation o...
for (x1, y1), (x2, y2), num_point in zip(previous_boundary, polygon, num_points): x.append(np.linspace(x1, x2, num=num_point, endpoint=False)) y.append(np.linspace(y1, y2, num=num_point, endpoint=False))
FOR
prefix_suffix_full_complete_current_block_no_evidence
<filename>open-iris/src/iris/nodes/normalization/perspective_normalization.py<fim_prefix>from typing import Collection, List, Tuple import cv2 import numpy as np from pydantic import Field, validator from iris.io.class_configs import Algorithm from iris.io.dataclasses import EyeOrientation, GeometryPolygons, IRImage,...
for radius in self.params.intermediate_radiuses: ring = pupil_points + radius * (iris_points - pupil_points) ring = np.vstack([ring, ring[0]]) src_points.append(ring)
FOR
prefix_suffix_full_complete_current_block_no_evidence
<filename>open-iris/src/iris/nodes/normalization/perspective_normalization.py<fim_prefix>from typing import Collection, List, Tuple import cv2 import numpy as np from pydantic import Field, validator from iris.io.class_configs import Algorithm from iris.io.dataclasses import EyeOrientation, GeometryPolygons, IRImage,...
for angle_point_idx in range(src_points.shape[1] - 1): for ring_idx in range(src_points.shape[0] - 1): current_src, current_dst = self._correspondence_rois_coords( angle_idx=angle_point_idx, ring_idx=ring_idx, src_points=src_points,...
FOR
prefix_suffix_full_complete_current_block_no_evidence
<filename>open-iris/src/iris/nodes/encoder/iris_encoder.py<fim_prefix>from typing import List import numpy as np from pydantic import Field from iris.callbacks.callback_interface import Callback from iris.io.class_configs import Algorithm from iris.io.dataclasses import IrisFilterResponse, IrisTemplate class IrisEn...
for iris_response, mask_response in zip(response.iris_responses, response.mask_responses): mask_code = mask_response >= self.params.mask_threshold iris_code = np.stack([iris_response.real > 0, iris_response.imag > 0], axis=-1) mask_code = np.stack([mask_code, mask_code], axis=-1) ...
FOR
prefix_suffix_full_complete_current_block_no_evidence
<filename>open-iris/src/iris/nodes/iris_response/image_filters/gabor_filters.py<fim_prefix>from typing import Any, Dict, Tuple import numpy as np from pydantic import Field, conint, root_validator, validator import iris.io.validators as pydantic_v from iris.io.errors import ImageFilterError from iris.nodes.iris_respo...
if self.params.dc_correction: # Step 1: calculate mean value of Gabor Wavelet g_mean = np.mean(np.real(kernel_values), axis=-1) # Step 2: define gaussian offset correction_term_mean = np.mean(envelope, axis=-1) # Step 3: substract gaussian kernel_v...
IF
prefix_suffix_full_complete_current_block_no_evidence
<filename>open-iris/src/iris/nodes/normalization/common.py<fim_prefix>from typing import Tuple import numpy as np from pydantic import NonNegativeInt from iris.io.dataclasses import GeometryPolygons from iris.utils import common def generate_iris_mask(extrapolated_contours: GeometryPolygons, noise_mask: np.ndarray)...
if xmin == xmax and not xmax == img_w - 1: xmax += 1
IF
prefix_suffix_full_complete_current_block_no_evidence
<filename>open-iris/src/iris/nodes/iris_response/probe_schemas/regular_probe_schema.py<fim_prefix>from typing import List, Literal, Optional, Tuple, Union import numpy as np from pydantic import Field, PositiveInt, confloat, fields, validator from iris.io.errors import ProbeSchemaError from iris.nodes.iris_response.p...
if isinstance(self.params.boundary_phi, List): phi = np.linspace( 0 + self.params.boundary_phi[0], 1 - self.params.boundary_phi[1], self.params.n_cols, endpoint=True )
IF
prefix_suffix_full_complete_current_block_no_evidence
<filename>open-iris/src/iris/nodes/geometry_refinement/smoothing.py<fim_prefix>from typing import List, Tuple import numpy as np from pydantic import Field from iris.io.class_configs import Algorithm from iris.io.dataclasses import EyeCenters, GeometryPolygons from iris.io.errors import GeometryRefinementError from i...
if gap_indices.size < 2: return [polygon], gap_indices.size
IF
prefix_suffix_full_complete_current_block_no_evidence
<filename>open-iris/src/iris/io/dataclasses.py<fim_prefix>from __future__ import annotations from typing import Any, Dict, List, Literal, Tuple import numpy as np from pydantic import Field, NonNegativeInt, root_validator, validator from iris.io import validators as v from iris.io.class_configs import ImmutableModel...
if name == class_name: return index
IF
prefix_suffix_full_complete_current_block_no_evidence
<filename>open-iris/src/iris/nodes/normalization/common.py<fim_prefix>from typing import Tuple import numpy as np from pydantic import NonNegativeInt from iris.io.dataclasses import GeometryPolygons from iris.utils import common def generate_iris_mask(extrapolated_contours: GeometryPolygons, noise_mask: np.ndarray)...
if xmin == xmax and not xmax == img_w - 1: xmax += 1
IF
prefix_suffix_full_complete_current_block_no_evidence
<filename>open-iris/src/iris/nodes/iris_response/probe_schemas/regular_probe_schema.py<fim_prefix>from typing import List, Literal, Optional, Tuple, Union import numpy as np from pydantic import Field, PositiveInt, confloat, fields, validator from iris.io.errors import ProbeSchemaError from iris.nodes.iris_response.p...
if self.params.boundary_phi == "periodic-left": phi = np.linspace(0, 1, self.params.n_cols, endpoint=False)
IF
prefix_suffix_full_complete_current_block_no_evidence
<filename>open-iris/src/iris/io/validators.py<fim_prefix>from typing import Any, Callable, Dict, Iterable, List import numpy as np from pydantic import fields # ----- validators ----- def is_odd(cls: type, v: int, field: fields.ModelField) -> int: """Check that kernel size are odd numbers. Args: cl...
if len(values[field1]) != len(values[field2]): raise ValueError( f"{cls.__name__}: {field1} and {field2} length mismatch, " f"resp. {len(values[field1])} and {len(values[field2])}" )
IF
prefix_suffix_full_complete_current_block_no_evidence
<filename>open-iris/src/iris/nodes/normalization/common.py<fim_prefix>from typing import Tuple import numpy as np from pydantic import NonNegativeInt from iris.io.dataclasses import GeometryPolygons from iris.utils import common def generate_iris_mask(extrapolated_contours: GeometryPolygons, noise_mask: np.ndarray)...
if ymin == ymax and not ymax == img_h - 1: ymax += 1
IF
prefix_suffix_full_complete_current_block_no_evidence
<filename>open-iris/src/iris/nodes/iris_response/image_filters/gabor_filters.py<fim_prefix>from typing import Any, Dict, Tuple import numpy as np from pydantic import Field, conint, root_validator, validator import iris.io.validators as pydantic_v from iris.io.errors import ImageFilterError from iris.nodes.iris_respo...
if norm_real > 0: kernel_values.real /= norm_real
IF
prefix_suffix_full_complete_current_block_no_evidence
<filename>open-iris/src/iris/nodes/iris_response/image_filters/gabor_filters.py<fim_prefix>from typing import Any, Dict, Tuple import numpy as np from pydantic import Field, conint, root_validator, validator import iris.io.validators as pydantic_v from iris.io.errors import ImageFilterError from iris.nodes.iris_respo...
# normalize kernel values
LINE_COMMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>open-iris/src/iris/utils/math.py<fim_prefix>import math from typing import Dict, Tuple import numpy as np def area(array: np.ndarray) -> float: """Shoelace formula for simple polygon area calculation. WARNING: This formula only works for simple polygons, i.e planar polygon without self-intersectio...
# fmt: on
LINE_COMMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>open-iris/src/iris/nodes/iris_response/image_filters/gabor_filters.py<fim_prefix>from typing import Any, Dict, Tuple import numpy as np from pydantic import Field, conint, root_validator, validator import iris.io.validators as pydantic_v from iris.io.errors import ImageFilterError from iris.nodes.iris_respo...
# calculate kernel values
LINE_COMMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>open-iris/src/iris/nodes/eye_properties_estimation/bisectors_method.py<fim_prefix>from typing import Tuple import numpy as np from pydantic import Field from iris.io.class_configs import Algorithm from iris.io.dataclasses import EyeCenters, GeometryPolygons from iris.io.errors import EyeCentersEstimationErr...
# Add perpendicular vector to center and normalize
LINE_COMMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>open-iris/src/iris/nodes/geometry_estimation/lsq_ellipse_fit_with_refinement.py<fim_prefix>from typing import List import cv2 import numpy as np from pydantic import Field from iris.callbacks.callback_interface import Callback from iris.io.class_configs import Algorithm from iris.io.dataclasses import Geome...
# Rotate such that 0 degree is parallel with x-axis and array is clockwise
LINE_COMMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>open-iris/src/iris/nodes/iris_response/image_filters/gabor_filters.py<fim_prefix>from typing import Any, Dict, Tuple import numpy as np from pydantic import Field, conint, root_validator, validator import iris.io.validators as pydantic_v from iris.io.errors import ImageFilterError from iris.nodes.iris_respo...
# convert to polar coordinates
LINE_COMMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>open-iris/src/iris/nodes/iris_response/probe_schemas/regular_probe_schema.py<fim_prefix>from typing import List, Literal, Optional, Tuple, Union import numpy as np from pydantic import Field, PositiveInt, confloat, fields, validator from iris.io.errors import ProbeSchemaError from iris.nodes.iris_response.p...
# if image_shape provided: verify that values lie on pixel values
LINE_COMMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>open-iris/src/iris/nodes/eye_properties_estimation/bisectors_method.py<fim_prefix>from typing import Tuple import numpy as np from pydantic import Field from iris.io.class_configs import Algorithm from iris.io.dataclasses import EyeCenters, GeometryPolygons from iris.io.errors import EyeCentersEstimationErr...
# Flip xs with ys and flip sign of on of them to create a 90deg rotation
LINE_COMMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>open-iris/src/iris/nodes/iris_response/image_filters/gabor_filters.py<fim_prefix>from typing import Any, Dict, Tuple import numpy as np from pydantic import Field, conint, root_validator, validator import iris.io.validators as pydantic_v from iris.io.errors import ImageFilterError from iris.nodes.iris_respo...
# Step 1: calculate mean value of Gabor Wavelet
LINE_COMMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>open-iris/src/iris/nodes/iris_response/image_filters/gabor_filters.py<fim_prefix>from typing import Any, Dict, Tuple import numpy as np from pydantic import Field, conint, root_validator, validator import iris.io.validators as pydantic_v from iris.io.errors import ImageFilterError from iris.nodes.iris_respo...
# calculate carrier and envelope
LINE_COMMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>open-iris/src/iris/nodes/normalization/perspective_normalization.py<fim_prefix>from typing import Collection, List, Tuple import cv2 import numpy as np from pydantic import Field, validator from iris.io.class_configs import Algorithm from iris.io.dataclasses import EyeOrientation, GeometryPolygons, IRImage,...
try: img_x, img_y = map(int, image_xy) normalized_mask_roi[shifted_y, shifted_x] = iris_mask[img_y, img_x]
TRY
prefix_suffix_full_complete_current_block_no_evidence
<filename>open-iris/src/iris/nodes/normalization/common.py<fim_prefix>from typing import Tuple import numpy as np from pydantic import NonNegativeInt from iris.io.dataclasses import GeometryPolygons from iris.utils import common def generate_iris_mask(extrapolated_contours: GeometryPolygons, noise_mask: np.ndarray)...
try: return image[int(pixel_y), int(pixel_x)]
TRY
prefix_suffix_full_complete_current_block_no_evidence
<filename>open-iris/src/iris/nodes/normalization/common.py<fim_prefix>from typing import Tuple import numpy as np from pydantic import NonNegativeInt from iris.io.dataclasses import GeometryPolygons from iris.utils import common def generate_iris_mask(extrapolated_contours: GeometryPolygons, noise_mask: np.ndarray)...
try: return image[int(pixel_y), int(pixel_x)]
TRY
prefix_suffix_full_complete_current_block_no_evidence
<filename>open-iris/src/iris/nodes/normalization/common.py<fim_prefix>from typing import Tuple import numpy as np from pydantic import NonNegativeInt from iris.io.dataclasses import GeometryPolygons from iris.utils import common def generate_iris_mask(extrapolated_contours: GeometryPolygons, noise_mask: np.ndarray)...
except IndexError: return 0.0
CATCH
prefix_suffix_full_complete_current_block_no_evidence
<filename>open-iris/src/iris/nodes/normalization/perspective_normalization.py<fim_prefix>from typing import Collection, List, Tuple import cv2 import numpy as np from pydantic import Field, validator from iris.io.class_configs import Algorithm from iris.io.dataclasses import EyeOrientation, GeometryPolygons, IRImage,...
except IndexError: normalized_mask_roi[shifted_y, shifted_x] = False
CATCH
prefix_suffix_full_complete_current_block_no_evidence
<filename>open-iris/src/iris/nodes/normalization/common.py<fim_prefix>from typing import Tuple import numpy as np from pydantic import NonNegativeInt from iris.io.dataclasses import GeometryPolygons from iris.utils import common def generate_iris_mask(extrapolated_contours: GeometryPolygons, noise_mask: np.ndarray)...
except IndexError: return 0.0
CATCH
prefix_suffix_full_complete_current_block_no_evidence
<filename>open-iris/src/iris/nodes/normalization/common.py<fim_prefix>from typing import Tuple import numpy as np from pydantic import NonNegativeInt from iris.io.dataclasses import GeometryPolygons from iris.utils import common def generate_iris_mask(extrapolated_contours: GeometryPolygons, noise_mask: np.ndarray)...
def get_pixel_intensity(image: np.ndarray, pixel_x: float, pixel_y: float) -> float: """Get the intensity value of a pixel from an intensity image. Args: image (np.ndarray): Intensity image. pixel_x (float): Pixel x coordinate. pixel_y (float): Pixel y coordinate. ...
METHOD
prefix_suffix_full_complete_current_block_no_evidence
<filename>open-iris/src/iris/nodes/normalization/common.py<fim_prefix>from typing import Tuple import numpy as np from pydantic import NonNegativeInt from iris.io.dataclasses import GeometryPolygons from iris.utils import common def generate_iris_mask(extrapolated_contours: GeometryPolygons, noise_mask: np.ndarray)...
def get_interpolation_points_coords( image: np.ndarray, pixel_x: float, pixel_y: float ) -> Tuple[float, float, float, float]: """Extract interpolation points coordinates. Args: image (np.ndarray): Original, not normalized image. pixel_x (float): Pixel x coordinate. ...
METHOD
prefix_suffix_full_complete_current_block_no_evidence
<filename>searcharray/searcharray/postings.py<fim_prefix>"""Tokenized, searchable text as a pandas dtype.""" import pandas as pd import numbers from pandas.api.extensions import ExtensionDtype, ExtensionArray, register_extension_dtype from pandas.api.types import is_list_like from pandas.api.extensions import take impo...
"""Return a mask on the postings array indicating which elements contain all terms."""
BLOCK_COMMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>searcharray/searcharray/utils/roaringish.py<fim_prefix>"""Roaring-ish bit array for storing sorted integers in numpy array. See - https://softwaredoug.com/blog/2024/01/21/search-array-phrase-algorithm """ import numpy as np import sortednp as snp import logging import numbers from typing import Optional, Tup...
"""Decode an encoded bit array into keys / payloads."""
BLOCK_COMMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>searcharray/searcharray/utils/roaringish.py<fim_prefix>"""Roaring-ish bit array for storing sorted integers in numpy array. See - https://softwaredoug.com/blog/2024/01/21/search-array-phrase-algorithm """ import numpy as np import sortednp as snp import logging import numbers from typing import Optional, Tup...
"""Convert keys to range or np.ndarray of uint64."""
BLOCK_COMMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>searcharray/searcharray/phrase/middle_out.py<fim_prefix>"""Encode positions in bits along with some neighboring information for wrapping. See this notebook for motivation: https://colab.research.google.com/drive/10tIEkdlCE_1J_CcgEcV0jkLfBc-0H4am?authuser=1#scrollTo=XWzy-n9dF3PG """ import numpy as np impor...
"""Count bigram matches between two encoded arrays. Returns: -------- count: number of matches per doc rhs_next: the next rhs array to continue matching """
BLOCK_COMMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>searcharray/searcharray/postings.py<fim_prefix>"""Tokenized, searchable text as a pandas dtype.""" import pandas as pd import numbers from pandas.api.extensions import ExtensionDtype, ExtensionArray, register_extension_dtype from pandas.api.types import is_list_like from pandas.api.extensions import take impo...
"""Return a boolean numpy array indicating which elements contain the given term."""
BLOCK_COMMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>searcharray/searcharray/utils/roaringish.py<fim_prefix>"""Roaring-ish bit array for storing sorted integers in numpy array. See - https://softwaredoug.com/blog/2024/01/21/search-array-phrase-algorithm """ import numpy as np import sortednp as snp import logging import numbers from typing import Optional, Tup...
"""Return payload MSBs from encoded."""
BLOCK_COMMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>searcharray/searcharray/solr.py<fim_prefix>"""Utility functions for Solr users of searcharray.""" import re import pandas as pd import numpy as np from typing import List, Optional, Dict, Tuple from searcharray.postings import SearchArray from searcharray.similarity import Similarity, default_bm25 def parse...
"""Run edismax search over dataframe with searcharray fields. Parameters ---------- q : str The query string mm : str The minimum should match spec qf : list The fields to search pf : list The fields to search for phrase matches pf2 : list The fields ...
BLOCK_COMMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>searcharray/searcharray/utils/row_viewable_matrix.py<fim_prefix>import numbers import numpy as np from searcharray.utils.mat_set import SparseMatSet from typing import Optional, Union, Dict, List def rowwise_eq(mat: SparseMatSet, other: SparseMatSet) -> Union[bool, np.ndarray]: <fim_suffix> if len(m...
"""Check equals on a row-by-row basis."""
BLOCK_COMMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>searcharray/searcharray/solr.py<fim_prefix>"""Utility functions for Solr users of searcharray.""" import re import pandas as pd import numpy as np from typing import List, Optional, Dict, Tuple from searcharray.postings import SearchArray from searcharray.similarity import Similarity, default_bm25 def parse...
"""Parse Solr's qf, pf, pf2, pf3 field boosts."""
BLOCK_COMMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>searcharray/searcharray/postings.py<fim_prefix>"""Tokenized, searchable text as a pandas dtype.""" import pandas as pd import numbers from pandas.api.extensions import ExtensionDtype, ExtensionArray, register_extension_dtype from pandas.api.types import is_list_like from pandas.api.extensions import take impo...
"""Score each doc using a similarity function. Parameters ---------- token : str or list of str of what to search (already tokenized) similarity : How to score the documents. Default is BM25. """
BLOCK_COMMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>searcharray/searcharray/phrase/middle_out.py<fim_prefix>"""Encode positions in bits along with some neighboring information for wrapping. See this notebook for motivation: https://colab.research.google.com/drive/10tIEkdlCE_1J_CcgEcV0jkLfBc-0H4am?authuser=1#scrollTo=XWzy-n9dF3PG """ import numpy as np impor...
# Find adjacent matches
LINE_COMMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>searcharray/searcharray/postings.py<fim_prefix>"""Tokenized, searchable text as a pandas dtype.""" import pandas as pd import numbers from pandas.api.extensions import ExtensionDtype, ExtensionArray, register_extension_dtype from pandas.api.types import is_list_like from pandas.api.extensions import take impo...
# When other is a scalar value
LINE_COMMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>searcharray/searcharray/postings.py<fim_prefix>"""Tokenized, searchable text as a pandas dtype.""" import pandas as pd import numbers from pandas.api.extensions import ExtensionDtype, ExtensionArray, register_extension_dtype from pandas.api.types import is_list_like from pandas.api.extensions import take impo...
# For expensive toknes, we compute doc freq first, so we
LINE_COMMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>searcharray/searcharray/solr.py<fim_prefix>"""Utility functions for Solr users of searcharray.""" import re import pandas as pd import numpy as np from typing import List, Optional, Dict, Tuple from searcharray.postings import SearchArray from searcharray.similarity import Similarity, default_bm25 def parse...
# otherwise, simple expression
LINE_COMMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>searcharray/searcharray/phrase/middle_out.py<fim_prefix>"""Encode positions in bits along with some neighboring information for wrapping. See this notebook for motivation: https://colab.research.google.com/drive/10tIEkdlCE_1J_CcgEcV0jkLfBc-0H4am?authuser=1#scrollTo=XWzy-n9dF3PG """ import numpy as np impor...
# ceiling divide
LINE_COMMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>searcharray/searcharray/postings.py<fim_prefix>"""Tokenized, searchable text as a pandas dtype.""" import pandas as pd import numbers from pandas.api.extensions import ExtensionDtype, ExtensionArray, register_extension_dtype from pandas.api.types import is_list_like from pandas.api.extensions import take impo...
# (not looking at positions, maybe we should?)
LINE_COMMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>searcharray/searcharray/similarity.py<fim_prefix>"""Similarity functions given term stats.""" from typing import Protocol import numpy as np class Similarity(Protocol): """Similarity function protocol.""" def __call__(self, term_freqs: np.ndarray, doc_freqs: np.ndarray, doc_lens: np.ndarray, ...
# Sum doc freqs
LINE_COMMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>searcharray/searcharray/postings.py<fim_prefix>"""Tokenized, searchable text as a pandas dtype.""" import pandas as pd import numbers from pandas.api.extensions import ExtensionDtype, ExtensionArray, register_extension_dtype from pandas.api.types import is_list_like from pandas.api.extensions import take impo...
# Count number of rows where the term appears
LINE_COMMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>searcharray/searcharray/phrase/middle_out.py<fim_prefix>"""Encode positions in bits along with some neighboring information for wrapping. See this notebook for motivation: https://colab.research.google.com/drive/10tIEkdlCE_1J_CcgEcV0jkLfBc-0H4am?authuser=1#scrollTo=XWzy-n9dF3PG """ import numpy as np impor...
# Combine lhs and rhs matches from two strategies
LINE_COMMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>searcharray/searcharray/utils/roaringish.py<fim_prefix>"""Roaring-ish bit array for storing sorted integers in numpy array. See - https://softwaredoug.com/blog/2024/01/21/search-array-phrase-algorithm """ import numpy as np import sortednp as snp import logging import numbers from typing import Optional, Tup...
# UNFORTUNATE COPY
LINE_COMMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>searcharray/searcharray/postings.py<fim_prefix>"""Tokenized, searchable text as a pandas dtype.""" import pandas as pd import numbers from pandas.api.extensions import ExtensionDtype, ExtensionArray, register_extension_dtype from pandas.api.types import is_list_like from pandas.api.extensions import take impo...
if slop == 1 and len(tokens) == len(set(tokens)): phrase_freqs = np.zeros(len(self)) try: doc_ids = self.term_mat.rows term_ids = [self.term_dict.get_term_id(token) for token in tokens] return self.posns.phrase_freqs(term_ids, doc_ids=doc_ids, ...
IF
prefix_suffix_full_complete_current_block_no_evidence
<filename>searcharray/searcharray/postings.py<fim_prefix>"""Tokenized, searchable text as a pandas dtype.""" import pandas as pd import numbers from pandas.api.extensions import ExtensionDtype, ExtensionArray, register_extension_dtype from pandas.api.types import is_list_like from pandas.api.extensions import take impo...
if isinstance(token, list): term_freq = self.phrase_freq(token) else: term_freq = self.termfreqs(token)
IF
prefix_suffix_full_complete_current_block_no_evidence
<filename>searcharray/searcharray/utils/roaringish.py<fim_prefix>"""Roaring-ish bit array for storing sorted integers in numpy array. See - https://softwaredoug.com/blog/2024/01/21/search-array-phrase-algorithm """ import numpy as np import sortednp as snp import logging import numbers from typing import Optional, Tup...
if isinstance(keys, numbers.Number): return np.asarray([keys], dtype=np.uint64) elif isinstance(keys, list): return np.asarray(keys, dtype=np.uint64) elif isinstance(keys, np.ndarray): return keys.astype(np.uint64) elif isinstance(keys, range) and len(keys) > 0: # UNFORTUNATE...
IF
prefix_suffix_full_complete_current_block_no_evidence
<filename>searcharray/searcharray/postings.py<fim_prefix>"""Tokenized, searchable text as a pandas dtype.""" import pandas as pd import numbers from pandas.api.extensions import ExtensionDtype, ExtensionArray, register_extension_dtype from pandas.api.types import is_list_like from pandas.api.extensions import take impo...
if isinstance(token, str): return token elif isinstance(token, list) and len(token) == 1: return token[0] elif isinstance(token, list): return token else: raise TypeError("Expected a string or list of strings for phrases")
IF
prefix_suffix_full_complete_current_block_no_evidence