id
int32
0
252k
repo
stringlengths
7
55
path
stringlengths
4
127
func_name
stringlengths
1
88
original_string
stringlengths
75
19.8k
language
stringclasses
1 value
code
stringlengths
51
19.8k
code_tokens
list
docstring
stringlengths
3
17.3k
docstring_tokens
list
sha
stringlengths
40
40
url
stringlengths
87
242
244,200
orbingol/NURBS-Python
geomdl/construct.py
extract_curves
def extract_curves(psurf, **kwargs): """ Extracts curves from a surface. The return value is a ``dict`` object containing the following keys: * ``u``: the curves which generate u-direction (or which lie on the v-direction) * ``v``: the curves which generate v-direction (or which lie on the u-direction) As an example; if a curve lies on the u-direction, then its knotvector is equal to surface's knotvector on the v-direction and vice versa. The curve extraction process can be controlled via ``extract_u`` and ``extract_v`` boolean keyword arguments. :param psurf: input surface :type psurf: abstract.Surface :return: extracted curves :rtype: dict """ if psurf.pdimension != 2: raise GeomdlException("The input should be a spline surface") if len(psurf) != 1: raise GeomdlException("Can only operate on single spline surfaces") # Get keyword arguments extract_u = kwargs.get('extract_u', True) extract_v = kwargs.get('extract_v', True) # Get data from the surface object surf_data = psurf.data rational = surf_data['rational'] degree_u = surf_data['degree'][0] degree_v = surf_data['degree'][1] kv_u = surf_data['knotvector'][0] kv_v = surf_data['knotvector'][1] size_u = surf_data['size'][0] size_v = surf_data['size'][1] cpts = surf_data['control_points'] # Determine object type obj = shortcuts.generate_curve(rational) # v-direction crvlist_v = [] if extract_v: for u in range(size_u): curve = obj.__class__() curve.degree = degree_v curve.set_ctrlpts([cpts[v + (size_v * u)] for v in range(size_v)]) curve.knotvector = kv_v crvlist_v.append(curve) # u-direction crvlist_u = [] if extract_u: for v in range(size_v): curve = obj.__class__() curve.degree = degree_u curve.set_ctrlpts([cpts[v + (size_v * u)] for u in range(size_u)]) curve.knotvector = kv_u crvlist_u.append(curve) # Return shapes as a dict object return dict(u=crvlist_u, v=crvlist_v)
python
def extract_curves(psurf, **kwargs): if psurf.pdimension != 2: raise GeomdlException("The input should be a spline surface") if len(psurf) != 1: raise GeomdlException("Can only operate on single spline surfaces") # Get keyword arguments extract_u = kwargs.get('extract_u', True) extract_v = kwargs.get('extract_v', True) # Get data from the surface object surf_data = psurf.data rational = surf_data['rational'] degree_u = surf_data['degree'][0] degree_v = surf_data['degree'][1] kv_u = surf_data['knotvector'][0] kv_v = surf_data['knotvector'][1] size_u = surf_data['size'][0] size_v = surf_data['size'][1] cpts = surf_data['control_points'] # Determine object type obj = shortcuts.generate_curve(rational) # v-direction crvlist_v = [] if extract_v: for u in range(size_u): curve = obj.__class__() curve.degree = degree_v curve.set_ctrlpts([cpts[v + (size_v * u)] for v in range(size_v)]) curve.knotvector = kv_v crvlist_v.append(curve) # u-direction crvlist_u = [] if extract_u: for v in range(size_v): curve = obj.__class__() curve.degree = degree_u curve.set_ctrlpts([cpts[v + (size_v * u)] for u in range(size_u)]) curve.knotvector = kv_u crvlist_u.append(curve) # Return shapes as a dict object return dict(u=crvlist_u, v=crvlist_v)
[ "def", "extract_curves", "(", "psurf", ",", "*", "*", "kwargs", ")", ":", "if", "psurf", ".", "pdimension", "!=", "2", ":", "raise", "GeomdlException", "(", "\"The input should be a spline surface\"", ")", "if", "len", "(", "psurf", ")", "!=", "1", ":", "r...
Extracts curves from a surface. The return value is a ``dict`` object containing the following keys: * ``u``: the curves which generate u-direction (or which lie on the v-direction) * ``v``: the curves which generate v-direction (or which lie on the u-direction) As an example; if a curve lies on the u-direction, then its knotvector is equal to surface's knotvector on the v-direction and vice versa. The curve extraction process can be controlled via ``extract_u`` and ``extract_v`` boolean keyword arguments. :param psurf: input surface :type psurf: abstract.Surface :return: extracted curves :rtype: dict
[ "Extracts", "curves", "from", "a", "surface", "." ]
b1c6a8b51cf143ff58761438e93ba6baef470627
https://github.com/orbingol/NURBS-Python/blob/b1c6a8b51cf143ff58761438e93ba6baef470627/geomdl/construct.py#L208-L270
244,201
orbingol/NURBS-Python
geomdl/construct.py
extract_surfaces
def extract_surfaces(pvol): """ Extracts surfaces from a volume. :param pvol: input volume :type pvol: abstract.Volume :return: extracted surface :rtype: dict """ if pvol.pdimension != 3: raise GeomdlException("The input should be a spline volume") if len(pvol) != 1: raise GeomdlException("Can only operate on single spline volumes") # Get data from the volume object vol_data = pvol.data rational = vol_data['rational'] degree_u = vol_data['degree'][0] degree_v = vol_data['degree'][1] degree_w = vol_data['degree'][2] kv_u = vol_data['knotvector'][0] kv_v = vol_data['knotvector'][1] kv_w = vol_data['knotvector'][2] size_u = vol_data['size'][0] size_v = vol_data['size'][1] size_w = vol_data['size'][2] cpts = vol_data['control_points'] # Determine object type obj = shortcuts.generate_surface(rational) # u-v plane surflist_uv = [] for w in range(size_w): surf = obj.__class__() surf.degree_u = degree_u surf.degree_v = degree_v surf.ctrlpts_size_u = size_u surf.ctrlpts_size_v = size_v surf.ctrlpts2d = [[cpts[v + (size_v * (u + (size_u * w)))] for v in range(size_v)] for u in range(size_u)] surf.knotvector_u = kv_u surf.knotvector_v = kv_v surflist_uv.append(surf) # u-w plane surflist_uw = [] for v in range(size_v): surf = obj.__class__() surf.degree_u = degree_u surf.degree_v = degree_w surf.ctrlpts_size_u = size_u surf.ctrlpts_size_v = size_w surf.ctrlpts2d = [[cpts[v + (size_v * (u + (size_u * w)))] for w in range(size_w)] for u in range(size_u)] surf.knotvector_u = kv_u surf.knotvector_v = kv_w surflist_uw.append(surf) # v-w plane surflist_vw = [] for u in range(size_u): surf = obj.__class__() surf.degree_u = degree_v surf.degree_v = degree_w surf.ctrlpts_size_u = size_v surf.ctrlpts_size_v = size_w surf.ctrlpts2d = [[cpts[v + (size_v * (u + (size_u * w)))] for w in range(size_w)] for v in range(size_v)] surf.knotvector_u = kv_v surf.knotvector_v = kv_w surflist_vw.append(surf) # Return shapes as a dict object return dict(uv=surflist_uv, uw=surflist_uw, vw=surflist_vw)
python
def extract_surfaces(pvol): if pvol.pdimension != 3: raise GeomdlException("The input should be a spline volume") if len(pvol) != 1: raise GeomdlException("Can only operate on single spline volumes") # Get data from the volume object vol_data = pvol.data rational = vol_data['rational'] degree_u = vol_data['degree'][0] degree_v = vol_data['degree'][1] degree_w = vol_data['degree'][2] kv_u = vol_data['knotvector'][0] kv_v = vol_data['knotvector'][1] kv_w = vol_data['knotvector'][2] size_u = vol_data['size'][0] size_v = vol_data['size'][1] size_w = vol_data['size'][2] cpts = vol_data['control_points'] # Determine object type obj = shortcuts.generate_surface(rational) # u-v plane surflist_uv = [] for w in range(size_w): surf = obj.__class__() surf.degree_u = degree_u surf.degree_v = degree_v surf.ctrlpts_size_u = size_u surf.ctrlpts_size_v = size_v surf.ctrlpts2d = [[cpts[v + (size_v * (u + (size_u * w)))] for v in range(size_v)] for u in range(size_u)] surf.knotvector_u = kv_u surf.knotvector_v = kv_v surflist_uv.append(surf) # u-w plane surflist_uw = [] for v in range(size_v): surf = obj.__class__() surf.degree_u = degree_u surf.degree_v = degree_w surf.ctrlpts_size_u = size_u surf.ctrlpts_size_v = size_w surf.ctrlpts2d = [[cpts[v + (size_v * (u + (size_u * w)))] for w in range(size_w)] for u in range(size_u)] surf.knotvector_u = kv_u surf.knotvector_v = kv_w surflist_uw.append(surf) # v-w plane surflist_vw = [] for u in range(size_u): surf = obj.__class__() surf.degree_u = degree_v surf.degree_v = degree_w surf.ctrlpts_size_u = size_v surf.ctrlpts_size_v = size_w surf.ctrlpts2d = [[cpts[v + (size_v * (u + (size_u * w)))] for w in range(size_w)] for v in range(size_v)] surf.knotvector_u = kv_v surf.knotvector_v = kv_w surflist_vw.append(surf) # Return shapes as a dict object return dict(uv=surflist_uv, uw=surflist_uw, vw=surflist_vw)
[ "def", "extract_surfaces", "(", "pvol", ")", ":", "if", "pvol", ".", "pdimension", "!=", "3", ":", "raise", "GeomdlException", "(", "\"The input should be a spline volume\"", ")", "if", "len", "(", "pvol", ")", "!=", "1", ":", "raise", "GeomdlException", "(", ...
Extracts surfaces from a volume. :param pvol: input volume :type pvol: abstract.Volume :return: extracted surface :rtype: dict
[ "Extracts", "surfaces", "from", "a", "volume", "." ]
b1c6a8b51cf143ff58761438e93ba6baef470627
https://github.com/orbingol/NURBS-Python/blob/b1c6a8b51cf143ff58761438e93ba6baef470627/geomdl/construct.py#L273-L343
244,202
orbingol/NURBS-Python
geomdl/construct.py
extract_isosurface
def extract_isosurface(pvol): """ Extracts the largest isosurface from a volume. The following example illustrates one of the usage scenarios: .. code-block:: python :linenos: from geomdl import construct, multi from geomdl.visualization import VisMPL # Assuming that "myvol" variable stores your spline volume information isosrf = construct.extract_isosurface(myvol) # Create a surface container to store extracted isosurface msurf = multi.SurfaceContainer(isosrf) # Set visualization components msurf.vis = VisMPL.VisSurface(VisMPL.VisConfig(ctrlpts=False)) # Render isosurface msurf.render() :param pvol: input volume :type pvol: abstract.Volume :return: isosurface (as a tuple of surfaces) :rtype: tuple """ if pvol.pdimension != 3: raise GeomdlException("The input should be a spline volume") if len(pvol) != 1: raise GeomdlException("Can only operate on single spline volumes") # Extract surfaces from the parametric volume isosrf = extract_surfaces(pvol) # Return the isosurface return isosrf['uv'][0], isosrf['uv'][-1], isosrf['uw'][0], isosrf['uw'][-1], isosrf['vw'][0], isosrf['vw'][-1]
python
def extract_isosurface(pvol): if pvol.pdimension != 3: raise GeomdlException("The input should be a spline volume") if len(pvol) != 1: raise GeomdlException("Can only operate on single spline volumes") # Extract surfaces from the parametric volume isosrf = extract_surfaces(pvol) # Return the isosurface return isosrf['uv'][0], isosrf['uv'][-1], isosrf['uw'][0], isosrf['uw'][-1], isosrf['vw'][0], isosrf['vw'][-1]
[ "def", "extract_isosurface", "(", "pvol", ")", ":", "if", "pvol", ".", "pdimension", "!=", "3", ":", "raise", "GeomdlException", "(", "\"The input should be a spline volume\"", ")", "if", "len", "(", "pvol", ")", "!=", "1", ":", "raise", "GeomdlException", "("...
Extracts the largest isosurface from a volume. The following example illustrates one of the usage scenarios: .. code-block:: python :linenos: from geomdl import construct, multi from geomdl.visualization import VisMPL # Assuming that "myvol" variable stores your spline volume information isosrf = construct.extract_isosurface(myvol) # Create a surface container to store extracted isosurface msurf = multi.SurfaceContainer(isosrf) # Set visualization components msurf.vis = VisMPL.VisSurface(VisMPL.VisConfig(ctrlpts=False)) # Render isosurface msurf.render() :param pvol: input volume :type pvol: abstract.Volume :return: isosurface (as a tuple of surfaces) :rtype: tuple
[ "Extracts", "the", "largest", "isosurface", "from", "a", "volume", "." ]
b1c6a8b51cf143ff58761438e93ba6baef470627
https://github.com/orbingol/NURBS-Python/blob/b1c6a8b51cf143ff58761438e93ba6baef470627/geomdl/construct.py#L346-L383
244,203
orbingol/NURBS-Python
geomdl/trimming.py
check_trim_curve
def check_trim_curve(curve, parbox, **kwargs): """ Checks if the trim curve was closed and sense was set. :param curve: trim curve :param parbox: parameter space bounding box of the underlying surface :return: a tuple containing the status of the operation and list of extra trim curves generated :rtype: tuple """ def next_idx(edge_idx, direction): tmp = edge_idx + direction if tmp < 0: return 3 if tmp > 3: return 0 return tmp # Keyword arguments tol = kwargs.get('tol', 10e-8) # First, check if the curve is closed dist = linalg.point_distance(curve.evalpts[0], curve.evalpts[-1]) if dist <= tol: # Curve is closed return detect_sense(curve, tol), [] else: # Define start and end points of the trim curve pt_start = curve.evalpts[0] pt_end = curve.evalpts[-1] # Search for intersections idx_spt = -1 idx_ept = -1 for idx in range(len(parbox) - 1): if detect_intersection(parbox[idx], parbox[idx + 1], pt_start, tol): idx_spt = idx if detect_intersection(parbox[idx], parbox[idx + 1], pt_end, tol): idx_ept = idx # Check result of the intersection if idx_spt < 0 or idx_ept < 0: # Curve does not intersect any edges of the parametric space # TODO: Extrapolate the curve using the tangent vector and find intersections return False, [] else: # Get sense of the original curve c_sense = curve.opt_get('reversed') # If sense is None, then detect sense if c_sense is None: # Get evaluated points pts = curve.evalpts num_pts = len(pts) # Find sense tmp_sense = 0 for pti in range(1, num_pts - 1): tmp_sense = detect_ccw(pts[pti - 1], pts[pti], pts[pti + 1], tol) if tmp_sense != 0: break if tmp_sense == 0: tmp_sense2 = detect_ccw(pts[int(num_pts/3)], pts[int(2*num_pts/3)], pts[-int(num_pts/3)], tol) if tmp_sense2 != 0: tmp_sense = -tmp_sense2 else: # We cannot decide which region to trim. Therefore, ignore this curve. return False, [] c_sense = 0 if tmp_sense > 0 else 1 # Update sense of the original curve curve.opt = ['reversed', c_sense] # Generate a curve container and add the original curve cont = [curve] move_dir = -1 if c_sense == 0 else 1 # Curve intersects with the edges of the parametric space counter = 0 while counter < 4: if idx_ept == idx_spt: counter = 5 pt_start = curve.evalpts[0] else: # Find next index idx_ept = next_idx(idx_ept, move_dir) # Update tracked last point pt_start = parbox[idx_ept + 1 - c_sense] # Increment counter counter += 1 # Generate the curve segment crv = shortcuts.generate_curve() crv.degree = 1 crv.ctrlpts = [pt_end, pt_start] crv.knotvector = [0, 0, 1, 1] crv.opt = ['reversed', c_sense] pt_end = pt_start # Add it to the container cont.append(crv) # Update curve return True, cont
python
def check_trim_curve(curve, parbox, **kwargs): def next_idx(edge_idx, direction): tmp = edge_idx + direction if tmp < 0: return 3 if tmp > 3: return 0 return tmp # Keyword arguments tol = kwargs.get('tol', 10e-8) # First, check if the curve is closed dist = linalg.point_distance(curve.evalpts[0], curve.evalpts[-1]) if dist <= tol: # Curve is closed return detect_sense(curve, tol), [] else: # Define start and end points of the trim curve pt_start = curve.evalpts[0] pt_end = curve.evalpts[-1] # Search for intersections idx_spt = -1 idx_ept = -1 for idx in range(len(parbox) - 1): if detect_intersection(parbox[idx], parbox[idx + 1], pt_start, tol): idx_spt = idx if detect_intersection(parbox[idx], parbox[idx + 1], pt_end, tol): idx_ept = idx # Check result of the intersection if idx_spt < 0 or idx_ept < 0: # Curve does not intersect any edges of the parametric space # TODO: Extrapolate the curve using the tangent vector and find intersections return False, [] else: # Get sense of the original curve c_sense = curve.opt_get('reversed') # If sense is None, then detect sense if c_sense is None: # Get evaluated points pts = curve.evalpts num_pts = len(pts) # Find sense tmp_sense = 0 for pti in range(1, num_pts - 1): tmp_sense = detect_ccw(pts[pti - 1], pts[pti], pts[pti + 1], tol) if tmp_sense != 0: break if tmp_sense == 0: tmp_sense2 = detect_ccw(pts[int(num_pts/3)], pts[int(2*num_pts/3)], pts[-int(num_pts/3)], tol) if tmp_sense2 != 0: tmp_sense = -tmp_sense2 else: # We cannot decide which region to trim. Therefore, ignore this curve. return False, [] c_sense = 0 if tmp_sense > 0 else 1 # Update sense of the original curve curve.opt = ['reversed', c_sense] # Generate a curve container and add the original curve cont = [curve] move_dir = -1 if c_sense == 0 else 1 # Curve intersects with the edges of the parametric space counter = 0 while counter < 4: if idx_ept == idx_spt: counter = 5 pt_start = curve.evalpts[0] else: # Find next index idx_ept = next_idx(idx_ept, move_dir) # Update tracked last point pt_start = parbox[idx_ept + 1 - c_sense] # Increment counter counter += 1 # Generate the curve segment crv = shortcuts.generate_curve() crv.degree = 1 crv.ctrlpts = [pt_end, pt_start] crv.knotvector = [0, 0, 1, 1] crv.opt = ['reversed', c_sense] pt_end = pt_start # Add it to the container cont.append(crv) # Update curve return True, cont
[ "def", "check_trim_curve", "(", "curve", ",", "parbox", ",", "*", "*", "kwargs", ")", ":", "def", "next_idx", "(", "edge_idx", ",", "direction", ")", ":", "tmp", "=", "edge_idx", "+", "direction", "if", "tmp", "<", "0", ":", "return", "3", "if", "tmp...
Checks if the trim curve was closed and sense was set. :param curve: trim curve :param parbox: parameter space bounding box of the underlying surface :return: a tuple containing the status of the operation and list of extra trim curves generated :rtype: tuple
[ "Checks", "if", "the", "trim", "curve", "was", "closed", "and", "sense", "was", "set", "." ]
b1c6a8b51cf143ff58761438e93ba6baef470627
https://github.com/orbingol/NURBS-Python/blob/b1c6a8b51cf143ff58761438e93ba6baef470627/geomdl/trimming.py#L174-L278
244,204
orbingol/NURBS-Python
geomdl/trimming.py
get_par_box
def get_par_box(domain, last=False): """ Returns the bounding box of the surface parametric domain in ccw direction. :param domain: parametric domain :type domain: list, tuple :param last: if True, adds the first vertex to the end of the return list :type last: bool :return: edges of the parametric domain :rtype: tuple """ u_range = domain[0] v_range = domain[1] verts = [(u_range[0], v_range[0]), (u_range[1], v_range[0]), (u_range[1], v_range[1]), (u_range[0], v_range[1])] if last: verts.append(verts[0]) return tuple(verts)
python
def get_par_box(domain, last=False): u_range = domain[0] v_range = domain[1] verts = [(u_range[0], v_range[0]), (u_range[1], v_range[0]), (u_range[1], v_range[1]), (u_range[0], v_range[1])] if last: verts.append(verts[0]) return tuple(verts)
[ "def", "get_par_box", "(", "domain", ",", "last", "=", "False", ")", ":", "u_range", "=", "domain", "[", "0", "]", "v_range", "=", "domain", "[", "1", "]", "verts", "=", "[", "(", "u_range", "[", "0", "]", ",", "v_range", "[", "0", "]", ")", ",...
Returns the bounding box of the surface parametric domain in ccw direction. :param domain: parametric domain :type domain: list, tuple :param last: if True, adds the first vertex to the end of the return list :type last: bool :return: edges of the parametric domain :rtype: tuple
[ "Returns", "the", "bounding", "box", "of", "the", "surface", "parametric", "domain", "in", "ccw", "direction", "." ]
b1c6a8b51cf143ff58761438e93ba6baef470627
https://github.com/orbingol/NURBS-Python/blob/b1c6a8b51cf143ff58761438e93ba6baef470627/geomdl/trimming.py#L281-L296
244,205
orbingol/NURBS-Python
geomdl/trimming.py
detect_sense
def detect_sense(curve, tol): """ Detects the sense, i.e. clockwise or counter-clockwise, of the curve. :param curve: 2-dimensional trim curve :type curve: abstract.Curve :param tol: tolerance value :type tol: float :return: True if detection is successful, False otherwise :rtype: bool """ if curve.opt_get('reversed') is None: # Detect sense since it is unset pts = curve.evalpts num_pts = len(pts) for idx in range(1, num_pts - 1): sense = detect_ccw(pts[idx - 1], pts[idx], pts[idx + 1], tol) if sense < 0: # cw curve.opt = ['reversed', 0] return True elif sense > 0: # ccw curve.opt = ['reversed', 1] return True else: continue # One final test with random points to determine the orientation sense = detect_ccw(pts[int(num_pts/3)], pts[int(2*num_pts/3)], pts[-int(num_pts/3)], tol) if sense < 0: # cw curve.opt = ['reversed', 0] return True elif sense > 0: # ccw curve.opt = ['reversed', 1] return True else: # Cannot determine the sense return False else: # Don't touch the sense value as it has been already set return True
python
def detect_sense(curve, tol): if curve.opt_get('reversed') is None: # Detect sense since it is unset pts = curve.evalpts num_pts = len(pts) for idx in range(1, num_pts - 1): sense = detect_ccw(pts[idx - 1], pts[idx], pts[idx + 1], tol) if sense < 0: # cw curve.opt = ['reversed', 0] return True elif sense > 0: # ccw curve.opt = ['reversed', 1] return True else: continue # One final test with random points to determine the orientation sense = detect_ccw(pts[int(num_pts/3)], pts[int(2*num_pts/3)], pts[-int(num_pts/3)], tol) if sense < 0: # cw curve.opt = ['reversed', 0] return True elif sense > 0: # ccw curve.opt = ['reversed', 1] return True else: # Cannot determine the sense return False else: # Don't touch the sense value as it has been already set return True
[ "def", "detect_sense", "(", "curve", ",", "tol", ")", ":", "if", "curve", ".", "opt_get", "(", "'reversed'", ")", "is", "None", ":", "# Detect sense since it is unset", "pts", "=", "curve", ".", "evalpts", "num_pts", "=", "len", "(", "pts", ")", "for", "...
Detects the sense, i.e. clockwise or counter-clockwise, of the curve. :param curve: 2-dimensional trim curve :type curve: abstract.Curve :param tol: tolerance value :type tol: float :return: True if detection is successful, False otherwise :rtype: bool
[ "Detects", "the", "sense", "i", ".", "e", ".", "clockwise", "or", "counter", "-", "clockwise", "of", "the", "curve", "." ]
b1c6a8b51cf143ff58761438e93ba6baef470627
https://github.com/orbingol/NURBS-Python/blob/b1c6a8b51cf143ff58761438e93ba6baef470627/geomdl/trimming.py#L299-L336
244,206
orbingol/NURBS-Python
geomdl/ray.py
intersect
def intersect(ray1, ray2, **kwargs): """ Finds intersection of 2 rays. This functions finds the parameter values for the 1st and 2nd input rays and returns a tuple of ``(parameter for ray1, parameter for ray2, intersection status)``. ``status`` value is a enum type which reports the case which the intersection operation encounters. The intersection operation can encounter 3 different cases: * Intersecting: This is the anticipated solution. Returns ``(t1, t2, RayIntersection.INTERSECT)`` * Colinear: The rays can be parallel or coincident. Returns ``(t1, t2, RayIntersection.COLINEAR)`` * Skew: The rays are neither parallel nor intersecting. Returns ``(t1, t2, RayIntersection.SKEW)`` For the colinear case, ``t1`` and ``t2`` are the parameter values that give the starting point of the ray2 and ray1, respectively. Therefore; .. code-block:: python ray1.eval(t1) == ray2.p ray2.eval(t2) == ray1.p Please note that this operation is only implemented for 2- and 3-dimensional rays. :param ray1: 1st ray :param ray2: 2nd ray :return: a tuple of the parameter (t) for ray1 and ray2, and status of the intersection :rtype: tuple """ if not isinstance(ray1, Ray) or not isinstance(ray2, Ray): raise TypeError("The input arguments must be instances of the Ray object") if ray1.dimension != ray2.dimension: raise ValueError("Dimensions of the input rays must be the same") # Keyword arguments tol = kwargs.get('tol', 10e-17) # Call intersection method if ray1.dimension == 2: return _intersect2d(ray1, ray2, tol) elif ray1.dimension == 3: return _intersect3d(ray1, ray2, tol) else: raise NotImplementedError("Intersection operation for the current type of rays has not been implemented yet")
python
def intersect(ray1, ray2, **kwargs): if not isinstance(ray1, Ray) or not isinstance(ray2, Ray): raise TypeError("The input arguments must be instances of the Ray object") if ray1.dimension != ray2.dimension: raise ValueError("Dimensions of the input rays must be the same") # Keyword arguments tol = kwargs.get('tol', 10e-17) # Call intersection method if ray1.dimension == 2: return _intersect2d(ray1, ray2, tol) elif ray1.dimension == 3: return _intersect3d(ray1, ray2, tol) else: raise NotImplementedError("Intersection operation for the current type of rays has not been implemented yet")
[ "def", "intersect", "(", "ray1", ",", "ray2", ",", "*", "*", "kwargs", ")", ":", "if", "not", "isinstance", "(", "ray1", ",", "Ray", ")", "or", "not", "isinstance", "(", "ray2", ",", "Ray", ")", ":", "raise", "TypeError", "(", "\"The input arguments mu...
Finds intersection of 2 rays. This functions finds the parameter values for the 1st and 2nd input rays and returns a tuple of ``(parameter for ray1, parameter for ray2, intersection status)``. ``status`` value is a enum type which reports the case which the intersection operation encounters. The intersection operation can encounter 3 different cases: * Intersecting: This is the anticipated solution. Returns ``(t1, t2, RayIntersection.INTERSECT)`` * Colinear: The rays can be parallel or coincident. Returns ``(t1, t2, RayIntersection.COLINEAR)`` * Skew: The rays are neither parallel nor intersecting. Returns ``(t1, t2, RayIntersection.SKEW)`` For the colinear case, ``t1`` and ``t2`` are the parameter values that give the starting point of the ray2 and ray1, respectively. Therefore; .. code-block:: python ray1.eval(t1) == ray2.p ray2.eval(t2) == ray1.p Please note that this operation is only implemented for 2- and 3-dimensional rays. :param ray1: 1st ray :param ray2: 2nd ray :return: a tuple of the parameter (t) for ray1 and ray2, and status of the intersection :rtype: tuple
[ "Finds", "intersection", "of", "2", "rays", "." ]
b1c6a8b51cf143ff58761438e93ba6baef470627
https://github.com/orbingol/NURBS-Python/blob/b1c6a8b51cf143ff58761438e93ba6baef470627/geomdl/ray.py#L107-L150
244,207
orbingol/NURBS-Python
geomdl/freeform.py
Freeform.evaluate
def evaluate(self, **kwargs): """ Sets points that form the geometry. Keyword Arguments: * ``points``: sets the points """ self._eval_points = kwargs.get('points', self._init_array()) self._dimension = len(self._eval_points[0])
python
def evaluate(self, **kwargs): self._eval_points = kwargs.get('points', self._init_array()) self._dimension = len(self._eval_points[0])
[ "def", "evaluate", "(", "self", ",", "*", "*", "kwargs", ")", ":", "self", ".", "_eval_points", "=", "kwargs", ".", "get", "(", "'points'", ",", "self", ".", "_init_array", "(", ")", ")", "self", ".", "_dimension", "=", "len", "(", "self", ".", "_e...
Sets points that form the geometry. Keyword Arguments: * ``points``: sets the points
[ "Sets", "points", "that", "form", "the", "geometry", "." ]
b1c6a8b51cf143ff58761438e93ba6baef470627
https://github.com/orbingol/NURBS-Python/blob/b1c6a8b51cf143ff58761438e93ba6baef470627/geomdl/freeform.py#L20-L27
244,208
orbingol/NURBS-Python
geomdl/exchange_vtk.py
export_polydata
def export_polydata(obj, file_name, **kwargs): """ Exports control points or evaluated points in VTK Polydata format. Please see the following document for details: http://www.vtk.org/VTK/img/file-formats.pdf Keyword Arguments: * ``point_type``: **ctrlpts** for control points or **evalpts** for evaluated points * ``tessellate``: tessellates the points (works only for surfaces) :param obj: geometry object :type obj: abstract.SplineGeometry, multi.AbstractContainer :param file_name: output file name :type file_name: str :raises GeomdlException: an error occurred writing the file """ content = export_polydata_str(obj, **kwargs) return exch.write_file(file_name, content)
python
def export_polydata(obj, file_name, **kwargs): content = export_polydata_str(obj, **kwargs) return exch.write_file(file_name, content)
[ "def", "export_polydata", "(", "obj", ",", "file_name", ",", "*", "*", "kwargs", ")", ":", "content", "=", "export_polydata_str", "(", "obj", ",", "*", "*", "kwargs", ")", "return", "exch", ".", "write_file", "(", "file_name", ",", "content", ")" ]
Exports control points or evaluated points in VTK Polydata format. Please see the following document for details: http://www.vtk.org/VTK/img/file-formats.pdf Keyword Arguments: * ``point_type``: **ctrlpts** for control points or **evalpts** for evaluated points * ``tessellate``: tessellates the points (works only for surfaces) :param obj: geometry object :type obj: abstract.SplineGeometry, multi.AbstractContainer :param file_name: output file name :type file_name: str :raises GeomdlException: an error occurred writing the file
[ "Exports", "control", "points", "or", "evaluated", "points", "in", "VTK", "Polydata", "format", "." ]
b1c6a8b51cf143ff58761438e93ba6baef470627
https://github.com/orbingol/NURBS-Python/blob/b1c6a8b51cf143ff58761438e93ba6baef470627/geomdl/exchange_vtk.py#L125-L141
244,209
orbingol/NURBS-Python
geomdl/utilities.py
make_zigzag
def make_zigzag(points, num_cols): """ Converts linear sequence of points into a zig-zag shape. This function is designed to create input for the visualization software. It orders the points to draw a zig-zag shape which enables generating properly connected lines without any scanlines. Please see the below sketch on the functionality of the ``num_cols`` parameter:: num cols <-=============-> ------->>-------| |------<<-------| |------>>-------| -------<<-------| Please note that this function does not detect the ordering of the input points to detect the input points have already been processed to generate a zig-zag shape. :param points: list of points to be ordered :type points: list :param num_cols: number of elements in a row which the zig-zag is generated :type num_cols: int :return: re-ordered points :rtype: list """ new_points = [] points_size = len(points) forward = True idx = 0 rev_idx = -1 while idx < points_size: if forward: new_points.append(points[idx]) else: new_points.append(points[rev_idx]) rev_idx -= 1 idx += 1 if idx % num_cols == 0: forward = False if forward else True rev_idx = idx + num_cols - 1 return new_points
python
def make_zigzag(points, num_cols): new_points = [] points_size = len(points) forward = True idx = 0 rev_idx = -1 while idx < points_size: if forward: new_points.append(points[idx]) else: new_points.append(points[rev_idx]) rev_idx -= 1 idx += 1 if idx % num_cols == 0: forward = False if forward else True rev_idx = idx + num_cols - 1 return new_points
[ "def", "make_zigzag", "(", "points", ",", "num_cols", ")", ":", "new_points", "=", "[", "]", "points_size", "=", "len", "(", "points", ")", "forward", "=", "True", "idx", "=", "0", "rev_idx", "=", "-", "1", "while", "idx", "<", "points_size", ":", "i...
Converts linear sequence of points into a zig-zag shape. This function is designed to create input for the visualization software. It orders the points to draw a zig-zag shape which enables generating properly connected lines without any scanlines. Please see the below sketch on the functionality of the ``num_cols`` parameter:: num cols <-=============-> ------->>-------| |------<<-------| |------>>-------| -------<<-------| Please note that this function does not detect the ordering of the input points to detect the input points have already been processed to generate a zig-zag shape. :param points: list of points to be ordered :type points: list :param num_cols: number of elements in a row which the zig-zag is generated :type num_cols: int :return: re-ordered points :rtype: list
[ "Converts", "linear", "sequence", "of", "points", "into", "a", "zig", "-", "zag", "shape", "." ]
b1c6a8b51cf143ff58761438e93ba6baef470627
https://github.com/orbingol/NURBS-Python/blob/b1c6a8b51cf143ff58761438e93ba6baef470627/geomdl/utilities.py#L40-L80
244,210
orbingol/NURBS-Python
geomdl/utilities.py
make_quad
def make_quad(points, size_u, size_v): """ Converts linear sequence of input points into a quad structure. :param points: list of points to be ordered :type points: list, tuple :param size_v: number of elements in a row :type size_v: int :param size_u: number of elements in a column :type size_u: int :return: re-ordered points :rtype: list """ # Start with generating a zig-zag shape in row direction and then take its reverse new_points = make_zigzag(points, size_v) new_points.reverse() # Start generating a zig-zag shape in col direction forward = True for row in range(0, size_v): temp = [] for col in range(0, size_u): temp.append(points[row + (col * size_v)]) if forward: forward = False else: forward = True temp.reverse() new_points += temp return new_points
python
def make_quad(points, size_u, size_v): # Start with generating a zig-zag shape in row direction and then take its reverse new_points = make_zigzag(points, size_v) new_points.reverse() # Start generating a zig-zag shape in col direction forward = True for row in range(0, size_v): temp = [] for col in range(0, size_u): temp.append(points[row + (col * size_v)]) if forward: forward = False else: forward = True temp.reverse() new_points += temp return new_points
[ "def", "make_quad", "(", "points", ",", "size_u", ",", "size_v", ")", ":", "# Start with generating a zig-zag shape in row direction and then take its reverse", "new_points", "=", "make_zigzag", "(", "points", ",", "size_v", ")", "new_points", ".", "reverse", "(", ")", ...
Converts linear sequence of input points into a quad structure. :param points: list of points to be ordered :type points: list, tuple :param size_v: number of elements in a row :type size_v: int :param size_u: number of elements in a column :type size_u: int :return: re-ordered points :rtype: list
[ "Converts", "linear", "sequence", "of", "input", "points", "into", "a", "quad", "structure", "." ]
b1c6a8b51cf143ff58761438e93ba6baef470627
https://github.com/orbingol/NURBS-Python/blob/b1c6a8b51cf143ff58761438e93ba6baef470627/geomdl/utilities.py#L83-L112
244,211
orbingol/NURBS-Python
geomdl/utilities.py
make_quadtree
def make_quadtree(points, size_u, size_v, **kwargs): """ Generates a quadtree-like structure from surface control points. This function generates a 2-dimensional list of control point coordinates. Considering the object-oriented representation of a quadtree data structure, first dimension of the generated list corresponds to a list of *QuadTree* classes. Second dimension of the generated list corresponds to a *QuadTree* data structure. The first element of the 2nd dimension is the mid-point of the bounding box and the remaining elements are corner points of the bounding box organized in counter-clockwise order. To maintain stability for the data structure on the edges and corners, the function accepts ``extrapolate`` keyword argument. If it is *True*, then the function extrapolates the surface on the corners and edges to complete the quad-like structure for each control point. If it is *False*, no extrapolation will be applied. By default, ``extrapolate`` is set to *True*. Please note that this function's intention is not generating a real quadtree structure but reorganizing the control points in a very similar fashion to make them available for various geometric operations. :param points: 1-dimensional array of surface control points :type points: list, tuple :param size_u: number of control points on the u-direction :type size_u: int :param size_v: number of control points on the v-direction :type size_v: int :return: control points organized in a quadtree-like structure :rtype: tuple """ # Get keyword arguments extrapolate = kwargs.get('extrapolate', True) # Convert control points array into 2-dimensional form points2d = [] for i in range(0, size_u): row_list = [] for j in range(0, size_v): row_list.append(points[j + (i * size_v)]) points2d.append(row_list) # Traverse 2-dimensional control points to find neighbors qtree = [] for u in range(size_u): for v in range(size_v): temp = [points2d[u][v]] # Note: negative indexing actually works in Python, so we need explicit checking if u + 1 < size_u: temp.append(points2d[u+1][v]) else: if extrapolate: extrapolated_edge = linalg.vector_generate(points2d[u - 1][v], points2d[u][v]) translated_point = linalg.point_translate(points2d[u][v], extrapolated_edge) temp.append(translated_point) if v + 1 < size_v: temp.append(points2d[u][v+1]) else: if extrapolate: extrapolated_edge = linalg.vector_generate(points2d[u][v - 1], points2d[u][v]) translated_point = linalg.point_translate(points2d[u][v], extrapolated_edge) temp.append(translated_point) if u - 1 >= 0: temp.append(points2d[u-1][v]) else: if extrapolate: extrapolated_edge = linalg.vector_generate(points2d[u + 1][v], points2d[u][v]) translated_point = linalg.point_translate(points2d[u][v], extrapolated_edge) temp.append(translated_point) if v - 1 >= 0: temp.append(points2d[u][v-1]) else: if extrapolate: extrapolated_edge = linalg.vector_generate(points2d[u][v + 1], points2d[u][v]) translated_point = linalg.point_translate(points2d[u][v], extrapolated_edge) temp.append(translated_point) qtree.append(tuple(temp)) # Return generated quad-tree return tuple(qtree)
python
def make_quadtree(points, size_u, size_v, **kwargs): # Get keyword arguments extrapolate = kwargs.get('extrapolate', True) # Convert control points array into 2-dimensional form points2d = [] for i in range(0, size_u): row_list = [] for j in range(0, size_v): row_list.append(points[j + (i * size_v)]) points2d.append(row_list) # Traverse 2-dimensional control points to find neighbors qtree = [] for u in range(size_u): for v in range(size_v): temp = [points2d[u][v]] # Note: negative indexing actually works in Python, so we need explicit checking if u + 1 < size_u: temp.append(points2d[u+1][v]) else: if extrapolate: extrapolated_edge = linalg.vector_generate(points2d[u - 1][v], points2d[u][v]) translated_point = linalg.point_translate(points2d[u][v], extrapolated_edge) temp.append(translated_point) if v + 1 < size_v: temp.append(points2d[u][v+1]) else: if extrapolate: extrapolated_edge = linalg.vector_generate(points2d[u][v - 1], points2d[u][v]) translated_point = linalg.point_translate(points2d[u][v], extrapolated_edge) temp.append(translated_point) if u - 1 >= 0: temp.append(points2d[u-1][v]) else: if extrapolate: extrapolated_edge = linalg.vector_generate(points2d[u + 1][v], points2d[u][v]) translated_point = linalg.point_translate(points2d[u][v], extrapolated_edge) temp.append(translated_point) if v - 1 >= 0: temp.append(points2d[u][v-1]) else: if extrapolate: extrapolated_edge = linalg.vector_generate(points2d[u][v + 1], points2d[u][v]) translated_point = linalg.point_translate(points2d[u][v], extrapolated_edge) temp.append(translated_point) qtree.append(tuple(temp)) # Return generated quad-tree return tuple(qtree)
[ "def", "make_quadtree", "(", "points", ",", "size_u", ",", "size_v", ",", "*", "*", "kwargs", ")", ":", "# Get keyword arguments", "extrapolate", "=", "kwargs", ".", "get", "(", "'extrapolate'", ",", "True", ")", "# Convert control points array into 2-dimensional fo...
Generates a quadtree-like structure from surface control points. This function generates a 2-dimensional list of control point coordinates. Considering the object-oriented representation of a quadtree data structure, first dimension of the generated list corresponds to a list of *QuadTree* classes. Second dimension of the generated list corresponds to a *QuadTree* data structure. The first element of the 2nd dimension is the mid-point of the bounding box and the remaining elements are corner points of the bounding box organized in counter-clockwise order. To maintain stability for the data structure on the edges and corners, the function accepts ``extrapolate`` keyword argument. If it is *True*, then the function extrapolates the surface on the corners and edges to complete the quad-like structure for each control point. If it is *False*, no extrapolation will be applied. By default, ``extrapolate`` is set to *True*. Please note that this function's intention is not generating a real quadtree structure but reorganizing the control points in a very similar fashion to make them available for various geometric operations. :param points: 1-dimensional array of surface control points :type points: list, tuple :param size_u: number of control points on the u-direction :type size_u: int :param size_v: number of control points on the v-direction :type size_v: int :return: control points organized in a quadtree-like structure :rtype: tuple
[ "Generates", "a", "quadtree", "-", "like", "structure", "from", "surface", "control", "points", "." ]
b1c6a8b51cf143ff58761438e93ba6baef470627
https://github.com/orbingol/NURBS-Python/blob/b1c6a8b51cf143ff58761438e93ba6baef470627/geomdl/utilities.py#L115-L189
244,212
orbingol/NURBS-Python
geomdl/utilities.py
evaluate_bounding_box
def evaluate_bounding_box(ctrlpts): """ Computes the minimum bounding box of the point set. The (minimum) bounding box is the smallest enclosure in which all the input points lie. :param ctrlpts: points :type ctrlpts: list, tuple :return: bounding box in the format [min, max] :rtype: tuple """ # Estimate dimension from the first element of the control points dimension = len(ctrlpts[0]) # Evaluate bounding box bbmin = [float('inf') for _ in range(0, dimension)] bbmax = [float('-inf') for _ in range(0, dimension)] for cpt in ctrlpts: for i, arr in enumerate(zip(cpt, bbmin)): if arr[0] < arr[1]: bbmin[i] = arr[0] for i, arr in enumerate(zip(cpt, bbmax)): if arr[0] > arr[1]: bbmax[i] = arr[0] return tuple(bbmin), tuple(bbmax)
python
def evaluate_bounding_box(ctrlpts): # Estimate dimension from the first element of the control points dimension = len(ctrlpts[0]) # Evaluate bounding box bbmin = [float('inf') for _ in range(0, dimension)] bbmax = [float('-inf') for _ in range(0, dimension)] for cpt in ctrlpts: for i, arr in enumerate(zip(cpt, bbmin)): if arr[0] < arr[1]: bbmin[i] = arr[0] for i, arr in enumerate(zip(cpt, bbmax)): if arr[0] > arr[1]: bbmax[i] = arr[0] return tuple(bbmin), tuple(bbmax)
[ "def", "evaluate_bounding_box", "(", "ctrlpts", ")", ":", "# Estimate dimension from the first element of the control points", "dimension", "=", "len", "(", "ctrlpts", "[", "0", "]", ")", "# Evaluate bounding box", "bbmin", "=", "[", "float", "(", "'inf'", ")", "for",...
Computes the minimum bounding box of the point set. The (minimum) bounding box is the smallest enclosure in which all the input points lie. :param ctrlpts: points :type ctrlpts: list, tuple :return: bounding box in the format [min, max] :rtype: tuple
[ "Computes", "the", "minimum", "bounding", "box", "of", "the", "point", "set", "." ]
b1c6a8b51cf143ff58761438e93ba6baef470627
https://github.com/orbingol/NURBS-Python/blob/b1c6a8b51cf143ff58761438e93ba6baef470627/geomdl/utilities.py#L192-L216
244,213
orbingol/NURBS-Python
geomdl/_tessellate.py
make_triangle_mesh
def make_triangle_mesh(points, size_u, size_v, **kwargs): """ Generates a triangular mesh from an array of points. This function generates a triangular mesh for a NURBS or B-Spline surface on its parametric space. The input is the surface points and the number of points on the parametric dimensions u and v, indicated as row and column sizes in the function signature. This function should operate correctly if row and column sizes are input correctly, no matter what the points are v-ordered or u-ordered. Please see the documentation of ``ctrlpts`` and ``ctrlpts2d`` properties of the Surface class for more details on point ordering for the surfaces. This function accepts the following keyword arguments: * ``vertex_spacing``: Defines the size of the triangles via setting the jump value between points * ``trims``: List of trim curves passed to the tessellation function * ``tessellate_func``: Function called for tessellation. *Default:* :func:`.tessellate.surface_tessellate` * ``tessellate_args``: Arguments passed to the tessellation function (as a dict) The tessellation function is designed to generate triangles from 4 vertices. It takes 4 :py:class:`.Vertex` objects, index values for setting the triangle and vertex IDs and additional parameters as its function arguments. It returns a tuple of :py:class:`.Vertex` and :py:class:`.Triangle` object lists generated from the input vertices. A default triangle generator is provided as a prototype for implementation in the source code. The return value of this function is a tuple containing two lists. First one is the list of vertices and the second one is the list of triangles. :param points: input points :type points: list, tuple :param size_u: number of elements on the u-direction :type size_u: int :param size_v: number of elements on the v-direction :type size_v: int :return: a tuple containing lists of vertices and triangles :rtype: tuple """ def fix_numbering(vertex_list, triangle_list): # Initialize variables final_vertices = [] # Get all vertices inside the triangle list tri_vertex_ids = [] for tri in triangle_list: for td in tri.data: if td not in tri_vertex_ids: tri_vertex_ids.append(td) # Find vertices used in triangles seen_vertices = [] for vertex in vertex_list: if vertex.id in tri_vertex_ids and vertex.id not in seen_vertices: final_vertices.append(vertex) seen_vertices.append(vertex.id) # Fix vertex numbering (automatically fixes triangle vertex numbering) vert_new_id = 0 for vertex in final_vertices: vertex.id = vert_new_id vert_new_id += 1 return final_vertices, triangle_list # Vertex spacing for triangulation vertex_spacing = kwargs.get('vertex_spacing', 1) # defines the size of the triangles trim_curves = kwargs.get('trims', []) # Tessellation algorithm tsl_func = kwargs.get('tessellate_func') if tsl_func is None: tsl_func = surface_tessellate tsl_args = kwargs.get('tessellate_args', dict()) # Numbering vrt_idx = 0 # vertex index numbering start tri_idx = 0 # triangle index numbering start # Variable initialization u_jump = (1.0 / float(size_u - 1)) * vertex_spacing # for computing vertex parametric u value v_jump = (1.0 / float(size_v - 1)) * vertex_spacing # for computing vertex parametric v value varr_size_u = int(round((float(size_u) / float(vertex_spacing)) + 10e-8)) # vertex array size on the u-direction varr_size_v = int(round((float(size_v) / float(vertex_spacing)) + 10e-8)) # vertex array size on the v-direction # Generate vertices directly from input points (preliminary evaluation) vertices = [Vertex() for _ in range(varr_size_v * varr_size_u)] u = 0.0 for i in range(0, size_u, vertex_spacing): v = 0.0 for j in range(0, size_v, vertex_spacing): idx = j + (i * size_v) vertices[vrt_idx].id = vrt_idx vertices[vrt_idx].data = points[idx] vertices[vrt_idx].uv = [u, v] vrt_idx += 1 v += v_jump u += u_jump # # Organization of vertices in a quad element on the parametric space: # # v4 v3 # o-------o i # | | | # | | | # | | |_ _ _ # o-------o j # v1 v2 # # Generate triangles and final vertices triangles = [] for i in range(varr_size_u - 1): for j in range(varr_size_v - 1): # Find vertex indices for a quad element vertex1 = vertices[j + (i * varr_size_v)] vertex2 = vertices[j + ((i + 1) * varr_size_v)] vertex3 = vertices[j + 1 + ((i + 1) * varr_size_v)] vertex4 = vertices[j + 1 + (i * varr_size_v)] # Call tessellation function vlst, tlst = tsl_func(vertex1, vertex2, vertex3, vertex4, vrt_idx, tri_idx, trim_curves, tsl_args) # Add tessellation results to the return lists vertices += vlst triangles += tlst # Increment index values vrt_idx += len(vlst) tri_idx += len(tlst) # Fix vertex and triangle numbering (ID values) vertices, triangles = fix_numbering(vertices, triangles) return vertices, triangles
python
def make_triangle_mesh(points, size_u, size_v, **kwargs): def fix_numbering(vertex_list, triangle_list): # Initialize variables final_vertices = [] # Get all vertices inside the triangle list tri_vertex_ids = [] for tri in triangle_list: for td in tri.data: if td not in tri_vertex_ids: tri_vertex_ids.append(td) # Find vertices used in triangles seen_vertices = [] for vertex in vertex_list: if vertex.id in tri_vertex_ids and vertex.id not in seen_vertices: final_vertices.append(vertex) seen_vertices.append(vertex.id) # Fix vertex numbering (automatically fixes triangle vertex numbering) vert_new_id = 0 for vertex in final_vertices: vertex.id = vert_new_id vert_new_id += 1 return final_vertices, triangle_list # Vertex spacing for triangulation vertex_spacing = kwargs.get('vertex_spacing', 1) # defines the size of the triangles trim_curves = kwargs.get('trims', []) # Tessellation algorithm tsl_func = kwargs.get('tessellate_func') if tsl_func is None: tsl_func = surface_tessellate tsl_args = kwargs.get('tessellate_args', dict()) # Numbering vrt_idx = 0 # vertex index numbering start tri_idx = 0 # triangle index numbering start # Variable initialization u_jump = (1.0 / float(size_u - 1)) * vertex_spacing # for computing vertex parametric u value v_jump = (1.0 / float(size_v - 1)) * vertex_spacing # for computing vertex parametric v value varr_size_u = int(round((float(size_u) / float(vertex_spacing)) + 10e-8)) # vertex array size on the u-direction varr_size_v = int(round((float(size_v) / float(vertex_spacing)) + 10e-8)) # vertex array size on the v-direction # Generate vertices directly from input points (preliminary evaluation) vertices = [Vertex() for _ in range(varr_size_v * varr_size_u)] u = 0.0 for i in range(0, size_u, vertex_spacing): v = 0.0 for j in range(0, size_v, vertex_spacing): idx = j + (i * size_v) vertices[vrt_idx].id = vrt_idx vertices[vrt_idx].data = points[idx] vertices[vrt_idx].uv = [u, v] vrt_idx += 1 v += v_jump u += u_jump # # Organization of vertices in a quad element on the parametric space: # # v4 v3 # o-------o i # | | | # | | | # | | |_ _ _ # o-------o j # v1 v2 # # Generate triangles and final vertices triangles = [] for i in range(varr_size_u - 1): for j in range(varr_size_v - 1): # Find vertex indices for a quad element vertex1 = vertices[j + (i * varr_size_v)] vertex2 = vertices[j + ((i + 1) * varr_size_v)] vertex3 = vertices[j + 1 + ((i + 1) * varr_size_v)] vertex4 = vertices[j + 1 + (i * varr_size_v)] # Call tessellation function vlst, tlst = tsl_func(vertex1, vertex2, vertex3, vertex4, vrt_idx, tri_idx, trim_curves, tsl_args) # Add tessellation results to the return lists vertices += vlst triangles += tlst # Increment index values vrt_idx += len(vlst) tri_idx += len(tlst) # Fix vertex and triangle numbering (ID values) vertices, triangles = fix_numbering(vertices, triangles) return vertices, triangles
[ "def", "make_triangle_mesh", "(", "points", ",", "size_u", ",", "size_v", ",", "*", "*", "kwargs", ")", ":", "def", "fix_numbering", "(", "vertex_list", ",", "triangle_list", ")", ":", "# Initialize variables", "final_vertices", "=", "[", "]", "# Get all vertice...
Generates a triangular mesh from an array of points. This function generates a triangular mesh for a NURBS or B-Spline surface on its parametric space. The input is the surface points and the number of points on the parametric dimensions u and v, indicated as row and column sizes in the function signature. This function should operate correctly if row and column sizes are input correctly, no matter what the points are v-ordered or u-ordered. Please see the documentation of ``ctrlpts`` and ``ctrlpts2d`` properties of the Surface class for more details on point ordering for the surfaces. This function accepts the following keyword arguments: * ``vertex_spacing``: Defines the size of the triangles via setting the jump value between points * ``trims``: List of trim curves passed to the tessellation function * ``tessellate_func``: Function called for tessellation. *Default:* :func:`.tessellate.surface_tessellate` * ``tessellate_args``: Arguments passed to the tessellation function (as a dict) The tessellation function is designed to generate triangles from 4 vertices. It takes 4 :py:class:`.Vertex` objects, index values for setting the triangle and vertex IDs and additional parameters as its function arguments. It returns a tuple of :py:class:`.Vertex` and :py:class:`.Triangle` object lists generated from the input vertices. A default triangle generator is provided as a prototype for implementation in the source code. The return value of this function is a tuple containing two lists. First one is the list of vertices and the second one is the list of triangles. :param points: input points :type points: list, tuple :param size_u: number of elements on the u-direction :type size_u: int :param size_v: number of elements on the v-direction :type size_v: int :return: a tuple containing lists of vertices and triangles :rtype: tuple
[ "Generates", "a", "triangular", "mesh", "from", "an", "array", "of", "points", "." ]
b1c6a8b51cf143ff58761438e93ba6baef470627
https://github.com/orbingol/NURBS-Python/blob/b1c6a8b51cf143ff58761438e93ba6baef470627/geomdl/_tessellate.py#L18-L148
244,214
orbingol/NURBS-Python
geomdl/_tessellate.py
polygon_triangulate
def polygon_triangulate(tri_idx, *args): """ Triangulates a monotone polygon defined by a list of vertices. The input vertices must form a convex polygon and must be arranged in counter-clockwise order. :param tri_idx: triangle numbering start value :type tri_idx: int :param args: list of Vertex objects :type args: Vertex :return: list of Triangle objects :rtype: list """ # Initialize variables tidx = 0 triangles = [] # Generate triangles for idx in range(1, len(args) - 1): tri = Triangle() tri.id = tri_idx + tidx tri.add_vertex(args[0], args[idx], args[idx + 1]) triangles.append(tri) tidx += 1 # Return generated triangles return triangles
python
def polygon_triangulate(tri_idx, *args): # Initialize variables tidx = 0 triangles = [] # Generate triangles for idx in range(1, len(args) - 1): tri = Triangle() tri.id = tri_idx + tidx tri.add_vertex(args[0], args[idx], args[idx + 1]) triangles.append(tri) tidx += 1 # Return generated triangles return triangles
[ "def", "polygon_triangulate", "(", "tri_idx", ",", "*", "args", ")", ":", "# Initialize variables", "tidx", "=", "0", "triangles", "=", "[", "]", "# Generate triangles", "for", "idx", "in", "range", "(", "1", ",", "len", "(", "args", ")", "-", "1", ")", ...
Triangulates a monotone polygon defined by a list of vertices. The input vertices must form a convex polygon and must be arranged in counter-clockwise order. :param tri_idx: triangle numbering start value :type tri_idx: int :param args: list of Vertex objects :type args: Vertex :return: list of Triangle objects :rtype: list
[ "Triangulates", "a", "monotone", "polygon", "defined", "by", "a", "list", "of", "vertices", "." ]
b1c6a8b51cf143ff58761438e93ba6baef470627
https://github.com/orbingol/NURBS-Python/blob/b1c6a8b51cf143ff58761438e93ba6baef470627/geomdl/_tessellate.py#L151-L176
244,215
orbingol/NURBS-Python
geomdl/_tessellate.py
make_quad_mesh
def make_quad_mesh(points, size_u, size_v): """ Generates a mesh of quadrilateral elements. :param points: list of points :type points: list, tuple :param size_u: number of points on the u-direction (column) :type size_u: int :param size_v: number of points on the v-direction (row) :type size_v: int :return: a tuple containing lists of vertices and quads :rtype: tuple """ # Numbering vertex_idx = 0 quad_idx = 0 # Generate vertices vertices = [] for pt in points: vrt = Vertex(*pt, id=vertex_idx) vertices.append(vrt) vertex_idx += 1 # Generate quads quads = [] for i in range(0, size_u - 1): for j in range(0, size_v - 1): v1 = vertices[j + (size_v * i)] v2 = vertices[j + (size_v * (i + 1))] v3 = vertices[j + 1 + (size_v * (i + 1))] v4 = vertices[j + 1 + (size_v * i)] qd = Quad(v1, v2, v3, v4, id=quad_idx) quads.append(qd) quad_idx += 1 return vertices, quads
python
def make_quad_mesh(points, size_u, size_v): # Numbering vertex_idx = 0 quad_idx = 0 # Generate vertices vertices = [] for pt in points: vrt = Vertex(*pt, id=vertex_idx) vertices.append(vrt) vertex_idx += 1 # Generate quads quads = [] for i in range(0, size_u - 1): for j in range(0, size_v - 1): v1 = vertices[j + (size_v * i)] v2 = vertices[j + (size_v * (i + 1))] v3 = vertices[j + 1 + (size_v * (i + 1))] v4 = vertices[j + 1 + (size_v * i)] qd = Quad(v1, v2, v3, v4, id=quad_idx) quads.append(qd) quad_idx += 1 return vertices, quads
[ "def", "make_quad_mesh", "(", "points", ",", "size_u", ",", "size_v", ")", ":", "# Numbering", "vertex_idx", "=", "0", "quad_idx", "=", "0", "# Generate vertices", "vertices", "=", "[", "]", "for", "pt", "in", "points", ":", "vrt", "=", "Vertex", "(", "*...
Generates a mesh of quadrilateral elements. :param points: list of points :type points: list, tuple :param size_u: number of points on the u-direction (column) :type size_u: int :param size_v: number of points on the v-direction (row) :type size_v: int :return: a tuple containing lists of vertices and quads :rtype: tuple
[ "Generates", "a", "mesh", "of", "quadrilateral", "elements", "." ]
b1c6a8b51cf143ff58761438e93ba6baef470627
https://github.com/orbingol/NURBS-Python/blob/b1c6a8b51cf143ff58761438e93ba6baef470627/geomdl/_tessellate.py#L179-L214
244,216
orbingol/NURBS-Python
geomdl/_tessellate.py
surface_tessellate
def surface_tessellate(v1, v2, v3, v4, vidx, tidx, trim_curves, tessellate_args): """ Triangular tessellation algorithm for surfaces with no trims. This function can be directly used as an input to :func:`.make_triangle_mesh` using ``tessellate_func`` keyword argument. :param v1: vertex 1 :type v1: Vertex :param v2: vertex 2 :type v2: Vertex :param v3: vertex 3 :type v3: Vertex :param v4: vertex 4 :type v4: Vertex :param vidx: vertex numbering start value :type vidx: int :param tidx: triangle numbering start value :type tidx: int :param trim_curves: trim curves :type: list, tuple :param tessellate_args: tessellation arguments :type tessellate_args: dict :return: lists of vertex and triangle objects in (vertex_list, triangle_list) format :type: tuple """ # Triangulate vertices tris = polygon_triangulate(tidx, v1, v2, v3, v4) # Return vertex and triangle lists return [], tris
python
def surface_tessellate(v1, v2, v3, v4, vidx, tidx, trim_curves, tessellate_args): # Triangulate vertices tris = polygon_triangulate(tidx, v1, v2, v3, v4) # Return vertex and triangle lists return [], tris
[ "def", "surface_tessellate", "(", "v1", ",", "v2", ",", "v3", ",", "v4", ",", "vidx", ",", "tidx", ",", "trim_curves", ",", "tessellate_args", ")", ":", "# Triangulate vertices", "tris", "=", "polygon_triangulate", "(", "tidx", ",", "v1", ",", "v2", ",", ...
Triangular tessellation algorithm for surfaces with no trims. This function can be directly used as an input to :func:`.make_triangle_mesh` using ``tessellate_func`` keyword argument. :param v1: vertex 1 :type v1: Vertex :param v2: vertex 2 :type v2: Vertex :param v3: vertex 3 :type v3: Vertex :param v4: vertex 4 :type v4: Vertex :param vidx: vertex numbering start value :type vidx: int :param tidx: triangle numbering start value :type tidx: int :param trim_curves: trim curves :type: list, tuple :param tessellate_args: tessellation arguments :type tessellate_args: dict :return: lists of vertex and triangle objects in (vertex_list, triangle_list) format :type: tuple
[ "Triangular", "tessellation", "algorithm", "for", "surfaces", "with", "no", "trims", "." ]
b1c6a8b51cf143ff58761438e93ba6baef470627
https://github.com/orbingol/NURBS-Python/blob/b1c6a8b51cf143ff58761438e93ba6baef470627/geomdl/_tessellate.py#L217-L246
244,217
Bearle/django-private-chat
django_private_chat/handlers.py
gone_online
def gone_online(stream): """ Distributes the users online status to everyone he has dialog with """ while True: packet = yield from stream.get() session_id = packet.get('session_key') if session_id: user_owner = get_user_from_session(session_id) if user_owner: logger.debug('User ' + user_owner.username + ' gone online') # find all connections including user_owner as opponent, # send them a message that the user has gone online online_opponents = list(filter(lambda x: x[1] == user_owner.username, ws_connections)) online_opponents_sockets = [ws_connections[i] for i in online_opponents] yield from fanout_message(online_opponents_sockets, {'type': 'gone-online', 'usernames': [user_owner.username]}) else: pass # invalid session id else: pass
python
def gone_online(stream): while True: packet = yield from stream.get() session_id = packet.get('session_key') if session_id: user_owner = get_user_from_session(session_id) if user_owner: logger.debug('User ' + user_owner.username + ' gone online') # find all connections including user_owner as opponent, # send them a message that the user has gone online online_opponents = list(filter(lambda x: x[1] == user_owner.username, ws_connections)) online_opponents_sockets = [ws_connections[i] for i in online_opponents] yield from fanout_message(online_opponents_sockets, {'type': 'gone-online', 'usernames': [user_owner.username]}) else: pass # invalid session id else: pass
[ "def", "gone_online", "(", "stream", ")", ":", "while", "True", ":", "packet", "=", "yield", "from", "stream", ".", "get", "(", ")", "session_id", "=", "packet", ".", "get", "(", "'session_key'", ")", "if", "session_id", ":", "user_owner", "=", "get_user...
Distributes the users online status to everyone he has dialog with
[ "Distributes", "the", "users", "online", "status", "to", "everyone", "he", "has", "dialog", "with" ]
5b51e65875795c5c0ce21bb631c53bd3aac4c26b
https://github.com/Bearle/django-private-chat/blob/5b51e65875795c5c0ce21bb631c53bd3aac4c26b/django_private_chat/handlers.py#L40-L60
244,218
Bearle/django-private-chat
django_private_chat/handlers.py
new_messages_handler
def new_messages_handler(stream): """ Saves a new chat message to db and distributes msg to connected users """ # TODO: handle no user found exception while True: packet = yield from stream.get() session_id = packet.get('session_key') msg = packet.get('message') username_opponent = packet.get('username') if session_id and msg and username_opponent: user_owner = get_user_from_session(session_id) if user_owner: user_opponent = get_user_model().objects.get(username=username_opponent) dialog = get_dialogs_with_user(user_owner, user_opponent) if len(dialog) > 0: # Save the message msg = models.Message.objects.create( dialog=dialog[0], sender=user_owner, text=packet['message'], read=False ) packet['created'] = msg.get_formatted_create_datetime() packet['sender_name'] = msg.sender.username packet['message_id'] = msg.id # Send the message connections = [] # Find socket of the user which sent the message if (user_owner.username, user_opponent.username) in ws_connections: connections.append(ws_connections[(user_owner.username, user_opponent.username)]) # Find socket of the opponent if (user_opponent.username, user_owner.username) in ws_connections: connections.append(ws_connections[(user_opponent.username, user_owner.username)]) else: # Find sockets of people who the opponent is talking with opponent_connections = list(filter(lambda x: x[0] == user_opponent.username, ws_connections)) opponent_connections_sockets = [ws_connections[i] for i in opponent_connections] connections.extend(opponent_connections_sockets) yield from fanout_message(connections, packet) else: pass # no dialog found else: pass # no user_owner else: pass
python
def new_messages_handler(stream): # TODO: handle no user found exception while True: packet = yield from stream.get() session_id = packet.get('session_key') msg = packet.get('message') username_opponent = packet.get('username') if session_id and msg and username_opponent: user_owner = get_user_from_session(session_id) if user_owner: user_opponent = get_user_model().objects.get(username=username_opponent) dialog = get_dialogs_with_user(user_owner, user_opponent) if len(dialog) > 0: # Save the message msg = models.Message.objects.create( dialog=dialog[0], sender=user_owner, text=packet['message'], read=False ) packet['created'] = msg.get_formatted_create_datetime() packet['sender_name'] = msg.sender.username packet['message_id'] = msg.id # Send the message connections = [] # Find socket of the user which sent the message if (user_owner.username, user_opponent.username) in ws_connections: connections.append(ws_connections[(user_owner.username, user_opponent.username)]) # Find socket of the opponent if (user_opponent.username, user_owner.username) in ws_connections: connections.append(ws_connections[(user_opponent.username, user_owner.username)]) else: # Find sockets of people who the opponent is talking with opponent_connections = list(filter(lambda x: x[0] == user_opponent.username, ws_connections)) opponent_connections_sockets = [ws_connections[i] for i in opponent_connections] connections.extend(opponent_connections_sockets) yield from fanout_message(connections, packet) else: pass # no dialog found else: pass # no user_owner else: pass
[ "def", "new_messages_handler", "(", "stream", ")", ":", "# TODO: handle no user found exception\r", "while", "True", ":", "packet", "=", "yield", "from", "stream", ".", "get", "(", ")", "session_id", "=", "packet", ".", "get", "(", "'session_key'", ")", "msg", ...
Saves a new chat message to db and distributes msg to connected users
[ "Saves", "a", "new", "chat", "message", "to", "db", "and", "distributes", "msg", "to", "connected", "users" ]
5b51e65875795c5c0ce21bb631c53bd3aac4c26b
https://github.com/Bearle/django-private-chat/blob/5b51e65875795c5c0ce21bb631c53bd3aac4c26b/django_private_chat/handlers.py#L118-L165
244,219
Bearle/django-private-chat
django_private_chat/handlers.py
users_changed_handler
def users_changed_handler(stream): """ Sends connected client list of currently active users in the chatroom """ while True: yield from stream.get() # Get list list of current active users users = [ {'username': username, 'uuid': uuid_str} for username, uuid_str in ws_connections.values() ] # Make packet with list of new users (sorted by username) packet = { 'type': 'users-changed', 'value': sorted(users, key=lambda i: i['username']) } logger.debug(packet) yield from fanout_message(ws_connections.keys(), packet)
python
def users_changed_handler(stream): while True: yield from stream.get() # Get list list of current active users users = [ {'username': username, 'uuid': uuid_str} for username, uuid_str in ws_connections.values() ] # Make packet with list of new users (sorted by username) packet = { 'type': 'users-changed', 'value': sorted(users, key=lambda i: i['username']) } logger.debug(packet) yield from fanout_message(ws_connections.keys(), packet)
[ "def", "users_changed_handler", "(", "stream", ")", ":", "while", "True", ":", "yield", "from", "stream", ".", "get", "(", ")", "# Get list list of current active users\r", "users", "=", "[", "{", "'username'", ":", "username", ",", "'uuid'", ":", "uuid_str", ...
Sends connected client list of currently active users in the chatroom
[ "Sends", "connected", "client", "list", "of", "currently", "active", "users", "in", "the", "chatroom" ]
5b51e65875795c5c0ce21bb631c53bd3aac4c26b
https://github.com/Bearle/django-private-chat/blob/5b51e65875795c5c0ce21bb631c53bd3aac4c26b/django_private_chat/handlers.py#L169-L188
244,220
Bearle/django-private-chat
django_private_chat/handlers.py
is_typing_handler
def is_typing_handler(stream): """ Show message to opponent if user is typing message """ while True: packet = yield from stream.get() session_id = packet.get('session_key') user_opponent = packet.get('username') typing = packet.get('typing') if session_id and user_opponent and typing is not None: user_owner = get_user_from_session(session_id) if user_owner: opponent_socket = ws_connections.get((user_opponent, user_owner.username)) if typing and opponent_socket: yield from target_message(opponent_socket, {'type': 'opponent-typing', 'username': user_opponent}) else: pass # invalid session id else: pass
python
def is_typing_handler(stream): while True: packet = yield from stream.get() session_id = packet.get('session_key') user_opponent = packet.get('username') typing = packet.get('typing') if session_id and user_opponent and typing is not None: user_owner = get_user_from_session(session_id) if user_owner: opponent_socket = ws_connections.get((user_opponent, user_owner.username)) if typing and opponent_socket: yield from target_message(opponent_socket, {'type': 'opponent-typing', 'username': user_opponent}) else: pass # invalid session id else: pass
[ "def", "is_typing_handler", "(", "stream", ")", ":", "while", "True", ":", "packet", "=", "yield", "from", "stream", ".", "get", "(", ")", "session_id", "=", "packet", ".", "get", "(", "'session_key'", ")", "user_opponent", "=", "packet", ".", "get", "("...
Show message to opponent if user is typing message
[ "Show", "message", "to", "opponent", "if", "user", "is", "typing", "message" ]
5b51e65875795c5c0ce21bb631c53bd3aac4c26b
https://github.com/Bearle/django-private-chat/blob/5b51e65875795c5c0ce21bb631c53bd3aac4c26b/django_private_chat/handlers.py#L192-L211
244,221
Bearle/django-private-chat
django_private_chat/handlers.py
read_message_handler
def read_message_handler(stream): """ Send message to user if the opponent has read the message """ while True: packet = yield from stream.get() session_id = packet.get('session_key') user_opponent = packet.get('username') message_id = packet.get('message_id') if session_id and user_opponent and message_id is not None: user_owner = get_user_from_session(session_id) if user_owner: message = models.Message.objects.filter(id=message_id).first() if message: message.read = True message.save() logger.debug('Message ' + str(message_id) + ' is now read') opponent_socket = ws_connections.get((user_opponent, user_owner.username)) if opponent_socket: yield from target_message(opponent_socket, {'type': 'opponent-read-message', 'username': user_opponent, 'message_id': message_id}) else: pass # message not found else: pass # invalid session id else: pass
python
def read_message_handler(stream): while True: packet = yield from stream.get() session_id = packet.get('session_key') user_opponent = packet.get('username') message_id = packet.get('message_id') if session_id and user_opponent and message_id is not None: user_owner = get_user_from_session(session_id) if user_owner: message = models.Message.objects.filter(id=message_id).first() if message: message.read = True message.save() logger.debug('Message ' + str(message_id) + ' is now read') opponent_socket = ws_connections.get((user_opponent, user_owner.username)) if opponent_socket: yield from target_message(opponent_socket, {'type': 'opponent-read-message', 'username': user_opponent, 'message_id': message_id}) else: pass # message not found else: pass # invalid session id else: pass
[ "def", "read_message_handler", "(", "stream", ")", ":", "while", "True", ":", "packet", "=", "yield", "from", "stream", ".", "get", "(", ")", "session_id", "=", "packet", ".", "get", "(", "'session_key'", ")", "user_opponent", "=", "packet", ".", "get", ...
Send message to user if the opponent has read the message
[ "Send", "message", "to", "user", "if", "the", "opponent", "has", "read", "the", "message" ]
5b51e65875795c5c0ce21bb631c53bd3aac4c26b
https://github.com/Bearle/django-private-chat/blob/5b51e65875795c5c0ce21bb631c53bd3aac4c26b/django_private_chat/handlers.py#L215-L242
244,222
Bearle/django-private-chat
django_private_chat/handlers.py
main_handler
def main_handler(websocket, path): """ An Asyncio Task is created for every new websocket client connection that is established. This coroutine listens to messages from the connected client and routes the message to the proper queue. This coroutine can be thought of as a producer. """ # Get users name from the path path = path.split('/') username = path[2] session_id = path[1] user_owner = get_user_from_session(session_id) if user_owner: user_owner = user_owner.username # Persist users connection, associate user w/a unique ID ws_connections[(user_owner, username)] = websocket # While the websocket is open, listen for incoming messages/events # if unable to listening for messages/events, then disconnect the client try: while websocket.open: data = yield from websocket.recv() if not data: continue logger.debug(data) try: yield from router.MessageRouter(data)() except Exception as e: logger.error('could not route msg', e) except websockets.exceptions.InvalidState: # User disconnected pass finally: del ws_connections[(user_owner, username)] else: logger.info("Got invalid session_id attempt to connect " + session_id)
python
def main_handler(websocket, path): # Get users name from the path path = path.split('/') username = path[2] session_id = path[1] user_owner = get_user_from_session(session_id) if user_owner: user_owner = user_owner.username # Persist users connection, associate user w/a unique ID ws_connections[(user_owner, username)] = websocket # While the websocket is open, listen for incoming messages/events # if unable to listening for messages/events, then disconnect the client try: while websocket.open: data = yield from websocket.recv() if not data: continue logger.debug(data) try: yield from router.MessageRouter(data)() except Exception as e: logger.error('could not route msg', e) except websockets.exceptions.InvalidState: # User disconnected pass finally: del ws_connections[(user_owner, username)] else: logger.info("Got invalid session_id attempt to connect " + session_id)
[ "def", "main_handler", "(", "websocket", ",", "path", ")", ":", "# Get users name from the path\r", "path", "=", "path", ".", "split", "(", "'/'", ")", "username", "=", "path", "[", "2", "]", "session_id", "=", "path", "[", "1", "]", "user_owner", "=", "...
An Asyncio Task is created for every new websocket client connection that is established. This coroutine listens to messages from the connected client and routes the message to the proper queue. This coroutine can be thought of as a producer.
[ "An", "Asyncio", "Task", "is", "created", "for", "every", "new", "websocket", "client", "connection", "that", "is", "established", ".", "This", "coroutine", "listens", "to", "messages", "from", "the", "connected", "client", "and", "routes", "the", "message", "...
5b51e65875795c5c0ce21bb631c53bd3aac4c26b
https://github.com/Bearle/django-private-chat/blob/5b51e65875795c5c0ce21bb631c53bd3aac4c26b/django_private_chat/handlers.py#L246-L283
244,223
blockstack/blockstack-core
api/search/substring_search.py
anyword_substring_search_inner
def anyword_substring_search_inner(query_word, target_words): """ return True if ANY target_word matches a query_word """ for target_word in target_words: if(target_word.startswith(query_word)): return query_word return False
python
def anyword_substring_search_inner(query_word, target_words): for target_word in target_words: if(target_word.startswith(query_word)): return query_word return False
[ "def", "anyword_substring_search_inner", "(", "query_word", ",", "target_words", ")", ":", "for", "target_word", "in", "target_words", ":", "if", "(", "target_word", ".", "startswith", "(", "query_word", ")", ")", ":", "return", "query_word", "return", "False" ]
return True if ANY target_word matches a query_word
[ "return", "True", "if", "ANY", "target_word", "matches", "a", "query_word" ]
1dcfdd39b152d29ce13e736a6a1a0981401a0505
https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/api/search/substring_search.py#L41-L50
244,224
blockstack/blockstack-core
api/search/substring_search.py
anyword_substring_search
def anyword_substring_search(target_words, query_words): """ return True if all query_words match """ matches_required = len(query_words) matches_found = 0 for query_word in query_words: reply = anyword_substring_search_inner(query_word, target_words) if reply is not False: matches_found += 1 else: # this is imp, otherwise will keep checking # when the final answer is already False return False if(matches_found == matches_required): return True else: return False
python
def anyword_substring_search(target_words, query_words): matches_required = len(query_words) matches_found = 0 for query_word in query_words: reply = anyword_substring_search_inner(query_word, target_words) if reply is not False: matches_found += 1 else: # this is imp, otherwise will keep checking # when the final answer is already False return False if(matches_found == matches_required): return True else: return False
[ "def", "anyword_substring_search", "(", "target_words", ",", "query_words", ")", ":", "matches_required", "=", "len", "(", "query_words", ")", "matches_found", "=", "0", "for", "query_word", "in", "query_words", ":", "reply", "=", "anyword_substring_search_inner", "...
return True if all query_words match
[ "return", "True", "if", "all", "query_words", "match" ]
1dcfdd39b152d29ce13e736a6a1a0981401a0505
https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/api/search/substring_search.py#L53-L76
244,225
blockstack/blockstack-core
api/search/substring_search.py
substring_search
def substring_search(query, list_of_strings, limit_results=DEFAULT_LIMIT): """ main function to call for searching """ matching = [] query_words = query.split(' ') # sort by longest word (higest probability of not finding a match) query_words.sort(key=len, reverse=True) counter = 0 for s in list_of_strings: target_words = s.split(' ') # the anyword searching function is separate if(anyword_substring_search(target_words, query_words)): matching.append(s) # limit results counter += 1 if(counter == limit_results): break return matching
python
def substring_search(query, list_of_strings, limit_results=DEFAULT_LIMIT): matching = [] query_words = query.split(' ') # sort by longest word (higest probability of not finding a match) query_words.sort(key=len, reverse=True) counter = 0 for s in list_of_strings: target_words = s.split(' ') # the anyword searching function is separate if(anyword_substring_search(target_words, query_words)): matching.append(s) # limit results counter += 1 if(counter == limit_results): break return matching
[ "def", "substring_search", "(", "query", ",", "list_of_strings", ",", "limit_results", "=", "DEFAULT_LIMIT", ")", ":", "matching", "=", "[", "]", "query_words", "=", "query", ".", "split", "(", "' '", ")", "# sort by longest word (higest probability of not finding a m...
main function to call for searching
[ "main", "function", "to", "call", "for", "searching" ]
1dcfdd39b152d29ce13e736a6a1a0981401a0505
https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/api/search/substring_search.py#L79-L105
244,226
blockstack/blockstack-core
api/search/substring_search.py
search_people_by_bio
def search_people_by_bio(query, limit_results=DEFAULT_LIMIT, index=['onename_people_index']): """ queries lucene index to find a nearest match, output is profile username """ from pyes import QueryStringQuery, ES conn = ES() q = QueryStringQuery(query, search_fields=['username', 'profile_bio'], default_operator='and') results = conn.search(query=q, size=20, indices=index) count = conn.count(query=q) count = count.count # having 'or' gives more results but results quality goes down if(count == 0): q = QueryStringQuery(query, search_fields=['username', 'profile_bio'], default_operator='or') results = conn.search(query=q, size=20, indices=index) results_list = [] counter = 0 for profile in results: username = profile['username'] results_list.append(username) counter += 1 if(counter == limit_results): break return results_list
python
def search_people_by_bio(query, limit_results=DEFAULT_LIMIT, index=['onename_people_index']): from pyes import QueryStringQuery, ES conn = ES() q = QueryStringQuery(query, search_fields=['username', 'profile_bio'], default_operator='and') results = conn.search(query=q, size=20, indices=index) count = conn.count(query=q) count = count.count # having 'or' gives more results but results quality goes down if(count == 0): q = QueryStringQuery(query, search_fields=['username', 'profile_bio'], default_operator='or') results = conn.search(query=q, size=20, indices=index) results_list = [] counter = 0 for profile in results: username = profile['username'] results_list.append(username) counter += 1 if(counter == limit_results): break return results_list
[ "def", "search_people_by_bio", "(", "query", ",", "limit_results", "=", "DEFAULT_LIMIT", ",", "index", "=", "[", "'onename_people_index'", "]", ")", ":", "from", "pyes", "import", "QueryStringQuery", ",", "ES", "conn", "=", "ES", "(", ")", "q", "=", "QuerySt...
queries lucene index to find a nearest match, output is profile username
[ "queries", "lucene", "index", "to", "find", "a", "nearest", "match", "output", "is", "profile", "username" ]
1dcfdd39b152d29ce13e736a6a1a0981401a0505
https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/api/search/substring_search.py#L172-L210
244,227
blockstack/blockstack-core
api/search/substring_search.py
order_search_results
def order_search_results(query, search_results): """ order of results should be a) query in first name, b) query in last name """ results = search_results results_names = [] old_query = query query = query.split(' ') first_word = '' second_word = '' third_word = '' if(len(query) < 2): first_word = old_query else: first_word = query[0] second_word = query[1] if(len(query) > 2): third_word = query[2] # save results for multiple passes results_second = [] results_third = [] for result in results: result_list = result.split(' ') try: if(result_list[0].startswith(first_word)): results_names.append(result) else: results_second.append(result) except: results_second.append(result) for result in results_second: result_list = result.split(' ') try: if(result_list[1].startswith(first_word)): results_names.append(result) else: results_third.append(result) except: results_third.append(result) # results are either in results_names (filtered) # or unprocessed in results_third (last pass) return results_names + results_third
python
def order_search_results(query, search_results): results = search_results results_names = [] old_query = query query = query.split(' ') first_word = '' second_word = '' third_word = '' if(len(query) < 2): first_word = old_query else: first_word = query[0] second_word = query[1] if(len(query) > 2): third_word = query[2] # save results for multiple passes results_second = [] results_third = [] for result in results: result_list = result.split(' ') try: if(result_list[0].startswith(first_word)): results_names.append(result) else: results_second.append(result) except: results_second.append(result) for result in results_second: result_list = result.split(' ') try: if(result_list[1].startswith(first_word)): results_names.append(result) else: results_third.append(result) except: results_third.append(result) # results are either in results_names (filtered) # or unprocessed in results_third (last pass) return results_names + results_third
[ "def", "order_search_results", "(", "query", ",", "search_results", ")", ":", "results", "=", "search_results", "results_names", "=", "[", "]", "old_query", "=", "query", "query", "=", "query", ".", "split", "(", "' '", ")", "first_word", "=", "''", "second_...
order of results should be a) query in first name, b) query in last name
[ "order", "of", "results", "should", "be", "a", ")", "query", "in", "first", "name", "b", ")", "query", "in", "last", "name" ]
1dcfdd39b152d29ce13e736a6a1a0981401a0505
https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/api/search/substring_search.py#L242-L295
244,228
blockstack/blockstack-core
blockstack/lib/storage/auth.py
get_data_hash
def get_data_hash(data_txt): """ Generate a hash over data for immutable storage. Return the hex string. """ h = hashlib.sha256() h.update(data_txt) return h.hexdigest()
python
def get_data_hash(data_txt): h = hashlib.sha256() h.update(data_txt) return h.hexdigest()
[ "def", "get_data_hash", "(", "data_txt", ")", ":", "h", "=", "hashlib", ".", "sha256", "(", ")", "h", ".", "update", "(", "data_txt", ")", "return", "h", ".", "hexdigest", "(", ")" ]
Generate a hash over data for immutable storage. Return the hex string.
[ "Generate", "a", "hash", "over", "data", "for", "immutable", "storage", ".", "Return", "the", "hex", "string", "." ]
1dcfdd39b152d29ce13e736a6a1a0981401a0505
https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/storage/auth.py#L32-L39
244,229
blockstack/blockstack-core
blockstack/lib/storage/auth.py
verify_zonefile
def verify_zonefile( zonefile_str, value_hash ): """ Verify that a zonefile hashes to the given value hash @zonefile_str must be the zonefile as a serialized string """ zonefile_hash = get_zonefile_data_hash( zonefile_str ) if zonefile_hash != value_hash: log.debug("Zonefile hash mismatch: expected %s, got %s" % (value_hash, zonefile_hash)) return False return True
python
def verify_zonefile( zonefile_str, value_hash ): zonefile_hash = get_zonefile_data_hash( zonefile_str ) if zonefile_hash != value_hash: log.debug("Zonefile hash mismatch: expected %s, got %s" % (value_hash, zonefile_hash)) return False return True
[ "def", "verify_zonefile", "(", "zonefile_str", ",", "value_hash", ")", ":", "zonefile_hash", "=", "get_zonefile_data_hash", "(", "zonefile_str", ")", "if", "zonefile_hash", "!=", "value_hash", ":", "log", ".", "debug", "(", "\"Zonefile hash mismatch: expected %s, got %s...
Verify that a zonefile hashes to the given value hash @zonefile_str must be the zonefile as a serialized string
[ "Verify", "that", "a", "zonefile", "hashes", "to", "the", "given", "value", "hash" ]
1dcfdd39b152d29ce13e736a6a1a0981401a0505
https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/storage/auth.py#L50-L60
244,230
blockstack/blockstack-core
blockstack/lib/atlas.py
atlas_peer_table_lock
def atlas_peer_table_lock(): """ Lock the global health info table. Return the table. """ global PEER_TABLE_LOCK, PEER_TABLE, PEER_TABLE_LOCK_HOLDER, PEER_TABLE_LOCK_TRACEBACK if PEER_TABLE_LOCK_HOLDER is not None: assert PEER_TABLE_LOCK_HOLDER != threading.current_thread(), "DEADLOCK" # log.warning("\n\nPossible contention: lock from %s (but held by %s at)\n%s\n\n" % (threading.current_thread(), PEER_TABLE_LOCK_HOLDER, PEER_TABLE_LOCK_TRACEBACK)) PEER_TABLE_LOCK.acquire() PEER_TABLE_LOCK_HOLDER = threading.current_thread() PEER_TABLE_LOCK_TRACEBACK = traceback.format_stack() # log.debug("\n\npeer table lock held by %s at \n%s\n\n" % (PEER_TABLE_LOCK_HOLDER, PEER_TABLE_LOCK_TRACEBACK)) return PEER_TABLE
python
def atlas_peer_table_lock(): global PEER_TABLE_LOCK, PEER_TABLE, PEER_TABLE_LOCK_HOLDER, PEER_TABLE_LOCK_TRACEBACK if PEER_TABLE_LOCK_HOLDER is not None: assert PEER_TABLE_LOCK_HOLDER != threading.current_thread(), "DEADLOCK" # log.warning("\n\nPossible contention: lock from %s (but held by %s at)\n%s\n\n" % (threading.current_thread(), PEER_TABLE_LOCK_HOLDER, PEER_TABLE_LOCK_TRACEBACK)) PEER_TABLE_LOCK.acquire() PEER_TABLE_LOCK_HOLDER = threading.current_thread() PEER_TABLE_LOCK_TRACEBACK = traceback.format_stack() # log.debug("\n\npeer table lock held by %s at \n%s\n\n" % (PEER_TABLE_LOCK_HOLDER, PEER_TABLE_LOCK_TRACEBACK)) return PEER_TABLE
[ "def", "atlas_peer_table_lock", "(", ")", ":", "global", "PEER_TABLE_LOCK", ",", "PEER_TABLE", ",", "PEER_TABLE_LOCK_HOLDER", ",", "PEER_TABLE_LOCK_TRACEBACK", "if", "PEER_TABLE_LOCK_HOLDER", "is", "not", "None", ":", "assert", "PEER_TABLE_LOCK_HOLDER", "!=", "threading",...
Lock the global health info table. Return the table.
[ "Lock", "the", "global", "health", "info", "table", ".", "Return", "the", "table", "." ]
1dcfdd39b152d29ce13e736a6a1a0981401a0505
https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/atlas.py#L352-L368
244,231
blockstack/blockstack-core
blockstack/lib/atlas.py
atlas_peer_table_unlock
def atlas_peer_table_unlock(): """ Unlock the global health info table. """ global PEER_TABLE_LOCK, PEER_TABLE_LOCK_HOLDER, PEER_TABLE_LOCK_TRACEBACK try: assert PEER_TABLE_LOCK_HOLDER == threading.current_thread() except: log.error("Locked by %s, unlocked by %s" % (PEER_TABLE_LOCK_HOLDER, threading.current_thread())) log.error("Holder locked from:\n%s" % "".join(PEER_TABLE_LOCK_TRACEBACK)) log.error("Errant thread unlocked from:\n%s" % "".join(traceback.format_stack())) os.abort() # log.debug("\n\npeer table lock released by %s at \n%s\n\n" % (PEER_TABLE_LOCK_HOLDER, PEER_TABLE_LOCK_TRACEBACK)) PEER_TABLE_LOCK_HOLDER = None PEER_TABLE_LOCK_TRACEBACK = None PEER_TABLE_LOCK.release() return
python
def atlas_peer_table_unlock(): global PEER_TABLE_LOCK, PEER_TABLE_LOCK_HOLDER, PEER_TABLE_LOCK_TRACEBACK try: assert PEER_TABLE_LOCK_HOLDER == threading.current_thread() except: log.error("Locked by %s, unlocked by %s" % (PEER_TABLE_LOCK_HOLDER, threading.current_thread())) log.error("Holder locked from:\n%s" % "".join(PEER_TABLE_LOCK_TRACEBACK)) log.error("Errant thread unlocked from:\n%s" % "".join(traceback.format_stack())) os.abort() # log.debug("\n\npeer table lock released by %s at \n%s\n\n" % (PEER_TABLE_LOCK_HOLDER, PEER_TABLE_LOCK_TRACEBACK)) PEER_TABLE_LOCK_HOLDER = None PEER_TABLE_LOCK_TRACEBACK = None PEER_TABLE_LOCK.release() return
[ "def", "atlas_peer_table_unlock", "(", ")", ":", "global", "PEER_TABLE_LOCK", ",", "PEER_TABLE_LOCK_HOLDER", ",", "PEER_TABLE_LOCK_TRACEBACK", "try", ":", "assert", "PEER_TABLE_LOCK_HOLDER", "==", "threading", ".", "current_thread", "(", ")", "except", ":", "log", "."...
Unlock the global health info table.
[ "Unlock", "the", "global", "health", "info", "table", "." ]
1dcfdd39b152d29ce13e736a6a1a0981401a0505
https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/atlas.py#L387-L405
244,232
blockstack/blockstack-core
blockstack/lib/atlas.py
atlasdb_format_query
def atlasdb_format_query( query, values ): """ Turn a query into a string for printing. Useful for debugging. """ return "".join( ["%s %s" % (frag, "'%s'" % val if type(val) in [str, unicode] else val) for (frag, val) in zip(query.split("?"), values + ("",))] )
python
def atlasdb_format_query( query, values ): return "".join( ["%s %s" % (frag, "'%s'" % val if type(val) in [str, unicode] else val) for (frag, val) in zip(query.split("?"), values + ("",))] )
[ "def", "atlasdb_format_query", "(", "query", ",", "values", ")", ":", "return", "\"\"", ".", "join", "(", "[", "\"%s %s\"", "%", "(", "frag", ",", "\"'%s'\"", "%", "val", "if", "type", "(", "val", ")", "in", "[", "str", ",", "unicode", "]", "else", ...
Turn a query into a string for printing. Useful for debugging.
[ "Turn", "a", "query", "into", "a", "string", "for", "printing", ".", "Useful", "for", "debugging", "." ]
1dcfdd39b152d29ce13e736a6a1a0981401a0505
https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/atlas.py#L532-L537
244,233
blockstack/blockstack-core
blockstack/lib/atlas.py
atlasdb_open
def atlasdb_open( path ): """ Open the atlas db. Return a connection. Return None if it doesn't exist """ if not os.path.exists(path): log.debug("Atlas DB doesn't exist at %s" % path) return None con = sqlite3.connect( path, isolation_level=None ) con.row_factory = atlasdb_row_factory return con
python
def atlasdb_open( path ): if not os.path.exists(path): log.debug("Atlas DB doesn't exist at %s" % path) return None con = sqlite3.connect( path, isolation_level=None ) con.row_factory = atlasdb_row_factory return con
[ "def", "atlasdb_open", "(", "path", ")", ":", "if", "not", "os", ".", "path", ".", "exists", "(", "path", ")", ":", "log", ".", "debug", "(", "\"Atlas DB doesn't exist at %s\"", "%", "path", ")", "return", "None", "con", "=", "sqlite3", ".", "connect", ...
Open the atlas db. Return a connection. Return None if it doesn't exist
[ "Open", "the", "atlas", "db", ".", "Return", "a", "connection", ".", "Return", "None", "if", "it", "doesn", "t", "exist" ]
1dcfdd39b152d29ce13e736a6a1a0981401a0505
https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/atlas.py#L550-L562
244,234
blockstack/blockstack-core
blockstack/lib/atlas.py
atlasdb_add_zonefile_info
def atlasdb_add_zonefile_info( name, zonefile_hash, txid, present, tried_storage, block_height, con=None, path=None ): """ Add a zonefile to the database. Mark it as present or absent. Keep our in-RAM inventory vector up-to-date """ global ZONEFILE_INV, NUM_ZONEFILES, ZONEFILE_INV_LOCK with AtlasDBOpen( con=con, path=path ) as dbcon: with ZONEFILE_INV_LOCK: # need to lock here since someone could call atlasdb_cache_zonefile_info if present: present = 1 else: present = 0 if tried_storage: tried_storage = 1 else: tried_storage = 0 sql = "UPDATE zonefiles SET name = ?, zonefile_hash = ?, txid = ?, present = ?, tried_storage = ?, block_height = ? WHERE txid = ?;" args = (name, zonefile_hash, txid, present, tried_storage, block_height, txid ) cur = dbcon.cursor() update_res = atlasdb_query_execute( cur, sql, args ) dbcon.commit() if update_res.rowcount == 0: sql = "INSERT OR IGNORE INTO zonefiles (name, zonefile_hash, txid, present, tried_storage, block_height) VALUES (?,?,?,?,?,?);" args = (name, zonefile_hash, txid, present, tried_storage, block_height) cur = dbcon.cursor() atlasdb_query_execute( cur, sql, args ) dbcon.commit() # keep in-RAM zonefile inv coherent zfbits = atlasdb_get_zonefile_bits( zonefile_hash, con=dbcon, path=path ) inv_vec = None if ZONEFILE_INV is None: inv_vec = "" else: inv_vec = ZONEFILE_INV[:] ZONEFILE_INV = atlas_inventory_flip_zonefile_bits( inv_vec, zfbits, present ) log.debug('Set {} ({}) to {}'.format(zonefile_hash, ','.join(str(i) for i in zfbits), present)) # keep in-RAM zonefile count coherent NUM_ZONEFILES = atlasdb_zonefile_inv_length( con=dbcon, path=path ) return True
python
def atlasdb_add_zonefile_info( name, zonefile_hash, txid, present, tried_storage, block_height, con=None, path=None ): global ZONEFILE_INV, NUM_ZONEFILES, ZONEFILE_INV_LOCK with AtlasDBOpen( con=con, path=path ) as dbcon: with ZONEFILE_INV_LOCK: # need to lock here since someone could call atlasdb_cache_zonefile_info if present: present = 1 else: present = 0 if tried_storage: tried_storage = 1 else: tried_storage = 0 sql = "UPDATE zonefiles SET name = ?, zonefile_hash = ?, txid = ?, present = ?, tried_storage = ?, block_height = ? WHERE txid = ?;" args = (name, zonefile_hash, txid, present, tried_storage, block_height, txid ) cur = dbcon.cursor() update_res = atlasdb_query_execute( cur, sql, args ) dbcon.commit() if update_res.rowcount == 0: sql = "INSERT OR IGNORE INTO zonefiles (name, zonefile_hash, txid, present, tried_storage, block_height) VALUES (?,?,?,?,?,?);" args = (name, zonefile_hash, txid, present, tried_storage, block_height) cur = dbcon.cursor() atlasdb_query_execute( cur, sql, args ) dbcon.commit() # keep in-RAM zonefile inv coherent zfbits = atlasdb_get_zonefile_bits( zonefile_hash, con=dbcon, path=path ) inv_vec = None if ZONEFILE_INV is None: inv_vec = "" else: inv_vec = ZONEFILE_INV[:] ZONEFILE_INV = atlas_inventory_flip_zonefile_bits( inv_vec, zfbits, present ) log.debug('Set {} ({}) to {}'.format(zonefile_hash, ','.join(str(i) for i in zfbits), present)) # keep in-RAM zonefile count coherent NUM_ZONEFILES = atlasdb_zonefile_inv_length( con=dbcon, path=path ) return True
[ "def", "atlasdb_add_zonefile_info", "(", "name", ",", "zonefile_hash", ",", "txid", ",", "present", ",", "tried_storage", ",", "block_height", ",", "con", "=", "None", ",", "path", "=", "None", ")", ":", "global", "ZONEFILE_INV", ",", "NUM_ZONEFILES", ",", "...
Add a zonefile to the database. Mark it as present or absent. Keep our in-RAM inventory vector up-to-date
[ "Add", "a", "zonefile", "to", "the", "database", ".", "Mark", "it", "as", "present", "or", "absent", ".", "Keep", "our", "in", "-", "RAM", "inventory", "vector", "up", "-", "to", "-", "date" ]
1dcfdd39b152d29ce13e736a6a1a0981401a0505
https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/atlas.py#L565-L616
244,235
blockstack/blockstack-core
blockstack/lib/atlas.py
atlasdb_get_lastblock
def atlasdb_get_lastblock( con=None, path=None ): """ Get the highest block height in the atlas db """ row = None with AtlasDBOpen(con=con, path=path) as dbcon: sql = "SELECT MAX(block_height) FROM zonefiles;" args = () cur = dbcon.cursor() res = atlasdb_query_execute( cur, sql, args ) row = {} for r in res: row.update(r) break return row['MAX(block_height)']
python
def atlasdb_get_lastblock( con=None, path=None ): row = None with AtlasDBOpen(con=con, path=path) as dbcon: sql = "SELECT MAX(block_height) FROM zonefiles;" args = () cur = dbcon.cursor() res = atlasdb_query_execute( cur, sql, args ) row = {} for r in res: row.update(r) break return row['MAX(block_height)']
[ "def", "atlasdb_get_lastblock", "(", "con", "=", "None", ",", "path", "=", "None", ")", ":", "row", "=", "None", "with", "AtlasDBOpen", "(", "con", "=", "con", ",", "path", "=", "path", ")", "as", "dbcon", ":", "sql", "=", "\"SELECT MAX(block_height) FRO...
Get the highest block height in the atlas db
[ "Get", "the", "highest", "block", "height", "in", "the", "atlas", "db" ]
1dcfdd39b152d29ce13e736a6a1a0981401a0505
https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/atlas.py#L619-L637
244,236
blockstack/blockstack-core
blockstack/lib/atlas.py
atlasdb_get_zonefiles_missing_count_by_name
def atlasdb_get_zonefiles_missing_count_by_name(name, max_index=None, indexes_exclude=[], con=None, path=None): """ Get the number of missing zone files for a particular name, optionally up to a maximum zonefile index and optionally omitting particular zone files in the count. Returns an integer """ with AtlasDBOpen(con=con, path=path) as dbcon: sql = 'SELECT COUNT(*) FROM zonefiles WHERE name = ? AND present = 0 {} {};'.format( 'AND inv_index <= ?' if max_index is not None else '', 'AND inv_index NOT IN ({})'.format(','.join([str(int(i)) for i in indexes_exclude])) if len(indexes_exclude) > 0 else '' ) args = (name,) if max_index is not None: args += (max_index,) cur = dbcon.cursor() res = atlasdb_query_execute(cur, sql, args) for row in res: return row['COUNT(*)']
python
def atlasdb_get_zonefiles_missing_count_by_name(name, max_index=None, indexes_exclude=[], con=None, path=None): with AtlasDBOpen(con=con, path=path) as dbcon: sql = 'SELECT COUNT(*) FROM zonefiles WHERE name = ? AND present = 0 {} {};'.format( 'AND inv_index <= ?' if max_index is not None else '', 'AND inv_index NOT IN ({})'.format(','.join([str(int(i)) for i in indexes_exclude])) if len(indexes_exclude) > 0 else '' ) args = (name,) if max_index is not None: args += (max_index,) cur = dbcon.cursor() res = atlasdb_query_execute(cur, sql, args) for row in res: return row['COUNT(*)']
[ "def", "atlasdb_get_zonefiles_missing_count_by_name", "(", "name", ",", "max_index", "=", "None", ",", "indexes_exclude", "=", "[", "]", ",", "con", "=", "None", ",", "path", "=", "None", ")", ":", "with", "AtlasDBOpen", "(", "con", "=", "con", ",", "path"...
Get the number of missing zone files for a particular name, optionally up to a maximum zonefile index and optionally omitting particular zone files in the count. Returns an integer
[ "Get", "the", "number", "of", "missing", "zone", "files", "for", "a", "particular", "name", "optionally", "up", "to", "a", "maximum", "zonefile", "index", "and", "optionally", "omitting", "particular", "zone", "files", "in", "the", "count", ".", "Returns", "...
1dcfdd39b152d29ce13e736a6a1a0981401a0505
https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/atlas.py#L764-L783
244,237
blockstack/blockstack-core
blockstack/lib/atlas.py
atlasdb_get_zonefiles_by_hash
def atlasdb_get_zonefiles_by_hash(zonefile_hash, block_height=None, con=None, path=None): """ Find all instances of this zone file in the atlasdb. Optionally filter on block height Returns [{'name': ..., 'zonefile_hash': ..., 'txid': ..., 'inv_index': ..., 'block_height': ..., 'present': ..., 'tried_storage': ...}], in blockchain order Returns None if the zone file is not in the db, or if block_height is set, return None if the zone file is not at this block height. """ with AtlasDBOpen(con=con, path=path) as dbcon: sql = 'SELECT * FROM zonefiles WHERE zonefile_hash = ?' args = (zonefile_hash,) if block_height: sql += ' AND block_height = ?' args += (block_height,) sql += ' ORDER BY inv_index;' cur = dbcon.cursor() res = atlasdb_query_execute(cur, sql, args) ret = [] for zfinfo in res: row = {} row.update(zfinfo) ret.append(row) if len(ret) == 0: return None return ret
python
def atlasdb_get_zonefiles_by_hash(zonefile_hash, block_height=None, con=None, path=None): with AtlasDBOpen(con=con, path=path) as dbcon: sql = 'SELECT * FROM zonefiles WHERE zonefile_hash = ?' args = (zonefile_hash,) if block_height: sql += ' AND block_height = ?' args += (block_height,) sql += ' ORDER BY inv_index;' cur = dbcon.cursor() res = atlasdb_query_execute(cur, sql, args) ret = [] for zfinfo in res: row = {} row.update(zfinfo) ret.append(row) if len(ret) == 0: return None return ret
[ "def", "atlasdb_get_zonefiles_by_hash", "(", "zonefile_hash", ",", "block_height", "=", "None", ",", "con", "=", "None", ",", "path", "=", "None", ")", ":", "with", "AtlasDBOpen", "(", "con", "=", "con", ",", "path", "=", "path", ")", "as", "dbcon", ":",...
Find all instances of this zone file in the atlasdb. Optionally filter on block height Returns [{'name': ..., 'zonefile_hash': ..., 'txid': ..., 'inv_index': ..., 'block_height': ..., 'present': ..., 'tried_storage': ...}], in blockchain order Returns None if the zone file is not in the db, or if block_height is set, return None if the zone file is not at this block height.
[ "Find", "all", "instances", "of", "this", "zone", "file", "in", "the", "atlasdb", ".", "Optionally", "filter", "on", "block", "height" ]
1dcfdd39b152d29ce13e736a6a1a0981401a0505
https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/atlas.py#L786-L817
244,238
blockstack/blockstack-core
blockstack/lib/atlas.py
atlasdb_set_zonefile_tried_storage
def atlasdb_set_zonefile_tried_storage( zonefile_hash, tried_storage, con=None, path=None ): """ Make a note that we tried to get the zonefile from storage """ with AtlasDBOpen(con=con, path=path) as dbcon: if tried_storage: tried_storage = 1 else: tried_storage = 0 sql = "UPDATE zonefiles SET tried_storage = ? WHERE zonefile_hash = ?;" args = (tried_storage, zonefile_hash) cur = dbcon.cursor() res = atlasdb_query_execute( cur, sql, args ) dbcon.commit() return True
python
def atlasdb_set_zonefile_tried_storage( zonefile_hash, tried_storage, con=None, path=None ): with AtlasDBOpen(con=con, path=path) as dbcon: if tried_storage: tried_storage = 1 else: tried_storage = 0 sql = "UPDATE zonefiles SET tried_storage = ? WHERE zonefile_hash = ?;" args = (tried_storage, zonefile_hash) cur = dbcon.cursor() res = atlasdb_query_execute( cur, sql, args ) dbcon.commit() return True
[ "def", "atlasdb_set_zonefile_tried_storage", "(", "zonefile_hash", ",", "tried_storage", ",", "con", "=", "None", ",", "path", "=", "None", ")", ":", "with", "AtlasDBOpen", "(", "con", "=", "con", ",", "path", "=", "path", ")", "as", "dbcon", ":", "if", ...
Make a note that we tried to get the zonefile from storage
[ "Make", "a", "note", "that", "we", "tried", "to", "get", "the", "zonefile", "from", "storage" ]
1dcfdd39b152d29ce13e736a6a1a0981401a0505
https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/atlas.py#L865-L882
244,239
blockstack/blockstack-core
blockstack/lib/atlas.py
atlasdb_reset_zonefile_tried_storage
def atlasdb_reset_zonefile_tried_storage( con=None, path=None ): """ For zonefiles that we don't have, re-attempt to fetch them from storage. """ with AtlasDBOpen(con=con, path=path) as dbcon: sql = "UPDATE zonefiles SET tried_storage = ? WHERE present = ?;" args = (0, 0) cur = dbcon.cursor() res = atlasdb_query_execute( cur, sql, args ) dbcon.commit() return True
python
def atlasdb_reset_zonefile_tried_storage( con=None, path=None ): with AtlasDBOpen(con=con, path=path) as dbcon: sql = "UPDATE zonefiles SET tried_storage = ? WHERE present = ?;" args = (0, 0) cur = dbcon.cursor() res = atlasdb_query_execute( cur, sql, args ) dbcon.commit() return True
[ "def", "atlasdb_reset_zonefile_tried_storage", "(", "con", "=", "None", ",", "path", "=", "None", ")", ":", "with", "AtlasDBOpen", "(", "con", "=", "con", ",", "path", "=", "path", ")", "as", "dbcon", ":", "sql", "=", "\"UPDATE zonefiles SET tried_storage = ? ...
For zonefiles that we don't have, re-attempt to fetch them from storage.
[ "For", "zonefiles", "that", "we", "don", "t", "have", "re", "-", "attempt", "to", "fetch", "them", "from", "storage", "." ]
1dcfdd39b152d29ce13e736a6a1a0981401a0505
https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/atlas.py#L885-L899
244,240
blockstack/blockstack-core
blockstack/lib/atlas.py
atlasdb_cache_zonefile_info
def atlasdb_cache_zonefile_info( con=None, path=None ): """ Load up and cache our zonefile inventory from the database """ global ZONEFILE_INV, NUM_ZONEFILES, ZONEFILE_INV_LOCK inv = None with ZONEFILE_INV_LOCK: inv_len = atlasdb_zonefile_inv_length( con=con, path=path ) inv = atlas_make_zonefile_inventory( 0, inv_len, con=con, path=path ) ZONEFILE_INV = inv NUM_ZONEFILES = inv_len return inv
python
def atlasdb_cache_zonefile_info( con=None, path=None ): global ZONEFILE_INV, NUM_ZONEFILES, ZONEFILE_INV_LOCK inv = None with ZONEFILE_INV_LOCK: inv_len = atlasdb_zonefile_inv_length( con=con, path=path ) inv = atlas_make_zonefile_inventory( 0, inv_len, con=con, path=path ) ZONEFILE_INV = inv NUM_ZONEFILES = inv_len return inv
[ "def", "atlasdb_cache_zonefile_info", "(", "con", "=", "None", ",", "path", "=", "None", ")", ":", "global", "ZONEFILE_INV", ",", "NUM_ZONEFILES", ",", "ZONEFILE_INV_LOCK", "inv", "=", "None", "with", "ZONEFILE_INV_LOCK", ":", "inv_len", "=", "atlasdb_zonefile_inv...
Load up and cache our zonefile inventory from the database
[ "Load", "up", "and", "cache", "our", "zonefile", "inventory", "from", "the", "database" ]
1dcfdd39b152d29ce13e736a6a1a0981401a0505
https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/atlas.py#L902-L916
244,241
blockstack/blockstack-core
blockstack/lib/atlas.py
atlasdb_queue_zonefiles
def atlasdb_queue_zonefiles( con, db, start_block, zonefile_dir, recover=False, validate=True, end_block=None ): """ Queue all zonefile hashes in the BlockstackDB to the zonefile queue NOT THREAD SAFE Returns the list of zonefile infos queued, and whether or not they are present. """ # populate zonefile queue total = 0 if end_block is None: end_block = db.lastblock+1 ret = [] # map zonefile hash to zfinfo for block_height in range(start_block, end_block, 1): # TODO: can we do this transactionally? zonefile_info = db.get_atlas_zonefile_info_at( block_height ) for name_txid_zfhash in zonefile_info: name = str(name_txid_zfhash['name']) zfhash = str(name_txid_zfhash['value_hash']) txid = str(name_txid_zfhash['txid']) tried_storage = 0 present = is_zonefile_cached( zfhash, zonefile_dir, validate=validate ) zfinfo = atlasdb_get_zonefile( zfhash, con=con ) if zfinfo is not None: tried_storage = zfinfo['tried_storage'] if recover and present: log.debug('Recover: assume that {} is absent so we will reprocess it'.format(zfhash)) present = False log.debug("Add %s %s %s at %s (present: %s, tried_storage: %s)" % (name, zfhash, txid, block_height, present, tried_storage) ) atlasdb_add_zonefile_info( name, zfhash, txid, present, tried_storage, block_height, con=con ) total += 1 ret.append({ 'name': name, 'zonefile_hash': zfhash, 'txid': txid, 'block_height': block_height, 'present': present, 'tried_storage': tried_storage }) log.debug("Queued %s zonefiles from %s-%s" % (total, start_block, db.lastblock)) return ret
python
def atlasdb_queue_zonefiles( con, db, start_block, zonefile_dir, recover=False, validate=True, end_block=None ): # populate zonefile queue total = 0 if end_block is None: end_block = db.lastblock+1 ret = [] # map zonefile hash to zfinfo for block_height in range(start_block, end_block, 1): # TODO: can we do this transactionally? zonefile_info = db.get_atlas_zonefile_info_at( block_height ) for name_txid_zfhash in zonefile_info: name = str(name_txid_zfhash['name']) zfhash = str(name_txid_zfhash['value_hash']) txid = str(name_txid_zfhash['txid']) tried_storage = 0 present = is_zonefile_cached( zfhash, zonefile_dir, validate=validate ) zfinfo = atlasdb_get_zonefile( zfhash, con=con ) if zfinfo is not None: tried_storage = zfinfo['tried_storage'] if recover and present: log.debug('Recover: assume that {} is absent so we will reprocess it'.format(zfhash)) present = False log.debug("Add %s %s %s at %s (present: %s, tried_storage: %s)" % (name, zfhash, txid, block_height, present, tried_storage) ) atlasdb_add_zonefile_info( name, zfhash, txid, present, tried_storage, block_height, con=con ) total += 1 ret.append({ 'name': name, 'zonefile_hash': zfhash, 'txid': txid, 'block_height': block_height, 'present': present, 'tried_storage': tried_storage }) log.debug("Queued %s zonefiles from %s-%s" % (total, start_block, db.lastblock)) return ret
[ "def", "atlasdb_queue_zonefiles", "(", "con", ",", "db", ",", "start_block", ",", "zonefile_dir", ",", "recover", "=", "False", ",", "validate", "=", "True", ",", "end_block", "=", "None", ")", ":", "# populate zonefile queue", "total", "=", "0", "if", "end_...
Queue all zonefile hashes in the BlockstackDB to the zonefile queue NOT THREAD SAFE Returns the list of zonefile infos queued, and whether or not they are present.
[ "Queue", "all", "zonefile", "hashes", "in", "the", "BlockstackDB", "to", "the", "zonefile", "queue" ]
1dcfdd39b152d29ce13e736a6a1a0981401a0505
https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/atlas.py#L940-L989
244,242
blockstack/blockstack-core
blockstack/lib/atlas.py
atlasdb_sync_zonefiles
def atlasdb_sync_zonefiles( db, start_block, zonefile_dir, atlas_state, validate=True, end_block=None, path=None, con=None ): """ Synchronize atlas DB with name db NOT THREAD SAFE """ ret = None with AtlasDBOpen(con=con, path=path) as dbcon: ret = atlasdb_queue_zonefiles( dbcon, db, start_block, zonefile_dir, validate=validate, end_block=end_block ) atlasdb_cache_zonefile_info( con=dbcon ) if atlas_state: # it could have been the case that a zone file we already have was re-announced. # if so, then inform any storage listeners in the crawler thread that this has happened # (such as the subdomain system). crawler_thread = atlas_state['zonefile_crawler'] for zfinfo in filter(lambda zfi: zfi['present'], ret): log.debug('Store re-discovered zonefile {} at {}'.format(zfinfo['zonefile_hash'], zfinfo['block_height'])) crawler_thread.store_zonefile_cb(zfinfo['zonefile_hash'], zfinfo['block_height']) return ret
python
def atlasdb_sync_zonefiles( db, start_block, zonefile_dir, atlas_state, validate=True, end_block=None, path=None, con=None ): ret = None with AtlasDBOpen(con=con, path=path) as dbcon: ret = atlasdb_queue_zonefiles( dbcon, db, start_block, zonefile_dir, validate=validate, end_block=end_block ) atlasdb_cache_zonefile_info( con=dbcon ) if atlas_state: # it could have been the case that a zone file we already have was re-announced. # if so, then inform any storage listeners in the crawler thread that this has happened # (such as the subdomain system). crawler_thread = atlas_state['zonefile_crawler'] for zfinfo in filter(lambda zfi: zfi['present'], ret): log.debug('Store re-discovered zonefile {} at {}'.format(zfinfo['zonefile_hash'], zfinfo['block_height'])) crawler_thread.store_zonefile_cb(zfinfo['zonefile_hash'], zfinfo['block_height']) return ret
[ "def", "atlasdb_sync_zonefiles", "(", "db", ",", "start_block", ",", "zonefile_dir", ",", "atlas_state", ",", "validate", "=", "True", ",", "end_block", "=", "None", ",", "path", "=", "None", ",", "con", "=", "None", ")", ":", "ret", "=", "None", "with",...
Synchronize atlas DB with name db NOT THREAD SAFE
[ "Synchronize", "atlas", "DB", "with", "name", "db" ]
1dcfdd39b152d29ce13e736a6a1a0981401a0505
https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/atlas.py#L992-L1012
244,243
blockstack/blockstack-core
blockstack/lib/atlas.py
atlasdb_add_peer
def atlasdb_add_peer( peer_hostport, discovery_time=None, peer_table=None, con=None, path=None, ping_on_evict=True ): """ Add a peer to the peer table. If the peer conflicts with another peer, ping it first, and only insert the new peer if the old peer is dead. Keep the in-RAM peer table cache-coherent as well. Return True if this peer was added to the table (or preserved) Return False if not """ # bound the number of peers we add to PEER_MAX_DB assert len(peer_hostport) > 0 sk = random.randint(0, 2**32) peer_host, peer_port = url_to_host_port( peer_hostport ) assert len(peer_host) > 0 peer_slot = int( hashlib.sha256("%s%s" % (sk, peer_host)).hexdigest(), 16 ) % PEER_MAX_DB with AtlasDBOpen(con=con, path=path) as dbcon: if discovery_time is None: discovery_time = int(time.time()) do_evict_and_ping = False with AtlasPeerTableLocked(peer_table) as ptbl: # if the peer is already present, then we're done if peer_hostport in ptbl.keys(): log.debug("%s already in the peer table" % peer_hostport) return True # not in the table yet. See if we can evict someone if ping_on_evict: do_evict_and_ping = True if do_evict_and_ping: # evict someone # don't hold the peer table lock across network I/O sql = "SELECT peer_hostport FROM peers WHERE peer_slot = ?;" args = (peer_slot,) cur = dbcon.cursor() res = atlasdb_query_execute( cur, sql, args ) old_hostports = [] for row in res: old_hostport = res['peer_hostport'] old_hostports.append( old_hostport ) for old_hostport in old_hostports: # is this other peer still alive? # is this other peer part of the same mainnet history? # res = atlas_peer_ping( old_hostport ) res = atlas_peer_getinfo(old_hostport) if res: log.debug("Peer %s is still alive; will not replace" % (old_hostport)) return False # insert new peer with AtlasPeerTableLocked(peer_table) as ptbl: log.debug("Add peer '%s' discovered at %s (slot %s)" % (peer_hostport, discovery_time, peer_slot)) # peer is dead (or we don't care). Can insert or update sql = "INSERT OR REPLACE INTO peers (peer_hostport, peer_slot, discovery_time) VALUES (?,?,?);" args = (peer_hostport, peer_slot, discovery_time) cur = dbcon.cursor() res = atlasdb_query_execute( cur, sql, args ) dbcon.commit() # add to peer table as well atlas_init_peer_info( ptbl, peer_hostport, blacklisted=False, whitelisted=False ) return True
python
def atlasdb_add_peer( peer_hostport, discovery_time=None, peer_table=None, con=None, path=None, ping_on_evict=True ): # bound the number of peers we add to PEER_MAX_DB assert len(peer_hostport) > 0 sk = random.randint(0, 2**32) peer_host, peer_port = url_to_host_port( peer_hostport ) assert len(peer_host) > 0 peer_slot = int( hashlib.sha256("%s%s" % (sk, peer_host)).hexdigest(), 16 ) % PEER_MAX_DB with AtlasDBOpen(con=con, path=path) as dbcon: if discovery_time is None: discovery_time = int(time.time()) do_evict_and_ping = False with AtlasPeerTableLocked(peer_table) as ptbl: # if the peer is already present, then we're done if peer_hostport in ptbl.keys(): log.debug("%s already in the peer table" % peer_hostport) return True # not in the table yet. See if we can evict someone if ping_on_evict: do_evict_and_ping = True if do_evict_and_ping: # evict someone # don't hold the peer table lock across network I/O sql = "SELECT peer_hostport FROM peers WHERE peer_slot = ?;" args = (peer_slot,) cur = dbcon.cursor() res = atlasdb_query_execute( cur, sql, args ) old_hostports = [] for row in res: old_hostport = res['peer_hostport'] old_hostports.append( old_hostport ) for old_hostport in old_hostports: # is this other peer still alive? # is this other peer part of the same mainnet history? # res = atlas_peer_ping( old_hostport ) res = atlas_peer_getinfo(old_hostport) if res: log.debug("Peer %s is still alive; will not replace" % (old_hostport)) return False # insert new peer with AtlasPeerTableLocked(peer_table) as ptbl: log.debug("Add peer '%s' discovered at %s (slot %s)" % (peer_hostport, discovery_time, peer_slot)) # peer is dead (or we don't care). Can insert or update sql = "INSERT OR REPLACE INTO peers (peer_hostport, peer_slot, discovery_time) VALUES (?,?,?);" args = (peer_hostport, peer_slot, discovery_time) cur = dbcon.cursor() res = atlasdb_query_execute( cur, sql, args ) dbcon.commit() # add to peer table as well atlas_init_peer_info( ptbl, peer_hostport, blacklisted=False, whitelisted=False ) return True
[ "def", "atlasdb_add_peer", "(", "peer_hostport", ",", "discovery_time", "=", "None", ",", "peer_table", "=", "None", ",", "con", "=", "None", ",", "path", "=", "None", ",", "ping_on_evict", "=", "True", ")", ":", "# bound the number of peers we add to PEER_MAX_DB"...
Add a peer to the peer table. If the peer conflicts with another peer, ping it first, and only insert the new peer if the old peer is dead. Keep the in-RAM peer table cache-coherent as well. Return True if this peer was added to the table (or preserved) Return False if not
[ "Add", "a", "peer", "to", "the", "peer", "table", ".", "If", "the", "peer", "conflicts", "with", "another", "peer", "ping", "it", "first", "and", "only", "insert", "the", "new", "peer", "if", "the", "old", "peer", "is", "dead", "." ]
1dcfdd39b152d29ce13e736a6a1a0981401a0505
https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/atlas.py#L1015-L1095
244,244
blockstack/blockstack-core
blockstack/lib/atlas.py
atlasdb_num_peers
def atlasdb_num_peers( con=None, path=None ): """ How many peers are there in the db? """ with AtlasDBOpen(con=con, path=path) as dbcon: sql = "SELECT MAX(peer_index) FROM peers;" args = () cur = dbcon.cursor() res = atlasdb_query_execute( cur, sql, args ) ret = [] for row in res: tmp = {} tmp.update(row) ret.append(tmp) assert len(ret) == 1 return ret[0]['MAX(peer_index)']
python
def atlasdb_num_peers( con=None, path=None ): with AtlasDBOpen(con=con, path=path) as dbcon: sql = "SELECT MAX(peer_index) FROM peers;" args = () cur = dbcon.cursor() res = atlasdb_query_execute( cur, sql, args ) ret = [] for row in res: tmp = {} tmp.update(row) ret.append(tmp) assert len(ret) == 1 return ret[0]['MAX(peer_index)']
[ "def", "atlasdb_num_peers", "(", "con", "=", "None", ",", "path", "=", "None", ")", ":", "with", "AtlasDBOpen", "(", "con", "=", "con", ",", "path", "=", "path", ")", "as", "dbcon", ":", "sql", "=", "\"SELECT MAX(peer_index) FROM peers;\"", "args", "=", ...
How many peers are there in the db?
[ "How", "many", "peers", "are", "there", "in", "the", "db?" ]
1dcfdd39b152d29ce13e736a6a1a0981401a0505
https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/atlas.py#L1124-L1144
244,245
blockstack/blockstack-core
blockstack/lib/atlas.py
atlas_get_peer
def atlas_get_peer( peer_hostport, peer_table=None ): """ Get the given peer's info """ ret = None with AtlasPeerTableLocked(peer_table) as ptbl: ret = ptbl.get(peer_hostport, None) return ret
python
def atlas_get_peer( peer_hostport, peer_table=None ): ret = None with AtlasPeerTableLocked(peer_table) as ptbl: ret = ptbl.get(peer_hostport, None) return ret
[ "def", "atlas_get_peer", "(", "peer_hostport", ",", "peer_table", "=", "None", ")", ":", "ret", "=", "None", "with", "AtlasPeerTableLocked", "(", "peer_table", ")", "as", "ptbl", ":", "ret", "=", "ptbl", ".", "get", "(", "peer_hostport", ",", "None", ")", ...
Get the given peer's info
[ "Get", "the", "given", "peer", "s", "info" ]
1dcfdd39b152d29ce13e736a6a1a0981401a0505
https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/atlas.py#L1147-L1156
244,246
blockstack/blockstack-core
blockstack/lib/atlas.py
atlasdb_get_random_peer
def atlasdb_get_random_peer( con=None, path=None ): """ Select a peer from the db at random Return None if the table is empty """ ret = {} with AtlasDBOpen(con=con, path=path) as dbcon: num_peers = atlasdb_num_peers( con=con, path=path ) if num_peers is None or num_peers == 0: # no peers ret['peer_hostport'] = None else: r = random.randint(1, num_peers) sql = "SELECT * FROM peers WHERE peer_index = ?;" args = (r,) cur = dbcon.cursor() res = atlasdb_query_execute( cur, sql, args ) ret = {'peer_hostport': None} for row in res: ret.update( row ) break return ret['peer_hostport']
python
def atlasdb_get_random_peer( con=None, path=None ): ret = {} with AtlasDBOpen(con=con, path=path) as dbcon: num_peers = atlasdb_num_peers( con=con, path=path ) if num_peers is None or num_peers == 0: # no peers ret['peer_hostport'] = None else: r = random.randint(1, num_peers) sql = "SELECT * FROM peers WHERE peer_index = ?;" args = (r,) cur = dbcon.cursor() res = atlasdb_query_execute( cur, sql, args ) ret = {'peer_hostport': None} for row in res: ret.update( row ) break return ret['peer_hostport']
[ "def", "atlasdb_get_random_peer", "(", "con", "=", "None", ",", "path", "=", "None", ")", ":", "ret", "=", "{", "}", "with", "AtlasDBOpen", "(", "con", "=", "con", ",", "path", "=", "path", ")", "as", "dbcon", ":", "num_peers", "=", "atlasdb_num_peers"...
Select a peer from the db at random Return None if the table is empty
[ "Select", "a", "peer", "from", "the", "db", "at", "random", "Return", "None", "if", "the", "table", "is", "empty" ]
1dcfdd39b152d29ce13e736a6a1a0981401a0505
https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/atlas.py#L1159-L1188
244,247
blockstack/blockstack-core
blockstack/lib/atlas.py
atlasdb_get_old_peers
def atlasdb_get_old_peers( now, con=None, path=None ): """ Get peers older than now - PEER_LIFETIME """ with AtlasDBOpen(con=con, path=path) as dbcon: if now is None: now = time.time() expire = now - atlas_peer_max_age() sql = "SELECT * FROM peers WHERE discovery_time < ?"; args = (expire,) cur = dbcon.cursor() res = atlasdb_query_execute( cur, sql, args ) rows = [] for row in res: tmp = {} tmp.update(row) rows.append(tmp) return rows
python
def atlasdb_get_old_peers( now, con=None, path=None ): with AtlasDBOpen(con=con, path=path) as dbcon: if now is None: now = time.time() expire = now - atlas_peer_max_age() sql = "SELECT * FROM peers WHERE discovery_time < ?"; args = (expire,) cur = dbcon.cursor() res = atlasdb_query_execute( cur, sql, args ) rows = [] for row in res: tmp = {} tmp.update(row) rows.append(tmp) return rows
[ "def", "atlasdb_get_old_peers", "(", "now", ",", "con", "=", "None", ",", "path", "=", "None", ")", ":", "with", "AtlasDBOpen", "(", "con", "=", "con", ",", "path", "=", "path", ")", "as", "dbcon", ":", "if", "now", "is", "None", ":", "now", "=", ...
Get peers older than now - PEER_LIFETIME
[ "Get", "peers", "older", "than", "now", "-", "PEER_LIFETIME" ]
1dcfdd39b152d29ce13e736a6a1a0981401a0505
https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/atlas.py#L1191-L1213
244,248
blockstack/blockstack-core
blockstack/lib/atlas.py
atlasdb_renew_peer
def atlasdb_renew_peer( peer_hostport, now, con=None, path=None ): """ Renew a peer's discovery time """ with AtlasDBOpen(con=con, path=path) as dbcon: if now is None: now = time.time() sql = "UPDATE peers SET discovery_time = ? WHERE peer_hostport = ?;" args = (now, peer_hostport) cur = dbcon.cursor() res = atlasdb_query_execute( cur, sql, args ) dbcon.commit() return True
python
def atlasdb_renew_peer( peer_hostport, now, con=None, path=None ): with AtlasDBOpen(con=con, path=path) as dbcon: if now is None: now = time.time() sql = "UPDATE peers SET discovery_time = ? WHERE peer_hostport = ?;" args = (now, peer_hostport) cur = dbcon.cursor() res = atlasdb_query_execute( cur, sql, args ) dbcon.commit() return True
[ "def", "atlasdb_renew_peer", "(", "peer_hostport", ",", "now", ",", "con", "=", "None", ",", "path", "=", "None", ")", ":", "with", "AtlasDBOpen", "(", "con", "=", "con", ",", "path", "=", "path", ")", "as", "dbcon", ":", "if", "now", "is", "None", ...
Renew a peer's discovery time
[ "Renew", "a", "peer", "s", "discovery", "time" ]
1dcfdd39b152d29ce13e736a6a1a0981401a0505
https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/atlas.py#L1216-L1231
244,249
blockstack/blockstack-core
blockstack/lib/atlas.py
atlasdb_load_peer_table
def atlasdb_load_peer_table( con=None, path=None ): """ Create a peer table from the peer DB """ peer_table = {} with AtlasDBOpen(con=con, path=path) as dbcon: sql = "SELECT * FROM peers;" args = () cur = dbcon.cursor() res = atlasdb_query_execute( cur, sql, args ) # build it up count = 0 for row in res: if count > 0 and count % 100 == 0: log.debug("Loaded %s peers..." % count) atlas_init_peer_info( peer_table, row['peer_hostport'] ) count += 1 return peer_table
python
def atlasdb_load_peer_table( con=None, path=None ): peer_table = {} with AtlasDBOpen(con=con, path=path) as dbcon: sql = "SELECT * FROM peers;" args = () cur = dbcon.cursor() res = atlasdb_query_execute( cur, sql, args ) # build it up count = 0 for row in res: if count > 0 and count % 100 == 0: log.debug("Loaded %s peers..." % count) atlas_init_peer_info( peer_table, row['peer_hostport'] ) count += 1 return peer_table
[ "def", "atlasdb_load_peer_table", "(", "con", "=", "None", ",", "path", "=", "None", ")", ":", "peer_table", "=", "{", "}", "with", "AtlasDBOpen", "(", "con", "=", "con", ",", "path", "=", "path", ")", "as", "dbcon", ":", "sql", "=", "\"SELECT * FROM p...
Create a peer table from the peer DB
[ "Create", "a", "peer", "table", "from", "the", "peer", "DB" ]
1dcfdd39b152d29ce13e736a6a1a0981401a0505
https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/atlas.py#L1234-L1257
244,250
blockstack/blockstack-core
blockstack/lib/atlas.py
atlasdb_zonefile_inv_list
def atlasdb_zonefile_inv_list( bit_offset, bit_length, con=None, path=None ): """ Get an inventory listing. offset and length are in bits. Return the list of zonefile information. The list may be less than length elements. """ with AtlasDBOpen(con=con, path=path) as dbcon: sql = "SELECT * FROM zonefiles LIMIT ? OFFSET ?;" args = (bit_length, bit_offset) cur = dbcon.cursor() res = atlasdb_query_execute( cur, sql, args ) ret = [] for row in res: tmp = {} tmp.update(row) ret.append(tmp) return ret
python
def atlasdb_zonefile_inv_list( bit_offset, bit_length, con=None, path=None ): with AtlasDBOpen(con=con, path=path) as dbcon: sql = "SELECT * FROM zonefiles LIMIT ? OFFSET ?;" args = (bit_length, bit_offset) cur = dbcon.cursor() res = atlasdb_query_execute( cur, sql, args ) ret = [] for row in res: tmp = {} tmp.update(row) ret.append(tmp) return ret
[ "def", "atlasdb_zonefile_inv_list", "(", "bit_offset", ",", "bit_length", ",", "con", "=", "None", ",", "path", "=", "None", ")", ":", "with", "AtlasDBOpen", "(", "con", "=", "con", ",", "path", "=", "path", ")", "as", "dbcon", ":", "sql", "=", "\"SELE...
Get an inventory listing. offset and length are in bits. Return the list of zonefile information. The list may be less than length elements.
[ "Get", "an", "inventory", "listing", ".", "offset", "and", "length", "are", "in", "bits", "." ]
1dcfdd39b152d29ce13e736a6a1a0981401a0505
https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/atlas.py#L1364-L1386
244,251
blockstack/blockstack-core
blockstack/lib/atlas.py
atlas_init_peer_info
def atlas_init_peer_info( peer_table, peer_hostport, blacklisted=False, whitelisted=False ): """ Initialize peer info table entry """ peer_table[peer_hostport] = { "time": [], "zonefile_inv": "", "blacklisted": blacklisted, "whitelisted": whitelisted }
python
def atlas_init_peer_info( peer_table, peer_hostport, blacklisted=False, whitelisted=False ): peer_table[peer_hostport] = { "time": [], "zonefile_inv": "", "blacklisted": blacklisted, "whitelisted": whitelisted }
[ "def", "atlas_init_peer_info", "(", "peer_table", ",", "peer_hostport", ",", "blacklisted", "=", "False", ",", "whitelisted", "=", "False", ")", ":", "peer_table", "[", "peer_hostport", "]", "=", "{", "\"time\"", ":", "[", "]", ",", "\"zonefile_inv\"", ":", ...
Initialize peer info table entry
[ "Initialize", "peer", "info", "table", "entry" ]
1dcfdd39b152d29ce13e736a6a1a0981401a0505
https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/atlas.py#L1546-L1555
244,252
blockstack/blockstack-core
blockstack/lib/atlas.py
atlas_log_socket_error
def atlas_log_socket_error( method_invocation, peer_hostport, se ): """ Log a socket exception tastefully """ if isinstance( se, socket.timeout ): log.debug("%s %s: timed out (socket.timeout)" % (method_invocation, peer_hostport)) elif isinstance( se, socket.gaierror ): log.debug("%s %s: failed to query address or info (socket.gaierror)" % (method_invocation, peer_hostport )) elif isinstance( se, socket.herror ): log.debug("%s %s: failed to query host info (socket.herror)" % (method_invocation, peer_hostport )) elif isinstance( se, socket.error ): if se.errno == errno.ECONNREFUSED: log.debug("%s %s: is unreachable (socket.error ECONNREFUSED)" % (method_invocation, peer_hostport)) elif se.errno == errno.ETIMEDOUT: log.debug("%s %s: timed out (socket.error ETIMEDOUT)" % (method_invocation, peer_hostport)) else: log.debug("%s %s: socket error" % (method_invocation, peer_hostport)) log.exception(se) else: log.debug("%s %s: general exception" % (method_invocation, peer_hostport)) log.exception(se)
python
def atlas_log_socket_error( method_invocation, peer_hostport, se ): if isinstance( se, socket.timeout ): log.debug("%s %s: timed out (socket.timeout)" % (method_invocation, peer_hostport)) elif isinstance( se, socket.gaierror ): log.debug("%s %s: failed to query address or info (socket.gaierror)" % (method_invocation, peer_hostport )) elif isinstance( se, socket.herror ): log.debug("%s %s: failed to query host info (socket.herror)" % (method_invocation, peer_hostport )) elif isinstance( se, socket.error ): if se.errno == errno.ECONNREFUSED: log.debug("%s %s: is unreachable (socket.error ECONNREFUSED)" % (method_invocation, peer_hostport)) elif se.errno == errno.ETIMEDOUT: log.debug("%s %s: timed out (socket.error ETIMEDOUT)" % (method_invocation, peer_hostport)) else: log.debug("%s %s: socket error" % (method_invocation, peer_hostport)) log.exception(se) else: log.debug("%s %s: general exception" % (method_invocation, peer_hostport)) log.exception(se)
[ "def", "atlas_log_socket_error", "(", "method_invocation", ",", "peer_hostport", ",", "se", ")", ":", "if", "isinstance", "(", "se", ",", "socket", ".", "timeout", ")", ":", "log", ".", "debug", "(", "\"%s %s: timed out (socket.timeout)\"", "%", "(", "method_inv...
Log a socket exception tastefully
[ "Log", "a", "socket", "exception", "tastefully" ]
1dcfdd39b152d29ce13e736a6a1a0981401a0505
https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/atlas.py#L1558-L1582
244,253
blockstack/blockstack-core
blockstack/lib/atlas.py
atlas_peer_ping
def atlas_peer_ping( peer_hostport, timeout=None, peer_table=None ): """ Ping a host Return True if alive Return False if not """ if timeout is None: timeout = atlas_ping_timeout() assert not atlas_peer_table_is_locked_by_me() host, port = url_to_host_port( peer_hostport ) RPC = get_rpc_client_class() rpc = RPC( host, port, timeout=timeout ) log.debug("Ping %s" % peer_hostport) ret = False try: res = blockstack_ping( proxy=rpc ) if 'error' not in res: ret = True except (socket.timeout, socket.gaierror, socket.herror, socket.error), se: atlas_log_socket_error( "ping(%s)" % peer_hostport, peer_hostport, se ) pass except Exception, e: log.exception(e) pass # update health with AtlasPeerTableLocked(peer_table) as ptbl: atlas_peer_update_health( peer_hostport, ret, peer_table=ptbl ) return ret
python
def atlas_peer_ping( peer_hostport, timeout=None, peer_table=None ): if timeout is None: timeout = atlas_ping_timeout() assert not atlas_peer_table_is_locked_by_me() host, port = url_to_host_port( peer_hostport ) RPC = get_rpc_client_class() rpc = RPC( host, port, timeout=timeout ) log.debug("Ping %s" % peer_hostport) ret = False try: res = blockstack_ping( proxy=rpc ) if 'error' not in res: ret = True except (socket.timeout, socket.gaierror, socket.herror, socket.error), se: atlas_log_socket_error( "ping(%s)" % peer_hostport, peer_hostport, se ) pass except Exception, e: log.exception(e) pass # update health with AtlasPeerTableLocked(peer_table) as ptbl: atlas_peer_update_health( peer_hostport, ret, peer_table=ptbl ) return ret
[ "def", "atlas_peer_ping", "(", "peer_hostport", ",", "timeout", "=", "None", ",", "peer_table", "=", "None", ")", ":", "if", "timeout", "is", "None", ":", "timeout", "=", "atlas_ping_timeout", "(", ")", "assert", "not", "atlas_peer_table_is_locked_by_me", "(", ...
Ping a host Return True if alive Return False if not
[ "Ping", "a", "host", "Return", "True", "if", "alive", "Return", "False", "if", "not" ]
1dcfdd39b152d29ce13e736a6a1a0981401a0505
https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/atlas.py#L1585-L1621
244,254
blockstack/blockstack-core
blockstack/lib/atlas.py
atlas_inventory_count_missing
def atlas_inventory_count_missing( inv1, inv2 ): """ Find out how many bits are set in inv2 that are not set in inv1. """ count = 0 common = min(len(inv1), len(inv2)) for i in xrange(0, common): for j in xrange(0, 8): if ((1 << (7 - j)) & ord(inv2[i])) != 0 and ((1 << (7 - j)) & ord(inv1[i])) == 0: count += 1 if len(inv1) < len(inv2): for i in xrange(len(inv1), len(inv2)): for j in xrange(0, 8): if ((1 << (7 - j)) & ord(inv2[i])) != 0: count += 1 return count
python
def atlas_inventory_count_missing( inv1, inv2 ): count = 0 common = min(len(inv1), len(inv2)) for i in xrange(0, common): for j in xrange(0, 8): if ((1 << (7 - j)) & ord(inv2[i])) != 0 and ((1 << (7 - j)) & ord(inv1[i])) == 0: count += 1 if len(inv1) < len(inv2): for i in xrange(len(inv1), len(inv2)): for j in xrange(0, 8): if ((1 << (7 - j)) & ord(inv2[i])) != 0: count += 1 return count
[ "def", "atlas_inventory_count_missing", "(", "inv1", ",", "inv2", ")", ":", "count", "=", "0", "common", "=", "min", "(", "len", "(", "inv1", ")", ",", "len", "(", "inv2", ")", ")", "for", "i", "in", "xrange", "(", "0", ",", "common", ")", ":", "...
Find out how many bits are set in inv2 that are not set in inv1.
[ "Find", "out", "how", "many", "bits", "are", "set", "in", "inv2", "that", "are", "not", "set", "in", "inv1", "." ]
1dcfdd39b152d29ce13e736a6a1a0981401a0505
https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/atlas.py#L1707-L1725
244,255
blockstack/blockstack-core
blockstack/lib/atlas.py
atlas_revalidate_peers
def atlas_revalidate_peers( con=None, path=None, now=None, peer_table=None ): """ Revalidate peers that are older than the maximum peer age. Ping them, and if they don't respond, remove them. """ global MIN_PEER_HEALTH if now is None: now = time_now() old_peer_infos = atlasdb_get_old_peers( now, con=con, path=path ) for old_peer_info in old_peer_infos: res = atlas_peer_getinfo( old_peer_info['peer_hostport'] ) if not res: log.debug("Failed to revalidate %s" % (old_peer_info['peer_hostport'])) if atlas_peer_is_whitelisted( old_peer_info['peer_hostport'], peer_table=peer_table ): continue if atlas_peer_is_blacklisted( old_peer_info['peer_hostport'], peer_table=peer_table ): continue if atlas_peer_get_health( old_peer_info['peer_hostport'], peer_table=peer_table ) < MIN_PEER_HEALTH: atlasdb_remove_peer( old_peer_info['peer_hostport'], con=con, path=path, peer_table=peer_table ) else: # renew atlasdb_renew_peer( old_peer_info['peer_hostport'], now, con=con, path=path ) return True
python
def atlas_revalidate_peers( con=None, path=None, now=None, peer_table=None ): global MIN_PEER_HEALTH if now is None: now = time_now() old_peer_infos = atlasdb_get_old_peers( now, con=con, path=path ) for old_peer_info in old_peer_infos: res = atlas_peer_getinfo( old_peer_info['peer_hostport'] ) if not res: log.debug("Failed to revalidate %s" % (old_peer_info['peer_hostport'])) if atlas_peer_is_whitelisted( old_peer_info['peer_hostport'], peer_table=peer_table ): continue if atlas_peer_is_blacklisted( old_peer_info['peer_hostport'], peer_table=peer_table ): continue if atlas_peer_get_health( old_peer_info['peer_hostport'], peer_table=peer_table ) < MIN_PEER_HEALTH: atlasdb_remove_peer( old_peer_info['peer_hostport'], con=con, path=path, peer_table=peer_table ) else: # renew atlasdb_renew_peer( old_peer_info['peer_hostport'], now, con=con, path=path ) return True
[ "def", "atlas_revalidate_peers", "(", "con", "=", "None", ",", "path", "=", "None", ",", "now", "=", "None", ",", "peer_table", "=", "None", ")", ":", "global", "MIN_PEER_HEALTH", "if", "now", "is", "None", ":", "now", "=", "time_now", "(", ")", "old_p...
Revalidate peers that are older than the maximum peer age. Ping them, and if they don't respond, remove them.
[ "Revalidate", "peers", "that", "are", "older", "than", "the", "maximum", "peer", "age", ".", "Ping", "them", "and", "if", "they", "don", "t", "respond", "remove", "them", "." ]
1dcfdd39b152d29ce13e736a6a1a0981401a0505
https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/atlas.py#L1775-L1803
244,256
blockstack/blockstack-core
blockstack/lib/atlas.py
atlas_peer_get_request_count
def atlas_peer_get_request_count( peer_hostport, peer_table=None ): """ How many times have we contacted this peer? """ with AtlasPeerTableLocked(peer_table) as ptbl: if peer_hostport not in ptbl.keys(): return 0 count = 0 for (t, r) in ptbl[peer_hostport]['time']: if r: count += 1 return count
python
def atlas_peer_get_request_count( peer_hostport, peer_table=None ): with AtlasPeerTableLocked(peer_table) as ptbl: if peer_hostport not in ptbl.keys(): return 0 count = 0 for (t, r) in ptbl[peer_hostport]['time']: if r: count += 1 return count
[ "def", "atlas_peer_get_request_count", "(", "peer_hostport", ",", "peer_table", "=", "None", ")", ":", "with", "AtlasPeerTableLocked", "(", "peer_table", ")", "as", "ptbl", ":", "if", "peer_hostport", "not", "in", "ptbl", ".", "keys", "(", ")", ":", "return", ...
How many times have we contacted this peer?
[ "How", "many", "times", "have", "we", "contacted", "this", "peer?" ]
1dcfdd39b152d29ce13e736a6a1a0981401a0505
https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/atlas.py#L1828-L1841
244,257
blockstack/blockstack-core
blockstack/lib/atlas.py
atlas_peer_get_zonefile_inventory
def atlas_peer_get_zonefile_inventory( peer_hostport, peer_table=None ): """ What's the zonefile inventory vector for this peer? Return None if not defined """ inv = None with AtlasPeerTableLocked(peer_table) as ptbl: if peer_hostport not in ptbl.keys(): return None inv = ptbl[peer_hostport]['zonefile_inv'] return inv
python
def atlas_peer_get_zonefile_inventory( peer_hostport, peer_table=None ): inv = None with AtlasPeerTableLocked(peer_table) as ptbl: if peer_hostport not in ptbl.keys(): return None inv = ptbl[peer_hostport]['zonefile_inv'] return inv
[ "def", "atlas_peer_get_zonefile_inventory", "(", "peer_hostport", ",", "peer_table", "=", "None", ")", ":", "inv", "=", "None", "with", "AtlasPeerTableLocked", "(", "peer_table", ")", "as", "ptbl", ":", "if", "peer_hostport", "not", "in", "ptbl", ".", "keys", ...
What's the zonefile inventory vector for this peer? Return None if not defined
[ "What", "s", "the", "zonefile", "inventory", "vector", "for", "this", "peer?", "Return", "None", "if", "not", "defined" ]
1dcfdd39b152d29ce13e736a6a1a0981401a0505
https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/atlas.py#L1844-L1857
244,258
blockstack/blockstack-core
blockstack/lib/atlas.py
atlas_peer_set_zonefile_inventory
def atlas_peer_set_zonefile_inventory( peer_hostport, peer_inv, peer_table=None ): """ Set this peer's zonefile inventory """ with AtlasPeerTableLocked(peer_table) as ptbl: if peer_hostport not in ptbl.keys(): return None ptbl[peer_hostport]['zonefile_inv'] = peer_inv return peer_inv
python
def atlas_peer_set_zonefile_inventory( peer_hostport, peer_inv, peer_table=None ): with AtlasPeerTableLocked(peer_table) as ptbl: if peer_hostport not in ptbl.keys(): return None ptbl[peer_hostport]['zonefile_inv'] = peer_inv return peer_inv
[ "def", "atlas_peer_set_zonefile_inventory", "(", "peer_hostport", ",", "peer_inv", ",", "peer_table", "=", "None", ")", ":", "with", "AtlasPeerTableLocked", "(", "peer_table", ")", "as", "ptbl", ":", "if", "peer_hostport", "not", "in", "ptbl", ".", "keys", "(", ...
Set this peer's zonefile inventory
[ "Set", "this", "peer", "s", "zonefile", "inventory" ]
1dcfdd39b152d29ce13e736a6a1a0981401a0505
https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/atlas.py#L1860-L1870
244,259
blockstack/blockstack-core
blockstack/lib/atlas.py
atlas_peer_is_whitelisted
def atlas_peer_is_whitelisted( peer_hostport, peer_table=None ): """ Is a peer whitelisted """ ret = None with AtlasPeerTableLocked(peer_table) as ptbl: if peer_hostport not in ptbl.keys(): return None ret = ptbl[peer_hostport].get("whitelisted", False) return ret
python
def atlas_peer_is_whitelisted( peer_hostport, peer_table=None ): ret = None with AtlasPeerTableLocked(peer_table) as ptbl: if peer_hostport not in ptbl.keys(): return None ret = ptbl[peer_hostport].get("whitelisted", False) return ret
[ "def", "atlas_peer_is_whitelisted", "(", "peer_hostport", ",", "peer_table", "=", "None", ")", ":", "ret", "=", "None", "with", "AtlasPeerTableLocked", "(", "peer_table", ")", "as", "ptbl", ":", "if", "peer_hostport", "not", "in", "ptbl", ".", "keys", "(", "...
Is a peer whitelisted
[ "Is", "a", "peer", "whitelisted" ]
1dcfdd39b152d29ce13e736a6a1a0981401a0505
https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/atlas.py#L1888-L1899
244,260
blockstack/blockstack-core
blockstack/lib/atlas.py
atlas_peer_update_health
def atlas_peer_update_health( peer_hostport, received_response, peer_table=None ): """ Mark the given peer as alive at this time. Update times at which we contacted it, and update its health score. Use the global health table by default, or use the given health info if set. """ with AtlasPeerTableLocked(peer_table) as ptbl: if peer_hostport not in ptbl.keys(): return False # record that we contacted this peer, and whether or not we useful info from it now = time_now() # update timestamps; remove old data new_times = [] for (t, r) in ptbl[peer_hostport]['time']: if t + atlas_peer_lifetime_interval() < now: continue new_times.append((t, r)) new_times.append((now, received_response)) ptbl[peer_hostport]['time'] = new_times return True
python
def atlas_peer_update_health( peer_hostport, received_response, peer_table=None ): with AtlasPeerTableLocked(peer_table) as ptbl: if peer_hostport not in ptbl.keys(): return False # record that we contacted this peer, and whether or not we useful info from it now = time_now() # update timestamps; remove old data new_times = [] for (t, r) in ptbl[peer_hostport]['time']: if t + atlas_peer_lifetime_interval() < now: continue new_times.append((t, r)) new_times.append((now, received_response)) ptbl[peer_hostport]['time'] = new_times return True
[ "def", "atlas_peer_update_health", "(", "peer_hostport", ",", "received_response", ",", "peer_table", "=", "None", ")", ":", "with", "AtlasPeerTableLocked", "(", "peer_table", ")", "as", "ptbl", ":", "if", "peer_hostport", "not", "in", "ptbl", ".", "keys", "(", ...
Mark the given peer as alive at this time. Update times at which we contacted it, and update its health score. Use the global health table by default, or use the given health info if set.
[ "Mark", "the", "given", "peer", "as", "alive", "at", "this", "time", ".", "Update", "times", "at", "which", "we", "contacted", "it", "and", "update", "its", "health", "score", "." ]
1dcfdd39b152d29ce13e736a6a1a0981401a0505
https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/atlas.py#L1902-L1930
244,261
blockstack/blockstack-core
blockstack/lib/atlas.py
atlas_peer_download_zonefile_inventory
def atlas_peer_download_zonefile_inventory( my_hostport, peer_hostport, maxlen, bit_offset=0, timeout=None, peer_table={} ): """ Get the zonefile inventory from the remote peer Start from the given bit_offset NOTE: this doesn't update the peer table health by default; you'll have to explicitly pass in a peer table (i.e. setting to {} ensures that nothing happens). """ if timeout is None: timeout = atlas_inv_timeout() interval = 524288 # number of bits in 64KB peer_inv = "" log.debug("Download zonefile inventory %s-%s from %s" % (bit_offset, maxlen, peer_hostport)) if bit_offset > maxlen: # synced already return peer_inv for offset in xrange( bit_offset, maxlen, interval): next_inv = atlas_peer_get_zonefile_inventory_range( my_hostport, peer_hostport, offset, interval, timeout=timeout, peer_table=peer_table ) if next_inv is None: # partial failure log.debug("Failed to sync inventory for %s from %s to %s" % (peer_hostport, offset, offset+interval)) break peer_inv += next_inv if len(next_inv) < interval: # end-of-interval break return peer_inv
python
def atlas_peer_download_zonefile_inventory( my_hostport, peer_hostport, maxlen, bit_offset=0, timeout=None, peer_table={} ): if timeout is None: timeout = atlas_inv_timeout() interval = 524288 # number of bits in 64KB peer_inv = "" log.debug("Download zonefile inventory %s-%s from %s" % (bit_offset, maxlen, peer_hostport)) if bit_offset > maxlen: # synced already return peer_inv for offset in xrange( bit_offset, maxlen, interval): next_inv = atlas_peer_get_zonefile_inventory_range( my_hostport, peer_hostport, offset, interval, timeout=timeout, peer_table=peer_table ) if next_inv is None: # partial failure log.debug("Failed to sync inventory for %s from %s to %s" % (peer_hostport, offset, offset+interval)) break peer_inv += next_inv if len(next_inv) < interval: # end-of-interval break return peer_inv
[ "def", "atlas_peer_download_zonefile_inventory", "(", "my_hostport", ",", "peer_hostport", ",", "maxlen", ",", "bit_offset", "=", "0", ",", "timeout", "=", "None", ",", "peer_table", "=", "{", "}", ")", ":", "if", "timeout", "is", "None", ":", "timeout", "="...
Get the zonefile inventory from the remote peer Start from the given bit_offset NOTE: this doesn't update the peer table health by default; you'll have to explicitly pass in a peer table (i.e. setting to {} ensures that nothing happens).
[ "Get", "the", "zonefile", "inventory", "from", "the", "remote", "peer", "Start", "from", "the", "given", "bit_offset" ]
1dcfdd39b152d29ce13e736a6a1a0981401a0505
https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/atlas.py#L1993-L2027
244,262
blockstack/blockstack-core
blockstack/lib/atlas.py
atlas_peer_sync_zonefile_inventory
def atlas_peer_sync_zonefile_inventory( my_hostport, peer_hostport, maxlen, timeout=None, peer_table=None ): """ Synchronize our knowledge of a peer's zonefiles up to a given byte length NOT THREAD SAFE; CALL FROM ONLY ONE THREAD. maxlen is the maximum length in bits of the expected zonefile. Return the new inv vector if we synced it (updating the peer table in the process) Return None if not """ if timeout is None: timeout = atlas_inv_timeout() peer_inv = "" bit_offset = None with AtlasPeerTableLocked(peer_table) as ptbl: if peer_hostport not in ptbl.keys(): return None peer_inv = atlas_peer_get_zonefile_inventory( peer_hostport, peer_table=ptbl ) bit_offset = (len(peer_inv) - 1) * 8 # i.e. re-obtain the last byte if bit_offset < 0: bit_offset = 0 else: peer_inv = peer_inv[:-1] peer_inv = atlas_peer_download_zonefile_inventory( my_hostport, peer_hostport, maxlen, bit_offset=bit_offset, timeout=timeout, peer_table=peer_table ) with AtlasPeerTableLocked(peer_table) as ptbl: if peer_hostport not in ptbl.keys(): log.debug("%s no longer a peer" % peer_hostport) return None inv_str = atlas_inventory_to_string(peer_inv) if len(inv_str) > 40: inv_str = inv_str[:40] + "..." log.debug("Set zonefile inventory %s: %s" % (peer_hostport, inv_str)) atlas_peer_set_zonefile_inventory( peer_hostport, peer_inv, peer_table=ptbl ) # NOTE: may have trailing 0's for padding return peer_inv
python
def atlas_peer_sync_zonefile_inventory( my_hostport, peer_hostport, maxlen, timeout=None, peer_table=None ): if timeout is None: timeout = atlas_inv_timeout() peer_inv = "" bit_offset = None with AtlasPeerTableLocked(peer_table) as ptbl: if peer_hostport not in ptbl.keys(): return None peer_inv = atlas_peer_get_zonefile_inventory( peer_hostport, peer_table=ptbl ) bit_offset = (len(peer_inv) - 1) * 8 # i.e. re-obtain the last byte if bit_offset < 0: bit_offset = 0 else: peer_inv = peer_inv[:-1] peer_inv = atlas_peer_download_zonefile_inventory( my_hostport, peer_hostport, maxlen, bit_offset=bit_offset, timeout=timeout, peer_table=peer_table ) with AtlasPeerTableLocked(peer_table) as ptbl: if peer_hostport not in ptbl.keys(): log.debug("%s no longer a peer" % peer_hostport) return None inv_str = atlas_inventory_to_string(peer_inv) if len(inv_str) > 40: inv_str = inv_str[:40] + "..." log.debug("Set zonefile inventory %s: %s" % (peer_hostport, inv_str)) atlas_peer_set_zonefile_inventory( peer_hostport, peer_inv, peer_table=ptbl ) # NOTE: may have trailing 0's for padding return peer_inv
[ "def", "atlas_peer_sync_zonefile_inventory", "(", "my_hostport", ",", "peer_hostport", ",", "maxlen", ",", "timeout", "=", "None", ",", "peer_table", "=", "None", ")", ":", "if", "timeout", "is", "None", ":", "timeout", "=", "atlas_inv_timeout", "(", ")", "pee...
Synchronize our knowledge of a peer's zonefiles up to a given byte length NOT THREAD SAFE; CALL FROM ONLY ONE THREAD. maxlen is the maximum length in bits of the expected zonefile. Return the new inv vector if we synced it (updating the peer table in the process) Return None if not
[ "Synchronize", "our", "knowledge", "of", "a", "peer", "s", "zonefiles", "up", "to", "a", "given", "byte", "length", "NOT", "THREAD", "SAFE", ";", "CALL", "FROM", "ONLY", "ONE", "THREAD", "." ]
1dcfdd39b152d29ce13e736a6a1a0981401a0505
https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/atlas.py#L2031-L2074
244,263
blockstack/blockstack-core
blockstack/lib/atlas.py
atlas_peer_refresh_zonefile_inventory
def atlas_peer_refresh_zonefile_inventory( my_hostport, peer_hostport, byte_offset, timeout=None, peer_table=None, con=None, path=None, local_inv=None ): """ Refresh a peer's zonefile recent inventory vector entries, by removing every bit after byte_offset and re-synchronizing them. The intuition here is that recent zonefiles are much rarer than older zonefiles (which will have been near-100% replicated), meaning the tail of the peer's zonefile inventory is a lot less stable than the head (since peers will be actively distributing recent zonefiles). NOT THREAD SAFE; CALL FROM ONLY ONE THREAD. Return True if we synced all the way up to the expected inventory length, and update the refresh time in the peer table. Return False if not. """ if timeout is None: timeout = atlas_inv_timeout() if local_inv is None: # get local zonefile inv inv_len = atlasdb_zonefile_inv_length( con=con, path=path ) local_inv = atlas_make_zonefile_inventory( 0, inv_len, con=con, path=path ) maxlen = len(local_inv) with AtlasPeerTableLocked(peer_table) as ptbl: if peer_hostport not in ptbl.keys(): return False # reset the peer's zonefile inventory, back to offset cur_inv = atlas_peer_get_zonefile_inventory( peer_hostport, peer_table=ptbl ) atlas_peer_set_zonefile_inventory( peer_hostport, cur_inv[:byte_offset], peer_table=ptbl ) inv = atlas_peer_sync_zonefile_inventory( my_hostport, peer_hostport, maxlen, timeout=timeout, peer_table=peer_table ) with AtlasPeerTableLocked(peer_table) as ptbl: if peer_hostport not in ptbl.keys(): return False # Update refresh time (even if we fail) ptbl[peer_hostport]['zonefile_inventory_last_refresh'] = time_now() if inv is not None: inv_str = atlas_inventory_to_string(inv) if len(inv_str) > 40: inv_str = inv_str[:40] + "..." log.debug("%s: inventory of %s is now '%s'" % (my_hostport, peer_hostport, inv_str)) if inv is None: return False else: return True
python
def atlas_peer_refresh_zonefile_inventory( my_hostport, peer_hostport, byte_offset, timeout=None, peer_table=None, con=None, path=None, local_inv=None ): if timeout is None: timeout = atlas_inv_timeout() if local_inv is None: # get local zonefile inv inv_len = atlasdb_zonefile_inv_length( con=con, path=path ) local_inv = atlas_make_zonefile_inventory( 0, inv_len, con=con, path=path ) maxlen = len(local_inv) with AtlasPeerTableLocked(peer_table) as ptbl: if peer_hostport not in ptbl.keys(): return False # reset the peer's zonefile inventory, back to offset cur_inv = atlas_peer_get_zonefile_inventory( peer_hostport, peer_table=ptbl ) atlas_peer_set_zonefile_inventory( peer_hostport, cur_inv[:byte_offset], peer_table=ptbl ) inv = atlas_peer_sync_zonefile_inventory( my_hostport, peer_hostport, maxlen, timeout=timeout, peer_table=peer_table ) with AtlasPeerTableLocked(peer_table) as ptbl: if peer_hostport not in ptbl.keys(): return False # Update refresh time (even if we fail) ptbl[peer_hostport]['zonefile_inventory_last_refresh'] = time_now() if inv is not None: inv_str = atlas_inventory_to_string(inv) if len(inv_str) > 40: inv_str = inv_str[:40] + "..." log.debug("%s: inventory of %s is now '%s'" % (my_hostport, peer_hostport, inv_str)) if inv is None: return False else: return True
[ "def", "atlas_peer_refresh_zonefile_inventory", "(", "my_hostport", ",", "peer_hostport", ",", "byte_offset", ",", "timeout", "=", "None", ",", "peer_table", "=", "None", ",", "con", "=", "None", ",", "path", "=", "None", ",", "local_inv", "=", "None", ")", ...
Refresh a peer's zonefile recent inventory vector entries, by removing every bit after byte_offset and re-synchronizing them. The intuition here is that recent zonefiles are much rarer than older zonefiles (which will have been near-100% replicated), meaning the tail of the peer's zonefile inventory is a lot less stable than the head (since peers will be actively distributing recent zonefiles). NOT THREAD SAFE; CALL FROM ONLY ONE THREAD. Return True if we synced all the way up to the expected inventory length, and update the refresh time in the peer table. Return False if not.
[ "Refresh", "a", "peer", "s", "zonefile", "recent", "inventory", "vector", "entries", "by", "removing", "every", "bit", "after", "byte_offset", "and", "re", "-", "synchronizing", "them", "." ]
1dcfdd39b152d29ce13e736a6a1a0981401a0505
https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/atlas.py#L2077-L2131
244,264
blockstack/blockstack-core
blockstack/lib/atlas.py
atlas_peer_has_fresh_zonefile_inventory
def atlas_peer_has_fresh_zonefile_inventory( peer_hostport, peer_table=None ): """ Does the given atlas node have a fresh zonefile inventory? """ fresh = False with AtlasPeerTableLocked(peer_table) as ptbl: if peer_hostport not in ptbl.keys(): return False now = time_now() peer_inv = atlas_peer_get_zonefile_inventory( peer_hostport, peer_table=ptbl ) # NOTE: zero-length or None peer inventory means the peer is simply dead, but we've pinged it if ptbl[peer_hostport].has_key('zonefile_inventory_last_refresh') and \ ptbl[peer_hostport]['zonefile_inventory_last_refresh'] + atlas_peer_ping_interval() > now: fresh = True return fresh
python
def atlas_peer_has_fresh_zonefile_inventory( peer_hostport, peer_table=None ): fresh = False with AtlasPeerTableLocked(peer_table) as ptbl: if peer_hostport not in ptbl.keys(): return False now = time_now() peer_inv = atlas_peer_get_zonefile_inventory( peer_hostport, peer_table=ptbl ) # NOTE: zero-length or None peer inventory means the peer is simply dead, but we've pinged it if ptbl[peer_hostport].has_key('zonefile_inventory_last_refresh') and \ ptbl[peer_hostport]['zonefile_inventory_last_refresh'] + atlas_peer_ping_interval() > now: fresh = True return fresh
[ "def", "atlas_peer_has_fresh_zonefile_inventory", "(", "peer_hostport", ",", "peer_table", "=", "None", ")", ":", "fresh", "=", "False", "with", "AtlasPeerTableLocked", "(", "peer_table", ")", "as", "ptbl", ":", "if", "peer_hostport", "not", "in", "ptbl", ".", "...
Does the given atlas node have a fresh zonefile inventory?
[ "Does", "the", "given", "atlas", "node", "have", "a", "fresh", "zonefile", "inventory?" ]
1dcfdd39b152d29ce13e736a6a1a0981401a0505
https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/atlas.py#L2134-L2153
244,265
blockstack/blockstack-core
blockstack/lib/atlas.py
atlas_find_missing_zonefile_availability
def atlas_find_missing_zonefile_availability( peer_table=None, con=None, path=None, missing_zonefile_info=None ): """ Find the set of missing zonefiles, as well as their popularity amongst our neighbors. Only consider zonefiles that are known by at least one peer; otherwise they're missing from our clique (and we'll re-sync our neighborss' inventories every so often to make sure we detect when zonefiles become available). Return a dict, structured as: { 'zonefile hash': { 'names': [names], 'txid': first txid that set it, 'indexes': [...], 'popularity': ..., 'peers': [...], 'tried_storage': True|False } } """ # which zonefiles do we have? bit_offset = 0 bit_count = 10000 missing = [] ret = {} if missing_zonefile_info is None: while True: zfinfo = atlasdb_zonefile_find_missing( bit_offset, bit_count, con=con, path=path ) if len(zfinfo) == 0: break missing += zfinfo bit_offset += len(zfinfo) if len(missing) > 0: log.debug("Missing %s zonefiles" % len(missing)) else: missing = missing_zonefile_info if len(missing) == 0: # none! return ret with AtlasPeerTableLocked(peer_table) as ptbl: # do any other peers have this zonefile? for zfinfo in missing: popularity = 0 byte_index = (zfinfo['inv_index'] - 1) / 8 bit_index = 7 - ((zfinfo['inv_index'] - 1) % 8) peers = [] if not ret.has_key(zfinfo['zonefile_hash']): ret[zfinfo['zonefile_hash']] = { 'names': [], 'txid': zfinfo['txid'], 'indexes': [], 'block_heights': [], 'popularity': 0, 'peers': [], 'tried_storage': False } for peer_hostport in ptbl.keys(): peer_inv = atlas_peer_get_zonefile_inventory( peer_hostport, peer_table=ptbl ) if len(peer_inv) <= byte_index: # too new for this peer continue if (ord(peer_inv[byte_index]) & (1 << bit_index)) == 0: # this peer doesn't have it continue if peer_hostport not in ret[zfinfo['zonefile_hash']]['peers']: popularity += 1 peers.append( peer_hostport ) ret[zfinfo['zonefile_hash']]['names'].append( zfinfo['name'] ) ret[zfinfo['zonefile_hash']]['indexes'].append( zfinfo['inv_index']-1 ) ret[zfinfo['zonefile_hash']]['block_heights'].append( zfinfo['block_height'] ) ret[zfinfo['zonefile_hash']]['popularity'] += popularity ret[zfinfo['zonefile_hash']]['peers'] += peers ret[zfinfo['zonefile_hash']]['tried_storage'] = zfinfo['tried_storage'] return ret
python
def atlas_find_missing_zonefile_availability( peer_table=None, con=None, path=None, missing_zonefile_info=None ): # which zonefiles do we have? bit_offset = 0 bit_count = 10000 missing = [] ret = {} if missing_zonefile_info is None: while True: zfinfo = atlasdb_zonefile_find_missing( bit_offset, bit_count, con=con, path=path ) if len(zfinfo) == 0: break missing += zfinfo bit_offset += len(zfinfo) if len(missing) > 0: log.debug("Missing %s zonefiles" % len(missing)) else: missing = missing_zonefile_info if len(missing) == 0: # none! return ret with AtlasPeerTableLocked(peer_table) as ptbl: # do any other peers have this zonefile? for zfinfo in missing: popularity = 0 byte_index = (zfinfo['inv_index'] - 1) / 8 bit_index = 7 - ((zfinfo['inv_index'] - 1) % 8) peers = [] if not ret.has_key(zfinfo['zonefile_hash']): ret[zfinfo['zonefile_hash']] = { 'names': [], 'txid': zfinfo['txid'], 'indexes': [], 'block_heights': [], 'popularity': 0, 'peers': [], 'tried_storage': False } for peer_hostport in ptbl.keys(): peer_inv = atlas_peer_get_zonefile_inventory( peer_hostport, peer_table=ptbl ) if len(peer_inv) <= byte_index: # too new for this peer continue if (ord(peer_inv[byte_index]) & (1 << bit_index)) == 0: # this peer doesn't have it continue if peer_hostport not in ret[zfinfo['zonefile_hash']]['peers']: popularity += 1 peers.append( peer_hostport ) ret[zfinfo['zonefile_hash']]['names'].append( zfinfo['name'] ) ret[zfinfo['zonefile_hash']]['indexes'].append( zfinfo['inv_index']-1 ) ret[zfinfo['zonefile_hash']]['block_heights'].append( zfinfo['block_height'] ) ret[zfinfo['zonefile_hash']]['popularity'] += popularity ret[zfinfo['zonefile_hash']]['peers'] += peers ret[zfinfo['zonefile_hash']]['tried_storage'] = zfinfo['tried_storage'] return ret
[ "def", "atlas_find_missing_zonefile_availability", "(", "peer_table", "=", "None", ",", "con", "=", "None", ",", "path", "=", "None", ",", "missing_zonefile_info", "=", "None", ")", ":", "# which zonefiles do we have?", "bit_offset", "=", "0", "bit_count", "=", "1...
Find the set of missing zonefiles, as well as their popularity amongst our neighbors. Only consider zonefiles that are known by at least one peer; otherwise they're missing from our clique (and we'll re-sync our neighborss' inventories every so often to make sure we detect when zonefiles become available). Return a dict, structured as: { 'zonefile hash': { 'names': [names], 'txid': first txid that set it, 'indexes': [...], 'popularity': ..., 'peers': [...], 'tried_storage': True|False } }
[ "Find", "the", "set", "of", "missing", "zonefiles", "as", "well", "as", "their", "popularity", "amongst", "our", "neighbors", "." ]
1dcfdd39b152d29ce13e736a6a1a0981401a0505
https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/atlas.py#L2177-L2266
244,266
blockstack/blockstack-core
blockstack/lib/atlas.py
atlas_peer_has_zonefile
def atlas_peer_has_zonefile( peer_hostport, zonefile_hash, zonefile_bits=None, con=None, path=None, peer_table=None ): """ Does the given peer have the given zonefile defined? Check its inventory vector Return True if present Return False if not present Return None if we don't know about the zonefile ourselves, or if we don't know about the peer """ bits = None if zonefile_bits is None: bits = atlasdb_get_zonefile_bits( zonefile_hash, con=con, path=path ) if len(bits) == 0: return None else: bits = zonefile_bits zonefile_inv = None with AtlasPeerTableLocked(peer_table) as ptbl: if peer_hostport not in ptbl.keys(): return False zonefile_inv = atlas_peer_get_zonefile_inventory( peer_hostport, peer_table=ptbl ) res = atlas_inventory_test_zonefile_bits( zonefile_inv, bits ) return res
python
def atlas_peer_has_zonefile( peer_hostport, zonefile_hash, zonefile_bits=None, con=None, path=None, peer_table=None ): bits = None if zonefile_bits is None: bits = atlasdb_get_zonefile_bits( zonefile_hash, con=con, path=path ) if len(bits) == 0: return None else: bits = zonefile_bits zonefile_inv = None with AtlasPeerTableLocked(peer_table) as ptbl: if peer_hostport not in ptbl.keys(): return False zonefile_inv = atlas_peer_get_zonefile_inventory( peer_hostport, peer_table=ptbl ) res = atlas_inventory_test_zonefile_bits( zonefile_inv, bits ) return res
[ "def", "atlas_peer_has_zonefile", "(", "peer_hostport", ",", "zonefile_hash", ",", "zonefile_bits", "=", "None", ",", "con", "=", "None", ",", "path", "=", "None", ",", "peer_table", "=", "None", ")", ":", "bits", "=", "None", "if", "zonefile_bits", "is", ...
Does the given peer have the given zonefile defined? Check its inventory vector Return True if present Return False if not present Return None if we don't know about the zonefile ourselves, or if we don't know about the peer
[ "Does", "the", "given", "peer", "have", "the", "given", "zonefile", "defined?", "Check", "its", "inventory", "vector" ]
1dcfdd39b152d29ce13e736a6a1a0981401a0505
https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/atlas.py#L2269-L2297
244,267
blockstack/blockstack-core
blockstack/lib/atlas.py
atlas_peer_get_neighbors
def atlas_peer_get_neighbors( my_hostport, peer_hostport, timeout=None, peer_table=None, con=None, path=None ): """ Ask the peer server at the given URL for its neighbors. Update the health info in peer_table (if not given, the global peer table will be used instead) Return the list on success Return None on failure to contact Raise on invalid URL """ if timeout is None: timeout = atlas_neighbors_timeout() peer_list = None host, port = url_to_host_port( peer_hostport ) RPC = get_rpc_client_class() rpc = RPC( host, port, timeout=timeout, src=my_hostport ) # sane limits max_neighbors = atlas_max_neighbors() assert not atlas_peer_table_is_locked_by_me() try: peer_list = blockstack_atlas_peer_exchange( peer_hostport, my_hostport, timeout=timeout, proxy=rpc ) if json_is_exception(peer_list): # fall back to legacy method peer_list = blockstack_get_atlas_peers(peer_hostport, timeout=timeout, proxy=rpc) except (socket.timeout, socket.gaierror, socket.herror, socket.error), se: atlas_log_socket_error( "atlas_peer_exchange(%s)" % peer_hostport, peer_hostport, se) log.error("Socket error in response from '%s'" % peer_hostport) except Exception, e: if os.environ.get("BLOCKSTACK_DEBUG") == "1": log.exception(e) log.error("Failed to talk to '%s'" % peer_hostport) if peer_list is None: log.error("Failed to query remote peer %s" % peer_hostport) atlas_peer_update_health( peer_hostport, False, peer_table=peer_table ) return None if 'error' in peer_list: log.debug("Remote peer error: %s" % peer_list['error']) log.error("Remote peer error on %s" % peer_hostport) atlas_peer_update_health( peer_hostport, False, peer_table=peer_table ) return None ret = peer_list['peers'] atlas_peer_update_health( peer_hostport, True, peer_table=peer_table ) return ret
python
def atlas_peer_get_neighbors( my_hostport, peer_hostport, timeout=None, peer_table=None, con=None, path=None ): if timeout is None: timeout = atlas_neighbors_timeout() peer_list = None host, port = url_to_host_port( peer_hostport ) RPC = get_rpc_client_class() rpc = RPC( host, port, timeout=timeout, src=my_hostport ) # sane limits max_neighbors = atlas_max_neighbors() assert not atlas_peer_table_is_locked_by_me() try: peer_list = blockstack_atlas_peer_exchange( peer_hostport, my_hostport, timeout=timeout, proxy=rpc ) if json_is_exception(peer_list): # fall back to legacy method peer_list = blockstack_get_atlas_peers(peer_hostport, timeout=timeout, proxy=rpc) except (socket.timeout, socket.gaierror, socket.herror, socket.error), se: atlas_log_socket_error( "atlas_peer_exchange(%s)" % peer_hostport, peer_hostport, se) log.error("Socket error in response from '%s'" % peer_hostport) except Exception, e: if os.environ.get("BLOCKSTACK_DEBUG") == "1": log.exception(e) log.error("Failed to talk to '%s'" % peer_hostport) if peer_list is None: log.error("Failed to query remote peer %s" % peer_hostport) atlas_peer_update_health( peer_hostport, False, peer_table=peer_table ) return None if 'error' in peer_list: log.debug("Remote peer error: %s" % peer_list['error']) log.error("Remote peer error on %s" % peer_hostport) atlas_peer_update_health( peer_hostport, False, peer_table=peer_table ) return None ret = peer_list['peers'] atlas_peer_update_health( peer_hostport, True, peer_table=peer_table ) return ret
[ "def", "atlas_peer_get_neighbors", "(", "my_hostport", ",", "peer_hostport", ",", "timeout", "=", "None", ",", "peer_table", "=", "None", ",", "con", "=", "None", ",", "path", "=", "None", ")", ":", "if", "timeout", "is", "None", ":", "timeout", "=", "at...
Ask the peer server at the given URL for its neighbors. Update the health info in peer_table (if not given, the global peer table will be used instead) Return the list on success Return None on failure to contact Raise on invalid URL
[ "Ask", "the", "peer", "server", "at", "the", "given", "URL", "for", "its", "neighbors", "." ]
1dcfdd39b152d29ce13e736a6a1a0981401a0505
https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/atlas.py#L2300-L2354
244,268
blockstack/blockstack-core
blockstack/lib/atlas.py
atlas_get_zonefiles
def atlas_get_zonefiles( my_hostport, peer_hostport, zonefile_hashes, timeout=None, peer_table=None ): """ Given a list of zonefile hashes. go and get them from the given host. Update node health Return the newly-fetched zonefiles on success (as a dict mapping hashes to zonefile data) Return None on error. """ if timeout is None: timeout = atlas_zonefiles_timeout() zf_payload = None zonefile_datas = {} host, port = url_to_host_port( peer_hostport ) RPC = get_rpc_client_class() rpc = RPC( host, port, timeout=timeout, src=my_hostport ) assert not atlas_peer_table_is_locked_by_me() # get in batches of 100 or less zf_batches = [] for i in xrange(0, len(zonefile_hashes), 100): zf_batches.append(zonefile_hashes[i:i+100]) for zf_batch in zf_batches: zf_payload = None try: zf_payload = blockstack_get_zonefiles( peer_hostport, zf_batch, timeout=timeout, my_hostport=my_hostport, proxy=rpc ) except (socket.timeout, socket.gaierror, socket.herror, socket.error), se: atlas_log_socket_error( "get_zonefiles(%s)" % peer_hostport, peer_hostport, se) except Exception, e: if os.environ.get("BLOCKSTACK_DEBUG") is not None: log.exception(e) log.error("Invalid zonefile data from %s" % peer_hostport) if zf_payload is None: log.error("Failed to fetch zonefile data from %s" % peer_hostport) atlas_peer_update_health( peer_hostport, False, peer_table=peer_table ) zonefile_datas = None break if 'error' in zf_payload.keys(): log.error("Failed to fetch zonefile data from %s: %s" % (peer_hostport, zf_payload['error'])) atlas_peer_update_health( peer_hostport, False, peer_table=peer_table ) zonefile_datas = None break # success! zonefile_datas.update( zf_payload['zonefiles'] ) atlas_peer_update_health( peer_hostport, True, peer_table=peer_table ) return zonefile_datas
python
def atlas_get_zonefiles( my_hostport, peer_hostport, zonefile_hashes, timeout=None, peer_table=None ): if timeout is None: timeout = atlas_zonefiles_timeout() zf_payload = None zonefile_datas = {} host, port = url_to_host_port( peer_hostport ) RPC = get_rpc_client_class() rpc = RPC( host, port, timeout=timeout, src=my_hostport ) assert not atlas_peer_table_is_locked_by_me() # get in batches of 100 or less zf_batches = [] for i in xrange(0, len(zonefile_hashes), 100): zf_batches.append(zonefile_hashes[i:i+100]) for zf_batch in zf_batches: zf_payload = None try: zf_payload = blockstack_get_zonefiles( peer_hostport, zf_batch, timeout=timeout, my_hostport=my_hostport, proxy=rpc ) except (socket.timeout, socket.gaierror, socket.herror, socket.error), se: atlas_log_socket_error( "get_zonefiles(%s)" % peer_hostport, peer_hostport, se) except Exception, e: if os.environ.get("BLOCKSTACK_DEBUG") is not None: log.exception(e) log.error("Invalid zonefile data from %s" % peer_hostport) if zf_payload is None: log.error("Failed to fetch zonefile data from %s" % peer_hostport) atlas_peer_update_health( peer_hostport, False, peer_table=peer_table ) zonefile_datas = None break if 'error' in zf_payload.keys(): log.error("Failed to fetch zonefile data from %s: %s" % (peer_hostport, zf_payload['error'])) atlas_peer_update_health( peer_hostport, False, peer_table=peer_table ) zonefile_datas = None break # success! zonefile_datas.update( zf_payload['zonefiles'] ) atlas_peer_update_health( peer_hostport, True, peer_table=peer_table ) return zonefile_datas
[ "def", "atlas_get_zonefiles", "(", "my_hostport", ",", "peer_hostport", ",", "zonefile_hashes", ",", "timeout", "=", "None", ",", "peer_table", "=", "None", ")", ":", "if", "timeout", "is", "None", ":", "timeout", "=", "atlas_zonefiles_timeout", "(", ")", "zf_...
Given a list of zonefile hashes. go and get them from the given host. Update node health Return the newly-fetched zonefiles on success (as a dict mapping hashes to zonefile data) Return None on error.
[ "Given", "a", "list", "of", "zonefile", "hashes", ".", "go", "and", "get", "them", "from", "the", "given", "host", "." ]
1dcfdd39b152d29ce13e736a6a1a0981401a0505
https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/atlas.py#L2357-L2417
244,269
blockstack/blockstack-core
blockstack/lib/atlas.py
atlas_rank_peers_by_data_availability
def atlas_rank_peers_by_data_availability( peer_list=None, peer_table=None, local_inv=None, con=None, path=None ): """ Get a ranking of peers to contact for a zonefile. Peers are ranked by the number of zonefiles they have which we don't have. This is used to select neighbors. """ with AtlasPeerTableLocked(peer_table) as ptbl: if peer_list is None: peer_list = ptbl.keys()[:] if local_inv is None: # what's my inventory? inv_len = atlasdb_zonefile_inv_length( con=con, path=path ) local_inv = atlas_make_zonefile_inventory( 0, inv_len, con=con, path=path ) peer_availability_ranking = [] # (health score, peer hostport) for peer_hostport in peer_list: peer_inv = atlas_peer_get_zonefile_inventory( peer_hostport, peer_table=ptbl ) # ignore peers that we don't have an inventory for if len(peer_inv) == 0: continue availability_score = atlas_inventory_count_missing( local_inv, peer_inv ) peer_availability_ranking.append( (availability_score, peer_hostport) ) # sort on availability peer_availability_ranking.sort() peer_availability_ranking.reverse() return [peer_hp for _, peer_hp in peer_availability_ranking]
python
def atlas_rank_peers_by_data_availability( peer_list=None, peer_table=None, local_inv=None, con=None, path=None ): with AtlasPeerTableLocked(peer_table) as ptbl: if peer_list is None: peer_list = ptbl.keys()[:] if local_inv is None: # what's my inventory? inv_len = atlasdb_zonefile_inv_length( con=con, path=path ) local_inv = atlas_make_zonefile_inventory( 0, inv_len, con=con, path=path ) peer_availability_ranking = [] # (health score, peer hostport) for peer_hostport in peer_list: peer_inv = atlas_peer_get_zonefile_inventory( peer_hostport, peer_table=ptbl ) # ignore peers that we don't have an inventory for if len(peer_inv) == 0: continue availability_score = atlas_inventory_count_missing( local_inv, peer_inv ) peer_availability_ranking.append( (availability_score, peer_hostport) ) # sort on availability peer_availability_ranking.sort() peer_availability_ranking.reverse() return [peer_hp for _, peer_hp in peer_availability_ranking]
[ "def", "atlas_rank_peers_by_data_availability", "(", "peer_list", "=", "None", ",", "peer_table", "=", "None", ",", "local_inv", "=", "None", ",", "con", "=", "None", ",", "path", "=", "None", ")", ":", "with", "AtlasPeerTableLocked", "(", "peer_table", ")", ...
Get a ranking of peers to contact for a zonefile. Peers are ranked by the number of zonefiles they have which we don't have. This is used to select neighbors.
[ "Get", "a", "ranking", "of", "peers", "to", "contact", "for", "a", "zonefile", ".", "Peers", "are", "ranked", "by", "the", "number", "of", "zonefiles", "they", "have", "which", "we", "don", "t", "have", "." ]
1dcfdd39b152d29ce13e736a6a1a0981401a0505
https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/atlas.py#L2453-L2488
244,270
blockstack/blockstack-core
blockstack/lib/atlas.py
atlas_peer_dequeue_all
def atlas_peer_dequeue_all( peer_queue=None ): """ Get all queued peers """ peers = [] with AtlasPeerQueueLocked(peer_queue) as pq: while len(pq) > 0: peers.append( pq.pop(0) ) return peers
python
def atlas_peer_dequeue_all( peer_queue=None ): peers = [] with AtlasPeerQueueLocked(peer_queue) as pq: while len(pq) > 0: peers.append( pq.pop(0) ) return peers
[ "def", "atlas_peer_dequeue_all", "(", "peer_queue", "=", "None", ")", ":", "peers", "=", "[", "]", "with", "AtlasPeerQueueLocked", "(", "peer_queue", ")", "as", "pq", ":", "while", "len", "(", "pq", ")", ">", "0", ":", "peers", ".", "append", "(", "pq"...
Get all queued peers
[ "Get", "all", "queued", "peers" ]
1dcfdd39b152d29ce13e736a6a1a0981401a0505
https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/atlas.py#L2524-L2534
244,271
blockstack/blockstack-core
blockstack/lib/atlas.py
atlas_zonefile_push_enqueue
def atlas_zonefile_push_enqueue( zonefile_hash, name, txid, zonefile_data, zonefile_queue=None, con=None, path=None ): """ Enqueue the given zonefile into our "push" queue, from which it will be replicated to storage and sent out to other peers who don't have it. Return True if we enqueued it Return False if not """ res = False bits = atlasdb_get_zonefile_bits( zonefile_hash, path=path, con=con ) if len(bits) == 0: # invalid hash return with AtlasZonefileQueueLocked(zonefile_queue) as zfq: if len(zfq) < MAX_QUEUED_ZONEFILES: zfdata = { 'zonefile_hash': zonefile_hash, 'zonefile': zonefile_data, 'name': name, 'txid': txid } zfq.append( zfdata ) res = True return res
python
def atlas_zonefile_push_enqueue( zonefile_hash, name, txid, zonefile_data, zonefile_queue=None, con=None, path=None ): res = False bits = atlasdb_get_zonefile_bits( zonefile_hash, path=path, con=con ) if len(bits) == 0: # invalid hash return with AtlasZonefileQueueLocked(zonefile_queue) as zfq: if len(zfq) < MAX_QUEUED_ZONEFILES: zfdata = { 'zonefile_hash': zonefile_hash, 'zonefile': zonefile_data, 'name': name, 'txid': txid } zfq.append( zfdata ) res = True return res
[ "def", "atlas_zonefile_push_enqueue", "(", "zonefile_hash", ",", "name", ",", "txid", ",", "zonefile_data", ",", "zonefile_queue", "=", "None", ",", "con", "=", "None", ",", "path", "=", "None", ")", ":", "res", "=", "False", "bits", "=", "atlasdb_get_zonefi...
Enqueue the given zonefile into our "push" queue, from which it will be replicated to storage and sent out to other peers who don't have it. Return True if we enqueued it Return False if not
[ "Enqueue", "the", "given", "zonefile", "into", "our", "push", "queue", "from", "which", "it", "will", "be", "replicated", "to", "storage", "and", "sent", "out", "to", "other", "peers", "who", "don", "t", "have", "it", "." ]
1dcfdd39b152d29ce13e736a6a1a0981401a0505
https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/atlas.py#L2560-L2589
244,272
blockstack/blockstack-core
blockstack/lib/atlas.py
atlas_zonefile_push_dequeue
def atlas_zonefile_push_dequeue( zonefile_queue=None ): """ Dequeue a zonefile's information to replicate Return None if there are none queued """ ret = None with AtlasZonefileQueueLocked(zonefile_queue) as zfq: if len(zfq) > 0: ret = zfq.pop(0) return ret
python
def atlas_zonefile_push_dequeue( zonefile_queue=None ): ret = None with AtlasZonefileQueueLocked(zonefile_queue) as zfq: if len(zfq) > 0: ret = zfq.pop(0) return ret
[ "def", "atlas_zonefile_push_dequeue", "(", "zonefile_queue", "=", "None", ")", ":", "ret", "=", "None", "with", "AtlasZonefileQueueLocked", "(", "zonefile_queue", ")", "as", "zfq", ":", "if", "len", "(", "zfq", ")", ">", "0", ":", "ret", "=", "zfq", ".", ...
Dequeue a zonefile's information to replicate Return None if there are none queued
[ "Dequeue", "a", "zonefile", "s", "information", "to", "replicate", "Return", "None", "if", "there", "are", "none", "queued" ]
1dcfdd39b152d29ce13e736a6a1a0981401a0505
https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/atlas.py#L2592-L2602
244,273
blockstack/blockstack-core
blockstack/lib/atlas.py
atlas_zonefile_push
def atlas_zonefile_push( my_hostport, peer_hostport, zonefile_data, timeout=None, peer_table=None ): """ Push the given zonefile to the given peer Return True on success Return False on failure """ if timeout is None: timeout = atlas_push_zonefiles_timeout() zonefile_hash = get_zonefile_data_hash(zonefile_data) zonefile_data_b64 = base64.b64encode( zonefile_data ) host, port = url_to_host_port( peer_hostport ) RPC = get_rpc_client_class() rpc = RPC( host, port, timeout=timeout, src=my_hostport ) status = False assert not atlas_peer_table_is_locked_by_me() try: push_info = blockstack_put_zonefiles( peer_hostport, [zonefile_data_b64], timeout=timeout, my_hostport=my_hostport, proxy=rpc ) if 'error' not in push_info: if push_info['saved'] == 1: # woo! saved = True except (socket.timeout, socket.gaierror, socket.herror, socket.error), se: atlas_log_socket_error( "put_zonefiles(%s)" % peer_hostport, peer_hostport, se) except AssertionError, ae: log.exception(ae) log.error("Invalid server response from %s" % peer_hostport ) except Exception, e: log.exception(e) log.error("Failed to push zonefile %s to %s" % (zonefile_hash, peer_hostport)) with AtlasPeerTableLocked(peer_table) as ptbl: atlas_peer_update_health( peer_hostport, status, peer_table=ptbl ) return status
python
def atlas_zonefile_push( my_hostport, peer_hostport, zonefile_data, timeout=None, peer_table=None ): if timeout is None: timeout = atlas_push_zonefiles_timeout() zonefile_hash = get_zonefile_data_hash(zonefile_data) zonefile_data_b64 = base64.b64encode( zonefile_data ) host, port = url_to_host_port( peer_hostport ) RPC = get_rpc_client_class() rpc = RPC( host, port, timeout=timeout, src=my_hostport ) status = False assert not atlas_peer_table_is_locked_by_me() try: push_info = blockstack_put_zonefiles( peer_hostport, [zonefile_data_b64], timeout=timeout, my_hostport=my_hostport, proxy=rpc ) if 'error' not in push_info: if push_info['saved'] == 1: # woo! saved = True except (socket.timeout, socket.gaierror, socket.herror, socket.error), se: atlas_log_socket_error( "put_zonefiles(%s)" % peer_hostport, peer_hostport, se) except AssertionError, ae: log.exception(ae) log.error("Invalid server response from %s" % peer_hostport ) except Exception, e: log.exception(e) log.error("Failed to push zonefile %s to %s" % (zonefile_hash, peer_hostport)) with AtlasPeerTableLocked(peer_table) as ptbl: atlas_peer_update_health( peer_hostport, status, peer_table=ptbl ) return status
[ "def", "atlas_zonefile_push", "(", "my_hostport", ",", "peer_hostport", ",", "zonefile_data", ",", "timeout", "=", "None", ",", "peer_table", "=", "None", ")", ":", "if", "timeout", "is", "None", ":", "timeout", "=", "atlas_push_zonefiles_timeout", "(", ")", "...
Push the given zonefile to the given peer Return True on success Return False on failure
[ "Push", "the", "given", "zonefile", "to", "the", "given", "peer", "Return", "True", "on", "success", "Return", "False", "on", "failure" ]
1dcfdd39b152d29ce13e736a6a1a0981401a0505
https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/atlas.py#L2605-L2646
244,274
blockstack/blockstack-core
blockstack/lib/atlas.py
atlas_node_init
def atlas_node_init(my_hostname, my_portnum, atlasdb_path, zonefile_dir, working_dir): """ Start up the atlas node. Return a bundle of atlas state """ atlas_state = {} atlas_state['peer_crawler'] = AtlasPeerCrawler(my_hostname, my_portnum, atlasdb_path, working_dir) atlas_state['health_checker'] = AtlasHealthChecker(my_hostname, my_portnum, atlasdb_path) atlas_state['zonefile_crawler'] = AtlasZonefileCrawler(my_hostname, my_portnum, atlasdb_path, zonefile_dir) # atlas_state['zonefile_pusher'] = AtlasZonefilePusher(my_hostname, my_portnum, atlasdb_path, zonefile_dir) return atlas_state
python
def atlas_node_init(my_hostname, my_portnum, atlasdb_path, zonefile_dir, working_dir): atlas_state = {} atlas_state['peer_crawler'] = AtlasPeerCrawler(my_hostname, my_portnum, atlasdb_path, working_dir) atlas_state['health_checker'] = AtlasHealthChecker(my_hostname, my_portnum, atlasdb_path) atlas_state['zonefile_crawler'] = AtlasZonefileCrawler(my_hostname, my_portnum, atlasdb_path, zonefile_dir) # atlas_state['zonefile_pusher'] = AtlasZonefilePusher(my_hostname, my_portnum, atlasdb_path, zonefile_dir) return atlas_state
[ "def", "atlas_node_init", "(", "my_hostname", ",", "my_portnum", ",", "atlasdb_path", ",", "zonefile_dir", ",", "working_dir", ")", ":", "atlas_state", "=", "{", "}", "atlas_state", "[", "'peer_crawler'", "]", "=", "AtlasPeerCrawler", "(", "my_hostname", ",", "m...
Start up the atlas node. Return a bundle of atlas state
[ "Start", "up", "the", "atlas", "node", ".", "Return", "a", "bundle", "of", "atlas", "state" ]
1dcfdd39b152d29ce13e736a6a1a0981401a0505
https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/atlas.py#L3584-L3595
244,275
blockstack/blockstack-core
blockstack/lib/atlas.py
atlas_node_start
def atlas_node_start(atlas_state): """ Start up atlas threads """ for component in atlas_state.keys(): log.debug("Starting Atlas component '%s'" % component) atlas_state[component].start()
python
def atlas_node_start(atlas_state): for component in atlas_state.keys(): log.debug("Starting Atlas component '%s'" % component) atlas_state[component].start()
[ "def", "atlas_node_start", "(", "atlas_state", ")", ":", "for", "component", "in", "atlas_state", ".", "keys", "(", ")", ":", "log", ".", "debug", "(", "\"Starting Atlas component '%s'\"", "%", "component", ")", "atlas_state", "[", "component", "]", ".", "star...
Start up atlas threads
[ "Start", "up", "atlas", "threads" ]
1dcfdd39b152d29ce13e736a6a1a0981401a0505
https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/atlas.py#L3597-L3603
244,276
blockstack/blockstack-core
blockstack/lib/atlas.py
atlas_node_add_callback
def atlas_node_add_callback(atlas_state, callback_name, callback): """ Add a callback to the initialized atlas state """ if callback_name == 'store_zonefile': atlas_state['zonefile_crawler'].set_store_zonefile_callback(callback) else: raise ValueError("Unrecognized callback {}".format(callback_name))
python
def atlas_node_add_callback(atlas_state, callback_name, callback): if callback_name == 'store_zonefile': atlas_state['zonefile_crawler'].set_store_zonefile_callback(callback) else: raise ValueError("Unrecognized callback {}".format(callback_name))
[ "def", "atlas_node_add_callback", "(", "atlas_state", ",", "callback_name", ",", "callback", ")", ":", "if", "callback_name", "==", "'store_zonefile'", ":", "atlas_state", "[", "'zonefile_crawler'", "]", ".", "set_store_zonefile_callback", "(", "callback", ")", "else"...
Add a callback to the initialized atlas state
[ "Add", "a", "callback", "to", "the", "initialized", "atlas", "state" ]
1dcfdd39b152d29ce13e736a6a1a0981401a0505
https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/atlas.py#L3606-L3614
244,277
blockstack/blockstack-core
blockstack/lib/atlas.py
atlas_node_stop
def atlas_node_stop( atlas_state ): """ Stop the atlas node threads """ for component in atlas_state.keys(): log.debug("Stopping Atlas component '%s'" % component) atlas_state[component].ask_join() atlas_state[component].join() return True
python
def atlas_node_stop( atlas_state ): for component in atlas_state.keys(): log.debug("Stopping Atlas component '%s'" % component) atlas_state[component].ask_join() atlas_state[component].join() return True
[ "def", "atlas_node_stop", "(", "atlas_state", ")", ":", "for", "component", "in", "atlas_state", ".", "keys", "(", ")", ":", "log", ".", "debug", "(", "\"Stopping Atlas component '%s'\"", "%", "component", ")", "atlas_state", "[", "component", "]", ".", "ask_j...
Stop the atlas node threads
[ "Stop", "the", "atlas", "node", "threads" ]
1dcfdd39b152d29ce13e736a6a1a0981401a0505
https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/atlas.py#L3617-L3626
244,278
blockstack/blockstack-core
blockstack/lib/atlas.py
AtlasPeerCrawler.canonical_peer
def canonical_peer( self, peer ): """ Get the canonical peer name """ their_host, their_port = url_to_host_port( peer ) if their_host in ['127.0.0.1', '::1']: their_host = 'localhost' return "%s:%s" % (their_host, their_port)
python
def canonical_peer( self, peer ): their_host, their_port = url_to_host_port( peer ) if their_host in ['127.0.0.1', '::1']: their_host = 'localhost' return "%s:%s" % (their_host, their_port)
[ "def", "canonical_peer", "(", "self", ",", "peer", ")", ":", "their_host", ",", "their_port", "=", "url_to_host_port", "(", "peer", ")", "if", "their_host", "in", "[", "'127.0.0.1'", ",", "'::1'", "]", ":", "their_host", "=", "'localhost'", "return", "\"%s:%...
Get the canonical peer name
[ "Get", "the", "canonical", "peer", "name" ]
1dcfdd39b152d29ce13e736a6a1a0981401a0505
https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/atlas.py#L2713-L2722
244,279
blockstack/blockstack-core
blockstack/lib/atlas.py
AtlasPeerCrawler.remove_unhealthy_peers
def remove_unhealthy_peers( self, count, con=None, path=None, peer_table=None, min_request_count=10, min_health=MIN_PEER_HEALTH ): """ Remove up to @count unhealthy peers Return the list of peers we removed """ if path is None: path = self.atlasdb_path removed = [] rank_peer_list = atlas_rank_peers_by_health( peer_table=peer_table, with_rank=True ) for rank, peer in rank_peer_list: reqcount = atlas_peer_get_request_count( peer, peer_table=peer_table ) if reqcount >= min_request_count and rank < min_health and not atlas_peer_is_whitelisted( peer, peer_table=peer_table ) and not atlas_peer_is_blacklisted( peer, peer_table=peer_table ): removed.append( peer ) random.shuffle(removed) if len(removed) > count: removed = removed[:count] for peer in removed: log.debug("Remove unhealthy peer %s" % (peer)) atlasdb_remove_peer( peer, con=con, path=path, peer_table=peer_table ) return removed
python
def remove_unhealthy_peers( self, count, con=None, path=None, peer_table=None, min_request_count=10, min_health=MIN_PEER_HEALTH ): if path is None: path = self.atlasdb_path removed = [] rank_peer_list = atlas_rank_peers_by_health( peer_table=peer_table, with_rank=True ) for rank, peer in rank_peer_list: reqcount = atlas_peer_get_request_count( peer, peer_table=peer_table ) if reqcount >= min_request_count and rank < min_health and not atlas_peer_is_whitelisted( peer, peer_table=peer_table ) and not atlas_peer_is_blacklisted( peer, peer_table=peer_table ): removed.append( peer ) random.shuffle(removed) if len(removed) > count: removed = removed[:count] for peer in removed: log.debug("Remove unhealthy peer %s" % (peer)) atlasdb_remove_peer( peer, con=con, path=path, peer_table=peer_table ) return removed
[ "def", "remove_unhealthy_peers", "(", "self", ",", "count", ",", "con", "=", "None", ",", "path", "=", "None", ",", "peer_table", "=", "None", ",", "min_request_count", "=", "10", ",", "min_health", "=", "MIN_PEER_HEALTH", ")", ":", "if", "path", "is", "...
Remove up to @count unhealthy peers Return the list of peers we removed
[ "Remove", "up", "to" ]
1dcfdd39b152d29ce13e736a6a1a0981401a0505
https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/atlas.py#L2836-L2860
244,280
blockstack/blockstack-core
blockstack/lib/atlas.py
AtlasPeerCrawler.get_current_peers
def get_current_peers( self, peer_table=None ): """ Get the current set of peers """ # get current peers current_peers = None with AtlasPeerTableLocked(peer_table) as ptbl: current_peers = ptbl.keys()[:] return current_peers
python
def get_current_peers( self, peer_table=None ): # get current peers current_peers = None with AtlasPeerTableLocked(peer_table) as ptbl: current_peers = ptbl.keys()[:] return current_peers
[ "def", "get_current_peers", "(", "self", ",", "peer_table", "=", "None", ")", ":", "# get current peers", "current_peers", "=", "None", "with", "AtlasPeerTableLocked", "(", "peer_table", ")", "as", "ptbl", ":", "current_peers", "=", "ptbl", ".", "keys", "(", "...
Get the current set of peers
[ "Get", "the", "current", "set", "of", "peers" ]
1dcfdd39b152d29ce13e736a6a1a0981401a0505
https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/atlas.py#L2945-L2955
244,281
blockstack/blockstack-core
blockstack/lib/atlas.py
AtlasPeerCrawler.canonical_new_peer_list
def canonical_new_peer_list( self, peers_to_add ): """ Make a list of canonical new peers, using the self.new_peers and the given peers to add Return a shuffled list of canonicalized host:port strings. """ new_peers = list(set(self.new_peers + peers_to_add)) random.shuffle( new_peers ) # canonicalize tmp = [] for peer in new_peers: tmp.append( self.canonical_peer(peer) ) new_peers = tmp # don't talk to myself if self.my_hostport in new_peers: new_peers.remove(self.my_hostport) return new_peers
python
def canonical_new_peer_list( self, peers_to_add ): new_peers = list(set(self.new_peers + peers_to_add)) random.shuffle( new_peers ) # canonicalize tmp = [] for peer in new_peers: tmp.append( self.canonical_peer(peer) ) new_peers = tmp # don't talk to myself if self.my_hostport in new_peers: new_peers.remove(self.my_hostport) return new_peers
[ "def", "canonical_new_peer_list", "(", "self", ",", "peers_to_add", ")", ":", "new_peers", "=", "list", "(", "set", "(", "self", ".", "new_peers", "+", "peers_to_add", ")", ")", "random", ".", "shuffle", "(", "new_peers", ")", "# canonicalize", "tmp", "=", ...
Make a list of canonical new peers, using the self.new_peers and the given peers to add Return a shuffled list of canonicalized host:port strings.
[ "Make", "a", "list", "of", "canonical", "new", "peers", "using", "the", "self", ".", "new_peers", "and", "the", "given", "peers", "to", "add" ]
1dcfdd39b152d29ce13e736a6a1a0981401a0505
https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/atlas.py#L2958-L2980
244,282
blockstack/blockstack-core
blockstack/lib/atlas.py
AtlasHealthChecker.step
def step(self, con=None, path=None, peer_table=None, local_inv=None): """ Find peers with stale zonefile inventory data, and refresh them. Return True on success Return False on error """ if path is None: path = self.atlasdb_path peer_hostports = [] stale_peers = [] num_peers = None peer_hostports = None with AtlasPeerTableLocked(peer_table) as ptbl: num_peers = len(ptbl.keys()) peer_hostports = ptbl.keys()[:] # who are we going to ping? # someone we haven't pinged in a while, chosen at random for peer in peer_hostports: if not atlas_peer_has_fresh_zonefile_inventory( peer, peer_table=ptbl ): # haven't talked to this peer in a while stale_peers.append(peer) log.debug("Peer %s has a stale zonefile inventory" % peer) if len(stale_peers) > 0: log.debug("Refresh zonefile inventories for %s peers" % len(stale_peers)) for peer_hostport in stale_peers: # refresh everyone log.debug("%s: Refresh zonefile inventory for %s" % (self.hostport, peer_hostport)) res = atlas_peer_refresh_zonefile_inventory( self.hostport, peer_hostport, 0, con=con, path=path, peer_table=peer_table, local_inv=local_inv ) if res is None: log.warning("Failed to refresh zonefile inventory for %s" % peer_hostport) return
python
def step(self, con=None, path=None, peer_table=None, local_inv=None): if path is None: path = self.atlasdb_path peer_hostports = [] stale_peers = [] num_peers = None peer_hostports = None with AtlasPeerTableLocked(peer_table) as ptbl: num_peers = len(ptbl.keys()) peer_hostports = ptbl.keys()[:] # who are we going to ping? # someone we haven't pinged in a while, chosen at random for peer in peer_hostports: if not atlas_peer_has_fresh_zonefile_inventory( peer, peer_table=ptbl ): # haven't talked to this peer in a while stale_peers.append(peer) log.debug("Peer %s has a stale zonefile inventory" % peer) if len(stale_peers) > 0: log.debug("Refresh zonefile inventories for %s peers" % len(stale_peers)) for peer_hostport in stale_peers: # refresh everyone log.debug("%s: Refresh zonefile inventory for %s" % (self.hostport, peer_hostport)) res = atlas_peer_refresh_zonefile_inventory( self.hostport, peer_hostport, 0, con=con, path=path, peer_table=peer_table, local_inv=local_inv ) if res is None: log.warning("Failed to refresh zonefile inventory for %s" % peer_hostport) return
[ "def", "step", "(", "self", ",", "con", "=", "None", ",", "path", "=", "None", ",", "peer_table", "=", "None", ",", "local_inv", "=", "None", ")", ":", "if", "path", "is", "None", ":", "path", "=", "self", ".", "atlasdb_path", "peer_hostports", "=", ...
Find peers with stale zonefile inventory data, and refresh them. Return True on success Return False on error
[ "Find", "peers", "with", "stale", "zonefile", "inventory", "data", "and", "refresh", "them", "." ]
1dcfdd39b152d29ce13e736a6a1a0981401a0505
https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/atlas.py#L3171-L3210
244,283
blockstack/blockstack-core
blockstack/lib/atlas.py
AtlasHealthChecker.run
def run(self, peer_table=None): """ Loop forever, pinging someone every pass. """ self.running = True while self.running: local_inv = atlas_get_zonefile_inventory() t1 = time_now() self.step( peer_table=peer_table, local_inv=local_inv, path=self.atlasdb_path ) t2 = time_now() # don't go too fast if t2 - t1 < PEER_HEALTH_NEIGHBOR_WORK_INTERVAL: deadline = time_now() + PEER_HEALTH_NEIGHBOR_WORK_INTERVAL - (t2 - t1) while time_now() < deadline and self.running: time_sleep( self.hostport, self.__class__.__name__, 1.0 ) if not self.running: break
python
def run(self, peer_table=None): self.running = True while self.running: local_inv = atlas_get_zonefile_inventory() t1 = time_now() self.step( peer_table=peer_table, local_inv=local_inv, path=self.atlasdb_path ) t2 = time_now() # don't go too fast if t2 - t1 < PEER_HEALTH_NEIGHBOR_WORK_INTERVAL: deadline = time_now() + PEER_HEALTH_NEIGHBOR_WORK_INTERVAL - (t2 - t1) while time_now() < deadline and self.running: time_sleep( self.hostport, self.__class__.__name__, 1.0 ) if not self.running: break
[ "def", "run", "(", "self", ",", "peer_table", "=", "None", ")", ":", "self", ".", "running", "=", "True", "while", "self", ".", "running", ":", "local_inv", "=", "atlas_get_zonefile_inventory", "(", ")", "t1", "=", "time_now", "(", ")", "self", ".", "s...
Loop forever, pinging someone every pass.
[ "Loop", "forever", "pinging", "someone", "every", "pass", "." ]
1dcfdd39b152d29ce13e736a6a1a0981401a0505
https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/atlas.py#L3213-L3231
244,284
blockstack/blockstack-core
blockstack/lib/atlas.py
AtlasZonefileCrawler.set_zonefile_present
def set_zonefile_present(self, zfhash, block_height, con=None, path=None): """ Set a zonefile as present, and if it was previously absent, inform the storage listener """ was_present = atlasdb_set_zonefile_present( zfhash, True, con=con, path=path ) # tell anyone who cares that we got this zone file, if it was new if not was_present and self.store_zonefile_cb: log.debug('{} was new, so passing it along to zonefile storage watchers...'.format(zfhash)) self.store_zonefile_cb(zfhash, block_height) else: log.debug('{} was seen before, so not passing it along to zonefile storage watchers'.format(zfhash))
python
def set_zonefile_present(self, zfhash, block_height, con=None, path=None): was_present = atlasdb_set_zonefile_present( zfhash, True, con=con, path=path ) # tell anyone who cares that we got this zone file, if it was new if not was_present and self.store_zonefile_cb: log.debug('{} was new, so passing it along to zonefile storage watchers...'.format(zfhash)) self.store_zonefile_cb(zfhash, block_height) else: log.debug('{} was seen before, so not passing it along to zonefile storage watchers'.format(zfhash))
[ "def", "set_zonefile_present", "(", "self", ",", "zfhash", ",", "block_height", ",", "con", "=", "None", ",", "path", "=", "None", ")", ":", "was_present", "=", "atlasdb_set_zonefile_present", "(", "zfhash", ",", "True", ",", "con", "=", "con", ",", "path"...
Set a zonefile as present, and if it was previously absent, inform the storage listener
[ "Set", "a", "zonefile", "as", "present", "and", "if", "it", "was", "previously", "absent", "inform", "the", "storage", "listener" ]
1dcfdd39b152d29ce13e736a6a1a0981401a0505
https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/atlas.py#L3257-L3268
244,285
blockstack/blockstack-core
blockstack/lib/atlas.py
AtlasZonefileCrawler.find_zonefile_origins
def find_zonefile_origins( self, missing_zfinfo, peer_hostports ): """ Find out which peers can serve which zonefiles """ zonefile_origins = {} # map peer hostport to list of zonefile hashes # which peers can serve each zonefile? for zfhash in missing_zfinfo.keys(): for peer_hostport in peer_hostports: if not zonefile_origins.has_key(peer_hostport): zonefile_origins[peer_hostport] = [] if peer_hostport in missing_zfinfo[zfhash]['peers']: zonefile_origins[peer_hostport].append( zfhash ) return zonefile_origins
python
def find_zonefile_origins( self, missing_zfinfo, peer_hostports ): zonefile_origins = {} # map peer hostport to list of zonefile hashes # which peers can serve each zonefile? for zfhash in missing_zfinfo.keys(): for peer_hostport in peer_hostports: if not zonefile_origins.has_key(peer_hostport): zonefile_origins[peer_hostport] = [] if peer_hostport in missing_zfinfo[zfhash]['peers']: zonefile_origins[peer_hostport].append( zfhash ) return zonefile_origins
[ "def", "find_zonefile_origins", "(", "self", ",", "missing_zfinfo", ",", "peer_hostports", ")", ":", "zonefile_origins", "=", "{", "}", "# map peer hostport to list of zonefile hashes", "# which peers can serve each zonefile?", "for", "zfhash", "in", "missing_zfinfo", ".", ...
Find out which peers can serve which zonefiles
[ "Find", "out", "which", "peers", "can", "serve", "which", "zonefiles" ]
1dcfdd39b152d29ce13e736a6a1a0981401a0505
https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/atlas.py#L3316-L3331
244,286
blockstack/blockstack-core
blockstack/lib/atlas.py
AtlasZonefilePusher.step
def step( self, peer_table=None, zonefile_queue=None, path=None ): """ Run one step of this algorithm. Push the zonefile to all the peers that need it. Return the number of peers we sent to """ if path is None: path = self.atlasdb_path if BLOCKSTACK_TEST: log.debug("%s: %s step" % (self.hostport, self.__class__.__name__)) if self.push_timeout is None: self.push_timeout = atlas_push_zonefiles_timeout() zfinfo = atlas_zonefile_push_dequeue( zonefile_queue=zonefile_queue ) if zfinfo is None: return 0 zfhash = zfinfo['zonefile_hash'] zfdata_txt = zfinfo['zonefile'] name = zfinfo['name'] txid = zfinfo['txid'] zfbits = atlasdb_get_zonefile_bits( zfhash, path=path ) if len(zfbits) == 0: # not recognized return 0 # it's a valid zonefile. store it. rc = add_atlas_zonefile_data( str(zfdata_txt), self.zonefile_dir ) if not rc: log.error("Failed to replicate zonefile %s to external storage" % zfhash) peers = None # see if we can send this somewhere with AtlasPeerTableLocked(peer_table) as ptbl: peers = atlas_zonefile_find_push_peers( zfhash, peer_table=ptbl, zonefile_bits=zfbits ) if len(peers) == 0: # everyone has it log.debug("%s: All peers have zonefile %s" % (self.hostport, zfhash)) return 0 # push it off ret = 0 for peer in peers: log.debug("%s: Push to %s" % (self.hostport, peer)) atlas_zonefile_push( self.hostport, peer, zfdata_txt, timeout=self.push_timeout ) ret += 1 return ret
python
def step( self, peer_table=None, zonefile_queue=None, path=None ): if path is None: path = self.atlasdb_path if BLOCKSTACK_TEST: log.debug("%s: %s step" % (self.hostport, self.__class__.__name__)) if self.push_timeout is None: self.push_timeout = atlas_push_zonefiles_timeout() zfinfo = atlas_zonefile_push_dequeue( zonefile_queue=zonefile_queue ) if zfinfo is None: return 0 zfhash = zfinfo['zonefile_hash'] zfdata_txt = zfinfo['zonefile'] name = zfinfo['name'] txid = zfinfo['txid'] zfbits = atlasdb_get_zonefile_bits( zfhash, path=path ) if len(zfbits) == 0: # not recognized return 0 # it's a valid zonefile. store it. rc = add_atlas_zonefile_data( str(zfdata_txt), self.zonefile_dir ) if not rc: log.error("Failed to replicate zonefile %s to external storage" % zfhash) peers = None # see if we can send this somewhere with AtlasPeerTableLocked(peer_table) as ptbl: peers = atlas_zonefile_find_push_peers( zfhash, peer_table=ptbl, zonefile_bits=zfbits ) if len(peers) == 0: # everyone has it log.debug("%s: All peers have zonefile %s" % (self.hostport, zfhash)) return 0 # push it off ret = 0 for peer in peers: log.debug("%s: Push to %s" % (self.hostport, peer)) atlas_zonefile_push( self.hostport, peer, zfdata_txt, timeout=self.push_timeout ) ret += 1 return ret
[ "def", "step", "(", "self", ",", "peer_table", "=", "None", ",", "zonefile_queue", "=", "None", ",", "path", "=", "None", ")", ":", "if", "path", "is", "None", ":", "path", "=", "self", ".", "atlasdb_path", "if", "BLOCKSTACK_TEST", ":", "log", ".", "...
Run one step of this algorithm. Push the zonefile to all the peers that need it. Return the number of peers we sent to
[ "Run", "one", "step", "of", "this", "algorithm", ".", "Push", "the", "zonefile", "to", "all", "the", "peers", "that", "need", "it", ".", "Return", "the", "number", "of", "peers", "we", "sent", "to" ]
1dcfdd39b152d29ce13e736a6a1a0981401a0505
https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/atlas.py#L3509-L3561
244,287
blockstack/blockstack-core
blockstack/lib/queue.py
queuedb_create
def queuedb_create(path): """ Create a sqlite3 db at the given path. Create all the tables and indexes we need. Raises if the table already exists """ global QUEUE_SQL, ERROR_SQL lines = [l + ";" for l in QUEUE_SQL.split(";")] con = sqlite3.connect( path, isolation_level=None ) db_query_execute(con, 'pragma mmap_size=536870912', ()) for line in lines: db_query_execute(con, line, ()) con.commit() con.row_factory = queuedb_row_factory return con
python
def queuedb_create(path): global QUEUE_SQL, ERROR_SQL lines = [l + ";" for l in QUEUE_SQL.split(";")] con = sqlite3.connect( path, isolation_level=None ) db_query_execute(con, 'pragma mmap_size=536870912', ()) for line in lines: db_query_execute(con, line, ()) con.commit() con.row_factory = queuedb_row_factory return con
[ "def", "queuedb_create", "(", "path", ")", ":", "global", "QUEUE_SQL", ",", "ERROR_SQL", "lines", "=", "[", "l", "+", "\";\"", "for", "l", "in", "QUEUE_SQL", ".", "split", "(", "\";\"", ")", "]", "con", "=", "sqlite3", ".", "connect", "(", "path", ",...
Create a sqlite3 db at the given path. Create all the tables and indexes we need. Raises if the table already exists
[ "Create", "a", "sqlite3", "db", "at", "the", "given", "path", ".", "Create", "all", "the", "tables", "and", "indexes", "we", "need", ".", "Raises", "if", "the", "table", "already", "exists" ]
1dcfdd39b152d29ce13e736a6a1a0981401a0505
https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/queue.py#L46-L63
244,288
blockstack/blockstack-core
blockstack/lib/queue.py
queuedb_row_factory
def queuedb_row_factory(cursor, row): """ Dict row factory """ d = {} for idx, col in enumerate(cursor.description): d[col[0]] = row[idx] return d
python
def queuedb_row_factory(cursor, row): d = {} for idx, col in enumerate(cursor.description): d[col[0]] = row[idx] return d
[ "def", "queuedb_row_factory", "(", "cursor", ",", "row", ")", ":", "d", "=", "{", "}", "for", "idx", ",", "col", "in", "enumerate", "(", "cursor", ".", "description", ")", ":", "d", "[", "col", "[", "0", "]", "]", "=", "row", "[", "idx", "]", "...
Dict row factory
[ "Dict", "row", "factory" ]
1dcfdd39b152d29ce13e736a6a1a0981401a0505
https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/queue.py#L75-L83
244,289
blockstack/blockstack-core
blockstack/lib/queue.py
queuedb_findall
def queuedb_findall(path, queue_id, name=None, offset=None, limit=None): """ Get all queued entries for a queue and a name. If name is None, then find all queue entries Return the rows on success (empty list if not found) Raise on error """ sql = "SELECT * FROM queue WHERE queue_id = ? ORDER BY rowid ASC" args = (queue_id,) if name: sql += ' AND name = ?' args += (name,) if limit: sql += ' LIMIT ?' args += (limit,) if offset: sql += ' OFFSET ?' args += (offset,) sql += ';' db = queuedb_open(path) if db is None: raise Exception("Failed to open %s" % path) cur = db.cursor() rows = queuedb_query_execute(cur, sql, args) count = 0 ret = [] for row in rows: dat = {} dat.update(row) ret.append(dat) db.close() return ret
python
def queuedb_findall(path, queue_id, name=None, offset=None, limit=None): sql = "SELECT * FROM queue WHERE queue_id = ? ORDER BY rowid ASC" args = (queue_id,) if name: sql += ' AND name = ?' args += (name,) if limit: sql += ' LIMIT ?' args += (limit,) if offset: sql += ' OFFSET ?' args += (offset,) sql += ';' db = queuedb_open(path) if db is None: raise Exception("Failed to open %s" % path) cur = db.cursor() rows = queuedb_query_execute(cur, sql, args) count = 0 ret = [] for row in rows: dat = {} dat.update(row) ret.append(dat) db.close() return ret
[ "def", "queuedb_findall", "(", "path", ",", "queue_id", ",", "name", "=", "None", ",", "offset", "=", "None", ",", "limit", "=", "None", ")", ":", "sql", "=", "\"SELECT * FROM queue WHERE queue_id = ? ORDER BY rowid ASC\"", "args", "=", "(", "queue_id", ",", "...
Get all queued entries for a queue and a name. If name is None, then find all queue entries Return the rows on success (empty list if not found) Raise on error
[ "Get", "all", "queued", "entries", "for", "a", "queue", "and", "a", "name", ".", "If", "name", "is", "None", "then", "find", "all", "queue", "entries" ]
1dcfdd39b152d29ce13e736a6a1a0981401a0505
https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/queue.py#L103-L143
244,290
blockstack/blockstack-core
blockstack/lib/queue.py
queuedb_append
def queuedb_append(path, queue_id, name, data): """ Append an element to the back of the queue. Return True on success Raise on error """ sql = "INSERT INTO queue VALUES (?,?,?);" args = (name, queue_id, data) db = queuedb_open(path) if db is None: raise Exception("Failed to open %s" % path) cur = db.cursor() res = queuedb_query_execute(cur, sql, args) db.commit() db.close() return True
python
def queuedb_append(path, queue_id, name, data): sql = "INSERT INTO queue VALUES (?,?,?);" args = (name, queue_id, data) db = queuedb_open(path) if db is None: raise Exception("Failed to open %s" % path) cur = db.cursor() res = queuedb_query_execute(cur, sql, args) db.commit() db.close() return True
[ "def", "queuedb_append", "(", "path", ",", "queue_id", ",", "name", ",", "data", ")", ":", "sql", "=", "\"INSERT INTO queue VALUES (?,?,?);\"", "args", "=", "(", "name", ",", "queue_id", ",", "data", ")", "db", "=", "queuedb_open", "(", "path", ")", "if", ...
Append an element to the back of the queue. Return True on success Raise on error
[ "Append", "an", "element", "to", "the", "back", "of", "the", "queue", ".", "Return", "True", "on", "success", "Raise", "on", "error" ]
1dcfdd39b152d29ce13e736a6a1a0981401a0505
https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/queue.py#L146-L164
244,291
blockstack/blockstack-core
blockstack/lib/queue.py
queuedb_remove
def queuedb_remove(path, entry, cur=None): """ Remove an element from a queue. Return True on success Raise on error """ sql = "DELETE FROM queue WHERE queue_id = ? AND name = ?;" args = (entry['queue_id'], entry['name']) cursor = None if cur: cursor = cur else: db = queuedb_open(path) if db is None: raise Exception("Failed to open %s" % path) cursor = db.cursor() res = queuedb_query_execute(cursor, sql, args) if cur is None: db.commit() db.close() return True
python
def queuedb_remove(path, entry, cur=None): sql = "DELETE FROM queue WHERE queue_id = ? AND name = ?;" args = (entry['queue_id'], entry['name']) cursor = None if cur: cursor = cur else: db = queuedb_open(path) if db is None: raise Exception("Failed to open %s" % path) cursor = db.cursor() res = queuedb_query_execute(cursor, sql, args) if cur is None: db.commit() db.close() return True
[ "def", "queuedb_remove", "(", "path", ",", "entry", ",", "cur", "=", "None", ")", ":", "sql", "=", "\"DELETE FROM queue WHERE queue_id = ? AND name = ?;\"", "args", "=", "(", "entry", "[", "'queue_id'", "]", ",", "entry", "[", "'name'", "]", ")", "cursor", "...
Remove an element from a queue. Return True on success Raise on error
[ "Remove", "an", "element", "from", "a", "queue", ".", "Return", "True", "on", "success", "Raise", "on", "error" ]
1dcfdd39b152d29ce13e736a6a1a0981401a0505
https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/queue.py#L167-L192
244,292
blockstack/blockstack-core
blockstack/lib/queue.py
queuedb_removeall
def queuedb_removeall(path, entries): """ Remove all entries from a queue """ db = queuedb_open(path) if db is None: raise Exception("Failed to open %s" % path) cursor = db.cursor() queuedb_query_execute(cursor, 'BEGIN', ()) for entry in entries: queuedb_remove(path, entry, cur=cursor) queuedb_query_execute(cursor, 'END', ()) db.commit() db.close() return True
python
def queuedb_removeall(path, entries): db = queuedb_open(path) if db is None: raise Exception("Failed to open %s" % path) cursor = db.cursor() queuedb_query_execute(cursor, 'BEGIN', ()) for entry in entries: queuedb_remove(path, entry, cur=cursor) queuedb_query_execute(cursor, 'END', ()) db.commit() db.close() return True
[ "def", "queuedb_removeall", "(", "path", ",", "entries", ")", ":", "db", "=", "queuedb_open", "(", "path", ")", "if", "db", "is", "None", ":", "raise", "Exception", "(", "\"Failed to open %s\"", "%", "path", ")", "cursor", "=", "db", ".", "cursor", "(", ...
Remove all entries from a queue
[ "Remove", "all", "entries", "from", "a", "queue" ]
1dcfdd39b152d29ce13e736a6a1a0981401a0505
https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/queue.py#L195-L213
244,293
blockstack/blockstack-core
blockstack/lib/operations/register.py
check_payment_in_stacks
def check_payment_in_stacks(state_engine, nameop, state_op_type, fee_block_id): """ Verify that if tokens were paid for a name priced in BTC, that enough were paid. Does not check account balances or namespace types; it only inspects the transaction data. Returns {'status': True, 'tokens_paid': ..., 'token_units': ...} on success Returns {'status': False} on error """ name = nameop['name'] namespace_id = get_namespace_from_name(name) name_without_namespace = get_name_from_fq_name(name) namespace = state_engine.get_namespace( namespace_id ) stacks_payment_info = get_stacks_payment(state_engine, nameop, state_op_type) if stacks_payment_info['status']: # got a stacks payment! check price and make sure we paid the right amount tokens_paid = stacks_payment_info['tokens_paid'] token_units = stacks_payment_info['token_units'] log.debug('Transaction pays {} units of {} for {}, even though its namespace was priced in BTC'.format(tokens_paid, token_units, name)) stacks_price = price_name_stacks(name_without_namespace, namespace, fee_block_id) # price in Stacks, but following the BTC-given price curve res = check_token_payment(name, stacks_price, stacks_payment_info) if res['status']: # success return {'status': True, 'tokens_paid': tokens_paid, 'token_units': token_units} return {'status': False}
python
def check_payment_in_stacks(state_engine, nameop, state_op_type, fee_block_id): name = nameop['name'] namespace_id = get_namespace_from_name(name) name_without_namespace = get_name_from_fq_name(name) namespace = state_engine.get_namespace( namespace_id ) stacks_payment_info = get_stacks_payment(state_engine, nameop, state_op_type) if stacks_payment_info['status']: # got a stacks payment! check price and make sure we paid the right amount tokens_paid = stacks_payment_info['tokens_paid'] token_units = stacks_payment_info['token_units'] log.debug('Transaction pays {} units of {} for {}, even though its namespace was priced in BTC'.format(tokens_paid, token_units, name)) stacks_price = price_name_stacks(name_without_namespace, namespace, fee_block_id) # price in Stacks, but following the BTC-given price curve res = check_token_payment(name, stacks_price, stacks_payment_info) if res['status']: # success return {'status': True, 'tokens_paid': tokens_paid, 'token_units': token_units} return {'status': False}
[ "def", "check_payment_in_stacks", "(", "state_engine", ",", "nameop", ",", "state_op_type", ",", "fee_block_id", ")", ":", "name", "=", "nameop", "[", "'name'", "]", "namespace_id", "=", "get_namespace_from_name", "(", "name", ")", "name_without_namespace", "=", "...
Verify that if tokens were paid for a name priced in BTC, that enough were paid. Does not check account balances or namespace types; it only inspects the transaction data. Returns {'status': True, 'tokens_paid': ..., 'token_units': ...} on success Returns {'status': False} on error
[ "Verify", "that", "if", "tokens", "were", "paid", "for", "a", "name", "priced", "in", "BTC", "that", "enough", "were", "paid", ".", "Does", "not", "check", "account", "balances", "or", "namespace", "types", ";", "it", "only", "inspects", "the", "transactio...
1dcfdd39b152d29ce13e736a6a1a0981401a0505
https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/operations/register.py#L201-L228
244,294
blockstack/blockstack-core
blockstack/lib/operations/register.py
check_payment
def check_payment(state_engine, state_op_type, nameop, fee_block_id, token_address, burn_address, name_fee, block_id): """ Verify that the right payment was made, in the right cryptocurrency units. Does not check any accounts or modify the nameop in any way; it only checks that the name was paid for by the transaction. NOTE: if state_op_type is NAME_REGISTRATION, you will need to have called state_create_put_preorder() before calling this method! Returns {'status': True, 'tokens_paid': tokens_paid, 'token_units': ...} if the payment information is correct. Returns {'status': False} if not """ assert state_op_type in ['NAME_REGISTRATION', 'NAME_RENEWAL'], 'Invalid op type {}'.format(state_op_type) assert name_fee is not None assert isinstance(name_fee, (int,long)) name = nameop['name'] namespace_id = get_namespace_from_name(name) namespace = state_engine.get_namespace( namespace_id ) res = None log.debug('{} is a version-0x{} namespace'.format(namespace['namespace_id'], namespace['version'])) # check name fee, depending on which version. if namespace['version'] == NAMESPACE_VERSION_PAY_TO_BURN: res = check_payment_v1(state_engine, state_op_type, nameop, fee_block_id, token_address, burn_address, name_fee, block_id) elif namespace['version'] == NAMESPACE_VERSION_PAY_TO_CREATOR: res = check_payment_v2(state_engine, state_op_type, nameop, fee_block_id, token_address, burn_address, name_fee, block_id) elif namespace['version'] == NAMESPACE_VERSION_PAY_WITH_STACKS: res = check_payment_v3(state_engine, state_op_type, nameop, fee_block_id, token_address, burn_address, name_fee, block_id) else: # unrecognized namespace rules log.warning("Namespace {} has version bits 0x{:x}, which has unknown registration rules".format(namespace['namespace_id'], namespace['version'])) return {'status': False} if not res['status']: return res tokens_paid = res['tokens_paid'] token_units = res['token_units'] return {'status': True, 'tokens_paid': tokens_paid, 'token_units': token_units}
python
def check_payment(state_engine, state_op_type, nameop, fee_block_id, token_address, burn_address, name_fee, block_id): assert state_op_type in ['NAME_REGISTRATION', 'NAME_RENEWAL'], 'Invalid op type {}'.format(state_op_type) assert name_fee is not None assert isinstance(name_fee, (int,long)) name = nameop['name'] namespace_id = get_namespace_from_name(name) namespace = state_engine.get_namespace( namespace_id ) res = None log.debug('{} is a version-0x{} namespace'.format(namespace['namespace_id'], namespace['version'])) # check name fee, depending on which version. if namespace['version'] == NAMESPACE_VERSION_PAY_TO_BURN: res = check_payment_v1(state_engine, state_op_type, nameop, fee_block_id, token_address, burn_address, name_fee, block_id) elif namespace['version'] == NAMESPACE_VERSION_PAY_TO_CREATOR: res = check_payment_v2(state_engine, state_op_type, nameop, fee_block_id, token_address, burn_address, name_fee, block_id) elif namespace['version'] == NAMESPACE_VERSION_PAY_WITH_STACKS: res = check_payment_v3(state_engine, state_op_type, nameop, fee_block_id, token_address, burn_address, name_fee, block_id) else: # unrecognized namespace rules log.warning("Namespace {} has version bits 0x{:x}, which has unknown registration rules".format(namespace['namespace_id'], namespace['version'])) return {'status': False} if not res['status']: return res tokens_paid = res['tokens_paid'] token_units = res['token_units'] return {'status': True, 'tokens_paid': tokens_paid, 'token_units': token_units}
[ "def", "check_payment", "(", "state_engine", ",", "state_op_type", ",", "nameop", ",", "fee_block_id", ",", "token_address", ",", "burn_address", ",", "name_fee", ",", "block_id", ")", ":", "assert", "state_op_type", "in", "[", "'NAME_REGISTRATION'", ",", "'NAME_R...
Verify that the right payment was made, in the right cryptocurrency units. Does not check any accounts or modify the nameop in any way; it only checks that the name was paid for by the transaction. NOTE: if state_op_type is NAME_REGISTRATION, you will need to have called state_create_put_preorder() before calling this method! Returns {'status': True, 'tokens_paid': tokens_paid, 'token_units': ...} if the payment information is correct. Returns {'status': False} if not
[ "Verify", "that", "the", "right", "payment", "was", "made", "in", "the", "right", "cryptocurrency", "units", ".", "Does", "not", "check", "any", "accounts", "or", "modify", "the", "nameop", "in", "any", "way", ";", "it", "only", "checks", "that", "the", ...
1dcfdd39b152d29ce13e736a6a1a0981401a0505
https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/operations/register.py#L405-L448
244,295
blockstack/blockstack-core
blockstack/lib/operations/namespacepreorder.py
check
def check( state_engine, nameop, block_id, checked_ops ): """ Given a NAMESPACE_PREORDER nameop, see if we can preorder it. It must be unqiue. Return True if accepted. Return False if not. """ namespace_id_hash = nameop['preorder_hash'] consensus_hash = nameop['consensus_hash'] token_fee = nameop['token_fee'] # cannot be preordered already if not state_engine.is_new_namespace_preorder( namespace_id_hash ): log.warning("Namespace preorder '%s' already in use" % namespace_id_hash) return False # has to have a reasonable consensus hash if not state_engine.is_consensus_hash_valid( block_id, consensus_hash ): valid_consensus_hashes = state_engine.get_valid_consensus_hashes( block_id ) log.warning("Invalid consensus hash '%s': expected any of %s" % (consensus_hash, ",".join( valid_consensus_hashes )) ) return False # has to have paid a fee if not 'op_fee' in nameop: log.warning("Missing namespace preorder fee") return False # paid to the right burn address if nameop['burn_address'] != BLOCKSTACK_BURN_ADDRESS: log.warning("Invalid burn address: expected {}, got {}".format(BLOCKSTACK_BURN_ADDRESS, nameop['burn_address'])) return False # token burn fee must be present, if we're in the right epoch for it epoch_features = get_epoch_features(block_id) if EPOCH_FEATURE_STACKS_BUY_NAMESPACES in epoch_features: # must pay in STACKs if 'token_fee' not in nameop: log.warning("Missing token fee") return False token_fee = nameop['token_fee'] token_address = nameop['address'] token_type = TOKEN_TYPE_STACKS # was a token fee paid? if token_fee is None: log.warning("No tokens paid by this NAMESPACE_PREORDER") return False # does this account have enough balance? account_info = state_engine.get_account(token_address, token_type) if account_info is None: log.warning("No account for {} ({})".format(token_address, token_type)) return False account_balance = state_engine.get_account_balance(account_info) assert isinstance(account_balance, (int,long)), 'BUG: account_balance of {} is {} (type {})'.format(token_address, account_balance, type(account_balance)) assert isinstance(token_fee, (int,long)), 'BUG: token_fee is {} (type {})'.format(token_fee, type(token_fee)) if account_balance < token_fee: # can't afford log.warning("Account {} has balance {} {}, but needs to pay {} {}".format(token_address, account_balance, token_type, token_fee, token_type)) return False # debit this account when we commit state_preorder_put_account_payment_info(nameop, token_address, token_type, token_fee) # NOTE: must be a string, to avoid overflow nameop['token_fee'] = '{}'.format(token_fee) nameop['token_units'] = TOKEN_TYPE_STACKS else: # must pay in BTC # not paying in tokens, but say so! state_preorder_put_account_payment_info(nameop, None, None, None) nameop['token_fee'] = '0' nameop['token_units'] = 'BTC' return True
python
def check( state_engine, nameop, block_id, checked_ops ): namespace_id_hash = nameop['preorder_hash'] consensus_hash = nameop['consensus_hash'] token_fee = nameop['token_fee'] # cannot be preordered already if not state_engine.is_new_namespace_preorder( namespace_id_hash ): log.warning("Namespace preorder '%s' already in use" % namespace_id_hash) return False # has to have a reasonable consensus hash if not state_engine.is_consensus_hash_valid( block_id, consensus_hash ): valid_consensus_hashes = state_engine.get_valid_consensus_hashes( block_id ) log.warning("Invalid consensus hash '%s': expected any of %s" % (consensus_hash, ",".join( valid_consensus_hashes )) ) return False # has to have paid a fee if not 'op_fee' in nameop: log.warning("Missing namespace preorder fee") return False # paid to the right burn address if nameop['burn_address'] != BLOCKSTACK_BURN_ADDRESS: log.warning("Invalid burn address: expected {}, got {}".format(BLOCKSTACK_BURN_ADDRESS, nameop['burn_address'])) return False # token burn fee must be present, if we're in the right epoch for it epoch_features = get_epoch_features(block_id) if EPOCH_FEATURE_STACKS_BUY_NAMESPACES in epoch_features: # must pay in STACKs if 'token_fee' not in nameop: log.warning("Missing token fee") return False token_fee = nameop['token_fee'] token_address = nameop['address'] token_type = TOKEN_TYPE_STACKS # was a token fee paid? if token_fee is None: log.warning("No tokens paid by this NAMESPACE_PREORDER") return False # does this account have enough balance? account_info = state_engine.get_account(token_address, token_type) if account_info is None: log.warning("No account for {} ({})".format(token_address, token_type)) return False account_balance = state_engine.get_account_balance(account_info) assert isinstance(account_balance, (int,long)), 'BUG: account_balance of {} is {} (type {})'.format(token_address, account_balance, type(account_balance)) assert isinstance(token_fee, (int,long)), 'BUG: token_fee is {} (type {})'.format(token_fee, type(token_fee)) if account_balance < token_fee: # can't afford log.warning("Account {} has balance {} {}, but needs to pay {} {}".format(token_address, account_balance, token_type, token_fee, token_type)) return False # debit this account when we commit state_preorder_put_account_payment_info(nameop, token_address, token_type, token_fee) # NOTE: must be a string, to avoid overflow nameop['token_fee'] = '{}'.format(token_fee) nameop['token_units'] = TOKEN_TYPE_STACKS else: # must pay in BTC # not paying in tokens, but say so! state_preorder_put_account_payment_info(nameop, None, None, None) nameop['token_fee'] = '0' nameop['token_units'] = 'BTC' return True
[ "def", "check", "(", "state_engine", ",", "nameop", ",", "block_id", ",", "checked_ops", ")", ":", "namespace_id_hash", "=", "nameop", "[", "'preorder_hash'", "]", "consensus_hash", "=", "nameop", "[", "'consensus_hash'", "]", "token_fee", "=", "nameop", "[", ...
Given a NAMESPACE_PREORDER nameop, see if we can preorder it. It must be unqiue. Return True if accepted. Return False if not.
[ "Given", "a", "NAMESPACE_PREORDER", "nameop", "see", "if", "we", "can", "preorder", "it", ".", "It", "must", "be", "unqiue", "." ]
1dcfdd39b152d29ce13e736a6a1a0981401a0505
https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/operations/namespacepreorder.py#L53-L134
244,296
blockstack/blockstack-core
blockstack/lib/fast_sync.py
snapshot_peek_number
def snapshot_peek_number( fd, off ): """ Read the last 8 bytes of fd and interpret it as an int. """ # read number of 8 bytes fd.seek( off - 8, os.SEEK_SET ) value_hex = fd.read(8) if len(value_hex) != 8: return None try: value = int(value_hex, 16) except ValueError: return None return value
python
def snapshot_peek_number( fd, off ): # read number of 8 bytes fd.seek( off - 8, os.SEEK_SET ) value_hex = fd.read(8) if len(value_hex) != 8: return None try: value = int(value_hex, 16) except ValueError: return None return value
[ "def", "snapshot_peek_number", "(", "fd", ",", "off", ")", ":", "# read number of 8 bytes ", "fd", ".", "seek", "(", "off", "-", "8", ",", "os", ".", "SEEK_SET", ")", "value_hex", "=", "fd", ".", "read", "(", "8", ")", "if", "len", "(", "value_hex", ...
Read the last 8 bytes of fd and interpret it as an int.
[ "Read", "the", "last", "8", "bytes", "of", "fd", "and", "interpret", "it", "as", "an", "int", "." ]
1dcfdd39b152d29ce13e736a6a1a0981401a0505
https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/fast_sync.py#L52-L67
244,297
blockstack/blockstack-core
blockstack/lib/fast_sync.py
get_file_hash
def get_file_hash( fd, hashfunc, fd_len=None ): """ Get the hex-encoded hash of the fd's data """ h = hashfunc() fd.seek(0, os.SEEK_SET) count = 0 while True: buf = fd.read(65536) if len(buf) == 0: break if fd_len is not None: if count + len(buf) > fd_len: buf = buf[:fd_len - count] h.update(buf) count += len(buf) hashed = h.hexdigest() return hashed
python
def get_file_hash( fd, hashfunc, fd_len=None ): h = hashfunc() fd.seek(0, os.SEEK_SET) count = 0 while True: buf = fd.read(65536) if len(buf) == 0: break if fd_len is not None: if count + len(buf) > fd_len: buf = buf[:fd_len - count] h.update(buf) count += len(buf) hashed = h.hexdigest() return hashed
[ "def", "get_file_hash", "(", "fd", ",", "hashfunc", ",", "fd_len", "=", "None", ")", ":", "h", "=", "hashfunc", "(", ")", "fd", ".", "seek", "(", "0", ",", "os", ".", "SEEK_SET", ")", "count", "=", "0", "while", "True", ":", "buf", "=", "fd", "...
Get the hex-encoded hash of the fd's data
[ "Get", "the", "hex", "-", "encoded", "hash", "of", "the", "fd", "s", "data" ]
1dcfdd39b152d29ce13e736a6a1a0981401a0505
https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/fast_sync.py#L89-L111
244,298
blockstack/blockstack-core
blockstack/lib/fast_sync.py
fast_sync_sign_snapshot
def fast_sync_sign_snapshot( snapshot_path, private_key, first=False ): """ Append a signature to the end of a snapshot path with the given private key. If first is True, then don't expect the signature trailer. Return True on success Return False on error """ if not os.path.exists(snapshot_path): log.error("No such file or directory: {}".format(snapshot_path)) return False file_size = 0 payload_size = 0 write_offset = 0 try: sb = os.stat(snapshot_path) file_size = sb.st_size assert file_size > 8 except Exception as e: log.exception(e) return False num_sigs = 0 snapshot_hash = None with open(snapshot_path, 'r+') as f: if not first: info = fast_sync_inspect(f) if 'error' in info: log.error("Failed to inspect {}: {}".format(snapshot_path, info['error'])) return False num_sigs = len(info['signatures']) write_offset = info['sig_append_offset'] payload_size = info['payload_size'] else: # no one has signed yet. write_offset = file_size num_sigs = 0 payload_size = file_size # hash the file and sign the (bin-encoded) hash privkey_hex = keylib.ECPrivateKey(private_key).to_hex() hash_hex = get_file_hash( f, hashlib.sha256, fd_len=payload_size ) sigb64 = sign_digest( hash_hex, privkey_hex, hashfunc=hashlib.sha256 ) if BLOCKSTACK_TEST: log.debug("Signed {} with {} to make {}".format(hash_hex, keylib.ECPrivateKey(private_key).public_key().to_hex(), sigb64)) # append f.seek(write_offset, os.SEEK_SET) f.write(sigb64) f.write('{:08x}'.format(len(sigb64))) # append number of signatures num_sigs += 1 f.write('{:08x}'.format(num_sigs)) f.flush() os.fsync(f.fileno()) return True
python
def fast_sync_sign_snapshot( snapshot_path, private_key, first=False ): if not os.path.exists(snapshot_path): log.error("No such file or directory: {}".format(snapshot_path)) return False file_size = 0 payload_size = 0 write_offset = 0 try: sb = os.stat(snapshot_path) file_size = sb.st_size assert file_size > 8 except Exception as e: log.exception(e) return False num_sigs = 0 snapshot_hash = None with open(snapshot_path, 'r+') as f: if not first: info = fast_sync_inspect(f) if 'error' in info: log.error("Failed to inspect {}: {}".format(snapshot_path, info['error'])) return False num_sigs = len(info['signatures']) write_offset = info['sig_append_offset'] payload_size = info['payload_size'] else: # no one has signed yet. write_offset = file_size num_sigs = 0 payload_size = file_size # hash the file and sign the (bin-encoded) hash privkey_hex = keylib.ECPrivateKey(private_key).to_hex() hash_hex = get_file_hash( f, hashlib.sha256, fd_len=payload_size ) sigb64 = sign_digest( hash_hex, privkey_hex, hashfunc=hashlib.sha256 ) if BLOCKSTACK_TEST: log.debug("Signed {} with {} to make {}".format(hash_hex, keylib.ECPrivateKey(private_key).public_key().to_hex(), sigb64)) # append f.seek(write_offset, os.SEEK_SET) f.write(sigb64) f.write('{:08x}'.format(len(sigb64))) # append number of signatures num_sigs += 1 f.write('{:08x}'.format(num_sigs)) f.flush() os.fsync(f.fileno()) return True
[ "def", "fast_sync_sign_snapshot", "(", "snapshot_path", ",", "private_key", ",", "first", "=", "False", ")", ":", "if", "not", "os", ".", "path", ".", "exists", "(", "snapshot_path", ")", ":", "log", ".", "error", "(", "\"No such file or directory: {}\"", ".",...
Append a signature to the end of a snapshot path with the given private key. If first is True, then don't expect the signature trailer. Return True on success Return False on error
[ "Append", "a", "signature", "to", "the", "end", "of", "a", "snapshot", "path", "with", "the", "given", "private", "key", "." ]
1dcfdd39b152d29ce13e736a6a1a0981401a0505
https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/fast_sync.py#L114-L180
244,299
blockstack/blockstack-core
blockstack/lib/fast_sync.py
fast_sync_snapshot_compress
def fast_sync_snapshot_compress( snapshot_dir, export_path ): """ Given the path to a directory, compress it and export it to the given path. Return {'status': True} on success Return {'error': ...} on failure """ snapshot_dir = os.path.abspath(snapshot_dir) export_path = os.path.abspath(export_path) if os.path.exists(export_path): return {'error': 'Snapshot path exists: {}'.format(export_path)} old_dir = os.getcwd() count_ref = [0] def print_progress(tarinfo): count_ref[0] += 1 if count_ref[0] % 100 == 0: log.debug("{} files compressed...".format(count_ref[0])) return tarinfo try: os.chdir(snapshot_dir) with tarfile.TarFile.bz2open(export_path, "w") as f: f.add(".", filter=print_progress) except: os.chdir(old_dir) raise finally: os.chdir(old_dir) return {'status': True}
python
def fast_sync_snapshot_compress( snapshot_dir, export_path ): snapshot_dir = os.path.abspath(snapshot_dir) export_path = os.path.abspath(export_path) if os.path.exists(export_path): return {'error': 'Snapshot path exists: {}'.format(export_path)} old_dir = os.getcwd() count_ref = [0] def print_progress(tarinfo): count_ref[0] += 1 if count_ref[0] % 100 == 0: log.debug("{} files compressed...".format(count_ref[0])) return tarinfo try: os.chdir(snapshot_dir) with tarfile.TarFile.bz2open(export_path, "w") as f: f.add(".", filter=print_progress) except: os.chdir(old_dir) raise finally: os.chdir(old_dir) return {'status': True}
[ "def", "fast_sync_snapshot_compress", "(", "snapshot_dir", ",", "export_path", ")", ":", "snapshot_dir", "=", "os", ".", "path", ".", "abspath", "(", "snapshot_dir", ")", "export_path", "=", "os", ".", "path", ".", "abspath", "(", "export_path", ")", "if", "...
Given the path to a directory, compress it and export it to the given path. Return {'status': True} on success Return {'error': ...} on failure
[ "Given", "the", "path", "to", "a", "directory", "compress", "it", "and", "export", "it", "to", "the", "given", "path", "." ]
1dcfdd39b152d29ce13e736a6a1a0981401a0505
https://github.com/blockstack/blockstack-core/blob/1dcfdd39b152d29ce13e736a6a1a0981401a0505/blockstack/lib/fast_sync.py#L183-L220