repo
stringlengths 7
55
| path
stringlengths 4
223
| func_name
stringlengths 1
134
| original_string
stringlengths 75
104k
| language
stringclasses 1
value | code
stringlengths 75
104k
| code_tokens
listlengths 19
28.4k
| docstring
stringlengths 1
46.9k
| docstring_tokens
listlengths 1
1.97k
| sha
stringlengths 40
40
| url
stringlengths 87
315
| partition
stringclasses 1
value |
|---|---|---|---|---|---|---|---|---|---|---|---|
philgyford/django-spectator
|
spectator/events/templatetags/spectator_events.py
|
most_seen_creators_by_works_card
|
def most_seen_creators_by_works_card(work_kind=None, role_name=None, num=10):
"""
Displays a card showing the Creators that are associated with the most Works.
e.g.:
{% most_seen_creators_by_works_card work_kind='movie' role_name='Director' num=5 %}
"""
object_list = most_seen_creators_by_works(
work_kind=work_kind, role_name=role_name, num=num)
object_list = chartify(object_list, 'num_works', cutoff=1)
# Attempt to create a sensible card title...
if role_name:
# Yes, this pluralization is going to break at some point:
creators_name = '{}s'.format(role_name.capitalize())
else:
creators_name = 'People/groups'
if work_kind:
works_name = Work.get_kind_name_plural(work_kind).lower()
else:
works_name = 'works'
card_title = '{} with most {}'.format(creators_name, works_name)
return {
'card_title': card_title,
'score_attr': 'num_works',
'object_list': object_list,
}
|
python
|
def most_seen_creators_by_works_card(work_kind=None, role_name=None, num=10):
"""
Displays a card showing the Creators that are associated with the most Works.
e.g.:
{% most_seen_creators_by_works_card work_kind='movie' role_name='Director' num=5 %}
"""
object_list = most_seen_creators_by_works(
work_kind=work_kind, role_name=role_name, num=num)
object_list = chartify(object_list, 'num_works', cutoff=1)
# Attempt to create a sensible card title...
if role_name:
# Yes, this pluralization is going to break at some point:
creators_name = '{}s'.format(role_name.capitalize())
else:
creators_name = 'People/groups'
if work_kind:
works_name = Work.get_kind_name_plural(work_kind).lower()
else:
works_name = 'works'
card_title = '{} with most {}'.format(creators_name, works_name)
return {
'card_title': card_title,
'score_attr': 'num_works',
'object_list': object_list,
}
|
[
"def",
"most_seen_creators_by_works_card",
"(",
"work_kind",
"=",
"None",
",",
"role_name",
"=",
"None",
",",
"num",
"=",
"10",
")",
":",
"object_list",
"=",
"most_seen_creators_by_works",
"(",
"work_kind",
"=",
"work_kind",
",",
"role_name",
"=",
"role_name",
",",
"num",
"=",
"num",
")",
"object_list",
"=",
"chartify",
"(",
"object_list",
",",
"'num_works'",
",",
"cutoff",
"=",
"1",
")",
"# Attempt to create a sensible card title...",
"if",
"role_name",
":",
"# Yes, this pluralization is going to break at some point:",
"creators_name",
"=",
"'{}s'",
".",
"format",
"(",
"role_name",
".",
"capitalize",
"(",
")",
")",
"else",
":",
"creators_name",
"=",
"'People/groups'",
"if",
"work_kind",
":",
"works_name",
"=",
"Work",
".",
"get_kind_name_plural",
"(",
"work_kind",
")",
".",
"lower",
"(",
")",
"else",
":",
"works_name",
"=",
"'works'",
"card_title",
"=",
"'{} with most {}'",
".",
"format",
"(",
"creators_name",
",",
"works_name",
")",
"return",
"{",
"'card_title'",
":",
"card_title",
",",
"'score_attr'",
":",
"'num_works'",
",",
"'object_list'",
":",
"object_list",
",",
"}"
] |
Displays a card showing the Creators that are associated with the most Works.
e.g.:
{% most_seen_creators_by_works_card work_kind='movie' role_name='Director' num=5 %}
|
[
"Displays",
"a",
"card",
"showing",
"the",
"Creators",
"that",
"are",
"associated",
"with",
"the",
"most",
"Works",
"."
] |
f3c72004f9caa1fde0f5a3b2f0d2bf285fc01ada
|
https://github.com/philgyford/django-spectator/blob/f3c72004f9caa1fde0f5a3b2f0d2bf285fc01ada/spectator/events/templatetags/spectator_events.py#L203-L235
|
train
|
philgyford/django-spectator
|
spectator/events/templatetags/spectator_events.py
|
most_seen_works_card
|
def most_seen_works_card(kind=None, num=10):
"""
Displays a card showing the Works that are associated with the most Events.
"""
object_list = most_seen_works(kind=kind, num=num)
object_list = chartify(object_list, 'num_views', cutoff=1)
if kind:
card_title = 'Most seen {}'.format(
Work.get_kind_name_plural(kind).lower())
else:
card_title = 'Most seen works'
return {
'card_title': card_title,
'score_attr': 'num_views',
'object_list': object_list,
'name_attr': 'title',
'use_cite': True,
}
|
python
|
def most_seen_works_card(kind=None, num=10):
"""
Displays a card showing the Works that are associated with the most Events.
"""
object_list = most_seen_works(kind=kind, num=num)
object_list = chartify(object_list, 'num_views', cutoff=1)
if kind:
card_title = 'Most seen {}'.format(
Work.get_kind_name_plural(kind).lower())
else:
card_title = 'Most seen works'
return {
'card_title': card_title,
'score_attr': 'num_views',
'object_list': object_list,
'name_attr': 'title',
'use_cite': True,
}
|
[
"def",
"most_seen_works_card",
"(",
"kind",
"=",
"None",
",",
"num",
"=",
"10",
")",
":",
"object_list",
"=",
"most_seen_works",
"(",
"kind",
"=",
"kind",
",",
"num",
"=",
"num",
")",
"object_list",
"=",
"chartify",
"(",
"object_list",
",",
"'num_views'",
",",
"cutoff",
"=",
"1",
")",
"if",
"kind",
":",
"card_title",
"=",
"'Most seen {}'",
".",
"format",
"(",
"Work",
".",
"get_kind_name_plural",
"(",
"kind",
")",
".",
"lower",
"(",
")",
")",
"else",
":",
"card_title",
"=",
"'Most seen works'",
"return",
"{",
"'card_title'",
":",
"card_title",
",",
"'score_attr'",
":",
"'num_views'",
",",
"'object_list'",
":",
"object_list",
",",
"'name_attr'",
":",
"'title'",
",",
"'use_cite'",
":",
"True",
",",
"}"
] |
Displays a card showing the Works that are associated with the most Events.
|
[
"Displays",
"a",
"card",
"showing",
"the",
"Works",
"that",
"are",
"associated",
"with",
"the",
"most",
"Events",
"."
] |
f3c72004f9caa1fde0f5a3b2f0d2bf285fc01ada
|
https://github.com/philgyford/django-spectator/blob/f3c72004f9caa1fde0f5a3b2f0d2bf285fc01ada/spectator/events/templatetags/spectator_events.py#L247-L267
|
train
|
philgyford/django-spectator
|
spectator/core/models.py
|
SluggedModelMixin._generate_slug
|
def _generate_slug(self, value):
"""
Generates a slug using a Hashid of `value`.
"""
alphabet = app_settings.SLUG_ALPHABET
salt = app_settings.SLUG_SALT
hashids = Hashids(alphabet=alphabet, salt=salt, min_length=5)
return hashids.encode(value)
|
python
|
def _generate_slug(self, value):
"""
Generates a slug using a Hashid of `value`.
"""
alphabet = app_settings.SLUG_ALPHABET
salt = app_settings.SLUG_SALT
hashids = Hashids(alphabet=alphabet, salt=salt, min_length=5)
return hashids.encode(value)
|
[
"def",
"_generate_slug",
"(",
"self",
",",
"value",
")",
":",
"alphabet",
"=",
"app_settings",
".",
"SLUG_ALPHABET",
"salt",
"=",
"app_settings",
".",
"SLUG_SALT",
"hashids",
"=",
"Hashids",
"(",
"alphabet",
"=",
"alphabet",
",",
"salt",
"=",
"salt",
",",
"min_length",
"=",
"5",
")",
"return",
"hashids",
".",
"encode",
"(",
"value",
")"
] |
Generates a slug using a Hashid of `value`.
|
[
"Generates",
"a",
"slug",
"using",
"a",
"Hashid",
"of",
"value",
"."
] |
f3c72004f9caa1fde0f5a3b2f0d2bf285fc01ada
|
https://github.com/philgyford/django-spectator/blob/f3c72004f9caa1fde0f5a3b2f0d2bf285fc01ada/spectator/core/models.py#L46-L55
|
train
|
frictionlessdata/tableschema-pandas-py
|
tableschema_pandas/storage.py
|
Storage.create
|
def create(self, bucket, descriptor, force=False):
"""https://github.com/frictionlessdata/tableschema-pandas-py#storage
"""
# Make lists
buckets = bucket
if isinstance(bucket, six.string_types):
buckets = [bucket]
descriptors = descriptor
if isinstance(descriptor, dict):
descriptors = [descriptor]
# Check buckets for existence
for bucket in buckets:
if bucket in self.buckets:
if not force:
message = 'Bucket "%s" already exists' % bucket
raise tableschema.exceptions.StorageError(message)
self.delete(bucket)
# Define dataframes
for bucket, descriptor in zip(buckets, descriptors):
tableschema.validate(descriptor)
self.__descriptors[bucket] = descriptor
self.__dataframes[bucket] = pd.DataFrame()
|
python
|
def create(self, bucket, descriptor, force=False):
"""https://github.com/frictionlessdata/tableschema-pandas-py#storage
"""
# Make lists
buckets = bucket
if isinstance(bucket, six.string_types):
buckets = [bucket]
descriptors = descriptor
if isinstance(descriptor, dict):
descriptors = [descriptor]
# Check buckets for existence
for bucket in buckets:
if bucket in self.buckets:
if not force:
message = 'Bucket "%s" already exists' % bucket
raise tableschema.exceptions.StorageError(message)
self.delete(bucket)
# Define dataframes
for bucket, descriptor in zip(buckets, descriptors):
tableschema.validate(descriptor)
self.__descriptors[bucket] = descriptor
self.__dataframes[bucket] = pd.DataFrame()
|
[
"def",
"create",
"(",
"self",
",",
"bucket",
",",
"descriptor",
",",
"force",
"=",
"False",
")",
":",
"# Make lists",
"buckets",
"=",
"bucket",
"if",
"isinstance",
"(",
"bucket",
",",
"six",
".",
"string_types",
")",
":",
"buckets",
"=",
"[",
"bucket",
"]",
"descriptors",
"=",
"descriptor",
"if",
"isinstance",
"(",
"descriptor",
",",
"dict",
")",
":",
"descriptors",
"=",
"[",
"descriptor",
"]",
"# Check buckets for existence",
"for",
"bucket",
"in",
"buckets",
":",
"if",
"bucket",
"in",
"self",
".",
"buckets",
":",
"if",
"not",
"force",
":",
"message",
"=",
"'Bucket \"%s\" already exists'",
"%",
"bucket",
"raise",
"tableschema",
".",
"exceptions",
".",
"StorageError",
"(",
"message",
")",
"self",
".",
"delete",
"(",
"bucket",
")",
"# Define dataframes",
"for",
"bucket",
",",
"descriptor",
"in",
"zip",
"(",
"buckets",
",",
"descriptors",
")",
":",
"tableschema",
".",
"validate",
"(",
"descriptor",
")",
"self",
".",
"__descriptors",
"[",
"bucket",
"]",
"=",
"descriptor",
"self",
".",
"__dataframes",
"[",
"bucket",
"]",
"=",
"pd",
".",
"DataFrame",
"(",
")"
] |
https://github.com/frictionlessdata/tableschema-pandas-py#storage
|
[
"https",
":",
"//",
"github",
".",
"com",
"/",
"frictionlessdata",
"/",
"tableschema",
"-",
"pandas",
"-",
"py#storage"
] |
ef941dbc12f5d346e9612f8fec1b4b356b8493ca
|
https://github.com/frictionlessdata/tableschema-pandas-py/blob/ef941dbc12f5d346e9612f8fec1b4b356b8493ca/tableschema_pandas/storage.py#L47-L71
|
train
|
frictionlessdata/tableschema-pandas-py
|
tableschema_pandas/storage.py
|
Storage.delete
|
def delete(self, bucket=None, ignore=False):
"""https://github.com/frictionlessdata/tableschema-pandas-py#storage
"""
# Make lists
buckets = bucket
if isinstance(bucket, six.string_types):
buckets = [bucket]
elif bucket is None:
buckets = reversed(self.buckets)
# Iterate over buckets
for bucket in buckets:
# Non existent bucket
if bucket not in self.buckets:
if not ignore:
message = 'Bucket "%s" doesn\'t exist' % bucket
raise tableschema.exceptions.StorageError(message)
return
# Remove from descriptors
if bucket in self.__descriptors:
del self.__descriptors[bucket]
# Remove from dataframes
if bucket in self.__dataframes:
del self.__dataframes[bucket]
|
python
|
def delete(self, bucket=None, ignore=False):
"""https://github.com/frictionlessdata/tableschema-pandas-py#storage
"""
# Make lists
buckets = bucket
if isinstance(bucket, six.string_types):
buckets = [bucket]
elif bucket is None:
buckets = reversed(self.buckets)
# Iterate over buckets
for bucket in buckets:
# Non existent bucket
if bucket not in self.buckets:
if not ignore:
message = 'Bucket "%s" doesn\'t exist' % bucket
raise tableschema.exceptions.StorageError(message)
return
# Remove from descriptors
if bucket in self.__descriptors:
del self.__descriptors[bucket]
# Remove from dataframes
if bucket in self.__dataframes:
del self.__dataframes[bucket]
|
[
"def",
"delete",
"(",
"self",
",",
"bucket",
"=",
"None",
",",
"ignore",
"=",
"False",
")",
":",
"# Make lists",
"buckets",
"=",
"bucket",
"if",
"isinstance",
"(",
"bucket",
",",
"six",
".",
"string_types",
")",
":",
"buckets",
"=",
"[",
"bucket",
"]",
"elif",
"bucket",
"is",
"None",
":",
"buckets",
"=",
"reversed",
"(",
"self",
".",
"buckets",
")",
"# Iterate over buckets",
"for",
"bucket",
"in",
"buckets",
":",
"# Non existent bucket",
"if",
"bucket",
"not",
"in",
"self",
".",
"buckets",
":",
"if",
"not",
"ignore",
":",
"message",
"=",
"'Bucket \"%s\" doesn\\'t exist'",
"%",
"bucket",
"raise",
"tableschema",
".",
"exceptions",
".",
"StorageError",
"(",
"message",
")",
"return",
"# Remove from descriptors",
"if",
"bucket",
"in",
"self",
".",
"__descriptors",
":",
"del",
"self",
".",
"__descriptors",
"[",
"bucket",
"]",
"# Remove from dataframes",
"if",
"bucket",
"in",
"self",
".",
"__dataframes",
":",
"del",
"self",
".",
"__dataframes",
"[",
"bucket",
"]"
] |
https://github.com/frictionlessdata/tableschema-pandas-py#storage
|
[
"https",
":",
"//",
"github",
".",
"com",
"/",
"frictionlessdata",
"/",
"tableschema",
"-",
"pandas",
"-",
"py#storage"
] |
ef941dbc12f5d346e9612f8fec1b4b356b8493ca
|
https://github.com/frictionlessdata/tableschema-pandas-py/blob/ef941dbc12f5d346e9612f8fec1b4b356b8493ca/tableschema_pandas/storage.py#L73-L100
|
train
|
frictionlessdata/tableschema-pandas-py
|
tableschema_pandas/storage.py
|
Storage.describe
|
def describe(self, bucket, descriptor=None):
"""https://github.com/frictionlessdata/tableschema-pandas-py#storage
"""
# Set descriptor
if descriptor is not None:
self.__descriptors[bucket] = descriptor
# Get descriptor
else:
descriptor = self.__descriptors.get(bucket)
if descriptor is None:
dataframe = self.__dataframes[bucket]
descriptor = self.__mapper.restore_descriptor(dataframe)
return descriptor
|
python
|
def describe(self, bucket, descriptor=None):
"""https://github.com/frictionlessdata/tableschema-pandas-py#storage
"""
# Set descriptor
if descriptor is not None:
self.__descriptors[bucket] = descriptor
# Get descriptor
else:
descriptor = self.__descriptors.get(bucket)
if descriptor is None:
dataframe = self.__dataframes[bucket]
descriptor = self.__mapper.restore_descriptor(dataframe)
return descriptor
|
[
"def",
"describe",
"(",
"self",
",",
"bucket",
",",
"descriptor",
"=",
"None",
")",
":",
"# Set descriptor",
"if",
"descriptor",
"is",
"not",
"None",
":",
"self",
".",
"__descriptors",
"[",
"bucket",
"]",
"=",
"descriptor",
"# Get descriptor",
"else",
":",
"descriptor",
"=",
"self",
".",
"__descriptors",
".",
"get",
"(",
"bucket",
")",
"if",
"descriptor",
"is",
"None",
":",
"dataframe",
"=",
"self",
".",
"__dataframes",
"[",
"bucket",
"]",
"descriptor",
"=",
"self",
".",
"__mapper",
".",
"restore_descriptor",
"(",
"dataframe",
")",
"return",
"descriptor"
] |
https://github.com/frictionlessdata/tableschema-pandas-py#storage
|
[
"https",
":",
"//",
"github",
".",
"com",
"/",
"frictionlessdata",
"/",
"tableschema",
"-",
"pandas",
"-",
"py#storage"
] |
ef941dbc12f5d346e9612f8fec1b4b356b8493ca
|
https://github.com/frictionlessdata/tableschema-pandas-py/blob/ef941dbc12f5d346e9612f8fec1b4b356b8493ca/tableschema_pandas/storage.py#L102-L117
|
train
|
frictionlessdata/tableschema-pandas-py
|
tableschema_pandas/storage.py
|
Storage.iter
|
def iter(self, bucket):
"""https://github.com/frictionlessdata/tableschema-pandas-py#storage
"""
# Check existense
if bucket not in self.buckets:
message = 'Bucket "%s" doesn\'t exist.' % bucket
raise tableschema.exceptions.StorageError(message)
# Prepare
descriptor = self.describe(bucket)
schema = tableschema.Schema(descriptor)
# Yield rows
for pk, row in self.__dataframes[bucket].iterrows():
row = self.__mapper.restore_row(row, schema=schema, pk=pk)
yield row
|
python
|
def iter(self, bucket):
"""https://github.com/frictionlessdata/tableschema-pandas-py#storage
"""
# Check existense
if bucket not in self.buckets:
message = 'Bucket "%s" doesn\'t exist.' % bucket
raise tableschema.exceptions.StorageError(message)
# Prepare
descriptor = self.describe(bucket)
schema = tableschema.Schema(descriptor)
# Yield rows
for pk, row in self.__dataframes[bucket].iterrows():
row = self.__mapper.restore_row(row, schema=schema, pk=pk)
yield row
|
[
"def",
"iter",
"(",
"self",
",",
"bucket",
")",
":",
"# Check existense",
"if",
"bucket",
"not",
"in",
"self",
".",
"buckets",
":",
"message",
"=",
"'Bucket \"%s\" doesn\\'t exist.'",
"%",
"bucket",
"raise",
"tableschema",
".",
"exceptions",
".",
"StorageError",
"(",
"message",
")",
"# Prepare",
"descriptor",
"=",
"self",
".",
"describe",
"(",
"bucket",
")",
"schema",
"=",
"tableschema",
".",
"Schema",
"(",
"descriptor",
")",
"# Yield rows",
"for",
"pk",
",",
"row",
"in",
"self",
".",
"__dataframes",
"[",
"bucket",
"]",
".",
"iterrows",
"(",
")",
":",
"row",
"=",
"self",
".",
"__mapper",
".",
"restore_row",
"(",
"row",
",",
"schema",
"=",
"schema",
",",
"pk",
"=",
"pk",
")",
"yield",
"row"
] |
https://github.com/frictionlessdata/tableschema-pandas-py#storage
|
[
"https",
":",
"//",
"github",
".",
"com",
"/",
"frictionlessdata",
"/",
"tableschema",
"-",
"pandas",
"-",
"py#storage"
] |
ef941dbc12f5d346e9612f8fec1b4b356b8493ca
|
https://github.com/frictionlessdata/tableschema-pandas-py/blob/ef941dbc12f5d346e9612f8fec1b4b356b8493ca/tableschema_pandas/storage.py#L119-L135
|
train
|
frictionlessdata/tableschema-pandas-py
|
tableschema_pandas/storage.py
|
Storage.write
|
def write(self, bucket, rows):
"""https://github.com/frictionlessdata/tableschema-pandas-py#storage
"""
# Prepare
descriptor = self.describe(bucket)
new_data_frame = self.__mapper.convert_descriptor_and_rows(descriptor, rows)
# Just set new DataFrame if current is empty
if self.__dataframes[bucket].size == 0:
self.__dataframes[bucket] = new_data_frame
# Append new data frame to the old one setting new data frame
# containing data from both old and new data frames
else:
self.__dataframes[bucket] = pd.concat([
self.__dataframes[bucket],
new_data_frame,
])
|
python
|
def write(self, bucket, rows):
"""https://github.com/frictionlessdata/tableschema-pandas-py#storage
"""
# Prepare
descriptor = self.describe(bucket)
new_data_frame = self.__mapper.convert_descriptor_and_rows(descriptor, rows)
# Just set new DataFrame if current is empty
if self.__dataframes[bucket].size == 0:
self.__dataframes[bucket] = new_data_frame
# Append new data frame to the old one setting new data frame
# containing data from both old and new data frames
else:
self.__dataframes[bucket] = pd.concat([
self.__dataframes[bucket],
new_data_frame,
])
|
[
"def",
"write",
"(",
"self",
",",
"bucket",
",",
"rows",
")",
":",
"# Prepare",
"descriptor",
"=",
"self",
".",
"describe",
"(",
"bucket",
")",
"new_data_frame",
"=",
"self",
".",
"__mapper",
".",
"convert_descriptor_and_rows",
"(",
"descriptor",
",",
"rows",
")",
"# Just set new DataFrame if current is empty",
"if",
"self",
".",
"__dataframes",
"[",
"bucket",
"]",
".",
"size",
"==",
"0",
":",
"self",
".",
"__dataframes",
"[",
"bucket",
"]",
"=",
"new_data_frame",
"# Append new data frame to the old one setting new data frame",
"# containing data from both old and new data frames",
"else",
":",
"self",
".",
"__dataframes",
"[",
"bucket",
"]",
"=",
"pd",
".",
"concat",
"(",
"[",
"self",
".",
"__dataframes",
"[",
"bucket",
"]",
",",
"new_data_frame",
",",
"]",
")"
] |
https://github.com/frictionlessdata/tableschema-pandas-py#storage
|
[
"https",
":",
"//",
"github",
".",
"com",
"/",
"frictionlessdata",
"/",
"tableschema",
"-",
"pandas",
"-",
"py#storage"
] |
ef941dbc12f5d346e9612f8fec1b4b356b8493ca
|
https://github.com/frictionlessdata/tableschema-pandas-py/blob/ef941dbc12f5d346e9612f8fec1b4b356b8493ca/tableschema_pandas/storage.py#L143-L161
|
train
|
philgyford/django-spectator
|
spectator/events/migrations/0030_movies_to_works.py
|
forwards
|
def forwards(apps, schema_editor):
"""
Change all Movie objects into Work objects, and their associated
data into WorkRole and WorkSelection models, then delete the Movie.
"""
Movie = apps.get_model('spectator_events', 'Movie')
Work = apps.get_model('spectator_events', 'Work')
WorkRole = apps.get_model('spectator_events', 'WorkRole')
WorkSelection = apps.get_model('spectator_events', 'WorkSelection')
for m in Movie.objects.all():
work = Work.objects.create(
kind='movie',
title=m.title,
title_sort=m.title_sort,
year=m.year,
imdb_id=m.imdb_id
)
for role in m.roles.all():
WorkRole.objects.create(
creator=role.creator,
work=work,
role_name=role.role_name,
role_order=role.role_order
)
for selection in m.events.all():
WorkSelection.objects.create(
event=selection.event,
work=work,
order=selection.order
)
m.delete()
|
python
|
def forwards(apps, schema_editor):
"""
Change all Movie objects into Work objects, and their associated
data into WorkRole and WorkSelection models, then delete the Movie.
"""
Movie = apps.get_model('spectator_events', 'Movie')
Work = apps.get_model('spectator_events', 'Work')
WorkRole = apps.get_model('spectator_events', 'WorkRole')
WorkSelection = apps.get_model('spectator_events', 'WorkSelection')
for m in Movie.objects.all():
work = Work.objects.create(
kind='movie',
title=m.title,
title_sort=m.title_sort,
year=m.year,
imdb_id=m.imdb_id
)
for role in m.roles.all():
WorkRole.objects.create(
creator=role.creator,
work=work,
role_name=role.role_name,
role_order=role.role_order
)
for selection in m.events.all():
WorkSelection.objects.create(
event=selection.event,
work=work,
order=selection.order
)
m.delete()
|
[
"def",
"forwards",
"(",
"apps",
",",
"schema_editor",
")",
":",
"Movie",
"=",
"apps",
".",
"get_model",
"(",
"'spectator_events'",
",",
"'Movie'",
")",
"Work",
"=",
"apps",
".",
"get_model",
"(",
"'spectator_events'",
",",
"'Work'",
")",
"WorkRole",
"=",
"apps",
".",
"get_model",
"(",
"'spectator_events'",
",",
"'WorkRole'",
")",
"WorkSelection",
"=",
"apps",
".",
"get_model",
"(",
"'spectator_events'",
",",
"'WorkSelection'",
")",
"for",
"m",
"in",
"Movie",
".",
"objects",
".",
"all",
"(",
")",
":",
"work",
"=",
"Work",
".",
"objects",
".",
"create",
"(",
"kind",
"=",
"'movie'",
",",
"title",
"=",
"m",
".",
"title",
",",
"title_sort",
"=",
"m",
".",
"title_sort",
",",
"year",
"=",
"m",
".",
"year",
",",
"imdb_id",
"=",
"m",
".",
"imdb_id",
")",
"for",
"role",
"in",
"m",
".",
"roles",
".",
"all",
"(",
")",
":",
"WorkRole",
".",
"objects",
".",
"create",
"(",
"creator",
"=",
"role",
".",
"creator",
",",
"work",
"=",
"work",
",",
"role_name",
"=",
"role",
".",
"role_name",
",",
"role_order",
"=",
"role",
".",
"role_order",
")",
"for",
"selection",
"in",
"m",
".",
"events",
".",
"all",
"(",
")",
":",
"WorkSelection",
".",
"objects",
".",
"create",
"(",
"event",
"=",
"selection",
".",
"event",
",",
"work",
"=",
"work",
",",
"order",
"=",
"selection",
".",
"order",
")",
"m",
".",
"delete",
"(",
")"
] |
Change all Movie objects into Work objects, and their associated
data into WorkRole and WorkSelection models, then delete the Movie.
|
[
"Change",
"all",
"Movie",
"objects",
"into",
"Work",
"objects",
"and",
"their",
"associated",
"data",
"into",
"WorkRole",
"and",
"WorkSelection",
"models",
"then",
"delete",
"the",
"Movie",
"."
] |
f3c72004f9caa1fde0f5a3b2f0d2bf285fc01ada
|
https://github.com/philgyford/django-spectator/blob/f3c72004f9caa1fde0f5a3b2f0d2bf285fc01ada/spectator/events/migrations/0030_movies_to_works.py#L6-L41
|
train
|
philgyford/django-spectator
|
spectator/core/views.py
|
PaginatedListView.paginate_queryset
|
def paginate_queryset(self, queryset, page_size):
"""
Paginate the queryset, if needed.
This is EXACTLY the same as the standard ListView.paginate_queryset()
except for this line:
page = paginator.page(page_number, softlimit=True)
Because we want to use the DiggPaginator's softlimit option.
So that if you're viewing a page of, say, Flickr photos, and you switch
from viewing by Uploaded Time to viewing by Taken Time, the new
ordering might have fewer pages. In that case we want to see the final
page, not a 404. The softlimit does that, but I can't see how to use
it without copying all of this...
"""
paginator = self.get_paginator(
queryset,
page_size,
orphans = self.get_paginate_orphans(),
allow_empty_first_page = self.get_allow_empty(),
body = self.paginator_body,
margin = self.paginator_margin,
padding = self.paginator_padding,
tail = self.paginator_tail,
)
page_kwarg = self.page_kwarg
page = self.kwargs.get(page_kwarg) or self.request.GET.get(page_kwarg) or 1
try:
page_number = int(page)
except ValueError:
if page == 'last':
page_number = paginator.num_pages
else:
raise Http404(_("Page is not 'last', nor can it be converted to an int."))
try:
page = paginator.page(page_number, softlimit=False)
return (paginator, page, page.object_list, page.has_other_pages())
except InvalidPage as e:
raise Http404(_('Invalid page (%(page_number)s): %(message)s') % {
'page_number': page_number,
'message': str(e)
})
|
python
|
def paginate_queryset(self, queryset, page_size):
"""
Paginate the queryset, if needed.
This is EXACTLY the same as the standard ListView.paginate_queryset()
except for this line:
page = paginator.page(page_number, softlimit=True)
Because we want to use the DiggPaginator's softlimit option.
So that if you're viewing a page of, say, Flickr photos, and you switch
from viewing by Uploaded Time to viewing by Taken Time, the new
ordering might have fewer pages. In that case we want to see the final
page, not a 404. The softlimit does that, but I can't see how to use
it without copying all of this...
"""
paginator = self.get_paginator(
queryset,
page_size,
orphans = self.get_paginate_orphans(),
allow_empty_first_page = self.get_allow_empty(),
body = self.paginator_body,
margin = self.paginator_margin,
padding = self.paginator_padding,
tail = self.paginator_tail,
)
page_kwarg = self.page_kwarg
page = self.kwargs.get(page_kwarg) or self.request.GET.get(page_kwarg) or 1
try:
page_number = int(page)
except ValueError:
if page == 'last':
page_number = paginator.num_pages
else:
raise Http404(_("Page is not 'last', nor can it be converted to an int."))
try:
page = paginator.page(page_number, softlimit=False)
return (paginator, page, page.object_list, page.has_other_pages())
except InvalidPage as e:
raise Http404(_('Invalid page (%(page_number)s): %(message)s') % {
'page_number': page_number,
'message': str(e)
})
|
[
"def",
"paginate_queryset",
"(",
"self",
",",
"queryset",
",",
"page_size",
")",
":",
"paginator",
"=",
"self",
".",
"get_paginator",
"(",
"queryset",
",",
"page_size",
",",
"orphans",
"=",
"self",
".",
"get_paginate_orphans",
"(",
")",
",",
"allow_empty_first_page",
"=",
"self",
".",
"get_allow_empty",
"(",
")",
",",
"body",
"=",
"self",
".",
"paginator_body",
",",
"margin",
"=",
"self",
".",
"paginator_margin",
",",
"padding",
"=",
"self",
".",
"paginator_padding",
",",
"tail",
"=",
"self",
".",
"paginator_tail",
",",
")",
"page_kwarg",
"=",
"self",
".",
"page_kwarg",
"page",
"=",
"self",
".",
"kwargs",
".",
"get",
"(",
"page_kwarg",
")",
"or",
"self",
".",
"request",
".",
"GET",
".",
"get",
"(",
"page_kwarg",
")",
"or",
"1",
"try",
":",
"page_number",
"=",
"int",
"(",
"page",
")",
"except",
"ValueError",
":",
"if",
"page",
"==",
"'last'",
":",
"page_number",
"=",
"paginator",
".",
"num_pages",
"else",
":",
"raise",
"Http404",
"(",
"_",
"(",
"\"Page is not 'last', nor can it be converted to an int.\"",
")",
")",
"try",
":",
"page",
"=",
"paginator",
".",
"page",
"(",
"page_number",
",",
"softlimit",
"=",
"False",
")",
"return",
"(",
"paginator",
",",
"page",
",",
"page",
".",
"object_list",
",",
"page",
".",
"has_other_pages",
"(",
")",
")",
"except",
"InvalidPage",
"as",
"e",
":",
"raise",
"Http404",
"(",
"_",
"(",
"'Invalid page (%(page_number)s): %(message)s'",
")",
"%",
"{",
"'page_number'",
":",
"page_number",
",",
"'message'",
":",
"str",
"(",
"e",
")",
"}",
")"
] |
Paginate the queryset, if needed.
This is EXACTLY the same as the standard ListView.paginate_queryset()
except for this line:
page = paginator.page(page_number, softlimit=True)
Because we want to use the DiggPaginator's softlimit option.
So that if you're viewing a page of, say, Flickr photos, and you switch
from viewing by Uploaded Time to viewing by Taken Time, the new
ordering might have fewer pages. In that case we want to see the final
page, not a 404. The softlimit does that, but I can't see how to use
it without copying all of this...
|
[
"Paginate",
"the",
"queryset",
"if",
"needed",
"."
] |
f3c72004f9caa1fde0f5a3b2f0d2bf285fc01ada
|
https://github.com/philgyford/django-spectator/blob/f3c72004f9caa1fde0f5a3b2f0d2bf285fc01ada/spectator/core/views.py#L33-L73
|
train
|
philgyford/django-spectator
|
spectator/reading/templatetags/spectator_reading.py
|
annual_reading_counts_card
|
def annual_reading_counts_card(kind='all', current_year=None):
"""
Displays years and the number of books/periodicals read per year.
kind is one of 'book', 'periodical', 'all' (default).
current_year is an optional date object representing the year we're already
showing information about.
"""
if kind == 'book':
card_title = 'Books per year'
elif kind == 'periodical':
card_title = 'Periodicals per year'
else:
card_title = 'Reading per year'
return {
'card_title': card_title,
'kind': kind,
'years': utils.annual_reading_counts(kind),
'current_year': current_year
}
|
python
|
def annual_reading_counts_card(kind='all', current_year=None):
"""
Displays years and the number of books/periodicals read per year.
kind is one of 'book', 'periodical', 'all' (default).
current_year is an optional date object representing the year we're already
showing information about.
"""
if kind == 'book':
card_title = 'Books per year'
elif kind == 'periodical':
card_title = 'Periodicals per year'
else:
card_title = 'Reading per year'
return {
'card_title': card_title,
'kind': kind,
'years': utils.annual_reading_counts(kind),
'current_year': current_year
}
|
[
"def",
"annual_reading_counts_card",
"(",
"kind",
"=",
"'all'",
",",
"current_year",
"=",
"None",
")",
":",
"if",
"kind",
"==",
"'book'",
":",
"card_title",
"=",
"'Books per year'",
"elif",
"kind",
"==",
"'periodical'",
":",
"card_title",
"=",
"'Periodicals per year'",
"else",
":",
"card_title",
"=",
"'Reading per year'",
"return",
"{",
"'card_title'",
":",
"card_title",
",",
"'kind'",
":",
"kind",
",",
"'years'",
":",
"utils",
".",
"annual_reading_counts",
"(",
"kind",
")",
",",
"'current_year'",
":",
"current_year",
"}"
] |
Displays years and the number of books/periodicals read per year.
kind is one of 'book', 'periodical', 'all' (default).
current_year is an optional date object representing the year we're already
showing information about.
|
[
"Displays",
"years",
"and",
"the",
"number",
"of",
"books",
"/",
"periodicals",
"read",
"per",
"year",
"."
] |
f3c72004f9caa1fde0f5a3b2f0d2bf285fc01ada
|
https://github.com/philgyford/django-spectator/blob/f3c72004f9caa1fde0f5a3b2f0d2bf285fc01ada/spectator/reading/templatetags/spectator_reading.py#L33-L53
|
train
|
philgyford/django-spectator
|
spectator/reading/templatetags/spectator_reading.py
|
day_publications
|
def day_publications(date):
"""
Returns a QuerySet of Publications that were being read on `date`.
`date` is a date tobject.
"""
readings = Reading.objects \
.filter(start_date__lte=date) \
.filter(
Q(end_date__gte=date)
|
Q(end_date__isnull=True)
)
if readings:
return Publication.objects.filter(reading__in=readings) \
.select_related('series') \
.prefetch_related('roles__creator') \
.distinct()
else:
return Publication.objects.none()
|
python
|
def day_publications(date):
"""
Returns a QuerySet of Publications that were being read on `date`.
`date` is a date tobject.
"""
readings = Reading.objects \
.filter(start_date__lte=date) \
.filter(
Q(end_date__gte=date)
|
Q(end_date__isnull=True)
)
if readings:
return Publication.objects.filter(reading__in=readings) \
.select_related('series') \
.prefetch_related('roles__creator') \
.distinct()
else:
return Publication.objects.none()
|
[
"def",
"day_publications",
"(",
"date",
")",
":",
"readings",
"=",
"Reading",
".",
"objects",
".",
"filter",
"(",
"start_date__lte",
"=",
"date",
")",
".",
"filter",
"(",
"Q",
"(",
"end_date__gte",
"=",
"date",
")",
"|",
"Q",
"(",
"end_date__isnull",
"=",
"True",
")",
")",
"if",
"readings",
":",
"return",
"Publication",
".",
"objects",
".",
"filter",
"(",
"reading__in",
"=",
"readings",
")",
".",
"select_related",
"(",
"'series'",
")",
".",
"prefetch_related",
"(",
"'roles__creator'",
")",
".",
"distinct",
"(",
")",
"else",
":",
"return",
"Publication",
".",
"objects",
".",
"none",
"(",
")"
] |
Returns a QuerySet of Publications that were being read on `date`.
`date` is a date tobject.
|
[
"Returns",
"a",
"QuerySet",
"of",
"Publications",
"that",
"were",
"being",
"read",
"on",
"date",
".",
"date",
"is",
"a",
"date",
"tobject",
"."
] |
f3c72004f9caa1fde0f5a3b2f0d2bf285fc01ada
|
https://github.com/philgyford/django-spectator/blob/f3c72004f9caa1fde0f5a3b2f0d2bf285fc01ada/spectator/reading/templatetags/spectator_reading.py#L78-L96
|
train
|
philgyford/django-spectator
|
spectator/reading/templatetags/spectator_reading.py
|
day_publications_card
|
def day_publications_card(date):
"""
Displays Publications that were being read on `date`.
`date` is a date tobject.
"""
d = date.strftime(app_settings.DATE_FORMAT)
card_title = 'Reading on {}'.format(d)
return {
'card_title': card_title,
'publication_list': day_publications(date=date),
}
|
python
|
def day_publications_card(date):
"""
Displays Publications that were being read on `date`.
`date` is a date tobject.
"""
d = date.strftime(app_settings.DATE_FORMAT)
card_title = 'Reading on {}'.format(d)
return {
'card_title': card_title,
'publication_list': day_publications(date=date),
}
|
[
"def",
"day_publications_card",
"(",
"date",
")",
":",
"d",
"=",
"date",
".",
"strftime",
"(",
"app_settings",
".",
"DATE_FORMAT",
")",
"card_title",
"=",
"'Reading on {}'",
".",
"format",
"(",
"d",
")",
"return",
"{",
"'card_title'",
":",
"card_title",
",",
"'publication_list'",
":",
"day_publications",
"(",
"date",
"=",
"date",
")",
",",
"}"
] |
Displays Publications that were being read on `date`.
`date` is a date tobject.
|
[
"Displays",
"Publications",
"that",
"were",
"being",
"read",
"on",
"date",
".",
"date",
"is",
"a",
"date",
"tobject",
"."
] |
f3c72004f9caa1fde0f5a3b2f0d2bf285fc01ada
|
https://github.com/philgyford/django-spectator/blob/f3c72004f9caa1fde0f5a3b2f0d2bf285fc01ada/spectator/reading/templatetags/spectator_reading.py#L100-L110
|
train
|
philgyford/django-spectator
|
spectator/reading/templatetags/spectator_reading.py
|
reading_dates
|
def reading_dates(reading):
"""
Given a Reading, with start and end dates and granularities[1] it returns
an HTML string representing that period. eg:
* '1–6 Feb 2017'
* '1 Feb to 3 Mar 2017'
* 'Feb 2017 to Mar 2018'
* '2017–2018'
etc.
[1] https://www.flickr.com/services/api/misc.dates.html
"""
# 3 September 2017
full_format = '<time datetime="%Y-%m-%d">{}</time>'.format('%-d %B %Y')
# September 2017
month_year_format = '<time datetime="%Y-%m">{}</time>'.format('%B %Y')
# 2017
year_format = '<time datetime="%Y">{}</time>'.format('%Y')
# 3
day_format = '<time datetime="%Y-%m-%d">{}</time>'.format('%-d')
# 3 September
day_month_format = '<time datetime="%Y-%m-%d">{}</time>'.format('%-d %B')
# September
month_format = '<time datetime="%Y-%m">{}</time>'.format('%B')
period_format_short = '{}–{}'
period_format_long = '{} to {}'
# For brevity:
start_date = reading.start_date
end_date = reading.end_date
start_gran = reading.start_granularity
end_gran = reading.end_granularity
# Are start and end in the same day, year or month?
same_day = False
same_month = False
same_year = False
if start_date and end_date:
if start_date.strftime('%Y') == end_date.strftime('%Y'):
same_year = True
if start_date.strftime('%m%Y') == end_date.strftime('%m%Y'):
same_month = True
if start_date.strftime('%d%m%Y') == end_date.strftime('%d%m%Y'):
same_day = True
start_str = ''
end_str = ''
output = ''
# Make some basic start and end strings, which we might use...
if start_date:
if start_gran == 3:
start_str = start_date.strftime(full_format)
elif start_gran == 4:
start_str = start_date.strftime(month_year_format)
else:
start_str = start_date.strftime(year_format)
if end_date:
if end_gran == 3:
end_str = end_date.strftime(full_format)
elif end_gran == 4:
end_str = end_date.strftime(month_year_format)
else:
end_str = end_date.strftime(year_format)
# Now make the final strings we'll return:
if start_date and end_date:
# A default which will be overridden in many cases. This covers:
# 1 February 2017 to 3 March 2018
# 1 February 2017 to March 2018
# 1 February 2017 to 2018
# February 2017 to 3 March 2018
# February 2017 to March 2018
# February 2017 to 2018
# 2017 to 3 March 2018
# 2017 to March 2018
# 2017 to 2018
output = period_format_long.format(start_str, end_str)
if (start_gran == 4 or end_gran == 4) and same_month:
# Only have enough to output 'February 2017'.
output = start_str
elif (start_gran == 6 or end_gran == 6) and same_year:
# Only have enough to output '2017'.
output = start_str
elif start_gran == 3:
if end_gran == 3:
if same_day:
# 1 February 2017
output = start_str
elif same_month:
# 1–6 February 2017
output = period_format_short.format(
start_date.strftime(day_format),
end_str)
elif same_year:
# 1 February to 3 March 2017
output = period_format_long.format(
start_date.strftime(day_month_format),
end_str)
elif end_gran == 4:
if same_year:
# 1 February to March 2017
output = period_format_long.format(
start_date.strftime(day_month_format),
end_str)
elif start_gran == 4:
if end_gran == 3:
if same_year:
# February to 3 March 2017
output = period_format_long.format(
start_date.strftime(month_format),
end_str)
elif end_gran == 4:
if same_year:
# February to March 2017
output = period_format_long.format(
start_date.strftime(month_format),
end_str)
elif end_date:
# Only an end_date.
if end_gran == 3:
# Finished on 1 February 2017
output = "Finished on {}".format(end_str)
else:
# Finished in February 2017
# Finished in 2017
output = "Finished in {}".format(end_str)
else:
# No end_date: the reading has started, but not ended.
if start_gran == 3:
# Started on 1 February 2017
output = "Started on {}".format(start_str)
else:
# Started in February 2017
# Started in 2017
output = "Started in {}".format(start_str)
return format_html(output)
|
python
|
def reading_dates(reading):
"""
Given a Reading, with start and end dates and granularities[1] it returns
an HTML string representing that period. eg:
* '1–6 Feb 2017'
* '1 Feb to 3 Mar 2017'
* 'Feb 2017 to Mar 2018'
* '2017–2018'
etc.
[1] https://www.flickr.com/services/api/misc.dates.html
"""
# 3 September 2017
full_format = '<time datetime="%Y-%m-%d">{}</time>'.format('%-d %B %Y')
# September 2017
month_year_format = '<time datetime="%Y-%m">{}</time>'.format('%B %Y')
# 2017
year_format = '<time datetime="%Y">{}</time>'.format('%Y')
# 3
day_format = '<time datetime="%Y-%m-%d">{}</time>'.format('%-d')
# 3 September
day_month_format = '<time datetime="%Y-%m-%d">{}</time>'.format('%-d %B')
# September
month_format = '<time datetime="%Y-%m">{}</time>'.format('%B')
period_format_short = '{}–{}'
period_format_long = '{} to {}'
# For brevity:
start_date = reading.start_date
end_date = reading.end_date
start_gran = reading.start_granularity
end_gran = reading.end_granularity
# Are start and end in the same day, year or month?
same_day = False
same_month = False
same_year = False
if start_date and end_date:
if start_date.strftime('%Y') == end_date.strftime('%Y'):
same_year = True
if start_date.strftime('%m%Y') == end_date.strftime('%m%Y'):
same_month = True
if start_date.strftime('%d%m%Y') == end_date.strftime('%d%m%Y'):
same_day = True
start_str = ''
end_str = ''
output = ''
# Make some basic start and end strings, which we might use...
if start_date:
if start_gran == 3:
start_str = start_date.strftime(full_format)
elif start_gran == 4:
start_str = start_date.strftime(month_year_format)
else:
start_str = start_date.strftime(year_format)
if end_date:
if end_gran == 3:
end_str = end_date.strftime(full_format)
elif end_gran == 4:
end_str = end_date.strftime(month_year_format)
else:
end_str = end_date.strftime(year_format)
# Now make the final strings we'll return:
if start_date and end_date:
# A default which will be overridden in many cases. This covers:
# 1 February 2017 to 3 March 2018
# 1 February 2017 to March 2018
# 1 February 2017 to 2018
# February 2017 to 3 March 2018
# February 2017 to March 2018
# February 2017 to 2018
# 2017 to 3 March 2018
# 2017 to March 2018
# 2017 to 2018
output = period_format_long.format(start_str, end_str)
if (start_gran == 4 or end_gran == 4) and same_month:
# Only have enough to output 'February 2017'.
output = start_str
elif (start_gran == 6 or end_gran == 6) and same_year:
# Only have enough to output '2017'.
output = start_str
elif start_gran == 3:
if end_gran == 3:
if same_day:
# 1 February 2017
output = start_str
elif same_month:
# 1–6 February 2017
output = period_format_short.format(
start_date.strftime(day_format),
end_str)
elif same_year:
# 1 February to 3 March 2017
output = period_format_long.format(
start_date.strftime(day_month_format),
end_str)
elif end_gran == 4:
if same_year:
# 1 February to March 2017
output = period_format_long.format(
start_date.strftime(day_month_format),
end_str)
elif start_gran == 4:
if end_gran == 3:
if same_year:
# February to 3 March 2017
output = period_format_long.format(
start_date.strftime(month_format),
end_str)
elif end_gran == 4:
if same_year:
# February to March 2017
output = period_format_long.format(
start_date.strftime(month_format),
end_str)
elif end_date:
# Only an end_date.
if end_gran == 3:
# Finished on 1 February 2017
output = "Finished on {}".format(end_str)
else:
# Finished in February 2017
# Finished in 2017
output = "Finished in {}".format(end_str)
else:
# No end_date: the reading has started, but not ended.
if start_gran == 3:
# Started on 1 February 2017
output = "Started on {}".format(start_str)
else:
# Started in February 2017
# Started in 2017
output = "Started in {}".format(start_str)
return format_html(output)
|
[
"def",
"reading_dates",
"(",
"reading",
")",
":",
"# 3 September 2017",
"full_format",
"=",
"'<time datetime=\"%Y-%m-%d\">{}</time>'",
".",
"format",
"(",
"'%-d %B %Y'",
")",
"# September 2017",
"month_year_format",
"=",
"'<time datetime=\"%Y-%m\">{}</time>'",
".",
"format",
"(",
"'%B %Y'",
")",
"# 2017",
"year_format",
"=",
"'<time datetime=\"%Y\">{}</time>'",
".",
"format",
"(",
"'%Y'",
")",
"# 3",
"day_format",
"=",
"'<time datetime=\"%Y-%m-%d\">{}</time>'",
".",
"format",
"(",
"'%-d'",
")",
"# 3 September",
"day_month_format",
"=",
"'<time datetime=\"%Y-%m-%d\">{}</time>'",
".",
"format",
"(",
"'%-d %B'",
")",
"# September",
"month_format",
"=",
"'<time datetime=\"%Y-%m\">{}</time>'",
".",
"format",
"(",
"'%B'",
")",
"period_format_short",
"=",
"'{}–{}'",
"period_format_long",
"=",
"'{} to {}'",
"# For brevity:",
"start_date",
"=",
"reading",
".",
"start_date",
"end_date",
"=",
"reading",
".",
"end_date",
"start_gran",
"=",
"reading",
".",
"start_granularity",
"end_gran",
"=",
"reading",
".",
"end_granularity",
"# Are start and end in the same day, year or month?",
"same_day",
"=",
"False",
"same_month",
"=",
"False",
"same_year",
"=",
"False",
"if",
"start_date",
"and",
"end_date",
":",
"if",
"start_date",
".",
"strftime",
"(",
"'%Y'",
")",
"==",
"end_date",
".",
"strftime",
"(",
"'%Y'",
")",
":",
"same_year",
"=",
"True",
"if",
"start_date",
".",
"strftime",
"(",
"'%m%Y'",
")",
"==",
"end_date",
".",
"strftime",
"(",
"'%m%Y'",
")",
":",
"same_month",
"=",
"True",
"if",
"start_date",
".",
"strftime",
"(",
"'%d%m%Y'",
")",
"==",
"end_date",
".",
"strftime",
"(",
"'%d%m%Y'",
")",
":",
"same_day",
"=",
"True",
"start_str",
"=",
"''",
"end_str",
"=",
"''",
"output",
"=",
"''",
"# Make some basic start and end strings, which we might use...",
"if",
"start_date",
":",
"if",
"start_gran",
"==",
"3",
":",
"start_str",
"=",
"start_date",
".",
"strftime",
"(",
"full_format",
")",
"elif",
"start_gran",
"==",
"4",
":",
"start_str",
"=",
"start_date",
".",
"strftime",
"(",
"month_year_format",
")",
"else",
":",
"start_str",
"=",
"start_date",
".",
"strftime",
"(",
"year_format",
")",
"if",
"end_date",
":",
"if",
"end_gran",
"==",
"3",
":",
"end_str",
"=",
"end_date",
".",
"strftime",
"(",
"full_format",
")",
"elif",
"end_gran",
"==",
"4",
":",
"end_str",
"=",
"end_date",
".",
"strftime",
"(",
"month_year_format",
")",
"else",
":",
"end_str",
"=",
"end_date",
".",
"strftime",
"(",
"year_format",
")",
"# Now make the final strings we'll return:",
"if",
"start_date",
"and",
"end_date",
":",
"# A default which will be overridden in many cases. This covers:",
"# 1 February 2017 to 3 March 2018",
"# 1 February 2017 to March 2018",
"# 1 February 2017 to 2018",
"# February 2017 to 3 March 2018",
"# February 2017 to March 2018",
"# February 2017 to 2018",
"# 2017 to 3 March 2018",
"# 2017 to March 2018",
"# 2017 to 2018",
"output",
"=",
"period_format_long",
".",
"format",
"(",
"start_str",
",",
"end_str",
")",
"if",
"(",
"start_gran",
"==",
"4",
"or",
"end_gran",
"==",
"4",
")",
"and",
"same_month",
":",
"# Only have enough to output 'February 2017'.",
"output",
"=",
"start_str",
"elif",
"(",
"start_gran",
"==",
"6",
"or",
"end_gran",
"==",
"6",
")",
"and",
"same_year",
":",
"# Only have enough to output '2017'.",
"output",
"=",
"start_str",
"elif",
"start_gran",
"==",
"3",
":",
"if",
"end_gran",
"==",
"3",
":",
"if",
"same_day",
":",
"# 1 February 2017",
"output",
"=",
"start_str",
"elif",
"same_month",
":",
"# 1–6 February 2017",
"output",
"=",
"period_format_short",
".",
"format",
"(",
"start_date",
".",
"strftime",
"(",
"day_format",
")",
",",
"end_str",
")",
"elif",
"same_year",
":",
"# 1 February to 3 March 2017",
"output",
"=",
"period_format_long",
".",
"format",
"(",
"start_date",
".",
"strftime",
"(",
"day_month_format",
")",
",",
"end_str",
")",
"elif",
"end_gran",
"==",
"4",
":",
"if",
"same_year",
":",
"# 1 February to March 2017",
"output",
"=",
"period_format_long",
".",
"format",
"(",
"start_date",
".",
"strftime",
"(",
"day_month_format",
")",
",",
"end_str",
")",
"elif",
"start_gran",
"==",
"4",
":",
"if",
"end_gran",
"==",
"3",
":",
"if",
"same_year",
":",
"# February to 3 March 2017",
"output",
"=",
"period_format_long",
".",
"format",
"(",
"start_date",
".",
"strftime",
"(",
"month_format",
")",
",",
"end_str",
")",
"elif",
"end_gran",
"==",
"4",
":",
"if",
"same_year",
":",
"# February to March 2017",
"output",
"=",
"period_format_long",
".",
"format",
"(",
"start_date",
".",
"strftime",
"(",
"month_format",
")",
",",
"end_str",
")",
"elif",
"end_date",
":",
"# Only an end_date.",
"if",
"end_gran",
"==",
"3",
":",
"# Finished on 1 February 2017",
"output",
"=",
"\"Finished on {}\"",
".",
"format",
"(",
"end_str",
")",
"else",
":",
"# Finished in February 2017",
"# Finished in 2017",
"output",
"=",
"\"Finished in {}\"",
".",
"format",
"(",
"end_str",
")",
"else",
":",
"# No end_date: the reading has started, but not ended.",
"if",
"start_gran",
"==",
"3",
":",
"# Started on 1 February 2017",
"output",
"=",
"\"Started on {}\"",
".",
"format",
"(",
"start_str",
")",
"else",
":",
"# Started in February 2017",
"# Started in 2017",
"output",
"=",
"\"Started in {}\"",
".",
"format",
"(",
"start_str",
")",
"return",
"format_html",
"(",
"output",
")"
] |
Given a Reading, with start and end dates and granularities[1] it returns
an HTML string representing that period. eg:
* '1–6 Feb 2017'
* '1 Feb to 3 Mar 2017'
* 'Feb 2017 to Mar 2018'
* '2017–2018'
etc.
[1] https://www.flickr.com/services/api/misc.dates.html
|
[
"Given",
"a",
"Reading",
"with",
"start",
"and",
"end",
"dates",
"and",
"granularities",
"[",
"1",
"]",
"it",
"returns",
"an",
"HTML",
"string",
"representing",
"that",
"period",
".",
"eg",
":"
] |
f3c72004f9caa1fde0f5a3b2f0d2bf285fc01ada
|
https://github.com/philgyford/django-spectator/blob/f3c72004f9caa1fde0f5a3b2f0d2bf285fc01ada/spectator/reading/templatetags/spectator_reading.py#L136-L288
|
train
|
philgyford/django-spectator
|
spectator/events/migrations/0035_change_event_kinds.py
|
forwards
|
def forwards(apps, schema_editor):
"""
Change Events with kind 'movie' to 'cinema'
and Events with kind 'play' to 'theatre'.
Purely for more consistency.
"""
Event = apps.get_model('spectator_events', 'Event')
for ev in Event.objects.filter(kind='movie'):
ev.kind = 'cinema'
ev.save()
for ev in Event.objects.filter(kind='play'):
ev.kind = 'theatre'
ev.save()
|
python
|
def forwards(apps, schema_editor):
"""
Change Events with kind 'movie' to 'cinema'
and Events with kind 'play' to 'theatre'.
Purely for more consistency.
"""
Event = apps.get_model('spectator_events', 'Event')
for ev in Event.objects.filter(kind='movie'):
ev.kind = 'cinema'
ev.save()
for ev in Event.objects.filter(kind='play'):
ev.kind = 'theatre'
ev.save()
|
[
"def",
"forwards",
"(",
"apps",
",",
"schema_editor",
")",
":",
"Event",
"=",
"apps",
".",
"get_model",
"(",
"'spectator_events'",
",",
"'Event'",
")",
"for",
"ev",
"in",
"Event",
".",
"objects",
".",
"filter",
"(",
"kind",
"=",
"'movie'",
")",
":",
"ev",
".",
"kind",
"=",
"'cinema'",
"ev",
".",
"save",
"(",
")",
"for",
"ev",
"in",
"Event",
".",
"objects",
".",
"filter",
"(",
"kind",
"=",
"'play'",
")",
":",
"ev",
".",
"kind",
"=",
"'theatre'",
"ev",
".",
"save",
"(",
")"
] |
Change Events with kind 'movie' to 'cinema'
and Events with kind 'play' to 'theatre'.
Purely for more consistency.
|
[
"Change",
"Events",
"with",
"kind",
"movie",
"to",
"cinema",
"and",
"Events",
"with",
"kind",
"play",
"to",
"theatre",
"."
] |
f3c72004f9caa1fde0f5a3b2f0d2bf285fc01ada
|
https://github.com/philgyford/django-spectator/blob/f3c72004f9caa1fde0f5a3b2f0d2bf285fc01ada/spectator/events/migrations/0035_change_event_kinds.py#L6-L21
|
train
|
philgyford/django-spectator
|
devproject/devproject/settings.py
|
get_env_variable
|
def get_env_variable(var_name, default=None):
"""Get the environment variable or return exception."""
try:
return os.environ[var_name]
except KeyError:
if default is None:
error_msg = "Set the %s environment variable" % var_name
raise ImproperlyConfigured(error_msg)
else:
return default
|
python
|
def get_env_variable(var_name, default=None):
"""Get the environment variable or return exception."""
try:
return os.environ[var_name]
except KeyError:
if default is None:
error_msg = "Set the %s environment variable" % var_name
raise ImproperlyConfigured(error_msg)
else:
return default
|
[
"def",
"get_env_variable",
"(",
"var_name",
",",
"default",
"=",
"None",
")",
":",
"try",
":",
"return",
"os",
".",
"environ",
"[",
"var_name",
"]",
"except",
"KeyError",
":",
"if",
"default",
"is",
"None",
":",
"error_msg",
"=",
"\"Set the %s environment variable\"",
"%",
"var_name",
"raise",
"ImproperlyConfigured",
"(",
"error_msg",
")",
"else",
":",
"return",
"default"
] |
Get the environment variable or return exception.
|
[
"Get",
"the",
"environment",
"variable",
"or",
"return",
"exception",
"."
] |
f3c72004f9caa1fde0f5a3b2f0d2bf285fc01ada
|
https://github.com/philgyford/django-spectator/blob/f3c72004f9caa1fde0f5a3b2f0d2bf285fc01ada/devproject/devproject/settings.py#L21-L30
|
train
|
philgyford/django-spectator
|
spectator/events/migrations/0039_populate_exhibitions.py
|
generate_slug
|
def generate_slug(value):
"""
Generates a slug using a Hashid of `value`.
COPIED from spectator.core.models.SluggedModelMixin() because migrations
don't make this happen automatically and perhaps the least bad thing is
to copy the method here, ugh.
"""
alphabet = app_settings.SLUG_ALPHABET
salt = app_settings.SLUG_SALT
hashids = Hashids(alphabet=alphabet, salt=salt, min_length=5)
return hashids.encode(value)
|
python
|
def generate_slug(value):
"""
Generates a slug using a Hashid of `value`.
COPIED from spectator.core.models.SluggedModelMixin() because migrations
don't make this happen automatically and perhaps the least bad thing is
to copy the method here, ugh.
"""
alphabet = app_settings.SLUG_ALPHABET
salt = app_settings.SLUG_SALT
hashids = Hashids(alphabet=alphabet, salt=salt, min_length=5)
return hashids.encode(value)
|
[
"def",
"generate_slug",
"(",
"value",
")",
":",
"alphabet",
"=",
"app_settings",
".",
"SLUG_ALPHABET",
"salt",
"=",
"app_settings",
".",
"SLUG_SALT",
"hashids",
"=",
"Hashids",
"(",
"alphabet",
"=",
"alphabet",
",",
"salt",
"=",
"salt",
",",
"min_length",
"=",
"5",
")",
"return",
"hashids",
".",
"encode",
"(",
"value",
")"
] |
Generates a slug using a Hashid of `value`.
COPIED from spectator.core.models.SluggedModelMixin() because migrations
don't make this happen automatically and perhaps the least bad thing is
to copy the method here, ugh.
|
[
"Generates",
"a",
"slug",
"using",
"a",
"Hashid",
"of",
"value",
"."
] |
f3c72004f9caa1fde0f5a3b2f0d2bf285fc01ada
|
https://github.com/philgyford/django-spectator/blob/f3c72004f9caa1fde0f5a3b2f0d2bf285fc01ada/spectator/events/migrations/0039_populate_exhibitions.py#L11-L24
|
train
|
philgyford/django-spectator
|
spectator/events/migrations/0039_populate_exhibitions.py
|
forwards
|
def forwards(apps, schema_editor):
"""
Having added the new 'exhibition' Work type, we're going to assume that
every Event of type 'museum' should actually have one Exhibition attached.
So, we'll add one, with the same title as the Event.
And we'll move all Creators from the Event to the Exhibition.
"""
Event = apps.get_model('spectator_events', 'Event')
Work = apps.get_model('spectator_events', 'Work')
WorkRole = apps.get_model('spectator_events', 'WorkRole')
WorkSelection = apps.get_model('spectator_events', 'WorkSelection')
for event in Event.objects.filter(kind='museum'):
# Create a new Work based on this Event's details.
work = Work.objects.create(
kind='exhibition',
title=event.title,
title_sort=event.title_sort
)
# This doesn't generate the slug field automatically because Django.
# So we'll have to do it manually. Graarhhh.
work.slug = generate_slug(work.pk)
work.save()
# Associate the new Work with the Event.
WorkSelection.objects.create(
event=event,
work=work
)
# Associate any Creators on the Event with the new Work.
for role in event.roles.all():
WorkRole.objects.create(
creator=role.creator,
work=work,
role_name=role.role_name,
role_order=role.role_order
)
# Remove Creators from the Event.
role.delete()
|
python
|
def forwards(apps, schema_editor):
"""
Having added the new 'exhibition' Work type, we're going to assume that
every Event of type 'museum' should actually have one Exhibition attached.
So, we'll add one, with the same title as the Event.
And we'll move all Creators from the Event to the Exhibition.
"""
Event = apps.get_model('spectator_events', 'Event')
Work = apps.get_model('spectator_events', 'Work')
WorkRole = apps.get_model('spectator_events', 'WorkRole')
WorkSelection = apps.get_model('spectator_events', 'WorkSelection')
for event in Event.objects.filter(kind='museum'):
# Create a new Work based on this Event's details.
work = Work.objects.create(
kind='exhibition',
title=event.title,
title_sort=event.title_sort
)
# This doesn't generate the slug field automatically because Django.
# So we'll have to do it manually. Graarhhh.
work.slug = generate_slug(work.pk)
work.save()
# Associate the new Work with the Event.
WorkSelection.objects.create(
event=event,
work=work
)
# Associate any Creators on the Event with the new Work.
for role in event.roles.all():
WorkRole.objects.create(
creator=role.creator,
work=work,
role_name=role.role_name,
role_order=role.role_order
)
# Remove Creators from the Event.
role.delete()
|
[
"def",
"forwards",
"(",
"apps",
",",
"schema_editor",
")",
":",
"Event",
"=",
"apps",
".",
"get_model",
"(",
"'spectator_events'",
",",
"'Event'",
")",
"Work",
"=",
"apps",
".",
"get_model",
"(",
"'spectator_events'",
",",
"'Work'",
")",
"WorkRole",
"=",
"apps",
".",
"get_model",
"(",
"'spectator_events'",
",",
"'WorkRole'",
")",
"WorkSelection",
"=",
"apps",
".",
"get_model",
"(",
"'spectator_events'",
",",
"'WorkSelection'",
")",
"for",
"event",
"in",
"Event",
".",
"objects",
".",
"filter",
"(",
"kind",
"=",
"'museum'",
")",
":",
"# Create a new Work based on this Event's details.",
"work",
"=",
"Work",
".",
"objects",
".",
"create",
"(",
"kind",
"=",
"'exhibition'",
",",
"title",
"=",
"event",
".",
"title",
",",
"title_sort",
"=",
"event",
".",
"title_sort",
")",
"# This doesn't generate the slug field automatically because Django.",
"# So we'll have to do it manually. Graarhhh.",
"work",
".",
"slug",
"=",
"generate_slug",
"(",
"work",
".",
"pk",
")",
"work",
".",
"save",
"(",
")",
"# Associate the new Work with the Event.",
"WorkSelection",
".",
"objects",
".",
"create",
"(",
"event",
"=",
"event",
",",
"work",
"=",
"work",
")",
"# Associate any Creators on the Event with the new Work.",
"for",
"role",
"in",
"event",
".",
"roles",
".",
"all",
"(",
")",
":",
"WorkRole",
".",
"objects",
".",
"create",
"(",
"creator",
"=",
"role",
".",
"creator",
",",
"work",
"=",
"work",
",",
"role_name",
"=",
"role",
".",
"role_name",
",",
"role_order",
"=",
"role",
".",
"role_order",
")",
"# Remove Creators from the Event.",
"role",
".",
"delete",
"(",
")"
] |
Having added the new 'exhibition' Work type, we're going to assume that
every Event of type 'museum' should actually have one Exhibition attached.
So, we'll add one, with the same title as the Event.
And we'll move all Creators from the Event to the Exhibition.
|
[
"Having",
"added",
"the",
"new",
"exhibition",
"Work",
"type",
"we",
"re",
"going",
"to",
"assume",
"that",
"every",
"Event",
"of",
"type",
"museum",
"should",
"actually",
"have",
"one",
"Exhibition",
"attached",
"."
] |
f3c72004f9caa1fde0f5a3b2f0d2bf285fc01ada
|
https://github.com/philgyford/django-spectator/blob/f3c72004f9caa1fde0f5a3b2f0d2bf285fc01ada/spectator/events/migrations/0039_populate_exhibitions.py#L27-L70
|
train
|
philgyford/django-spectator
|
spectator/core/managers.py
|
CreatorManager.by_publications
|
def by_publications(self):
"""
The Creators who have been most-read, ordered by number of read
publications (ignoring if any of those publicatinos have been read
multiple times.)
Each Creator will have a `num_publications` attribute.
"""
if not spectator_apps.is_enabled('reading'):
raise ImproperlyConfigured("To use the CreatorManager.by_publications() method, 'spectator.reading' must by in INSTALLED_APPS.")
qs = self.get_queryset()
qs = qs.exclude(publications__reading__isnull=True) \
.annotate(num_publications=Count('publications')) \
.order_by('-num_publications', 'name_sort')
return qs
|
python
|
def by_publications(self):
"""
The Creators who have been most-read, ordered by number of read
publications (ignoring if any of those publicatinos have been read
multiple times.)
Each Creator will have a `num_publications` attribute.
"""
if not spectator_apps.is_enabled('reading'):
raise ImproperlyConfigured("To use the CreatorManager.by_publications() method, 'spectator.reading' must by in INSTALLED_APPS.")
qs = self.get_queryset()
qs = qs.exclude(publications__reading__isnull=True) \
.annotate(num_publications=Count('publications')) \
.order_by('-num_publications', 'name_sort')
return qs
|
[
"def",
"by_publications",
"(",
"self",
")",
":",
"if",
"not",
"spectator_apps",
".",
"is_enabled",
"(",
"'reading'",
")",
":",
"raise",
"ImproperlyConfigured",
"(",
"\"To use the CreatorManager.by_publications() method, 'spectator.reading' must by in INSTALLED_APPS.\"",
")",
"qs",
"=",
"self",
".",
"get_queryset",
"(",
")",
"qs",
"=",
"qs",
".",
"exclude",
"(",
"publications__reading__isnull",
"=",
"True",
")",
".",
"annotate",
"(",
"num_publications",
"=",
"Count",
"(",
"'publications'",
")",
")",
".",
"order_by",
"(",
"'-num_publications'",
",",
"'name_sort'",
")",
"return",
"qs"
] |
The Creators who have been most-read, ordered by number of read
publications (ignoring if any of those publicatinos have been read
multiple times.)
Each Creator will have a `num_publications` attribute.
|
[
"The",
"Creators",
"who",
"have",
"been",
"most",
"-",
"read",
"ordered",
"by",
"number",
"of",
"read",
"publications",
"(",
"ignoring",
"if",
"any",
"of",
"those",
"publicatinos",
"have",
"been",
"read",
"multiple",
"times",
".",
")"
] |
f3c72004f9caa1fde0f5a3b2f0d2bf285fc01ada
|
https://github.com/philgyford/django-spectator/blob/f3c72004f9caa1fde0f5a3b2f0d2bf285fc01ada/spectator/core/managers.py#L10-L27
|
train
|
philgyford/django-spectator
|
spectator/core/managers.py
|
CreatorManager.by_readings
|
def by_readings(self, role_names=['', 'Author']):
"""
The Creators who have been most-read, ordered by number of readings.
By default it will only include Creators whose role was left empty,
or is 'Author'.
Each Creator will have a `num_readings` attribute.
"""
if not spectator_apps.is_enabled('reading'):
raise ImproperlyConfigured("To use the CreatorManager.by_readings() method, 'spectator.reading' must by in INSTALLED_APPS.")
qs = self.get_queryset()
qs = qs.filter(publication_roles__role_name__in=role_names) \
.exclude(publications__reading__isnull=True) \
.annotate(num_readings=Count('publications__reading')) \
.order_by('-num_readings', 'name_sort')
return qs
|
python
|
def by_readings(self, role_names=['', 'Author']):
"""
The Creators who have been most-read, ordered by number of readings.
By default it will only include Creators whose role was left empty,
or is 'Author'.
Each Creator will have a `num_readings` attribute.
"""
if not spectator_apps.is_enabled('reading'):
raise ImproperlyConfigured("To use the CreatorManager.by_readings() method, 'spectator.reading' must by in INSTALLED_APPS.")
qs = self.get_queryset()
qs = qs.filter(publication_roles__role_name__in=role_names) \
.exclude(publications__reading__isnull=True) \
.annotate(num_readings=Count('publications__reading')) \
.order_by('-num_readings', 'name_sort')
return qs
|
[
"def",
"by_readings",
"(",
"self",
",",
"role_names",
"=",
"[",
"''",
",",
"'Author'",
"]",
")",
":",
"if",
"not",
"spectator_apps",
".",
"is_enabled",
"(",
"'reading'",
")",
":",
"raise",
"ImproperlyConfigured",
"(",
"\"To use the CreatorManager.by_readings() method, 'spectator.reading' must by in INSTALLED_APPS.\"",
")",
"qs",
"=",
"self",
".",
"get_queryset",
"(",
")",
"qs",
"=",
"qs",
".",
"filter",
"(",
"publication_roles__role_name__in",
"=",
"role_names",
")",
".",
"exclude",
"(",
"publications__reading__isnull",
"=",
"True",
")",
".",
"annotate",
"(",
"num_readings",
"=",
"Count",
"(",
"'publications__reading'",
")",
")",
".",
"order_by",
"(",
"'-num_readings'",
",",
"'name_sort'",
")",
"return",
"qs"
] |
The Creators who have been most-read, ordered by number of readings.
By default it will only include Creators whose role was left empty,
or is 'Author'.
Each Creator will have a `num_readings` attribute.
|
[
"The",
"Creators",
"who",
"have",
"been",
"most",
"-",
"read",
"ordered",
"by",
"number",
"of",
"readings",
"."
] |
f3c72004f9caa1fde0f5a3b2f0d2bf285fc01ada
|
https://github.com/philgyford/django-spectator/blob/f3c72004f9caa1fde0f5a3b2f0d2bf285fc01ada/spectator/core/managers.py#L29-L48
|
train
|
philgyford/django-spectator
|
spectator/core/managers.py
|
CreatorManager.by_events
|
def by_events(self, kind=None):
"""
Get the Creators involved in the most Events.
This only counts Creators directly involved in an Event.
i.e. if a Creator is the director of a movie Work, and an Event was
a viewing of that movie, that Event wouldn't count. Unless they were
also directly involved in the Event (e.g. speaking after the movie).
kind - If supplied, only Events with that `kind` value will be counted.
"""
if not spectator_apps.is_enabled('events'):
raise ImproperlyConfigured("To use the CreatorManager.by_events() method, 'spectator.events' must by in INSTALLED_APPS.")
qs = self.get_queryset()
if kind is not None:
qs = qs.filter(events__kind=kind)
qs = qs.annotate(num_events=Count('events', distinct=True)) \
.order_by('-num_events', 'name_sort')
return qs
|
python
|
def by_events(self, kind=None):
"""
Get the Creators involved in the most Events.
This only counts Creators directly involved in an Event.
i.e. if a Creator is the director of a movie Work, and an Event was
a viewing of that movie, that Event wouldn't count. Unless they were
also directly involved in the Event (e.g. speaking after the movie).
kind - If supplied, only Events with that `kind` value will be counted.
"""
if not spectator_apps.is_enabled('events'):
raise ImproperlyConfigured("To use the CreatorManager.by_events() method, 'spectator.events' must by in INSTALLED_APPS.")
qs = self.get_queryset()
if kind is not None:
qs = qs.filter(events__kind=kind)
qs = qs.annotate(num_events=Count('events', distinct=True)) \
.order_by('-num_events', 'name_sort')
return qs
|
[
"def",
"by_events",
"(",
"self",
",",
"kind",
"=",
"None",
")",
":",
"if",
"not",
"spectator_apps",
".",
"is_enabled",
"(",
"'events'",
")",
":",
"raise",
"ImproperlyConfigured",
"(",
"\"To use the CreatorManager.by_events() method, 'spectator.events' must by in INSTALLED_APPS.\"",
")",
"qs",
"=",
"self",
".",
"get_queryset",
"(",
")",
"if",
"kind",
"is",
"not",
"None",
":",
"qs",
"=",
"qs",
".",
"filter",
"(",
"events__kind",
"=",
"kind",
")",
"qs",
"=",
"qs",
".",
"annotate",
"(",
"num_events",
"=",
"Count",
"(",
"'events'",
",",
"distinct",
"=",
"True",
")",
")",
".",
"order_by",
"(",
"'-num_events'",
",",
"'name_sort'",
")",
"return",
"qs"
] |
Get the Creators involved in the most Events.
This only counts Creators directly involved in an Event.
i.e. if a Creator is the director of a movie Work, and an Event was
a viewing of that movie, that Event wouldn't count. Unless they were
also directly involved in the Event (e.g. speaking after the movie).
kind - If supplied, only Events with that `kind` value will be counted.
|
[
"Get",
"the",
"Creators",
"involved",
"in",
"the",
"most",
"Events",
"."
] |
f3c72004f9caa1fde0f5a3b2f0d2bf285fc01ada
|
https://github.com/philgyford/django-spectator/blob/f3c72004f9caa1fde0f5a3b2f0d2bf285fc01ada/spectator/core/managers.py#L50-L72
|
train
|
philgyford/django-spectator
|
spectator/core/managers.py
|
CreatorManager.by_works
|
def by_works(self, kind=None, role_name=None):
"""
Get the Creators involved in the most Works.
kind - If supplied, only Works with that `kind` value will be counted.
role_name - If supplied, only Works on which the role is that will be counted.
e.g. To get all 'movie' Works on which the Creators had the role 'Director':
Creator.objects.by_works(kind='movie', role_name='Director')
"""
if not spectator_apps.is_enabled('events'):
raise ImproperlyConfigured("To use the CreatorManager.by_works() method, 'spectator.events' must by in INSTALLED_APPS.")
qs = self.get_queryset()
filter_kwargs = {}
if kind is not None:
filter_kwargs['works__kind'] = kind
if role_name is not None:
filter_kwargs['work_roles__role_name'] = role_name
if filter_kwargs:
qs = qs.filter(**filter_kwargs)
qs = qs.annotate(num_works=Count('works', distinct=True)) \
.order_by('-num_works', 'name_sort')
return qs
|
python
|
def by_works(self, kind=None, role_name=None):
"""
Get the Creators involved in the most Works.
kind - If supplied, only Works with that `kind` value will be counted.
role_name - If supplied, only Works on which the role is that will be counted.
e.g. To get all 'movie' Works on which the Creators had the role 'Director':
Creator.objects.by_works(kind='movie', role_name='Director')
"""
if not spectator_apps.is_enabled('events'):
raise ImproperlyConfigured("To use the CreatorManager.by_works() method, 'spectator.events' must by in INSTALLED_APPS.")
qs = self.get_queryset()
filter_kwargs = {}
if kind is not None:
filter_kwargs['works__kind'] = kind
if role_name is not None:
filter_kwargs['work_roles__role_name'] = role_name
if filter_kwargs:
qs = qs.filter(**filter_kwargs)
qs = qs.annotate(num_works=Count('works', distinct=True)) \
.order_by('-num_works', 'name_sort')
return qs
|
[
"def",
"by_works",
"(",
"self",
",",
"kind",
"=",
"None",
",",
"role_name",
"=",
"None",
")",
":",
"if",
"not",
"spectator_apps",
".",
"is_enabled",
"(",
"'events'",
")",
":",
"raise",
"ImproperlyConfigured",
"(",
"\"To use the CreatorManager.by_works() method, 'spectator.events' must by in INSTALLED_APPS.\"",
")",
"qs",
"=",
"self",
".",
"get_queryset",
"(",
")",
"filter_kwargs",
"=",
"{",
"}",
"if",
"kind",
"is",
"not",
"None",
":",
"filter_kwargs",
"[",
"'works__kind'",
"]",
"=",
"kind",
"if",
"role_name",
"is",
"not",
"None",
":",
"filter_kwargs",
"[",
"'work_roles__role_name'",
"]",
"=",
"role_name",
"if",
"filter_kwargs",
":",
"qs",
"=",
"qs",
".",
"filter",
"(",
"*",
"*",
"filter_kwargs",
")",
"qs",
"=",
"qs",
".",
"annotate",
"(",
"num_works",
"=",
"Count",
"(",
"'works'",
",",
"distinct",
"=",
"True",
")",
")",
".",
"order_by",
"(",
"'-num_works'",
",",
"'name_sort'",
")",
"return",
"qs"
] |
Get the Creators involved in the most Works.
kind - If supplied, only Works with that `kind` value will be counted.
role_name - If supplied, only Works on which the role is that will be counted.
e.g. To get all 'movie' Works on which the Creators had the role 'Director':
Creator.objects.by_works(kind='movie', role_name='Director')
|
[
"Get",
"the",
"Creators",
"involved",
"in",
"the",
"most",
"Works",
"."
] |
f3c72004f9caa1fde0f5a3b2f0d2bf285fc01ada
|
https://github.com/philgyford/django-spectator/blob/f3c72004f9caa1fde0f5a3b2f0d2bf285fc01ada/spectator/core/managers.py#L74-L104
|
train
|
inveniosoftware/invenio-search
|
examples/app.py
|
index
|
def index():
"""Query Elasticsearch using Invenio query syntax."""
page = request.values.get('page', 1, type=int)
size = request.values.get('size', 2, type=int)
search = ExampleSearch()[(page - 1) * size:page * size]
if 'q' in request.values:
search = search.query(QueryString(query=request.values.get('q')))
search = search.sort(
request.values.get('sort', 'title')
)
search = ExampleSearch.faceted_search(search=search)
results = search.execute().to_dict()
return jsonify({'hits': results.get('hits')})
|
python
|
def index():
"""Query Elasticsearch using Invenio query syntax."""
page = request.values.get('page', 1, type=int)
size = request.values.get('size', 2, type=int)
search = ExampleSearch()[(page - 1) * size:page * size]
if 'q' in request.values:
search = search.query(QueryString(query=request.values.get('q')))
search = search.sort(
request.values.get('sort', 'title')
)
search = ExampleSearch.faceted_search(search=search)
results = search.execute().to_dict()
return jsonify({'hits': results.get('hits')})
|
[
"def",
"index",
"(",
")",
":",
"page",
"=",
"request",
".",
"values",
".",
"get",
"(",
"'page'",
",",
"1",
",",
"type",
"=",
"int",
")",
"size",
"=",
"request",
".",
"values",
".",
"get",
"(",
"'size'",
",",
"2",
",",
"type",
"=",
"int",
")",
"search",
"=",
"ExampleSearch",
"(",
")",
"[",
"(",
"page",
"-",
"1",
")",
"*",
"size",
":",
"page",
"*",
"size",
"]",
"if",
"'q'",
"in",
"request",
".",
"values",
":",
"search",
"=",
"search",
".",
"query",
"(",
"QueryString",
"(",
"query",
"=",
"request",
".",
"values",
".",
"get",
"(",
"'q'",
")",
")",
")",
"search",
"=",
"search",
".",
"sort",
"(",
"request",
".",
"values",
".",
"get",
"(",
"'sort'",
",",
"'title'",
")",
")",
"search",
"=",
"ExampleSearch",
".",
"faceted_search",
"(",
"search",
"=",
"search",
")",
"results",
"=",
"search",
".",
"execute",
"(",
")",
".",
"to_dict",
"(",
")",
"return",
"jsonify",
"(",
"{",
"'hits'",
":",
"results",
".",
"get",
"(",
"'hits'",
")",
"}",
")"
] |
Query Elasticsearch using Invenio query syntax.
|
[
"Query",
"Elasticsearch",
"using",
"Invenio",
"query",
"syntax",
"."
] |
19c073d608d4c811f1c5aecb6622402d39715228
|
https://github.com/inveniosoftware/invenio-search/blob/19c073d608d4c811f1c5aecb6622402d39715228/examples/app.py#L106-L119
|
train
|
krbcontext/python-krbcontext
|
krbcontext/context.py
|
krbContext.clean_options
|
def clean_options(self,
using_keytab=False, principal=None,
keytab_file=None, ccache_file=None,
password=None):
"""Clean argument to related object
:param bool using_keytab: refer to ``krbContext.__init__``.
:param str principal: refer to ``krbContext.__init__``.
:param str keytab_file: refer to ``krbContext.__init__``.
:param str ccache_file: refer to ``krbContext.__init__``.
:param str password: refer to ``krbContext.__init__``.
:return: a mapping containing cleaned names and values, which are used
internally.
:rtype: dict
:raises ValueError: principal is missing or given keytab file does not
exist, when initialize from a keytab.
"""
cleaned = {}
if using_keytab:
if principal is None:
raise ValueError('Principal is required when using key table.')
princ_name = gssapi.names.Name(
principal, gssapi.names.NameType.kerberos_principal)
if keytab_file is None:
cleaned['keytab'] = DEFAULT_KEYTAB
elif not os.path.exists(keytab_file):
raise ValueError(
'Keytab file {0} does not exist.'.format(keytab_file))
else:
cleaned['keytab'] = keytab_file
else:
if principal is None:
principal = get_login()
princ_name = gssapi.names.Name(principal,
gssapi.names.NameType.user)
cleaned['using_keytab'] = using_keytab
cleaned['principal'] = princ_name
cleaned['ccache'] = ccache_file or DEFAULT_CCACHE
cleaned['password'] = password
return cleaned
|
python
|
def clean_options(self,
using_keytab=False, principal=None,
keytab_file=None, ccache_file=None,
password=None):
"""Clean argument to related object
:param bool using_keytab: refer to ``krbContext.__init__``.
:param str principal: refer to ``krbContext.__init__``.
:param str keytab_file: refer to ``krbContext.__init__``.
:param str ccache_file: refer to ``krbContext.__init__``.
:param str password: refer to ``krbContext.__init__``.
:return: a mapping containing cleaned names and values, which are used
internally.
:rtype: dict
:raises ValueError: principal is missing or given keytab file does not
exist, when initialize from a keytab.
"""
cleaned = {}
if using_keytab:
if principal is None:
raise ValueError('Principal is required when using key table.')
princ_name = gssapi.names.Name(
principal, gssapi.names.NameType.kerberos_principal)
if keytab_file is None:
cleaned['keytab'] = DEFAULT_KEYTAB
elif not os.path.exists(keytab_file):
raise ValueError(
'Keytab file {0} does not exist.'.format(keytab_file))
else:
cleaned['keytab'] = keytab_file
else:
if principal is None:
principal = get_login()
princ_name = gssapi.names.Name(principal,
gssapi.names.NameType.user)
cleaned['using_keytab'] = using_keytab
cleaned['principal'] = princ_name
cleaned['ccache'] = ccache_file or DEFAULT_CCACHE
cleaned['password'] = password
return cleaned
|
[
"def",
"clean_options",
"(",
"self",
",",
"using_keytab",
"=",
"False",
",",
"principal",
"=",
"None",
",",
"keytab_file",
"=",
"None",
",",
"ccache_file",
"=",
"None",
",",
"password",
"=",
"None",
")",
":",
"cleaned",
"=",
"{",
"}",
"if",
"using_keytab",
":",
"if",
"principal",
"is",
"None",
":",
"raise",
"ValueError",
"(",
"'Principal is required when using key table.'",
")",
"princ_name",
"=",
"gssapi",
".",
"names",
".",
"Name",
"(",
"principal",
",",
"gssapi",
".",
"names",
".",
"NameType",
".",
"kerberos_principal",
")",
"if",
"keytab_file",
"is",
"None",
":",
"cleaned",
"[",
"'keytab'",
"]",
"=",
"DEFAULT_KEYTAB",
"elif",
"not",
"os",
".",
"path",
".",
"exists",
"(",
"keytab_file",
")",
":",
"raise",
"ValueError",
"(",
"'Keytab file {0} does not exist.'",
".",
"format",
"(",
"keytab_file",
")",
")",
"else",
":",
"cleaned",
"[",
"'keytab'",
"]",
"=",
"keytab_file",
"else",
":",
"if",
"principal",
"is",
"None",
":",
"principal",
"=",
"get_login",
"(",
")",
"princ_name",
"=",
"gssapi",
".",
"names",
".",
"Name",
"(",
"principal",
",",
"gssapi",
".",
"names",
".",
"NameType",
".",
"user",
")",
"cleaned",
"[",
"'using_keytab'",
"]",
"=",
"using_keytab",
"cleaned",
"[",
"'principal'",
"]",
"=",
"princ_name",
"cleaned",
"[",
"'ccache'",
"]",
"=",
"ccache_file",
"or",
"DEFAULT_CCACHE",
"cleaned",
"[",
"'password'",
"]",
"=",
"password",
"return",
"cleaned"
] |
Clean argument to related object
:param bool using_keytab: refer to ``krbContext.__init__``.
:param str principal: refer to ``krbContext.__init__``.
:param str keytab_file: refer to ``krbContext.__init__``.
:param str ccache_file: refer to ``krbContext.__init__``.
:param str password: refer to ``krbContext.__init__``.
:return: a mapping containing cleaned names and values, which are used
internally.
:rtype: dict
:raises ValueError: principal is missing or given keytab file does not
exist, when initialize from a keytab.
|
[
"Clean",
"argument",
"to",
"related",
"object"
] |
9dc93765a4723fa04c2ca57125294d422a6356cd
|
https://github.com/krbcontext/python-krbcontext/blob/9dc93765a4723fa04c2ca57125294d422a6356cd/krbcontext/context.py#L104-L148
|
train
|
krbcontext/python-krbcontext
|
krbcontext/context.py
|
krbContext.init_with_keytab
|
def init_with_keytab(self):
"""Initialize credential cache with keytab"""
creds_opts = {
'usage': 'initiate',
'name': self._cleaned_options['principal'],
}
store = {}
if self._cleaned_options['keytab'] != DEFAULT_KEYTAB:
store['client_keytab'] = self._cleaned_options['keytab']
if self._cleaned_options['ccache'] != DEFAULT_CCACHE:
store['ccache'] = self._cleaned_options['ccache']
if store:
creds_opts['store'] = store
creds = gssapi.creds.Credentials(**creds_opts)
try:
creds.lifetime
except gssapi.exceptions.ExpiredCredentialsError:
new_creds_opts = copy.deepcopy(creds_opts)
# Get new credential and put it into a temporary ccache
if 'store' in new_creds_opts:
new_creds_opts['store']['ccache'] = _get_temp_ccache()
else:
new_creds_opts['store'] = {'ccache': _get_temp_ccache()}
creds = gssapi.creds.Credentials(**new_creds_opts)
# Then, store new credential back to original specified ccache,
# whatever a given ccache file or the default one.
_store = None
# If default cccache is used, no need to specify ccache in store
# parameter passed to ``creds.store``.
if self._cleaned_options['ccache'] != DEFAULT_CCACHE:
_store = {'ccache': store['ccache']}
creds.store(usage='initiate', store=_store, overwrite=True)
|
python
|
def init_with_keytab(self):
"""Initialize credential cache with keytab"""
creds_opts = {
'usage': 'initiate',
'name': self._cleaned_options['principal'],
}
store = {}
if self._cleaned_options['keytab'] != DEFAULT_KEYTAB:
store['client_keytab'] = self._cleaned_options['keytab']
if self._cleaned_options['ccache'] != DEFAULT_CCACHE:
store['ccache'] = self._cleaned_options['ccache']
if store:
creds_opts['store'] = store
creds = gssapi.creds.Credentials(**creds_opts)
try:
creds.lifetime
except gssapi.exceptions.ExpiredCredentialsError:
new_creds_opts = copy.deepcopy(creds_opts)
# Get new credential and put it into a temporary ccache
if 'store' in new_creds_opts:
new_creds_opts['store']['ccache'] = _get_temp_ccache()
else:
new_creds_opts['store'] = {'ccache': _get_temp_ccache()}
creds = gssapi.creds.Credentials(**new_creds_opts)
# Then, store new credential back to original specified ccache,
# whatever a given ccache file or the default one.
_store = None
# If default cccache is used, no need to specify ccache in store
# parameter passed to ``creds.store``.
if self._cleaned_options['ccache'] != DEFAULT_CCACHE:
_store = {'ccache': store['ccache']}
creds.store(usage='initiate', store=_store, overwrite=True)
|
[
"def",
"init_with_keytab",
"(",
"self",
")",
":",
"creds_opts",
"=",
"{",
"'usage'",
":",
"'initiate'",
",",
"'name'",
":",
"self",
".",
"_cleaned_options",
"[",
"'principal'",
"]",
",",
"}",
"store",
"=",
"{",
"}",
"if",
"self",
".",
"_cleaned_options",
"[",
"'keytab'",
"]",
"!=",
"DEFAULT_KEYTAB",
":",
"store",
"[",
"'client_keytab'",
"]",
"=",
"self",
".",
"_cleaned_options",
"[",
"'keytab'",
"]",
"if",
"self",
".",
"_cleaned_options",
"[",
"'ccache'",
"]",
"!=",
"DEFAULT_CCACHE",
":",
"store",
"[",
"'ccache'",
"]",
"=",
"self",
".",
"_cleaned_options",
"[",
"'ccache'",
"]",
"if",
"store",
":",
"creds_opts",
"[",
"'store'",
"]",
"=",
"store",
"creds",
"=",
"gssapi",
".",
"creds",
".",
"Credentials",
"(",
"*",
"*",
"creds_opts",
")",
"try",
":",
"creds",
".",
"lifetime",
"except",
"gssapi",
".",
"exceptions",
".",
"ExpiredCredentialsError",
":",
"new_creds_opts",
"=",
"copy",
".",
"deepcopy",
"(",
"creds_opts",
")",
"# Get new credential and put it into a temporary ccache",
"if",
"'store'",
"in",
"new_creds_opts",
":",
"new_creds_opts",
"[",
"'store'",
"]",
"[",
"'ccache'",
"]",
"=",
"_get_temp_ccache",
"(",
")",
"else",
":",
"new_creds_opts",
"[",
"'store'",
"]",
"=",
"{",
"'ccache'",
":",
"_get_temp_ccache",
"(",
")",
"}",
"creds",
"=",
"gssapi",
".",
"creds",
".",
"Credentials",
"(",
"*",
"*",
"new_creds_opts",
")",
"# Then, store new credential back to original specified ccache,",
"# whatever a given ccache file or the default one.",
"_store",
"=",
"None",
"# If default cccache is used, no need to specify ccache in store",
"# parameter passed to ``creds.store``.",
"if",
"self",
".",
"_cleaned_options",
"[",
"'ccache'",
"]",
"!=",
"DEFAULT_CCACHE",
":",
"_store",
"=",
"{",
"'ccache'",
":",
"store",
"[",
"'ccache'",
"]",
"}",
"creds",
".",
"store",
"(",
"usage",
"=",
"'initiate'",
",",
"store",
"=",
"_store",
",",
"overwrite",
"=",
"True",
")"
] |
Initialize credential cache with keytab
|
[
"Initialize",
"credential",
"cache",
"with",
"keytab"
] |
9dc93765a4723fa04c2ca57125294d422a6356cd
|
https://github.com/krbcontext/python-krbcontext/blob/9dc93765a4723fa04c2ca57125294d422a6356cd/krbcontext/context.py#L150-L183
|
train
|
krbcontext/python-krbcontext
|
krbcontext/context.py
|
krbContext.init_with_password
|
def init_with_password(self):
"""Initialize credential cache with password
**Causion:** once you enter password from command line, or pass it to
API directly, the given password is not encrypted always. Although
getting credential with password works, from security point of view, it
is strongly recommended **NOT** use it in any formal production
environment. If you need to initialize credential in an application to
application Kerberos authentication context, keytab has to be used.
:raises IOError: when trying to prompt to input password from command
line but no attry is available.
"""
creds_opts = {
'usage': 'initiate',
'name': self._cleaned_options['principal'],
}
if self._cleaned_options['ccache'] != DEFAULT_CCACHE:
creds_opts['store'] = {'ccache': self._cleaned_options['ccache']}
cred = gssapi.creds.Credentials(**creds_opts)
try:
cred.lifetime
except gssapi.exceptions.ExpiredCredentialsError:
password = self._cleaned_options['password']
if not password:
if not sys.stdin.isatty():
raise IOError(
'krbContext is not running from a terminal. So, you '
'need to run kinit with your principal manually before'
' anything goes.')
# If there is no password specified via API call, prompt to
# enter one in order to continue to get credential. BUT, in
# some cases, blocking program and waiting for input of
# password is really bad, which may be only suitable for some
# simple use cases, for example, writing some scripts to test
# something that need Kerberos authentication. Anyway, whether
# it is really to enter a password from command line, it
# depends on concrete use cases totally.
password = getpass.getpass()
cred = gssapi.raw.acquire_cred_with_password(
self._cleaned_options['principal'], password)
ccache = self._cleaned_options['ccache']
if ccache == DEFAULT_CCACHE:
gssapi.raw.store_cred(cred.creds,
usage='initiate',
overwrite=True)
else:
gssapi.raw.store_cred_into({'ccache': ccache},
cred.creds,
usage='initiate',
overwrite=True)
|
python
|
def init_with_password(self):
"""Initialize credential cache with password
**Causion:** once you enter password from command line, or pass it to
API directly, the given password is not encrypted always. Although
getting credential with password works, from security point of view, it
is strongly recommended **NOT** use it in any formal production
environment. If you need to initialize credential in an application to
application Kerberos authentication context, keytab has to be used.
:raises IOError: when trying to prompt to input password from command
line but no attry is available.
"""
creds_opts = {
'usage': 'initiate',
'name': self._cleaned_options['principal'],
}
if self._cleaned_options['ccache'] != DEFAULT_CCACHE:
creds_opts['store'] = {'ccache': self._cleaned_options['ccache']}
cred = gssapi.creds.Credentials(**creds_opts)
try:
cred.lifetime
except gssapi.exceptions.ExpiredCredentialsError:
password = self._cleaned_options['password']
if not password:
if not sys.stdin.isatty():
raise IOError(
'krbContext is not running from a terminal. So, you '
'need to run kinit with your principal manually before'
' anything goes.')
# If there is no password specified via API call, prompt to
# enter one in order to continue to get credential. BUT, in
# some cases, blocking program and waiting for input of
# password is really bad, which may be only suitable for some
# simple use cases, for example, writing some scripts to test
# something that need Kerberos authentication. Anyway, whether
# it is really to enter a password from command line, it
# depends on concrete use cases totally.
password = getpass.getpass()
cred = gssapi.raw.acquire_cred_with_password(
self._cleaned_options['principal'], password)
ccache = self._cleaned_options['ccache']
if ccache == DEFAULT_CCACHE:
gssapi.raw.store_cred(cred.creds,
usage='initiate',
overwrite=True)
else:
gssapi.raw.store_cred_into({'ccache': ccache},
cred.creds,
usage='initiate',
overwrite=True)
|
[
"def",
"init_with_password",
"(",
"self",
")",
":",
"creds_opts",
"=",
"{",
"'usage'",
":",
"'initiate'",
",",
"'name'",
":",
"self",
".",
"_cleaned_options",
"[",
"'principal'",
"]",
",",
"}",
"if",
"self",
".",
"_cleaned_options",
"[",
"'ccache'",
"]",
"!=",
"DEFAULT_CCACHE",
":",
"creds_opts",
"[",
"'store'",
"]",
"=",
"{",
"'ccache'",
":",
"self",
".",
"_cleaned_options",
"[",
"'ccache'",
"]",
"}",
"cred",
"=",
"gssapi",
".",
"creds",
".",
"Credentials",
"(",
"*",
"*",
"creds_opts",
")",
"try",
":",
"cred",
".",
"lifetime",
"except",
"gssapi",
".",
"exceptions",
".",
"ExpiredCredentialsError",
":",
"password",
"=",
"self",
".",
"_cleaned_options",
"[",
"'password'",
"]",
"if",
"not",
"password",
":",
"if",
"not",
"sys",
".",
"stdin",
".",
"isatty",
"(",
")",
":",
"raise",
"IOError",
"(",
"'krbContext is not running from a terminal. So, you '",
"'need to run kinit with your principal manually before'",
"' anything goes.'",
")",
"# If there is no password specified via API call, prompt to",
"# enter one in order to continue to get credential. BUT, in",
"# some cases, blocking program and waiting for input of",
"# password is really bad, which may be only suitable for some",
"# simple use cases, for example, writing some scripts to test",
"# something that need Kerberos authentication. Anyway, whether",
"# it is really to enter a password from command line, it",
"# depends on concrete use cases totally.",
"password",
"=",
"getpass",
".",
"getpass",
"(",
")",
"cred",
"=",
"gssapi",
".",
"raw",
".",
"acquire_cred_with_password",
"(",
"self",
".",
"_cleaned_options",
"[",
"'principal'",
"]",
",",
"password",
")",
"ccache",
"=",
"self",
".",
"_cleaned_options",
"[",
"'ccache'",
"]",
"if",
"ccache",
"==",
"DEFAULT_CCACHE",
":",
"gssapi",
".",
"raw",
".",
"store_cred",
"(",
"cred",
".",
"creds",
",",
"usage",
"=",
"'initiate'",
",",
"overwrite",
"=",
"True",
")",
"else",
":",
"gssapi",
".",
"raw",
".",
"store_cred_into",
"(",
"{",
"'ccache'",
":",
"ccache",
"}",
",",
"cred",
".",
"creds",
",",
"usage",
"=",
"'initiate'",
",",
"overwrite",
"=",
"True",
")"
] |
Initialize credential cache with password
**Causion:** once you enter password from command line, or pass it to
API directly, the given password is not encrypted always. Although
getting credential with password works, from security point of view, it
is strongly recommended **NOT** use it in any formal production
environment. If you need to initialize credential in an application to
application Kerberos authentication context, keytab has to be used.
:raises IOError: when trying to prompt to input password from command
line but no attry is available.
|
[
"Initialize",
"credential",
"cache",
"with",
"password"
] |
9dc93765a4723fa04c2ca57125294d422a6356cd
|
https://github.com/krbcontext/python-krbcontext/blob/9dc93765a4723fa04c2ca57125294d422a6356cd/krbcontext/context.py#L185-L240
|
train
|
krbcontext/python-krbcontext
|
krbcontext/context.py
|
krbContext._prepare_context
|
def _prepare_context(self):
"""Prepare context
Initialize credential cache with keytab or password according to
``using_keytab`` parameter. Then, ``KRB5CCNAME`` is set properly so
that Kerberos library called in current context is able to get
credential from correct cache.
Internal use only.
"""
ccache = self._cleaned_options['ccache']
# Whatever there is KRB5CCNAME was set in current process,
# original_krb5ccname will contain current value even if None if
# that variable wasn't set, and when leave krbcontext, it can be
# handled properly.
self._original_krb5ccname = os.environ.get(ENV_KRB5CCNAME)
if ccache == DEFAULT_CCACHE:
# If user wants to use default ccache, existing KRB5CCNAME in
# current environment variable should be removed.
if self._original_krb5ccname:
del os.environ[ENV_KRB5CCNAME]
else:
# When not using default ccache to initialize new credential, let
# us point to the given ccache by KRB5CCNAME.
os.environ[ENV_KRB5CCNAME] = ccache
if self._cleaned_options['using_keytab']:
self.init_with_keytab()
else:
self.init_with_password()
|
python
|
def _prepare_context(self):
"""Prepare context
Initialize credential cache with keytab or password according to
``using_keytab`` parameter. Then, ``KRB5CCNAME`` is set properly so
that Kerberos library called in current context is able to get
credential from correct cache.
Internal use only.
"""
ccache = self._cleaned_options['ccache']
# Whatever there is KRB5CCNAME was set in current process,
# original_krb5ccname will contain current value even if None if
# that variable wasn't set, and when leave krbcontext, it can be
# handled properly.
self._original_krb5ccname = os.environ.get(ENV_KRB5CCNAME)
if ccache == DEFAULT_CCACHE:
# If user wants to use default ccache, existing KRB5CCNAME in
# current environment variable should be removed.
if self._original_krb5ccname:
del os.environ[ENV_KRB5CCNAME]
else:
# When not using default ccache to initialize new credential, let
# us point to the given ccache by KRB5CCNAME.
os.environ[ENV_KRB5CCNAME] = ccache
if self._cleaned_options['using_keytab']:
self.init_with_keytab()
else:
self.init_with_password()
|
[
"def",
"_prepare_context",
"(",
"self",
")",
":",
"ccache",
"=",
"self",
".",
"_cleaned_options",
"[",
"'ccache'",
"]",
"# Whatever there is KRB5CCNAME was set in current process,",
"# original_krb5ccname will contain current value even if None if",
"# that variable wasn't set, and when leave krbcontext, it can be",
"# handled properly.",
"self",
".",
"_original_krb5ccname",
"=",
"os",
".",
"environ",
".",
"get",
"(",
"ENV_KRB5CCNAME",
")",
"if",
"ccache",
"==",
"DEFAULT_CCACHE",
":",
"# If user wants to use default ccache, existing KRB5CCNAME in",
"# current environment variable should be removed.",
"if",
"self",
".",
"_original_krb5ccname",
":",
"del",
"os",
".",
"environ",
"[",
"ENV_KRB5CCNAME",
"]",
"else",
":",
"# When not using default ccache to initialize new credential, let",
"# us point to the given ccache by KRB5CCNAME.",
"os",
".",
"environ",
"[",
"ENV_KRB5CCNAME",
"]",
"=",
"ccache",
"if",
"self",
".",
"_cleaned_options",
"[",
"'using_keytab'",
"]",
":",
"self",
".",
"init_with_keytab",
"(",
")",
"else",
":",
"self",
".",
"init_with_password",
"(",
")"
] |
Prepare context
Initialize credential cache with keytab or password according to
``using_keytab`` parameter. Then, ``KRB5CCNAME`` is set properly so
that Kerberos library called in current context is able to get
credential from correct cache.
Internal use only.
|
[
"Prepare",
"context"
] |
9dc93765a4723fa04c2ca57125294d422a6356cd
|
https://github.com/krbcontext/python-krbcontext/blob/9dc93765a4723fa04c2ca57125294d422a6356cd/krbcontext/context.py#L242-L273
|
train
|
inveniosoftware/invenio-search
|
invenio_search/ext.py
|
_SearchState.templates
|
def templates(self):
"""Generate a dictionary with template names and file paths."""
templates = {}
result = []
if self.entry_point_group_templates:
result = self.load_entry_point_group_templates(
self.entry_point_group_templates) or []
for template in result:
for name, path in template.items():
templates[name] = path
return templates
|
python
|
def templates(self):
"""Generate a dictionary with template names and file paths."""
templates = {}
result = []
if self.entry_point_group_templates:
result = self.load_entry_point_group_templates(
self.entry_point_group_templates) or []
for template in result:
for name, path in template.items():
templates[name] = path
return templates
|
[
"def",
"templates",
"(",
"self",
")",
":",
"templates",
"=",
"{",
"}",
"result",
"=",
"[",
"]",
"if",
"self",
".",
"entry_point_group_templates",
":",
"result",
"=",
"self",
".",
"load_entry_point_group_templates",
"(",
"self",
".",
"entry_point_group_templates",
")",
"or",
"[",
"]",
"for",
"template",
"in",
"result",
":",
"for",
"name",
",",
"path",
"in",
"template",
".",
"items",
"(",
")",
":",
"templates",
"[",
"name",
"]",
"=",
"path",
"return",
"templates"
] |
Generate a dictionary with template names and file paths.
|
[
"Generate",
"a",
"dictionary",
"with",
"template",
"names",
"and",
"file",
"paths",
"."
] |
19c073d608d4c811f1c5aecb6622402d39715228
|
https://github.com/inveniosoftware/invenio-search/blob/19c073d608d4c811f1c5aecb6622402d39715228/invenio_search/ext.py#L71-L83
|
train
|
inveniosoftware/invenio-search
|
invenio_search/ext.py
|
_SearchState.register_mappings
|
def register_mappings(self, alias, package_name):
"""Register mappings from a package under given alias.
:param alias: The alias.
:param package_name: The package name.
"""
# For backwards compatibility, we also allow for ES2 mappings to be
# placed at the root level of the specified package path, and not in
# the `<package-path>/v2` directory.
if ES_VERSION[0] == 2:
try:
resource_listdir(package_name, 'v2')
package_name += '.v2'
except (OSError, IOError) as ex:
if getattr(ex, 'errno', 0) != errno.ENOENT:
raise
warnings.warn(
"Having mappings in a path which doesn't specify the "
"Elasticsearch version is deprecated. Please move your "
"mappings to a subfolder named according to the "
"Elasticsearch version which your mappings are intended "
"for. (e.g. '{}/v2/{}')".format(
package_name, alias),
PendingDeprecationWarning)
else:
package_name = '{}.v{}'.format(package_name, ES_VERSION[0])
def _walk_dir(aliases, *parts):
root_name = build_index_name(self.app, *parts)
resource_name = os.path.join(*parts)
if root_name not in aliases:
self.number_of_indexes += 1
data = aliases.get(root_name, {})
for filename in resource_listdir(package_name, resource_name):
index_name = build_index_name(
self.app,
*(parts + (filename, ))
)
file_path = os.path.join(resource_name, filename)
if resource_isdir(package_name, file_path):
_walk_dir(data, *(parts + (filename, )))
continue
ext = os.path.splitext(filename)[1]
if ext not in {'.json', }:
continue
assert index_name not in data, 'Duplicate index'
data[index_name] = self.mappings[index_name] = \
resource_filename(
package_name, os.path.join(resource_name, filename))
self.number_of_indexes += 1
aliases[root_name] = data
# Start the recursion here:
_walk_dir(self.aliases, alias)
|
python
|
def register_mappings(self, alias, package_name):
"""Register mappings from a package under given alias.
:param alias: The alias.
:param package_name: The package name.
"""
# For backwards compatibility, we also allow for ES2 mappings to be
# placed at the root level of the specified package path, and not in
# the `<package-path>/v2` directory.
if ES_VERSION[0] == 2:
try:
resource_listdir(package_name, 'v2')
package_name += '.v2'
except (OSError, IOError) as ex:
if getattr(ex, 'errno', 0) != errno.ENOENT:
raise
warnings.warn(
"Having mappings in a path which doesn't specify the "
"Elasticsearch version is deprecated. Please move your "
"mappings to a subfolder named according to the "
"Elasticsearch version which your mappings are intended "
"for. (e.g. '{}/v2/{}')".format(
package_name, alias),
PendingDeprecationWarning)
else:
package_name = '{}.v{}'.format(package_name, ES_VERSION[0])
def _walk_dir(aliases, *parts):
root_name = build_index_name(self.app, *parts)
resource_name = os.path.join(*parts)
if root_name not in aliases:
self.number_of_indexes += 1
data = aliases.get(root_name, {})
for filename in resource_listdir(package_name, resource_name):
index_name = build_index_name(
self.app,
*(parts + (filename, ))
)
file_path = os.path.join(resource_name, filename)
if resource_isdir(package_name, file_path):
_walk_dir(data, *(parts + (filename, )))
continue
ext = os.path.splitext(filename)[1]
if ext not in {'.json', }:
continue
assert index_name not in data, 'Duplicate index'
data[index_name] = self.mappings[index_name] = \
resource_filename(
package_name, os.path.join(resource_name, filename))
self.number_of_indexes += 1
aliases[root_name] = data
# Start the recursion here:
_walk_dir(self.aliases, alias)
|
[
"def",
"register_mappings",
"(",
"self",
",",
"alias",
",",
"package_name",
")",
":",
"# For backwards compatibility, we also allow for ES2 mappings to be",
"# placed at the root level of the specified package path, and not in",
"# the `<package-path>/v2` directory.",
"if",
"ES_VERSION",
"[",
"0",
"]",
"==",
"2",
":",
"try",
":",
"resource_listdir",
"(",
"package_name",
",",
"'v2'",
")",
"package_name",
"+=",
"'.v2'",
"except",
"(",
"OSError",
",",
"IOError",
")",
"as",
"ex",
":",
"if",
"getattr",
"(",
"ex",
",",
"'errno'",
",",
"0",
")",
"!=",
"errno",
".",
"ENOENT",
":",
"raise",
"warnings",
".",
"warn",
"(",
"\"Having mappings in a path which doesn't specify the \"",
"\"Elasticsearch version is deprecated. Please move your \"",
"\"mappings to a subfolder named according to the \"",
"\"Elasticsearch version which your mappings are intended \"",
"\"for. (e.g. '{}/v2/{}')\"",
".",
"format",
"(",
"package_name",
",",
"alias",
")",
",",
"PendingDeprecationWarning",
")",
"else",
":",
"package_name",
"=",
"'{}.v{}'",
".",
"format",
"(",
"package_name",
",",
"ES_VERSION",
"[",
"0",
"]",
")",
"def",
"_walk_dir",
"(",
"aliases",
",",
"*",
"parts",
")",
":",
"root_name",
"=",
"build_index_name",
"(",
"self",
".",
"app",
",",
"*",
"parts",
")",
"resource_name",
"=",
"os",
".",
"path",
".",
"join",
"(",
"*",
"parts",
")",
"if",
"root_name",
"not",
"in",
"aliases",
":",
"self",
".",
"number_of_indexes",
"+=",
"1",
"data",
"=",
"aliases",
".",
"get",
"(",
"root_name",
",",
"{",
"}",
")",
"for",
"filename",
"in",
"resource_listdir",
"(",
"package_name",
",",
"resource_name",
")",
":",
"index_name",
"=",
"build_index_name",
"(",
"self",
".",
"app",
",",
"*",
"(",
"parts",
"+",
"(",
"filename",
",",
")",
")",
")",
"file_path",
"=",
"os",
".",
"path",
".",
"join",
"(",
"resource_name",
",",
"filename",
")",
"if",
"resource_isdir",
"(",
"package_name",
",",
"file_path",
")",
":",
"_walk_dir",
"(",
"data",
",",
"*",
"(",
"parts",
"+",
"(",
"filename",
",",
")",
")",
")",
"continue",
"ext",
"=",
"os",
".",
"path",
".",
"splitext",
"(",
"filename",
")",
"[",
"1",
"]",
"if",
"ext",
"not",
"in",
"{",
"'.json'",
",",
"}",
":",
"continue",
"assert",
"index_name",
"not",
"in",
"data",
",",
"'Duplicate index'",
"data",
"[",
"index_name",
"]",
"=",
"self",
".",
"mappings",
"[",
"index_name",
"]",
"=",
"resource_filename",
"(",
"package_name",
",",
"os",
".",
"path",
".",
"join",
"(",
"resource_name",
",",
"filename",
")",
")",
"self",
".",
"number_of_indexes",
"+=",
"1",
"aliases",
"[",
"root_name",
"]",
"=",
"data",
"# Start the recursion here:",
"_walk_dir",
"(",
"self",
".",
"aliases",
",",
"alias",
")"
] |
Register mappings from a package under given alias.
:param alias: The alias.
:param package_name: The package name.
|
[
"Register",
"mappings",
"from",
"a",
"package",
"under",
"given",
"alias",
"."
] |
19c073d608d4c811f1c5aecb6622402d39715228
|
https://github.com/inveniosoftware/invenio-search/blob/19c073d608d4c811f1c5aecb6622402d39715228/invenio_search/ext.py#L85-L145
|
train
|
inveniosoftware/invenio-search
|
invenio_search/ext.py
|
_SearchState.register_templates
|
def register_templates(self, directory):
"""Register templates from the provided directory.
:param directory: The templates directory.
"""
try:
resource_listdir(directory, 'v{}'.format(ES_VERSION[0]))
directory = '{}/v{}'.format(directory, ES_VERSION[0])
except (OSError, IOError) as ex:
if getattr(ex, 'errno', 0) == errno.ENOENT:
raise OSError(
"Please move your templates to a subfolder named "
"according to the Elasticsearch version "
"which your templates are intended "
"for. (e.g. '{}.v{}')".format(directory,
ES_VERSION[0]))
result = {}
module_name, parts = directory.split('.')[0], directory.split('.')[1:]
parts = tuple(parts)
def _walk_dir(parts):
resource_name = os.path.join(*parts)
for filename in resource_listdir(module_name, resource_name):
template_name = build_index_name(
self.app,
*(parts[1:] + (filename, ))
)
file_path = os.path.join(resource_name, filename)
if resource_isdir(module_name, file_path):
_walk_dir((parts + (filename, )))
continue
ext = os.path.splitext(filename)[1]
if ext not in {'.json', }:
continue
result[template_name] = resource_filename(
module_name, os.path.join(resource_name, filename))
# Start the recursion here:
_walk_dir(parts)
return result
|
python
|
def register_templates(self, directory):
"""Register templates from the provided directory.
:param directory: The templates directory.
"""
try:
resource_listdir(directory, 'v{}'.format(ES_VERSION[0]))
directory = '{}/v{}'.format(directory, ES_VERSION[0])
except (OSError, IOError) as ex:
if getattr(ex, 'errno', 0) == errno.ENOENT:
raise OSError(
"Please move your templates to a subfolder named "
"according to the Elasticsearch version "
"which your templates are intended "
"for. (e.g. '{}.v{}')".format(directory,
ES_VERSION[0]))
result = {}
module_name, parts = directory.split('.')[0], directory.split('.')[1:]
parts = tuple(parts)
def _walk_dir(parts):
resource_name = os.path.join(*parts)
for filename in resource_listdir(module_name, resource_name):
template_name = build_index_name(
self.app,
*(parts[1:] + (filename, ))
)
file_path = os.path.join(resource_name, filename)
if resource_isdir(module_name, file_path):
_walk_dir((parts + (filename, )))
continue
ext = os.path.splitext(filename)[1]
if ext not in {'.json', }:
continue
result[template_name] = resource_filename(
module_name, os.path.join(resource_name, filename))
# Start the recursion here:
_walk_dir(parts)
return result
|
[
"def",
"register_templates",
"(",
"self",
",",
"directory",
")",
":",
"try",
":",
"resource_listdir",
"(",
"directory",
",",
"'v{}'",
".",
"format",
"(",
"ES_VERSION",
"[",
"0",
"]",
")",
")",
"directory",
"=",
"'{}/v{}'",
".",
"format",
"(",
"directory",
",",
"ES_VERSION",
"[",
"0",
"]",
")",
"except",
"(",
"OSError",
",",
"IOError",
")",
"as",
"ex",
":",
"if",
"getattr",
"(",
"ex",
",",
"'errno'",
",",
"0",
")",
"==",
"errno",
".",
"ENOENT",
":",
"raise",
"OSError",
"(",
"\"Please move your templates to a subfolder named \"",
"\"according to the Elasticsearch version \"",
"\"which your templates are intended \"",
"\"for. (e.g. '{}.v{}')\"",
".",
"format",
"(",
"directory",
",",
"ES_VERSION",
"[",
"0",
"]",
")",
")",
"result",
"=",
"{",
"}",
"module_name",
",",
"parts",
"=",
"directory",
".",
"split",
"(",
"'.'",
")",
"[",
"0",
"]",
",",
"directory",
".",
"split",
"(",
"'.'",
")",
"[",
"1",
":",
"]",
"parts",
"=",
"tuple",
"(",
"parts",
")",
"def",
"_walk_dir",
"(",
"parts",
")",
":",
"resource_name",
"=",
"os",
".",
"path",
".",
"join",
"(",
"*",
"parts",
")",
"for",
"filename",
"in",
"resource_listdir",
"(",
"module_name",
",",
"resource_name",
")",
":",
"template_name",
"=",
"build_index_name",
"(",
"self",
".",
"app",
",",
"*",
"(",
"parts",
"[",
"1",
":",
"]",
"+",
"(",
"filename",
",",
")",
")",
")",
"file_path",
"=",
"os",
".",
"path",
".",
"join",
"(",
"resource_name",
",",
"filename",
")",
"if",
"resource_isdir",
"(",
"module_name",
",",
"file_path",
")",
":",
"_walk_dir",
"(",
"(",
"parts",
"+",
"(",
"filename",
",",
")",
")",
")",
"continue",
"ext",
"=",
"os",
".",
"path",
".",
"splitext",
"(",
"filename",
")",
"[",
"1",
"]",
"if",
"ext",
"not",
"in",
"{",
"'.json'",
",",
"}",
":",
"continue",
"result",
"[",
"template_name",
"]",
"=",
"resource_filename",
"(",
"module_name",
",",
"os",
".",
"path",
".",
"join",
"(",
"resource_name",
",",
"filename",
")",
")",
"# Start the recursion here:",
"_walk_dir",
"(",
"parts",
")",
"return",
"result"
] |
Register templates from the provided directory.
:param directory: The templates directory.
|
[
"Register",
"templates",
"from",
"the",
"provided",
"directory",
"."
] |
19c073d608d4c811f1c5aecb6622402d39715228
|
https://github.com/inveniosoftware/invenio-search/blob/19c073d608d4c811f1c5aecb6622402d39715228/invenio_search/ext.py#L147-L190
|
train
|
inveniosoftware/invenio-search
|
invenio_search/ext.py
|
_SearchState.load_entry_point_group_mappings
|
def load_entry_point_group_mappings(self, entry_point_group_mappings):
"""Load actions from an entry point group."""
for ep in iter_entry_points(group=entry_point_group_mappings):
self.register_mappings(ep.name, ep.module_name)
|
python
|
def load_entry_point_group_mappings(self, entry_point_group_mappings):
"""Load actions from an entry point group."""
for ep in iter_entry_points(group=entry_point_group_mappings):
self.register_mappings(ep.name, ep.module_name)
|
[
"def",
"load_entry_point_group_mappings",
"(",
"self",
",",
"entry_point_group_mappings",
")",
":",
"for",
"ep",
"in",
"iter_entry_points",
"(",
"group",
"=",
"entry_point_group_mappings",
")",
":",
"self",
".",
"register_mappings",
"(",
"ep",
".",
"name",
",",
"ep",
".",
"module_name",
")"
] |
Load actions from an entry point group.
|
[
"Load",
"actions",
"from",
"an",
"entry",
"point",
"group",
"."
] |
19c073d608d4c811f1c5aecb6622402d39715228
|
https://github.com/inveniosoftware/invenio-search/blob/19c073d608d4c811f1c5aecb6622402d39715228/invenio_search/ext.py#L192-L195
|
train
|
inveniosoftware/invenio-search
|
invenio_search/ext.py
|
_SearchState.load_entry_point_group_templates
|
def load_entry_point_group_templates(self, entry_point_group_templates):
"""Load actions from an entry point group."""
result = []
for ep in iter_entry_points(group=entry_point_group_templates):
with self.app.app_context():
for template_dir in ep.load()():
result.append(self.register_templates(template_dir))
return result
|
python
|
def load_entry_point_group_templates(self, entry_point_group_templates):
"""Load actions from an entry point group."""
result = []
for ep in iter_entry_points(group=entry_point_group_templates):
with self.app.app_context():
for template_dir in ep.load()():
result.append(self.register_templates(template_dir))
return result
|
[
"def",
"load_entry_point_group_templates",
"(",
"self",
",",
"entry_point_group_templates",
")",
":",
"result",
"=",
"[",
"]",
"for",
"ep",
"in",
"iter_entry_points",
"(",
"group",
"=",
"entry_point_group_templates",
")",
":",
"with",
"self",
".",
"app",
".",
"app_context",
"(",
")",
":",
"for",
"template_dir",
"in",
"ep",
".",
"load",
"(",
")",
"(",
")",
":",
"result",
".",
"append",
"(",
"self",
".",
"register_templates",
"(",
"template_dir",
")",
")",
"return",
"result"
] |
Load actions from an entry point group.
|
[
"Load",
"actions",
"from",
"an",
"entry",
"point",
"group",
"."
] |
19c073d608d4c811f1c5aecb6622402d39715228
|
https://github.com/inveniosoftware/invenio-search/blob/19c073d608d4c811f1c5aecb6622402d39715228/invenio_search/ext.py#L197-L204
|
train
|
inveniosoftware/invenio-search
|
invenio_search/ext.py
|
_SearchState._client_builder
|
def _client_builder(self):
"""Build Elasticsearch client."""
client_config = self.app.config.get('SEARCH_CLIENT_CONFIG') or {}
client_config.setdefault(
'hosts', self.app.config.get('SEARCH_ELASTIC_HOSTS'))
client_config.setdefault('connection_class', RequestsHttpConnection)
return Elasticsearch(**client_config)
|
python
|
def _client_builder(self):
"""Build Elasticsearch client."""
client_config = self.app.config.get('SEARCH_CLIENT_CONFIG') or {}
client_config.setdefault(
'hosts', self.app.config.get('SEARCH_ELASTIC_HOSTS'))
client_config.setdefault('connection_class', RequestsHttpConnection)
return Elasticsearch(**client_config)
|
[
"def",
"_client_builder",
"(",
"self",
")",
":",
"client_config",
"=",
"self",
".",
"app",
".",
"config",
".",
"get",
"(",
"'SEARCH_CLIENT_CONFIG'",
")",
"or",
"{",
"}",
"client_config",
".",
"setdefault",
"(",
"'hosts'",
",",
"self",
".",
"app",
".",
"config",
".",
"get",
"(",
"'SEARCH_ELASTIC_HOSTS'",
")",
")",
"client_config",
".",
"setdefault",
"(",
"'connection_class'",
",",
"RequestsHttpConnection",
")",
"return",
"Elasticsearch",
"(",
"*",
"*",
"client_config",
")"
] |
Build Elasticsearch client.
|
[
"Build",
"Elasticsearch",
"client",
"."
] |
19c073d608d4c811f1c5aecb6622402d39715228
|
https://github.com/inveniosoftware/invenio-search/blob/19c073d608d4c811f1c5aecb6622402d39715228/invenio_search/ext.py#L206-L212
|
train
|
inveniosoftware/invenio-search
|
invenio_search/ext.py
|
_SearchState.client
|
def client(self):
"""Return client for current application."""
if self._client is None:
self._client = self._client_builder()
return self._client
|
python
|
def client(self):
"""Return client for current application."""
if self._client is None:
self._client = self._client_builder()
return self._client
|
[
"def",
"client",
"(",
"self",
")",
":",
"if",
"self",
".",
"_client",
"is",
"None",
":",
"self",
".",
"_client",
"=",
"self",
".",
"_client_builder",
"(",
")",
"return",
"self",
".",
"_client"
] |
Return client for current application.
|
[
"Return",
"client",
"for",
"current",
"application",
"."
] |
19c073d608d4c811f1c5aecb6622402d39715228
|
https://github.com/inveniosoftware/invenio-search/blob/19c073d608d4c811f1c5aecb6622402d39715228/invenio_search/ext.py#L215-L219
|
train
|
inveniosoftware/invenio-search
|
invenio_search/ext.py
|
_SearchState.flush_and_refresh
|
def flush_and_refresh(self, index):
"""Flush and refresh one or more indices.
.. warning::
Do not call this method unless you know what you are doing. This
method is only intended to be called during tests.
"""
self.client.indices.flush(wait_if_ongoing=True, index=index)
self.client.indices.refresh(index=index)
self.client.cluster.health(
wait_for_status='yellow', request_timeout=30)
return True
|
python
|
def flush_and_refresh(self, index):
"""Flush and refresh one or more indices.
.. warning::
Do not call this method unless you know what you are doing. This
method is only intended to be called during tests.
"""
self.client.indices.flush(wait_if_ongoing=True, index=index)
self.client.indices.refresh(index=index)
self.client.cluster.health(
wait_for_status='yellow', request_timeout=30)
return True
|
[
"def",
"flush_and_refresh",
"(",
"self",
",",
"index",
")",
":",
"self",
".",
"client",
".",
"indices",
".",
"flush",
"(",
"wait_if_ongoing",
"=",
"True",
",",
"index",
"=",
"index",
")",
"self",
".",
"client",
".",
"indices",
".",
"refresh",
"(",
"index",
"=",
"index",
")",
"self",
".",
"client",
".",
"cluster",
".",
"health",
"(",
"wait_for_status",
"=",
"'yellow'",
",",
"request_timeout",
"=",
"30",
")",
"return",
"True"
] |
Flush and refresh one or more indices.
.. warning::
Do not call this method unless you know what you are doing. This
method is only intended to be called during tests.
|
[
"Flush",
"and",
"refresh",
"one",
"or",
"more",
"indices",
"."
] |
19c073d608d4c811f1c5aecb6622402d39715228
|
https://github.com/inveniosoftware/invenio-search/blob/19c073d608d4c811f1c5aecb6622402d39715228/invenio_search/ext.py#L221-L233
|
train
|
inveniosoftware/invenio-search
|
invenio_search/ext.py
|
_SearchState.cluster_version
|
def cluster_version(self):
"""Get version of Elasticsearch running on the cluster."""
versionstr = self.client.info()['version']['number']
return [int(x) for x in versionstr.split('.')]
|
python
|
def cluster_version(self):
"""Get version of Elasticsearch running on the cluster."""
versionstr = self.client.info()['version']['number']
return [int(x) for x in versionstr.split('.')]
|
[
"def",
"cluster_version",
"(",
"self",
")",
":",
"versionstr",
"=",
"self",
".",
"client",
".",
"info",
"(",
")",
"[",
"'version'",
"]",
"[",
"'number'",
"]",
"return",
"[",
"int",
"(",
"x",
")",
"for",
"x",
"in",
"versionstr",
".",
"split",
"(",
"'.'",
")",
"]"
] |
Get version of Elasticsearch running on the cluster.
|
[
"Get",
"version",
"of",
"Elasticsearch",
"running",
"on",
"the",
"cluster",
"."
] |
19c073d608d4c811f1c5aecb6622402d39715228
|
https://github.com/inveniosoftware/invenio-search/blob/19c073d608d4c811f1c5aecb6622402d39715228/invenio_search/ext.py#L236-L239
|
train
|
inveniosoftware/invenio-search
|
invenio_search/ext.py
|
_SearchState.active_aliases
|
def active_aliases(self):
"""Get a filtered list of aliases based on configuration.
Returns aliases and their mappings that are defined in the
`SEARCH_MAPPINGS` config variable. If the `SEARCH_MAPPINGS` is set to
`None` (the default), all aliases are included.
"""
whitelisted_aliases = self.app.config.get('SEARCH_MAPPINGS')
if whitelisted_aliases is None:
return self.aliases
else:
return {k: v for k, v in self.aliases.items()
if k in whitelisted_aliases}
|
python
|
def active_aliases(self):
"""Get a filtered list of aliases based on configuration.
Returns aliases and their mappings that are defined in the
`SEARCH_MAPPINGS` config variable. If the `SEARCH_MAPPINGS` is set to
`None` (the default), all aliases are included.
"""
whitelisted_aliases = self.app.config.get('SEARCH_MAPPINGS')
if whitelisted_aliases is None:
return self.aliases
else:
return {k: v for k, v in self.aliases.items()
if k in whitelisted_aliases}
|
[
"def",
"active_aliases",
"(",
"self",
")",
":",
"whitelisted_aliases",
"=",
"self",
".",
"app",
".",
"config",
".",
"get",
"(",
"'SEARCH_MAPPINGS'",
")",
"if",
"whitelisted_aliases",
"is",
"None",
":",
"return",
"self",
".",
"aliases",
"else",
":",
"return",
"{",
"k",
":",
"v",
"for",
"k",
",",
"v",
"in",
"self",
".",
"aliases",
".",
"items",
"(",
")",
"if",
"k",
"in",
"whitelisted_aliases",
"}"
] |
Get a filtered list of aliases based on configuration.
Returns aliases and their mappings that are defined in the
`SEARCH_MAPPINGS` config variable. If the `SEARCH_MAPPINGS` is set to
`None` (the default), all aliases are included.
|
[
"Get",
"a",
"filtered",
"list",
"of",
"aliases",
"based",
"on",
"configuration",
"."
] |
19c073d608d4c811f1c5aecb6622402d39715228
|
https://github.com/inveniosoftware/invenio-search/blob/19c073d608d4c811f1c5aecb6622402d39715228/invenio_search/ext.py#L242-L254
|
train
|
inveniosoftware/invenio-search
|
invenio_search/ext.py
|
_SearchState.create
|
def create(self, ignore=None):
"""Yield tuple with created index name and responses from a client."""
ignore = ignore or []
def _create(tree_or_filename, alias=None):
"""Create indices and aliases by walking DFS."""
# Iterate over aliases:
for name, value in tree_or_filename.items():
if isinstance(value, dict):
for result in _create(value, alias=name):
yield result
else:
with open(value, 'r') as body:
yield name, self.client.indices.create(
index=name,
body=json.load(body),
ignore=ignore,
)
if alias:
yield alias, self.client.indices.put_alias(
index=list(_get_indices(tree_or_filename)),
name=alias,
ignore=ignore,
)
for result in _create(self.active_aliases):
yield result
|
python
|
def create(self, ignore=None):
"""Yield tuple with created index name and responses from a client."""
ignore = ignore or []
def _create(tree_or_filename, alias=None):
"""Create indices and aliases by walking DFS."""
# Iterate over aliases:
for name, value in tree_or_filename.items():
if isinstance(value, dict):
for result in _create(value, alias=name):
yield result
else:
with open(value, 'r') as body:
yield name, self.client.indices.create(
index=name,
body=json.load(body),
ignore=ignore,
)
if alias:
yield alias, self.client.indices.put_alias(
index=list(_get_indices(tree_or_filename)),
name=alias,
ignore=ignore,
)
for result in _create(self.active_aliases):
yield result
|
[
"def",
"create",
"(",
"self",
",",
"ignore",
"=",
"None",
")",
":",
"ignore",
"=",
"ignore",
"or",
"[",
"]",
"def",
"_create",
"(",
"tree_or_filename",
",",
"alias",
"=",
"None",
")",
":",
"\"\"\"Create indices and aliases by walking DFS.\"\"\"",
"# Iterate over aliases:",
"for",
"name",
",",
"value",
"in",
"tree_or_filename",
".",
"items",
"(",
")",
":",
"if",
"isinstance",
"(",
"value",
",",
"dict",
")",
":",
"for",
"result",
"in",
"_create",
"(",
"value",
",",
"alias",
"=",
"name",
")",
":",
"yield",
"result",
"else",
":",
"with",
"open",
"(",
"value",
",",
"'r'",
")",
"as",
"body",
":",
"yield",
"name",
",",
"self",
".",
"client",
".",
"indices",
".",
"create",
"(",
"index",
"=",
"name",
",",
"body",
"=",
"json",
".",
"load",
"(",
"body",
")",
",",
"ignore",
"=",
"ignore",
",",
")",
"if",
"alias",
":",
"yield",
"alias",
",",
"self",
".",
"client",
".",
"indices",
".",
"put_alias",
"(",
"index",
"=",
"list",
"(",
"_get_indices",
"(",
"tree_or_filename",
")",
")",
",",
"name",
"=",
"alias",
",",
"ignore",
"=",
"ignore",
",",
")",
"for",
"result",
"in",
"_create",
"(",
"self",
".",
"active_aliases",
")",
":",
"yield",
"result"
] |
Yield tuple with created index name and responses from a client.
|
[
"Yield",
"tuple",
"with",
"created",
"index",
"name",
"and",
"responses",
"from",
"a",
"client",
"."
] |
19c073d608d4c811f1c5aecb6622402d39715228
|
https://github.com/inveniosoftware/invenio-search/blob/19c073d608d4c811f1c5aecb6622402d39715228/invenio_search/ext.py#L256-L283
|
train
|
inveniosoftware/invenio-search
|
invenio_search/ext.py
|
_SearchState.put_templates
|
def put_templates(self, ignore=None):
"""Yield tuple with registered template and response from client."""
ignore = ignore or []
def _replace_prefix(template_path, body):
"""Replace index prefix in template request body."""
pattern = '__SEARCH_INDEX_PREFIX__'
prefix = self.app.config['SEARCH_INDEX_PREFIX'] or ''
if prefix:
assert pattern in body, "You are using the prefix `{0}`, "
"but the template `{1}` does not contain the "
"pattern `{2}`.".format(prefix, template_path, pattern)
return body.replace(pattern, prefix)
def _put_template(template):
"""Put template in search client."""
with open(self.templates[template], 'r') as fp:
body = fp.read()
replaced_body = _replace_prefix(self.templates[template], body)
return self.templates[template],\
current_search_client.indices.put_template(
name=template,
body=json.loads(replaced_body),
ignore=ignore,
)
for template in self.templates:
yield _put_template(template)
|
python
|
def put_templates(self, ignore=None):
"""Yield tuple with registered template and response from client."""
ignore = ignore or []
def _replace_prefix(template_path, body):
"""Replace index prefix in template request body."""
pattern = '__SEARCH_INDEX_PREFIX__'
prefix = self.app.config['SEARCH_INDEX_PREFIX'] or ''
if prefix:
assert pattern in body, "You are using the prefix `{0}`, "
"but the template `{1}` does not contain the "
"pattern `{2}`.".format(prefix, template_path, pattern)
return body.replace(pattern, prefix)
def _put_template(template):
"""Put template in search client."""
with open(self.templates[template], 'r') as fp:
body = fp.read()
replaced_body = _replace_prefix(self.templates[template], body)
return self.templates[template],\
current_search_client.indices.put_template(
name=template,
body=json.loads(replaced_body),
ignore=ignore,
)
for template in self.templates:
yield _put_template(template)
|
[
"def",
"put_templates",
"(",
"self",
",",
"ignore",
"=",
"None",
")",
":",
"ignore",
"=",
"ignore",
"or",
"[",
"]",
"def",
"_replace_prefix",
"(",
"template_path",
",",
"body",
")",
":",
"\"\"\"Replace index prefix in template request body.\"\"\"",
"pattern",
"=",
"'__SEARCH_INDEX_PREFIX__'",
"prefix",
"=",
"self",
".",
"app",
".",
"config",
"[",
"'SEARCH_INDEX_PREFIX'",
"]",
"or",
"''",
"if",
"prefix",
":",
"assert",
"pattern",
"in",
"body",
",",
"\"You are using the prefix `{0}`, \"",
"\"but the template `{1}` does not contain the \"",
"\"pattern `{2}`.\"",
".",
"format",
"(",
"prefix",
",",
"template_path",
",",
"pattern",
")",
"return",
"body",
".",
"replace",
"(",
"pattern",
",",
"prefix",
")",
"def",
"_put_template",
"(",
"template",
")",
":",
"\"\"\"Put template in search client.\"\"\"",
"with",
"open",
"(",
"self",
".",
"templates",
"[",
"template",
"]",
",",
"'r'",
")",
"as",
"fp",
":",
"body",
"=",
"fp",
".",
"read",
"(",
")",
"replaced_body",
"=",
"_replace_prefix",
"(",
"self",
".",
"templates",
"[",
"template",
"]",
",",
"body",
")",
"return",
"self",
".",
"templates",
"[",
"template",
"]",
",",
"current_search_client",
".",
"indices",
".",
"put_template",
"(",
"name",
"=",
"template",
",",
"body",
"=",
"json",
".",
"loads",
"(",
"replaced_body",
")",
",",
"ignore",
"=",
"ignore",
",",
")",
"for",
"template",
"in",
"self",
".",
"templates",
":",
"yield",
"_put_template",
"(",
"template",
")"
] |
Yield tuple with registered template and response from client.
|
[
"Yield",
"tuple",
"with",
"registered",
"template",
"and",
"response",
"from",
"client",
"."
] |
19c073d608d4c811f1c5aecb6622402d39715228
|
https://github.com/inveniosoftware/invenio-search/blob/19c073d608d4c811f1c5aecb6622402d39715228/invenio_search/ext.py#L285-L314
|
train
|
inveniosoftware/invenio-search
|
invenio_search/ext.py
|
_SearchState.delete
|
def delete(self, ignore=None):
"""Yield tuple with deleted index name and responses from a client."""
ignore = ignore or []
def _delete(tree_or_filename, alias=None):
"""Delete indexes and aliases by walking DFS."""
if alias:
yield alias, self.client.indices.delete_alias(
index=list(_get_indices(tree_or_filename)),
name=alias,
ignore=ignore,
)
# Iterate over aliases:
for name, value in tree_or_filename.items():
if isinstance(value, dict):
for result in _delete(value, alias=name):
yield result
else:
yield name, self.client.indices.delete(
index=name,
ignore=ignore,
)
for result in _delete(self.active_aliases):
yield result
|
python
|
def delete(self, ignore=None):
"""Yield tuple with deleted index name and responses from a client."""
ignore = ignore or []
def _delete(tree_or_filename, alias=None):
"""Delete indexes and aliases by walking DFS."""
if alias:
yield alias, self.client.indices.delete_alias(
index=list(_get_indices(tree_or_filename)),
name=alias,
ignore=ignore,
)
# Iterate over aliases:
for name, value in tree_or_filename.items():
if isinstance(value, dict):
for result in _delete(value, alias=name):
yield result
else:
yield name, self.client.indices.delete(
index=name,
ignore=ignore,
)
for result in _delete(self.active_aliases):
yield result
|
[
"def",
"delete",
"(",
"self",
",",
"ignore",
"=",
"None",
")",
":",
"ignore",
"=",
"ignore",
"or",
"[",
"]",
"def",
"_delete",
"(",
"tree_or_filename",
",",
"alias",
"=",
"None",
")",
":",
"\"\"\"Delete indexes and aliases by walking DFS.\"\"\"",
"if",
"alias",
":",
"yield",
"alias",
",",
"self",
".",
"client",
".",
"indices",
".",
"delete_alias",
"(",
"index",
"=",
"list",
"(",
"_get_indices",
"(",
"tree_or_filename",
")",
")",
",",
"name",
"=",
"alias",
",",
"ignore",
"=",
"ignore",
",",
")",
"# Iterate over aliases:",
"for",
"name",
",",
"value",
"in",
"tree_or_filename",
".",
"items",
"(",
")",
":",
"if",
"isinstance",
"(",
"value",
",",
"dict",
")",
":",
"for",
"result",
"in",
"_delete",
"(",
"value",
",",
"alias",
"=",
"name",
")",
":",
"yield",
"result",
"else",
":",
"yield",
"name",
",",
"self",
".",
"client",
".",
"indices",
".",
"delete",
"(",
"index",
"=",
"name",
",",
"ignore",
"=",
"ignore",
",",
")",
"for",
"result",
"in",
"_delete",
"(",
"self",
".",
"active_aliases",
")",
":",
"yield",
"result"
] |
Yield tuple with deleted index name and responses from a client.
|
[
"Yield",
"tuple",
"with",
"deleted",
"index",
"name",
"and",
"responses",
"from",
"a",
"client",
"."
] |
19c073d608d4c811f1c5aecb6622402d39715228
|
https://github.com/inveniosoftware/invenio-search/blob/19c073d608d4c811f1c5aecb6622402d39715228/invenio_search/ext.py#L316-L341
|
train
|
inveniosoftware/invenio-search
|
invenio_search/ext.py
|
InvenioSearch.init_app
|
def init_app(self, app,
entry_point_group_mappings='invenio_search.mappings',
entry_point_group_templates='invenio_search.templates',
**kwargs):
"""Flask application initialization.
:param app: An instance of :class:`~flask.app.Flask`.
"""
self.init_config(app)
app.cli.add_command(index_cmd)
state = _SearchState(
app,
entry_point_group_mappings=entry_point_group_mappings,
entry_point_group_templates=entry_point_group_templates,
**kwargs
)
self._state = app.extensions['invenio-search'] = state
|
python
|
def init_app(self, app,
entry_point_group_mappings='invenio_search.mappings',
entry_point_group_templates='invenio_search.templates',
**kwargs):
"""Flask application initialization.
:param app: An instance of :class:`~flask.app.Flask`.
"""
self.init_config(app)
app.cli.add_command(index_cmd)
state = _SearchState(
app,
entry_point_group_mappings=entry_point_group_mappings,
entry_point_group_templates=entry_point_group_templates,
**kwargs
)
self._state = app.extensions['invenio-search'] = state
|
[
"def",
"init_app",
"(",
"self",
",",
"app",
",",
"entry_point_group_mappings",
"=",
"'invenio_search.mappings'",
",",
"entry_point_group_templates",
"=",
"'invenio_search.templates'",
",",
"*",
"*",
"kwargs",
")",
":",
"self",
".",
"init_config",
"(",
"app",
")",
"app",
".",
"cli",
".",
"add_command",
"(",
"index_cmd",
")",
"state",
"=",
"_SearchState",
"(",
"app",
",",
"entry_point_group_mappings",
"=",
"entry_point_group_mappings",
",",
"entry_point_group_templates",
"=",
"entry_point_group_templates",
",",
"*",
"*",
"kwargs",
")",
"self",
".",
"_state",
"=",
"app",
".",
"extensions",
"[",
"'invenio-search'",
"]",
"=",
"state"
] |
Flask application initialization.
:param app: An instance of :class:`~flask.app.Flask`.
|
[
"Flask",
"application",
"initialization",
"."
] |
19c073d608d4c811f1c5aecb6622402d39715228
|
https://github.com/inveniosoftware/invenio-search/blob/19c073d608d4c811f1c5aecb6622402d39715228/invenio_search/ext.py#L357-L375
|
train
|
ybrs/pydisque
|
example/poor_consumer.py
|
main
|
def main():
"""Start the poor_consumer."""
try:
opts, args = getopt.getopt(sys.argv[1:], "h:v", ["help", "nack=",
"servers=", "queues="])
except getopt.GetoptError as err:
print str(err)
usage()
sys.exit()
# defaults
nack = 0.0
verbose = False
servers = "localhost:7712,localhost:7711"
queues = "test"
for o, a in opts:
if o == "-v":
verbose = True
elif o in ("-h", "--help"):
usage()
sys.exit()
elif o in ("--nack"):
nack = float(a)
elif o in ("--servers"):
servers = a
elif o in ("--queues"):
queues = a
else:
assert False, "unhandled option"
# prepare servers and queus for pydisque
servers = servers.split(",")
queues = queues.split(",")
c = Client(servers)
c.connect()
while True:
jobs = c.get_job(queues)
for queue_name, job_id, job in jobs:
rnd = random.random()
# as this is a test processor, we don't do any validation on
# the actual job body, so lets just pay attention to id's
if rnd >= nack:
print ">>> received job:", job_id
c.ack_job(job_id)
else:
print ">>> bouncing job:", job_id
c.nack_job(job_id)
|
python
|
def main():
"""Start the poor_consumer."""
try:
opts, args = getopt.getopt(sys.argv[1:], "h:v", ["help", "nack=",
"servers=", "queues="])
except getopt.GetoptError as err:
print str(err)
usage()
sys.exit()
# defaults
nack = 0.0
verbose = False
servers = "localhost:7712,localhost:7711"
queues = "test"
for o, a in opts:
if o == "-v":
verbose = True
elif o in ("-h", "--help"):
usage()
sys.exit()
elif o in ("--nack"):
nack = float(a)
elif o in ("--servers"):
servers = a
elif o in ("--queues"):
queues = a
else:
assert False, "unhandled option"
# prepare servers and queus for pydisque
servers = servers.split(",")
queues = queues.split(",")
c = Client(servers)
c.connect()
while True:
jobs = c.get_job(queues)
for queue_name, job_id, job in jobs:
rnd = random.random()
# as this is a test processor, we don't do any validation on
# the actual job body, so lets just pay attention to id's
if rnd >= nack:
print ">>> received job:", job_id
c.ack_job(job_id)
else:
print ">>> bouncing job:", job_id
c.nack_job(job_id)
|
[
"def",
"main",
"(",
")",
":",
"try",
":",
"opts",
",",
"args",
"=",
"getopt",
".",
"getopt",
"(",
"sys",
".",
"argv",
"[",
"1",
":",
"]",
",",
"\"h:v\"",
",",
"[",
"\"help\"",
",",
"\"nack=\"",
",",
"\"servers=\"",
",",
"\"queues=\"",
"]",
")",
"except",
"getopt",
".",
"GetoptError",
"as",
"err",
":",
"print",
"str",
"(",
"err",
")",
"usage",
"(",
")",
"sys",
".",
"exit",
"(",
")",
"# defaults",
"nack",
"=",
"0.0",
"verbose",
"=",
"False",
"servers",
"=",
"\"localhost:7712,localhost:7711\"",
"queues",
"=",
"\"test\"",
"for",
"o",
",",
"a",
"in",
"opts",
":",
"if",
"o",
"==",
"\"-v\"",
":",
"verbose",
"=",
"True",
"elif",
"o",
"in",
"(",
"\"-h\"",
",",
"\"--help\"",
")",
":",
"usage",
"(",
")",
"sys",
".",
"exit",
"(",
")",
"elif",
"o",
"in",
"(",
"\"--nack\"",
")",
":",
"nack",
"=",
"float",
"(",
"a",
")",
"elif",
"o",
"in",
"(",
"\"--servers\"",
")",
":",
"servers",
"=",
"a",
"elif",
"o",
"in",
"(",
"\"--queues\"",
")",
":",
"queues",
"=",
"a",
"else",
":",
"assert",
"False",
",",
"\"unhandled option\"",
"# prepare servers and queus for pydisque",
"servers",
"=",
"servers",
".",
"split",
"(",
"\",\"",
")",
"queues",
"=",
"queues",
".",
"split",
"(",
"\",\"",
")",
"c",
"=",
"Client",
"(",
"servers",
")",
"c",
".",
"connect",
"(",
")",
"while",
"True",
":",
"jobs",
"=",
"c",
".",
"get_job",
"(",
"queues",
")",
"for",
"queue_name",
",",
"job_id",
",",
"job",
"in",
"jobs",
":",
"rnd",
"=",
"random",
".",
"random",
"(",
")",
"# as this is a test processor, we don't do any validation on",
"# the actual job body, so lets just pay attention to id's",
"if",
"rnd",
">=",
"nack",
":",
"print",
"\">>> received job:\"",
",",
"job_id",
"c",
".",
"ack_job",
"(",
"job_id",
")",
"else",
":",
"print",
"\">>> bouncing job:\"",
",",
"job_id",
"c",
".",
"nack_job",
"(",
"job_id",
")"
] |
Start the poor_consumer.
|
[
"Start",
"the",
"poor_consumer",
"."
] |
ea5ce1576b66398c1cce32cad0f15709b1ea8df8
|
https://github.com/ybrs/pydisque/blob/ea5ce1576b66398c1cce32cad0f15709b1ea8df8/example/poor_consumer.py#L27-L78
|
train
|
ybrs/pydisque
|
pydisque/client.py
|
Client.connect
|
def connect(self):
"""
Connect to one of the Disque nodes.
You can get current connection with connected_node property
:returns: nothing
"""
self.connected_node = None
for i, node in self.nodes.items():
host, port = i.split(':')
port = int(port)
redis_client = redis.Redis(host, port, **self.client_kw_args)
try:
ret = redis_client.execute_command('HELLO')
format_version, node_id = ret[0], ret[1]
others = ret[2:]
self.nodes[i] = Node(node_id, host, port, redis_client)
self.connected_node = self.nodes[i]
except redis.exceptions.ConnectionError:
pass
if not self.connected_node:
raise ConnectionError('couldnt connect to any nodes')
logger.info("connected to node %s" % self.connected_node)
|
python
|
def connect(self):
"""
Connect to one of the Disque nodes.
You can get current connection with connected_node property
:returns: nothing
"""
self.connected_node = None
for i, node in self.nodes.items():
host, port = i.split(':')
port = int(port)
redis_client = redis.Redis(host, port, **self.client_kw_args)
try:
ret = redis_client.execute_command('HELLO')
format_version, node_id = ret[0], ret[1]
others = ret[2:]
self.nodes[i] = Node(node_id, host, port, redis_client)
self.connected_node = self.nodes[i]
except redis.exceptions.ConnectionError:
pass
if not self.connected_node:
raise ConnectionError('couldnt connect to any nodes')
logger.info("connected to node %s" % self.connected_node)
|
[
"def",
"connect",
"(",
"self",
")",
":",
"self",
".",
"connected_node",
"=",
"None",
"for",
"i",
",",
"node",
"in",
"self",
".",
"nodes",
".",
"items",
"(",
")",
":",
"host",
",",
"port",
"=",
"i",
".",
"split",
"(",
"':'",
")",
"port",
"=",
"int",
"(",
"port",
")",
"redis_client",
"=",
"redis",
".",
"Redis",
"(",
"host",
",",
"port",
",",
"*",
"*",
"self",
".",
"client_kw_args",
")",
"try",
":",
"ret",
"=",
"redis_client",
".",
"execute_command",
"(",
"'HELLO'",
")",
"format_version",
",",
"node_id",
"=",
"ret",
"[",
"0",
"]",
",",
"ret",
"[",
"1",
"]",
"others",
"=",
"ret",
"[",
"2",
":",
"]",
"self",
".",
"nodes",
"[",
"i",
"]",
"=",
"Node",
"(",
"node_id",
",",
"host",
",",
"port",
",",
"redis_client",
")",
"self",
".",
"connected_node",
"=",
"self",
".",
"nodes",
"[",
"i",
"]",
"except",
"redis",
".",
"exceptions",
".",
"ConnectionError",
":",
"pass",
"if",
"not",
"self",
".",
"connected_node",
":",
"raise",
"ConnectionError",
"(",
"'couldnt connect to any nodes'",
")",
"logger",
".",
"info",
"(",
"\"connected to node %s\"",
"%",
"self",
".",
"connected_node",
")"
] |
Connect to one of the Disque nodes.
You can get current connection with connected_node property
:returns: nothing
|
[
"Connect",
"to",
"one",
"of",
"the",
"Disque",
"nodes",
"."
] |
ea5ce1576b66398c1cce32cad0f15709b1ea8df8
|
https://github.com/ybrs/pydisque/blob/ea5ce1576b66398c1cce32cad0f15709b1ea8df8/pydisque/client.py#L117-L143
|
train
|
ybrs/pydisque
|
pydisque/client.py
|
Client.execute_command
|
def execute_command(self, *args, **kwargs):
"""Execute a command on the connected server."""
try:
return self.get_connection().execute_command(*args, **kwargs)
except ConnectionError as e:
logger.warn('trying to reconnect')
self.connect()
logger.warn('connected')
raise
|
python
|
def execute_command(self, *args, **kwargs):
"""Execute a command on the connected server."""
try:
return self.get_connection().execute_command(*args, **kwargs)
except ConnectionError as e:
logger.warn('trying to reconnect')
self.connect()
logger.warn('connected')
raise
|
[
"def",
"execute_command",
"(",
"self",
",",
"*",
"args",
",",
"*",
"*",
"kwargs",
")",
":",
"try",
":",
"return",
"self",
".",
"get_connection",
"(",
")",
".",
"execute_command",
"(",
"*",
"args",
",",
"*",
"*",
"kwargs",
")",
"except",
"ConnectionError",
"as",
"e",
":",
"logger",
".",
"warn",
"(",
"'trying to reconnect'",
")",
"self",
".",
"connect",
"(",
")",
"logger",
".",
"warn",
"(",
"'connected'",
")",
"raise"
] |
Execute a command on the connected server.
|
[
"Execute",
"a",
"command",
"on",
"the",
"connected",
"server",
"."
] |
ea5ce1576b66398c1cce32cad0f15709b1ea8df8
|
https://github.com/ybrs/pydisque/blob/ea5ce1576b66398c1cce32cad0f15709b1ea8df8/pydisque/client.py#L157-L165
|
train
|
ybrs/pydisque
|
pydisque/client.py
|
Client.add_job
|
def add_job(self, queue_name, job, timeout=200, replicate=None, delay=None,
retry=None, ttl=None, maxlen=None, asynchronous=None):
"""
Add a job to a queue.
ADDJOB queue_name job <ms-timeout> [REPLICATE <count>] [DELAY <sec>]
[RETRY <sec>] [TTL <sec>] [MAXLEN <count>] [ASYNC]
:param queue_name: is the name of the queue, any string, basically.
:param job: is a string representing the job.
:param timeout: is the command timeout in milliseconds.
:param replicate: count is the number of nodes the job should be
replicated to.
:param delay: sec is the number of seconds that should elapse
before the job is queued by any server.
:param retry: sec period after which, if no ACK is received, the
job is put again into the queue for delivery. If RETRY is 0,
the job has an at-most-once delivery semantics.
:param ttl: sec is the max job life in seconds. After this time,
the job is deleted even if it was not successfully delivered.
:param maxlen: count specifies that if there are already count
messages queued for the specified queue name, the message is
refused and an error reported to the client.
:param asynchronous: asks the server to let the command return ASAP and
replicate the job to other nodes in the background. The job
gets queued ASAP, while normally the job is put into the queue
only when the client gets a positive reply. Changing the name of this
argument as async is reserved keyword in python 3.7
:returns: job_id
"""
command = ['ADDJOB', queue_name, job, timeout]
if replicate:
command += ['REPLICATE', replicate]
if delay:
command += ['DELAY', delay]
if retry is not None:
command += ['RETRY', retry]
if ttl:
command += ['TTL', ttl]
if maxlen:
command += ['MAXLEN', maxlen]
if asynchronous:
command += ['ASYNC']
# TODO(canardleteer): we need to handle "-PAUSE" messages more
# appropriately, for now it's up to the person using the library
# to handle a generic ResponseError on their own.
logger.debug("sending job - %s", command)
job_id = self.execute_command(*command)
logger.debug("sent job - %s", command)
logger.debug("job_id: %s " % job_id)
return job_id
|
python
|
def add_job(self, queue_name, job, timeout=200, replicate=None, delay=None,
retry=None, ttl=None, maxlen=None, asynchronous=None):
"""
Add a job to a queue.
ADDJOB queue_name job <ms-timeout> [REPLICATE <count>] [DELAY <sec>]
[RETRY <sec>] [TTL <sec>] [MAXLEN <count>] [ASYNC]
:param queue_name: is the name of the queue, any string, basically.
:param job: is a string representing the job.
:param timeout: is the command timeout in milliseconds.
:param replicate: count is the number of nodes the job should be
replicated to.
:param delay: sec is the number of seconds that should elapse
before the job is queued by any server.
:param retry: sec period after which, if no ACK is received, the
job is put again into the queue for delivery. If RETRY is 0,
the job has an at-most-once delivery semantics.
:param ttl: sec is the max job life in seconds. After this time,
the job is deleted even if it was not successfully delivered.
:param maxlen: count specifies that if there are already count
messages queued for the specified queue name, the message is
refused and an error reported to the client.
:param asynchronous: asks the server to let the command return ASAP and
replicate the job to other nodes in the background. The job
gets queued ASAP, while normally the job is put into the queue
only when the client gets a positive reply. Changing the name of this
argument as async is reserved keyword in python 3.7
:returns: job_id
"""
command = ['ADDJOB', queue_name, job, timeout]
if replicate:
command += ['REPLICATE', replicate]
if delay:
command += ['DELAY', delay]
if retry is not None:
command += ['RETRY', retry]
if ttl:
command += ['TTL', ttl]
if maxlen:
command += ['MAXLEN', maxlen]
if asynchronous:
command += ['ASYNC']
# TODO(canardleteer): we need to handle "-PAUSE" messages more
# appropriately, for now it's up to the person using the library
# to handle a generic ResponseError on their own.
logger.debug("sending job - %s", command)
job_id = self.execute_command(*command)
logger.debug("sent job - %s", command)
logger.debug("job_id: %s " % job_id)
return job_id
|
[
"def",
"add_job",
"(",
"self",
",",
"queue_name",
",",
"job",
",",
"timeout",
"=",
"200",
",",
"replicate",
"=",
"None",
",",
"delay",
"=",
"None",
",",
"retry",
"=",
"None",
",",
"ttl",
"=",
"None",
",",
"maxlen",
"=",
"None",
",",
"asynchronous",
"=",
"None",
")",
":",
"command",
"=",
"[",
"'ADDJOB'",
",",
"queue_name",
",",
"job",
",",
"timeout",
"]",
"if",
"replicate",
":",
"command",
"+=",
"[",
"'REPLICATE'",
",",
"replicate",
"]",
"if",
"delay",
":",
"command",
"+=",
"[",
"'DELAY'",
",",
"delay",
"]",
"if",
"retry",
"is",
"not",
"None",
":",
"command",
"+=",
"[",
"'RETRY'",
",",
"retry",
"]",
"if",
"ttl",
":",
"command",
"+=",
"[",
"'TTL'",
",",
"ttl",
"]",
"if",
"maxlen",
":",
"command",
"+=",
"[",
"'MAXLEN'",
",",
"maxlen",
"]",
"if",
"asynchronous",
":",
"command",
"+=",
"[",
"'ASYNC'",
"]",
"# TODO(canardleteer): we need to handle \"-PAUSE\" messages more",
"# appropriately, for now it's up to the person using the library",
"# to handle a generic ResponseError on their own.",
"logger",
".",
"debug",
"(",
"\"sending job - %s\"",
",",
"command",
")",
"job_id",
"=",
"self",
".",
"execute_command",
"(",
"*",
"command",
")",
"logger",
".",
"debug",
"(",
"\"sent job - %s\"",
",",
"command",
")",
"logger",
".",
"debug",
"(",
"\"job_id: %s \"",
"%",
"job_id",
")",
"return",
"job_id"
] |
Add a job to a queue.
ADDJOB queue_name job <ms-timeout> [REPLICATE <count>] [DELAY <sec>]
[RETRY <sec>] [TTL <sec>] [MAXLEN <count>] [ASYNC]
:param queue_name: is the name of the queue, any string, basically.
:param job: is a string representing the job.
:param timeout: is the command timeout in milliseconds.
:param replicate: count is the number of nodes the job should be
replicated to.
:param delay: sec is the number of seconds that should elapse
before the job is queued by any server.
:param retry: sec period after which, if no ACK is received, the
job is put again into the queue for delivery. If RETRY is 0,
the job has an at-most-once delivery semantics.
:param ttl: sec is the max job life in seconds. After this time,
the job is deleted even if it was not successfully delivered.
:param maxlen: count specifies that if there are already count
messages queued for the specified queue name, the message is
refused and an error reported to the client.
:param asynchronous: asks the server to let the command return ASAP and
replicate the job to other nodes in the background. The job
gets queued ASAP, while normally the job is put into the queue
only when the client gets a positive reply. Changing the name of this
argument as async is reserved keyword in python 3.7
:returns: job_id
|
[
"Add",
"a",
"job",
"to",
"a",
"queue",
"."
] |
ea5ce1576b66398c1cce32cad0f15709b1ea8df8
|
https://github.com/ybrs/pydisque/blob/ea5ce1576b66398c1cce32cad0f15709b1ea8df8/pydisque/client.py#L182-L235
|
train
|
ybrs/pydisque
|
pydisque/client.py
|
Client.get_job
|
def get_job(self, queues, timeout=None, count=None, nohang=False, withcounters=False):
"""
Return some number of jobs from specified queues.
GETJOB [NOHANG] [TIMEOUT <ms-timeout>] [COUNT <count>] [WITHCOUNTERS] FROM
queue1 queue2 ... queueN
:param queues: name of queues
:returns: list of tuple(job_id, queue_name, job), tuple(job_id, queue_name, job, nacks, additional_deliveries) or empty list
:rtype: list
"""
assert queues
command = ['GETJOB']
if nohang:
command += ['NOHANG']
if timeout:
command += ['TIMEOUT', timeout]
if count:
command += ['COUNT', count]
if withcounters:
command += ['WITHCOUNTERS']
command += ['FROM'] + queues
results = self.execute_command(*command)
if not results:
return []
if withcounters:
return [(job_id, queue_name, job, nacks, additional_deliveries) for
job_id, queue_name, job, _, nacks, _, additional_deliveries in results]
else:
return [(job_id, queue_name, job) for
job_id, queue_name, job in results]
|
python
|
def get_job(self, queues, timeout=None, count=None, nohang=False, withcounters=False):
"""
Return some number of jobs from specified queues.
GETJOB [NOHANG] [TIMEOUT <ms-timeout>] [COUNT <count>] [WITHCOUNTERS] FROM
queue1 queue2 ... queueN
:param queues: name of queues
:returns: list of tuple(job_id, queue_name, job), tuple(job_id, queue_name, job, nacks, additional_deliveries) or empty list
:rtype: list
"""
assert queues
command = ['GETJOB']
if nohang:
command += ['NOHANG']
if timeout:
command += ['TIMEOUT', timeout]
if count:
command += ['COUNT', count]
if withcounters:
command += ['WITHCOUNTERS']
command += ['FROM'] + queues
results = self.execute_command(*command)
if not results:
return []
if withcounters:
return [(job_id, queue_name, job, nacks, additional_deliveries) for
job_id, queue_name, job, _, nacks, _, additional_deliveries in results]
else:
return [(job_id, queue_name, job) for
job_id, queue_name, job in results]
|
[
"def",
"get_job",
"(",
"self",
",",
"queues",
",",
"timeout",
"=",
"None",
",",
"count",
"=",
"None",
",",
"nohang",
"=",
"False",
",",
"withcounters",
"=",
"False",
")",
":",
"assert",
"queues",
"command",
"=",
"[",
"'GETJOB'",
"]",
"if",
"nohang",
":",
"command",
"+=",
"[",
"'NOHANG'",
"]",
"if",
"timeout",
":",
"command",
"+=",
"[",
"'TIMEOUT'",
",",
"timeout",
"]",
"if",
"count",
":",
"command",
"+=",
"[",
"'COUNT'",
",",
"count",
"]",
"if",
"withcounters",
":",
"command",
"+=",
"[",
"'WITHCOUNTERS'",
"]",
"command",
"+=",
"[",
"'FROM'",
"]",
"+",
"queues",
"results",
"=",
"self",
".",
"execute_command",
"(",
"*",
"command",
")",
"if",
"not",
"results",
":",
"return",
"[",
"]",
"if",
"withcounters",
":",
"return",
"[",
"(",
"job_id",
",",
"queue_name",
",",
"job",
",",
"nacks",
",",
"additional_deliveries",
")",
"for",
"job_id",
",",
"queue_name",
",",
"job",
",",
"_",
",",
"nacks",
",",
"_",
",",
"additional_deliveries",
"in",
"results",
"]",
"else",
":",
"return",
"[",
"(",
"job_id",
",",
"queue_name",
",",
"job",
")",
"for",
"job_id",
",",
"queue_name",
",",
"job",
"in",
"results",
"]"
] |
Return some number of jobs from specified queues.
GETJOB [NOHANG] [TIMEOUT <ms-timeout>] [COUNT <count>] [WITHCOUNTERS] FROM
queue1 queue2 ... queueN
:param queues: name of queues
:returns: list of tuple(job_id, queue_name, job), tuple(job_id, queue_name, job, nacks, additional_deliveries) or empty list
:rtype: list
|
[
"Return",
"some",
"number",
"of",
"jobs",
"from",
"specified",
"queues",
"."
] |
ea5ce1576b66398c1cce32cad0f15709b1ea8df8
|
https://github.com/ybrs/pydisque/blob/ea5ce1576b66398c1cce32cad0f15709b1ea8df8/pydisque/client.py#L237-L271
|
train
|
ybrs/pydisque
|
pydisque/client.py
|
Client.qstat
|
def qstat(self, queue_name, return_dict=False):
"""
Return the status of the queue (currently unimplemented).
Future support / testing of QSTAT support in Disque
QSTAT <qname>
Return produced ... consumed ... idle ... sources [...] ctime ...
"""
rtn = self.execute_command('QSTAT', queue_name)
if return_dict:
grouped = self._grouper(rtn, 2)
rtn = dict((a, b) for a, b in grouped)
return rtn
|
python
|
def qstat(self, queue_name, return_dict=False):
"""
Return the status of the queue (currently unimplemented).
Future support / testing of QSTAT support in Disque
QSTAT <qname>
Return produced ... consumed ... idle ... sources [...] ctime ...
"""
rtn = self.execute_command('QSTAT', queue_name)
if return_dict:
grouped = self._grouper(rtn, 2)
rtn = dict((a, b) for a, b in grouped)
return rtn
|
[
"def",
"qstat",
"(",
"self",
",",
"queue_name",
",",
"return_dict",
"=",
"False",
")",
":",
"rtn",
"=",
"self",
".",
"execute_command",
"(",
"'QSTAT'",
",",
"queue_name",
")",
"if",
"return_dict",
":",
"grouped",
"=",
"self",
".",
"_grouper",
"(",
"rtn",
",",
"2",
")",
"rtn",
"=",
"dict",
"(",
"(",
"a",
",",
"b",
")",
"for",
"a",
",",
"b",
"in",
"grouped",
")",
"return",
"rtn"
] |
Return the status of the queue (currently unimplemented).
Future support / testing of QSTAT support in Disque
QSTAT <qname>
Return produced ... consumed ... idle ... sources [...] ctime ...
|
[
"Return",
"the",
"status",
"of",
"the",
"queue",
"(",
"currently",
"unimplemented",
")",
"."
] |
ea5ce1576b66398c1cce32cad0f15709b1ea8df8
|
https://github.com/ybrs/pydisque/blob/ea5ce1576b66398c1cce32cad0f15709b1ea8df8/pydisque/client.py#L333-L349
|
train
|
ybrs/pydisque
|
pydisque/client.py
|
Client.show
|
def show(self, job_id, return_dict=False):
"""
Describe the job.
:param job_id:
"""
rtn = self.execute_command('SHOW', job_id)
if return_dict:
grouped = self._grouper(rtn, 2)
rtn = dict((a, b) for a, b in grouped)
return rtn
|
python
|
def show(self, job_id, return_dict=False):
"""
Describe the job.
:param job_id:
"""
rtn = self.execute_command('SHOW', job_id)
if return_dict:
grouped = self._grouper(rtn, 2)
rtn = dict((a, b) for a, b in grouped)
return rtn
|
[
"def",
"show",
"(",
"self",
",",
"job_id",
",",
"return_dict",
"=",
"False",
")",
":",
"rtn",
"=",
"self",
".",
"execute_command",
"(",
"'SHOW'",
",",
"job_id",
")",
"if",
"return_dict",
":",
"grouped",
"=",
"self",
".",
"_grouper",
"(",
"rtn",
",",
"2",
")",
"rtn",
"=",
"dict",
"(",
"(",
"a",
",",
"b",
")",
"for",
"a",
",",
"b",
"in",
"grouped",
")",
"return",
"rtn"
] |
Describe the job.
:param job_id:
|
[
"Describe",
"the",
"job",
"."
] |
ea5ce1576b66398c1cce32cad0f15709b1ea8df8
|
https://github.com/ybrs/pydisque/blob/ea5ce1576b66398c1cce32cad0f15709b1ea8df8/pydisque/client.py#L402-L415
|
train
|
ybrs/pydisque
|
pydisque/client.py
|
Client.pause
|
def pause(self, queue_name, kw_in=None, kw_out=None, kw_all=None,
kw_none=None, kw_state=None, kw_bcast=None):
"""
Pause a queue.
Unfortunately, the PAUSE keywords are mostly reserved words in Python,
so I've been a little creative in the function variable names. Open
to suggestions to change it (canardleteer)
:param queue_name: The job queue we are modifying.
:param kw_in: pause the queue in input.
:param kw_out: pause the queue in output.
:param kw_all: pause the queue in input and output (same as specifying
both the in and out options).
:param kw_none: clear the paused state in input and output.
:param kw_state: just report the current queue state.
:param kw_bcast: send a PAUSE command to all the reachable nodes of
the cluster to set the same queue in the other nodes
to the same state.
"""
command = ["PAUSE", queue_name]
if kw_in:
command += ["in"]
if kw_out:
command += ["out"]
if kw_all:
command += ["all"]
if kw_none:
command += ["none"]
if kw_state:
command += ["state"]
if kw_bcast:
command += ["bcast"]
return self.execute_command(*command)
|
python
|
def pause(self, queue_name, kw_in=None, kw_out=None, kw_all=None,
kw_none=None, kw_state=None, kw_bcast=None):
"""
Pause a queue.
Unfortunately, the PAUSE keywords are mostly reserved words in Python,
so I've been a little creative in the function variable names. Open
to suggestions to change it (canardleteer)
:param queue_name: The job queue we are modifying.
:param kw_in: pause the queue in input.
:param kw_out: pause the queue in output.
:param kw_all: pause the queue in input and output (same as specifying
both the in and out options).
:param kw_none: clear the paused state in input and output.
:param kw_state: just report the current queue state.
:param kw_bcast: send a PAUSE command to all the reachable nodes of
the cluster to set the same queue in the other nodes
to the same state.
"""
command = ["PAUSE", queue_name]
if kw_in:
command += ["in"]
if kw_out:
command += ["out"]
if kw_all:
command += ["all"]
if kw_none:
command += ["none"]
if kw_state:
command += ["state"]
if kw_bcast:
command += ["bcast"]
return self.execute_command(*command)
|
[
"def",
"pause",
"(",
"self",
",",
"queue_name",
",",
"kw_in",
"=",
"None",
",",
"kw_out",
"=",
"None",
",",
"kw_all",
"=",
"None",
",",
"kw_none",
"=",
"None",
",",
"kw_state",
"=",
"None",
",",
"kw_bcast",
"=",
"None",
")",
":",
"command",
"=",
"[",
"\"PAUSE\"",
",",
"queue_name",
"]",
"if",
"kw_in",
":",
"command",
"+=",
"[",
"\"in\"",
"]",
"if",
"kw_out",
":",
"command",
"+=",
"[",
"\"out\"",
"]",
"if",
"kw_all",
":",
"command",
"+=",
"[",
"\"all\"",
"]",
"if",
"kw_none",
":",
"command",
"+=",
"[",
"\"none\"",
"]",
"if",
"kw_state",
":",
"command",
"+=",
"[",
"\"state\"",
"]",
"if",
"kw_bcast",
":",
"command",
"+=",
"[",
"\"bcast\"",
"]",
"return",
"self",
".",
"execute_command",
"(",
"*",
"command",
")"
] |
Pause a queue.
Unfortunately, the PAUSE keywords are mostly reserved words in Python,
so I've been a little creative in the function variable names. Open
to suggestions to change it (canardleteer)
:param queue_name: The job queue we are modifying.
:param kw_in: pause the queue in input.
:param kw_out: pause the queue in output.
:param kw_all: pause the queue in input and output (same as specifying
both the in and out options).
:param kw_none: clear the paused state in input and output.
:param kw_state: just report the current queue state.
:param kw_bcast: send a PAUSE command to all the reachable nodes of
the cluster to set the same queue in the other nodes
to the same state.
|
[
"Pause",
"a",
"queue",
"."
] |
ea5ce1576b66398c1cce32cad0f15709b1ea8df8
|
https://github.com/ybrs/pydisque/blob/ea5ce1576b66398c1cce32cad0f15709b1ea8df8/pydisque/client.py#L417-L451
|
train
|
ybrs/pydisque
|
pydisque/client.py
|
Client.qscan
|
def qscan(self, cursor=0, count=None, busyloop=None, minlen=None,
maxlen=None, importrate=None):
"""
Iterate all the existing queues in the local node.
:param count: An hint about how much work to do per iteration.
:param busyloop: Block and return all the elements in a busy loop.
:param minlen: Don't return elements with less than count jobs queued.
:param maxlen: Don't return elements with more than count jobs queued.
:param importrate: Only return elements with an job import rate
(from other nodes) >= rate.
"""
command = ["QSCAN", cursor]
if count:
command += ["COUNT", count]
if busyloop:
command += ["BUSYLOOP"]
if minlen:
command += ["MINLEN", minlen]
if maxlen:
command += ["MAXLEN", maxlen]
if importrate:
command += ["IMPORTRATE", importrate]
return self.execute_command(*command)
|
python
|
def qscan(self, cursor=0, count=None, busyloop=None, minlen=None,
maxlen=None, importrate=None):
"""
Iterate all the existing queues in the local node.
:param count: An hint about how much work to do per iteration.
:param busyloop: Block and return all the elements in a busy loop.
:param minlen: Don't return elements with less than count jobs queued.
:param maxlen: Don't return elements with more than count jobs queued.
:param importrate: Only return elements with an job import rate
(from other nodes) >= rate.
"""
command = ["QSCAN", cursor]
if count:
command += ["COUNT", count]
if busyloop:
command += ["BUSYLOOP"]
if minlen:
command += ["MINLEN", minlen]
if maxlen:
command += ["MAXLEN", maxlen]
if importrate:
command += ["IMPORTRATE", importrate]
return self.execute_command(*command)
|
[
"def",
"qscan",
"(",
"self",
",",
"cursor",
"=",
"0",
",",
"count",
"=",
"None",
",",
"busyloop",
"=",
"None",
",",
"minlen",
"=",
"None",
",",
"maxlen",
"=",
"None",
",",
"importrate",
"=",
"None",
")",
":",
"command",
"=",
"[",
"\"QSCAN\"",
",",
"cursor",
"]",
"if",
"count",
":",
"command",
"+=",
"[",
"\"COUNT\"",
",",
"count",
"]",
"if",
"busyloop",
":",
"command",
"+=",
"[",
"\"BUSYLOOP\"",
"]",
"if",
"minlen",
":",
"command",
"+=",
"[",
"\"MINLEN\"",
",",
"minlen",
"]",
"if",
"maxlen",
":",
"command",
"+=",
"[",
"\"MAXLEN\"",
",",
"maxlen",
"]",
"if",
"importrate",
":",
"command",
"+=",
"[",
"\"IMPORTRATE\"",
",",
"importrate",
"]",
"return",
"self",
".",
"execute_command",
"(",
"*",
"command",
")"
] |
Iterate all the existing queues in the local node.
:param count: An hint about how much work to do per iteration.
:param busyloop: Block and return all the elements in a busy loop.
:param minlen: Don't return elements with less than count jobs queued.
:param maxlen: Don't return elements with more than count jobs queued.
:param importrate: Only return elements with an job import rate
(from other nodes) >= rate.
|
[
"Iterate",
"all",
"the",
"existing",
"queues",
"in",
"the",
"local",
"node",
"."
] |
ea5ce1576b66398c1cce32cad0f15709b1ea8df8
|
https://github.com/ybrs/pydisque/blob/ea5ce1576b66398c1cce32cad0f15709b1ea8df8/pydisque/client.py#L453-L477
|
train
|
ybrs/pydisque
|
pydisque/client.py
|
Client.jscan
|
def jscan(self, cursor=0, count=None, busyloop=None, queue=None,
state=None, reply=None):
"""Iterate all the existing jobs in the local node.
:param count: An hint about how much work to do per iteration.
:param busyloop: Block and return all the elements in a busy loop.
:param queue: Return only jobs in the specified queue.
:param state: Must be a list - Return jobs in the specified state.
Can be used multiple times for a logic OR.
:param reply: None or string {"all", "id"} - Job reply type. Type can
be all or id. Default is to report just the job ID. If all is
specified the full job state is returned like for the SHOW command.
"""
command = ["JSCAN", cursor]
if count:
command += ["COUNT", count]
if busyloop:
command += ["BUSYLOOP"]
if queue:
command += ["QUEUE", queue]
if type(state) is list:
for s in state:
command += ["STATE", s]
if reply:
command += ["REPLY", reply]
return self.execute_command(*command)
|
python
|
def jscan(self, cursor=0, count=None, busyloop=None, queue=None,
state=None, reply=None):
"""Iterate all the existing jobs in the local node.
:param count: An hint about how much work to do per iteration.
:param busyloop: Block and return all the elements in a busy loop.
:param queue: Return only jobs in the specified queue.
:param state: Must be a list - Return jobs in the specified state.
Can be used multiple times for a logic OR.
:param reply: None or string {"all", "id"} - Job reply type. Type can
be all or id. Default is to report just the job ID. If all is
specified the full job state is returned like for the SHOW command.
"""
command = ["JSCAN", cursor]
if count:
command += ["COUNT", count]
if busyloop:
command += ["BUSYLOOP"]
if queue:
command += ["QUEUE", queue]
if type(state) is list:
for s in state:
command += ["STATE", s]
if reply:
command += ["REPLY", reply]
return self.execute_command(*command)
|
[
"def",
"jscan",
"(",
"self",
",",
"cursor",
"=",
"0",
",",
"count",
"=",
"None",
",",
"busyloop",
"=",
"None",
",",
"queue",
"=",
"None",
",",
"state",
"=",
"None",
",",
"reply",
"=",
"None",
")",
":",
"command",
"=",
"[",
"\"JSCAN\"",
",",
"cursor",
"]",
"if",
"count",
":",
"command",
"+=",
"[",
"\"COUNT\"",
",",
"count",
"]",
"if",
"busyloop",
":",
"command",
"+=",
"[",
"\"BUSYLOOP\"",
"]",
"if",
"queue",
":",
"command",
"+=",
"[",
"\"QUEUE\"",
",",
"queue",
"]",
"if",
"type",
"(",
"state",
")",
"is",
"list",
":",
"for",
"s",
"in",
"state",
":",
"command",
"+=",
"[",
"\"STATE\"",
",",
"s",
"]",
"if",
"reply",
":",
"command",
"+=",
"[",
"\"REPLY\"",
",",
"reply",
"]",
"return",
"self",
".",
"execute_command",
"(",
"*",
"command",
")"
] |
Iterate all the existing jobs in the local node.
:param count: An hint about how much work to do per iteration.
:param busyloop: Block and return all the elements in a busy loop.
:param queue: Return only jobs in the specified queue.
:param state: Must be a list - Return jobs in the specified state.
Can be used multiple times for a logic OR.
:param reply: None or string {"all", "id"} - Job reply type. Type can
be all or id. Default is to report just the job ID. If all is
specified the full job state is returned like for the SHOW command.
|
[
"Iterate",
"all",
"the",
"existing",
"jobs",
"in",
"the",
"local",
"node",
"."
] |
ea5ce1576b66398c1cce32cad0f15709b1ea8df8
|
https://github.com/ybrs/pydisque/blob/ea5ce1576b66398c1cce32cad0f15709b1ea8df8/pydisque/client.py#L479-L505
|
train
|
inveniosoftware/invenio-search
|
invenio_search/utils.py
|
build_index_name
|
def build_index_name(app, *parts):
"""Build an index name from parts.
:param parts: Parts that should be combined to make an index name.
"""
base_index = os.path.splitext(
'-'.join([part for part in parts if part])
)[0]
return prefix_index(app=app, index=base_index)
|
python
|
def build_index_name(app, *parts):
"""Build an index name from parts.
:param parts: Parts that should be combined to make an index name.
"""
base_index = os.path.splitext(
'-'.join([part for part in parts if part])
)[0]
return prefix_index(app=app, index=base_index)
|
[
"def",
"build_index_name",
"(",
"app",
",",
"*",
"parts",
")",
":",
"base_index",
"=",
"os",
".",
"path",
".",
"splitext",
"(",
"'-'",
".",
"join",
"(",
"[",
"part",
"for",
"part",
"in",
"parts",
"if",
"part",
"]",
")",
")",
"[",
"0",
"]",
"return",
"prefix_index",
"(",
"app",
"=",
"app",
",",
"index",
"=",
"base_index",
")"
] |
Build an index name from parts.
:param parts: Parts that should be combined to make an index name.
|
[
"Build",
"an",
"index",
"name",
"from",
"parts",
"."
] |
19c073d608d4c811f1c5aecb6622402d39715228
|
https://github.com/inveniosoftware/invenio-search/blob/19c073d608d4c811f1c5aecb6622402d39715228/invenio_search/utils.py#L27-L36
|
train
|
inveniosoftware/invenio-search
|
invenio_search/utils.py
|
schema_to_index
|
def schema_to_index(schema, index_names=None):
"""Get index/doc_type given a schema URL.
:param schema: The schema name
:param index_names: A list of index name.
:returns: A tuple containing (index, doc_type).
"""
parts = schema.split('/')
doc_type = os.path.splitext(parts[-1])
if doc_type[1] not in {'.json', }:
return (None, None)
if index_names is None:
return (build_index_name(current_app, *parts), doc_type[0])
for start in range(len(parts)):
index_name = build_index_name(current_app, *parts[start:])
if index_name in index_names:
return (index_name, doc_type[0])
return (None, None)
|
python
|
def schema_to_index(schema, index_names=None):
"""Get index/doc_type given a schema URL.
:param schema: The schema name
:param index_names: A list of index name.
:returns: A tuple containing (index, doc_type).
"""
parts = schema.split('/')
doc_type = os.path.splitext(parts[-1])
if doc_type[1] not in {'.json', }:
return (None, None)
if index_names is None:
return (build_index_name(current_app, *parts), doc_type[0])
for start in range(len(parts)):
index_name = build_index_name(current_app, *parts[start:])
if index_name in index_names:
return (index_name, doc_type[0])
return (None, None)
|
[
"def",
"schema_to_index",
"(",
"schema",
",",
"index_names",
"=",
"None",
")",
":",
"parts",
"=",
"schema",
".",
"split",
"(",
"'/'",
")",
"doc_type",
"=",
"os",
".",
"path",
".",
"splitext",
"(",
"parts",
"[",
"-",
"1",
"]",
")",
"if",
"doc_type",
"[",
"1",
"]",
"not",
"in",
"{",
"'.json'",
",",
"}",
":",
"return",
"(",
"None",
",",
"None",
")",
"if",
"index_names",
"is",
"None",
":",
"return",
"(",
"build_index_name",
"(",
"current_app",
",",
"*",
"parts",
")",
",",
"doc_type",
"[",
"0",
"]",
")",
"for",
"start",
"in",
"range",
"(",
"len",
"(",
"parts",
")",
")",
":",
"index_name",
"=",
"build_index_name",
"(",
"current_app",
",",
"*",
"parts",
"[",
"start",
":",
"]",
")",
"if",
"index_name",
"in",
"index_names",
":",
"return",
"(",
"index_name",
",",
"doc_type",
"[",
"0",
"]",
")",
"return",
"(",
"None",
",",
"None",
")"
] |
Get index/doc_type given a schema URL.
:param schema: The schema name
:param index_names: A list of index name.
:returns: A tuple containing (index, doc_type).
|
[
"Get",
"index",
"/",
"doc_type",
"given",
"a",
"schema",
"URL",
"."
] |
19c073d608d4c811f1c5aecb6622402d39715228
|
https://github.com/inveniosoftware/invenio-search/blob/19c073d608d4c811f1c5aecb6622402d39715228/invenio_search/utils.py#L39-L60
|
train
|
inveniosoftware/invenio-search
|
invenio_search/cli.py
|
es_version_check
|
def es_version_check(f):
"""Decorator to check Elasticsearch version."""
@wraps(f)
def inner(*args, **kwargs):
cluster_ver = current_search.cluster_version[0]
client_ver = ES_VERSION[0]
if cluster_ver != client_ver:
raise click.ClickException(
'Elasticsearch version mismatch. Invenio was installed with '
'Elasticsearch v{client_ver}.x support, but the cluster runs '
'Elasticsearch v{cluster_ver}.x.'.format(
client_ver=client_ver,
cluster_ver=cluster_ver,
))
return f(*args, **kwargs)
return inner
|
python
|
def es_version_check(f):
"""Decorator to check Elasticsearch version."""
@wraps(f)
def inner(*args, **kwargs):
cluster_ver = current_search.cluster_version[0]
client_ver = ES_VERSION[0]
if cluster_ver != client_ver:
raise click.ClickException(
'Elasticsearch version mismatch. Invenio was installed with '
'Elasticsearch v{client_ver}.x support, but the cluster runs '
'Elasticsearch v{cluster_ver}.x.'.format(
client_ver=client_ver,
cluster_ver=cluster_ver,
))
return f(*args, **kwargs)
return inner
|
[
"def",
"es_version_check",
"(",
"f",
")",
":",
"@",
"wraps",
"(",
"f",
")",
"def",
"inner",
"(",
"*",
"args",
",",
"*",
"*",
"kwargs",
")",
":",
"cluster_ver",
"=",
"current_search",
".",
"cluster_version",
"[",
"0",
"]",
"client_ver",
"=",
"ES_VERSION",
"[",
"0",
"]",
"if",
"cluster_ver",
"!=",
"client_ver",
":",
"raise",
"click",
".",
"ClickException",
"(",
"'Elasticsearch version mismatch. Invenio was installed with '",
"'Elasticsearch v{client_ver}.x support, but the cluster runs '",
"'Elasticsearch v{cluster_ver}.x.'",
".",
"format",
"(",
"client_ver",
"=",
"client_ver",
",",
"cluster_ver",
"=",
"cluster_ver",
",",
")",
")",
"return",
"f",
"(",
"*",
"args",
",",
"*",
"*",
"kwargs",
")",
"return",
"inner"
] |
Decorator to check Elasticsearch version.
|
[
"Decorator",
"to",
"check",
"Elasticsearch",
"version",
"."
] |
19c073d608d4c811f1c5aecb6622402d39715228
|
https://github.com/inveniosoftware/invenio-search/blob/19c073d608d4c811f1c5aecb6622402d39715228/invenio_search/cli.py#L30-L45
|
train
|
inveniosoftware/invenio-search
|
invenio_search/cli.py
|
init
|
def init(force):
"""Initialize registered aliases and mappings."""
click.secho('Creating indexes...', fg='green', bold=True, file=sys.stderr)
with click.progressbar(
current_search.create(ignore=[400] if force else None),
length=current_search.number_of_indexes) as bar:
for name, response in bar:
bar.label = name
click.secho('Putting templates...', fg='green', bold=True, file=sys.stderr)
with click.progressbar(
current_search.put_templates(ignore=[400] if force else None),
length=len(current_search.templates.keys())) as bar:
for response in bar:
bar.label = response
|
python
|
def init(force):
"""Initialize registered aliases and mappings."""
click.secho('Creating indexes...', fg='green', bold=True, file=sys.stderr)
with click.progressbar(
current_search.create(ignore=[400] if force else None),
length=current_search.number_of_indexes) as bar:
for name, response in bar:
bar.label = name
click.secho('Putting templates...', fg='green', bold=True, file=sys.stderr)
with click.progressbar(
current_search.put_templates(ignore=[400] if force else None),
length=len(current_search.templates.keys())) as bar:
for response in bar:
bar.label = response
|
[
"def",
"init",
"(",
"force",
")",
":",
"click",
".",
"secho",
"(",
"'Creating indexes...'",
",",
"fg",
"=",
"'green'",
",",
"bold",
"=",
"True",
",",
"file",
"=",
"sys",
".",
"stderr",
")",
"with",
"click",
".",
"progressbar",
"(",
"current_search",
".",
"create",
"(",
"ignore",
"=",
"[",
"400",
"]",
"if",
"force",
"else",
"None",
")",
",",
"length",
"=",
"current_search",
".",
"number_of_indexes",
")",
"as",
"bar",
":",
"for",
"name",
",",
"response",
"in",
"bar",
":",
"bar",
".",
"label",
"=",
"name",
"click",
".",
"secho",
"(",
"'Putting templates...'",
",",
"fg",
"=",
"'green'",
",",
"bold",
"=",
"True",
",",
"file",
"=",
"sys",
".",
"stderr",
")",
"with",
"click",
".",
"progressbar",
"(",
"current_search",
".",
"put_templates",
"(",
"ignore",
"=",
"[",
"400",
"]",
"if",
"force",
"else",
"None",
")",
",",
"length",
"=",
"len",
"(",
"current_search",
".",
"templates",
".",
"keys",
"(",
")",
")",
")",
"as",
"bar",
":",
"for",
"response",
"in",
"bar",
":",
"bar",
".",
"label",
"=",
"response"
] |
Initialize registered aliases and mappings.
|
[
"Initialize",
"registered",
"aliases",
"and",
"mappings",
"."
] |
19c073d608d4c811f1c5aecb6622402d39715228
|
https://github.com/inveniosoftware/invenio-search/blob/19c073d608d4c811f1c5aecb6622402d39715228/invenio_search/cli.py#L68-L81
|
train
|
inveniosoftware/invenio-search
|
invenio_search/cli.py
|
destroy
|
def destroy(force):
"""Destroy all indexes."""
click.secho('Destroying indexes...', fg='red', bold=True, file=sys.stderr)
with click.progressbar(
current_search.delete(ignore=[400, 404] if force else None),
length=current_search.number_of_indexes) as bar:
for name, response in bar:
bar.label = name
|
python
|
def destroy(force):
"""Destroy all indexes."""
click.secho('Destroying indexes...', fg='red', bold=True, file=sys.stderr)
with click.progressbar(
current_search.delete(ignore=[400, 404] if force else None),
length=current_search.number_of_indexes) as bar:
for name, response in bar:
bar.label = name
|
[
"def",
"destroy",
"(",
"force",
")",
":",
"click",
".",
"secho",
"(",
"'Destroying indexes...'",
",",
"fg",
"=",
"'red'",
",",
"bold",
"=",
"True",
",",
"file",
"=",
"sys",
".",
"stderr",
")",
"with",
"click",
".",
"progressbar",
"(",
"current_search",
".",
"delete",
"(",
"ignore",
"=",
"[",
"400",
",",
"404",
"]",
"if",
"force",
"else",
"None",
")",
",",
"length",
"=",
"current_search",
".",
"number_of_indexes",
")",
"as",
"bar",
":",
"for",
"name",
",",
"response",
"in",
"bar",
":",
"bar",
".",
"label",
"=",
"name"
] |
Destroy all indexes.
|
[
"Destroy",
"all",
"indexes",
"."
] |
19c073d608d4c811f1c5aecb6622402d39715228
|
https://github.com/inveniosoftware/invenio-search/blob/19c073d608d4c811f1c5aecb6622402d39715228/invenio_search/cli.py#L91-L98
|
train
|
inveniosoftware/invenio-search
|
invenio_search/cli.py
|
create
|
def create(index_name, body, force, verbose):
"""Create a new index."""
result = current_search_client.indices.create(
index=index_name,
body=json.load(body),
ignore=[400] if force else None,
)
if verbose:
click.echo(json.dumps(result))
|
python
|
def create(index_name, body, force, verbose):
"""Create a new index."""
result = current_search_client.indices.create(
index=index_name,
body=json.load(body),
ignore=[400] if force else None,
)
if verbose:
click.echo(json.dumps(result))
|
[
"def",
"create",
"(",
"index_name",
",",
"body",
",",
"force",
",",
"verbose",
")",
":",
"result",
"=",
"current_search_client",
".",
"indices",
".",
"create",
"(",
"index",
"=",
"index_name",
",",
"body",
"=",
"json",
".",
"load",
"(",
"body",
")",
",",
"ignore",
"=",
"[",
"400",
"]",
"if",
"force",
"else",
"None",
",",
")",
"if",
"verbose",
":",
"click",
".",
"echo",
"(",
"json",
".",
"dumps",
"(",
"result",
")",
")"
] |
Create a new index.
|
[
"Create",
"a",
"new",
"index",
"."
] |
19c073d608d4c811f1c5aecb6622402d39715228
|
https://github.com/inveniosoftware/invenio-search/blob/19c073d608d4c811f1c5aecb6622402d39715228/invenio_search/cli.py#L108-L116
|
train
|
inveniosoftware/invenio-search
|
invenio_search/cli.py
|
list_cmd
|
def list_cmd(only_active, only_aliases, verbose):
"""List indices."""
def _tree_print(d, rec_list=None, verbose=False, indent=2):
# Note that on every recursion rec_list is copied,
# which might not be very effective for very deep dictionaries.
rec_list = rec_list or []
for idx, key in enumerate(sorted(d)):
line = (['│' + ' ' * indent
if i == 1 else ' ' * (indent+1) for i in rec_list])
line.append('└──' if len(d)-1 == idx else '├──')
click.echo(''.join(line), nl=False)
if isinstance(d[key], dict):
click.echo(key)
new_rec_list = rec_list + [0 if len(d)-1 == idx else 1]
_tree_print(d[key], new_rec_list, verbose)
else:
leaf_txt = '{} -> {}'.format(key, d[key]) if verbose else key
click.echo(leaf_txt)
aliases = (current_search.active_aliases
if only_active else current_search.aliases)
active_aliases = current_search.active_aliases
if only_aliases:
click.echo(json.dumps(list((aliases.keys())), indent=4))
else:
# Mark active indices for printout
aliases = {(k + (' *' if k in active_aliases else '')): v
for k, v in aliases.items()}
click.echo(_tree_print(aliases, verbose=verbose))
|
python
|
def list_cmd(only_active, only_aliases, verbose):
"""List indices."""
def _tree_print(d, rec_list=None, verbose=False, indent=2):
# Note that on every recursion rec_list is copied,
# which might not be very effective for very deep dictionaries.
rec_list = rec_list or []
for idx, key in enumerate(sorted(d)):
line = (['│' + ' ' * indent
if i == 1 else ' ' * (indent+1) for i in rec_list])
line.append('└──' if len(d)-1 == idx else '├──')
click.echo(''.join(line), nl=False)
if isinstance(d[key], dict):
click.echo(key)
new_rec_list = rec_list + [0 if len(d)-1 == idx else 1]
_tree_print(d[key], new_rec_list, verbose)
else:
leaf_txt = '{} -> {}'.format(key, d[key]) if verbose else key
click.echo(leaf_txt)
aliases = (current_search.active_aliases
if only_active else current_search.aliases)
active_aliases = current_search.active_aliases
if only_aliases:
click.echo(json.dumps(list((aliases.keys())), indent=4))
else:
# Mark active indices for printout
aliases = {(k + (' *' if k in active_aliases else '')): v
for k, v in aliases.items()}
click.echo(_tree_print(aliases, verbose=verbose))
|
[
"def",
"list_cmd",
"(",
"only_active",
",",
"only_aliases",
",",
"verbose",
")",
":",
"def",
"_tree_print",
"(",
"d",
",",
"rec_list",
"=",
"None",
",",
"verbose",
"=",
"False",
",",
"indent",
"=",
"2",
")",
":",
"# Note that on every recursion rec_list is copied,",
"# which might not be very effective for very deep dictionaries.",
"rec_list",
"=",
"rec_list",
"or",
"[",
"]",
"for",
"idx",
",",
"key",
"in",
"enumerate",
"(",
"sorted",
"(",
"d",
")",
")",
":",
"line",
"=",
"(",
"[",
"'│' +",
"'",
"' *",
"i",
"dent",
"if",
"i",
"==",
"1",
"else",
"' '",
"*",
"(",
"indent",
"+",
"1",
")",
"for",
"i",
"in",
"rec_list",
"]",
")",
"line",
".",
"append",
"(",
"'└──' if le",
"(d",
"-1 ",
"=",
"=",
" ",
"i",
"d",
" e",
"se ",
"├──'",
"",
"",
"click",
".",
"echo",
"(",
"''",
".",
"join",
"(",
"line",
")",
",",
"nl",
"=",
"False",
")",
"if",
"isinstance",
"(",
"d",
"[",
"key",
"]",
",",
"dict",
")",
":",
"click",
".",
"echo",
"(",
"key",
")",
"new_rec_list",
"=",
"rec_list",
"+",
"[",
"0",
"if",
"len",
"(",
"d",
")",
"-",
"1",
"==",
"idx",
"else",
"1",
"]",
"_tree_print",
"(",
"d",
"[",
"key",
"]",
",",
"new_rec_list",
",",
"verbose",
")",
"else",
":",
"leaf_txt",
"=",
"'{} -> {}'",
".",
"format",
"(",
"key",
",",
"d",
"[",
"key",
"]",
")",
"if",
"verbose",
"else",
"key",
"click",
".",
"echo",
"(",
"leaf_txt",
")",
"aliases",
"=",
"(",
"current_search",
".",
"active_aliases",
"if",
"only_active",
"else",
"current_search",
".",
"aliases",
")",
"active_aliases",
"=",
"current_search",
".",
"active_aliases",
"if",
"only_aliases",
":",
"click",
".",
"echo",
"(",
"json",
".",
"dumps",
"(",
"list",
"(",
"(",
"aliases",
".",
"keys",
"(",
")",
")",
")",
",",
"indent",
"=",
"4",
")",
")",
"else",
":",
"# Mark active indices for printout",
"aliases",
"=",
"{",
"(",
"k",
"+",
"(",
"' *'",
"if",
"k",
"in",
"active_aliases",
"else",
"''",
")",
")",
":",
"v",
"for",
"k",
",",
"v",
"in",
"aliases",
".",
"items",
"(",
")",
"}",
"click",
".",
"echo",
"(",
"_tree_print",
"(",
"aliases",
",",
"verbose",
"=",
"verbose",
")",
")"
] |
List indices.
|
[
"List",
"indices",
"."
] |
19c073d608d4c811f1c5aecb6622402d39715228
|
https://github.com/inveniosoftware/invenio-search/blob/19c073d608d4c811f1c5aecb6622402d39715228/invenio_search/cli.py#L124-L153
|
train
|
inveniosoftware/invenio-search
|
invenio_search/cli.py
|
delete
|
def delete(index_name, force, verbose):
"""Delete index by its name."""
result = current_search_client.indices.delete(
index=index_name,
ignore=[400, 404] if force else None,
)
if verbose:
click.echo(json.dumps(result))
|
python
|
def delete(index_name, force, verbose):
"""Delete index by its name."""
result = current_search_client.indices.delete(
index=index_name,
ignore=[400, 404] if force else None,
)
if verbose:
click.echo(json.dumps(result))
|
[
"def",
"delete",
"(",
"index_name",
",",
"force",
",",
"verbose",
")",
":",
"result",
"=",
"current_search_client",
".",
"indices",
".",
"delete",
"(",
"index",
"=",
"index_name",
",",
"ignore",
"=",
"[",
"400",
",",
"404",
"]",
"if",
"force",
"else",
"None",
",",
")",
"if",
"verbose",
":",
"click",
".",
"echo",
"(",
"json",
".",
"dumps",
"(",
"result",
")",
")"
] |
Delete index by its name.
|
[
"Delete",
"index",
"by",
"its",
"name",
"."
] |
19c073d608d4c811f1c5aecb6622402d39715228
|
https://github.com/inveniosoftware/invenio-search/blob/19c073d608d4c811f1c5aecb6622402d39715228/invenio_search/cli.py#L165-L172
|
train
|
inveniosoftware/invenio-search
|
invenio_search/cli.py
|
put
|
def put(index_name, doc_type, identifier, body, force, verbose):
"""Index input data."""
result = current_search_client.index(
index=index_name,
doc_type=doc_type or index_name,
id=identifier,
body=json.load(body),
op_type='index' if force or identifier is None else 'create',
)
if verbose:
click.echo(json.dumps(result))
|
python
|
def put(index_name, doc_type, identifier, body, force, verbose):
"""Index input data."""
result = current_search_client.index(
index=index_name,
doc_type=doc_type or index_name,
id=identifier,
body=json.load(body),
op_type='index' if force or identifier is None else 'create',
)
if verbose:
click.echo(json.dumps(result))
|
[
"def",
"put",
"(",
"index_name",
",",
"doc_type",
",",
"identifier",
",",
"body",
",",
"force",
",",
"verbose",
")",
":",
"result",
"=",
"current_search_client",
".",
"index",
"(",
"index",
"=",
"index_name",
",",
"doc_type",
"=",
"doc_type",
"or",
"index_name",
",",
"id",
"=",
"identifier",
",",
"body",
"=",
"json",
".",
"load",
"(",
"body",
")",
",",
"op_type",
"=",
"'index'",
"if",
"force",
"or",
"identifier",
"is",
"None",
"else",
"'create'",
",",
")",
"if",
"verbose",
":",
"click",
".",
"echo",
"(",
"json",
".",
"dumps",
"(",
"result",
")",
")"
] |
Index input data.
|
[
"Index",
"input",
"data",
"."
] |
19c073d608d4c811f1c5aecb6622402d39715228
|
https://github.com/inveniosoftware/invenio-search/blob/19c073d608d4c811f1c5aecb6622402d39715228/invenio_search/cli.py#L184-L194
|
train
|
inveniosoftware/invenio-search
|
invenio_search/api.py
|
RecordsSearch.get_records
|
def get_records(self, ids):
"""Return records by their identifiers.
:param ids: A list of record identifier.
:returns: A list of records.
"""
return self.query(Ids(values=[str(id_) for id_ in ids]))
|
python
|
def get_records(self, ids):
"""Return records by their identifiers.
:param ids: A list of record identifier.
:returns: A list of records.
"""
return self.query(Ids(values=[str(id_) for id_ in ids]))
|
[
"def",
"get_records",
"(",
"self",
",",
"ids",
")",
":",
"return",
"self",
".",
"query",
"(",
"Ids",
"(",
"values",
"=",
"[",
"str",
"(",
"id_",
")",
"for",
"id_",
"in",
"ids",
"]",
")",
")"
] |
Return records by their identifiers.
:param ids: A list of record identifier.
:returns: A list of records.
|
[
"Return",
"records",
"by",
"their",
"identifiers",
"."
] |
19c073d608d4c811f1c5aecb6622402d39715228
|
https://github.com/inveniosoftware/invenio-search/blob/19c073d608d4c811f1c5aecb6622402d39715228/invenio_search/api.py#L120-L126
|
train
|
inveniosoftware/invenio-search
|
invenio_search/api.py
|
RecordsSearch.faceted_search
|
def faceted_search(cls, query=None, filters=None, search=None):
"""Return faceted search instance with defaults set.
:param query: Elastic DSL query object (``Q``).
:param filters: Dictionary with selected facet values.
:param search: An instance of ``Search`` class. (default: ``cls()``).
"""
search_ = search or cls()
class RecordsFacetedSearch(FacetedSearch):
"""Pass defaults from ``cls.Meta`` object."""
index = prefix_index(app=current_app, index=search_._index[0])
doc_types = getattr(search_.Meta, 'doc_types', ['_all'])
fields = getattr(search_.Meta, 'fields', ('*', ))
facets = getattr(search_.Meta, 'facets', {})
def search(self):
"""Use ``search`` or ``cls()`` instead of default Search."""
# Later versions of `elasticsearch-dsl` (>=5.1.0) changed the
# Elasticsearch FacetedResponse class constructor signature.
if ES_VERSION[0] > 2:
return search_.response_class(FacetedResponse)
return search_.response_class(partial(FacetedResponse, self))
return RecordsFacetedSearch(query=query, filters=filters or {})
|
python
|
def faceted_search(cls, query=None, filters=None, search=None):
"""Return faceted search instance with defaults set.
:param query: Elastic DSL query object (``Q``).
:param filters: Dictionary with selected facet values.
:param search: An instance of ``Search`` class. (default: ``cls()``).
"""
search_ = search or cls()
class RecordsFacetedSearch(FacetedSearch):
"""Pass defaults from ``cls.Meta`` object."""
index = prefix_index(app=current_app, index=search_._index[0])
doc_types = getattr(search_.Meta, 'doc_types', ['_all'])
fields = getattr(search_.Meta, 'fields', ('*', ))
facets = getattr(search_.Meta, 'facets', {})
def search(self):
"""Use ``search`` or ``cls()`` instead of default Search."""
# Later versions of `elasticsearch-dsl` (>=5.1.0) changed the
# Elasticsearch FacetedResponse class constructor signature.
if ES_VERSION[0] > 2:
return search_.response_class(FacetedResponse)
return search_.response_class(partial(FacetedResponse, self))
return RecordsFacetedSearch(query=query, filters=filters or {})
|
[
"def",
"faceted_search",
"(",
"cls",
",",
"query",
"=",
"None",
",",
"filters",
"=",
"None",
",",
"search",
"=",
"None",
")",
":",
"search_",
"=",
"search",
"or",
"cls",
"(",
")",
"class",
"RecordsFacetedSearch",
"(",
"FacetedSearch",
")",
":",
"\"\"\"Pass defaults from ``cls.Meta`` object.\"\"\"",
"index",
"=",
"prefix_index",
"(",
"app",
"=",
"current_app",
",",
"index",
"=",
"search_",
".",
"_index",
"[",
"0",
"]",
")",
"doc_types",
"=",
"getattr",
"(",
"search_",
".",
"Meta",
",",
"'doc_types'",
",",
"[",
"'_all'",
"]",
")",
"fields",
"=",
"getattr",
"(",
"search_",
".",
"Meta",
",",
"'fields'",
",",
"(",
"'*'",
",",
")",
")",
"facets",
"=",
"getattr",
"(",
"search_",
".",
"Meta",
",",
"'facets'",
",",
"{",
"}",
")",
"def",
"search",
"(",
"self",
")",
":",
"\"\"\"Use ``search`` or ``cls()`` instead of default Search.\"\"\"",
"# Later versions of `elasticsearch-dsl` (>=5.1.0) changed the",
"# Elasticsearch FacetedResponse class constructor signature.",
"if",
"ES_VERSION",
"[",
"0",
"]",
">",
"2",
":",
"return",
"search_",
".",
"response_class",
"(",
"FacetedResponse",
")",
"return",
"search_",
".",
"response_class",
"(",
"partial",
"(",
"FacetedResponse",
",",
"self",
")",
")",
"return",
"RecordsFacetedSearch",
"(",
"query",
"=",
"query",
",",
"filters",
"=",
"filters",
"or",
"{",
"}",
")"
] |
Return faceted search instance with defaults set.
:param query: Elastic DSL query object (``Q``).
:param filters: Dictionary with selected facet values.
:param search: An instance of ``Search`` class. (default: ``cls()``).
|
[
"Return",
"faceted",
"search",
"instance",
"with",
"defaults",
"set",
"."
] |
19c073d608d4c811f1c5aecb6622402d39715228
|
https://github.com/inveniosoftware/invenio-search/blob/19c073d608d4c811f1c5aecb6622402d39715228/invenio_search/api.py#L129-L155
|
train
|
inveniosoftware/invenio-search
|
invenio_search/api.py
|
RecordsSearch.with_preference_param
|
def with_preference_param(self):
"""Add the preference param to the ES request and return a new Search.
The preference param avoids the bouncing effect with multiple
replicas, documented on ES documentation.
See: https://www.elastic.co/guide/en/elasticsearch/guide/current
/_search_options.html#_preference for more information.
"""
user_hash = self._get_user_hash()
if user_hash:
return self.params(preference=user_hash)
return self
|
python
|
def with_preference_param(self):
"""Add the preference param to the ES request and return a new Search.
The preference param avoids the bouncing effect with multiple
replicas, documented on ES documentation.
See: https://www.elastic.co/guide/en/elasticsearch/guide/current
/_search_options.html#_preference for more information.
"""
user_hash = self._get_user_hash()
if user_hash:
return self.params(preference=user_hash)
return self
|
[
"def",
"with_preference_param",
"(",
"self",
")",
":",
"user_hash",
"=",
"self",
".",
"_get_user_hash",
"(",
")",
"if",
"user_hash",
":",
"return",
"self",
".",
"params",
"(",
"preference",
"=",
"user_hash",
")",
"return",
"self"
] |
Add the preference param to the ES request and return a new Search.
The preference param avoids the bouncing effect with multiple
replicas, documented on ES documentation.
See: https://www.elastic.co/guide/en/elasticsearch/guide/current
/_search_options.html#_preference for more information.
|
[
"Add",
"the",
"preference",
"param",
"to",
"the",
"ES",
"request",
"and",
"return",
"a",
"new",
"Search",
"."
] |
19c073d608d4c811f1c5aecb6622402d39715228
|
https://github.com/inveniosoftware/invenio-search/blob/19c073d608d4c811f1c5aecb6622402d39715228/invenio_search/api.py#L157-L168
|
train
|
inveniosoftware/invenio-search
|
invenio_search/api.py
|
RecordsSearch._get_user_agent
|
def _get_user_agent(self):
"""Retrieve the request's User-Agent, if available.
Taken from Flask Login utils.py.
"""
user_agent = request.headers.get('User-Agent')
if user_agent:
user_agent = user_agent.encode('utf-8')
return user_agent or ''
|
python
|
def _get_user_agent(self):
"""Retrieve the request's User-Agent, if available.
Taken from Flask Login utils.py.
"""
user_agent = request.headers.get('User-Agent')
if user_agent:
user_agent = user_agent.encode('utf-8')
return user_agent or ''
|
[
"def",
"_get_user_agent",
"(",
"self",
")",
":",
"user_agent",
"=",
"request",
".",
"headers",
".",
"get",
"(",
"'User-Agent'",
")",
"if",
"user_agent",
":",
"user_agent",
"=",
"user_agent",
".",
"encode",
"(",
"'utf-8'",
")",
"return",
"user_agent",
"or",
"''"
] |
Retrieve the request's User-Agent, if available.
Taken from Flask Login utils.py.
|
[
"Retrieve",
"the",
"request",
"s",
"User",
"-",
"Agent",
"if",
"available",
"."
] |
19c073d608d4c811f1c5aecb6622402d39715228
|
https://github.com/inveniosoftware/invenio-search/blob/19c073d608d4c811f1c5aecb6622402d39715228/invenio_search/api.py#L170-L178
|
train
|
inveniosoftware/invenio-search
|
invenio_search/api.py
|
RecordsSearch._get_user_hash
|
def _get_user_hash(self):
"""Calculate a digest based on request's User-Agent and IP address."""
if request:
user_hash = '{ip}-{ua}'.format(ip=request.remote_addr,
ua=self._get_user_agent())
alg = hashlib.md5()
alg.update(user_hash.encode('utf8'))
return alg.hexdigest()
return None
|
python
|
def _get_user_hash(self):
"""Calculate a digest based on request's User-Agent and IP address."""
if request:
user_hash = '{ip}-{ua}'.format(ip=request.remote_addr,
ua=self._get_user_agent())
alg = hashlib.md5()
alg.update(user_hash.encode('utf8'))
return alg.hexdigest()
return None
|
[
"def",
"_get_user_hash",
"(",
"self",
")",
":",
"if",
"request",
":",
"user_hash",
"=",
"'{ip}-{ua}'",
".",
"format",
"(",
"ip",
"=",
"request",
".",
"remote_addr",
",",
"ua",
"=",
"self",
".",
"_get_user_agent",
"(",
")",
")",
"alg",
"=",
"hashlib",
".",
"md5",
"(",
")",
"alg",
".",
"update",
"(",
"user_hash",
".",
"encode",
"(",
"'utf8'",
")",
")",
"return",
"alg",
".",
"hexdigest",
"(",
")",
"return",
"None"
] |
Calculate a digest based on request's User-Agent and IP address.
|
[
"Calculate",
"a",
"digest",
"based",
"on",
"request",
"s",
"User",
"-",
"Agent",
"and",
"IP",
"address",
"."
] |
19c073d608d4c811f1c5aecb6622402d39715228
|
https://github.com/inveniosoftware/invenio-search/blob/19c073d608d4c811f1c5aecb6622402d39715228/invenio_search/api.py#L180-L188
|
train
|
chaoss/grimoirelab-sigils
|
src/migration/utils.py
|
beautify
|
def beautify(filename=None, json_str=None):
"""Beautify JSON string or file.
Keyword arguments:
:param filename: use its contents as json string instead of
json_str param.
:param json_str: json string to be beautified.
"""
if filename is not None:
with open(filename) as json_file:
json_str = json.load(json_file)
return json.dumps(json_str, indent=4, sort_keys=True)
|
python
|
def beautify(filename=None, json_str=None):
"""Beautify JSON string or file.
Keyword arguments:
:param filename: use its contents as json string instead of
json_str param.
:param json_str: json string to be beautified.
"""
if filename is not None:
with open(filename) as json_file:
json_str = json.load(json_file)
return json.dumps(json_str, indent=4, sort_keys=True)
|
[
"def",
"beautify",
"(",
"filename",
"=",
"None",
",",
"json_str",
"=",
"None",
")",
":",
"if",
"filename",
"is",
"not",
"None",
":",
"with",
"open",
"(",
"filename",
")",
"as",
"json_file",
":",
"json_str",
"=",
"json",
".",
"load",
"(",
"json_file",
")",
"return",
"json",
".",
"dumps",
"(",
"json_str",
",",
"indent",
"=",
"4",
",",
"sort_keys",
"=",
"True",
")"
] |
Beautify JSON string or file.
Keyword arguments:
:param filename: use its contents as json string instead of
json_str param.
:param json_str: json string to be beautified.
|
[
"Beautify",
"JSON",
"string",
"or",
"file",
"."
] |
33d395195acb316287143a535a2c6e4009bf0528
|
https://github.com/chaoss/grimoirelab-sigils/blob/33d395195acb316287143a535a2c6e4009bf0528/src/migration/utils.py#L26-L38
|
train
|
chaoss/grimoirelab-sigils
|
src/migration/utils.py
|
replace
|
def replace(pretty, old_str, new_str):
""" Replace strings giving some info on where
the replacement was done
"""
out_str = ''
line_number = 1
changes = 0
for line in pretty.splitlines(keepends=True):
new_line = line.replace(old_str, new_str)
if line.find(old_str) != -1:
logging.debug('%s', line_number)
logging.debug('< %s', line)
logging.debug('> %s', new_line)
changes += 1
out_str += new_line
line_number += 1
logging.info('Total changes(%s): %s', old_str, changes)
return out_str
|
python
|
def replace(pretty, old_str, new_str):
""" Replace strings giving some info on where
the replacement was done
"""
out_str = ''
line_number = 1
changes = 0
for line in pretty.splitlines(keepends=True):
new_line = line.replace(old_str, new_str)
if line.find(old_str) != -1:
logging.debug('%s', line_number)
logging.debug('< %s', line)
logging.debug('> %s', new_line)
changes += 1
out_str += new_line
line_number += 1
logging.info('Total changes(%s): %s', old_str, changes)
return out_str
|
[
"def",
"replace",
"(",
"pretty",
",",
"old_str",
",",
"new_str",
")",
":",
"out_str",
"=",
"''",
"line_number",
"=",
"1",
"changes",
"=",
"0",
"for",
"line",
"in",
"pretty",
".",
"splitlines",
"(",
"keepends",
"=",
"True",
")",
":",
"new_line",
"=",
"line",
".",
"replace",
"(",
"old_str",
",",
"new_str",
")",
"if",
"line",
".",
"find",
"(",
"old_str",
")",
"!=",
"-",
"1",
":",
"logging",
".",
"debug",
"(",
"'%s'",
",",
"line_number",
")",
"logging",
".",
"debug",
"(",
"'< %s'",
",",
"line",
")",
"logging",
".",
"debug",
"(",
"'> %s'",
",",
"new_line",
")",
"changes",
"+=",
"1",
"out_str",
"+=",
"new_line",
"line_number",
"+=",
"1",
"logging",
".",
"info",
"(",
"'Total changes(%s): %s'",
",",
"old_str",
",",
"changes",
")",
"return",
"out_str"
] |
Replace strings giving some info on where
the replacement was done
|
[
"Replace",
"strings",
"giving",
"some",
"info",
"on",
"where",
"the",
"replacement",
"was",
"done"
] |
33d395195acb316287143a535a2c6e4009bf0528
|
https://github.com/chaoss/grimoirelab-sigils/blob/33d395195acb316287143a535a2c6e4009bf0528/src/migration/utils.py#L40-L58
|
train
|
praw-dev/prawcore
|
examples/obtain_refresh_token.py
|
receive_connection
|
def receive_connection():
"""Wait for and then return a connected socket..
Opens a TCP connection on port 8080, and waits for a single client.
"""
server = socket.socket(socket.AF_INET, socket.SOCK_STREAM)
server.setsockopt(socket.SOL_SOCKET, socket.SO_REUSEADDR, 1)
server.bind(("localhost", 8080))
server.listen(1)
client = server.accept()[0]
server.close()
return client
|
python
|
def receive_connection():
"""Wait for and then return a connected socket..
Opens a TCP connection on port 8080, and waits for a single client.
"""
server = socket.socket(socket.AF_INET, socket.SOCK_STREAM)
server.setsockopt(socket.SOL_SOCKET, socket.SO_REUSEADDR, 1)
server.bind(("localhost", 8080))
server.listen(1)
client = server.accept()[0]
server.close()
return client
|
[
"def",
"receive_connection",
"(",
")",
":",
"server",
"=",
"socket",
".",
"socket",
"(",
"socket",
".",
"AF_INET",
",",
"socket",
".",
"SOCK_STREAM",
")",
"server",
".",
"setsockopt",
"(",
"socket",
".",
"SOL_SOCKET",
",",
"socket",
".",
"SO_REUSEADDR",
",",
"1",
")",
"server",
".",
"bind",
"(",
"(",
"\"localhost\"",
",",
"8080",
")",
")",
"server",
".",
"listen",
"(",
"1",
")",
"client",
"=",
"server",
".",
"accept",
"(",
")",
"[",
"0",
"]",
"server",
".",
"close",
"(",
")",
"return",
"client"
] |
Wait for and then return a connected socket..
Opens a TCP connection on port 8080, and waits for a single client.
|
[
"Wait",
"for",
"and",
"then",
"return",
"a",
"connected",
"socket",
".."
] |
b16ae88a1f2bf98095ed6fe64851cb7add7ed752
|
https://github.com/praw-dev/prawcore/blob/b16ae88a1f2bf98095ed6fe64851cb7add7ed752/examples/obtain_refresh_token.py#L19-L31
|
train
|
praw-dev/prawcore
|
examples/obtain_refresh_token.py
|
send_message
|
def send_message(client, message):
"""Send message to client and close the connection."""
print(message)
client.send("HTTP/1.1 200 OK\r\n\r\n{}".format(message).encode("utf-8"))
client.close()
|
python
|
def send_message(client, message):
"""Send message to client and close the connection."""
print(message)
client.send("HTTP/1.1 200 OK\r\n\r\n{}".format(message).encode("utf-8"))
client.close()
|
[
"def",
"send_message",
"(",
"client",
",",
"message",
")",
":",
"print",
"(",
"message",
")",
"client",
".",
"send",
"(",
"\"HTTP/1.1 200 OK\\r\\n\\r\\n{}\"",
".",
"format",
"(",
"message",
")",
".",
"encode",
"(",
"\"utf-8\"",
")",
")",
"client",
".",
"close",
"(",
")"
] |
Send message to client and close the connection.
|
[
"Send",
"message",
"to",
"client",
"and",
"close",
"the",
"connection",
"."
] |
b16ae88a1f2bf98095ed6fe64851cb7add7ed752
|
https://github.com/praw-dev/prawcore/blob/b16ae88a1f2bf98095ed6fe64851cb7add7ed752/examples/obtain_refresh_token.py#L34-L38
|
train
|
praw-dev/prawcore
|
examples/obtain_refresh_token.py
|
main
|
def main():
"""Provide the program's entry point when directly executed."""
if len(sys.argv) < 2:
print("Usage: {} SCOPE...".format(sys.argv[0]))
return 1
authenticator = prawcore.TrustedAuthenticator(
prawcore.Requestor("prawcore_refresh_token_example"),
os.environ["PRAWCORE_CLIENT_ID"],
os.environ["PRAWCORE_CLIENT_SECRET"],
os.environ["PRAWCORE_REDIRECT_URI"],
)
state = str(random.randint(0, 65000))
url = authenticator.authorize_url("permanent", sys.argv[1:], state)
print(url)
client = receive_connection()
data = client.recv(1024).decode("utf-8")
param_tokens = data.split(" ", 2)[1].split("?", 1)[1].split("&")
params = {
key: value
for (key, value) in [token.split("=") for token in param_tokens]
}
if state != params["state"]:
send_message(
client,
"State mismatch. Expected: {} Received: {}".format(
state, params["state"]
),
)
return 1
elif "error" in params:
send_message(client, params["error"])
return 1
authorizer = prawcore.Authorizer(authenticator)
authorizer.authorize(params["code"])
send_message(client, "Refresh token: {}".format(authorizer.refresh_token))
return 0
|
python
|
def main():
"""Provide the program's entry point when directly executed."""
if len(sys.argv) < 2:
print("Usage: {} SCOPE...".format(sys.argv[0]))
return 1
authenticator = prawcore.TrustedAuthenticator(
prawcore.Requestor("prawcore_refresh_token_example"),
os.environ["PRAWCORE_CLIENT_ID"],
os.environ["PRAWCORE_CLIENT_SECRET"],
os.environ["PRAWCORE_REDIRECT_URI"],
)
state = str(random.randint(0, 65000))
url = authenticator.authorize_url("permanent", sys.argv[1:], state)
print(url)
client = receive_connection()
data = client.recv(1024).decode("utf-8")
param_tokens = data.split(" ", 2)[1].split("?", 1)[1].split("&")
params = {
key: value
for (key, value) in [token.split("=") for token in param_tokens]
}
if state != params["state"]:
send_message(
client,
"State mismatch. Expected: {} Received: {}".format(
state, params["state"]
),
)
return 1
elif "error" in params:
send_message(client, params["error"])
return 1
authorizer = prawcore.Authorizer(authenticator)
authorizer.authorize(params["code"])
send_message(client, "Refresh token: {}".format(authorizer.refresh_token))
return 0
|
[
"def",
"main",
"(",
")",
":",
"if",
"len",
"(",
"sys",
".",
"argv",
")",
"<",
"2",
":",
"print",
"(",
"\"Usage: {} SCOPE...\"",
".",
"format",
"(",
"sys",
".",
"argv",
"[",
"0",
"]",
")",
")",
"return",
"1",
"authenticator",
"=",
"prawcore",
".",
"TrustedAuthenticator",
"(",
"prawcore",
".",
"Requestor",
"(",
"\"prawcore_refresh_token_example\"",
")",
",",
"os",
".",
"environ",
"[",
"\"PRAWCORE_CLIENT_ID\"",
"]",
",",
"os",
".",
"environ",
"[",
"\"PRAWCORE_CLIENT_SECRET\"",
"]",
",",
"os",
".",
"environ",
"[",
"\"PRAWCORE_REDIRECT_URI\"",
"]",
",",
")",
"state",
"=",
"str",
"(",
"random",
".",
"randint",
"(",
"0",
",",
"65000",
")",
")",
"url",
"=",
"authenticator",
".",
"authorize_url",
"(",
"\"permanent\"",
",",
"sys",
".",
"argv",
"[",
"1",
":",
"]",
",",
"state",
")",
"print",
"(",
"url",
")",
"client",
"=",
"receive_connection",
"(",
")",
"data",
"=",
"client",
".",
"recv",
"(",
"1024",
")",
".",
"decode",
"(",
"\"utf-8\"",
")",
"param_tokens",
"=",
"data",
".",
"split",
"(",
"\" \"",
",",
"2",
")",
"[",
"1",
"]",
".",
"split",
"(",
"\"?\"",
",",
"1",
")",
"[",
"1",
"]",
".",
"split",
"(",
"\"&\"",
")",
"params",
"=",
"{",
"key",
":",
"value",
"for",
"(",
"key",
",",
"value",
")",
"in",
"[",
"token",
".",
"split",
"(",
"\"=\"",
")",
"for",
"token",
"in",
"param_tokens",
"]",
"}",
"if",
"state",
"!=",
"params",
"[",
"\"state\"",
"]",
":",
"send_message",
"(",
"client",
",",
"\"State mismatch. Expected: {} Received: {}\"",
".",
"format",
"(",
"state",
",",
"params",
"[",
"\"state\"",
"]",
")",
",",
")",
"return",
"1",
"elif",
"\"error\"",
"in",
"params",
":",
"send_message",
"(",
"client",
",",
"params",
"[",
"\"error\"",
"]",
")",
"return",
"1",
"authorizer",
"=",
"prawcore",
".",
"Authorizer",
"(",
"authenticator",
")",
"authorizer",
".",
"authorize",
"(",
"params",
"[",
"\"code\"",
"]",
")",
"send_message",
"(",
"client",
",",
"\"Refresh token: {}\"",
".",
"format",
"(",
"authorizer",
".",
"refresh_token",
")",
")",
"return",
"0"
] |
Provide the program's entry point when directly executed.
|
[
"Provide",
"the",
"program",
"s",
"entry",
"point",
"when",
"directly",
"executed",
"."
] |
b16ae88a1f2bf98095ed6fe64851cb7add7ed752
|
https://github.com/praw-dev/prawcore/blob/b16ae88a1f2bf98095ed6fe64851cb7add7ed752/examples/obtain_refresh_token.py#L41-L82
|
train
|
neo4j-drivers/neobolt
|
neobolt/diagnostics.py
|
watch
|
def watch(logger_name, level=DEBUG, out=stdout):
""" Quick wrapper for using the Watcher.
:param logger_name: name of logger to watch
:param level: minimum log level to show (default INFO)
:param out: where to send output (default stdout)
:return: Watcher instance
"""
watcher = Watcher(logger_name)
watcher.watch(level, out)
return watcher
|
python
|
def watch(logger_name, level=DEBUG, out=stdout):
""" Quick wrapper for using the Watcher.
:param logger_name: name of logger to watch
:param level: minimum log level to show (default INFO)
:param out: where to send output (default stdout)
:return: Watcher instance
"""
watcher = Watcher(logger_name)
watcher.watch(level, out)
return watcher
|
[
"def",
"watch",
"(",
"logger_name",
",",
"level",
"=",
"DEBUG",
",",
"out",
"=",
"stdout",
")",
":",
"watcher",
"=",
"Watcher",
"(",
"logger_name",
")",
"watcher",
".",
"watch",
"(",
"level",
",",
"out",
")",
"return",
"watcher"
] |
Quick wrapper for using the Watcher.
:param logger_name: name of logger to watch
:param level: minimum log level to show (default INFO)
:param out: where to send output (default stdout)
:return: Watcher instance
|
[
"Quick",
"wrapper",
"for",
"using",
"the",
"Watcher",
"."
] |
724569d76e85777c4f5e30e8d0a18116bda4d8cd
|
https://github.com/neo4j-drivers/neobolt/blob/724569d76e85777c4f5e30e8d0a18116bda4d8cd/neobolt/diagnostics.py#L80-L90
|
train
|
neo4j-drivers/neobolt
|
neobolt/meta.py
|
get_user_agent
|
def get_user_agent():
""" Obtain the default user agent string sent to the server after
a successful handshake.
"""
from sys import platform, version_info
template = "neobolt/{} Python/{}.{}.{}-{}-{} ({})"
fields = (version,) + tuple(version_info) + (platform,)
return template.format(*fields)
|
python
|
def get_user_agent():
""" Obtain the default user agent string sent to the server after
a successful handshake.
"""
from sys import platform, version_info
template = "neobolt/{} Python/{}.{}.{}-{}-{} ({})"
fields = (version,) + tuple(version_info) + (platform,)
return template.format(*fields)
|
[
"def",
"get_user_agent",
"(",
")",
":",
"from",
"sys",
"import",
"platform",
",",
"version_info",
"template",
"=",
"\"neobolt/{} Python/{}.{}.{}-{}-{} ({})\"",
"fields",
"=",
"(",
"version",
",",
")",
"+",
"tuple",
"(",
"version_info",
")",
"+",
"(",
"platform",
",",
")",
"return",
"template",
".",
"format",
"(",
"*",
"fields",
")"
] |
Obtain the default user agent string sent to the server after
a successful handshake.
|
[
"Obtain",
"the",
"default",
"user",
"agent",
"string",
"sent",
"to",
"the",
"server",
"after",
"a",
"successful",
"handshake",
"."
] |
724569d76e85777c4f5e30e8d0a18116bda4d8cd
|
https://github.com/neo4j-drivers/neobolt/blob/724569d76e85777c4f5e30e8d0a18116bda4d8cd/neobolt/meta.py#L26-L33
|
train
|
neo4j-drivers/neobolt
|
neobolt/meta.py
|
import_best
|
def import_best(c_module, py_module):
""" Import the best available module,
with C preferred to pure Python.
"""
from importlib import import_module
from os import getenv
pure_python = getenv("PURE_PYTHON", "")
if pure_python:
return import_module(py_module)
else:
try:
return import_module(c_module)
except ImportError:
return import_module(py_module)
|
python
|
def import_best(c_module, py_module):
""" Import the best available module,
with C preferred to pure Python.
"""
from importlib import import_module
from os import getenv
pure_python = getenv("PURE_PYTHON", "")
if pure_python:
return import_module(py_module)
else:
try:
return import_module(c_module)
except ImportError:
return import_module(py_module)
|
[
"def",
"import_best",
"(",
"c_module",
",",
"py_module",
")",
":",
"from",
"importlib",
"import",
"import_module",
"from",
"os",
"import",
"getenv",
"pure_python",
"=",
"getenv",
"(",
"\"PURE_PYTHON\"",
",",
"\"\"",
")",
"if",
"pure_python",
":",
"return",
"import_module",
"(",
"py_module",
")",
"else",
":",
"try",
":",
"return",
"import_module",
"(",
"c_module",
")",
"except",
"ImportError",
":",
"return",
"import_module",
"(",
"py_module",
")"
] |
Import the best available module,
with C preferred to pure Python.
|
[
"Import",
"the",
"best",
"available",
"module",
"with",
"C",
"preferred",
"to",
"pure",
"Python",
"."
] |
724569d76e85777c4f5e30e8d0a18116bda4d8cd
|
https://github.com/neo4j-drivers/neobolt/blob/724569d76e85777c4f5e30e8d0a18116bda4d8cd/neobolt/meta.py#L58-L71
|
train
|
neo4j-drivers/neobolt
|
neobolt/types/__init__.py
|
PackStreamHydrator.hydrate
|
def hydrate(self, values):
""" Convert PackStream values into native values.
"""
def hydrate_(obj):
if isinstance(obj, Structure):
try:
f = self.hydration_functions[obj.tag]
except KeyError:
# If we don't recognise the structure type, just return it as-is
return obj
else:
return f(*map(hydrate_, obj.fields))
elif isinstance(obj, list):
return list(map(hydrate_, obj))
elif isinstance(obj, dict):
return {key: hydrate_(value) for key, value in obj.items()}
else:
return obj
return tuple(map(hydrate_, values))
|
python
|
def hydrate(self, values):
""" Convert PackStream values into native values.
"""
def hydrate_(obj):
if isinstance(obj, Structure):
try:
f = self.hydration_functions[obj.tag]
except KeyError:
# If we don't recognise the structure type, just return it as-is
return obj
else:
return f(*map(hydrate_, obj.fields))
elif isinstance(obj, list):
return list(map(hydrate_, obj))
elif isinstance(obj, dict):
return {key: hydrate_(value) for key, value in obj.items()}
else:
return obj
return tuple(map(hydrate_, values))
|
[
"def",
"hydrate",
"(",
"self",
",",
"values",
")",
":",
"def",
"hydrate_",
"(",
"obj",
")",
":",
"if",
"isinstance",
"(",
"obj",
",",
"Structure",
")",
":",
"try",
":",
"f",
"=",
"self",
".",
"hydration_functions",
"[",
"obj",
".",
"tag",
"]",
"except",
"KeyError",
":",
"# If we don't recognise the structure type, just return it as-is",
"return",
"obj",
"else",
":",
"return",
"f",
"(",
"*",
"map",
"(",
"hydrate_",
",",
"obj",
".",
"fields",
")",
")",
"elif",
"isinstance",
"(",
"obj",
",",
"list",
")",
":",
"return",
"list",
"(",
"map",
"(",
"hydrate_",
",",
"obj",
")",
")",
"elif",
"isinstance",
"(",
"obj",
",",
"dict",
")",
":",
"return",
"{",
"key",
":",
"hydrate_",
"(",
"value",
")",
"for",
"key",
",",
"value",
"in",
"obj",
".",
"items",
"(",
")",
"}",
"else",
":",
"return",
"obj",
"return",
"tuple",
"(",
"map",
"(",
"hydrate_",
",",
"values",
")",
")"
] |
Convert PackStream values into native values.
|
[
"Convert",
"PackStream",
"values",
"into",
"native",
"values",
"."
] |
724569d76e85777c4f5e30e8d0a18116bda4d8cd
|
https://github.com/neo4j-drivers/neobolt/blob/724569d76e85777c4f5e30e8d0a18116bda4d8cd/neobolt/types/__init__.py#L97-L117
|
train
|
praw-dev/prawcore
|
prawcore/auth.py
|
BaseAuthenticator.authorize_url
|
def authorize_url(self, duration, scopes, state, implicit=False):
"""Return the URL used out-of-band to grant access to your application.
:param duration: Either ``permanent`` or ``temporary``. ``temporary``
authorizations generate access tokens that last only 1
hour. ``permanent`` authorizations additionally generate a refresh
token that can be indefinitely used to generate new hour-long
access tokens. Only ``temporary`` can be specified if ``implicit``
is set to ``True``.
:param scopes: A list of OAuth scopes to request authorization for.
:param state: A string that will be reflected in the callback to
``redirect_uri``. This value should be temporarily unique to the
client for whom the URL was generated for.
:param implicit: (optional) Use the implicit grant flow (default:
False). This flow is only available for UntrustedAuthenticators.
"""
if self.redirect_uri is None:
raise InvalidInvocation("redirect URI not provided")
if implicit and not isinstance(self, UntrustedAuthenticator):
raise InvalidInvocation(
"Only UntrustedAuthentictor instances can "
"use the implicit grant flow."
)
if implicit and duration != "temporary":
raise InvalidInvocation(
"The implicit grant flow only supports "
"temporary access tokens."
)
params = {
"client_id": self.client_id,
"duration": duration,
"redirect_uri": self.redirect_uri,
"response_type": "token" if implicit else "code",
"scope": " ".join(scopes),
"state": state,
}
url = self._requestor.reddit_url + const.AUTHORIZATION_PATH
request = Request("GET", url, params=params)
return request.prepare().url
|
python
|
def authorize_url(self, duration, scopes, state, implicit=False):
"""Return the URL used out-of-band to grant access to your application.
:param duration: Either ``permanent`` or ``temporary``. ``temporary``
authorizations generate access tokens that last only 1
hour. ``permanent`` authorizations additionally generate a refresh
token that can be indefinitely used to generate new hour-long
access tokens. Only ``temporary`` can be specified if ``implicit``
is set to ``True``.
:param scopes: A list of OAuth scopes to request authorization for.
:param state: A string that will be reflected in the callback to
``redirect_uri``. This value should be temporarily unique to the
client for whom the URL was generated for.
:param implicit: (optional) Use the implicit grant flow (default:
False). This flow is only available for UntrustedAuthenticators.
"""
if self.redirect_uri is None:
raise InvalidInvocation("redirect URI not provided")
if implicit and not isinstance(self, UntrustedAuthenticator):
raise InvalidInvocation(
"Only UntrustedAuthentictor instances can "
"use the implicit grant flow."
)
if implicit and duration != "temporary":
raise InvalidInvocation(
"The implicit grant flow only supports "
"temporary access tokens."
)
params = {
"client_id": self.client_id,
"duration": duration,
"redirect_uri": self.redirect_uri,
"response_type": "token" if implicit else "code",
"scope": " ".join(scopes),
"state": state,
}
url = self._requestor.reddit_url + const.AUTHORIZATION_PATH
request = Request("GET", url, params=params)
return request.prepare().url
|
[
"def",
"authorize_url",
"(",
"self",
",",
"duration",
",",
"scopes",
",",
"state",
",",
"implicit",
"=",
"False",
")",
":",
"if",
"self",
".",
"redirect_uri",
"is",
"None",
":",
"raise",
"InvalidInvocation",
"(",
"\"redirect URI not provided\"",
")",
"if",
"implicit",
"and",
"not",
"isinstance",
"(",
"self",
",",
"UntrustedAuthenticator",
")",
":",
"raise",
"InvalidInvocation",
"(",
"\"Only UntrustedAuthentictor instances can \"",
"\"use the implicit grant flow.\"",
")",
"if",
"implicit",
"and",
"duration",
"!=",
"\"temporary\"",
":",
"raise",
"InvalidInvocation",
"(",
"\"The implicit grant flow only supports \"",
"\"temporary access tokens.\"",
")",
"params",
"=",
"{",
"\"client_id\"",
":",
"self",
".",
"client_id",
",",
"\"duration\"",
":",
"duration",
",",
"\"redirect_uri\"",
":",
"self",
".",
"redirect_uri",
",",
"\"response_type\"",
":",
"\"token\"",
"if",
"implicit",
"else",
"\"code\"",
",",
"\"scope\"",
":",
"\" \"",
".",
"join",
"(",
"scopes",
")",
",",
"\"state\"",
":",
"state",
",",
"}",
"url",
"=",
"self",
".",
"_requestor",
".",
"reddit_url",
"+",
"const",
".",
"AUTHORIZATION_PATH",
"request",
"=",
"Request",
"(",
"\"GET\"",
",",
"url",
",",
"params",
"=",
"params",
")",
"return",
"request",
".",
"prepare",
"(",
")",
".",
"url"
] |
Return the URL used out-of-band to grant access to your application.
:param duration: Either ``permanent`` or ``temporary``. ``temporary``
authorizations generate access tokens that last only 1
hour. ``permanent`` authorizations additionally generate a refresh
token that can be indefinitely used to generate new hour-long
access tokens. Only ``temporary`` can be specified if ``implicit``
is set to ``True``.
:param scopes: A list of OAuth scopes to request authorization for.
:param state: A string that will be reflected in the callback to
``redirect_uri``. This value should be temporarily unique to the
client for whom the URL was generated for.
:param implicit: (optional) Use the implicit grant flow (default:
False). This flow is only available for UntrustedAuthenticators.
|
[
"Return",
"the",
"URL",
"used",
"out",
"-",
"of",
"-",
"band",
"to",
"grant",
"access",
"to",
"your",
"application",
"."
] |
b16ae88a1f2bf98095ed6fe64851cb7add7ed752
|
https://github.com/praw-dev/prawcore/blob/b16ae88a1f2bf98095ed6fe64851cb7add7ed752/prawcore/auth.py#L35-L75
|
train
|
praw-dev/prawcore
|
prawcore/auth.py
|
BaseAuthenticator.revoke_token
|
def revoke_token(self, token, token_type=None):
"""Ask Reddit to revoke the provided token.
:param token: The access or refresh token to revoke.
:param token_type: (Optional) When provided, hint to Reddit what the
token type is for a possible efficiency gain. The value can be
either ``access_token`` or ``refresh_token``.
"""
data = {"token": token}
if token_type is not None:
data["token_type_hint"] = token_type
url = self._requestor.reddit_url + const.REVOKE_TOKEN_PATH
self._post(url, success_status=codes["no_content"], **data)
|
python
|
def revoke_token(self, token, token_type=None):
"""Ask Reddit to revoke the provided token.
:param token: The access or refresh token to revoke.
:param token_type: (Optional) When provided, hint to Reddit what the
token type is for a possible efficiency gain. The value can be
either ``access_token`` or ``refresh_token``.
"""
data = {"token": token}
if token_type is not None:
data["token_type_hint"] = token_type
url = self._requestor.reddit_url + const.REVOKE_TOKEN_PATH
self._post(url, success_status=codes["no_content"], **data)
|
[
"def",
"revoke_token",
"(",
"self",
",",
"token",
",",
"token_type",
"=",
"None",
")",
":",
"data",
"=",
"{",
"\"token\"",
":",
"token",
"}",
"if",
"token_type",
"is",
"not",
"None",
":",
"data",
"[",
"\"token_type_hint\"",
"]",
"=",
"token_type",
"url",
"=",
"self",
".",
"_requestor",
".",
"reddit_url",
"+",
"const",
".",
"REVOKE_TOKEN_PATH",
"self",
".",
"_post",
"(",
"url",
",",
"success_status",
"=",
"codes",
"[",
"\"no_content\"",
"]",
",",
"*",
"*",
"data",
")"
] |
Ask Reddit to revoke the provided token.
:param token: The access or refresh token to revoke.
:param token_type: (Optional) When provided, hint to Reddit what the
token type is for a possible efficiency gain. The value can be
either ``access_token`` or ``refresh_token``.
|
[
"Ask",
"Reddit",
"to",
"revoke",
"the",
"provided",
"token",
"."
] |
b16ae88a1f2bf98095ed6fe64851cb7add7ed752
|
https://github.com/praw-dev/prawcore/blob/b16ae88a1f2bf98095ed6fe64851cb7add7ed752/prawcore/auth.py#L77-L90
|
train
|
praw-dev/prawcore
|
prawcore/auth.py
|
BaseAuthorizer.revoke
|
def revoke(self):
"""Revoke the current Authorization."""
if self.access_token is None:
raise InvalidInvocation("no token available to revoke")
self._authenticator.revoke_token(self.access_token, "access_token")
self._clear_access_token()
|
python
|
def revoke(self):
"""Revoke the current Authorization."""
if self.access_token is None:
raise InvalidInvocation("no token available to revoke")
self._authenticator.revoke_token(self.access_token, "access_token")
self._clear_access_token()
|
[
"def",
"revoke",
"(",
"self",
")",
":",
"if",
"self",
".",
"access_token",
"is",
"None",
":",
"raise",
"InvalidInvocation",
"(",
"\"no token available to revoke\"",
")",
"self",
".",
"_authenticator",
".",
"revoke_token",
"(",
"self",
".",
"access_token",
",",
"\"access_token\"",
")",
"self",
".",
"_clear_access_token",
"(",
")"
] |
Revoke the current Authorization.
|
[
"Revoke",
"the",
"current",
"Authorization",
"."
] |
b16ae88a1f2bf98095ed6fe64851cb7add7ed752
|
https://github.com/praw-dev/prawcore/blob/b16ae88a1f2bf98095ed6fe64851cb7add7ed752/prawcore/auth.py#L184-L190
|
train
|
praw-dev/prawcore
|
prawcore/auth.py
|
Authorizer.authorize
|
def authorize(self, code):
"""Obtain and set authorization tokens based on ``code``.
:param code: The code obtained by an out-of-band authorization request
to Reddit.
"""
if self._authenticator.redirect_uri is None:
raise InvalidInvocation("redirect URI not provided")
self._request_token(
code=code,
grant_type="authorization_code",
redirect_uri=self._authenticator.redirect_uri,
)
|
python
|
def authorize(self, code):
"""Obtain and set authorization tokens based on ``code``.
:param code: The code obtained by an out-of-band authorization request
to Reddit.
"""
if self._authenticator.redirect_uri is None:
raise InvalidInvocation("redirect URI not provided")
self._request_token(
code=code,
grant_type="authorization_code",
redirect_uri=self._authenticator.redirect_uri,
)
|
[
"def",
"authorize",
"(",
"self",
",",
"code",
")",
":",
"if",
"self",
".",
"_authenticator",
".",
"redirect_uri",
"is",
"None",
":",
"raise",
"InvalidInvocation",
"(",
"\"redirect URI not provided\"",
")",
"self",
".",
"_request_token",
"(",
"code",
"=",
"code",
",",
"grant_type",
"=",
"\"authorization_code\"",
",",
"redirect_uri",
"=",
"self",
".",
"_authenticator",
".",
"redirect_uri",
",",
")"
] |
Obtain and set authorization tokens based on ``code``.
:param code: The code obtained by an out-of-band authorization request
to Reddit.
|
[
"Obtain",
"and",
"set",
"authorization",
"tokens",
"based",
"on",
"code",
"."
] |
b16ae88a1f2bf98095ed6fe64851cb7add7ed752
|
https://github.com/praw-dev/prawcore/blob/b16ae88a1f2bf98095ed6fe64851cb7add7ed752/prawcore/auth.py#L210-L223
|
train
|
praw-dev/prawcore
|
prawcore/auth.py
|
Authorizer.refresh
|
def refresh(self):
"""Obtain a new access token from the refresh_token."""
if self.refresh_token is None:
raise InvalidInvocation("refresh token not provided")
self._request_token(
grant_type="refresh_token", refresh_token=self.refresh_token
)
|
python
|
def refresh(self):
"""Obtain a new access token from the refresh_token."""
if self.refresh_token is None:
raise InvalidInvocation("refresh token not provided")
self._request_token(
grant_type="refresh_token", refresh_token=self.refresh_token
)
|
[
"def",
"refresh",
"(",
"self",
")",
":",
"if",
"self",
".",
"refresh_token",
"is",
"None",
":",
"raise",
"InvalidInvocation",
"(",
"\"refresh token not provided\"",
")",
"self",
".",
"_request_token",
"(",
"grant_type",
"=",
"\"refresh_token\"",
",",
"refresh_token",
"=",
"self",
".",
"refresh_token",
")"
] |
Obtain a new access token from the refresh_token.
|
[
"Obtain",
"a",
"new",
"access",
"token",
"from",
"the",
"refresh_token",
"."
] |
b16ae88a1f2bf98095ed6fe64851cb7add7ed752
|
https://github.com/praw-dev/prawcore/blob/b16ae88a1f2bf98095ed6fe64851cb7add7ed752/prawcore/auth.py#L225-L231
|
train
|
praw-dev/prawcore
|
prawcore/auth.py
|
Authorizer.revoke
|
def revoke(self, only_access=False):
"""Revoke the current Authorization.
:param only_access: (Optional) When explicitly set to True, do not
evict the refresh token if one is set.
Revoking a refresh token will in-turn revoke all access tokens
associated with that authorization.
"""
if only_access or self.refresh_token is None:
super(Authorizer, self).revoke()
else:
self._authenticator.revoke_token(
self.refresh_token, "refresh_token"
)
self._clear_access_token()
self.refresh_token = None
|
python
|
def revoke(self, only_access=False):
"""Revoke the current Authorization.
:param only_access: (Optional) When explicitly set to True, do not
evict the refresh token if one is set.
Revoking a refresh token will in-turn revoke all access tokens
associated with that authorization.
"""
if only_access or self.refresh_token is None:
super(Authorizer, self).revoke()
else:
self._authenticator.revoke_token(
self.refresh_token, "refresh_token"
)
self._clear_access_token()
self.refresh_token = None
|
[
"def",
"revoke",
"(",
"self",
",",
"only_access",
"=",
"False",
")",
":",
"if",
"only_access",
"or",
"self",
".",
"refresh_token",
"is",
"None",
":",
"super",
"(",
"Authorizer",
",",
"self",
")",
".",
"revoke",
"(",
")",
"else",
":",
"self",
".",
"_authenticator",
".",
"revoke_token",
"(",
"self",
".",
"refresh_token",
",",
"\"refresh_token\"",
")",
"self",
".",
"_clear_access_token",
"(",
")",
"self",
".",
"refresh_token",
"=",
"None"
] |
Revoke the current Authorization.
:param only_access: (Optional) When explicitly set to True, do not
evict the refresh token if one is set.
Revoking a refresh token will in-turn revoke all access tokens
associated with that authorization.
|
[
"Revoke",
"the",
"current",
"Authorization",
"."
] |
b16ae88a1f2bf98095ed6fe64851cb7add7ed752
|
https://github.com/praw-dev/prawcore/blob/b16ae88a1f2bf98095ed6fe64851cb7add7ed752/prawcore/auth.py#L233-L250
|
train
|
praw-dev/prawcore
|
prawcore/auth.py
|
DeviceIDAuthorizer.refresh
|
def refresh(self):
"""Obtain a new access token."""
grant_type = "https://oauth.reddit.com/grants/installed_client"
self._request_token(grant_type=grant_type, device_id=self._device_id)
|
python
|
def refresh(self):
"""Obtain a new access token."""
grant_type = "https://oauth.reddit.com/grants/installed_client"
self._request_token(grant_type=grant_type, device_id=self._device_id)
|
[
"def",
"refresh",
"(",
"self",
")",
":",
"grant_type",
"=",
"\"https://oauth.reddit.com/grants/installed_client\"",
"self",
".",
"_request_token",
"(",
"grant_type",
"=",
"grant_type",
",",
"device_id",
"=",
"self",
".",
"_device_id",
")"
] |
Obtain a new access token.
|
[
"Obtain",
"a",
"new",
"access",
"token",
"."
] |
b16ae88a1f2bf98095ed6fe64851cb7add7ed752
|
https://github.com/praw-dev/prawcore/blob/b16ae88a1f2bf98095ed6fe64851cb7add7ed752/prawcore/auth.py#L275-L278
|
train
|
praw-dev/prawcore
|
prawcore/auth.py
|
ScriptAuthorizer.refresh
|
def refresh(self):
"""Obtain a new personal-use script type access token."""
self._request_token(
grant_type="password",
username=self._username,
password=self._password,
)
|
python
|
def refresh(self):
"""Obtain a new personal-use script type access token."""
self._request_token(
grant_type="password",
username=self._username,
password=self._password,
)
|
[
"def",
"refresh",
"(",
"self",
")",
":",
"self",
".",
"_request_token",
"(",
"grant_type",
"=",
"\"password\"",
",",
"username",
"=",
"self",
".",
"_username",
",",
"password",
"=",
"self",
".",
"_password",
",",
")"
] |
Obtain a new personal-use script type access token.
|
[
"Obtain",
"a",
"new",
"personal",
"-",
"use",
"script",
"type",
"access",
"token",
"."
] |
b16ae88a1f2bf98095ed6fe64851cb7add7ed752
|
https://github.com/praw-dev/prawcore/blob/b16ae88a1f2bf98095ed6fe64851cb7add7ed752/prawcore/auth.py#L347-L353
|
train
|
praw-dev/prawcore
|
prawcore/requestor.py
|
Requestor.request
|
def request(self, *args, **kwargs):
"""Issue the HTTP request capturing any errors that may occur."""
try:
return self._http.request(*args, timeout=TIMEOUT, **kwargs)
except Exception as exc:
raise RequestException(exc, args, kwargs)
|
python
|
def request(self, *args, **kwargs):
"""Issue the HTTP request capturing any errors that may occur."""
try:
return self._http.request(*args, timeout=TIMEOUT, **kwargs)
except Exception as exc:
raise RequestException(exc, args, kwargs)
|
[
"def",
"request",
"(",
"self",
",",
"*",
"args",
",",
"*",
"*",
"kwargs",
")",
":",
"try",
":",
"return",
"self",
".",
"_http",
".",
"request",
"(",
"*",
"args",
",",
"timeout",
"=",
"TIMEOUT",
",",
"*",
"*",
"kwargs",
")",
"except",
"Exception",
"as",
"exc",
":",
"raise",
"RequestException",
"(",
"exc",
",",
"args",
",",
"kwargs",
")"
] |
Issue the HTTP request capturing any errors that may occur.
|
[
"Issue",
"the",
"HTTP",
"request",
"capturing",
"any",
"errors",
"that",
"may",
"occur",
"."
] |
b16ae88a1f2bf98095ed6fe64851cb7add7ed752
|
https://github.com/praw-dev/prawcore/blob/b16ae88a1f2bf98095ed6fe64851cb7add7ed752/prawcore/requestor.py#L50-L55
|
train
|
JDongian/python-jamo
|
jamo/jamo.py
|
_hangul_char_to_jamo
|
def _hangul_char_to_jamo(syllable):
"""Return a 3-tuple of lead, vowel, and tail jamo characters.
Note: Non-Hangul characters are echoed back.
"""
if is_hangul_char(syllable):
rem = ord(syllable) - _JAMO_OFFSET
tail = rem % 28
vowel = 1 + ((rem - tail) % 588) // 28
lead = 1 + rem // 588
if tail:
return (chr(lead + _JAMO_LEAD_OFFSET),
chr(vowel + _JAMO_VOWEL_OFFSET),
chr(tail + _JAMO_TAIL_OFFSET))
else:
return (chr(lead + _JAMO_LEAD_OFFSET),
chr(vowel + _JAMO_VOWEL_OFFSET))
else:
return syllable
|
python
|
def _hangul_char_to_jamo(syllable):
"""Return a 3-tuple of lead, vowel, and tail jamo characters.
Note: Non-Hangul characters are echoed back.
"""
if is_hangul_char(syllable):
rem = ord(syllable) - _JAMO_OFFSET
tail = rem % 28
vowel = 1 + ((rem - tail) % 588) // 28
lead = 1 + rem // 588
if tail:
return (chr(lead + _JAMO_LEAD_OFFSET),
chr(vowel + _JAMO_VOWEL_OFFSET),
chr(tail + _JAMO_TAIL_OFFSET))
else:
return (chr(lead + _JAMO_LEAD_OFFSET),
chr(vowel + _JAMO_VOWEL_OFFSET))
else:
return syllable
|
[
"def",
"_hangul_char_to_jamo",
"(",
"syllable",
")",
":",
"if",
"is_hangul_char",
"(",
"syllable",
")",
":",
"rem",
"=",
"ord",
"(",
"syllable",
")",
"-",
"_JAMO_OFFSET",
"tail",
"=",
"rem",
"%",
"28",
"vowel",
"=",
"1",
"+",
"(",
"(",
"rem",
"-",
"tail",
")",
"%",
"588",
")",
"//",
"28",
"lead",
"=",
"1",
"+",
"rem",
"//",
"588",
"if",
"tail",
":",
"return",
"(",
"chr",
"(",
"lead",
"+",
"_JAMO_LEAD_OFFSET",
")",
",",
"chr",
"(",
"vowel",
"+",
"_JAMO_VOWEL_OFFSET",
")",
",",
"chr",
"(",
"tail",
"+",
"_JAMO_TAIL_OFFSET",
")",
")",
"else",
":",
"return",
"(",
"chr",
"(",
"lead",
"+",
"_JAMO_LEAD_OFFSET",
")",
",",
"chr",
"(",
"vowel",
"+",
"_JAMO_VOWEL_OFFSET",
")",
")",
"else",
":",
"return",
"syllable"
] |
Return a 3-tuple of lead, vowel, and tail jamo characters.
Note: Non-Hangul characters are echoed back.
|
[
"Return",
"a",
"3",
"-",
"tuple",
"of",
"lead",
"vowel",
"and",
"tail",
"jamo",
"characters",
".",
"Note",
":",
"Non",
"-",
"Hangul",
"characters",
"are",
"echoed",
"back",
"."
] |
d087a9f5f52f066fb933ad1da8e9915703374c9a
|
https://github.com/JDongian/python-jamo/blob/d087a9f5f52f066fb933ad1da8e9915703374c9a/jamo/jamo.py#L54-L71
|
train
|
JDongian/python-jamo
|
jamo/jamo.py
|
_jamo_to_hangul_char
|
def _jamo_to_hangul_char(lead, vowel, tail=0):
"""Return the Hangul character for the given jamo characters.
"""
lead = ord(lead) - _JAMO_LEAD_OFFSET
vowel = ord(vowel) - _JAMO_VOWEL_OFFSET
tail = ord(tail) - _JAMO_TAIL_OFFSET if tail else 0
return chr(tail + (vowel - 1) * 28 + (lead - 1) * 588 + _JAMO_OFFSET)
|
python
|
def _jamo_to_hangul_char(lead, vowel, tail=0):
"""Return the Hangul character for the given jamo characters.
"""
lead = ord(lead) - _JAMO_LEAD_OFFSET
vowel = ord(vowel) - _JAMO_VOWEL_OFFSET
tail = ord(tail) - _JAMO_TAIL_OFFSET if tail else 0
return chr(tail + (vowel - 1) * 28 + (lead - 1) * 588 + _JAMO_OFFSET)
|
[
"def",
"_jamo_to_hangul_char",
"(",
"lead",
",",
"vowel",
",",
"tail",
"=",
"0",
")",
":",
"lead",
"=",
"ord",
"(",
"lead",
")",
"-",
"_JAMO_LEAD_OFFSET",
"vowel",
"=",
"ord",
"(",
"vowel",
")",
"-",
"_JAMO_VOWEL_OFFSET",
"tail",
"=",
"ord",
"(",
"tail",
")",
"-",
"_JAMO_TAIL_OFFSET",
"if",
"tail",
"else",
"0",
"return",
"chr",
"(",
"tail",
"+",
"(",
"vowel",
"-",
"1",
")",
"*",
"28",
"+",
"(",
"lead",
"-",
"1",
")",
"*",
"588",
"+",
"_JAMO_OFFSET",
")"
] |
Return the Hangul character for the given jamo characters.
|
[
"Return",
"the",
"Hangul",
"character",
"for",
"the",
"given",
"jamo",
"characters",
"."
] |
d087a9f5f52f066fb933ad1da8e9915703374c9a
|
https://github.com/JDongian/python-jamo/blob/d087a9f5f52f066fb933ad1da8e9915703374c9a/jamo/jamo.py#L74-L80
|
train
|
JDongian/python-jamo
|
jamo/jamo.py
|
_get_unicode_name
|
def _get_unicode_name(char):
"""Fetch the unicode name for jamo characters.
"""
if char not in _JAMO_TO_NAME.keys() and char not in _HCJ_TO_NAME.keys():
raise InvalidJamoError("Not jamo or nameless jamo character", char)
else:
if is_hcj(char):
return _HCJ_TO_NAME[char]
return _JAMO_TO_NAME[char]
|
python
|
def _get_unicode_name(char):
"""Fetch the unicode name for jamo characters.
"""
if char not in _JAMO_TO_NAME.keys() and char not in _HCJ_TO_NAME.keys():
raise InvalidJamoError("Not jamo or nameless jamo character", char)
else:
if is_hcj(char):
return _HCJ_TO_NAME[char]
return _JAMO_TO_NAME[char]
|
[
"def",
"_get_unicode_name",
"(",
"char",
")",
":",
"if",
"char",
"not",
"in",
"_JAMO_TO_NAME",
".",
"keys",
"(",
")",
"and",
"char",
"not",
"in",
"_HCJ_TO_NAME",
".",
"keys",
"(",
")",
":",
"raise",
"InvalidJamoError",
"(",
"\"Not jamo or nameless jamo character\"",
",",
"char",
")",
"else",
":",
"if",
"is_hcj",
"(",
"char",
")",
":",
"return",
"_HCJ_TO_NAME",
"[",
"char",
"]",
"return",
"_JAMO_TO_NAME",
"[",
"char",
"]"
] |
Fetch the unicode name for jamo characters.
|
[
"Fetch",
"the",
"unicode",
"name",
"for",
"jamo",
"characters",
"."
] |
d087a9f5f52f066fb933ad1da8e9915703374c9a
|
https://github.com/JDongian/python-jamo/blob/d087a9f5f52f066fb933ad1da8e9915703374c9a/jamo/jamo.py#L93-L101
|
train
|
JDongian/python-jamo
|
jamo/jamo.py
|
is_jamo
|
def is_jamo(character):
"""Test if a single character is a jamo character.
Valid jamo includes all modern and archaic jamo, as well as all HCJ.
Non-assigned code points are invalid.
"""
code = ord(character)
return 0x1100 <= code <= 0x11FF or\
0xA960 <= code <= 0xA97C or\
0xD7B0 <= code <= 0xD7C6 or 0xD7CB <= code <= 0xD7FB or\
is_hcj(character)
|
python
|
def is_jamo(character):
"""Test if a single character is a jamo character.
Valid jamo includes all modern and archaic jamo, as well as all HCJ.
Non-assigned code points are invalid.
"""
code = ord(character)
return 0x1100 <= code <= 0x11FF or\
0xA960 <= code <= 0xA97C or\
0xD7B0 <= code <= 0xD7C6 or 0xD7CB <= code <= 0xD7FB or\
is_hcj(character)
|
[
"def",
"is_jamo",
"(",
"character",
")",
":",
"code",
"=",
"ord",
"(",
"character",
")",
"return",
"0x1100",
"<=",
"code",
"<=",
"0x11FF",
"or",
"0xA960",
"<=",
"code",
"<=",
"0xA97C",
"or",
"0xD7B0",
"<=",
"code",
"<=",
"0xD7C6",
"or",
"0xD7CB",
"<=",
"code",
"<=",
"0xD7FB",
"or",
"is_hcj",
"(",
"character",
")"
] |
Test if a single character is a jamo character.
Valid jamo includes all modern and archaic jamo, as well as all HCJ.
Non-assigned code points are invalid.
|
[
"Test",
"if",
"a",
"single",
"character",
"is",
"a",
"jamo",
"character",
".",
"Valid",
"jamo",
"includes",
"all",
"modern",
"and",
"archaic",
"jamo",
"as",
"well",
"as",
"all",
"HCJ",
".",
"Non",
"-",
"assigned",
"code",
"points",
"are",
"invalid",
"."
] |
d087a9f5f52f066fb933ad1da8e9915703374c9a
|
https://github.com/JDongian/python-jamo/blob/d087a9f5f52f066fb933ad1da8e9915703374c9a/jamo/jamo.py#L104-L113
|
train
|
JDongian/python-jamo
|
jamo/jamo.py
|
is_jamo_modern
|
def is_jamo_modern(character):
"""Test if a single character is a modern jamo character.
Modern jamo includes all U+11xx jamo in addition to HCJ in modern usage,
as defined in Unicode 7.0.
WARNING: U+1160 is NOT considered a modern jamo character, but it is listed
under 'Medial Vowels' in the Unicode 7.0 spec.
"""
code = ord(character)
return 0x1100 <= code <= 0x1112 or\
0x1161 <= code <= 0x1175 or\
0x11A8 <= code <= 0x11C2 or\
is_hcj_modern(character)
|
python
|
def is_jamo_modern(character):
"""Test if a single character is a modern jamo character.
Modern jamo includes all U+11xx jamo in addition to HCJ in modern usage,
as defined in Unicode 7.0.
WARNING: U+1160 is NOT considered a modern jamo character, but it is listed
under 'Medial Vowels' in the Unicode 7.0 spec.
"""
code = ord(character)
return 0x1100 <= code <= 0x1112 or\
0x1161 <= code <= 0x1175 or\
0x11A8 <= code <= 0x11C2 or\
is_hcj_modern(character)
|
[
"def",
"is_jamo_modern",
"(",
"character",
")",
":",
"code",
"=",
"ord",
"(",
"character",
")",
"return",
"0x1100",
"<=",
"code",
"<=",
"0x1112",
"or",
"0x1161",
"<=",
"code",
"<=",
"0x1175",
"or",
"0x11A8",
"<=",
"code",
"<=",
"0x11C2",
"or",
"is_hcj_modern",
"(",
"character",
")"
] |
Test if a single character is a modern jamo character.
Modern jamo includes all U+11xx jamo in addition to HCJ in modern usage,
as defined in Unicode 7.0.
WARNING: U+1160 is NOT considered a modern jamo character, but it is listed
under 'Medial Vowels' in the Unicode 7.0 spec.
|
[
"Test",
"if",
"a",
"single",
"character",
"is",
"a",
"modern",
"jamo",
"character",
".",
"Modern",
"jamo",
"includes",
"all",
"U",
"+",
"11xx",
"jamo",
"in",
"addition",
"to",
"HCJ",
"in",
"modern",
"usage",
"as",
"defined",
"in",
"Unicode",
"7",
".",
"0",
".",
"WARNING",
":",
"U",
"+",
"1160",
"is",
"NOT",
"considered",
"a",
"modern",
"jamo",
"character",
"but",
"it",
"is",
"listed",
"under",
"Medial",
"Vowels",
"in",
"the",
"Unicode",
"7",
".",
"0",
"spec",
"."
] |
d087a9f5f52f066fb933ad1da8e9915703374c9a
|
https://github.com/JDongian/python-jamo/blob/d087a9f5f52f066fb933ad1da8e9915703374c9a/jamo/jamo.py#L116-L127
|
train
|
JDongian/python-jamo
|
jamo/jamo.py
|
is_jamo_compound
|
def is_jamo_compound(character):
"""Test if a single character is a compound, i.e., a consonant
cluster, double consonant, or dipthong.
"""
if len(character) != 1:
return False
# Consider instead:
# raise TypeError('is_jamo_compound() expected a single character')
if is_jamo(character):
return character in JAMO_COMPOUNDS
return False
|
python
|
def is_jamo_compound(character):
"""Test if a single character is a compound, i.e., a consonant
cluster, double consonant, or dipthong.
"""
if len(character) != 1:
return False
# Consider instead:
# raise TypeError('is_jamo_compound() expected a single character')
if is_jamo(character):
return character in JAMO_COMPOUNDS
return False
|
[
"def",
"is_jamo_compound",
"(",
"character",
")",
":",
"if",
"len",
"(",
"character",
")",
"!=",
"1",
":",
"return",
"False",
"# Consider instead:",
"# raise TypeError('is_jamo_compound() expected a single character')",
"if",
"is_jamo",
"(",
"character",
")",
":",
"return",
"character",
"in",
"JAMO_COMPOUNDS",
"return",
"False"
] |
Test if a single character is a compound, i.e., a consonant
cluster, double consonant, or dipthong.
|
[
"Test",
"if",
"a",
"single",
"character",
"is",
"a",
"compound",
"i",
".",
"e",
".",
"a",
"consonant",
"cluster",
"double",
"consonant",
"or",
"dipthong",
"."
] |
d087a9f5f52f066fb933ad1da8e9915703374c9a
|
https://github.com/JDongian/python-jamo/blob/d087a9f5f52f066fb933ad1da8e9915703374c9a/jamo/jamo.py#L155-L165
|
train
|
JDongian/python-jamo
|
jamo/jamo.py
|
get_jamo_class
|
def get_jamo_class(jamo):
"""Determine if a jamo character is a lead, vowel, or tail.
Integers and U+11xx characters are valid arguments. HCJ consonants are not
valid here.
get_jamo_class should return the class ["lead" | "vowel" | "tail"] of a
given character or integer.
Note: jamo class directly corresponds to the Unicode 7.0 specification,
thus includes filler characters as having a class.
"""
# TODO: Perhaps raise a separate error for U+3xxx jamo.
if jamo in JAMO_LEADS or jamo == chr(0x115F):
return "lead"
if jamo in JAMO_VOWELS or jamo == chr(0x1160) or\
0x314F <= ord(jamo) <= 0x3163:
return "vowel"
if jamo in JAMO_TAILS:
return "tail"
else:
raise InvalidJamoError("Invalid or classless jamo argument.", jamo)
|
python
|
def get_jamo_class(jamo):
"""Determine if a jamo character is a lead, vowel, or tail.
Integers and U+11xx characters are valid arguments. HCJ consonants are not
valid here.
get_jamo_class should return the class ["lead" | "vowel" | "tail"] of a
given character or integer.
Note: jamo class directly corresponds to the Unicode 7.0 specification,
thus includes filler characters as having a class.
"""
# TODO: Perhaps raise a separate error for U+3xxx jamo.
if jamo in JAMO_LEADS or jamo == chr(0x115F):
return "lead"
if jamo in JAMO_VOWELS or jamo == chr(0x1160) or\
0x314F <= ord(jamo) <= 0x3163:
return "vowel"
if jamo in JAMO_TAILS:
return "tail"
else:
raise InvalidJamoError("Invalid or classless jamo argument.", jamo)
|
[
"def",
"get_jamo_class",
"(",
"jamo",
")",
":",
"# TODO: Perhaps raise a separate error for U+3xxx jamo.",
"if",
"jamo",
"in",
"JAMO_LEADS",
"or",
"jamo",
"==",
"chr",
"(",
"0x115F",
")",
":",
"return",
"\"lead\"",
"if",
"jamo",
"in",
"JAMO_VOWELS",
"or",
"jamo",
"==",
"chr",
"(",
"0x1160",
")",
"or",
"0x314F",
"<=",
"ord",
"(",
"jamo",
")",
"<=",
"0x3163",
":",
"return",
"\"vowel\"",
"if",
"jamo",
"in",
"JAMO_TAILS",
":",
"return",
"\"tail\"",
"else",
":",
"raise",
"InvalidJamoError",
"(",
"\"Invalid or classless jamo argument.\"",
",",
"jamo",
")"
] |
Determine if a jamo character is a lead, vowel, or tail.
Integers and U+11xx characters are valid arguments. HCJ consonants are not
valid here.
get_jamo_class should return the class ["lead" | "vowel" | "tail"] of a
given character or integer.
Note: jamo class directly corresponds to the Unicode 7.0 specification,
thus includes filler characters as having a class.
|
[
"Determine",
"if",
"a",
"jamo",
"character",
"is",
"a",
"lead",
"vowel",
"or",
"tail",
".",
"Integers",
"and",
"U",
"+",
"11xx",
"characters",
"are",
"valid",
"arguments",
".",
"HCJ",
"consonants",
"are",
"not",
"valid",
"here",
"."
] |
d087a9f5f52f066fb933ad1da8e9915703374c9a
|
https://github.com/JDongian/python-jamo/blob/d087a9f5f52f066fb933ad1da8e9915703374c9a/jamo/jamo.py#L168-L188
|
train
|
JDongian/python-jamo
|
jamo/jamo.py
|
hcj_to_jamo
|
def hcj_to_jamo(hcj_char, position="vowel"):
"""Convert a HCJ character to a jamo character.
Arguments may be single characters along with the desired jamo class
(lead, vowel, tail). Non-mappable input will raise an InvalidJamoError.
"""
if position == "lead":
jamo_class = "CHOSEONG"
elif position == "vowel":
jamo_class = "JUNGSEONG"
elif position == "tail":
jamo_class = "JONGSEONG"
else:
raise InvalidJamoError("No mapping from input to jamo.", hcj_char)
jamo_name = re.sub("(?<=HANGUL )(\w+)",
jamo_class,
_get_unicode_name(hcj_char))
# TODO: add tests that test non entries.
if jamo_name in _JAMO_REVERSE_LOOKUP.keys():
return _JAMO_REVERSE_LOOKUP[jamo_name]
return hcj_char
|
python
|
def hcj_to_jamo(hcj_char, position="vowel"):
"""Convert a HCJ character to a jamo character.
Arguments may be single characters along with the desired jamo class
(lead, vowel, tail). Non-mappable input will raise an InvalidJamoError.
"""
if position == "lead":
jamo_class = "CHOSEONG"
elif position == "vowel":
jamo_class = "JUNGSEONG"
elif position == "tail":
jamo_class = "JONGSEONG"
else:
raise InvalidJamoError("No mapping from input to jamo.", hcj_char)
jamo_name = re.sub("(?<=HANGUL )(\w+)",
jamo_class,
_get_unicode_name(hcj_char))
# TODO: add tests that test non entries.
if jamo_name in _JAMO_REVERSE_LOOKUP.keys():
return _JAMO_REVERSE_LOOKUP[jamo_name]
return hcj_char
|
[
"def",
"hcj_to_jamo",
"(",
"hcj_char",
",",
"position",
"=",
"\"vowel\"",
")",
":",
"if",
"position",
"==",
"\"lead\"",
":",
"jamo_class",
"=",
"\"CHOSEONG\"",
"elif",
"position",
"==",
"\"vowel\"",
":",
"jamo_class",
"=",
"\"JUNGSEONG\"",
"elif",
"position",
"==",
"\"tail\"",
":",
"jamo_class",
"=",
"\"JONGSEONG\"",
"else",
":",
"raise",
"InvalidJamoError",
"(",
"\"No mapping from input to jamo.\"",
",",
"hcj_char",
")",
"jamo_name",
"=",
"re",
".",
"sub",
"(",
"\"(?<=HANGUL )(\\w+)\"",
",",
"jamo_class",
",",
"_get_unicode_name",
"(",
"hcj_char",
")",
")",
"# TODO: add tests that test non entries.",
"if",
"jamo_name",
"in",
"_JAMO_REVERSE_LOOKUP",
".",
"keys",
"(",
")",
":",
"return",
"_JAMO_REVERSE_LOOKUP",
"[",
"jamo_name",
"]",
"return",
"hcj_char"
] |
Convert a HCJ character to a jamo character.
Arguments may be single characters along with the desired jamo class
(lead, vowel, tail). Non-mappable input will raise an InvalidJamoError.
|
[
"Convert",
"a",
"HCJ",
"character",
"to",
"a",
"jamo",
"character",
".",
"Arguments",
"may",
"be",
"single",
"characters",
"along",
"with",
"the",
"desired",
"jamo",
"class",
"(",
"lead",
"vowel",
"tail",
")",
".",
"Non",
"-",
"mappable",
"input",
"will",
"raise",
"an",
"InvalidJamoError",
"."
] |
d087a9f5f52f066fb933ad1da8e9915703374c9a
|
https://github.com/JDongian/python-jamo/blob/d087a9f5f52f066fb933ad1da8e9915703374c9a/jamo/jamo.py#L216-L235
|
train
|
JDongian/python-jamo
|
jamo/jamo.py
|
hangul_to_jamo
|
def hangul_to_jamo(hangul_string):
"""Convert a string of Hangul to jamo.
Arguments may be iterables of characters.
hangul_to_jamo should split every Hangul character into U+11xx jamo
characters for any given string. Non-hangul characters are not changed.
hangul_to_jamo is the generator version of h2j, the string version.
"""
return (_ for _ in
chain.from_iterable(_hangul_char_to_jamo(_) for _ in
hangul_string))
|
python
|
def hangul_to_jamo(hangul_string):
"""Convert a string of Hangul to jamo.
Arguments may be iterables of characters.
hangul_to_jamo should split every Hangul character into U+11xx jamo
characters for any given string. Non-hangul characters are not changed.
hangul_to_jamo is the generator version of h2j, the string version.
"""
return (_ for _ in
chain.from_iterable(_hangul_char_to_jamo(_) for _ in
hangul_string))
|
[
"def",
"hangul_to_jamo",
"(",
"hangul_string",
")",
":",
"return",
"(",
"_",
"for",
"_",
"in",
"chain",
".",
"from_iterable",
"(",
"_hangul_char_to_jamo",
"(",
"_",
")",
"for",
"_",
"in",
"hangul_string",
")",
")"
] |
Convert a string of Hangul to jamo.
Arguments may be iterables of characters.
hangul_to_jamo should split every Hangul character into U+11xx jamo
characters for any given string. Non-hangul characters are not changed.
hangul_to_jamo is the generator version of h2j, the string version.
|
[
"Convert",
"a",
"string",
"of",
"Hangul",
"to",
"jamo",
".",
"Arguments",
"may",
"be",
"iterables",
"of",
"characters",
"."
] |
d087a9f5f52f066fb933ad1da8e9915703374c9a
|
https://github.com/JDongian/python-jamo/blob/d087a9f5f52f066fb933ad1da8e9915703374c9a/jamo/jamo.py#L245-L256
|
train
|
JDongian/python-jamo
|
jamo/jamo.py
|
jamo_to_hangul
|
def jamo_to_hangul(lead, vowel, tail=''):
"""Return the Hangul character for the given jamo input.
Integers corresponding to U+11xx jamo codepoints, U+11xx jamo characters,
or HCJ are valid inputs.
Outputs a one-character Hangul string.
This function is identical to j2h.
"""
# Internally, we convert everything to a jamo char,
# then pass it to _jamo_to_hangul_char
lead = hcj_to_jamo(lead, "lead")
vowel = hcj_to_jamo(vowel, "vowel")
if not tail or ord(tail) == 0:
tail = None
elif is_hcj(tail):
tail = hcj_to_jamo(tail, "tail")
if (is_jamo(lead) and get_jamo_class(lead) == "lead") and\
(is_jamo(vowel) and get_jamo_class(vowel) == "vowel") and\
((not tail) or (is_jamo(tail) and get_jamo_class(tail) == "tail")):
result = _jamo_to_hangul_char(lead, vowel, tail)
if is_hangul_char(result):
return result
raise InvalidJamoError("Could not synthesize characters to Hangul.",
'\x00')
|
python
|
def jamo_to_hangul(lead, vowel, tail=''):
"""Return the Hangul character for the given jamo input.
Integers corresponding to U+11xx jamo codepoints, U+11xx jamo characters,
or HCJ are valid inputs.
Outputs a one-character Hangul string.
This function is identical to j2h.
"""
# Internally, we convert everything to a jamo char,
# then pass it to _jamo_to_hangul_char
lead = hcj_to_jamo(lead, "lead")
vowel = hcj_to_jamo(vowel, "vowel")
if not tail or ord(tail) == 0:
tail = None
elif is_hcj(tail):
tail = hcj_to_jamo(tail, "tail")
if (is_jamo(lead) and get_jamo_class(lead) == "lead") and\
(is_jamo(vowel) and get_jamo_class(vowel) == "vowel") and\
((not tail) or (is_jamo(tail) and get_jamo_class(tail) == "tail")):
result = _jamo_to_hangul_char(lead, vowel, tail)
if is_hangul_char(result):
return result
raise InvalidJamoError("Could not synthesize characters to Hangul.",
'\x00')
|
[
"def",
"jamo_to_hangul",
"(",
"lead",
",",
"vowel",
",",
"tail",
"=",
"''",
")",
":",
"# Internally, we convert everything to a jamo char,",
"# then pass it to _jamo_to_hangul_char",
"lead",
"=",
"hcj_to_jamo",
"(",
"lead",
",",
"\"lead\"",
")",
"vowel",
"=",
"hcj_to_jamo",
"(",
"vowel",
",",
"\"vowel\"",
")",
"if",
"not",
"tail",
"or",
"ord",
"(",
"tail",
")",
"==",
"0",
":",
"tail",
"=",
"None",
"elif",
"is_hcj",
"(",
"tail",
")",
":",
"tail",
"=",
"hcj_to_jamo",
"(",
"tail",
",",
"\"tail\"",
")",
"if",
"(",
"is_jamo",
"(",
"lead",
")",
"and",
"get_jamo_class",
"(",
"lead",
")",
"==",
"\"lead\"",
")",
"and",
"(",
"is_jamo",
"(",
"vowel",
")",
"and",
"get_jamo_class",
"(",
"vowel",
")",
"==",
"\"vowel\"",
")",
"and",
"(",
"(",
"not",
"tail",
")",
"or",
"(",
"is_jamo",
"(",
"tail",
")",
"and",
"get_jamo_class",
"(",
"tail",
")",
"==",
"\"tail\"",
")",
")",
":",
"result",
"=",
"_jamo_to_hangul_char",
"(",
"lead",
",",
"vowel",
",",
"tail",
")",
"if",
"is_hangul_char",
"(",
"result",
")",
":",
"return",
"result",
"raise",
"InvalidJamoError",
"(",
"\"Could not synthesize characters to Hangul.\"",
",",
"'\\x00'",
")"
] |
Return the Hangul character for the given jamo input.
Integers corresponding to U+11xx jamo codepoints, U+11xx jamo characters,
or HCJ are valid inputs.
Outputs a one-character Hangul string.
This function is identical to j2h.
|
[
"Return",
"the",
"Hangul",
"character",
"for",
"the",
"given",
"jamo",
"input",
".",
"Integers",
"corresponding",
"to",
"U",
"+",
"11xx",
"jamo",
"codepoints",
"U",
"+",
"11xx",
"jamo",
"characters",
"or",
"HCJ",
"are",
"valid",
"inputs",
"."
] |
d087a9f5f52f066fb933ad1da8e9915703374c9a
|
https://github.com/JDongian/python-jamo/blob/d087a9f5f52f066fb933ad1da8e9915703374c9a/jamo/jamo.py#L271-L295
|
train
|
JDongian/python-jamo
|
jamo/jamo.py
|
decompose_jamo
|
def decompose_jamo(compound):
"""Return a tuple of jamo character constituents of a compound.
Note: Non-compound characters are echoed back.
WARNING: Archaic jamo compounds will raise NotImplementedError.
"""
if len(compound) != 1:
raise TypeError("decompose_jamo() expects a single character,",
"but received", type(compound), "length",
len(compound))
if compound not in JAMO_COMPOUNDS:
# Strict version:
# raise TypeError("decompose_jamo() expects a compound jamo,",
# "but received", compound)
return compound
return _JAMO_TO_COMPONENTS.get(compound, compound)
|
python
|
def decompose_jamo(compound):
"""Return a tuple of jamo character constituents of a compound.
Note: Non-compound characters are echoed back.
WARNING: Archaic jamo compounds will raise NotImplementedError.
"""
if len(compound) != 1:
raise TypeError("decompose_jamo() expects a single character,",
"but received", type(compound), "length",
len(compound))
if compound not in JAMO_COMPOUNDS:
# Strict version:
# raise TypeError("decompose_jamo() expects a compound jamo,",
# "but received", compound)
return compound
return _JAMO_TO_COMPONENTS.get(compound, compound)
|
[
"def",
"decompose_jamo",
"(",
"compound",
")",
":",
"if",
"len",
"(",
"compound",
")",
"!=",
"1",
":",
"raise",
"TypeError",
"(",
"\"decompose_jamo() expects a single character,\"",
",",
"\"but received\"",
",",
"type",
"(",
"compound",
")",
",",
"\"length\"",
",",
"len",
"(",
"compound",
")",
")",
"if",
"compound",
"not",
"in",
"JAMO_COMPOUNDS",
":",
"# Strict version:",
"# raise TypeError(\"decompose_jamo() expects a compound jamo,\",",
"# \"but received\", compound)",
"return",
"compound",
"return",
"_JAMO_TO_COMPONENTS",
".",
"get",
"(",
"compound",
",",
"compound",
")"
] |
Return a tuple of jamo character constituents of a compound.
Note: Non-compound characters are echoed back.
WARNING: Archaic jamo compounds will raise NotImplementedError.
|
[
"Return",
"a",
"tuple",
"of",
"jamo",
"character",
"constituents",
"of",
"a",
"compound",
".",
"Note",
":",
"Non",
"-",
"compound",
"characters",
"are",
"echoed",
"back",
"."
] |
d087a9f5f52f066fb933ad1da8e9915703374c9a
|
https://github.com/JDongian/python-jamo/blob/d087a9f5f52f066fb933ad1da8e9915703374c9a/jamo/jamo.py#L310-L325
|
train
|
JDongian/python-jamo
|
jamo/jamo.py
|
compose_jamo
|
def compose_jamo(*parts):
"""Return the compound jamo for the given jamo input.
Integers corresponding to U+11xx jamo codepoints, U+11xx jamo
characters, or HCJ are valid inputs.
Outputs a one-character jamo string.
"""
# Internally, we convert everything to a jamo char,
# then pass it to _jamo_to_hangul_char
# NOTE: Relies on hcj_to_jamo not strictly requiring "position" arg.
for p in parts:
if not (type(p) == str and len(p) == 1 and 2 <= len(parts) <= 3):
raise TypeError("compose_jamo() expected 2-3 single characters " +
"but received " + str(parts),
'\x00')
hcparts = [j2hcj(_) for _ in parts]
hcparts = tuple(hcparts)
if hcparts in _COMPONENTS_REVERSE_LOOKUP:
return _COMPONENTS_REVERSE_LOOKUP[hcparts]
raise InvalidJamoError(
"Could not synthesize characters to compound: " + ", ".join(
str(_) + "(U+" + str(hex(ord(_)))[2:] +
")" for _ in parts), '\x00')
|
python
|
def compose_jamo(*parts):
"""Return the compound jamo for the given jamo input.
Integers corresponding to U+11xx jamo codepoints, U+11xx jamo
characters, or HCJ are valid inputs.
Outputs a one-character jamo string.
"""
# Internally, we convert everything to a jamo char,
# then pass it to _jamo_to_hangul_char
# NOTE: Relies on hcj_to_jamo not strictly requiring "position" arg.
for p in parts:
if not (type(p) == str and len(p) == 1 and 2 <= len(parts) <= 3):
raise TypeError("compose_jamo() expected 2-3 single characters " +
"but received " + str(parts),
'\x00')
hcparts = [j2hcj(_) for _ in parts]
hcparts = tuple(hcparts)
if hcparts in _COMPONENTS_REVERSE_LOOKUP:
return _COMPONENTS_REVERSE_LOOKUP[hcparts]
raise InvalidJamoError(
"Could not synthesize characters to compound: " + ", ".join(
str(_) + "(U+" + str(hex(ord(_)))[2:] +
")" for _ in parts), '\x00')
|
[
"def",
"compose_jamo",
"(",
"*",
"parts",
")",
":",
"# Internally, we convert everything to a jamo char,",
"# then pass it to _jamo_to_hangul_char",
"# NOTE: Relies on hcj_to_jamo not strictly requiring \"position\" arg.",
"for",
"p",
"in",
"parts",
":",
"if",
"not",
"(",
"type",
"(",
"p",
")",
"==",
"str",
"and",
"len",
"(",
"p",
")",
"==",
"1",
"and",
"2",
"<=",
"len",
"(",
"parts",
")",
"<=",
"3",
")",
":",
"raise",
"TypeError",
"(",
"\"compose_jamo() expected 2-3 single characters \"",
"+",
"\"but received \"",
"+",
"str",
"(",
"parts",
")",
",",
"'\\x00'",
")",
"hcparts",
"=",
"[",
"j2hcj",
"(",
"_",
")",
"for",
"_",
"in",
"parts",
"]",
"hcparts",
"=",
"tuple",
"(",
"hcparts",
")",
"if",
"hcparts",
"in",
"_COMPONENTS_REVERSE_LOOKUP",
":",
"return",
"_COMPONENTS_REVERSE_LOOKUP",
"[",
"hcparts",
"]",
"raise",
"InvalidJamoError",
"(",
"\"Could not synthesize characters to compound: \"",
"+",
"\", \"",
".",
"join",
"(",
"str",
"(",
"_",
")",
"+",
"\"(U+\"",
"+",
"str",
"(",
"hex",
"(",
"ord",
"(",
"_",
")",
")",
")",
"[",
"2",
":",
"]",
"+",
"\")\"",
"for",
"_",
"in",
"parts",
")",
",",
"'\\x00'",
")"
] |
Return the compound jamo for the given jamo input.
Integers corresponding to U+11xx jamo codepoints, U+11xx jamo
characters, or HCJ are valid inputs.
Outputs a one-character jamo string.
|
[
"Return",
"the",
"compound",
"jamo",
"for",
"the",
"given",
"jamo",
"input",
".",
"Integers",
"corresponding",
"to",
"U",
"+",
"11xx",
"jamo",
"codepoints",
"U",
"+",
"11xx",
"jamo",
"characters",
"or",
"HCJ",
"are",
"valid",
"inputs",
"."
] |
d087a9f5f52f066fb933ad1da8e9915703374c9a
|
https://github.com/JDongian/python-jamo/blob/d087a9f5f52f066fb933ad1da8e9915703374c9a/jamo/jamo.py#L328-L350
|
train
|
JDongian/python-jamo
|
jamo/jamo.py
|
synth_hangul
|
def synth_hangul(string):
"""Convert jamo characters in a string into hcj as much as possible."""
raise NotImplementedError
return ''.join([''.join(''.join(jamo_to_hcj(_)) for _ in string)])
|
python
|
def synth_hangul(string):
"""Convert jamo characters in a string into hcj as much as possible."""
raise NotImplementedError
return ''.join([''.join(''.join(jamo_to_hcj(_)) for _ in string)])
|
[
"def",
"synth_hangul",
"(",
"string",
")",
":",
"raise",
"NotImplementedError",
"return",
"''",
".",
"join",
"(",
"[",
"''",
".",
"join",
"(",
"''",
".",
"join",
"(",
"jamo_to_hcj",
"(",
"_",
")",
")",
"for",
"_",
"in",
"string",
")",
"]",
")"
] |
Convert jamo characters in a string into hcj as much as possible.
|
[
"Convert",
"jamo",
"characters",
"in",
"a",
"string",
"into",
"hcj",
"as",
"much",
"as",
"possible",
"."
] |
d087a9f5f52f066fb933ad1da8e9915703374c9a
|
https://github.com/JDongian/python-jamo/blob/d087a9f5f52f066fb933ad1da8e9915703374c9a/jamo/jamo.py#L353-L356
|
train
|
praw-dev/prawcore
|
prawcore/util.py
|
authorization_error_class
|
def authorization_error_class(response):
"""Return an exception instance that maps to the OAuth Error.
:param response: The HTTP response containing a www-authenticate error.
"""
message = response.headers.get("www-authenticate")
if message:
error = message.replace('"', "").rsplit("=", 1)[1]
else:
error = response.status_code
return _auth_error_mapping[error](response)
|
python
|
def authorization_error_class(response):
"""Return an exception instance that maps to the OAuth Error.
:param response: The HTTP response containing a www-authenticate error.
"""
message = response.headers.get("www-authenticate")
if message:
error = message.replace('"', "").rsplit("=", 1)[1]
else:
error = response.status_code
return _auth_error_mapping[error](response)
|
[
"def",
"authorization_error_class",
"(",
"response",
")",
":",
"message",
"=",
"response",
".",
"headers",
".",
"get",
"(",
"\"www-authenticate\"",
")",
"if",
"message",
":",
"error",
"=",
"message",
".",
"replace",
"(",
"'\"'",
",",
"\"\"",
")",
".",
"rsplit",
"(",
"\"=\"",
",",
"1",
")",
"[",
"1",
"]",
"else",
":",
"error",
"=",
"response",
".",
"status_code",
"return",
"_auth_error_mapping",
"[",
"error",
"]",
"(",
"response",
")"
] |
Return an exception instance that maps to the OAuth Error.
:param response: The HTTP response containing a www-authenticate error.
|
[
"Return",
"an",
"exception",
"instance",
"that",
"maps",
"to",
"the",
"OAuth",
"Error",
"."
] |
b16ae88a1f2bf98095ed6fe64851cb7add7ed752
|
https://github.com/praw-dev/prawcore/blob/b16ae88a1f2bf98095ed6fe64851cb7add7ed752/prawcore/util.py#L12-L23
|
train
|
neo4j-drivers/neobolt
|
neobolt/impl/python/direct.py
|
_last_bookmark
|
def _last_bookmark(b0, b1):
""" Return the latest of two bookmarks by looking for the maximum
integer value following the last colon in the bookmark string.
"""
n = [None, None]
_, _, n[0] = b0.rpartition(":")
_, _, n[1] = b1.rpartition(":")
for i in range(2):
try:
n[i] = int(n[i])
except ValueError:
raise ValueError("Invalid bookmark: {}".format(b0))
return b0 if n[0] > n[1] else b1
|
python
|
def _last_bookmark(b0, b1):
""" Return the latest of two bookmarks by looking for the maximum
integer value following the last colon in the bookmark string.
"""
n = [None, None]
_, _, n[0] = b0.rpartition(":")
_, _, n[1] = b1.rpartition(":")
for i in range(2):
try:
n[i] = int(n[i])
except ValueError:
raise ValueError("Invalid bookmark: {}".format(b0))
return b0 if n[0] > n[1] else b1
|
[
"def",
"_last_bookmark",
"(",
"b0",
",",
"b1",
")",
":",
"n",
"=",
"[",
"None",
",",
"None",
"]",
"_",
",",
"_",
",",
"n",
"[",
"0",
"]",
"=",
"b0",
".",
"rpartition",
"(",
"\":\"",
")",
"_",
",",
"_",
",",
"n",
"[",
"1",
"]",
"=",
"b1",
".",
"rpartition",
"(",
"\":\"",
")",
"for",
"i",
"in",
"range",
"(",
"2",
")",
":",
"try",
":",
"n",
"[",
"i",
"]",
"=",
"int",
"(",
"n",
"[",
"i",
"]",
")",
"except",
"ValueError",
":",
"raise",
"ValueError",
"(",
"\"Invalid bookmark: {}\"",
".",
"format",
"(",
"b0",
")",
")",
"return",
"b0",
"if",
"n",
"[",
"0",
"]",
">",
"n",
"[",
"1",
"]",
"else",
"b1"
] |
Return the latest of two bookmarks by looking for the maximum
integer value following the last colon in the bookmark string.
|
[
"Return",
"the",
"latest",
"of",
"two",
"bookmarks",
"by",
"looking",
"for",
"the",
"maximum",
"integer",
"value",
"following",
"the",
"last",
"colon",
"in",
"the",
"bookmark",
"string",
"."
] |
724569d76e85777c4f5e30e8d0a18116bda4d8cd
|
https://github.com/neo4j-drivers/neobolt/blob/724569d76e85777c4f5e30e8d0a18116bda4d8cd/neobolt/impl/python/direct.py#L705-L717
|
train
|
neo4j-drivers/neobolt
|
neobolt/impl/python/direct.py
|
last_bookmark
|
def last_bookmark(bookmarks):
""" The bookmark returned by the last :class:`.Transaction`.
"""
last = None
for bookmark in bookmarks:
if last is None:
last = bookmark
else:
last = _last_bookmark(last, bookmark)
return last
|
python
|
def last_bookmark(bookmarks):
""" The bookmark returned by the last :class:`.Transaction`.
"""
last = None
for bookmark in bookmarks:
if last is None:
last = bookmark
else:
last = _last_bookmark(last, bookmark)
return last
|
[
"def",
"last_bookmark",
"(",
"bookmarks",
")",
":",
"last",
"=",
"None",
"for",
"bookmark",
"in",
"bookmarks",
":",
"if",
"last",
"is",
"None",
":",
"last",
"=",
"bookmark",
"else",
":",
"last",
"=",
"_last_bookmark",
"(",
"last",
",",
"bookmark",
")",
"return",
"last"
] |
The bookmark returned by the last :class:`.Transaction`.
|
[
"The",
"bookmark",
"returned",
"by",
"the",
"last",
":",
"class",
":",
".",
"Transaction",
"."
] |
724569d76e85777c4f5e30e8d0a18116bda4d8cd
|
https://github.com/neo4j-drivers/neobolt/blob/724569d76e85777c4f5e30e8d0a18116bda4d8cd/neobolt/impl/python/direct.py#L721-L730
|
train
|
neo4j-drivers/neobolt
|
neobolt/impl/python/direct.py
|
connect
|
def connect(address, **config):
""" Connect and perform a handshake and return a valid Connection object, assuming
a protocol version can be agreed.
"""
ssl_context = make_ssl_context(**config)
last_error = None
# Establish a connection to the host and port specified
# Catches refused connections see:
# https://docs.python.org/2/library/errno.html
log_debug("[#0000] C: <RESOLVE> %s", address)
resolver = Resolver(custom_resolver=config.get("resolver"))
resolver.addresses.append(address)
resolver.custom_resolve()
resolver.dns_resolve()
for resolved_address in resolver.addresses:
try:
s = _connect(resolved_address, **config)
s, der_encoded_server_certificate = _secure(s, address[0], ssl_context)
connection = _handshake(s, resolved_address, der_encoded_server_certificate, **config)
except Exception as error:
last_error = error
else:
return connection
if last_error is None:
raise ServiceUnavailable("Failed to resolve addresses for %s" % address)
else:
raise last_error
|
python
|
def connect(address, **config):
""" Connect and perform a handshake and return a valid Connection object, assuming
a protocol version can be agreed.
"""
ssl_context = make_ssl_context(**config)
last_error = None
# Establish a connection to the host and port specified
# Catches refused connections see:
# https://docs.python.org/2/library/errno.html
log_debug("[#0000] C: <RESOLVE> %s", address)
resolver = Resolver(custom_resolver=config.get("resolver"))
resolver.addresses.append(address)
resolver.custom_resolve()
resolver.dns_resolve()
for resolved_address in resolver.addresses:
try:
s = _connect(resolved_address, **config)
s, der_encoded_server_certificate = _secure(s, address[0], ssl_context)
connection = _handshake(s, resolved_address, der_encoded_server_certificate, **config)
except Exception as error:
last_error = error
else:
return connection
if last_error is None:
raise ServiceUnavailable("Failed to resolve addresses for %s" % address)
else:
raise last_error
|
[
"def",
"connect",
"(",
"address",
",",
"*",
"*",
"config",
")",
":",
"ssl_context",
"=",
"make_ssl_context",
"(",
"*",
"*",
"config",
")",
"last_error",
"=",
"None",
"# Establish a connection to the host and port specified",
"# Catches refused connections see:",
"# https://docs.python.org/2/library/errno.html",
"log_debug",
"(",
"\"[#0000] C: <RESOLVE> %s\"",
",",
"address",
")",
"resolver",
"=",
"Resolver",
"(",
"custom_resolver",
"=",
"config",
".",
"get",
"(",
"\"resolver\"",
")",
")",
"resolver",
".",
"addresses",
".",
"append",
"(",
"address",
")",
"resolver",
".",
"custom_resolve",
"(",
")",
"resolver",
".",
"dns_resolve",
"(",
")",
"for",
"resolved_address",
"in",
"resolver",
".",
"addresses",
":",
"try",
":",
"s",
"=",
"_connect",
"(",
"resolved_address",
",",
"*",
"*",
"config",
")",
"s",
",",
"der_encoded_server_certificate",
"=",
"_secure",
"(",
"s",
",",
"address",
"[",
"0",
"]",
",",
"ssl_context",
")",
"connection",
"=",
"_handshake",
"(",
"s",
",",
"resolved_address",
",",
"der_encoded_server_certificate",
",",
"*",
"*",
"config",
")",
"except",
"Exception",
"as",
"error",
":",
"last_error",
"=",
"error",
"else",
":",
"return",
"connection",
"if",
"last_error",
"is",
"None",
":",
"raise",
"ServiceUnavailable",
"(",
"\"Failed to resolve addresses for %s\"",
"%",
"address",
")",
"else",
":",
"raise",
"last_error"
] |
Connect and perform a handshake and return a valid Connection object, assuming
a protocol version can be agreed.
|
[
"Connect",
"and",
"perform",
"a",
"handshake",
"and",
"return",
"a",
"valid",
"Connection",
"object",
"assuming",
"a",
"protocol",
"version",
"can",
"be",
"agreed",
"."
] |
724569d76e85777c4f5e30e8d0a18116bda4d8cd
|
https://github.com/neo4j-drivers/neobolt/blob/724569d76e85777c4f5e30e8d0a18116bda4d8cd/neobolt/impl/python/direct.py#L858-L884
|
train
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.