{"repo": "talkiq/gcloud-aio", "pull_number": 24, "instance_id": "talkiq__gcloud-aio-24", "issue_numbers": "", "base_commit": "02d7ce2526855c16a36d7cced824a9c7b7d9bd6b", "patch": "diff --git a/auth/gcloud/aio/auth/auth.py b/auth/gcloud/aio/auth/auth.py\n--- a/auth/gcloud/aio/auth/auth.py\n+++ b/auth/gcloud/aio/auth/auth.py\n@@ -1,15 +1,16 @@\n \"\"\"\n Google Cloud auth via service account file\n \"\"\"\n+import asyncio\n import datetime\n import json\n import time\n import typing\n+from urllib.parse import quote_plus\n+from urllib.parse import urlencode\n \n import aiohttp\n import jwt\n-from gcloud.aio.core.aio import auto\n-from gcloud.aio.core.http import post\n \n \n ScopeList = typing.List[str]\n@@ -20,29 +21,24 @@\n 'project_id.'\n \n \n-async def acquire_token(session: aiohttp.ClientSession,\n- service_data: dict,\n+async def acquire_token(session: aiohttp.ClientSession, service_data: dict,\n scopes: ScopeList = None):\n-\n url, assertion = generate_assertion(service_data, scopes)\n \n- payload = {\n- 'grant_type': JWT_GRANT_TYPE,\n- 'assertion': assertion\n+ headers = {\n+ 'Content-Type': 'application/x-www-form-urlencoded',\n }\n+ payload = urlencode({\n+ 'assertion': assertion,\n+ 'grant_type': JWT_GRANT_TYPE,\n+ }, quote_via=quote_plus)\n \n- _status, content = await post(\n- url,\n- payload,\n- headers={'content-type': 'application/x-www-form-urlencoded'},\n- timeout=60,\n- urlencoded=True,\n- json_response=True,\n- session=session\n- )\n+ response = await session.post(url, data=payload, headers=headers,\n+ params=None, timeout=60)\n+ content = await response.json()\n \n if 'error' in content:\n- raise Exception('{}'.format(content))\n+ raise Exception(f'got error acquiring token: {content}')\n \n return {\n 'access_token': str(content['access_token']),\n@@ -100,7 +96,7 @@ def __init__(self, project: str, service_file: str,\n \n self.scopes = scopes or []\n \n- self.session = session\n+ self.session = session or aiohttp.ClientSession()\n self.access_token = None\n self.access_token_duration = None\n self.access_token_acquired_at = None\n@@ -121,7 +117,7 @@ async def ensure_token(self):\n \n elif not self.access_token:\n \n- self.acquiring = self.acquire_access_token()\n+ self.acquiring = asyncio.ensure_future(self.acquire_access_token())\n \n await self.acquiring\n \n@@ -132,11 +128,11 @@ async def ensure_token(self):\n \n if delta > self.access_token_duration / 2:\n \n- self.acquiring = self.acquire_access_token()\n+ self.acquiring = asyncio.ensure_future(\n+ self.acquire_access_token())\n \n await self.acquiring\n \n- @auto\n async def acquire_access_token(self):\n \n data = await acquire_token(\ndiff --git a/auth/nox.py b/auth/nox.py\n--- a/auth/nox.py\n+++ b/auth/nox.py\n@@ -4,16 +4,13 @@\n import nox\n \n \n-LOCAL_DEPS = ('../core/',)\n-\n-\n @nox.session\n @nox.parametrize('python_version', ['3.6'])\n def unit_tests(session, python_version):\n- session.interpreter = 'python{}'.format(python_version)\n- session.virtualenv_dirname = 'unit-' + python_version\n+ session.interpreter = f'python{python_version}'\n+ session.virtualenv_dirname = f'unit-{python_version}'\n \n- session.install('pytest', 'pytest-cov', *LOCAL_DEPS)\n+ session.install('pytest', 'pytest-cov')\n session.install('-e', '.')\n \n session.run(\n@@ -23,7 +20,7 @@ def unit_tests(session, python_version):\n '--cov=tests.unit',\n '--cov-append',\n '--cov-report=',\n- '--cov-fail-under=39',\n+ '--cov-fail-under=38',\n os.path.join('tests', 'unit'),\n *session.posargs)\n \n@@ -34,10 +31,10 @@ def integration_tests(session, python_version):\n if not os.environ.get('GOOGLE_APPLICATION_CREDENTIALS', ''):\n session.skip('Credentials must be set via environment variable.')\n \n- session.interpreter = 'python{}'.format(python_version)\n- session.virtualenv_dirname = 'integration-' + python_version\n+ session.interpreter = f'python{python_version}'\n+ session.virtualenv_dirname = f'integration-{python_version}'\n \n- session.install('aiohttp', 'pytest', *LOCAL_DEPS)\n+ session.install('aiohttp', 'pytest')\n session.install('.')\n \n session.run('py.test', '--quiet', 'tests/integration')\n@@ -46,7 +43,7 @@ def integration_tests(session, python_version):\n @nox.session\n @nox.parametrize('python_version', ['3.6'])\n def lint_setup_py(session, python_version):\n- session.interpreter = 'python{}'.format(python_version)\n+ session.interpreter = f'python{python_version}'\n session.virtualenv_dirname = 'setup'\n \n session.install('docutils', 'Pygments')\n@@ -61,7 +58,7 @@ def lint_setup_py(session, python_version):\n @nox.session\n @nox.parametrize('python_version', ['3.6'])\n def cover(session, python_version):\n- session.interpreter = 'python{}'.format(python_version)\n+ session.interpreter = f'python{python_version}'\n session.virtualenv_dirname = 'cover'\n \n session.install('codecov', 'coverage', 'pytest-cov')\ndiff --git a/bigquery/gcloud/aio/bigquery/bigquery.py b/bigquery/gcloud/aio/bigquery/bigquery.py\n--- a/bigquery/gcloud/aio/bigquery/bigquery.py\n+++ b/bigquery/gcloud/aio/bigquery/bigquery.py\n@@ -2,9 +2,12 @@\n import logging\n import uuid\n \n-import ujson\n+import aiohttp\n from gcloud.aio.auth import Token\n-from gcloud.aio.core.http import post\n+try:\n+ import ujson as json\n+except ModuleNotFoundError:\n+ import json\n \n \n API_ROOT = 'https://www.googleapis.com/bigquery/v2'\n@@ -63,51 +66,43 @@ async def headers(self):\n token = await self.token.get()\n \n return {\n- 'Authorization': 'Bearer {}'.format(token)\n+ 'Authorization': f'Bearer {token}',\n }\n \n async def insert(self, rows, skip_invalid=False, ignore_unknown=True,\n session=None):\n-\n session = session or self.session\n \n- body = make_insert_body(\n- rows,\n- skip_invalid=skip_invalid,\n- ignore_unknown=ignore_unknown\n- )\n-\n- headers = await self.headers()\n-\n- url = '{}/{}'.format(\n- API_ROOT,\n- INSERT_TEMPLATE.format(\n- proj=self.project,\n- dataset=self.dataset_name,\n- table=self.table_name\n- )\n- )\n-\n+ insert_url = INSERT_TEMPLATE.format(proj=self.project,\n+ dataset=self.dataset_name,\n+ table=self.table_name)\n+ url = f'{API_ROOT}/{insert_url}'\n log.info('Inserting %d rows to %s', len(rows), url)\n \n- status, content = await post(\n- url,\n- payload=body,\n- headers=headers\n- )\n+ body = make_insert_body(rows, skip_invalid=skip_invalid,\n+ ignore_unknown=ignore_unknown)\n+ payload = json.dumps(body).encode('utf-8')\n+\n+ headers = await self.headers()\n+ headers.update({\n+ 'Content-Length': str(len(payload)),\n+ 'Content-Type': 'application/json'\n+ })\n \n- success = 299 >= status >= 200 and 'insertErrors' not in content\n+ async with aiohttp.ClientSession() as s:\n+ response = await s.post(url, data=payload, headers=headers,\n+ params=None, timeout=60)\n+ content = await response.json()\n \n- if success:\n- return success\n+ if 299 >= response.status >= 200 and 'insertErrors' not in content:\n+ return True\n \n- log.debug('response code: %d', status)\n+ log.debug('response code: %d', response.status)\n log.debug('url: %s', url)\n- log.debug('body:\\n%s\\n', body)\n+ log.debug('body:\\n%s\\n', payload)\n \n- raise Exception('Could not insert: {}'.format(ujson.dumps(\n- content, sort_keys=True\n- )))\n+ content_blob = json.dumps(content, sort_keys=True)\n+ raise Exception(f'could not insert: {content_blob}')\n \n \n async def stream_insert(table, rows):\ndiff --git a/bigquery/nox.py b/bigquery/nox.py\n--- a/bigquery/nox.py\n+++ b/bigquery/nox.py\n@@ -4,14 +4,14 @@\n import nox\n \n \n-LOCAL_DEPS = ('../core/', '../auth/')\n+LOCAL_DEPS = ('../auth/', )\n \n \n @nox.session\n @nox.parametrize('python_version', ['3.6'])\n def unit_tests(session, python_version):\n- session.interpreter = 'python{}'.format(python_version)\n- session.virtualenv_dirname = 'unit-' + python_version\n+ session.interpreter = f'python{python_version}'\n+ session.virtualenv_dirname = f'unit-{python_version}'\n \n session.install('pytest', 'pytest-cov', *LOCAL_DEPS)\n session.install('-e', '.')\n@@ -23,7 +23,7 @@ def unit_tests(session, python_version):\n '--cov=tests.unit',\n '--cov-append',\n '--cov-report=',\n- '--cov-fail-under=47',\n+ '--cov-fail-under=46',\n os.path.join('tests', 'unit'),\n *session.posargs)\n \n@@ -34,8 +34,8 @@ def integration_tests(session, python_version):\n if not os.environ.get('GOOGLE_APPLICATION_CREDENTIALS', ''):\n session.skip('Credentials must be set via environment variable.')\n \n- session.interpreter = 'python{}'.format(python_version)\n- session.virtualenv_dirname = 'integration-' + python_version\n+ session.interpreter = f'python{python_version}'\n+ session.virtualenv_dirname = f'integration-{python_version}'\n \n session.install('aiohttp', 'pytest', *LOCAL_DEPS)\n session.install('.')\n@@ -46,7 +46,7 @@ def integration_tests(session, python_version):\n @nox.session\n @nox.parametrize('python_version', ['3.6'])\n def lint_setup_py(session, python_version):\n- session.interpreter = 'python{}'.format(python_version)\n+ session.interpreter = f'python{python_version}'\n session.virtualenv_dirname = 'setup'\n \n session.install('docutils', 'Pygments')\n@@ -61,11 +61,11 @@ def lint_setup_py(session, python_version):\n @nox.session\n @nox.parametrize('python_version', ['3.6'])\n def cover(session, python_version):\n- session.interpreter = 'python{}'.format(python_version)\n+ session.interpreter = f'python{python_version}'\n session.virtualenv_dirname = 'cover'\n \n session.install('codecov', 'coverage', 'pytest-cov')\n \n- session.run('coverage', 'report', '--show-missing', '--fail-under=47')\n+ session.run('coverage', 'report', '--show-missing', '--fail-under=46')\n session.run('codecov')\n session.run('coverage', 'erase')\ndiff --git a/core/gcloud/__init__.py b/core/gcloud/__init__.py\ndeleted file mode 100644\n--- a/core/gcloud/__init__.py\n+++ /dev/null\n@@ -1,6 +0,0 @@\n-try:\n- import pkg_resources\n- pkg_resources.declare_namespace(__name__)\n-except ImportError:\n- import pkgutil\n- __path__ = pkgutil.extend_path(__path__, __name__)\ndiff --git a/core/gcloud/aio/__init__.py b/core/gcloud/aio/__init__.py\ndeleted file mode 100644\n--- a/core/gcloud/aio/__init__.py\n+++ /dev/null\n@@ -1,6 +0,0 @@\n-try:\n- import pkg_resources\n- pkg_resources.declare_namespace(__name__)\n-except ImportError:\n- import pkgutil\n- __path__ = pkgutil.extend_path(__path__, __name__)\ndiff --git a/core/gcloud/aio/core/__init__.py b/core/gcloud/aio/core/__init__.py\ndeleted file mode 100644\n--- a/core/gcloud/aio/core/__init__.py\n+++ /dev/null\n@@ -1,5 +0,0 @@\n-from pkg_resources import get_distribution\n-__version__ = get_distribution('gcloud-aio-core').version\n-\n-\n-__all__ = ['__version__']\ndiff --git a/core/gcloud/aio/core/aio.py b/core/gcloud/aio/core/aio.py\ndeleted file mode 100644\n--- a/core/gcloud/aio/core/aio.py\n+++ /dev/null\n@@ -1,66 +0,0 @@\n-import asyncio\n-import functools\n-\n-\n-def maybe_async(callable_, *args, **kwargs):\n-\n- \"\"\"\n- Turn a callable into a coroutine if it isn't\n- \"\"\"\n-\n- if asyncio.iscoroutine(callable_):\n- return callable_\n-\n- return asyncio.coroutine(callable_)(*args, **kwargs)\n-\n-\n-def fire(callable_, *args, **kwargs):\n-\n- \"\"\"\n- Start a callable as a coroutine, and return it's future. The cool thing\n- about this function is that (via maybe_async) it lets you treat synchronous\n- and asynchronous callables the same (both as async), which simplifies code.\n- \"\"\"\n-\n- return asyncio.ensure_future(maybe_async(callable_, *args, **kwargs))\n-\n-\n-def auto(fn):\n-\n- \"\"\"\n- Decorate a function or method with this, and it will become a callable\n- that can be scheduled in the event loop just by calling it. Normally you'd\n- have to do an `asyncio.ensure_future(my_callable())`. Not you can just do\n- `my_callable()`. Twisted has always let you do this, and now you can let\n- asyncio do it as well (with a decorator, albeit...)\n- \"\"\"\n-\n- @functools.wraps(fn)\n- def wrapper(*args, **kwargs):\n-\n- return fire(fn, *args, **kwargs)\n-\n- return wrapper\n-\n-\n-async def _call_later(delay, callable_, *args, **kwargs):\n-\n- \"\"\"\n- The bus stop, where we wait.\n- \"\"\"\n-\n- await asyncio.sleep(delay)\n-\n- fire(callable_, *args, **kwargs)\n-\n-\n-def call_later(delay, callable_, *args, **kwargs):\n-\n- \"\"\"\n- After :delay seconds, call :callable with :args and :kwargs; :callable can\n- be a synchronous or asynchronous callable (a coroutine). Note that _this_\n- function is synchronous - mission accomplished - it can be used from within\n- any synchronous or asynchronous callable.\n- \"\"\"\n-\n- return fire(_call_later, delay, callable_, *args, **kwargs)\ndiff --git a/core/gcloud/aio/core/astate.py b/core/gcloud/aio/core/astate.py\ndeleted file mode 100644\n--- a/core/gcloud/aio/core/astate.py\n+++ /dev/null\n@@ -1,83 +0,0 @@\n-import asyncio\n-import logging\n-\n-from gcloud.aio.core.aio import fire\n-\n-\n-log = logging.getLogger(__name__)\n-\n-\n-class AwaitableState:\n- # pylint: disable=too-few-public-methods\n-\n- \"\"\"\n- Wrap a :future with a name and data. If :future is a coroutine, turn it\n- into a future by firing it.\n-\n- Use instances of AwaitableState as named states in state machines. Use\n- :data for arbitrary context beyond :name.\n- \"\"\"\n-\n- def __init__(self, name, future, data=None):\n-\n- self.name = name\n- self.future = future\n- self.data = data\n-\n- if asyncio.iscoroutine(self.future):\n- self.future = fire(self.future)\n-\n- def __await__(self):\n-\n- return self.future.__await__()\n-\n- def __str__(self):\n-\n- return self.__repr__()\n-\n- def __repr__(self):\n-\n- return ''.format(\n- self.name,\n- id(self)\n- )\n-\n- def __getattr__(self, attr):\n-\n- return getattr(self.future, attr)\n-\n- def __hash__(self):\n-\n- return hash(self.name)\n-\n- def __eq__(self, other):\n-\n- return hash(self) == hash(other)\n-\n-\n-def make_stepper(default_step, state_step, name='sm'):\n-\n- \"\"\"\n- `default_step`: a callable that takes no args\n- `state_step`: a mapping between AwaitableState.name -> callable\n- \"\"\"\n-\n- async def step(state, args):\n-\n- state_name = getattr(state, 'name', None)\n- step = state_step.get(state_name, default_step)\n- next_state = step(args) if args is not None else step()\n-\n- if next_state:\n- args = await next_state\n- else:\n- args = tuple()\n-\n- if next_state != state:\n- log.debug('%s state change: %s -> %s', name,\n- getattr(state, 'name', None),\n- getattr(next_state, 'name', None))\n-\n- return next_state, args\n-\n- return step\ndiff --git a/core/gcloud/aio/core/http.py b/core/gcloud/aio/core/http.py\ndeleted file mode 100644\n--- a/core/gcloud/aio/core/http.py\n+++ /dev/null\n@@ -1,140 +0,0 @@\n-from urllib.parse import quote_plus\n-from urllib.parse import urlencode\n-\n-import aiohttp\n-import ujson\n-from asyncio_extras.contextmanager import async_contextmanager\n-\n-\n-class HttpError(Exception):\n- pass\n-\n-\n-@async_contextmanager\n-async def ensure_session(session):\n-\n- if session:\n- yield session\n- else:\n- async with aiohttp.ClientSession() as session:\n- yield session\n-\n-\n-async def delete(url, headers=None, params=None, timeout=60, session=None):\n-\n- async with ensure_session(session) as s: # pylint: disable=not-async-context-manager\n-\n- response = await s.delete(\n- url,\n- headers=headers,\n- params=params,\n- timeout=timeout\n- )\n-\n- phrase = await response.text()\n-\n- return response.status, phrase\n-\n-\n-async def post(url, payload=None, timeout=60, urlencoded=False,\n- json_response=True, session=None, headers=None, params=None):\n- # pylint: disable=too-many-arguments\n-\n- headers = headers or {}\n-\n- if urlencoded:\n-\n- if payload:\n- payload = urlencode(payload, quote_via=quote_plus)\n-\n- headers['content-type'] = 'application/x-www-form-urlencoded'\n-\n- else:\n-\n- if payload:\n- payload = ujson.dumps(payload)\n- payload = payload.encode('utf-8')\n- content_length = str(len(payload))\n- else:\n- content_length = '0'\n-\n- headers.update({\n- 'content-length': content_length,\n- 'content-type': 'application/json'\n- })\n-\n- async with ensure_session(session) as s: # pylint: disable=not-async-context-manager\n-\n- response = await s.post(\n- url,\n- data=payload,\n- headers=headers,\n- params=params,\n- timeout=timeout\n- )\n-\n- if json_response:\n- content = await response.json()\n- else:\n- content = await response.text()\n-\n- return response.status, content\n-\n-\n-async def get(url, timeout=60, json_response=True, session=None, headers=None,\n- params=None):\n- # pylint: disable=too-many-arguments\n-\n- async with ensure_session(session) as s: # pylint: disable=not-async-context-manager\n-\n- response = await s.get(\n- url,\n- headers=headers,\n- params=params,\n- timeout=timeout\n- )\n-\n- if json_response:\n- content = await response.json()\n- else:\n- content = await response.text()\n-\n- return response.status, content\n-\n-\n-async def put(*args, **kwargs): # pylint: disable=unused-argument\n-\n- raise Exception('Not implemented.')\n-\n-\n-async def patch(url, payload=None, timeout=60, session=None, headers=None,\n- params=None):\n- # pylint: disable=too-many-arguments\n-\n- headers = headers or {}\n-\n- if payload:\n- payload = ujson.dumps(payload)\n- payload = payload.encode('utf-8')\n- content_length = str(len(payload))\n- else:\n- content_length = '0'\n-\n- headers.update({\n- 'content-length': content_length,\n- 'content-type': 'application/json'\n- })\n-\n- async with ensure_session(session) as s: # pylint: disable=not-async-context-manager\n-\n- response = await s.patch(\n- url,\n- data=payload,\n- headers=headers,\n- params=params,\n- timeout=timeout\n- )\n-\n- phrase = await response.text()\n-\n- return response.status, phrase\ndiff --git a/core/nox.py b/core/nox.py\ndeleted file mode 100644\n--- a/core/nox.py\n+++ /dev/null\n@@ -1,53 +0,0 @@\n-# pylint: disable=import-self,no-member\n-import os\n-\n-import nox\n-\n-\n-@nox.session\n-@nox.parametrize('python_version', ['3.6'])\n-def unit_tests(session, python_version):\n- session.interpreter = 'python{}'.format(python_version)\n- session.virtualenv_dirname = 'unit-' + python_version\n-\n- session.install('pytest', 'pytest-cov')\n- session.install('-e', '.')\n-\n- session.run(\n- 'py.test',\n- '--quiet',\n- '--cov=gcloud.aio.core',\n- '--cov=tests.unit',\n- '--cov-append',\n- '--cov-report=',\n- '--cov-fail-under=37',\n- os.path.join('tests', 'unit'),\n- *session.posargs)\n-\n-\n-@nox.session\n-@nox.parametrize('python_version', ['3.6'])\n-def lint_setup_py(session, python_version):\n- session.interpreter = 'python{}'.format(python_version)\n- session.virtualenv_dirname = 'setup'\n-\n- session.install('docutils', 'Pygments')\n- session.run(\n- 'python',\n- 'setup.py',\n- 'check',\n- '--restructuredtext',\n- '--strict')\n-\n-\n-@nox.session\n-@nox.parametrize('python_version', ['3.6'])\n-def cover(session, python_version):\n- session.interpreter = 'python{}'.format(python_version)\n- session.virtualenv_dirname = 'cover'\n-\n- session.install('codecov', 'coverage', 'pytest-cov')\n-\n- session.run('coverage', 'report', '--show-missing', '--fail-under=37')\n- session.run('codecov')\n- session.run('coverage', 'erase')\ndiff --git a/core/setup.py b/core/setup.py\ndeleted file mode 100644\n--- a/core/setup.py\n+++ /dev/null\n@@ -1,41 +0,0 @@\n-import os\n-\n-import setuptools\n-\n-\n-PACKAGE_ROOT = os.path.abspath(os.path.dirname(__file__))\n-with open(os.path.join(PACKAGE_ROOT, 'README.rst')) as f:\n- README = f.read()\n-\n-with open(os.path.join(PACKAGE_ROOT, 'requirements.txt')) as f:\n- REQUIREMENTS = [r.strip() for r in f.readlines()]\n-\n-\n-setuptools.setup(\n- name='gcloud-aio-core',\n- version='0.7.2',\n- description='Core Helpers for Asyncio Google Cloud Library',\n- long_description=README,\n- namespace_packages=[\n- 'gcloud',\n- 'gcloud.aio',\n- ],\n- packages=setuptools.find_packages(exclude=('tests',)),\n- install_requires=REQUIREMENTS,\n- author='TalkIQ',\n- author_email='engineering@talkiq.com',\n- url='https://github.com/talkiq/gcloud-aio',\n- platforms='Posix; MacOS X; Windows',\n- include_package_data=True,\n- zip_safe=False,\n- license='MIT License',\n- classifiers=[\n- 'Development Status :: 4 - Beta',\n- 'Intended Audience :: Developers',\n- 'License :: OSI Approved :: MIT License',\n- 'Operating System :: OS Independent',\n- 'Programming Language :: Python :: 3',\n- 'Programming Language :: Python :: 3.6',\n- 'Topic :: Internet',\n- ],\n-)\ndiff --git a/datastore/gcloud/aio/datastore/datastore.py b/datastore/gcloud/aio/datastore/datastore.py\n--- a/datastore/gcloud/aio/datastore/datastore.py\n+++ b/datastore/gcloud/aio/datastore/datastore.py\n@@ -1,11 +1,15 @@\n import datetime\n import logging\n \n+import aiohttp\n from gcloud.aio.auth import Token\n-from gcloud.aio.core.http import post\n from gcloud.aio.datastore.constants import Mode\n from gcloud.aio.datastore.constants import Operation\n from gcloud.aio.datastore.constants import TypeName\n+try:\n+ import ujson as json\n+except ModuleNotFoundError:\n+ import json\n \n \n API_ROOT = 'https://datastore.googleapis.com/v1/projects'\n@@ -30,9 +34,7 @@ def infer_type(value):\n }.get(type(value))\n \n if not type_name:\n- raise Exception('Type {} not supported for DS insert. :('.format(\n- type(value)\n- ))\n+ raise Exception(f'type {type(value)} not supported for DS insert')\n \n return type_name\n \n@@ -109,45 +111,54 @@ async def headers(self):\n token = await self.token.get()\n \n return {\n- 'Authorization': 'Bearer {}'.format(token),\n+ 'Authorization': f'Bearer {token}',\n }\n \n async def transact(self):\n- url = '{}/{}:beginTransaction'.format(API_ROOT, self.project)\n+ url = f'{API_ROOT}/{self.project}:beginTransaction'\n headers = await self.headers()\n- body = {}\n+ headers.update({\n+ 'Content-Length': '0',\n+ 'Content-Type': 'application/json'\n+ })\n \n- status, content = await post(url, payload={}, headers=headers)\n+ async with aiohttp.ClientSession() as s:\n+ response = await s.post(url, data={}, headers=headers, params=None,\n+ timeout=60)\n+ content = await response.json()\n \n # TODO: make this raise_for_status-able.\n- success = 299 >= status >= 200\n-\n- if success:\n+ if 299 >= response.status >= 200:\n transaction = content['transaction']\n return transaction\n \n- log.debug('response code: %d', status)\n+ log.debug('response code: %d', response.status)\n log.debug('url: %s', url)\n- log.debug('body:\\n%s\\n', body)\n \n- raise Exception('Could not transact: {}'.format(content))\n+ raise Exception(f'could not transact: {content}')\n \n async def commit(self, transaction, mutations, mode=Mode.TRANSACTIONAL):\n- url = '{}/{}:commit'.format(API_ROOT, self.project)\n+ url = f'{API_ROOT}/{self.project}:commit'\n \n body = make_commit_body(transaction, mode, mutations)\n+ payload = json.dumps(body).encode('utf-8')\n \n headers = await self.headers()\n+ headers.update({\n+ 'Content-Length': str(len(payload)),\n+ 'Content-Type': 'application/json'\n+ })\n \n- status, content = await post(url, payload=body, headers=headers)\n+ async with aiohttp.ClientSession() as s:\n+ response = await s.post(url, data=payload, headers=headers,\n+ params=None, timeout=60)\n+ content = await response.json()\n \n # TODO: make this raise_for_status-able.\n- success = 299 >= status >= 200 and 'insertErrors' not in content\n-\n- if success:\n- return success\n+ if 299 >= response.status >= 200 and 'insertErrors' not in content:\n+ return True\n \n- raise Exception('{}: {} > {}'.format(status, url, content))\n+ raise Exception(f'{response.status}: {url} > {content}')\n \n # TODO: look into deletion payload format\n \ndiff --git a/datastore/nox.py b/datastore/nox.py\n--- a/datastore/nox.py\n+++ b/datastore/nox.py\n@@ -4,14 +4,14 @@\n import nox\n \n \n-LOCAL_DEPS = ('../core/', '../auth/')\n+LOCAL_DEPS = ('../auth/', )\n \n \n @nox.session\n @nox.parametrize('python_version', ['3.6'])\n def unit_tests(session, python_version):\n- session.interpreter = 'python{}'.format(python_version)\n- session.virtualenv_dirname = 'unit-' + python_version\n+ session.interpreter = f'python{python_version}'\n+ session.virtualenv_dirname = f'unit-{python_version}'\n \n session.install('pytest', 'pytest-cov', *LOCAL_DEPS)\n session.install('-e', '.')\n@@ -34,8 +34,8 @@ def integration_tests(session, python_version):\n if not os.environ.get('GOOGLE_APPLICATION_CREDENTIALS', ''):\n session.skip('Credentials must be set via environment variable.')\n \n- session.interpreter = 'python{}'.format(python_version)\n- session.virtualenv_dirname = 'integration-' + python_version\n+ session.interpreter = f'python{python_version}'\n+ session.virtualenv_dirname = f'integration-{python_version}'\n \n session.install('pytest', *LOCAL_DEPS)\n session.install('.')\n@@ -46,7 +46,7 @@ def integration_tests(session, python_version):\n @nox.session\n @nox.parametrize('python_version', ['3.6'])\n def lint_setup_py(session, python_version):\n- session.interpreter = 'python{}'.format(python_version)\n+ session.interpreter = f'python{python_version}'\n session.virtualenv_dirname = 'setup'\n \n session.install('docutils', 'Pygments')\n@@ -61,7 +61,7 @@ def lint_setup_py(session, python_version):\n @nox.session\n @nox.parametrize('python_version', ['3.6'])\n def cover(session, python_version):\n- session.interpreter = 'python{}'.format(python_version)\n+ session.interpreter = f'python{python_version}'\n session.virtualenv_dirname = 'cover'\n \n session.install('codecov', 'coverage', 'pytest-cov')\ndiff --git a/pubsub/nox.py b/pubsub/nox.py\n--- a/pubsub/nox.py\n+++ b/pubsub/nox.py\n@@ -7,8 +7,8 @@\n @nox.session\n @nox.parametrize('python_version', ['3.6'])\n def unit_tests(session, python_version):\n- session.interpreter = 'python{}'.format(python_version)\n- session.virtualenv_dirname = 'unit-' + python_version\n+ session.interpreter = f'python{python_version}'\n+ session.virtualenv_dirname = f'unit-{python_version}'\n \n session.install('pytest', 'pytest-cov')\n session.install('-e', '.')\n@@ -31,8 +31,8 @@ def integration_tests(session, python_version):\n if not os.environ.get('GOOGLE_APPLICATION_CREDENTIALS', ''):\n session.skip('Credentials must be set via environment variable.')\n \n- session.interpreter = 'python{}'.format(python_version)\n- session.virtualenv_dirname = 'integration-' + python_version\n+ session.interpreter = f'python{python_version}'\n+ session.virtualenv_dirname = f'integration-{python_version}'\n \n session.install('pytest')\n session.install('.')\n@@ -43,7 +43,7 @@ def integration_tests(session, python_version):\n @nox.session\n @nox.parametrize('python_version', ['3.6'])\n def lint_setup_py(session, python_version):\n- session.interpreter = 'python{}'.format(python_version)\n+ session.interpreter = f'python{python_version}'\n session.virtualenv_dirname = 'setup'\n \n session.install('docutils', 'Pygments')\n@@ -58,7 +58,7 @@ def lint_setup_py(session, python_version):\n @nox.session\n @nox.parametrize('python_version', ['3.6'])\n def cover(session, python_version):\n- session.interpreter = 'python{}'.format(python_version)\n+ session.interpreter = f'python{python_version}'\n session.virtualenv_dirname = 'cover'\n \n session.install('codecov', 'coverage', 'pytest-cov')\ndiff --git a/storage/gcloud/aio/storage/__init__.py b/storage/gcloud/aio/storage/__init__.py\n--- a/storage/gcloud/aio/storage/__init__.py\n+++ b/storage/gcloud/aio/storage/__init__.py\n@@ -1,10 +1,10 @@\n from pkg_resources import get_distribution\n __version__ = get_distribution('gcloud-aio-storage').version\n \n-from gcloud.aio.storage.storage import Blob\n-from gcloud.aio.storage.storage import Bucket\n-from gcloud.aio.storage.storage import make_download\n+from gcloud.aio.storage.blob import Blob\n+from gcloud.aio.storage.bucket import Bucket\n from gcloud.aio.storage.storage import Storage\n+from gcloud.aio.storage.utils import make_download\n \n \n-__all__ = ['__version__', 'Blob', 'Bucket', 'make_download', 'Storage']\n+__all__ = ['__version__', 'Blob', 'Bucket', 'Storage', 'make_download']\ndiff --git a/storage/gcloud/aio/storage/blob.py b/storage/gcloud/aio/storage/blob.py\nnew file mode 100644\n--- /dev/null\n+++ b/storage/gcloud/aio/storage/blob.py\n@@ -0,0 +1,33 @@\n+try:\n+ import ujson as json\n+except ModuleNotFoundError:\n+ import json\n+\n+\n+class Blob:\n+ def __init__(self, bucket, name, data):\n+ self.__dict__.update(**data)\n+\n+ self.bucket = bucket\n+ self.name = name\n+ self.size = int(self.size)\n+\n+ @property\n+ def chunk_size(self):\n+ return self.size + (262144 - (self.size % 262144))\n+\n+ async def download_as_string(self, session=None):\n+ return await self.bucket.storage.download_as_string(self.bucket.name,\n+ self.name,\n+ session=session)\n+\n+ async def upload_from_string(self, data, session=None):\n+ status, content = await self.bucket.storage.upload(self.bucket.name,\n+ self.name, data,\n+ session=session)\n+\n+ if status < 200 or status >= 300:\n+ raise Exception(f'{status}: {json.dumps(content)}')\n+\n+ self.__dict__.update(content)\n+ return content\ndiff --git a/storage/gcloud/aio/storage/bucket.py b/storage/gcloud/aio/storage/bucket.py\nnew file mode 100644\n--- /dev/null\n+++ b/storage/gcloud/aio/storage/bucket.py\n@@ -0,0 +1,47 @@\n+import logging\n+\n+from gcloud.aio.storage.blob import Blob\n+try:\n+ import ujson as json\n+except ModuleNotFoundError:\n+ import json\n+\n+\n+log = logging.getLogger(__name__)\n+\n+\n+class Bucket:\n+ def __init__(self, storage, name):\n+ self.storage = storage\n+ self.name = name\n+\n+ async def get_blob(self, blob_name, session=None):\n+ blob_name = blob_name.replace('/', '%2F')\n+\n+ status, content = await self.storage.download(self.name, blob_name,\n+ session=session)\n+\n+ if status < 200 or status >= 300:\n+ log.error('Could not download %s/%s: %s', self.name, blob_name,\n+ content)\n+ return\n+\n+ content = json.loads(content)\n+\n+ return Blob(self, blob_name, content)\n+\n+ async def list_blobs(self, prefix='', session=None):\n+ params = {'prefix': prefix}\n+\n+ status, content = await self.storage.list_objects(self.name,\n+ params=params,\n+ session=session)\n+\n+ if status < 200 or status >= 300:\n+ log.error('Could not list %s/%s: %s', self.name, prefix, content)\n+ return\n+\n+ return [x['name'] for x in content.get('items', list())]\n+\n+ def new_blob(self, blob_name):\n+ return Blob(self, blob_name, {'size': 0})\ndiff --git a/storage/gcloud/aio/storage/storage.py b/storage/gcloud/aio/storage/storage.py\n--- a/storage/gcloud/aio/storage/storage.py\n+++ b/storage/gcloud/aio/storage/storage.py\n@@ -1,18 +1,16 @@\n-import functools\n import logging\n-import mimetypes\n \n import aiohttp\n-import ujson\n from gcloud.aio.auth import Token\n-from gcloud.aio.core.http import get\n-from gcloud.aio.core.http import HttpError\n-from gcloud.aio.core.http import post\n+from gcloud.aio.storage.bucket import Bucket\n+try:\n+ import ujson as json\n+except ModuleNotFoundError:\n+ import json\n \n \n STORAGE_API_ROOT = 'https://www.googleapis.com/storage/v1/b'\n STORAGE_UPLOAD_API_ROOT = 'https://www.googleapis.com/upload/storage/v1/b'\n-READ_ONLY_SCOPE = 'https://www.googleapis.com/auth/devstorage.read_only'\n READ_WRITE_SCOPE = 'https://www.googleapis.com/auth/devstorage.read_write'\n \n log = logging.getLogger(__name__)\n@@ -28,64 +26,68 @@ def __init__(self, project, service_file, token=None, session=None):\n scopes=[READ_WRITE_SCOPE])\n \n async def download(self, bucket, object_name, params=None, session=None):\n- session = session or self.session\n-\n token = await self.token.get()\n- url = '{}/{}/o/{}'.format(STORAGE_API_ROOT, bucket, object_name)\n+ url = f'{STORAGE_API_ROOT}/{bucket}/o/{object_name}'\n headers = {\n- 'Authorization': 'Bearer {}'.format(token),\n+ 'Authorization': f'Bearer {token}',\n }\n \n- return await get(url, params=params or {}, headers=headers,\n- session=self.session, json_response=False)\n-\n- async def list_objects(self, bucket, params=None, session=None):\n session = session or self.session\n+ response = await session.get(url, headers=headers, params=params or {},\n+ timeout=60)\n+ content = await response.text()\n+\n+ return response.status, content\n \n+ async def list_objects(self, bucket, params=None, session=None):\n token = await self.token.get()\n- url = '{}/{}/o'.format(STORAGE_API_ROOT, bucket)\n+ url = f'{STORAGE_API_ROOT}/{bucket}/o'\n headers = {\n- 'Authorization': 'Bearer {}'.format(token),\n+ 'Authorization': f'Bearer {token}',\n }\n \n- return await get(url, params=params or {}, headers=headers,\n- session=self.session, json_response=True)\n+ session = session or self.session\n+ response = await session.get(url, headers=headers, params=params or {},\n+ timeout=60)\n+ content = await response.json()\n+\n+ return response.status, content\n \n async def upload(self, bucket, object_name, file_data, headers=None,\n session=None):\n # pylint: disable=too-many-arguments\n # https://cloud.google.com/storage/docs/json_api/v1/how-tos/simple-upload\n- session = session or self.session\n-\n token = await self.token.get()\n- url = '{}/{}/o'.format(STORAGE_UPLOAD_API_ROOT, bucket)\n+ url = f'{STORAGE_UPLOAD_API_ROOT}/{bucket}/o'\n headers = headers or {}\n \n- # TODO: verify this\n- if not isinstance(file_data, bytes):\n- body = file_data.encode('utf-8')\n- else:\n- body = file_data\n-\n- body_length = str(len(body))\n-\n params = {\n 'name': object_name,\n 'uploadType': 'media',\n }\n \n- content_type = mimetypes.guess_type(object_name)[0]\n- content_type = content_type or 'application/octet-stream'\n+ if not isinstance(file_data, bytes):\n+ file_data = file_data.encode('utf-8')\n+\n+ if file_data:\n+ file_data = json.dumps(file_data).encode('utf-8')\n+ content_length = str(len(file_data))\n+ else:\n+ content_length = '0'\n \n headers.update({\n- 'accept': 'application/json',\n- 'Authorization': 'Bearer {}'.format(token),\n- 'Content-Length': body_length,\n- 'Content-Type': content_type,\n+ 'Accept': 'application/json',\n+ 'Authorization': f'Bearer {token}',\n+ 'Content-Length': content_length,\n+ 'Content-Type': 'application/json',\n })\n \n- return await post(url, params=params, payload=body, headers=headers,\n- timeout=120, session=session)\n+ session = session or self.session\n+ response = await session.post(url, data=file_data, headers=headers,\n+ params=params, timeout=120)\n+ content = await response.json()\n+\n+ return response.status, content\n \n async def download_as_string(self, bucket, object_name, session=None):\n object_name = object_name.replace('/', '%2F')\n@@ -98,88 +100,3 @@ async def download_as_string(self, bucket, object_name, session=None):\n \n def get_bucket(self, bucket_name):\n return Bucket(self, bucket_name)\n-\n-\n-class Bucket:\n- def __init__(self, storage, name):\n- self.storage = storage\n- self.name = name\n-\n- async def get_blob(self, blob_name, session=None):\n- blob_name = blob_name.replace('/', '%2F')\n-\n- status, content = await self.storage.download(self.name, blob_name,\n- session=session)\n-\n- if status < 200 or status >= 300:\n- log.error('Could not download %s/%s: %s', self.name, blob_name,\n- content)\n- return\n-\n- content = ujson.loads(content)\n-\n- return Blob(self, blob_name, content)\n-\n- async def list_blobs(self, prefix='', session=None):\n- params = {'prefix': prefix}\n-\n- status, content = await self.storage.list_objects(self.name,\n- params=params,\n- session=session)\n-\n- if status < 200 or status >= 300:\n- log.error('Could not list %s/%s: %s', self.name, prefix, content)\n- return\n-\n- return [x['name'] for x in content.get('items', list())]\n-\n- def new_blob(self, blob_name):\n- return Blob(self, blob_name, {'size': 0})\n-\n-\n-class Blob:\n- def __init__(self, bucket, name, data):\n- self.__dict__.update(**data)\n-\n- self.bucket = bucket\n- self.name = name\n- self.size = int(self.size)\n-\n- @property\n- def chunk_size(self):\n- return self.size + (262144 - (self.size % 262144))\n-\n- async def download_as_string(self, session=None):\n- return await self.bucket.storage.download_as_string(self.bucket.name,\n- self.name,\n- session=session)\n-\n- async def upload_from_string(self, data, session=None):\n- status, content = await self.bucket.storage.upload(self.bucket.name,\n- self.name, data,\n- session=session)\n-\n- if status < 200 or status >= 300:\n- raise HttpError('{}: {}'.format(status, ujson.dumps(content)))\n-\n- self.__dict__.update(content)\n- return content\n-\n-\n-async def download(bucket, object_name):\n- blob = await bucket.get_blob(object_name)\n- if not blob:\n- raise Exception('No such object \"{}/{}\"'.format(bucket.name,\n- object_name))\n-\n- return await blob.download_as_string()\n-\n-\n-def make_download(project, service_file, bucket_name, session=None,\n- token=None):\n- token = token or Token(project, service_file, scopes=[READ_ONLY_SCOPE])\n-\n- storage = Storage(project, service_file, session=session, token=token)\n- bucket = storage.get_bucket(bucket_name)\n-\n- return functools.partial(download, bucket)\ndiff --git a/storage/gcloud/aio/storage/utils.py b/storage/gcloud/aio/storage/utils.py\nnew file mode 100644\n--- /dev/null\n+++ b/storage/gcloud/aio/storage/utils.py\n@@ -0,0 +1,25 @@\n+import functools\n+\n+from gcloud.aio.auth import Token\n+from gcloud.aio.storage.storage import Storage\n+\n+\n+READ_ONLY_SCOPE = 'https://www.googleapis.com/auth/devstorage.read_only'\n+\n+\n+async def download(bucket, object_name):\n+ blob = await bucket.get_blob(object_name)\n+ if not blob:\n+ raise Exception(f'No such object \"{bucket.name}/{object_name}\"')\n+\n+ return await blob.download_as_string()\n+\n+\n+def make_download(project, service_file, bucket_name, session=None,\n+ token=None):\n+ token = token or Token(project, service_file, scopes=[READ_ONLY_SCOPE])\n+\n+ storage = Storage(project, service_file, session=session, token=token)\n+ bucket = storage.get_bucket(bucket_name)\n+\n+ return functools.partial(download, bucket)\ndiff --git a/storage/nox.py b/storage/nox.py\n--- a/storage/nox.py\n+++ b/storage/nox.py\n@@ -4,14 +4,14 @@\n import nox\n \n \n-LOCAL_DEPS = ('../core/', '../auth/')\n+LOCAL_DEPS = ('../auth/', )\n \n \n @nox.session\n @nox.parametrize('python_version', ['3.6'])\n def unit_tests(session, python_version):\n- session.interpreter = 'python{}'.format(python_version)\n- session.virtualenv_dirname = 'unit-' + python_version\n+ session.interpreter = f'python{python_version}'\n+ session.virtualenv_dirname = f'unit-{python_version}'\n \n session.install('pytest', 'pytest-cov', *LOCAL_DEPS)\n session.install('-e', '.')\n@@ -34,8 +34,8 @@ def integration_tests(session, python_version):\n if not os.environ.get('GOOGLE_APPLICATION_CREDENTIALS', ''):\n session.skip('Credentials must be set via environment variable.')\n \n- session.interpreter = 'python{}'.format(python_version)\n- session.virtualenv_dirname = 'integration-' + python_version\n+ session.interpreter = f'python{python_version}'\n+ session.virtualenv_dirname = f'integration-{python_version}'\n \n session.install('aiohttp', 'pytest', *LOCAL_DEPS)\n session.install('.')\n@@ -46,7 +46,7 @@ def integration_tests(session, python_version):\n @nox.session\n @nox.parametrize('python_version', ['3.6'])\n def lint_setup_py(session, python_version):\n- session.interpreter = 'python{}'.format(python_version)\n+ session.interpreter = f'python{python_version}'\n session.virtualenv_dirname = 'setup'\n \n session.install('docutils', 'Pygments')\n@@ -61,7 +61,7 @@ def lint_setup_py(session, python_version):\n @nox.session\n @nox.parametrize('python_version', ['3.6'])\n def cover(session, python_version):\n- session.interpreter = 'python{}'.format(python_version)\n+ session.interpreter = f'python{python_version}'\n session.virtualenv_dirname = 'cover'\n \n session.install('codecov', 'coverage', 'pytest-cov')\ndiff --git a/taskqueue/gcloud/aio/taskqueue/taskmanager.py b/taskqueue/gcloud/aio/taskqueue/taskmanager.py\n--- a/taskqueue/gcloud/aio/taskqueue/taskmanager.py\n+++ b/taskqueue/gcloud/aio/taskqueue/taskmanager.py\n@@ -59,9 +59,9 @@ async def stop(self):\n \n @staticmethod\n def autorenew(event, headers, task, lease_seconds):\n- url = '{}/{}:renewLease'.format(API_ROOT, task['name'])\n+ url = f'{API_ROOT}/{task[\"name\"]}:renewLease'\n body = {\n- 'leaseDuration': '{}s'.format(lease_seconds),\n+ 'leaseDuration': f'{lease_seconds}s',\n 'responseView': 'FULL',\n }\n \ndiff --git a/taskqueue/gcloud/aio/taskqueue/taskqueue.py b/taskqueue/gcloud/aio/taskqueue/taskqueue.py\n--- a/taskqueue/gcloud/aio/taskqueue/taskqueue.py\n+++ b/taskqueue/gcloud/aio/taskqueue/taskqueue.py\n@@ -28,21 +28,22 @@ def __init__(self, project, service_file, taskqueue, location=LOCATION,\n self.session = session or aiohttp.ClientSession(conn_timeout=10,\n read_timeout=10)\n \n- self.api_root = '{}/projects/{}/locations/{}/queues/{}'.format(\n- API_ROOT, project, location, taskqueue)\n+ self.api_root = (f'{API_ROOT}/projects/{project}/'\n+ f'locations/{location}/queues/{taskqueue}')\n \n self.token = token or Token(project, service_file, scopes=SCOPES,\n session=self.session)\n \n async def headers(self):\n+ token = await self.token.get()\n return {\n- 'Authorization': 'Bearer {}'.format(await self.token.get()),\n+ 'Authorization': f'Bearer {token}',\n 'Content-Type': 'application/json',\n }\n \n # https://cloud.google.com/cloud-tasks/docs/reference/rest/v2beta2/projects.locations.queues.tasks/acknowledge\n async def ack(self, task, session=None):\n- url = '{}/{}:acknowledge'.format(API_ROOT, task['name'])\n+ url = f'{API_ROOT}/{task[\"name\"]}:acknowledge'\n body = {\n 'scheduleTime': task['scheduleTime'],\n }\n@@ -55,7 +56,7 @@ async def ack(self, task, session=None):\n \n # https://cloud.google.com/cloud-tasks/docs/reference/rest/v2beta2/projects.locations.queues.tasks/cancelLease\n async def cancel(self, task, session=None):\n- url = '{}/{}:cancelLease'.format(API_ROOT, task['name'])\n+ url = f'{API_ROOT}/{task[\"name\"]}:cancelLease'\n body = {\n 'scheduleTime': task['scheduleTime'],\n 'responseView': 'BASIC',\n@@ -69,7 +70,7 @@ async def cancel(self, task, session=None):\n \n # https://cloud.google.com/cloud-tasks/docs/reference/rest/v2beta2/projects.locations.queues.tasks/delete\n async def delete(self, tname, session=None):\n- url = '{}/{}'.format(API_ROOT, tname)\n+ url = f'{API_ROOT}/{tname}'\n \n s = session or self.session\n resp = await retry(s.delete(url, headers=await self.headers()))\n@@ -84,7 +85,7 @@ async def drain(self):\n \n # https://cloud.google.com/cloud-tasks/docs/reference/rest/v2beta2/projects.locations.queues.tasks/get\n async def get(self, tname, full=False, session=None):\n- url = '{}/{}'.format(API_ROOT, tname)\n+ url = f'{API_ROOT}/{tname}'\n params = {\n 'responseView': 'FULL' if full else 'BASIC',\n }\n@@ -97,7 +98,7 @@ async def get(self, tname, full=False, session=None):\n \n # https://cloud.google.com/cloud-tasks/docs/reference/rest/v2beta2/projects.locations.queues.tasks/create\n async def insert(self, payload, tag=None, session=None):\n- url = '{}/tasks'.format(self.api_root)\n+ url = f'{self.api_root}/tasks'\n body = {\n 'task': {\n 'pullMessage': {\n@@ -117,10 +118,10 @@ async def insert(self, payload, tag=None, session=None):\n # https://cloud.google.com/cloud-tasks/docs/reference/rest/v2beta2/projects.locations.queues.tasks/lease\n async def lease(self, num_tasks=1, lease_seconds=60, task_filter=None,\n session=None):\n- url = '{}/tasks:lease'.format(self.api_root)\n+ url = f'{self.api_root}/tasks:lease'\n body = {\n 'maxTasks': min(num_tasks, 1000),\n- 'leaseDuration': '{}s'.format(lease_seconds),\n+ 'leaseDuration': f'{lease_seconds}s',\n 'responseView': 'FULL',\n }\n if task_filter:\n@@ -135,7 +136,7 @@ async def lease(self, num_tasks=1, lease_seconds=60, task_filter=None,\n # https://cloud.google.com/cloud-tasks/docs/reference/rest/v2beta2/projects.locations.queues.tasks/list\n async def list(self, full=False, page_size=1000, page_token='',\n session=None):\n- url = '{}/tasks'.format(self.api_root)\n+ url = f'{self.api_root}/tasks'\n params = {\n 'responseView': 'FULL' if full else 'BASIC',\n 'pageSize': page_size,\n@@ -150,10 +151,10 @@ async def list(self, full=False, page_size=1000, page_token='',\n \n # https://cloud.google.com/cloud-tasks/docs/reference/rest/v2beta2/projects.locations.queues.tasks/renewLease\n async def renew(self, task, lease_seconds=60, session=None):\n- url = '{}/{}:renewLease'.format(API_ROOT, task['name'])\n+ url = f'{API_ROOT}/{task[\"name\"]}:renewLease'\n body = {\n 'scheduleTime': task['scheduleTime'],\n- 'leaseDuration': '{}s'.format(lease_seconds),\n+ 'leaseDuration': f'{lease_seconds}s',\n 'responseView': 'FULL',\n }\n \ndiff --git a/taskqueue/nox.py b/taskqueue/nox.py\n--- a/taskqueue/nox.py\n+++ b/taskqueue/nox.py\n@@ -10,8 +10,8 @@\n @nox.session\n @nox.parametrize('python_version', ['3.6'])\n def unit_tests(session, python_version):\n- session.interpreter = 'python{}'.format(python_version)\n- session.virtualenv_dirname = 'unit-' + python_version\n+ session.interpreter = f'python{python_version}'\n+ session.virtualenv_dirname = f'unit-{python_version}'\n \n session.install('pytest', 'pytest-cov', *LOCAL_DEPS)\n session.install('-e', '.')\n@@ -34,8 +34,8 @@ def integration_tests(session, python_version):\n if not os.environ.get('GOOGLE_APPLICATION_CREDENTIALS', ''):\n session.skip('Credentials must be set via environment variable.')\n \n- session.interpreter = 'python{}'.format(python_version)\n- session.virtualenv_dirname = 'integration-' + python_version\n+ session.interpreter = f'python{python_version}'\n+ session.virtualenv_dirname = f'integration-{python_version}'\n \n session.install('pytest', 'pytest-mock', *LOCAL_DEPS)\n session.install('.')\n@@ -46,7 +46,7 @@ def integration_tests(session, python_version):\n @nox.session\n @nox.parametrize('python_version', ['3.6'])\n def lint_setup_py(session, python_version):\n- session.interpreter = 'python{}'.format(python_version)\n+ session.interpreter = f'python{python_version}'\n session.virtualenv_dirname = 'setup'\n \n session.install('docutils', 'Pygments')\n@@ -61,7 +61,7 @@ def lint_setup_py(session, python_version):\n @nox.session\n @nox.parametrize('python_version', ['3.6'])\n def cover(session, python_version):\n- session.interpreter = 'python{}'.format(python_version)\n+ session.interpreter = f'python{python_version}'\n session.virtualenv_dirname = 'cover'\n \n session.install('codecov', 'coverage', 'pytest-cov')\n", "test_patch": "diff --git a/core/tests/__init__.py b/core/tests/__init__.py\ndeleted file mode 100644\ndiff --git a/core/tests/unit/__init__.py b/core/tests/unit/__init__.py\ndeleted file mode 100644\ndiff --git a/core/tests/unit/aio_test.py b/core/tests/unit/aio_test.py\ndeleted file mode 100644\n--- a/core/tests/unit/aio_test.py\n+++ /dev/null\n@@ -1,5 +0,0 @@\n-import gcloud.aio.core.aio as aio # pylint: disable=unused-import\n-\n-\n-def test_importable():\n- assert True\ndiff --git a/core/tests/unit/astate_test.py b/core/tests/unit/astate_test.py\ndeleted file mode 100644\n--- a/core/tests/unit/astate_test.py\n+++ /dev/null\n@@ -1,5 +0,0 @@\n-import gcloud.aio.core.astate as astate # pylint: disable=unused-import\n-\n-\n-def test_importable():\n- assert True\ndiff --git a/core/tests/unit/http_test.py b/core/tests/unit/http_test.py\ndeleted file mode 100644\n--- a/core/tests/unit/http_test.py\n+++ /dev/null\n@@ -1,5 +0,0 @@\n-import gcloud.aio.core.http as http # pylint: disable=unused-import\n-\n-\n-def test_importable():\n- assert True\ndiff --git a/datastore/tests/integration/smoke_test.py b/datastore/tests/integration/smoke_test.py\n--- a/datastore/tests/integration/smoke_test.py\n+++ b/datastore/tests/integration/smoke_test.py\n@@ -24,7 +24,7 @@ def test_item_lifecycle():\n creds = os.environ['GOOGLE_APPLICATION_CREDENTIALS']\n \n kind_name = 'gcloud-aio-test'\n- object_name = 'test_record_{}'.format(uuid.uuid4())\n+ object_name = f'test_record_{uuid.uuid4()}'\n \n loop = asyncio.get_event_loop()\n loop.run_until_complete(\ndiff --git a/storage/tests/integration/smoke_test.py b/storage/tests/integration/smoke_test.py\n--- a/storage/tests/integration/smoke_test.py\n+++ b/storage/tests/integration/smoke_test.py\n@@ -21,7 +21,7 @@ def test_object_is_downloaded():\n call_id = '07fbe0cc-7f87-1235-06b0-0cc47a392728'\n side = 'callee'\n link = 0\n- object_name = '{}/{}/{}/rtp.pcap.wav.ctm'.format(call_id, side, link)\n+ object_name = f'{call_id}/{side}/{link}/rtp.pcap.wav.ctm'\n \n loop = asyncio.get_event_loop()\n loop.run_until_complete(\ndiff --git a/storage/tests/unit/blob_test.py b/storage/tests/unit/blob_test.py\nnew file mode 100644\n--- /dev/null\n+++ b/storage/tests/unit/blob_test.py\n@@ -0,0 +1,5 @@\n+import gcloud.aio.storage.blob as blob # pylint: disable=unused-import\n+\n+\n+def test_importable():\n+ assert True\ndiff --git a/storage/tests/unit/bucket_test.py b/storage/tests/unit/bucket_test.py\nnew file mode 100644\n--- /dev/null\n+++ b/storage/tests/unit/bucket_test.py\n@@ -0,0 +1,5 @@\n+import gcloud.aio.storage.bucket as bucket # pylint: disable=unused-import\n+\n+\n+def test_importable():\n+ assert True\ndiff --git a/storage/tests/unit/storage_test.py b/storage/tests/unit/storage_test.py\n--- a/storage/tests/unit/storage_test.py\n+++ b/storage/tests/unit/storage_test.py\n@@ -1,4 +1,4 @@\n-import gcloud.aio.storage as storage # pylint: disable=unused-import\n+import gcloud.aio.storage.storage as storage # pylint: disable=unused-import\n \n \n def test_importable():\ndiff --git a/storage/tests/unit/utils_test.py b/storage/tests/unit/utils_test.py\nnew file mode 100644\n--- /dev/null\n+++ b/storage/tests/unit/utils_test.py\n@@ -0,0 +1,5 @@\n+import gcloud.aio.storage.utils as utils # pylint: disable=unused-import\n+\n+\n+def test_importable():\n+ assert True\ndiff --git a/taskqueue/tests/integration/taskqueue_test.py b/taskqueue/tests/integration/taskqueue_test.py\n--- a/taskqueue/tests/integration/taskqueue_test.py\n+++ b/taskqueue/tests/integration/taskqueue_test.py\n@@ -26,8 +26,7 @@ async def do_task_lifecycle(project, creds, task_queue):\n assert inserted\n \n # GET\n- got = await tq.get(inserted['name'], full=True)\n- assert got == inserted\n+ assert inserted == await tq.get(inserted['name'], full=True)\n \n # LIST\n listed = await tq.list(full=True)\n@@ -36,14 +35,12 @@ async def do_task_lifecycle(project, creds, task_queue):\n \n # LEASE\n leased = await tq.lease(num_tasks=1, lease_seconds=10,\n- task_filter='tag={}'.format(encode(tag)))\n+ task_filter=f'tag={encode(tag)}')\n assert leased.get('tasks') and len(leased['tasks']) == 1\n \n leased_message = leased['tasks'][0]['pullMessage']\n- leased_payload = json.loads(decode(leased_message['payload']))\n- leased_tag = decode(leased_message['tag'])\n- assert leased_payload == payload\n- assert leased_tag == tag\n+ assert payload == json.loads(decode(leased_message['payload']))\n+ assert tag == decode(leased_message['tag'])\n \n # RENEW\n renewed = await tq.renew(leased['tasks'][0], lease_seconds=10)\n@@ -57,8 +54,7 @@ async def do_task_lifecycle(project, creds, task_queue):\n # cancel?\n \n # DELETE\n- result = await tq.delete(renewed['name'])\n- assert not result\n+ assert not await tq.delete(renewed['name'])\n \n \n def test_task_lifecycle():\n", "problem_statement": "", "hints_text": "", "created_at": "2018-05-01T17:56:36Z"}