File size: 58,446 Bytes
5980447
1
2
{"repo": "talkiq/gcloud-aio", "pull_number": 24, "instance_id": "talkiq__gcloud-aio-24", "issue_numbers": "", "base_commit": "02d7ce2526855c16a36d7cced824a9c7b7d9bd6b", "patch": "diff --git a/auth/gcloud/aio/auth/auth.py b/auth/gcloud/aio/auth/auth.py\n--- a/auth/gcloud/aio/auth/auth.py\n+++ b/auth/gcloud/aio/auth/auth.py\n@@ -1,15 +1,16 @@\n \"\"\"\n Google Cloud auth via service account file\n \"\"\"\n+import asyncio\n import datetime\n import json\n import time\n import typing\n+from urllib.parse import quote_plus\n+from urllib.parse import urlencode\n \n import aiohttp\n import jwt\n-from gcloud.aio.core.aio import auto\n-from gcloud.aio.core.http import post\n \n \n ScopeList = typing.List[str]\n@@ -20,29 +21,24 @@\n            'project_id.'\n \n \n-async def acquire_token(session: aiohttp.ClientSession,\n-                        service_data: dict,\n+async def acquire_token(session: aiohttp.ClientSession, service_data: dict,\n                         scopes: ScopeList = None):\n-\n     url, assertion = generate_assertion(service_data, scopes)\n \n-    payload = {\n-        'grant_type': JWT_GRANT_TYPE,\n-        'assertion': assertion\n+    headers = {\n+        'Content-Type': 'application/x-www-form-urlencoded',\n     }\n+    payload = urlencode({\n+        'assertion': assertion,\n+        'grant_type': JWT_GRANT_TYPE,\n+    }, quote_via=quote_plus)\n \n-    _status, content = await post(\n-        url,\n-        payload,\n-        headers={'content-type': 'application/x-www-form-urlencoded'},\n-        timeout=60,\n-        urlencoded=True,\n-        json_response=True,\n-        session=session\n-    )\n+    response = await session.post(url, data=payload, headers=headers,\n+                                  params=None, timeout=60)\n+    content = await response.json()\n \n     if 'error' in content:\n-        raise Exception('{}'.format(content))\n+        raise Exception(f'got error acquiring token: {content}')\n \n     return {\n         'access_token': str(content['access_token']),\n@@ -100,7 +96,7 @@ def __init__(self, project: str, service_file: str,\n \n         self.scopes = scopes or []\n \n-        self.session = session\n+        self.session = session or aiohttp.ClientSession()\n         self.access_token = None\n         self.access_token_duration = None\n         self.access_token_acquired_at = None\n@@ -121,7 +117,7 @@ async def ensure_token(self):\n \n         elif not self.access_token:\n \n-            self.acquiring = self.acquire_access_token()\n+            self.acquiring = asyncio.ensure_future(self.acquire_access_token())\n \n             await self.acquiring\n \n@@ -132,11 +128,11 @@ async def ensure_token(self):\n \n             if delta > self.access_token_duration / 2:\n \n-                self.acquiring = self.acquire_access_token()\n+                self.acquiring = asyncio.ensure_future(\n+                    self.acquire_access_token())\n \n                 await self.acquiring\n \n-    @auto\n     async def acquire_access_token(self):\n \n         data = await acquire_token(\ndiff --git a/auth/nox.py b/auth/nox.py\n--- a/auth/nox.py\n+++ b/auth/nox.py\n@@ -4,16 +4,13 @@\n import nox\n \n \n-LOCAL_DEPS = ('../core/',)\n-\n-\n @nox.session\n @nox.parametrize('python_version', ['3.6'])\n def unit_tests(session, python_version):\n-    session.interpreter = 'python{}'.format(python_version)\n-    session.virtualenv_dirname = 'unit-' + python_version\n+    session.interpreter = f'python{python_version}'\n+    session.virtualenv_dirname = f'unit-{python_version}'\n \n-    session.install('pytest', 'pytest-cov', *LOCAL_DEPS)\n+    session.install('pytest', 'pytest-cov')\n     session.install('-e', '.')\n \n     session.run(\n@@ -23,7 +20,7 @@ def unit_tests(session, python_version):\n         '--cov=tests.unit',\n         '--cov-append',\n         '--cov-report=',\n-        '--cov-fail-under=39',\n+        '--cov-fail-under=38',\n         os.path.join('tests', 'unit'),\n         *session.posargs)\n \n@@ -34,10 +31,10 @@ def integration_tests(session, python_version):\n     if not os.environ.get('GOOGLE_APPLICATION_CREDENTIALS', ''):\n         session.skip('Credentials must be set via environment variable.')\n \n-    session.interpreter = 'python{}'.format(python_version)\n-    session.virtualenv_dirname = 'integration-' + python_version\n+    session.interpreter = f'python{python_version}'\n+    session.virtualenv_dirname = f'integration-{python_version}'\n \n-    session.install('aiohttp', 'pytest', *LOCAL_DEPS)\n+    session.install('aiohttp', 'pytest')\n     session.install('.')\n \n     session.run('py.test', '--quiet', 'tests/integration')\n@@ -46,7 +43,7 @@ def integration_tests(session, python_version):\n @nox.session\n @nox.parametrize('python_version', ['3.6'])\n def lint_setup_py(session, python_version):\n-    session.interpreter = 'python{}'.format(python_version)\n+    session.interpreter = f'python{python_version}'\n     session.virtualenv_dirname = 'setup'\n \n     session.install('docutils', 'Pygments')\n@@ -61,7 +58,7 @@ def lint_setup_py(session, python_version):\n @nox.session\n @nox.parametrize('python_version', ['3.6'])\n def cover(session, python_version):\n-    session.interpreter = 'python{}'.format(python_version)\n+    session.interpreter = f'python{python_version}'\n     session.virtualenv_dirname = 'cover'\n \n     session.install('codecov', 'coverage', 'pytest-cov')\ndiff --git a/bigquery/gcloud/aio/bigquery/bigquery.py b/bigquery/gcloud/aio/bigquery/bigquery.py\n--- a/bigquery/gcloud/aio/bigquery/bigquery.py\n+++ b/bigquery/gcloud/aio/bigquery/bigquery.py\n@@ -2,9 +2,12 @@\n import logging\n import uuid\n \n-import ujson\n+import aiohttp\n from gcloud.aio.auth import Token\n-from gcloud.aio.core.http import post\n+try:\n+    import ujson as json\n+except ModuleNotFoundError:\n+    import json\n \n \n API_ROOT = 'https://www.googleapis.com/bigquery/v2'\n@@ -63,51 +66,43 @@ async def headers(self):\n         token = await self.token.get()\n \n         return {\n-            'Authorization': 'Bearer {}'.format(token)\n+            'Authorization': f'Bearer {token}',\n         }\n \n     async def insert(self, rows, skip_invalid=False, ignore_unknown=True,\n                      session=None):\n-\n         session = session or self.session\n \n-        body = make_insert_body(\n-            rows,\n-            skip_invalid=skip_invalid,\n-            ignore_unknown=ignore_unknown\n-        )\n-\n-        headers = await self.headers()\n-\n-        url = '{}/{}'.format(\n-            API_ROOT,\n-            INSERT_TEMPLATE.format(\n-                proj=self.project,\n-                dataset=self.dataset_name,\n-                table=self.table_name\n-            )\n-        )\n-\n+        insert_url = INSERT_TEMPLATE.format(proj=self.project,\n+                                            dataset=self.dataset_name,\n+                                            table=self.table_name)\n+        url = f'{API_ROOT}/{insert_url}'\n         log.info('Inserting %d rows to %s', len(rows), url)\n \n-        status, content = await post(\n-            url,\n-            payload=body,\n-            headers=headers\n-        )\n+        body = make_insert_body(rows, skip_invalid=skip_invalid,\n+                                ignore_unknown=ignore_unknown)\n+        payload = json.dumps(body).encode('utf-8')\n+\n+        headers = await self.headers()\n+        headers.update({\n+            'Content-Length': str(len(payload)),\n+            'Content-Type': 'application/json'\n+        })\n \n-        success = 299 >= status >= 200 and 'insertErrors' not in content\n+        async with aiohttp.ClientSession() as s:\n+            response = await s.post(url, data=payload, headers=headers,\n+                                    params=None, timeout=60)\n+            content = await response.json()\n \n-        if success:\n-            return success\n+        if 299 >= response.status >= 200 and 'insertErrors' not in content:\n+            return True\n \n-        log.debug('response code: %d', status)\n+        log.debug('response code: %d', response.status)\n         log.debug('url: %s', url)\n-        log.debug('body:\\n%s\\n', body)\n+        log.debug('body:\\n%s\\n', payload)\n \n-        raise Exception('Could not insert: {}'.format(ujson.dumps(\n-            content, sort_keys=True\n-        )))\n+        content_blob = json.dumps(content, sort_keys=True)\n+        raise Exception(f'could not insert: {content_blob}')\n \n \n async def stream_insert(table, rows):\ndiff --git a/bigquery/nox.py b/bigquery/nox.py\n--- a/bigquery/nox.py\n+++ b/bigquery/nox.py\n@@ -4,14 +4,14 @@\n import nox\n \n \n-LOCAL_DEPS = ('../core/', '../auth/')\n+LOCAL_DEPS = ('../auth/', )\n \n \n @nox.session\n @nox.parametrize('python_version', ['3.6'])\n def unit_tests(session, python_version):\n-    session.interpreter = 'python{}'.format(python_version)\n-    session.virtualenv_dirname = 'unit-' + python_version\n+    session.interpreter = f'python{python_version}'\n+    session.virtualenv_dirname = f'unit-{python_version}'\n \n     session.install('pytest', 'pytest-cov', *LOCAL_DEPS)\n     session.install('-e', '.')\n@@ -23,7 +23,7 @@ def unit_tests(session, python_version):\n         '--cov=tests.unit',\n         '--cov-append',\n         '--cov-report=',\n-        '--cov-fail-under=47',\n+        '--cov-fail-under=46',\n         os.path.join('tests', 'unit'),\n         *session.posargs)\n \n@@ -34,8 +34,8 @@ def integration_tests(session, python_version):\n     if not os.environ.get('GOOGLE_APPLICATION_CREDENTIALS', ''):\n         session.skip('Credentials must be set via environment variable.')\n \n-    session.interpreter = 'python{}'.format(python_version)\n-    session.virtualenv_dirname = 'integration-' + python_version\n+    session.interpreter = f'python{python_version}'\n+    session.virtualenv_dirname = f'integration-{python_version}'\n \n     session.install('aiohttp', 'pytest', *LOCAL_DEPS)\n     session.install('.')\n@@ -46,7 +46,7 @@ def integration_tests(session, python_version):\n @nox.session\n @nox.parametrize('python_version', ['3.6'])\n def lint_setup_py(session, python_version):\n-    session.interpreter = 'python{}'.format(python_version)\n+    session.interpreter = f'python{python_version}'\n     session.virtualenv_dirname = 'setup'\n \n     session.install('docutils', 'Pygments')\n@@ -61,11 +61,11 @@ def lint_setup_py(session, python_version):\n @nox.session\n @nox.parametrize('python_version', ['3.6'])\n def cover(session, python_version):\n-    session.interpreter = 'python{}'.format(python_version)\n+    session.interpreter = f'python{python_version}'\n     session.virtualenv_dirname = 'cover'\n \n     session.install('codecov', 'coverage', 'pytest-cov')\n \n-    session.run('coverage', 'report', '--show-missing', '--fail-under=47')\n+    session.run('coverage', 'report', '--show-missing', '--fail-under=46')\n     session.run('codecov')\n     session.run('coverage', 'erase')\ndiff --git a/core/gcloud/__init__.py b/core/gcloud/__init__.py\ndeleted file mode 100644\n--- a/core/gcloud/__init__.py\n+++ /dev/null\n@@ -1,6 +0,0 @@\n-try:\n-    import pkg_resources\n-    pkg_resources.declare_namespace(__name__)\n-except ImportError:\n-    import pkgutil\n-    __path__ = pkgutil.extend_path(__path__, __name__)\ndiff --git a/core/gcloud/aio/__init__.py b/core/gcloud/aio/__init__.py\ndeleted file mode 100644\n--- a/core/gcloud/aio/__init__.py\n+++ /dev/null\n@@ -1,6 +0,0 @@\n-try:\n-    import pkg_resources\n-    pkg_resources.declare_namespace(__name__)\n-except ImportError:\n-    import pkgutil\n-    __path__ = pkgutil.extend_path(__path__, __name__)\ndiff --git a/core/gcloud/aio/core/__init__.py b/core/gcloud/aio/core/__init__.py\ndeleted file mode 100644\n--- a/core/gcloud/aio/core/__init__.py\n+++ /dev/null\n@@ -1,5 +0,0 @@\n-from pkg_resources import get_distribution\n-__version__ = get_distribution('gcloud-aio-core').version\n-\n-\n-__all__ = ['__version__']\ndiff --git a/core/gcloud/aio/core/aio.py b/core/gcloud/aio/core/aio.py\ndeleted file mode 100644\n--- a/core/gcloud/aio/core/aio.py\n+++ /dev/null\n@@ -1,66 +0,0 @@\n-import asyncio\n-import functools\n-\n-\n-def maybe_async(callable_, *args, **kwargs):\n-\n-    \"\"\"\n-    Turn a callable into a coroutine if it isn't\n-    \"\"\"\n-\n-    if asyncio.iscoroutine(callable_):\n-        return callable_\n-\n-    return asyncio.coroutine(callable_)(*args, **kwargs)\n-\n-\n-def fire(callable_, *args, **kwargs):\n-\n-    \"\"\"\n-    Start a callable as a coroutine, and return it's future. The cool thing\n-    about this function is that (via maybe_async) it lets you treat synchronous\n-    and asynchronous callables the same (both as async), which simplifies code.\n-    \"\"\"\n-\n-    return asyncio.ensure_future(maybe_async(callable_, *args, **kwargs))\n-\n-\n-def auto(fn):\n-\n-    \"\"\"\n-    Decorate a function or method with this, and it will become a callable\n-    that can be scheduled in the event loop just by calling it. Normally you'd\n-    have to do an `asyncio.ensure_future(my_callable())`. Not you can just do\n-    `my_callable()`. Twisted has always let you do this, and now you can let\n-    asyncio do it as well (with a decorator, albeit...)\n-    \"\"\"\n-\n-    @functools.wraps(fn)\n-    def wrapper(*args, **kwargs):\n-\n-        return fire(fn, *args, **kwargs)\n-\n-    return wrapper\n-\n-\n-async def _call_later(delay, callable_, *args, **kwargs):\n-\n-    \"\"\"\n-    The bus stop, where we wait.\n-    \"\"\"\n-\n-    await asyncio.sleep(delay)\n-\n-    fire(callable_, *args, **kwargs)\n-\n-\n-def call_later(delay, callable_, *args, **kwargs):\n-\n-    \"\"\"\n-    After :delay seconds, call :callable with :args and :kwargs; :callable can\n-    be a synchronous or asynchronous callable (a coroutine). Note that _this_\n-    function is synchronous - mission accomplished - it can be used from within\n-    any synchronous or asynchronous callable.\n-    \"\"\"\n-\n-    return fire(_call_later, delay, callable_, *args, **kwargs)\ndiff --git a/core/gcloud/aio/core/astate.py b/core/gcloud/aio/core/astate.py\ndeleted file mode 100644\n--- a/core/gcloud/aio/core/astate.py\n+++ /dev/null\n@@ -1,83 +0,0 @@\n-import asyncio\n-import logging\n-\n-from gcloud.aio.core.aio import fire\n-\n-\n-log = logging.getLogger(__name__)\n-\n-\n-class AwaitableState:\n-    # pylint: disable=too-few-public-methods\n-\n-    \"\"\"\n-    Wrap a :future with a name and data. If :future is a coroutine, turn it\n-    into a future by firing it.\n-\n-    Use instances of AwaitableState as named states in state machines. Use\n-    :data for arbitrary context beyond :name.\n-    \"\"\"\n-\n-    def __init__(self, name, future, data=None):\n-\n-        self.name = name\n-        self.future = future\n-        self.data = data\n-\n-        if asyncio.iscoroutine(self.future):\n-            self.future = fire(self.future)\n-\n-    def __await__(self):\n-\n-        return self.future.__await__()\n-\n-    def __str__(self):\n-\n-        return self.__repr__()\n-\n-    def __repr__(self):\n-\n-        return '<awaitable state: {} at 0x{}>'.format(\n-            self.name,\n-            id(self)\n-        )\n-\n-    def __getattr__(self, attr):\n-\n-        return getattr(self.future, attr)\n-\n-    def __hash__(self):\n-\n-        return hash(self.name)\n-\n-    def __eq__(self, other):\n-\n-        return hash(self) == hash(other)\n-\n-\n-def make_stepper(default_step, state_step, name='sm'):\n-\n-    \"\"\"\n-    `default_step`: a callable that takes no args\n-    `state_step`: a mapping between AwaitableState.name -> callable\n-    \"\"\"\n-\n-    async def step(state, args):\n-\n-        state_name = getattr(state, 'name', None)\n-        step = state_step.get(state_name, default_step)\n-        next_state = step(args) if args is not None else step()\n-\n-        if next_state:\n-            args = await next_state\n-        else:\n-            args = tuple()\n-\n-        if next_state != state:\n-            log.debug('%s state change: %s -> %s', name,\n-                      getattr(state, 'name', None),\n-                      getattr(next_state, 'name', None))\n-\n-        return next_state, args\n-\n-    return step\ndiff --git a/core/gcloud/aio/core/http.py b/core/gcloud/aio/core/http.py\ndeleted file mode 100644\n--- a/core/gcloud/aio/core/http.py\n+++ /dev/null\n@@ -1,140 +0,0 @@\n-from urllib.parse import quote_plus\n-from urllib.parse import urlencode\n-\n-import aiohttp\n-import ujson\n-from asyncio_extras.contextmanager import async_contextmanager\n-\n-\n-class HttpError(Exception):\n-    pass\n-\n-\n-@async_contextmanager\n-async def ensure_session(session):\n-\n-    if session:\n-        yield session\n-    else:\n-        async with aiohttp.ClientSession() as session:\n-            yield session\n-\n-\n-async def delete(url, headers=None, params=None, timeout=60, session=None):\n-\n-    async with ensure_session(session) as s:  # pylint: disable=not-async-context-manager\n-\n-        response = await s.delete(\n-            url,\n-            headers=headers,\n-            params=params,\n-            timeout=timeout\n-        )\n-\n-        phrase = await response.text()\n-\n-    return response.status, phrase\n-\n-\n-async def post(url, payload=None, timeout=60, urlencoded=False,\n-               json_response=True, session=None, headers=None, params=None):\n-    # pylint: disable=too-many-arguments\n-\n-    headers = headers or {}\n-\n-    if urlencoded:\n-\n-        if payload:\n-            payload = urlencode(payload, quote_via=quote_plus)\n-\n-        headers['content-type'] = 'application/x-www-form-urlencoded'\n-\n-    else:\n-\n-        if payload:\n-            payload = ujson.dumps(payload)\n-            payload = payload.encode('utf-8')\n-            content_length = str(len(payload))\n-        else:\n-            content_length = '0'\n-\n-        headers.update({\n-            'content-length': content_length,\n-            'content-type': 'application/json'\n-        })\n-\n-    async with ensure_session(session) as s:  # pylint: disable=not-async-context-manager\n-\n-        response = await s.post(\n-            url,\n-            data=payload,\n-            headers=headers,\n-            params=params,\n-            timeout=timeout\n-        )\n-\n-        if json_response:\n-            content = await response.json()\n-        else:\n-            content = await response.text()\n-\n-    return response.status, content\n-\n-\n-async def get(url, timeout=60, json_response=True, session=None, headers=None,\n-              params=None):\n-    # pylint: disable=too-many-arguments\n-\n-    async with ensure_session(session) as s:  # pylint: disable=not-async-context-manager\n-\n-        response = await s.get(\n-            url,\n-            headers=headers,\n-            params=params,\n-            timeout=timeout\n-        )\n-\n-        if json_response:\n-            content = await response.json()\n-        else:\n-            content = await response.text()\n-\n-    return response.status, content\n-\n-\n-async def put(*args, **kwargs):  # pylint: disable=unused-argument\n-\n-    raise Exception('Not implemented.')\n-\n-\n-async def patch(url, payload=None, timeout=60, session=None, headers=None,\n-                params=None):\n-    # pylint: disable=too-many-arguments\n-\n-    headers = headers or {}\n-\n-    if payload:\n-        payload = ujson.dumps(payload)\n-        payload = payload.encode('utf-8')\n-        content_length = str(len(payload))\n-    else:\n-        content_length = '0'\n-\n-    headers.update({\n-        'content-length': content_length,\n-        'content-type': 'application/json'\n-    })\n-\n-    async with ensure_session(session) as s:  # pylint: disable=not-async-context-manager\n-\n-        response = await s.patch(\n-            url,\n-            data=payload,\n-            headers=headers,\n-            params=params,\n-            timeout=timeout\n-        )\n-\n-        phrase = await response.text()\n-\n-    return response.status, phrase\ndiff --git a/core/nox.py b/core/nox.py\ndeleted file mode 100644\n--- a/core/nox.py\n+++ /dev/null\n@@ -1,53 +0,0 @@\n-# pylint: disable=import-self,no-member\n-import os\n-\n-import nox\n-\n-\n-@nox.session\n-@nox.parametrize('python_version', ['3.6'])\n-def unit_tests(session, python_version):\n-    session.interpreter = 'python{}'.format(python_version)\n-    session.virtualenv_dirname = 'unit-' + python_version\n-\n-    session.install('pytest', 'pytest-cov')\n-    session.install('-e', '.')\n-\n-    session.run(\n-        'py.test',\n-        '--quiet',\n-        '--cov=gcloud.aio.core',\n-        '--cov=tests.unit',\n-        '--cov-append',\n-        '--cov-report=',\n-        '--cov-fail-under=37',\n-        os.path.join('tests', 'unit'),\n-        *session.posargs)\n-\n-\n-@nox.session\n-@nox.parametrize('python_version', ['3.6'])\n-def lint_setup_py(session, python_version):\n-    session.interpreter = 'python{}'.format(python_version)\n-    session.virtualenv_dirname = 'setup'\n-\n-    session.install('docutils', 'Pygments')\n-    session.run(\n-        'python',\n-        'setup.py',\n-        'check',\n-        '--restructuredtext',\n-        '--strict')\n-\n-\n-@nox.session\n-@nox.parametrize('python_version', ['3.6'])\n-def cover(session, python_version):\n-    session.interpreter = 'python{}'.format(python_version)\n-    session.virtualenv_dirname = 'cover'\n-\n-    session.install('codecov', 'coverage', 'pytest-cov')\n-\n-    session.run('coverage', 'report', '--show-missing', '--fail-under=37')\n-    session.run('codecov')\n-    session.run('coverage', 'erase')\ndiff --git a/core/setup.py b/core/setup.py\ndeleted file mode 100644\n--- a/core/setup.py\n+++ /dev/null\n@@ -1,41 +0,0 @@\n-import os\n-\n-import setuptools\n-\n-\n-PACKAGE_ROOT = os.path.abspath(os.path.dirname(__file__))\n-with open(os.path.join(PACKAGE_ROOT, 'README.rst')) as f:\n-    README = f.read()\n-\n-with open(os.path.join(PACKAGE_ROOT, 'requirements.txt')) as f:\n-    REQUIREMENTS = [r.strip() for r in f.readlines()]\n-\n-\n-setuptools.setup(\n-    name='gcloud-aio-core',\n-    version='0.7.2',\n-    description='Core Helpers for Asyncio Google Cloud Library',\n-    long_description=README,\n-    namespace_packages=[\n-        'gcloud',\n-        'gcloud.aio',\n-    ],\n-    packages=setuptools.find_packages(exclude=('tests',)),\n-    install_requires=REQUIREMENTS,\n-    author='TalkIQ',\n-    author_email='engineering@talkiq.com',\n-    url='https://github.com/talkiq/gcloud-aio',\n-    platforms='Posix; MacOS X; Windows',\n-    include_package_data=True,\n-    zip_safe=False,\n-    license='MIT License',\n-    classifiers=[\n-        'Development Status :: 4 - Beta',\n-        'Intended Audience :: Developers',\n-        'License :: OSI Approved :: MIT License',\n-        'Operating System :: OS Independent',\n-        'Programming Language :: Python :: 3',\n-        'Programming Language :: Python :: 3.6',\n-        'Topic :: Internet',\n-    ],\n-)\ndiff --git a/datastore/gcloud/aio/datastore/datastore.py b/datastore/gcloud/aio/datastore/datastore.py\n--- a/datastore/gcloud/aio/datastore/datastore.py\n+++ b/datastore/gcloud/aio/datastore/datastore.py\n@@ -1,11 +1,15 @@\n import datetime\n import logging\n \n+import aiohttp\n from gcloud.aio.auth import Token\n-from gcloud.aio.core.http import post\n from gcloud.aio.datastore.constants import Mode\n from gcloud.aio.datastore.constants import Operation\n from gcloud.aio.datastore.constants import TypeName\n+try:\n+    import ujson as json\n+except ModuleNotFoundError:\n+    import json\n \n \n API_ROOT = 'https://datastore.googleapis.com/v1/projects'\n@@ -30,9 +34,7 @@ def infer_type(value):\n     }.get(type(value))\n \n     if not type_name:\n-        raise Exception('Type {} not supported for DS insert. :('.format(\n-            type(value)\n-        ))\n+        raise Exception(f'type {type(value)} not supported for DS insert')\n \n     return type_name\n \n@@ -109,45 +111,54 @@ async def headers(self):\n         token = await self.token.get()\n \n         return {\n-            'Authorization': 'Bearer {}'.format(token),\n+            'Authorization': f'Bearer {token}',\n         }\n \n     async def transact(self):\n-        url = '{}/{}:beginTransaction'.format(API_ROOT, self.project)\n+        url = f'{API_ROOT}/{self.project}:beginTransaction'\n         headers = await self.headers()\n-        body = {}\n+        headers.update({\n+            'Content-Length': '0',\n+            'Content-Type': 'application/json'\n+        })\n \n-        status, content = await post(url, payload={}, headers=headers)\n+        async with aiohttp.ClientSession() as s:\n+            response = await s.post(url, data={}, headers=headers, params=None,\n+                                    timeout=60)\n+            content = await response.json()\n \n         # TODO: make this raise_for_status-able.\n-        success = 299 >= status >= 200\n-\n-        if success:\n+        if 299 >= response.status >= 200:\n             transaction = content['transaction']\n             return transaction\n \n-        log.debug('response code: %d', status)\n+        log.debug('response code: %d', response.status)\n         log.debug('url: %s', url)\n-        log.debug('body:\\n%s\\n', body)\n \n-        raise Exception('Could not transact: {}'.format(content))\n+        raise Exception(f'could not transact: {content}')\n \n     async def commit(self, transaction, mutations, mode=Mode.TRANSACTIONAL):\n-        url = '{}/{}:commit'.format(API_ROOT, self.project)\n+        url = f'{API_ROOT}/{self.project}:commit'\n \n         body = make_commit_body(transaction, mode, mutations)\n+        payload = json.dumps(body).encode('utf-8')\n \n         headers = await self.headers()\n+        headers.update({\n+            'Content-Length': str(len(payload)),\n+            'Content-Type': 'application/json'\n+        })\n \n-        status, content = await post(url, payload=body, headers=headers)\n+        async with aiohttp.ClientSession() as s:\n+            response = await s.post(url, data=payload, headers=headers,\n+                                    params=None, timeout=60)\n+            content = await response.json()\n \n         # TODO: make this raise_for_status-able.\n-        success = 299 >= status >= 200 and 'insertErrors' not in content\n-\n-        if success:\n-            return success\n+        if 299 >= response.status >= 200 and 'insertErrors' not in content:\n+            return True\n \n-        raise Exception('{}: {} > {}'.format(status, url, content))\n+        raise Exception(f'{response.status}: {url} > {content}')\n \n     # TODO: look into deletion payload format\n \ndiff --git a/datastore/nox.py b/datastore/nox.py\n--- a/datastore/nox.py\n+++ b/datastore/nox.py\n@@ -4,14 +4,14 @@\n import nox\n \n \n-LOCAL_DEPS = ('../core/', '../auth/')\n+LOCAL_DEPS = ('../auth/', )\n \n \n @nox.session\n @nox.parametrize('python_version', ['3.6'])\n def unit_tests(session, python_version):\n-    session.interpreter = 'python{}'.format(python_version)\n-    session.virtualenv_dirname = 'unit-' + python_version\n+    session.interpreter = f'python{python_version}'\n+    session.virtualenv_dirname = f'unit-{python_version}'\n \n     session.install('pytest', 'pytest-cov', *LOCAL_DEPS)\n     session.install('-e', '.')\n@@ -34,8 +34,8 @@ def integration_tests(session, python_version):\n     if not os.environ.get('GOOGLE_APPLICATION_CREDENTIALS', ''):\n         session.skip('Credentials must be set via environment variable.')\n \n-    session.interpreter = 'python{}'.format(python_version)\n-    session.virtualenv_dirname = 'integration-' + python_version\n+    session.interpreter = f'python{python_version}'\n+    session.virtualenv_dirname = f'integration-{python_version}'\n \n     session.install('pytest', *LOCAL_DEPS)\n     session.install('.')\n@@ -46,7 +46,7 @@ def integration_tests(session, python_version):\n @nox.session\n @nox.parametrize('python_version', ['3.6'])\n def lint_setup_py(session, python_version):\n-    session.interpreter = 'python{}'.format(python_version)\n+    session.interpreter = f'python{python_version}'\n     session.virtualenv_dirname = 'setup'\n \n     session.install('docutils', 'Pygments')\n@@ -61,7 +61,7 @@ def lint_setup_py(session, python_version):\n @nox.session\n @nox.parametrize('python_version', ['3.6'])\n def cover(session, python_version):\n-    session.interpreter = 'python{}'.format(python_version)\n+    session.interpreter = f'python{python_version}'\n     session.virtualenv_dirname = 'cover'\n \n     session.install('codecov', 'coverage', 'pytest-cov')\ndiff --git a/pubsub/nox.py b/pubsub/nox.py\n--- a/pubsub/nox.py\n+++ b/pubsub/nox.py\n@@ -7,8 +7,8 @@\n @nox.session\n @nox.parametrize('python_version', ['3.6'])\n def unit_tests(session, python_version):\n-    session.interpreter = 'python{}'.format(python_version)\n-    session.virtualenv_dirname = 'unit-' + python_version\n+    session.interpreter = f'python{python_version}'\n+    session.virtualenv_dirname = f'unit-{python_version}'\n \n     session.install('pytest', 'pytest-cov')\n     session.install('-e', '.')\n@@ -31,8 +31,8 @@ def integration_tests(session, python_version):\n     if not os.environ.get('GOOGLE_APPLICATION_CREDENTIALS', ''):\n         session.skip('Credentials must be set via environment variable.')\n \n-    session.interpreter = 'python{}'.format(python_version)\n-    session.virtualenv_dirname = 'integration-' + python_version\n+    session.interpreter = f'python{python_version}'\n+    session.virtualenv_dirname = f'integration-{python_version}'\n \n     session.install('pytest')\n     session.install('.')\n@@ -43,7 +43,7 @@ def integration_tests(session, python_version):\n @nox.session\n @nox.parametrize('python_version', ['3.6'])\n def lint_setup_py(session, python_version):\n-    session.interpreter = 'python{}'.format(python_version)\n+    session.interpreter = f'python{python_version}'\n     session.virtualenv_dirname = 'setup'\n \n     session.install('docutils', 'Pygments')\n@@ -58,7 +58,7 @@ def lint_setup_py(session, python_version):\n @nox.session\n @nox.parametrize('python_version', ['3.6'])\n def cover(session, python_version):\n-    session.interpreter = 'python{}'.format(python_version)\n+    session.interpreter = f'python{python_version}'\n     session.virtualenv_dirname = 'cover'\n \n     session.install('codecov', 'coverage', 'pytest-cov')\ndiff --git a/storage/gcloud/aio/storage/__init__.py b/storage/gcloud/aio/storage/__init__.py\n--- a/storage/gcloud/aio/storage/__init__.py\n+++ b/storage/gcloud/aio/storage/__init__.py\n@@ -1,10 +1,10 @@\n from pkg_resources import get_distribution\n __version__ = get_distribution('gcloud-aio-storage').version\n \n-from gcloud.aio.storage.storage import Blob\n-from gcloud.aio.storage.storage import Bucket\n-from gcloud.aio.storage.storage import make_download\n+from gcloud.aio.storage.blob import Blob\n+from gcloud.aio.storage.bucket import Bucket\n from gcloud.aio.storage.storage import Storage\n+from gcloud.aio.storage.utils import make_download\n \n \n-__all__ = ['__version__', 'Blob', 'Bucket', 'make_download', 'Storage']\n+__all__ = ['__version__', 'Blob', 'Bucket', 'Storage', 'make_download']\ndiff --git a/storage/gcloud/aio/storage/blob.py b/storage/gcloud/aio/storage/blob.py\nnew file mode 100644\n--- /dev/null\n+++ b/storage/gcloud/aio/storage/blob.py\n@@ -0,0 +1,33 @@\n+try:\n+    import ujson as json\n+except ModuleNotFoundError:\n+    import json\n+\n+\n+class Blob:\n+    def __init__(self, bucket, name, data):\n+        self.__dict__.update(**data)\n+\n+        self.bucket = bucket\n+        self.name = name\n+        self.size = int(self.size)\n+\n+    @property\n+    def chunk_size(self):\n+        return self.size + (262144 - (self.size % 262144))\n+\n+    async def download_as_string(self, session=None):\n+        return await self.bucket.storage.download_as_string(self.bucket.name,\n+                                                            self.name,\n+                                                            session=session)\n+\n+    async def upload_from_string(self, data, session=None):\n+        status, content = await self.bucket.storage.upload(self.bucket.name,\n+                                                           self.name, data,\n+                                                           session=session)\n+\n+        if status < 200 or status >= 300:\n+            raise Exception(f'{status}: {json.dumps(content)}')\n+\n+        self.__dict__.update(content)\n+        return content\ndiff --git a/storage/gcloud/aio/storage/bucket.py b/storage/gcloud/aio/storage/bucket.py\nnew file mode 100644\n--- /dev/null\n+++ b/storage/gcloud/aio/storage/bucket.py\n@@ -0,0 +1,47 @@\n+import logging\n+\n+from gcloud.aio.storage.blob import Blob\n+try:\n+    import ujson as json\n+except ModuleNotFoundError:\n+    import json\n+\n+\n+log = logging.getLogger(__name__)\n+\n+\n+class Bucket:\n+    def __init__(self, storage, name):\n+        self.storage = storage\n+        self.name = name\n+\n+    async def get_blob(self, blob_name, session=None):\n+        blob_name = blob_name.replace('/', '%2F')\n+\n+        status, content = await self.storage.download(self.name, blob_name,\n+                                                      session=session)\n+\n+        if status < 200 or status >= 300:\n+            log.error('Could not download %s/%s: %s', self.name, blob_name,\n+                      content)\n+            return\n+\n+        content = json.loads(content)\n+\n+        return Blob(self, blob_name, content)\n+\n+    async def list_blobs(self, prefix='', session=None):\n+        params = {'prefix': prefix}\n+\n+        status, content = await self.storage.list_objects(self.name,\n+                                                          params=params,\n+                                                          session=session)\n+\n+        if status < 200 or status >= 300:\n+            log.error('Could not list %s/%s: %s', self.name, prefix, content)\n+            return\n+\n+        return [x['name'] for x in content.get('items', list())]\n+\n+    def new_blob(self, blob_name):\n+        return Blob(self, blob_name, {'size': 0})\ndiff --git a/storage/gcloud/aio/storage/storage.py b/storage/gcloud/aio/storage/storage.py\n--- a/storage/gcloud/aio/storage/storage.py\n+++ b/storage/gcloud/aio/storage/storage.py\n@@ -1,18 +1,16 @@\n-import functools\n import logging\n-import mimetypes\n \n import aiohttp\n-import ujson\n from gcloud.aio.auth import Token\n-from gcloud.aio.core.http import get\n-from gcloud.aio.core.http import HttpError\n-from gcloud.aio.core.http import post\n+from gcloud.aio.storage.bucket import Bucket\n+try:\n+    import ujson as json\n+except ModuleNotFoundError:\n+    import json\n \n \n STORAGE_API_ROOT = 'https://www.googleapis.com/storage/v1/b'\n STORAGE_UPLOAD_API_ROOT = 'https://www.googleapis.com/upload/storage/v1/b'\n-READ_ONLY_SCOPE = 'https://www.googleapis.com/auth/devstorage.read_only'\n READ_WRITE_SCOPE = 'https://www.googleapis.com/auth/devstorage.read_write'\n \n log = logging.getLogger(__name__)\n@@ -28,64 +26,68 @@ def __init__(self, project, service_file, token=None, session=None):\n                                     scopes=[READ_WRITE_SCOPE])\n \n     async def download(self, bucket, object_name, params=None, session=None):\n-        session = session or self.session\n-\n         token = await self.token.get()\n-        url = '{}/{}/o/{}'.format(STORAGE_API_ROOT, bucket, object_name)\n+        url = f'{STORAGE_API_ROOT}/{bucket}/o/{object_name}'\n         headers = {\n-            'Authorization': 'Bearer {}'.format(token),\n+            'Authorization': f'Bearer {token}',\n         }\n \n-        return await get(url, params=params or {}, headers=headers,\n-                         session=self.session, json_response=False)\n-\n-    async def list_objects(self, bucket, params=None, session=None):\n         session = session or self.session\n+        response = await session.get(url, headers=headers, params=params or {},\n+                                     timeout=60)\n+        content = await response.text()\n+\n+        return response.status, content\n \n+    async def list_objects(self, bucket, params=None, session=None):\n         token = await self.token.get()\n-        url = '{}/{}/o'.format(STORAGE_API_ROOT, bucket)\n+        url = f'{STORAGE_API_ROOT}/{bucket}/o'\n         headers = {\n-            'Authorization': 'Bearer {}'.format(token),\n+            'Authorization': f'Bearer {token}',\n         }\n \n-        return await get(url, params=params or {}, headers=headers,\n-                         session=self.session, json_response=True)\n+        session = session or self.session\n+        response = await session.get(url, headers=headers, params=params or {},\n+                                     timeout=60)\n+        content = await response.json()\n+\n+        return response.status, content\n \n     async def upload(self, bucket, object_name, file_data, headers=None,\n                      session=None):\n         # pylint: disable=too-many-arguments\n         # https://cloud.google.com/storage/docs/json_api/v1/how-tos/simple-upload\n-        session = session or self.session\n-\n         token = await self.token.get()\n-        url = '{}/{}/o'.format(STORAGE_UPLOAD_API_ROOT, bucket)\n+        url = f'{STORAGE_UPLOAD_API_ROOT}/{bucket}/o'\n         headers = headers or {}\n \n-        # TODO: verify this\n-        if not isinstance(file_data, bytes):\n-            body = file_data.encode('utf-8')\n-        else:\n-            body = file_data\n-\n-        body_length = str(len(body))\n-\n         params = {\n             'name': object_name,\n             'uploadType': 'media',\n         }\n \n-        content_type = mimetypes.guess_type(object_name)[0]\n-        content_type = content_type or 'application/octet-stream'\n+        if not isinstance(file_data, bytes):\n+            file_data = file_data.encode('utf-8')\n+\n+        if file_data:\n+            file_data = json.dumps(file_data).encode('utf-8')\n+            content_length = str(len(file_data))\n+        else:\n+            content_length = '0'\n \n         headers.update({\n-            'accept': 'application/json',\n-            'Authorization': 'Bearer {}'.format(token),\n-            'Content-Length': body_length,\n-            'Content-Type': content_type,\n+            'Accept': 'application/json',\n+            'Authorization': f'Bearer {token}',\n+            'Content-Length': content_length,\n+            'Content-Type': 'application/json',\n         })\n \n-        return await post(url, params=params, payload=body, headers=headers,\n-                          timeout=120, session=session)\n+        session = session or self.session\n+        response = await session.post(url, data=file_data, headers=headers,\n+                                      params=params, timeout=120)\n+        content = await response.json()\n+\n+        return response.status, content\n \n     async def download_as_string(self, bucket, object_name, session=None):\n         object_name = object_name.replace('/', '%2F')\n@@ -98,88 +100,3 @@ async def download_as_string(self, bucket, object_name, session=None):\n \n     def get_bucket(self, bucket_name):\n         return Bucket(self, bucket_name)\n-\n-\n-class Bucket:\n-    def __init__(self, storage, name):\n-        self.storage = storage\n-        self.name = name\n-\n-    async def get_blob(self, blob_name, session=None):\n-        blob_name = blob_name.replace('/', '%2F')\n-\n-        status, content = await self.storage.download(self.name, blob_name,\n-                                                      session=session)\n-\n-        if status < 200 or status >= 300:\n-            log.error('Could not download %s/%s: %s', self.name, blob_name,\n-                      content)\n-            return\n-\n-        content = ujson.loads(content)\n-\n-        return Blob(self, blob_name, content)\n-\n-    async def list_blobs(self, prefix='', session=None):\n-        params = {'prefix': prefix}\n-\n-        status, content = await self.storage.list_objects(self.name,\n-                                                          params=params,\n-                                                          session=session)\n-\n-        if status < 200 or status >= 300:\n-            log.error('Could not list %s/%s: %s', self.name, prefix, content)\n-            return\n-\n-        return [x['name'] for x in content.get('items', list())]\n-\n-    def new_blob(self, blob_name):\n-        return Blob(self, blob_name, {'size': 0})\n-\n-\n-class Blob:\n-    def __init__(self, bucket, name, data):\n-        self.__dict__.update(**data)\n-\n-        self.bucket = bucket\n-        self.name = name\n-        self.size = int(self.size)\n-\n-    @property\n-    def chunk_size(self):\n-        return self.size + (262144 - (self.size % 262144))\n-\n-    async def download_as_string(self, session=None):\n-        return await self.bucket.storage.download_as_string(self.bucket.name,\n-                                                            self.name,\n-                                                            session=session)\n-\n-    async def upload_from_string(self, data, session=None):\n-        status, content = await self.bucket.storage.upload(self.bucket.name,\n-                                                           self.name, data,\n-                                                           session=session)\n-\n-        if status < 200 or status >= 300:\n-            raise HttpError('{}: {}'.format(status, ujson.dumps(content)))\n-\n-        self.__dict__.update(content)\n-        return content\n-\n-\n-async def download(bucket, object_name):\n-    blob = await bucket.get_blob(object_name)\n-    if not blob:\n-        raise Exception('No such object \"{}/{}\"'.format(bucket.name,\n-                                                        object_name))\n-\n-    return await blob.download_as_string()\n-\n-\n-def make_download(project, service_file, bucket_name, session=None,\n-                  token=None):\n-    token = token or Token(project, service_file, scopes=[READ_ONLY_SCOPE])\n-\n-    storage = Storage(project, service_file, session=session, token=token)\n-    bucket = storage.get_bucket(bucket_name)\n-\n-    return functools.partial(download, bucket)\ndiff --git a/storage/gcloud/aio/storage/utils.py b/storage/gcloud/aio/storage/utils.py\nnew file mode 100644\n--- /dev/null\n+++ b/storage/gcloud/aio/storage/utils.py\n@@ -0,0 +1,25 @@\n+import functools\n+\n+from gcloud.aio.auth import Token\n+from gcloud.aio.storage.storage import Storage\n+\n+\n+READ_ONLY_SCOPE = 'https://www.googleapis.com/auth/devstorage.read_only'\n+\n+\n+async def download(bucket, object_name):\n+    blob = await bucket.get_blob(object_name)\n+    if not blob:\n+        raise Exception(f'No such object \"{bucket.name}/{object_name}\"')\n+\n+    return await blob.download_as_string()\n+\n+\n+def make_download(project, service_file, bucket_name, session=None,\n+                  token=None):\n+    token = token or Token(project, service_file, scopes=[READ_ONLY_SCOPE])\n+\n+    storage = Storage(project, service_file, session=session, token=token)\n+    bucket = storage.get_bucket(bucket_name)\n+\n+    return functools.partial(download, bucket)\ndiff --git a/storage/nox.py b/storage/nox.py\n--- a/storage/nox.py\n+++ b/storage/nox.py\n@@ -4,14 +4,14 @@\n import nox\n \n \n-LOCAL_DEPS = ('../core/', '../auth/')\n+LOCAL_DEPS = ('../auth/', )\n \n \n @nox.session\n @nox.parametrize('python_version', ['3.6'])\n def unit_tests(session, python_version):\n-    session.interpreter = 'python{}'.format(python_version)\n-    session.virtualenv_dirname = 'unit-' + python_version\n+    session.interpreter = f'python{python_version}'\n+    session.virtualenv_dirname = f'unit-{python_version}'\n \n     session.install('pytest', 'pytest-cov', *LOCAL_DEPS)\n     session.install('-e', '.')\n@@ -34,8 +34,8 @@ def integration_tests(session, python_version):\n     if not os.environ.get('GOOGLE_APPLICATION_CREDENTIALS', ''):\n         session.skip('Credentials must be set via environment variable.')\n \n-    session.interpreter = 'python{}'.format(python_version)\n-    session.virtualenv_dirname = 'integration-' + python_version\n+    session.interpreter = f'python{python_version}'\n+    session.virtualenv_dirname = f'integration-{python_version}'\n \n     session.install('aiohttp', 'pytest', *LOCAL_DEPS)\n     session.install('.')\n@@ -46,7 +46,7 @@ def integration_tests(session, python_version):\n @nox.session\n @nox.parametrize('python_version', ['3.6'])\n def lint_setup_py(session, python_version):\n-    session.interpreter = 'python{}'.format(python_version)\n+    session.interpreter = f'python{python_version}'\n     session.virtualenv_dirname = 'setup'\n \n     session.install('docutils', 'Pygments')\n@@ -61,7 +61,7 @@ def lint_setup_py(session, python_version):\n @nox.session\n @nox.parametrize('python_version', ['3.6'])\n def cover(session, python_version):\n-    session.interpreter = 'python{}'.format(python_version)\n+    session.interpreter = f'python{python_version}'\n     session.virtualenv_dirname = 'cover'\n \n     session.install('codecov', 'coverage', 'pytest-cov')\ndiff --git a/taskqueue/gcloud/aio/taskqueue/taskmanager.py b/taskqueue/gcloud/aio/taskqueue/taskmanager.py\n--- a/taskqueue/gcloud/aio/taskqueue/taskmanager.py\n+++ b/taskqueue/gcloud/aio/taskqueue/taskmanager.py\n@@ -59,9 +59,9 @@ async def stop(self):\n \n     @staticmethod\n     def autorenew(event, headers, task, lease_seconds):\n-        url = '{}/{}:renewLease'.format(API_ROOT, task['name'])\n+        url = f'{API_ROOT}/{task[\"name\"]}:renewLease'\n         body = {\n-            'leaseDuration': '{}s'.format(lease_seconds),\n+            'leaseDuration': f'{lease_seconds}s',\n             'responseView': 'FULL',\n         }\n \ndiff --git a/taskqueue/gcloud/aio/taskqueue/taskqueue.py b/taskqueue/gcloud/aio/taskqueue/taskqueue.py\n--- a/taskqueue/gcloud/aio/taskqueue/taskqueue.py\n+++ b/taskqueue/gcloud/aio/taskqueue/taskqueue.py\n@@ -28,21 +28,22 @@ def __init__(self, project, service_file, taskqueue, location=LOCATION,\n         self.session = session or aiohttp.ClientSession(conn_timeout=10,\n                                                         read_timeout=10)\n \n-        self.api_root = '{}/projects/{}/locations/{}/queues/{}'.format(\n-            API_ROOT, project, location, taskqueue)\n+        self.api_root = (f'{API_ROOT}/projects/{project}/'\n+                         f'locations/{location}/queues/{taskqueue}')\n \n         self.token = token or Token(project, service_file, scopes=SCOPES,\n                                     session=self.session)\n \n     async def headers(self):\n+        token = await self.token.get()\n         return {\n-            'Authorization': 'Bearer {}'.format(await self.token.get()),\n+            'Authorization': f'Bearer {token}',\n             'Content-Type': 'application/json',\n         }\n \n     # https://cloud.google.com/cloud-tasks/docs/reference/rest/v2beta2/projects.locations.queues.tasks/acknowledge\n     async def ack(self, task, session=None):\n-        url = '{}/{}:acknowledge'.format(API_ROOT, task['name'])\n+        url = f'{API_ROOT}/{task[\"name\"]}:acknowledge'\n         body = {\n             'scheduleTime': task['scheduleTime'],\n         }\n@@ -55,7 +56,7 @@ async def ack(self, task, session=None):\n \n     # https://cloud.google.com/cloud-tasks/docs/reference/rest/v2beta2/projects.locations.queues.tasks/cancelLease\n     async def cancel(self, task, session=None):\n-        url = '{}/{}:cancelLease'.format(API_ROOT, task['name'])\n+        url = f'{API_ROOT}/{task[\"name\"]}:cancelLease'\n         body = {\n             'scheduleTime': task['scheduleTime'],\n             'responseView': 'BASIC',\n@@ -69,7 +70,7 @@ async def cancel(self, task, session=None):\n \n     # https://cloud.google.com/cloud-tasks/docs/reference/rest/v2beta2/projects.locations.queues.tasks/delete\n     async def delete(self, tname, session=None):\n-        url = '{}/{}'.format(API_ROOT, tname)\n+        url = f'{API_ROOT}/{tname}'\n \n         s = session or self.session\n         resp = await retry(s.delete(url, headers=await self.headers()))\n@@ -84,7 +85,7 @@ async def drain(self):\n \n     # https://cloud.google.com/cloud-tasks/docs/reference/rest/v2beta2/projects.locations.queues.tasks/get\n     async def get(self, tname, full=False, session=None):\n-        url = '{}/{}'.format(API_ROOT, tname)\n+        url = f'{API_ROOT}/{tname}'\n         params = {\n             'responseView': 'FULL' if full else 'BASIC',\n         }\n@@ -97,7 +98,7 @@ async def get(self, tname, full=False, session=None):\n \n     # https://cloud.google.com/cloud-tasks/docs/reference/rest/v2beta2/projects.locations.queues.tasks/create\n     async def insert(self, payload, tag=None, session=None):\n-        url = '{}/tasks'.format(self.api_root)\n+        url = f'{self.api_root}/tasks'\n         body = {\n             'task': {\n                 'pullMessage': {\n@@ -117,10 +118,10 @@ async def insert(self, payload, tag=None, session=None):\n     # https://cloud.google.com/cloud-tasks/docs/reference/rest/v2beta2/projects.locations.queues.tasks/lease\n     async def lease(self, num_tasks=1, lease_seconds=60, task_filter=None,\n                     session=None):\n-        url = '{}/tasks:lease'.format(self.api_root)\n+        url = f'{self.api_root}/tasks:lease'\n         body = {\n             'maxTasks': min(num_tasks, 1000),\n-            'leaseDuration': '{}s'.format(lease_seconds),\n+            'leaseDuration': f'{lease_seconds}s',\n             'responseView': 'FULL',\n         }\n         if task_filter:\n@@ -135,7 +136,7 @@ async def lease(self, num_tasks=1, lease_seconds=60, task_filter=None,\n     # https://cloud.google.com/cloud-tasks/docs/reference/rest/v2beta2/projects.locations.queues.tasks/list\n     async def list(self, full=False, page_size=1000, page_token='',\n                    session=None):\n-        url = '{}/tasks'.format(self.api_root)\n+        url = f'{self.api_root}/tasks'\n         params = {\n             'responseView': 'FULL' if full else 'BASIC',\n             'pageSize': page_size,\n@@ -150,10 +151,10 @@ async def list(self, full=False, page_size=1000, page_token='',\n \n     # https://cloud.google.com/cloud-tasks/docs/reference/rest/v2beta2/projects.locations.queues.tasks/renewLease\n     async def renew(self, task, lease_seconds=60, session=None):\n-        url = '{}/{}:renewLease'.format(API_ROOT, task['name'])\n+        url = f'{API_ROOT}/{task[\"name\"]}:renewLease'\n         body = {\n             'scheduleTime': task['scheduleTime'],\n-            'leaseDuration': '{}s'.format(lease_seconds),\n+            'leaseDuration': f'{lease_seconds}s',\n             'responseView': 'FULL',\n         }\n \ndiff --git a/taskqueue/nox.py b/taskqueue/nox.py\n--- a/taskqueue/nox.py\n+++ b/taskqueue/nox.py\n@@ -10,8 +10,8 @@\n @nox.session\n @nox.parametrize('python_version', ['3.6'])\n def unit_tests(session, python_version):\n-    session.interpreter = 'python{}'.format(python_version)\n-    session.virtualenv_dirname = 'unit-' + python_version\n+    session.interpreter = f'python{python_version}'\n+    session.virtualenv_dirname = f'unit-{python_version}'\n \n     session.install('pytest', 'pytest-cov', *LOCAL_DEPS)\n     session.install('-e', '.')\n@@ -34,8 +34,8 @@ def integration_tests(session, python_version):\n     if not os.environ.get('GOOGLE_APPLICATION_CREDENTIALS', ''):\n         session.skip('Credentials must be set via environment variable.')\n \n-    session.interpreter = 'python{}'.format(python_version)\n-    session.virtualenv_dirname = 'integration-' + python_version\n+    session.interpreter = f'python{python_version}'\n+    session.virtualenv_dirname = f'integration-{python_version}'\n \n     session.install('pytest', 'pytest-mock', *LOCAL_DEPS)\n     session.install('.')\n@@ -46,7 +46,7 @@ def integration_tests(session, python_version):\n @nox.session\n @nox.parametrize('python_version', ['3.6'])\n def lint_setup_py(session, python_version):\n-    session.interpreter = 'python{}'.format(python_version)\n+    session.interpreter = f'python{python_version}'\n     session.virtualenv_dirname = 'setup'\n \n     session.install('docutils', 'Pygments')\n@@ -61,7 +61,7 @@ def lint_setup_py(session, python_version):\n @nox.session\n @nox.parametrize('python_version', ['3.6'])\n def cover(session, python_version):\n-    session.interpreter = 'python{}'.format(python_version)\n+    session.interpreter = f'python{python_version}'\n     session.virtualenv_dirname = 'cover'\n \n     session.install('codecov', 'coverage', 'pytest-cov')\n", "test_patch": "diff --git a/core/tests/__init__.py b/core/tests/__init__.py\ndeleted file mode 100644\ndiff --git a/core/tests/unit/__init__.py b/core/tests/unit/__init__.py\ndeleted file mode 100644\ndiff --git a/core/tests/unit/aio_test.py b/core/tests/unit/aio_test.py\ndeleted file mode 100644\n--- a/core/tests/unit/aio_test.py\n+++ /dev/null\n@@ -1,5 +0,0 @@\n-import gcloud.aio.core.aio as aio  # pylint: disable=unused-import\n-\n-\n-def test_importable():\n-    assert True\ndiff --git a/core/tests/unit/astate_test.py b/core/tests/unit/astate_test.py\ndeleted file mode 100644\n--- a/core/tests/unit/astate_test.py\n+++ /dev/null\n@@ -1,5 +0,0 @@\n-import gcloud.aio.core.astate as astate  # pylint: disable=unused-import\n-\n-\n-def test_importable():\n-    assert True\ndiff --git a/core/tests/unit/http_test.py b/core/tests/unit/http_test.py\ndeleted file mode 100644\n--- a/core/tests/unit/http_test.py\n+++ /dev/null\n@@ -1,5 +0,0 @@\n-import gcloud.aio.core.http as http  # pylint: disable=unused-import\n-\n-\n-def test_importable():\n-    assert True\ndiff --git a/datastore/tests/integration/smoke_test.py b/datastore/tests/integration/smoke_test.py\n--- a/datastore/tests/integration/smoke_test.py\n+++ b/datastore/tests/integration/smoke_test.py\n@@ -24,7 +24,7 @@ def test_item_lifecycle():\n     creds = os.environ['GOOGLE_APPLICATION_CREDENTIALS']\n \n     kind_name = 'gcloud-aio-test'\n-    object_name = 'test_record_{}'.format(uuid.uuid4())\n+    object_name = f'test_record_{uuid.uuid4()}'\n \n     loop = asyncio.get_event_loop()\n     loop.run_until_complete(\ndiff --git a/storage/tests/integration/smoke_test.py b/storage/tests/integration/smoke_test.py\n--- a/storage/tests/integration/smoke_test.py\n+++ b/storage/tests/integration/smoke_test.py\n@@ -21,7 +21,7 @@ def test_object_is_downloaded():\n     call_id = '07fbe0cc-7f87-1235-06b0-0cc47a392728'\n     side = 'callee'\n     link = 0\n-    object_name = '{}/{}/{}/rtp.pcap.wav.ctm'.format(call_id, side, link)\n+    object_name = f'{call_id}/{side}/{link}/rtp.pcap.wav.ctm'\n \n     loop = asyncio.get_event_loop()\n     loop.run_until_complete(\ndiff --git a/storage/tests/unit/blob_test.py b/storage/tests/unit/blob_test.py\nnew file mode 100644\n--- /dev/null\n+++ b/storage/tests/unit/blob_test.py\n@@ -0,0 +1,5 @@\n+import gcloud.aio.storage.blob as blob  # pylint: disable=unused-import\n+\n+\n+def test_importable():\n+    assert True\ndiff --git a/storage/tests/unit/bucket_test.py b/storage/tests/unit/bucket_test.py\nnew file mode 100644\n--- /dev/null\n+++ b/storage/tests/unit/bucket_test.py\n@@ -0,0 +1,5 @@\n+import gcloud.aio.storage.bucket as bucket  # pylint: disable=unused-import\n+\n+\n+def test_importable():\n+    assert True\ndiff --git a/storage/tests/unit/storage_test.py b/storage/tests/unit/storage_test.py\n--- a/storage/tests/unit/storage_test.py\n+++ b/storage/tests/unit/storage_test.py\n@@ -1,4 +1,4 @@\n-import gcloud.aio.storage as storage  # pylint: disable=unused-import\n+import gcloud.aio.storage.storage as storage  # pylint: disable=unused-import\n \n \n def test_importable():\ndiff --git a/storage/tests/unit/utils_test.py b/storage/tests/unit/utils_test.py\nnew file mode 100644\n--- /dev/null\n+++ b/storage/tests/unit/utils_test.py\n@@ -0,0 +1,5 @@\n+import gcloud.aio.storage.utils as utils  # pylint: disable=unused-import\n+\n+\n+def test_importable():\n+    assert True\ndiff --git a/taskqueue/tests/integration/taskqueue_test.py b/taskqueue/tests/integration/taskqueue_test.py\n--- a/taskqueue/tests/integration/taskqueue_test.py\n+++ b/taskqueue/tests/integration/taskqueue_test.py\n@@ -26,8 +26,7 @@ async def do_task_lifecycle(project, creds, task_queue):\n         assert inserted\n \n         # GET\n-        got = await tq.get(inserted['name'], full=True)\n-        assert got == inserted\n+        assert inserted == await tq.get(inserted['name'], full=True)\n \n         # LIST\n         listed = await tq.list(full=True)\n@@ -36,14 +35,12 @@ async def do_task_lifecycle(project, creds, task_queue):\n \n         # LEASE\n         leased = await tq.lease(num_tasks=1, lease_seconds=10,\n-                                task_filter='tag={}'.format(encode(tag)))\n+                                task_filter=f'tag={encode(tag)}')\n         assert leased.get('tasks') and len(leased['tasks']) == 1\n \n         leased_message = leased['tasks'][0]['pullMessage']\n-        leased_payload = json.loads(decode(leased_message['payload']))\n-        leased_tag = decode(leased_message['tag'])\n-        assert leased_payload == payload\n-        assert leased_tag == tag\n+        assert payload == json.loads(decode(leased_message['payload']))\n+        assert tag == decode(leased_message['tag'])\n \n         # RENEW\n         renewed = await tq.renew(leased['tasks'][0], lease_seconds=10)\n@@ -57,8 +54,7 @@ async def do_task_lifecycle(project, creds, task_queue):\n         # cancel?\n \n         # DELETE\n-        result = await tq.delete(renewed['name'])\n-        assert not result\n+        assert not await tq.delete(renewed['name'])\n \n \n def test_task_lifecycle():\n", "problem_statement": "", "hints_text": "", "created_at": "2018-05-01T17:56:36Z"}