{"repo": "googleapis/google-cloud-python", "pull_number": 2223, "instance_id": "googleapis__google-cloud-python-2223", "issue_numbers": "", "base_commit": "89eed0e4251f44b0ae1b0e33d3d9ffdac1d3c330", "patch": "diff --git a/docs/bigquery_snippets.py b/docs/bigquery_snippets.py\n--- a/docs/bigquery_snippets.py\n+++ b/docs/bigquery_snippets.py\n@@ -15,8 +15,8 @@\n \"\"\"Testable usage examples for Google Cloud BigQuery API wrapper\n \n Each example function takes a ``client`` argument (which must be an instance\n-of :class:`gcloud.bigquery.client.Client`) and uses it to perform a task with\n-the API.\n+of :class:`google.cloud.bigquery.client.Client`) and uses it to perform a task\n+with the API.\n \n To facilitate running the examples as system tests, each example is also passed\n a ``to_delete`` list; the function adds to the list any objects created which\n@@ -26,8 +26,8 @@\n import operator\n import time\n \n-from gcloud.bigquery import SchemaField\n-from gcloud.bigquery.client import Client\n+from google.cloud.bigquery import SchemaField\n+from google.cloud.bigquery.client import Client\n \n ORIGINAL_FRIENDLY_NAME = 'Original friendly name'\n ORIGINAL_DESCRIPTION = 'Original description'\n@@ -162,7 +162,7 @@ def dataset_update(client, to_delete):\n dataset.reload()\n \n # [START dataset_update]\n- from gcloud.bigquery import AccessGrant\n+ from google.cloud.bigquery import AccessGrant\n assert dataset.description == ORIGINAL_DESCRIPTION\n assert dataset.default_table_expiration_ms is None\n grant = AccessGrant(\ndiff --git a/docs/conf.py b/docs/conf.py\n--- a/docs/conf.py\n+++ b/docs/conf.py\n@@ -1,6 +1,6 @@\n # -*- coding: utf-8 -*-\n #\n-# gcloud documentation build configuration file, created by\n+# google-cloud documentation build configuration file, created by\n # sphinx-quickstart on Tue Jan 21 22:24:47 2014.\n #\n # This file is execfile()d with the current directory set to its containing dir.\n@@ -56,7 +56,7 @@\n master_doc = 'index'\n \n # General information about the project.\n-project = u'gcloud'\n+project = u'google-cloud'\n copyright = u'2014, Google'\n \n # The version info for the project you're documenting, acts as replacement for\n@@ -64,7 +64,7 @@\n # built documents.\n #\n # The short X.Y version.\n-distro = get_distribution('gcloud')\n+distro = get_distribution('google-cloud')\n release = os.getenv('SPHINX_RELEASE', distro.version)\n \n # The language for content autogenerated by Sphinx. Refer to documentation\n@@ -184,7 +184,7 @@\n #html_file_suffix = None\n \n # Output file base name for HTML help builder.\n-htmlhelp_basename = 'gclouddoc'\n+htmlhelp_basename = 'google-cloud-doc'\n \n html_context = {}\n \n@@ -207,7 +207,7 @@\n # Grouping the document tree into LaTeX files. List of tuples\n # (source start file, target name, title, author, documentclass [howto/manual]).\n latex_documents = [\n- ('index', 'gcloud.tex', u'gCloud Documentation',\n+ ('index', 'google-cloud.tex', u'google-cloud Documentation',\n author, 'manual'),\n ]\n \n@@ -237,7 +237,7 @@\n # One entry per manual page. List of tuples\n # (source start file, name, description, authors, manual section).\n man_pages = [\n- ('index', 'gcloud', u'gCloud Documentation',\n+ ('index', 'google-cloud', u'google-cloud Documentation',\n [author], 1)\n ]\n \n@@ -251,8 +251,8 @@\n # (source start file, target name, title, author,\n # dir menu entry, description, category)\n texinfo_documents = [\n- ('index', 'gcloud', u'gCloud Documentation',\n- author, 'gcloud', 'Python API for Google Cloud.',\n+ ('index', 'google-cloud', u'google-cloud Documentation',\n+ author, 'google-cloud', 'Python API for Google Cloud.',\n 'Miscellaneous'),\n ]\n \n@@ -269,7 +269,7 @@\n # and parameter definitions from the __init__ docstring.\n autoclass_content = 'both'\n \n-issue_uri = ('https://github.com/GoogleCloudPlatform/gcloud-python/issues/'\n+issue_uri = ('https://github.com/GoogleCloudPlatform/google-cloud-python/issues/'\n 'new?' + urllib.urlencode({'title': '[Documentation Issue] '}))\n issue_uri_template = (\n issue_uri + '&' + urllib.urlencode({'body': 'Page Name: '}) + '{0}' +\ndiff --git a/docs/pubsub_snippets.py b/docs/pubsub_snippets.py\n--- a/docs/pubsub_snippets.py\n+++ b/docs/pubsub_snippets.py\n@@ -15,8 +15,8 @@\n \"\"\"Testable usage examples for Google Cloud Pubsub API wrapper\n \n Each example function takes a ``client`` argument (which must be an instance\n-of :class:`gcloud.pubsub.client.Client`) and uses it to perform a task with\n-the API.\n+of :class:`google.cloud.pubsub.client.Client`) and uses it to perform a task\n+with the API.\n \n To facilitate running the examples as system tests, each example is also passed\n a ``to_delete`` list; the function adds to the list any objects created which\n@@ -25,7 +25,7 @@\n \n import time\n \n-from gcloud.pubsub.client import Client\n+from google.cloud.pubsub.client import Client\n \n \n def snippet(func):\n@@ -154,7 +154,7 @@ def topic_check_iam_permissions(client, to_delete):\n to_delete.append(topic)\n \n # [START topic_check_iam_permissions]\n- from gcloud.pubsub.iam import OWNER_ROLE, EDITOR_ROLE, VIEWER_ROLE\n+ from google.cloud.pubsub.iam import OWNER_ROLE, EDITOR_ROLE, VIEWER_ROLE\n TO_CHECK = [OWNER_ROLE, EDITOR_ROLE, VIEWER_ROLE]\n ALLOWED = topic.check_iam_permissions(TO_CHECK)\n assert set(ALLOWED) == set(TO_CHECK)\n@@ -376,7 +376,7 @@ def do_something_with(message): # pylint: disable=unused-argument\n extras.append(message.attributes)\n \n # [START subscription_pull_autoack]\n- from gcloud.pubsub.subscription import AutoAck\n+ from google.cloud.pubsub.subscription import AutoAck\n with AutoAck(subscription, max_messages=10) as ack:\n for ack_id, message in list(ack.items()):\n try:\n@@ -436,7 +436,7 @@ def subscription_check_iam_permissions(client, to_delete):\n to_delete.append(subscription)\n \n # [START subscription_check_iam_permissions]\n- from gcloud.pubsub.iam import OWNER_ROLE, EDITOR_ROLE, VIEWER_ROLE\n+ from google.cloud.pubsub.iam import OWNER_ROLE, EDITOR_ROLE, VIEWER_ROLE\n TO_CHECK = [OWNER_ROLE, EDITOR_ROLE, VIEWER_ROLE]\n ALLOWED = subscription.check_iam_permissions(TO_CHECK)\n assert set(ALLOWED) == set(TO_CHECK)\ndiff --git a/google/__init__.py b/google/__init__.py\nnew file mode 100644\n--- /dev/null\n+++ b/google/__init__.py\n@@ -0,0 +1,22 @@\n+# Copyright 2016 Google Inc. All rights reserved.\n+#\n+# Licensed under the Apache License, Version 2.0 (the \"License\");\n+# you may not use this file except in compliance with the License.\n+# You may obtain a copy of the License at\n+#\n+# http://www.apache.org/licenses/LICENSE-2.0\n+#\n+# Unless required by applicable law or agreed to in writing, software\n+# distributed under the License is distributed on an \"AS IS\" BASIS,\n+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n+# See the License for the specific language governing permissions and\n+# limitations under the License.\n+\n+\"\"\"Base ``google`` namespace package.\"\"\"\n+\n+try:\n+ import pkg_resources\n+ pkg_resources.declare_namespace(__name__)\n+except ImportError:\n+ import pkgutil\n+ __path__ = pkgutil.extend_path(__path__, __name__)\ndiff --git a/gcloud/__init__.py b/google/cloud/__init__.py\nsimilarity index 73%\nrename from gcloud/__init__.py\nrename to google/cloud/__init__.py\n--- a/gcloud/__init__.py\n+++ b/google/cloud/__init__.py\n@@ -12,8 +12,11 @@\n # See the License for the specific language governing permissions and\n # limitations under the License.\n \n-\"\"\"GCloud API access in idiomatic Python.\"\"\"\n+\"\"\"Google Cloud API access in idiomatic Python.\"\"\"\n \n-from pkg_resources import get_distribution\n-\n-__version__ = get_distribution('gcloud').version\n+try:\n+ import pkg_resources\n+ pkg_resources.declare_namespace(__name__)\n+except ImportError:\n+ import pkgutil\n+ __path__ = pkgutil.extend_path(__path__, __name__)\ndiff --git a/gcloud/_helpers.py b/google/cloud/_helpers.py\nsimilarity index 97%\nrename from gcloud/_helpers.py\nrename to google/cloud/_helpers.py\n--- a/gcloud/_helpers.py\n+++ b/google/cloud/_helpers.py\n@@ -11,9 +11,10 @@\n # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n # See the License for the specific language governing permissions and\n # limitations under the License.\n-\"\"\"Thread-local resource stack.\n \n-This module is not part of the public API surface of `gcloud`.\n+\"\"\"Shared helpers for Google Cloud packages.\n+\n+This module is not part of the public API surface.\n \"\"\"\n \n import calendar\n@@ -41,8 +42,10 @@\n from six.moves.http_client import HTTPConnection\n from six.moves import configparser\n \n-from gcloud.environment_vars import PROJECT\n-from gcloud.environment_vars import CREDENTIALS\n+# pylint: disable=ungrouped-imports\n+from google.cloud.environment_vars import PROJECT\n+from google.cloud.environment_vars import CREDENTIALS\n+# pylint: enable=ungrouped-imports\n \n \n _NOW = datetime.datetime.utcnow # To be replaced by tests.\n@@ -72,8 +75,8 @@\n class _LocalStack(Local):\n \"\"\"Manage a thread-local LIFO stack of resources.\n \n- Intended for use in :class:`gcloud.datastore.batch.Batch.__enter__`,\n- :class:`gcloud.storage.batch.Batch.__enter__`, etc.\n+ Intended for use in :class:`google.cloud.datastore.batch.Batch.__enter__`,\n+ :class:`google.cloud.storage.batch.Batch.__enter__`, etc.\n \"\"\"\n def __init__(self):\n super(_LocalStack, self).__init__()\n@@ -290,7 +293,7 @@ def _determine_default_project(project=None):\n In implicit case, supports three environments. In order of precedence, the\n implicit environments are:\n \n- * GCLOUD_PROJECT environment variable\n+ * GOOGLE_CLOUD_PROJECT environment variable\n * GOOGLE_APPLICATION_CREDENTIALS JSON file\n * Get default service project from\n ``$ gcloud beta auth application-default login``\ndiff --git a/gcloud/bigquery/__init__.py b/google/cloud/bigquery/__init__.py\nsimilarity index 59%\nrename from gcloud/bigquery/__init__.py\nrename to google/cloud/bigquery/__init__.py\n--- a/gcloud/bigquery/__init__.py\n+++ b/google/cloud/bigquery/__init__.py\n@@ -16,17 +16,19 @@\n \n The main concepts with this API are:\n \n-- :class:`gcloud.bigquery.dataset.Dataset` represents an collection of tables.\n+- :class:`~google.cloud.bigquery.dataset.Dataset` represents a\n+ collection of tables.\n \n-- :class:`gcloud.bigquery.table.Table` represents a single \"relation\".\n+- :class:`~google.cloud.bigquery.table.Table` represents a single \"relation\".\n \"\"\"\n \n-from gcloud.bigquery.client import Client\n-from gcloud.bigquery.connection import Connection\n-from gcloud.bigquery.dataset import AccessGrant\n-from gcloud.bigquery.dataset import Dataset\n-from gcloud.bigquery.schema import SchemaField\n-from gcloud.bigquery.table import Table\n+\n+from google.cloud.bigquery.client import Client\n+from google.cloud.bigquery.connection import Connection\n+from google.cloud.bigquery.dataset import AccessGrant\n+from google.cloud.bigquery.dataset import Dataset\n+from google.cloud.bigquery.schema import SchemaField\n+from google.cloud.bigquery.table import Table\n \n \n SCOPE = Connection.SCOPE\ndiff --git a/gcloud/bigquery/_helpers.py b/google/cloud/bigquery/_helpers.py\nsimilarity index 98%\nrename from gcloud/bigquery/_helpers.py\nrename to google/cloud/bigquery/_helpers.py\n--- a/gcloud/bigquery/_helpers.py\n+++ b/google/cloud/bigquery/_helpers.py\n@@ -14,7 +14,7 @@\n \n \"\"\"Shared helper functions for BigQuery API classes.\"\"\"\n \n-from gcloud._helpers import _datetime_from_microseconds\n+from google.cloud._helpers import _datetime_from_microseconds\n \n \n def _not_null(value, field):\ndiff --git a/gcloud/bigquery/client.py b/google/cloud/bigquery/client.py\nsimilarity index 85%\nrename from gcloud/bigquery/client.py\nrename to google/cloud/bigquery/client.py\n--- a/gcloud/bigquery/client.py\n+++ b/google/cloud/bigquery/client.py\n@@ -15,14 +15,14 @@\n \"\"\"Client for interacting with the Google BigQuery API.\"\"\"\n \n \n-from gcloud.client import JSONClient\n-from gcloud.bigquery.connection import Connection\n-from gcloud.bigquery.dataset import Dataset\n-from gcloud.bigquery.job import CopyJob\n-from gcloud.bigquery.job import ExtractTableToStorageJob\n-from gcloud.bigquery.job import LoadTableFromStorageJob\n-from gcloud.bigquery.job import QueryJob\n-from gcloud.bigquery.query import QueryResults\n+from google.cloud.client import JSONClient\n+from google.cloud.bigquery.connection import Connection\n+from google.cloud.bigquery.dataset import Dataset\n+from google.cloud.bigquery.job import CopyJob\n+from google.cloud.bigquery.job import ExtractTableToStorageJob\n+from google.cloud.bigquery.job import LoadTableFromStorageJob\n+from google.cloud.bigquery.job import QueryJob\n+from google.cloud.bigquery.query import QueryResults\n \n \n class Client(JSONClient):\n@@ -68,8 +68,8 @@ def list_datasets(self, include_all=False, max_results=None,\n datasets.\n \n :rtype: tuple, (list, str)\n- :returns: list of :class:`gcloud.bigquery.dataset.Dataset`, plus a\n- \"next page token\" string: if the token is not None,\n+ :returns: list of :class:`~google.cloud.bigquery.dataset.Dataset`,\n+ plus a \"next page token\" string: if the token is not None,\n indicates that more datasets can be retrieved with another\n call (pass that value as ``page_token``).\n \"\"\"\n@@ -97,7 +97,7 @@ def dataset(self, dataset_name):\n :type dataset_name: str\n :param dataset_name: Name of the dataset.\n \n- :rtype: :class:`gcloud.bigquery.dataset.Dataset`\n+ :rtype: :class:`google.cloud.bigquery.dataset.Dataset`\n :returns: a new ``Dataset`` instance\n \"\"\"\n return Dataset(dataset_name, client=self)\n@@ -109,11 +109,11 @@ def job_from_resource(self, resource):\n :param resource: one job resource from API response\n \n :rtype: One of:\n- :class:`gcloud.bigquery.job.LoadTableFromStorageJob`,\n- :class:`gcloud.bigquery.job.CopyJob`,\n- :class:`gcloud.bigquery.job.ExtractTableToStorageJob`,\n- :class:`gcloud.bigquery.job.QueryJob`,\n- :class:`gcloud.bigquery.job.RunSyncQueryJob`\n+ :class:`google.cloud.bigquery.job.LoadTableFromStorageJob`,\n+ :class:`google.cloud.bigquery.job.CopyJob`,\n+ :class:`google.cloud.bigquery.job.ExtractTableToStorageJob`,\n+ :class:`google.cloud.bigquery.job.QueryJob`,\n+ :class:`google.cloud.bigquery.job.RunSyncQueryJob`\n :returns: the job instance, constructed via the resource\n \"\"\"\n config = resource['configuration']\n@@ -191,14 +191,14 @@ def load_table_from_storage(self, job_name, destination, *source_uris):\n :type job_name: str\n :param job_name: Name of the job.\n \n- :type destination: :class:`gcloud.bigquery.table.Table`\n+ :type destination: :class:`google.cloud.bigquery.table.Table`\n :param destination: Table into which data is to be loaded.\n \n :type source_uris: sequence of string\n :param source_uris: URIs of data files to be loaded; in format\n ``gs:///``.\n \n- :rtype: :class:`gcloud.bigquery.job.LoadTableFromStorageJob`\n+ :rtype: :class:`google.cloud.bigquery.job.LoadTableFromStorageJob`\n :returns: a new ``LoadTableFromStorageJob`` instance\n \"\"\"\n return LoadTableFromStorageJob(job_name, destination, source_uris,\n@@ -213,13 +213,13 @@ def copy_table(self, job_name, destination, *sources):\n :type job_name: str\n :param job_name: Name of the job.\n \n- :type destination: :class:`gcloud.bigquery.table.Table`\n+ :type destination: :class:`google.cloud.bigquery.table.Table`\n :param destination: Table into which data is to be copied.\n \n- :type sources: sequence of :class:`gcloud.bigquery.table.Table`\n+ :type sources: sequence of :class:`google.cloud.bigquery.table.Table`\n :param sources: tables to be copied.\n \n- :rtype: :class:`gcloud.bigquery.job.CopyJob`\n+ :rtype: :class:`google.cloud.bigquery.job.CopyJob`\n :returns: a new ``CopyJob`` instance\n \"\"\"\n return CopyJob(job_name, destination, sources, client=self)\n@@ -233,7 +233,7 @@ def extract_table_to_storage(self, job_name, source, *destination_uris):\n :type job_name: str\n :param job_name: Name of the job.\n \n- :type source: :class:`gcloud.bigquery.table.Table`\n+ :type source: :class:`google.cloud.bigquery.table.Table`\n :param source: table to be extracted.\n \n :type destination_uris: sequence of string\n@@ -241,7 +241,7 @@ def extract_table_to_storage(self, job_name, source, *destination_uris):\n table data is to be extracted; in format\n ``gs:///``.\n \n- :rtype: :class:`gcloud.bigquery.job.ExtractTableToStorageJob`\n+ :rtype: :class:`google.cloud.bigquery.job.ExtractTableToStorageJob`\n :returns: a new ``ExtractTableToStorageJob`` instance\n \"\"\"\n return ExtractTableToStorageJob(job_name, source, destination_uris,\n@@ -259,7 +259,7 @@ def run_async_query(self, job_name, query):\n :type query: str\n :param query: SQL query to be executed\n \n- :rtype: :class:`gcloud.bigquery.job.QueryJob`\n+ :rtype: :class:`google.cloud.bigquery.job.QueryJob`\n :returns: a new ``QueryJob`` instance\n \"\"\"\n return QueryJob(job_name, query, client=self)\n@@ -270,7 +270,7 @@ def run_sync_query(self, query):\n :type query: str\n :param query: SQL query to be executed\n \n- :rtype: :class:`gcloud.bigquery.query.QueryResults`\n+ :rtype: :class:`google.cloud.bigquery.query.QueryResults`\n :returns: a new ``QueryResults`` instance\n \"\"\"\n return QueryResults(query, client=self)\ndiff --git a/gcloud/bigquery/connection.py b/google/cloud/bigquery/connection.py\nsimilarity index 91%\nrename from gcloud/bigquery/connection.py\nrename to google/cloud/bigquery/connection.py\n--- a/gcloud/bigquery/connection.py\n+++ b/google/cloud/bigquery/connection.py\n@@ -12,9 +12,9 @@\n # See the License for the specific language governing permissions and\n # limitations under the License.\n \n-\"\"\"Create / interact with gcloud bigquery connections.\"\"\"\n+\"\"\"Create / interact with Google Cloud BigQuery connections.\"\"\"\n \n-from gcloud import connection as base_connection\n+from google.cloud import connection as base_connection\n \n \n class Connection(base_connection.JSONConnection):\ndiff --git a/gcloud/bigquery/dataset.py b/google/cloud/bigquery/dataset.py\nsimilarity index 93%\nrename from gcloud/bigquery/dataset.py\nrename to google/cloud/bigquery/dataset.py\n--- a/gcloud/bigquery/dataset.py\n+++ b/google/cloud/bigquery/dataset.py\n@@ -15,9 +15,9 @@\n \"\"\"Define API Datasets.\"\"\"\n import six\n \n-from gcloud._helpers import _datetime_from_microseconds\n-from gcloud.exceptions import NotFound\n-from gcloud.bigquery.table import Table\n+from google.cloud._helpers import _datetime_from_microseconds\n+from google.cloud.exceptions import NotFound\n+from google.cloud.bigquery.table import Table\n \n \n class AccessGrant(object):\n@@ -94,7 +94,7 @@ class Dataset(object):\n :type name: string\n :param name: the name of the dataset\n \n- :type client: :class:`gcloud.bigquery.client.Client`\n+ :type client: :class:`google.cloud.bigquery.client.Client`\n :param client: A client which holds credentials and project configuration\n for the dataset (which requires a project).\n \n@@ -298,11 +298,11 @@ def from_api_repr(cls, resource, client):\n :type resource: dict\n :param resource: dataset resource representation returned from the API\n \n- :type client: :class:`gcloud.bigquery.client.Client`\n+ :type client: :class:`google.cloud.bigquery.client.Client`\n :param client: Client which holds credentials and project\n configuration for the dataset.\n \n- :rtype: :class:`gcloud.bigquery.dataset.Dataset`\n+ :rtype: :class:`google.cloud.bigquery.dataset.Dataset`\n :returns: Dataset parsed from ``resource``.\n \"\"\"\n if ('datasetReference' not in resource or\n@@ -317,11 +317,12 @@ def from_api_repr(cls, resource, client):\n def _require_client(self, client):\n \"\"\"Check client or verify over-ride.\n \n- :type client: :class:`gcloud.bigquery.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.bigquery.client.Client` or\n+ ``NoneType``\n :param client: the client to use. If not passed, falls back to the\n ``client`` stored on the current dataset.\n \n- :rtype: :class:`gcloud.bigquery.client.Client`\n+ :rtype: :class:`google.cloud.bigquery.client.Client`\n :returns: The client passed in or the currently bound client.\n \"\"\"\n if client is None:\n@@ -413,7 +414,8 @@ def create(self, client=None):\n See:\n https://cloud.google.com/bigquery/docs/reference/v2/tables/insert\n \n- :type client: :class:`gcloud.bigquery.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.bigquery.client.Client` or\n+ ``NoneType``\n :param client: the client to use. If not passed, falls back to the\n ``client`` stored on the current dataset.\n \"\"\"\n@@ -429,7 +431,8 @@ def exists(self, client=None):\n See\n https://cloud.google.com/bigquery/docs/reference/v2/datasets/get\n \n- :type client: :class:`gcloud.bigquery.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.bigquery.client.Client` or\n+ ``NoneType``\n :param client: the client to use. If not passed, falls back to the\n ``client`` stored on the current dataset.\n \n@@ -452,7 +455,8 @@ def reload(self, client=None):\n See\n https://cloud.google.com/bigquery/docs/reference/v2/datasets/get\n \n- :type client: :class:`gcloud.bigquery.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.bigquery.client.Client` or\n+ ``NoneType``\n :param client: the client to use. If not passed, falls back to the\n ``client`` stored on the current dataset.\n \"\"\"\n@@ -468,7 +472,8 @@ def patch(self, client=None, **kw):\n See\n https://cloud.google.com/bigquery/docs/reference/v2/datasets/patch\n \n- :type client: :class:`gcloud.bigquery.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.bigquery.client.Client` or\n+ ``NoneType``\n :param client: the client to use. If not passed, falls back to the\n ``client`` stored on the current dataset.\n \n@@ -506,7 +511,8 @@ def update(self, client=None):\n See\n https://cloud.google.com/bigquery/docs/reference/v2/datasets/update\n \n- :type client: :class:`gcloud.bigquery.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.bigquery.client.Client` or\n+ ``NoneType``\n :param client: the client to use. If not passed, falls back to the\n ``client`` stored on the current dataset.\n \"\"\"\n@@ -521,7 +527,8 @@ def delete(self, client=None):\n See:\n https://cloud.google.com/bigquery/docs/reference/v2/tables/delete\n \n- :type client: :class:`gcloud.bigquery.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.bigquery.client.Client` or\n+ ``NoneType``\n :param client: the client to use. If not passed, falls back to the\n ``client`` stored on the current dataset.\n \"\"\"\n@@ -544,7 +551,7 @@ def list_tables(self, max_results=None, page_token=None):\n datasets.\n \n :rtype: tuple, (list, str)\n- :returns: list of :class:`gcloud.bigquery.table.Table`, plus a\n+ :returns: list of :class:`google.cloud.bigquery.table.Table`, plus a\n \"next page token\" string: if not ``None``, indicates that\n more tables can be retrieved with another call (pass that\n value as ``page_token``).\n@@ -571,10 +578,10 @@ def table(self, name, schema=()):\n :type name: string\n :param name: Name of the table.\n \n- :type schema: list of :class:`gcloud.bigquery.table.SchemaField`\n+ :type schema: list of :class:`google.cloud.bigquery.table.SchemaField`\n :param schema: The table's schema\n \n- :rtype: :class:`gcloud.bigquery.table.Table`\n+ :rtype: :class:`google.cloud.bigquery.table.Table`\n :returns: a new ``Table`` instance\n \"\"\"\n return Table(name, dataset=self, schema=schema)\ndiff --git a/gcloud/bigquery/job.py b/google/cloud/bigquery/job.py\nsimilarity index 93%\nrename from gcloud/bigquery/job.py\nrename to google/cloud/bigquery/job.py\n--- a/gcloud/bigquery/job.py\n+++ b/google/cloud/bigquery/job.py\n@@ -16,15 +16,15 @@\n \n import six\n \n-from gcloud.exceptions import NotFound\n-from gcloud._helpers import _datetime_from_microseconds\n-from gcloud.bigquery.dataset import Dataset\n-from gcloud.bigquery.schema import SchemaField\n-from gcloud.bigquery.table import Table\n-from gcloud.bigquery.table import _build_schema_resource\n-from gcloud.bigquery.table import _parse_schema_resource\n-from gcloud.bigquery._helpers import _EnumProperty\n-from gcloud.bigquery._helpers import _TypedProperty\n+from google.cloud.exceptions import NotFound\n+from google.cloud._helpers import _datetime_from_microseconds\n+from google.cloud.bigquery.dataset import Dataset\n+from google.cloud.bigquery.schema import SchemaField\n+from google.cloud.bigquery.table import Table\n+from google.cloud.bigquery.table import _build_schema_resource\n+from google.cloud.bigquery.table import _parse_schema_resource\n+from google.cloud.bigquery._helpers import _EnumProperty\n+from google.cloud.bigquery._helpers import _TypedProperty\n \n \n class UDFResource(object):\n@@ -66,7 +66,7 @@ def _build_udf_resources(resources):\n class UDFResourcesProperty(object):\n \"\"\"Custom property type for :class:`QueryJob`.\n \n- Also used by :class:`~gcloud.bigquery.query.Query`.\n+ Also used by :class:`~google.cloud.bigquery.query.Query`.\n \"\"\"\n def __get__(self, instance, owner):\n \"\"\"Descriptor protocol: accessor\"\"\"\n@@ -136,7 +136,7 @@ class WriteDisposition(_EnumProperty):\n class _BaseJob(object):\n \"\"\"Base class for jobs.\n \n- :type client: :class:`gcloud.bigquery.client.Client`\n+ :type client: :class:`google.cloud.bigquery.client.Client`\n :param client: A client which holds credentials and project configuration\n for the dataset (which requires a project).\n \"\"\"\n@@ -156,11 +156,12 @@ def project(self):\n def _require_client(self, client):\n \"\"\"Check client or verify over-ride.\n \n- :type client: :class:`gcloud.bigquery.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.bigquery.client.Client` or\n+ ``NoneType``\n :param client: the client to use. If not passed, falls back to the\n ``client`` stored on the current dataset.\n \n- :rtype: :class:`gcloud.bigquery.client.Client`\n+ :rtype: :class:`google.cloud.bigquery.client.Client`\n :returns: The client passed in or the currently bound client.\n \"\"\"\n if client is None:\n@@ -174,7 +175,7 @@ class _AsyncJob(_BaseJob):\n :type name: string\n :param name: the name of the job\n \n- :type client: :class:`gcloud.bigquery.client.Client`\n+ :type client: :class:`google.cloud.bigquery.client.Client`\n :param client: A client which holds credentials and project configuration\n for the dataset (which requires a project).\n \"\"\"\n@@ -354,7 +355,8 @@ def begin(self, client=None):\n See:\n https://cloud.google.com/bigquery/docs/reference/v2/jobs/insert\n \n- :type client: :class:`gcloud.bigquery.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.bigquery.client.Client` or\n+ ``NoneType``\n :param client: the client to use. If not passed, falls back to the\n ``client`` stored on the current dataset.\n \"\"\"\n@@ -370,7 +372,8 @@ def exists(self, client=None):\n See\n https://cloud.google.com/bigquery/docs/reference/v2/jobs/get\n \n- :type client: :class:`gcloud.bigquery.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.bigquery.client.Client` or\n+ ``NoneType``\n :param client: the client to use. If not passed, falls back to the\n ``client`` stored on the current dataset.\n \n@@ -393,7 +396,8 @@ def reload(self, client=None):\n See\n https://cloud.google.com/bigquery/docs/reference/v2/jobs/get\n \n- :type client: :class:`gcloud.bigquery.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.bigquery.client.Client` or\n+ ``NoneType``\n :param client: the client to use. If not passed, falls back to the\n ``client`` stored on the current dataset.\n \"\"\"\n@@ -409,7 +413,8 @@ def cancel(self, client=None):\n See\n https://cloud.google.com/bigquery/docs/reference/v2/jobs/cancel\n \n- :type client: :class:`gcloud.bigquery.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.bigquery.client.Client` or\n+ ``NoneType``\n :param client: the client to use. If not passed, falls back to the\n ``client`` stored on the current dataset.\n \"\"\"\n@@ -444,18 +449,18 @@ class LoadTableFromStorageJob(_AsyncJob):\n :type name: string\n :param name: the name of the job\n \n- :type destination: :class:`gcloud.bigquery.table.Table`\n+ :type destination: :class:`google.cloud.bigquery.table.Table`\n :param destination: Table into which data is to be loaded.\n \n :type source_uris: sequence of string\n :param source_uris: URIs of one or more data files to be loaded, in\n format ``gs:///``.\n \n- :type client: :class:`gcloud.bigquery.client.Client`\n+ :type client: :class:`google.cloud.bigquery.client.Client`\n :param client: A client which holds credentials and project configuration\n for the dataset (which requires a project).\n \n- :type schema: list of :class:`gcloud.bigquery.table.SchemaField`\n+ :type schema: list of :class:`google.cloud.bigquery.table.SchemaField`\n :param schema: The job's schema\n \"\"\"\n \n@@ -661,11 +666,11 @@ def from_api_repr(cls, resource, client):\n :type resource: dict\n :param resource: dataset job representation returned from the API\n \n- :type client: :class:`gcloud.bigquery.client.Client`\n+ :type client: :class:`google.cloud.bigquery.client.Client`\n :param client: Client which holds credentials and project\n configuration for the dataset.\n \n- :rtype: :class:`gcloud.bigquery.job.LoadTableFromStorageJob`\n+ :rtype: :class:`google.cloud.bigquery.job.LoadTableFromStorageJob`\n :returns: Job parsed from ``resource``.\n \"\"\"\n name, config = cls._get_resource_config(resource)\n@@ -693,13 +698,13 @@ class CopyJob(_AsyncJob):\n :type name: string\n :param name: the name of the job\n \n- :type destination: :class:`gcloud.bigquery.table.Table`\n+ :type destination: :class:`google.cloud.bigquery.table.Table`\n :param destination: Table into which data is to be loaded.\n \n- :type sources: list of :class:`gcloud.bigquery.table.Table`\n+ :type sources: list of :class:`google.cloud.bigquery.table.Table`\n :param sources: Table into which data is to be loaded.\n \n- :type client: :class:`gcloud.bigquery.client.Client`\n+ :type client: :class:`google.cloud.bigquery.client.Client`\n :param client: A client which holds credentials and project configuration\n for the dataset (which requires a project).\n \"\"\"\n@@ -771,11 +776,11 @@ def from_api_repr(cls, resource, client):\n :type resource: dict\n :param resource: dataset job representation returned from the API\n \n- :type client: :class:`gcloud.bigquery.client.Client`\n+ :type client: :class:`google.cloud.bigquery.client.Client`\n :param client: Client which holds credentials and project\n configuration for the dataset.\n \n- :rtype: :class:`gcloud.bigquery.job.CopyJob`\n+ :rtype: :class:`google.cloud.bigquery.job.CopyJob`\n :returns: Job parsed from ``resource``.\n \"\"\"\n name, config = cls._get_resource_config(resource)\n@@ -808,7 +813,7 @@ class ExtractTableToStorageJob(_AsyncJob):\n :type name: string\n :param name: the name of the job\n \n- :type source: :class:`gcloud.bigquery.table.Table`\n+ :type source: :class:`google.cloud.bigquery.table.Table`\n :param source: Table into which data is to be loaded.\n \n :type destination_uris: list of string\n@@ -816,7 +821,7 @@ class ExtractTableToStorageJob(_AsyncJob):\n extracted data will be written, in format\n ``gs:///``.\n \n- :type client: :class:`gcloud.bigquery.client.Client`\n+ :type client: :class:`google.cloud.bigquery.client.Client`\n :param client: A client which holds credentials and project configuration\n for the dataset (which requires a project).\n \"\"\"\n@@ -897,11 +902,11 @@ def from_api_repr(cls, resource, client):\n :type resource: dict\n :param resource: dataset job representation returned from the API\n \n- :type client: :class:`gcloud.bigquery.client.Client`\n+ :type client: :class:`google.cloud.bigquery.client.Client`\n :param client: Client which holds credentials and project\n configuration for the dataset.\n \n- :rtype: :class:`gcloud.bigquery.job.ExtractTableToStorageJob`\n+ :rtype: :class:`google.cloud.bigquery.job.ExtractTableToStorageJob`\n :returns: Job parsed from ``resource``.\n \"\"\"\n name, config = cls._get_resource_config(resource)\n@@ -939,13 +944,13 @@ class QueryJob(_AsyncJob):\n :type query: string\n :param query: SQL query string\n \n- :type client: :class:`gcloud.bigquery.client.Client`\n+ :type client: :class:`google.cloud.bigquery.client.Client`\n :param client: A client which holds credentials and project configuration\n for the dataset (which requires a project).\n \n :type udf_resources: tuple\n :param udf_resources: An iterable of\n- :class:`gcloud.bigquery.job.UDFResource`\n+ :class:`google.cloud.bigquery.job.UDFResource`\n (empty by default)\n \"\"\"\n _JOB_TYPE = 'query'\n@@ -1092,11 +1097,11 @@ def from_api_repr(cls, resource, client):\n :type resource: dict\n :param resource: dataset job representation returned from the API\n \n- :type client: :class:`gcloud.bigquery.client.Client`\n+ :type client: :class:`google.cloud.bigquery.client.Client`\n :param client: Client which holds credentials and project\n configuration for the dataset.\n \n- :rtype: :class:`gcloud.bigquery.job.RunAsyncQueryJob`\n+ :rtype: :class:`google.cloud.bigquery.job.RunAsyncQueryJob`\n :returns: Job parsed from ``resource``.\n \"\"\"\n name, config = cls._get_resource_config(resource)\ndiff --git a/gcloud/bigquery/query.py b/google/cloud/bigquery/query.py\nsimilarity index 92%\nrename from gcloud/bigquery/query.py\nrename to google/cloud/bigquery/query.py\n--- a/gcloud/bigquery/query.py\n+++ b/google/cloud/bigquery/query.py\n@@ -16,13 +16,13 @@\n \n import six\n \n-from gcloud.bigquery._helpers import _TypedProperty\n-from gcloud.bigquery._helpers import _rows_from_json\n-from gcloud.bigquery.dataset import Dataset\n-from gcloud.bigquery.job import QueryJob\n-from gcloud.bigquery.job import UDFResourcesProperty\n-from gcloud.bigquery.job import _build_udf_resources\n-from gcloud.bigquery.table import _parse_schema_resource\n+from google.cloud.bigquery._helpers import _TypedProperty\n+from google.cloud.bigquery._helpers import _rows_from_json\n+from google.cloud.bigquery.dataset import Dataset\n+from google.cloud.bigquery.job import QueryJob\n+from google.cloud.bigquery.job import UDFResourcesProperty\n+from google.cloud.bigquery.job import _build_udf_resources\n+from google.cloud.bigquery.table import _parse_schema_resource\n \n \n class _SyncQueryConfiguration(object):\n@@ -45,13 +45,13 @@ class QueryResults(object):\n :type query: string\n :param query: SQL query string\n \n- :type client: :class:`gcloud.bigquery.client.Client`\n+ :type client: :class:`google.cloud.bigquery.client.Client`\n :param client: A client which holds credentials and project configuration\n for the dataset (which requires a project).\n \n :type udf_resources: tuple\n :param udf_resources: An iterable of\n- :class:`gcloud.bigquery.job.UDFResource`\n+ :class:`google.cloud.bigquery.job.UDFResource`\n (empty by default)\n \"\"\"\n \n@@ -77,11 +77,12 @@ def project(self):\n def _require_client(self, client):\n \"\"\"Check client or verify over-ride.\n \n- :type client: :class:`gcloud.bigquery.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.bigquery.client.Client` or\n+ ``NoneType``\n :param client: the client to use. If not passed, falls back to the\n ``client`` stored on the current dataset.\n \n- :rtype: :class:`gcloud.bigquery.client.Client`\n+ :rtype: :class:`google.cloud.bigquery.client.Client`\n :returns: The client passed in or the currently bound client.\n \"\"\"\n if client is None:\n@@ -144,7 +145,7 @@ def name(self):\n def job(self):\n \"\"\"Job instance used to run the query.\n \n- :rtype: :class:`gcloud.bigquery.job.QueryJob`, or ``NoneType``\n+ :rtype: :class:`google.cloud.bigquery.job.QueryJob`, or ``NoneType``\n :returns: Job instance used to run the query (None until\n ``jobReference`` property is set by the server).\n \"\"\"\n@@ -301,7 +302,8 @@ def run(self, client=None):\n See:\n https://cloud.google.com/bigquery/docs/reference/v2/jobs/query\n \n- :type client: :class:`gcloud.bigquery.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.bigquery.client.Client` or\n+ ``NoneType``\n :param client: the client to use. If not passed, falls back to the\n ``client`` stored on the current dataset.\n \"\"\"\n@@ -331,7 +333,8 @@ def fetch_data(self, max_results=None, page_token=None, start_index=None,\n :param timeout_ms: timeout, in milliseconds, to wait for query to\n complete\n \n- :type client: :class:`gcloud.bigquery.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.bigquery.client.Client` or\n+ ``NoneType``\n :param client: the client to use. If not passed, falls back to the\n ``client`` stored on the current dataset.\n \ndiff --git a/gcloud/bigquery/schema.py b/google/cloud/bigquery/schema.py\nsimilarity index 100%\nrename from gcloud/bigquery/schema.py\nrename to google/cloud/bigquery/schema.py\ndiff --git a/gcloud/bigquery/table.py b/google/cloud/bigquery/table.py\nsimilarity index 92%\nrename from gcloud/bigquery/table.py\nrename to google/cloud/bigquery/table.py\n--- a/gcloud/bigquery/table.py\n+++ b/google/cloud/bigquery/table.py\n@@ -20,16 +20,16 @@\n \n import six\n \n-from gcloud._helpers import _datetime_from_microseconds\n-from gcloud._helpers import _microseconds_from_datetime\n-from gcloud._helpers import _millis_from_datetime\n-from gcloud.exceptions import NotFound\n-from gcloud.streaming.http_wrapper import Request\n-from gcloud.streaming.http_wrapper import make_api_request\n-from gcloud.streaming.transfer import RESUMABLE_UPLOAD\n-from gcloud.streaming.transfer import Upload\n-from gcloud.bigquery.schema import SchemaField\n-from gcloud.bigquery._helpers import _rows_from_json\n+from google.cloud._helpers import _datetime_from_microseconds\n+from google.cloud._helpers import _microseconds_from_datetime\n+from google.cloud._helpers import _millis_from_datetime\n+from google.cloud.exceptions import NotFound\n+from google.cloud.streaming.http_wrapper import Request\n+from google.cloud.streaming.http_wrapper import make_api_request\n+from google.cloud.streaming.transfer import RESUMABLE_UPLOAD\n+from google.cloud.streaming.transfer import Upload\n+from google.cloud.bigquery.schema import SchemaField\n+from google.cloud.bigquery._helpers import _rows_from_json\n \n \n _MARKER = object()\n@@ -44,7 +44,7 @@ class Table(object):\n :type name: str\n :param name: the name of the table\n \n- :type dataset: :class:`gcloud.bigquery.dataset.Dataset`\n+ :type dataset: :class:`google.cloud.bigquery.dataset.Dataset`\n :param dataset: The dataset which contains the table.\n \n :type schema: list of :class:`SchemaField`\n@@ -368,7 +368,8 @@ def view_query(self):\n def list_partitions(self, client=None):\n \"\"\"List the partitions in a table.\n \n- :type client: :class:`gcloud.bigquery.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.bigquery.client.Client` or\n+ ``NoneType``\n :param client: the client to use. If not passed, falls back to the\n ``client`` stored on the current dataset.\n \n@@ -388,10 +389,10 @@ def from_api_repr(cls, resource, dataset):\n :type resource: dict\n :param resource: table resource representation returned from the API\n \n- :type dataset: :class:`gcloud.bigquery.dataset.Dataset`\n+ :type dataset: :class:`google.cloud.bigquery.dataset.Dataset`\n :param dataset: The dataset containing the table.\n \n- :rtype: :class:`gcloud.bigquery.table.Table`\n+ :rtype: :class:`google.cloud.bigquery.table.Table`\n :returns: Table parsed from ``resource``.\n \"\"\"\n if ('tableReference' not in resource or\n@@ -406,11 +407,12 @@ def from_api_repr(cls, resource, dataset):\n def _require_client(self, client):\n \"\"\"Check client or verify over-ride.\n \n- :type client: :class:`gcloud.bigquery.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.bigquery.client.Client` or\n+ ``NoneType``\n :param client: the client to use. If not passed, falls back to the\n ``client`` stored on the current dataset.\n \n- :rtype: :class:`gcloud.bigquery.client.Client`\n+ :rtype: :class:`google.cloud.bigquery.client.Client`\n :returns: The client passed in or the currently bound client.\n \"\"\"\n if client is None:\n@@ -477,7 +479,8 @@ def create(self, client=None):\n See:\n https://cloud.google.com/bigquery/docs/reference/v2/tables/insert\n \n- :type client: :class:`gcloud.bigquery.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.bigquery.client.Client` or\n+ ``NoneType``\n :param client: the client to use. If not passed, falls back to the\n ``client`` stored on the current dataset.\n \"\"\"\n@@ -494,7 +497,8 @@ def exists(self, client=None):\n See\n https://cloud.google.com/bigquery/docs/reference/v2/tables/get\n \n- :type client: :class:`gcloud.bigquery.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.bigquery.client.Client` or\n+ ``NoneType``\n :param client: the client to use. If not passed, falls back to the\n ``client`` stored on the current dataset.\n \n@@ -517,7 +521,8 @@ def reload(self, client=None):\n See\n https://cloud.google.com/bigquery/docs/reference/v2/tables/get\n \n- :type client: :class:`gcloud.bigquery.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.bigquery.client.Client` or\n+ ``NoneType``\n :param client: the client to use. If not passed, falls back to the\n ``client`` stored on the current dataset.\n \"\"\"\n@@ -540,7 +545,8 @@ def patch(self,\n See\n https://cloud.google.com/bigquery/docs/reference/v2/tables/patch\n \n- :type client: :class:`gcloud.bigquery.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.bigquery.client.Client` or\n+ ``NoneType``\n :param client: the client to use. If not passed, falls back to the\n ``client`` stored on the current dataset.\n \n@@ -606,7 +612,8 @@ def update(self, client=None):\n See\n https://cloud.google.com/bigquery/docs/reference/v2/tables/update\n \n- :type client: :class:`gcloud.bigquery.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.bigquery.client.Client` or\n+ ``NoneType``\n :param client: the client to use. If not passed, falls back to the\n ``client`` stored on the current dataset.\n \"\"\"\n@@ -621,7 +628,8 @@ def delete(self, client=None):\n See:\n https://cloud.google.com/bigquery/docs/reference/v2/tables/delete\n \n- :type client: :class:`gcloud.bigquery.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.bigquery.client.Client` or\n+ ``NoneType``\n :param client: the client to use. If not passed, falls back to the\n ``client`` stored on the current dataset.\n \"\"\"\n@@ -648,7 +656,8 @@ def fetch_data(self, max_results=None, page_token=None, client=None):\n :type page_token: str or ``NoneType``\n :param page_token: token representing a cursor into the table's rows.\n \n- :type client: :class:`gcloud.bigquery.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.bigquery.client.Client` or\n+ ``NoneType``\n :param client: the client to use. If not passed, falls back to the\n ``client`` stored on the current dataset.\n \n@@ -714,7 +723,8 @@ def insert_data(self,\n schema of the template table. See:\n https://cloud.google.com/bigquery/streaming-data-into-bigquery#template-tables\n \n- :type client: :class:`gcloud.bigquery.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.bigquery.client.Client` or\n+ ``NoneType``\n :param client: the client to use. If not passed, falls back to the\n ``client`` stored on the current dataset.\n \n@@ -796,7 +806,7 @@ def upload_from_file(self,\n :type source_format: str\n :param source_format: one of 'CSV' or 'NEWLINE_DELIMITED_JSON'.\n job configuration option; see\n- :meth:`gcloud.bigquery.job.LoadJob`\n+ :meth:`google.cloud.bigquery.job.LoadJob`\n \n :type rewind: boolean\n :param rewind: If True, seek to the beginning of the file handle before\n@@ -813,49 +823,50 @@ def upload_from_file(self,\n \n :type allow_jagged_rows: boolean\n :param allow_jagged_rows: job configuration option; see\n- :meth:`gcloud.bigquery.job.LoadJob`.\n+ :meth:`google.cloud.bigquery.job.LoadJob`.\n \n :type allow_quoted_newlines: boolean\n :param allow_quoted_newlines: job configuration option; see\n- :meth:`gcloud.bigquery.job.LoadJob`.\n+ :meth:`google.cloud.bigquery.job.LoadJob`.\n \n :type create_disposition: str\n :param create_disposition: job configuration option; see\n- :meth:`gcloud.bigquery.job.LoadJob`.\n+ :meth:`google.cloud.bigquery.job.LoadJob`.\n \n :type encoding: str\n :param encoding: job configuration option; see\n- :meth:`gcloud.bigquery.job.LoadJob`.\n+ :meth:`google.cloud.bigquery.job.LoadJob`.\n \n :type field_delimiter: str\n :param field_delimiter: job configuration option; see\n- :meth:`gcloud.bigquery.job.LoadJob`.\n+ :meth:`google.cloud.bigquery.job.LoadJob`.\n \n :type ignore_unknown_values: boolean\n :param ignore_unknown_values: job configuration option; see\n- :meth:`gcloud.bigquery.job.LoadJob`.\n+ :meth:`google.cloud.bigquery.job.LoadJob`.\n \n :type max_bad_records: integer\n :param max_bad_records: job configuration option; see\n- :meth:`gcloud.bigquery.job.LoadJob`.\n+ :meth:`google.cloud.bigquery.job.LoadJob`.\n \n :type quote_character: str\n :param quote_character: job configuration option; see\n- :meth:`gcloud.bigquery.job.LoadJob`.\n+ :meth:`google.cloud.bigquery.job.LoadJob`.\n \n :type skip_leading_rows: integer\n :param skip_leading_rows: job configuration option; see\n- :meth:`gcloud.bigquery.job.LoadJob`.\n+ :meth:`google.cloud.bigquery.job.LoadJob`.\n \n :type write_disposition: str\n :param write_disposition: job configuration option; see\n- :meth:`gcloud.bigquery.job.LoadJob`.\n+ :meth:`google.cloud.bigquery.job.LoadJob`.\n \n- :type client: :class:`gcloud.storage.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.storage.client.Client` or\n+ ``NoneType``\n :param client: Optional. The client to use. If not passed, falls back\n to the ``client`` stored on the current dataset.\n \n- :rtype: :class:`gcloud.bigquery.jobs.LoadTableFromStorageJob`\n+ :rtype: :class:`google.cloud.bigquery.jobs.LoadTableFromStorageJob`\n :returns: the job instance used to load the data (e.g., for\n querying status).\n :raises: :class:`ValueError` if ``size`` is not passed in and can not\ndiff --git a/gcloud/bigtable/__init__.py b/google/cloud/bigtable/__init__.py\nsimilarity index 96%\nrename from gcloud/bigtable/__init__.py\nrename to google/cloud/bigtable/__init__.py\n--- a/gcloud/bigtable/__init__.py\n+++ b/google/cloud/bigtable/__init__.py\n@@ -15,7 +15,7 @@\n \"\"\"Google Cloud Bigtable API package.\"\"\"\n \n \n-from gcloud.bigtable.client import Client\n+from google.cloud.bigtable.client import Client\n \n \n _ERR_MSG = \"\"\"\\\ndiff --git a/gcloud/bigtable/_generated/__init__.py b/google/cloud/bigtable/_generated/__init__.py\nsimilarity index 100%\nrename from gcloud/bigtable/_generated/__init__.py\nrename to google/cloud/bigtable/_generated/__init__.py\ndiff --git a/gcloud/bigtable/_generated/bigtable_instance_admin_pb2.py b/google/cloud/bigtable/_generated/bigtable_instance_admin_pb2.py\nsimilarity index 99%\nrename from gcloud/bigtable/_generated/bigtable_instance_admin_pb2.py\nrename to google/cloud/bigtable/_generated/bigtable_instance_admin_pb2.py\n--- a/gcloud/bigtable/_generated/bigtable_instance_admin_pb2.py\n+++ b/google/cloud/bigtable/_generated/bigtable_instance_admin_pb2.py\n@@ -14,7 +14,7 @@\n \n \n from google.api import annotations_pb2 as google_dot_api_dot_annotations__pb2\n-from gcloud.bigtable._generated import instance_pb2 as google_dot_bigtable_dot_admin_dot_v2_dot_instance__pb2\n+from google.cloud.bigtable._generated import instance_pb2 as google_dot_bigtable_dot_admin_dot_v2_dot_instance__pb2\n from google.longrunning import operations_pb2 as google_dot_longrunning_dot_operations__pb2\n from google.protobuf import empty_pb2 as google_dot_protobuf_dot_empty__pb2\n from google.protobuf import timestamp_pb2 as google_dot_protobuf_dot_timestamp__pb2\ndiff --git a/gcloud/bigtable/_generated/bigtable_pb2.py b/google/cloud/bigtable/_generated/bigtable_pb2.py\nsimilarity index 99%\nrename from gcloud/bigtable/_generated/bigtable_pb2.py\nrename to google/cloud/bigtable/_generated/bigtable_pb2.py\n--- a/gcloud/bigtable/_generated/bigtable_pb2.py\n+++ b/google/cloud/bigtable/_generated/bigtable_pb2.py\n@@ -14,7 +14,7 @@\n \n \n from google.api import annotations_pb2 as google_dot_api_dot_annotations__pb2\n-from gcloud.bigtable._generated import data_pb2 as google_dot_bigtable_dot_v2_dot_data__pb2\n+from google.cloud.bigtable._generated import data_pb2 as google_dot_bigtable_dot_v2_dot_data__pb2\n from google.protobuf import wrappers_pb2 as google_dot_protobuf_dot_wrappers__pb2\n from google.rpc import status_pb2 as google_dot_rpc_dot_status__pb2\n \ndiff --git a/gcloud/bigtable/_generated/bigtable_table_admin_pb2.py b/google/cloud/bigtable/_generated/bigtable_table_admin_pb2.py\nsimilarity index 99%\nrename from gcloud/bigtable/_generated/bigtable_table_admin_pb2.py\nrename to google/cloud/bigtable/_generated/bigtable_table_admin_pb2.py\n--- a/gcloud/bigtable/_generated/bigtable_table_admin_pb2.py\n+++ b/google/cloud/bigtable/_generated/bigtable_table_admin_pb2.py\n@@ -14,7 +14,7 @@\n \n \n from google.api import annotations_pb2 as google_dot_api_dot_annotations__pb2\n-from gcloud.bigtable._generated import table_pb2 as google_dot_bigtable_dot_admin_dot_v2_dot_table__pb2\n+from google.cloud.bigtable._generated import table_pb2 as google_dot_bigtable_dot_admin_dot_v2_dot_table__pb2\n from google.protobuf import empty_pb2 as google_dot_protobuf_dot_empty__pb2\n \n \ndiff --git a/gcloud/bigtable/_generated/common_pb2.py b/google/cloud/bigtable/_generated/common_pb2.py\nsimilarity index 100%\nrename from gcloud/bigtable/_generated/common_pb2.py\nrename to google/cloud/bigtable/_generated/common_pb2.py\ndiff --git a/gcloud/bigtable/_generated/data_pb2.py b/google/cloud/bigtable/_generated/data_pb2.py\nsimilarity index 100%\nrename from gcloud/bigtable/_generated/data_pb2.py\nrename to google/cloud/bigtable/_generated/data_pb2.py\ndiff --git a/gcloud/bigtable/_generated/instance_pb2.py b/google/cloud/bigtable/_generated/instance_pb2.py\nsimilarity index 98%\nrename from gcloud/bigtable/_generated/instance_pb2.py\nrename to google/cloud/bigtable/_generated/instance_pb2.py\n--- a/gcloud/bigtable/_generated/instance_pb2.py\n+++ b/google/cloud/bigtable/_generated/instance_pb2.py\n@@ -14,7 +14,7 @@\n \n \n from google.api import annotations_pb2 as google_dot_api_dot_annotations__pb2\n-from gcloud.bigtable._generated import common_pb2 as google_dot_bigtable_dot_admin_dot_v2_dot_common__pb2\n+from google.cloud.bigtable._generated import common_pb2 as google_dot_bigtable_dot_admin_dot_v2_dot_common__pb2\n \n \n DESCRIPTOR = _descriptor.FileDescriptor(\ndiff --git a/gcloud/bigtable/_generated/operations_grpc_pb2.py b/google/cloud/bigtable/_generated/operations_grpc_pb2.py\nsimilarity index 100%\nrename from gcloud/bigtable/_generated/operations_grpc_pb2.py\nrename to google/cloud/bigtable/_generated/operations_grpc_pb2.py\ndiff --git a/gcloud/bigtable/_generated/table_pb2.py b/google/cloud/bigtable/_generated/table_pb2.py\nsimilarity index 100%\nrename from gcloud/bigtable/_generated/table_pb2.py\nrename to google/cloud/bigtable/_generated/table_pb2.py\ndiff --git a/gcloud/bigtable/client.py b/google/cloud/bigtable/client.py\nsimilarity index 92%\nrename from gcloud/bigtable/client.py\nrename to google/cloud/bigtable/client.py\n--- a/gcloud/bigtable/client.py\n+++ b/google/cloud/bigtable/client.py\n@@ -19,27 +19,27 @@\n In the hierarchy of API concepts\n \n * a :class:`Client` owns an :class:`.Instance`\n-* a :class:`.Instance` owns a :class:`Table `\n-* a :class:`Table ` owns a\n- :class:`ColumnFamily <.column_family.ColumnFamily>`\n-* a :class:`Table ` owns a :class:`Row <.row.Row>`\n+* a :class:`.Instance` owns a :class:`~google.cloud.bigtable.table.Table`\n+* a :class:`~google.cloud.bigtable.table.Table` owns a\n+ :class:`~.column_family.ColumnFamily`\n+* a :class:`~google.cloud.bigtable.table.Table` owns a :class:`~.row.Row`\n (and all the cells in the row)\n \"\"\"\n \n \n from pkg_resources import get_distribution\n \n-from gcloud._helpers import make_stub\n-from gcloud.bigtable._generated import bigtable_instance_admin_pb2\n-from gcloud.bigtable._generated import bigtable_pb2\n-from gcloud.bigtable._generated import bigtable_table_admin_pb2\n-from gcloud.bigtable._generated import operations_grpc_pb2\n-from gcloud.bigtable.cluster import DEFAULT_SERVE_NODES\n-from gcloud.bigtable.instance import Instance\n-from gcloud.bigtable.instance import _EXISTING_INSTANCE_LOCATION_ID\n-from gcloud.client import _ClientFactoryMixin\n-from gcloud.client import _ClientProjectMixin\n-from gcloud.credentials import get_credentials\n+from google.cloud._helpers import make_stub\n+from google.cloud.bigtable._generated import bigtable_instance_admin_pb2\n+from google.cloud.bigtable._generated import bigtable_pb2\n+from google.cloud.bigtable._generated import bigtable_table_admin_pb2\n+from google.cloud.bigtable._generated import operations_grpc_pb2\n+from google.cloud.bigtable.cluster import DEFAULT_SERVE_NODES\n+from google.cloud.bigtable.instance import Instance\n+from google.cloud.bigtable.instance import _EXISTING_INSTANCE_LOCATION_ID\n+from google.cloud.client import _ClientFactoryMixin\n+from google.cloud.client import _ClientProjectMixin\n+from google.cloud.credentials import get_credentials\n \n \n TABLE_ADMIN_HOST = 'bigtableadmin.googleapis.com'\n@@ -67,8 +67,8 @@\n READ_ONLY_SCOPE = 'https://www.googleapis.com/auth/bigtable.data.readonly'\n \"\"\"Scope for reading table data.\"\"\"\n \n-DEFAULT_USER_AGENT = 'gcloud-python/{0}'.format(\n- get_distribution('gcloud').version)\n+DEFAULT_USER_AGENT = 'google-cloud-python/{0}'.format(\n+ get_distribution('google-cloud').version)\n \"\"\"The default user agent for API requests.\"\"\"\n \n \ndiff --git a/gcloud/bigtable/cluster.py b/google/cloud/bigtable/cluster.py\nsimilarity index 97%\nrename from gcloud/bigtable/cluster.py\nrename to google/cloud/bigtable/cluster.py\n--- a/gcloud/bigtable/cluster.py\n+++ b/google/cloud/bigtable/cluster.py\n@@ -17,13 +17,13 @@\n \n import re\n \n-from gcloud.bigtable._generated import (\n+from google.cloud.bigtable._generated import (\n instance_pb2 as data_v2_pb2)\n-from gcloud.bigtable._generated import (\n+from google.cloud.bigtable._generated import (\n bigtable_instance_admin_pb2 as messages_v2_pb2)\n-from gcloud.operation import Operation\n-from gcloud.operation import _compute_type_url\n-from gcloud.operation import _register_type_url\n+from google.cloud.operation import Operation\n+from google.cloud.operation import _compute_type_url\n+from google.cloud.operation import _register_type_url\n \n \n _CLUSTER_NAME_RE = re.compile(r'^projects/(?P[^/]+)/'\ndiff --git a/gcloud/bigtable/column_family.py b/google/cloud/bigtable/column_family.py\nsimilarity index 98%\nrename from gcloud/bigtable/column_family.py\nrename to google/cloud/bigtable/column_family.py\n--- a/gcloud/bigtable/column_family.py\n+++ b/google/cloud/bigtable/column_family.py\n@@ -19,9 +19,9 @@\n \n from google.protobuf import duration_pb2\n \n-from gcloud.bigtable._generated import (\n+from google.cloud.bigtable._generated import (\n table_pb2 as table_v2_pb2)\n-from gcloud.bigtable._generated import (\n+from google.cloud.bigtable._generated import (\n bigtable_table_admin_pb2 as table_admin_v2_pb2)\n \n \n@@ -206,7 +206,7 @@ class ColumnFamily(object):\n :param column_family_id: The ID of the column family. Must be of the\n form ``[_a-zA-Z0-9][-_.a-zA-Z0-9]*``.\n \n- :type table: :class:`Table `\n+ :type table: :class:`Table `\n :param table: The table that owns the column family.\n \n :type gc_rule: :class:`GarbageCollectionRule`\ndiff --git a/gcloud/bigtable/instance.py b/google/cloud/bigtable/instance.py\nsimilarity index 94%\nrename from gcloud/bigtable/instance.py\nrename to google/cloud/bigtable/instance.py\n--- a/gcloud/bigtable/instance.py\n+++ b/google/cloud/bigtable/instance.py\n@@ -17,18 +17,18 @@\n \n import re\n \n-from gcloud.bigtable._generated import (\n+from google.cloud.bigtable._generated import (\n instance_pb2 as data_v2_pb2)\n-from gcloud.bigtable._generated import (\n+from google.cloud.bigtable._generated import (\n bigtable_instance_admin_pb2 as messages_v2_pb2)\n-from gcloud.bigtable._generated import (\n+from google.cloud.bigtable._generated import (\n bigtable_table_admin_pb2 as table_messages_v2_pb2)\n-from gcloud.bigtable.cluster import Cluster\n-from gcloud.bigtable.cluster import DEFAULT_SERVE_NODES\n-from gcloud.bigtable.table import Table\n-from gcloud.operation import Operation\n-from gcloud.operation import _compute_type_url\n-from gcloud.operation import _register_type_url\n+from google.cloud.bigtable.cluster import Cluster\n+from google.cloud.bigtable.cluster import DEFAULT_SERVE_NODES\n+from google.cloud.bigtable.table import Table\n+from google.cloud.operation import Operation\n+from google.cloud.operation import _compute_type_url\n+from google.cloud.operation import _register_type_url\n \n \n _EXISTING_INSTANCE_LOCATION_ID = 'see-existing-cluster'\n@@ -85,7 +85,7 @@ class Instance(object):\n :type instance_id: str\n :param instance_id: The ID of the instance.\n \n- :type client: :class:`Client `\n+ :type client: :class:`Client `\n :param client: The client that owns the instance. Provides\n authorization and a project ID.\n \n@@ -131,7 +131,7 @@ def from_pb(cls, instance_pb, client):\n :type instance_pb: :class:`instance_pb2.Instance`\n :param instance_pb: A instance protobuf object.\n \n- :type client: :class:`Client `\n+ :type client: :class:`Client `\n :param client: The client that owns the instance.\n \n :rtype: :class:`Instance`\n@@ -327,7 +327,7 @@ def table(self, table_id):\n :type table_id: str\n :param table_id: The ID of the table.\n \n- :rtype: :class:`Table `\n+ :rtype: :class:`Table `\n :returns: The table owned by this instance.\n \"\"\"\n return Table(table_id, self)\n@@ -335,7 +335,7 @@ def table(self, table_id):\n def list_tables(self):\n \"\"\"List the tables in this instance.\n \n- :rtype: list of :class:`Table `\n+ :rtype: list of :class:`Table `\n :returns: The list of tables owned by the instance.\n :raises: :class:`ValueError ` if one of the\n returned tables has a name that is not of the expected format.\ndiff --git a/gcloud/bigtable/row.py b/google/cloud/bigtable/row.py\nsimilarity index 98%\nrename from gcloud/bigtable/row.py\nrename to google/cloud/bigtable/row.py\n--- a/gcloud/bigtable/row.py\n+++ b/google/cloud/bigtable/row.py\n@@ -19,12 +19,12 @@\n \n import six\n \n-from gcloud._helpers import _datetime_from_microseconds\n-from gcloud._helpers import _microseconds_from_datetime\n-from gcloud._helpers import _to_bytes\n-from gcloud.bigtable._generated import (\n+from google.cloud._helpers import _datetime_from_microseconds\n+from google.cloud._helpers import _microseconds_from_datetime\n+from google.cloud._helpers import _to_bytes\n+from google.cloud.bigtable._generated import (\n data_pb2 as data_v2_pb2)\n-from gcloud.bigtable._generated import (\n+from google.cloud.bigtable._generated import (\n bigtable_pb2 as messages_v2_pb2)\n \n \n@@ -47,7 +47,7 @@ class Row(object):\n :type row_key: bytes\n :param row_key: The key for the current row.\n \n- :type table: :class:`Table `\n+ :type table: :class:`Table `\n :param table: The table that owns the row.\n \"\"\"\n \n@@ -69,7 +69,7 @@ class _SetDeleteRow(Row):\n :type row_key: bytes\n :param row_key: The key for the current row.\n \n- :type table: :class:`Table `\n+ :type table: :class:`Table `\n :param table: The table that owns the row.\n \"\"\"\n \n@@ -244,7 +244,7 @@ class DirectRow(_SetDeleteRow):\n :type row_key: bytes\n :param row_key: The key for the current row.\n \n- :type table: :class:`Table `\n+ :type table: :class:`Table `\n :param table: The table that owns the row.\n \"\"\"\n \n@@ -434,7 +434,7 @@ class ConditionalRow(_SetDeleteRow):\n :type row_key: bytes\n :param row_key: The key for the current row.\n \n- :type table: :class:`Table `\n+ :type table: :class:`Table `\n :param table: The table that owns the row.\n \n :type filter_: :class:`.RowFilter`\n@@ -662,7 +662,7 @@ class AppendRow(Row):\n :type row_key: bytes\n :param row_key: The key for the current row.\n \n- :type table: :class:`Table `\n+ :type table: :class:`Table `\n :param table: The table that owns the row.\n \"\"\"\n \ndiff --git a/gcloud/bigtable/row_data.py b/google/cloud/bigtable/row_data.py\nsimilarity index 99%\nrename from gcloud/bigtable/row_data.py\nrename to google/cloud/bigtable/row_data.py\n--- a/gcloud/bigtable/row_data.py\n+++ b/google/cloud/bigtable/row_data.py\n@@ -18,8 +18,8 @@\n import copy\n import six\n \n-from gcloud._helpers import _datetime_from_microseconds\n-from gcloud._helpers import _to_bytes\n+from google.cloud._helpers import _datetime_from_microseconds\n+from google.cloud._helpers import _to_bytes\n \n \n class Cell(object):\ndiff --git a/gcloud/bigtable/row_filters.py b/google/cloud/bigtable/row_filters.py\nsimilarity index 99%\nrename from gcloud/bigtable/row_filters.py\nrename to google/cloud/bigtable/row_filters.py\n--- a/gcloud/bigtable/row_filters.py\n+++ b/google/cloud/bigtable/row_filters.py\n@@ -15,9 +15,9 @@\n \"\"\"Filters for Google Cloud Bigtable Row classes.\"\"\"\n \n \n-from gcloud._helpers import _microseconds_from_datetime\n-from gcloud._helpers import _to_bytes\n-from gcloud.bigtable._generated import (\n+from google.cloud._helpers import _microseconds_from_datetime\n+from google.cloud._helpers import _to_bytes\n+from google.cloud.bigtable._generated import (\n data_pb2 as data_v2_pb2)\n \n \ndiff --git a/gcloud/bigtable/table.py b/google/cloud/bigtable/table.py\nsimilarity index 96%\nrename from gcloud/bigtable/table.py\nrename to google/cloud/bigtable/table.py\n--- a/gcloud/bigtable/table.py\n+++ b/google/cloud/bigtable/table.py\n@@ -14,19 +14,19 @@\n \n \"\"\"User friendly container for Google Cloud Bigtable Table.\"\"\"\n \n-from gcloud._helpers import _to_bytes\n-from gcloud.bigtable._generated import (\n+from google.cloud._helpers import _to_bytes\n+from google.cloud.bigtable._generated import (\n bigtable_pb2 as data_messages_v2_pb2)\n-from gcloud.bigtable._generated import (\n+from google.cloud.bigtable._generated import (\n bigtable_table_admin_pb2 as table_admin_messages_v2_pb2)\n-from gcloud.bigtable._generated import (\n+from google.cloud.bigtable._generated import (\n table_pb2 as table_v2_pb2)\n-from gcloud.bigtable.column_family import _gc_rule_from_pb\n-from gcloud.bigtable.column_family import ColumnFamily\n-from gcloud.bigtable.row import AppendRow\n-from gcloud.bigtable.row import ConditionalRow\n-from gcloud.bigtable.row import DirectRow\n-from gcloud.bigtable.row_data import PartialRowsData\n+from google.cloud.bigtable.column_family import _gc_rule_from_pb\n+from google.cloud.bigtable.column_family import ColumnFamily\n+from google.cloud.bigtable.row import AppendRow\n+from google.cloud.bigtable.row import ConditionalRow\n+from google.cloud.bigtable.row import DirectRow\n+from google.cloud.bigtable.row_data import PartialRowsData\n \n \n class Table(object):\ndiff --git a/gcloud/client.py b/google/cloud/client.py\nsimilarity index 96%\nrename from gcloud/client.py\nrename to google/cloud/client.py\n--- a/gcloud/client.py\n+++ b/google/cloud/client.py\n@@ -17,9 +17,9 @@\n from oauth2client.service_account import ServiceAccountCredentials\n import six\n \n-from gcloud._helpers import _determine_default_project\n-from gcloud.connection import Connection\n-from gcloud.credentials import get_credentials\n+from google.cloud._helpers import _determine_default_project\n+from google.cloud.connection import Connection\n+from google.cloud.credentials import get_credentials\n \n \n class _ClientFactoryMixin(object):\n@@ -48,7 +48,7 @@ def from_service_account_json(cls, json_credentials_path, *args, **kwargs):\n :type kwargs: dict\n :param kwargs: Remaining keyword arguments to pass to constructor.\n \n- :rtype: :class:`gcloud.pubsub.client.Client`\n+ :rtype: :class:`google.cloud.pubsub.client.Client`\n :returns: The client created with the retrieved JSON credentials.\n :raises: :class:`TypeError` if there is a conflict with the kwargs\n and the credentials created by the factory.\n@@ -83,7 +83,7 @@ def from_service_account_p12(cls, client_email, private_key_path,\n :type kwargs: dict\n :param kwargs: Remaining keyword arguments to pass to constructor.\n \n- :rtype: :class:`gcloud.client.Client`\n+ :rtype: :class:`google.cloud.client.Client`\n :returns: The client created with the retrieved P12 credentials.\n :raises: :class:`TypeError` if there is a conflict with the kwargs\n and the credentials created by the factory.\ndiff --git a/gcloud/connection.py b/google/cloud/connection.py\nsimilarity index 98%\nrename from gcloud/connection.py\nrename to google/cloud/connection.py\n--- a/gcloud/connection.py\n+++ b/google/cloud/connection.py\n@@ -21,7 +21,7 @@\n \n import httplib2\n \n-from gcloud.exceptions import make_exception\n+from google.cloud.exceptions import make_exception\n \n \n API_BASE_URL = 'https://www.googleapis.com'\n@@ -63,8 +63,9 @@ class Connection(object):\n :param http: An optional HTTP object to make requests.\n \"\"\"\n \n- USER_AGENT = \"gcloud-python/{0}\".format(get_distribution('gcloud').version)\n- \"\"\"The user agent for gcloud-python requests.\"\"\"\n+ USER_AGENT = \"google-cloud-python/{0}\".format(\n+ get_distribution('google-cloud').version)\n+ \"\"\"The user agent for google-cloud-python requests.\"\"\"\n \n SCOPE = None\n \"\"\"The scopes required for authenticating with a service.\n@@ -310,7 +311,7 @@ def api_request(self, method, path, query_params=None,\n you shouldn't provide this and instead use\n the default for the library. Default is the\n latest API version supported by\n- gcloud-python.\n+ google-cloud-python.\n \n :type expect_json: bool\n :param expect_json: If True, this method will try to parse the\ndiff --git a/gcloud/credentials.py b/google/cloud/credentials.py\nsimilarity index 94%\nrename from gcloud/credentials.py\nrename to google/cloud/credentials.py\n--- a/gcloud/credentials.py\n+++ b/google/cloud/credentials.py\n@@ -21,9 +21,9 @@\n \n from oauth2client import client\n \n-from gcloud._helpers import UTC\n-from gcloud._helpers import _NOW\n-from gcloud._helpers import _microseconds_from_datetime\n+from google.cloud._helpers import UTC\n+from google.cloud._helpers import _NOW\n+from google.cloud._helpers import _microseconds_from_datetime\n \n \n def get_credentials():\n@@ -102,12 +102,12 @@ def _get_signed_query_params(credentials, expiration, string_to_sign):\n signed payload.\n \"\"\"\n if not hasattr(credentials, 'sign_blob'):\n+ auth_uri = ('http://gcloud-python.readthedocs.io/en/latest/'\n+ 'gcloud-auth.html#setting-up-a-service-account')\n raise AttributeError('you need a private key to sign credentials.'\n 'the credentials you are currently using %s '\n- 'just contains a token. see https://googlecloud'\n- 'platform.github.io/gcloud-python/stable/gcloud-'\n- 'auth.html#setting-up-a-service-account for more '\n- 'details.' % type(credentials))\n+ 'just contains a token. see %s for more '\n+ 'details.' % (type(credentials), auth_uri))\n \n _, signature_bytes = credentials.sign_blob(string_to_sign)\n signature = base64.b64encode(signature_bytes)\n@@ -172,7 +172,7 @@ def generate_signed_url(credentials, resource, expiration,\n See headers `reference`_ for more details on optional arguments.\n \n .. _Issue 922: https://github.com/GoogleCloudPlatform/\\\n- gcloud-python/issues/922\n+ google-cloud-python/issues/922\n .. _reference: https://cloud.google.com/storage/docs/reference-headers\n \n :type credentials: :class:`oauth2client.appengine.AppAssertionCredentials`\ndiff --git a/gcloud/datastore/__init__.py b/google/cloud/datastore/__init__.py\nsimilarity index 70%\nrename from gcloud/datastore/__init__.py\nrename to google/cloud/datastore/__init__.py\n--- a/gcloud/datastore/__init__.py\n+++ b/google/cloud/datastore/__init__.py\n@@ -16,7 +16,7 @@\n \n You'll typically use these to get started with the API::\n \n- >>> from gcloud import datastore\n+ >>> from google.cloud import datastore\n >>>\n >>> client = datastore.Client()\n >>> key = client.key('EntityKind', 1234)\n@@ -25,38 +25,39 @@\n \n The main concepts with this API are:\n \n-- :class:`gcloud.datastore.connection.Connection`\n+- :class:`google.cloud.datastore.connection.Connection`\n which represents a connection between your machine and the Cloud Datastore\n API.\n \n-- :class:`gcloud.datastore.client.Client`\n+- :class:`google.cloud.datastore.client.Client`\n which represents a project (string) and namespace (string) bundled with\n a connection and has convenience methods for constructing objects with that\n project / namespace.\n \n-- :class:`gcloud.datastore.entity.Entity`\n+- :class:`google.cloud.datastore.entity.Entity`\n which represents a single entity in the datastore\n (akin to a row in relational database world).\n \n-- :class:`gcloud.datastore.key.Key`\n+- :class:`google.cloud.datastore.key.Key`\n which represents a pointer to a particular entity in the datastore\n (akin to a unique identifier in relational database world).\n \n-- :class:`gcloud.datastore.query.Query`\n+- :class:`google.cloud.datastore.query.Query`\n which represents a lookup or search over the rows in the datastore.\n \n-- :class:`gcloud.datastore.transaction.Transaction`\n+- :class:`google.cloud.datastore.transaction.Transaction`\n which represents an all-or-none transaction and enables consistency\n when race conditions may occur.\n \"\"\"\n \n-from gcloud.datastore.batch import Batch\n-from gcloud.datastore.connection import Connection\n-from gcloud.datastore.client import Client\n-from gcloud.datastore.entity import Entity\n-from gcloud.datastore.key import Key\n-from gcloud.datastore.query import Query\n-from gcloud.datastore.transaction import Transaction\n+\n+from google.cloud.datastore.batch import Batch\n+from google.cloud.datastore.connection import Connection\n+from google.cloud.datastore.client import Client\n+from google.cloud.datastore.entity import Entity\n+from google.cloud.datastore.key import Key\n+from google.cloud.datastore.query import Query\n+from google.cloud.datastore.transaction import Transaction\n \n \n SCOPE = Connection.SCOPE\ndiff --git a/gcloud/datastore/_generated/__init__.py b/google/cloud/datastore/_generated/__init__.py\nsimilarity index 100%\nrename from gcloud/datastore/_generated/__init__.py\nrename to google/cloud/datastore/_generated/__init__.py\ndiff --git a/gcloud/datastore/_generated/datastore_grpc_pb2.py b/google/cloud/datastore/_generated/datastore_grpc_pb2.py\nsimilarity index 91%\nrename from gcloud/datastore/_generated/datastore_grpc_pb2.py\nrename to google/cloud/datastore/_generated/datastore_grpc_pb2.py\n--- a/gcloud/datastore/_generated/datastore_grpc_pb2.py\n+++ b/google/cloud/datastore/_generated/datastore_grpc_pb2.py\n@@ -1,19 +1,19 @@\n # BEGIN: Imports from datastore_pb2\n-from gcloud.datastore._generated.datastore_pb2 import AllocateIdsRequest\n-from gcloud.datastore._generated.datastore_pb2 import AllocateIdsResponse\n-from gcloud.datastore._generated.datastore_pb2 import BeginTransactionRequest\n-from gcloud.datastore._generated.datastore_pb2 import BeginTransactionResponse\n-from gcloud.datastore._generated.datastore_pb2 import CommitRequest\n-from gcloud.datastore._generated.datastore_pb2 import CommitResponse\n-from gcloud.datastore._generated.datastore_pb2 import LookupRequest\n-from gcloud.datastore._generated.datastore_pb2 import LookupResponse\n-from gcloud.datastore._generated.datastore_pb2 import Mutation\n-from gcloud.datastore._generated.datastore_pb2 import MutationResult\n-from gcloud.datastore._generated.datastore_pb2 import ReadOptions\n-from gcloud.datastore._generated.datastore_pb2 import RollbackRequest\n-from gcloud.datastore._generated.datastore_pb2 import RollbackResponse\n-from gcloud.datastore._generated.datastore_pb2 import RunQueryRequest\n-from gcloud.datastore._generated.datastore_pb2 import RunQueryResponse\n+from google.cloud.datastore._generated.datastore_pb2 import AllocateIdsRequest\n+from google.cloud.datastore._generated.datastore_pb2 import AllocateIdsResponse\n+from google.cloud.datastore._generated.datastore_pb2 import BeginTransactionRequest\n+from google.cloud.datastore._generated.datastore_pb2 import BeginTransactionResponse\n+from google.cloud.datastore._generated.datastore_pb2 import CommitRequest\n+from google.cloud.datastore._generated.datastore_pb2 import CommitResponse\n+from google.cloud.datastore._generated.datastore_pb2 import LookupRequest\n+from google.cloud.datastore._generated.datastore_pb2 import LookupResponse\n+from google.cloud.datastore._generated.datastore_pb2 import Mutation\n+from google.cloud.datastore._generated.datastore_pb2 import MutationResult\n+from google.cloud.datastore._generated.datastore_pb2 import ReadOptions\n+from google.cloud.datastore._generated.datastore_pb2 import RollbackRequest\n+from google.cloud.datastore._generated.datastore_pb2 import RollbackResponse\n+from google.cloud.datastore._generated.datastore_pb2 import RunQueryRequest\n+from google.cloud.datastore._generated.datastore_pb2 import RunQueryResponse\n # END: Imports from datastore_pb2\n import grpc\n from grpc.beta import implementations as beta_implementations\ndiff --git a/gcloud/datastore/_generated/datastore_pb2.py b/google/cloud/datastore/_generated/datastore_pb2.py\nsimilarity index 99%\nrename from gcloud/datastore/_generated/datastore_pb2.py\nrename to google/cloud/datastore/_generated/datastore_pb2.py\n--- a/gcloud/datastore/_generated/datastore_pb2.py\n+++ b/google/cloud/datastore/_generated/datastore_pb2.py\n@@ -14,8 +14,8 @@\n \n \n from google.api import annotations_pb2 as google_dot_api_dot_annotations__pb2\n-from gcloud.datastore._generated import entity_pb2 as google_dot_datastore_dot_v1_dot_entity__pb2\n-from gcloud.datastore._generated import query_pb2 as google_dot_datastore_dot_v1_dot_query__pb2\n+from google.cloud.datastore._generated import entity_pb2 as google_dot_datastore_dot_v1_dot_entity__pb2\n+from google.cloud.datastore._generated import query_pb2 as google_dot_datastore_dot_v1_dot_query__pb2\n \n \n DESCRIPTOR = _descriptor.FileDescriptor(\ndiff --git a/gcloud/datastore/_generated/entity_pb2.py b/google/cloud/datastore/_generated/entity_pb2.py\nsimilarity index 100%\nrename from gcloud/datastore/_generated/entity_pb2.py\nrename to google/cloud/datastore/_generated/entity_pb2.py\ndiff --git a/gcloud/datastore/_generated/query_pb2.py b/google/cloud/datastore/_generated/query_pb2.py\nsimilarity index 99%\nrename from gcloud/datastore/_generated/query_pb2.py\nrename to google/cloud/datastore/_generated/query_pb2.py\n--- a/gcloud/datastore/_generated/query_pb2.py\n+++ b/google/cloud/datastore/_generated/query_pb2.py\n@@ -14,7 +14,7 @@\n \n \n from google.api import annotations_pb2 as google_dot_api_dot_annotations__pb2\n-from gcloud.datastore._generated import entity_pb2 as google_dot_datastore_dot_v1_dot_entity__pb2\n+from google.cloud.datastore._generated import entity_pb2 as google_dot_datastore_dot_v1_dot_entity__pb2\n from google.protobuf import wrappers_pb2 as google_dot_protobuf_dot_wrappers__pb2\n from google.type import latlng_pb2 as google_dot_type_dot_latlng__pb2\n \ndiff --git a/gcloud/datastore/batch.py b/google/cloud/datastore/batch.py\nsimilarity index 91%\nrename from gcloud/datastore/batch.py\nrename to google/cloud/datastore/batch.py\n--- a/gcloud/datastore/batch.py\n+++ b/google/cloud/datastore/batch.py\n@@ -21,8 +21,8 @@\n https://cloud.google.com/datastore/docs/concepts/entities#Datastore_Batch_operations\n \"\"\"\n \n-from gcloud.datastore import helpers\n-from gcloud.datastore._generated import datastore_pb2 as _datastore_pb2\n+from google.cloud.datastore import helpers\n+from google.cloud.datastore._generated import datastore_pb2 as _datastore_pb2\n \n \n class Batch(object):\n@@ -34,7 +34,7 @@ class Batch(object):\n operations and the ``delete`` operation into the same mutation, and send\n them to the server in a single API request::\n \n- >>> from gcloud import datastore\n+ >>> from google.cloud import datastore\n >>> client = datastore.Client()\n >>> batch = client.batch()\n >>> batch.put(entity1)\n@@ -57,7 +57,7 @@ class Batch(object):\n ... do_some_work(batch)\n ... raise Exception() # rolls back\n \n- :type client: :class:`gcloud.datastore.client.Client`\n+ :type client: :class:`google.cloud.datastore.client.Client`\n :param client: The client used to connect to datastore.\n \"\"\"\n \n@@ -107,7 +107,7 @@ def namespace(self):\n def connection(self):\n \"\"\"Getter for connection over which the batch will run.\n \n- :rtype: :class:`gcloud.datastore.connection.Connection`\n+ :rtype: :class:`google.cloud.datastore.connection.Connection`\n :returns: The connection over which the batch will run.\n \"\"\"\n return self._client.connection\n@@ -115,7 +115,7 @@ def connection(self):\n def _add_partial_key_entity_pb(self):\n \"\"\"Adds a new mutation for an entity with a partial key.\n \n- :rtype: :class:`gcloud.datastore._generated.entity_pb2.Entity`\n+ :rtype: :class:`google.cloud.datastore._generated.entity_pb2.Entity`\n :returns: The newly created entity protobuf that will be\n updated and sent with a commit.\n \"\"\"\n@@ -125,7 +125,7 @@ def _add_partial_key_entity_pb(self):\n def _add_complete_key_entity_pb(self):\n \"\"\"Adds a new mutation for an entity with a completed key.\n \n- :rtype: :class:`gcloud.datastore._generated.entity_pb2.Entity`\n+ :rtype: :class:`google.cloud.datastore._generated.entity_pb2.Entity`\n :returns: The newly created entity protobuf that will be\n updated and sent with a commit.\n \"\"\"\n@@ -138,7 +138,7 @@ def _add_complete_key_entity_pb(self):\n def _add_delete_key_pb(self):\n \"\"\"Adds a new mutation for a key to be deleted.\n \n- :rtype: :class:`gcloud.datastore._generated.entity_pb2.Key`\n+ :rtype: :class:`google.cloud.datastore._generated.entity_pb2.Key`\n :returns: The newly created key protobuf that will be\n deleted when sent with a commit.\n \"\"\"\n@@ -180,7 +180,7 @@ def put(self, entity):\n the key for the ``entity`` passed in is updated to match the key ID\n assigned by the server.\n \n- :type entity: :class:`gcloud.datastore.entity.Entity`\n+ :type entity: :class:`google.cloud.datastore.entity.Entity`\n :param entity: the entity to be saved.\n \n :raises: ValueError if entity has no key assigned, or if the key's\n@@ -203,7 +203,7 @@ def put(self, entity):\n def delete(self, key):\n \"\"\"Remember a key to be deleted during :meth:`commit`.\n \n- :type key: :class:`gcloud.datastore.key.Key`\n+ :type key: :class:`google.cloud.datastore.key.Key`\n :param key: the key to be deleted.\n \n :raises: ValueError if key is not complete, or if the key's\n@@ -225,7 +225,7 @@ def begin(self):\n statement, however it can be called explicitly if you don't want\n to use a context manager.\n \n- Overridden by :class:`gcloud.datastore.transaction.Transaction`.\n+ Overridden by :class:`google.cloud.datastore.transaction.Transaction`.\n \n :raises: :class:`ValueError` if the batch has already begun.\n \"\"\"\n@@ -266,7 +266,7 @@ def rollback(self):\n \n Marks the batch as aborted (can't be used again).\n \n- Overridden by :class:`gcloud.datastore.transaction.Transaction`.\n+ Overridden by :class:`google.cloud.datastore.transaction.Transaction`.\n \"\"\"\n self._status = self._ABORTED\n \n@@ -290,10 +290,10 @@ def _assign_entity_to_pb(entity_pb, entity):\n \n Helper method for ``Batch.put``.\n \n- :type entity_pb: :class:`gcloud.datastore._generated.entity_pb2.Entity`\n+ :type entity_pb: :class:`.datastore._generated.entity_pb2.Entity`\n :param entity_pb: The entity owned by a mutation.\n \n- :type entity: :class:`gcloud.datastore.entity.Entity`\n+ :type entity: :class:`google.cloud.datastore.entity.Entity`\n :param entity: The entity being updated within the batch / transaction.\n \"\"\"\n bare_entity_pb = helpers.entity_to_protobuf(entity)\ndiff --git a/gcloud/datastore/client.py b/google/cloud/datastore/client.py\nsimilarity index 84%\nrename from gcloud/datastore/client.py\nrename to google/cloud/datastore/client.py\n--- a/gcloud/datastore/client.py\n+++ b/google/cloud/datastore/client.py\n@@ -15,18 +15,19 @@\n \n import os\n \n-from gcloud._helpers import _LocalStack\n-from gcloud._helpers import _determine_default_project as _base_default_project\n-from gcloud.client import _ClientProjectMixin\n-from gcloud.client import Client as _BaseClient\n-from gcloud.datastore import helpers\n-from gcloud.datastore.connection import Connection\n-from gcloud.datastore.batch import Batch\n-from gcloud.datastore.entity import Entity\n-from gcloud.datastore.key import Key\n-from gcloud.datastore.query import Query\n-from gcloud.datastore.transaction import Transaction\n-from gcloud.environment_vars import GCD_DATASET\n+from google.cloud._helpers import _LocalStack\n+from google.cloud._helpers import (\n+ _determine_default_project as _base_default_project)\n+from google.cloud.client import _ClientProjectMixin\n+from google.cloud.client import Client as _BaseClient\n+from google.cloud.datastore import helpers\n+from google.cloud.datastore.connection import Connection\n+from google.cloud.datastore.batch import Batch\n+from google.cloud.datastore.entity import Entity\n+from google.cloud.datastore.key import Key\n+from google.cloud.datastore.query import Query\n+from google.cloud.datastore.transaction import Transaction\n+from google.cloud.environment_vars import GCD_DATASET\n \n \n _MAX_LOOPS = 128\n@@ -45,7 +46,7 @@ def _determine_default_project(project=None):\n implicit environments are:\n \n * DATASTORE_DATASET environment variable (for ``gcd`` / emulator testing)\n- * GCLOUD_PROJECT environment variable\n+ * GOOGLE_CLOUD_PROJECT environment variable\n * Google App Engine application ID\n * Google Compute Engine project ID (from metadata server)\n \n@@ -71,13 +72,13 @@ def _extended_lookup(connection, project, key_pbs,\n \n Helper function for :meth:`Client.get_multi`.\n \n- :type connection: :class:`gcloud.datastore.connection.Connection`\n+ :type connection: :class:`google.cloud.datastore.connection.Connection`\n :param connection: The connection used to connect to datastore.\n \n :type project: string\n :param project: The project to make the request for.\n \n- :type key_pbs: list of :class:`gcloud.datastore._generated.entity_pb2.Key`\n+ :type key_pbs: list of :class:`.datastore._generated.entity_pb2.Key`\n :param key_pbs: The keys to retrieve from the datastore.\n \n :type missing: list\n@@ -99,7 +100,7 @@ def _extended_lookup(connection, project, key_pbs,\n the given transaction. Incompatible with\n ``eventual==True``.\n \n- :rtype: list of :class:`gcloud.datastore._generated.entity_pb2.Entity`\n+ :rtype: list of :class:`.datastore._generated.entity_pb2.Entity`\n :returns: The requested entities.\n :raises: :class:`ValueError` if missing / deferred are not null or\n empty list.\n@@ -182,7 +183,7 @@ def _push_batch(self, batch):\n \n \"Protected\", intended for use by batch / transaction context mgrs.\n \n- :type batch: :class:`gcloud.datastore.batch.Batch`, or an object\n+ :type batch: :class:`google.cloud.datastore.batch.Batch`, or an object\n implementing its API.\n :param batch: newly-active batch/transaction.\n \"\"\"\n@@ -194,7 +195,7 @@ def _pop_batch(self):\n \"Protected\", intended for use by batch / transaction context mgrs.\n \n :raises: IndexError if the stack is empty.\n- :rtype: :class:`gcloud.datastore.batch.Batch`, or an object\n+ :rtype: :class:`google.cloud.datastore.batch.Batch`, or an object\n implementing its API.\n :returns: the top-most batch/transaction, after removing it.\n \"\"\"\n@@ -204,7 +205,7 @@ def _pop_batch(self):\n def current_batch(self):\n \"\"\"Currently-active batch.\n \n- :rtype: :class:`gcloud.datastore.batch.Batch`, or an object\n+ :rtype: :class:`google.cloud.datastore.batch.Batch`, or an object\n implementing its API, or ``NoneType`` (if no batch is active).\n :returns: The batch/transaction at the top of the batch stack.\n \"\"\"\n@@ -214,9 +215,9 @@ def current_batch(self):\n def current_transaction(self):\n \"\"\"Currently-active transaction.\n \n- :rtype: :class:`gcloud.datastore.transaction.Transaction`, or an object\n- implementing its API, or ``NoneType`` (if no transaction is\n- active).\n+ :rtype: :class:`google.cloud.datastore.transaction.Transaction`, or an\n+ object implementing its API, or ``NoneType`` (if no transaction\n+ is active).\n :returns: The transaction at the top of the batch stack.\n \"\"\"\n transaction = self.current_batch\n@@ -232,7 +233,7 @@ def get(self, key, missing=None, deferred=None, transaction=None):\n The backend API does not make a distinction between a single key or\n multiple keys in a lookup request.\n \n- :type key: :class:`gcloud.datastore.key.Key`\n+ :type key: :class:`google.cloud.datastore.key.Key`\n :param key: The key to be retrieved from the datastore.\n \n :type missing: list\n@@ -244,11 +245,11 @@ def get(self, key, missing=None, deferred=None, transaction=None):\n :param deferred: (Optional) If a list is passed, the keys returned\n by the backend as \"deferred\" will be copied into it.\n \n- :type transaction: :class:`gcloud.datastore.transaction.Transaction`\n+ :type transaction: :class:`~.datastore.transaction.Transaction`\n :param transaction: (Optional) Transaction to use for read consistency.\n If not passed, uses current transaction, if set.\n \n- :rtype: :class:`gcloud.datastore.entity.Entity` or ``NoneType``\n+ :rtype: :class:`google.cloud.datastore.entity.Entity` or ``NoneType``\n :returns: The requested entity if it exists.\n \"\"\"\n entities = self.get_multi(keys=[key], missing=missing,\n@@ -259,7 +260,7 @@ def get(self, key, missing=None, deferred=None, transaction=None):\n def get_multi(self, keys, missing=None, deferred=None, transaction=None):\n \"\"\"Retrieve entities, along with their attributes.\n \n- :type keys: list of :class:`gcloud.datastore.key.Key`\n+ :type keys: list of :class:`google.cloud.datastore.key.Key`\n :param keys: The keys to be retrieved from the datastore.\n \n :type missing: list\n@@ -272,11 +273,11 @@ def get_multi(self, keys, missing=None, deferred=None, transaction=None):\n by the backend as \"deferred\" will be copied into it.\n If the list is not empty, an error will occur.\n \n- :type transaction: :class:`gcloud.datastore.transaction.Transaction`\n+ :type transaction: :class:`~.datastore.transaction.Transaction`\n :param transaction: (Optional) Transaction to use for read consistency.\n If not passed, uses current transaction, if set.\n \n- :rtype: list of :class:`gcloud.datastore.entity.Entity`\n+ :rtype: list of :class:`google.cloud.datastore.entity.Entity`\n :returns: The requested entities.\n :raises: :class:`ValueError` if one or more of ``keys`` has a project\n which does not match our project.\n@@ -323,7 +324,7 @@ def put(self, entity):\n The backend API does not make a distinction between a single\n entity or multiple entities in a commit request.\n \n- :type entity: :class:`gcloud.datastore.entity.Entity`\n+ :type entity: :class:`google.cloud.datastore.entity.Entity`\n :param entity: The entity to be saved to the datastore.\n \"\"\"\n self.put_multi(entities=[entity])\n@@ -331,7 +332,7 @@ def put(self, entity):\n def put_multi(self, entities):\n \"\"\"Save entities in the Cloud Datastore.\n \n- :type entities: list of :class:`gcloud.datastore.entity.Entity`\n+ :type entities: list of :class:`google.cloud.datastore.entity.Entity`\n :param entities: The entities to be saved to the datastore.\n \n :raises: :class:`ValueError` if ``entities`` is a single entity.\n@@ -363,7 +364,7 @@ def delete(self, key):\n The backend API does not make a distinction between a single key or\n multiple keys in a commit request.\n \n- :type key: :class:`gcloud.datastore.key.Key`\n+ :type key: :class:`google.cloud.datastore.key.Key`\n :param key: The key to be deleted from the datastore.\n \"\"\"\n self.delete_multi(keys=[key])\n@@ -371,7 +372,7 @@ def delete(self, key):\n def delete_multi(self, keys):\n \"\"\"Delete keys from the Cloud Datastore.\n \n- :type keys: list of :class:`gcloud.datastore.key.Key`\n+ :type keys: list of :class:`google.cloud.datastore.key.Key`\n :param keys: The keys to be deleted from the Datastore.\n \"\"\"\n if not keys:\n@@ -393,13 +394,13 @@ def delete_multi(self, keys):\n def allocate_ids(self, incomplete_key, num_ids):\n \"\"\"Allocate a list of IDs from a partial key.\n \n- :type incomplete_key: :class:`gcloud.datastore.key.Key`\n+ :type incomplete_key: :class:`google.cloud.datastore.key.Key`\n :param incomplete_key: Partial key to use as base for allocated IDs.\n \n :type num_ids: int\n :param num_ids: The number of IDs to allocate.\n \n- :rtype: list of :class:`gcloud.datastore.key.Key`\n+ :rtype: list of :class:`google.cloud.datastore.key.Key`\n :returns: The (complete) keys allocated with ``incomplete_key`` as\n root.\n :raises: :class:`ValueError` if ``incomplete_key`` is not a\n@@ -420,7 +421,7 @@ def allocate_ids(self, incomplete_key, num_ids):\n for allocated_id in allocated_ids]\n \n def key(self, *path_args, **kwargs):\n- \"\"\"Proxy to :class:`gcloud.datastore.key.Key`.\n+ \"\"\"Proxy to :class:`google.cloud.datastore.key.Key`.\n \n Passes our ``project``.\n \"\"\"\n@@ -432,27 +433,27 @@ def key(self, *path_args, **kwargs):\n return Key(*path_args, **kwargs)\n \n def batch(self):\n- \"\"\"Proxy to :class:`gcloud.datastore.batch.Batch`.\"\"\"\n+ \"\"\"Proxy to :class:`google.cloud.datastore.batch.Batch`.\"\"\"\n return Batch(self)\n \n def transaction(self):\n- \"\"\"Proxy to :class:`gcloud.datastore.transaction.Transaction`.\"\"\"\n+ \"\"\"Proxy to :class:`google.cloud.datastore.transaction.Transaction`.\"\"\"\n return Transaction(self)\n \n def query(self, **kwargs):\n- \"\"\"Proxy to :class:`gcloud.datastore.query.Query`.\n+ \"\"\"Proxy to :class:`google.cloud.datastore.query.Query`.\n \n Passes our ``project``.\n \n Using query to search a datastore::\n \n- >>> from gcloud import datastore\n+ >>> from google.cloud import datastore\n >>> client = datastore.Client()\n >>> query = client.query(kind='MyKind')\n >>> query.add_filter('property', '=', 'val')\n \n Using the query iterator's\n- :meth:`next_page() ` method:\n+ :meth:`~google.cloud.datastore.query.Iterator.next_page` method:\n \n >>> query_iter = query.fetch()\n >>> entities, more_results, cursor = query_iter.next_page()\n@@ -470,10 +471,10 @@ def query(self, **kwargs):\n \n :type kwargs: dict\n :param kwargs: Parameters for initializing and instance of\n- :class:`gcloud.datastore.query.Query`.\n+ :class:`google.cloud.datastore.query.Query`.\n \n- :rtype: :class:`gcloud.datastore.query.Query`\n- :returns: An instance of :class:`gcloud.datastore.query.Query`\n+ :rtype: :class:`google.cloud.datastore.query.Query`\n+ :returns: An instance of :class:`google.cloud.datastore.query.Query`\n \"\"\"\n if 'client' in kwargs:\n raise TypeError('Cannot pass client')\ndiff --git a/gcloud/datastore/connection.py b/google/cloud/datastore/connection.py\nsimilarity index 93%\nrename from gcloud/datastore/connection.py\nrename to google/cloud/datastore/connection.py\n--- a/gcloud/datastore/connection.py\n+++ b/google/cloud/datastore/connection.py\n@@ -12,23 +12,23 @@\n # See the License for the specific language governing permissions and\n # limitations under the License.\n \n-\"\"\"Connections to gcloud datastore API servers.\"\"\"\n+\"\"\"Connections to Google Cloud Datastore API servers.\"\"\"\n \n import os\n \n from google.rpc import status_pb2\n \n-from gcloud._helpers import make_stub\n-from gcloud import connection as connection_module\n-from gcloud.environment_vars import GCD_HOST\n-from gcloud.exceptions import Conflict\n-from gcloud.exceptions import make_exception\n-from gcloud.datastore._generated import datastore_pb2 as _datastore_pb2\n+from google.cloud._helpers import make_stub\n+from google.cloud import connection as connection_module\n+from google.cloud.environment_vars import GCD_HOST\n+from google.cloud.exceptions import Conflict\n+from google.cloud.exceptions import make_exception\n+from google.cloud.datastore._generated import datastore_pb2 as _datastore_pb2\n # pylint: disable=ungrouped-imports\n try:\n from grpc import StatusCode\n from grpc._channel import _Rendezvous\n- from gcloud.datastore._generated import datastore_grpc_pb2\n+ from google.cloud.datastore._generated import datastore_grpc_pb2\n except ImportError: # pragma: NO COVER\n _HAVE_GRPC = False\n datastore_grpc_pb2 = None\n@@ -53,7 +53,7 @@ class _DatastoreAPIOverHttp(object):\n Methods make bare API requests without any helpers for constructing\n the requests or parsing the responses.\n \n- :type connection: :class:`gcloud.datastore.connection.Connection`\n+ :type connection: :class:`google.cloud.datastore.connection.Connection`\n :param connection: A connection object that contains helpful\n information for making requests.\n \"\"\"\n@@ -77,8 +77,8 @@ def _request(self, project, method, data):\n \n :rtype: string\n :returns: The string response content from the API call.\n- :raises: :class:`gcloud.exceptions.GCloudError` if the response\n- code is not 200 OK.\n+ :raises: :class:`google.cloud.exceptions.GoogleCloudError` if the\n+ response code is not 200 OK.\n \"\"\"\n headers = {\n 'Content-Type': 'application/x-protobuf',\n@@ -227,7 +227,7 @@ class _DatastoreAPIOverGRPC(object):\n Methods make bare API requests without any helpers for constructing\n the requests or parsing the responses.\n \n- :type connection: :class:`gcloud.datastore.connection.Connection`\n+ :type connection: :class:`google.cloud.datastore.connection.Connection`\n :param connection: A connection object that contains helpful\n information for making requests.\n \"\"\"\n@@ -423,12 +423,12 @@ def lookup(self, project, key_pbs,\n Maps the ``DatastoreService.Lookup`` protobuf RPC.\n \n This uses mostly protobufs\n- (:class:`gcloud.datastore._generated.entity_pb2.Key` as input and\n- :class:`gcloud.datastore._generated.entity_pb2.Entity` as output). It\n- is used under the hood in\n+ (:class:`google.cloud.datastore._generated.entity_pb2.Key` as input\n+ and :class:`google.cloud.datastore._generated.entity_pb2.Entity`\n+ as output). It is used under the hood in\n :meth:`Client.get() <.datastore.client.Client.get>`:\n \n- >>> from gcloud import datastore\n+ >>> from google.cloud import datastore\n >>> client = datastore.Client(project='project')\n >>> key = client.key('MyKind', 1234)\n >>> client.get(key)\n@@ -443,7 +443,7 @@ def lookup(self, project, key_pbs,\n :param project: The project to look up the keys in.\n \n :type key_pbs: list of\n- :class:`gcloud.datastore._generated.entity_pb2.Key`\n+ :class:`google.cloud.datastore._generated.entity_pb2.Key`\n :param key_pbs: The keys to retrieve from the datastore.\n \n :type eventual: bool\n@@ -459,9 +459,9 @@ def lookup(self, project, key_pbs,\n :rtype: tuple\n :returns: A triple of (``results``, ``missing``, ``deferred``) where\n both ``results`` and ``missing`` are lists of\n- :class:`gcloud.datastore._generated.entity_pb2.Entity` and\n- ``deferred`` is a list of\n- :class:`gcloud.datastore._generated.entity_pb2.Key`.\n+ :class:`google.cloud.datastore._generated.entity_pb2.Entity`\n+ and ``deferred`` is a list of\n+ :class:`google.cloud.datastore._generated.entity_pb2.Key`.\n \"\"\"\n lookup_request = _datastore_pb2.LookupRequest()\n _set_read_options(lookup_request, eventual, transaction_id)\n@@ -485,15 +485,15 @@ def run_query(self, project, query_pb, namespace=None,\n matching the query.\n \n You typically wouldn't use this method directly, in favor of the\n- :meth:`gcloud.datastore.query.Query.fetch` method.\n+ :meth:`google.cloud.datastore.query.Query.fetch` method.\n \n- Under the hood, the :class:`gcloud.datastore.query.Query` class\n+ Under the hood, the :class:`google.cloud.datastore.query.Query` class\n uses this method to fetch data.\n \n :type project: string\n :param project: The project over which to run the query.\n \n- :type query_pb: :class:`gcloud.datastore._generated.query_pb2.Query`\n+ :type query_pb: :class:`.datastore._generated.query_pb2.Query`\n :param query_pb: The Protobuf representing the query to run.\n \n :type namespace: string\n@@ -604,10 +604,10 @@ def allocate_ids(self, project, key_pbs):\n :param project: The project to which the transaction belongs.\n \n :type key_pbs: list of\n- :class:`gcloud.datastore._generated.entity_pb2.Key`\n+ :class:`google.cloud.datastore._generated.entity_pb2.Key`\n :param key_pbs: The keys for which the backend should allocate IDs.\n \n- :rtype: list of :class:`gcloud.datastore._generated.entity_pb2.Key`\n+ :rtype: list of :class:`.datastore._generated.entity_pb2.Key`\n :returns: An equal number of keys, with IDs filled in by the backend.\n \"\"\"\n request = _datastore_pb2.AllocateIdsRequest()\n@@ -641,7 +641,7 @@ def _add_keys_to_request(request_field_pb, key_pbs):\n :type request_field_pb: `RepeatedCompositeFieldContainer`\n :param request_field_pb: A repeated proto field that contains keys.\n \n- :type key_pbs: list of :class:`gcloud.datastore._generated.entity_pb2.Key`\n+ :type key_pbs: list of :class:`.datastore._generated.entity_pb2.Key`\n :param key_pbs: The keys to add to a request.\n \"\"\"\n for key_pb in key_pbs:\ndiff --git a/gcloud/datastore/entity.py b/google/cloud/datastore/entity.py\nsimilarity index 91%\nrename from gcloud/datastore/entity.py\nrename to google/cloud/datastore/entity.py\n--- a/gcloud/datastore/entity.py\n+++ b/google/cloud/datastore/entity.py\n@@ -15,7 +15,7 @@\n \"\"\"Class for representing a single entity in the Cloud Datastore.\"\"\"\n \n \n-from gcloud._helpers import _ensure_tuple_or_list\n+from google.cloud._helpers import _ensure_tuple_or_list\n \n \n class Entity(dict):\n@@ -24,7 +24,7 @@ class Entity(dict):\n An entity storing the actual instance of data.\n \n Each entity is officially represented with a\n- :class:`gcloud.datastore.key.Key` class, however it is possible that\n+ :class:`google.cloud.datastore.key.Key` class, however it is possible that\n you might create an Entity with only a partial Key (that is, a Key\n with a Kind, and possibly a parent, but without an ID). In such a\n case, the datastore service will automatically assign an ID to the\n@@ -37,9 +37,9 @@ class Entity(dict):\n This means you could take an existing entity and change the key\n to duplicate the object.\n \n- Use :func:`gcloud.datastore.get` to retrieve an existing entity.\n+ Use :func:`google.cloud.datastore.get` to retrieve an existing entity.\n \n- >>> from gcloud import datastore\n+ >>> from google.cloud import datastore\n >>> client = datastore.Client()\n >>> client.get(key)\n \n@@ -68,7 +68,7 @@ class Entity(dict):\n Python3), will be saved using the 'blob_value' field, without\n any decoding / encoding step.\n \n- :type key: :class:`gcloud.datastore.key.Key`\n+ :type key: :class:`google.cloud.datastore.key.Key`\n :param key: Optional key to be set on entity.\n \n :type exclude_from_indexes: tuple of string\n@@ -82,7 +82,7 @@ def __init__(self, key=None, exclude_from_indexes=()):\n self._exclude_from_indexes = set(_ensure_tuple_or_list(\n 'exclude_from_indexes', exclude_from_indexes))\n # NOTE: This will be populated when parsing a protobuf in\n- # gcloud.datastore.helpers.entity_from_protobuf.\n+ # google.cloud.datastore.helpers.entity_from_protobuf.\n self._meanings = {}\n \n def __eq__(self, other):\n@@ -118,7 +118,7 @@ def kind(self):\n \"\"\"Get the kind of the current entity.\n \n .. note::\n- This relies entirely on the :class:`gcloud.datastore.key.Key`\n+ This relies entirely on the :class:`google.cloud.datastore.key.Key`\n set on the entity. That means that we're not storing the kind\n of the entity at all, just the properties and a pointer to a\n Key which knows its Kind.\ndiff --git a/gcloud/datastore/helpers.py b/google/cloud/datastore/helpers.py\nsimilarity index 90%\nrename from gcloud/datastore/helpers.py\nrename to google/cloud/datastore/helpers.py\n--- a/gcloud/datastore/helpers.py\n+++ b/google/cloud/datastore/helpers.py\n@@ -24,11 +24,13 @@\n from google.type import latlng_pb2\n import six\n \n-from gcloud._helpers import _datetime_to_pb_timestamp\n-from gcloud._helpers import _pb_timestamp_to_datetime\n-from gcloud.datastore._generated import entity_pb2 as _entity_pb2\n-from gcloud.datastore.entity import Entity\n-from gcloud.datastore.key import Key\n+# pylint: disable=ungrouped-imports\n+from google.cloud._helpers import _datetime_to_pb_timestamp\n+from google.cloud._helpers import _pb_timestamp_to_datetime\n+from google.cloud.datastore._generated import entity_pb2 as _entity_pb2\n+from google.cloud.datastore.entity import Entity\n+from google.cloud.datastore.key import Key\n+# pylint: enable=ungrouped-imports\n \n __all__ = ('entity_from_protobuf', 'key_from_protobuf')\n \n@@ -36,7 +38,7 @@\n def _get_meaning(value_pb, is_list=False):\n \"\"\"Get the meaning from a protobuf value.\n \n- :type value_pb: :class:`gcloud.datastore._generated.entity_pb2.Value`\n+ :type value_pb: :class:`google.cloud.datastore._generated.entity_pb2.Value`\n :param value_pb: The protobuf value to be checked for an\n associated meaning.\n \n@@ -77,13 +79,13 @@ def _get_meaning(value_pb, is_list=False):\n def _new_value_pb(entity_pb, name):\n \"\"\"Add (by name) a new ``Value`` protobuf to an entity protobuf.\n \n- :type entity_pb: :class:`gcloud.datastore._generated.entity_pb2.Entity`\n+ :type entity_pb: :class:`.datastore._generated.entity_pb2.Entity`\n :param entity_pb: An entity protobuf to add a new property to.\n \n :type name: string\n :param name: The name of the new property.\n \n- :rtype: :class:`gcloud.datastore._generated.entity_pb2.Value`\n+ :rtype: :class:`google.cloud.datastore._generated.entity_pb2.Value`\n :returns: The new ``Value`` protobuf that was added to the entity.\n \"\"\"\n return entity_pb.properties.get_or_create(name)\n@@ -92,7 +94,7 @@ def _new_value_pb(entity_pb, name):\n def _property_tuples(entity_pb):\n \"\"\"Iterator of name, ``Value`` tuples from entity properties.\n \n- :type entity_pb: :class:`gcloud.datastore._generated.entity_pb2.Entity`\n+ :type entity_pb: :class:`.datastore._generated.entity_pb2.Entity`\n :param entity_pb: An entity protobuf to add a new property to.\n \n :rtype: :class:`generator`\n@@ -108,10 +110,10 @@ def entity_from_protobuf(pb):\n The protobuf should be one returned from the Cloud Datastore\n Protobuf API.\n \n- :type pb: :class:`gcloud.datastore._generated.entity_pb2.Entity`\n+ :type pb: :class:`google.cloud.datastore._generated.entity_pb2.Entity`\n :param pb: The Protobuf representing the entity.\n \n- :rtype: :class:`gcloud.datastore.entity.Entity`\n+ :rtype: :class:`google.cloud.datastore.entity.Entity`\n :returns: The entity derived from the protobuf.\n \"\"\"\n key = None\n@@ -159,7 +161,7 @@ def _set_pb_meaning_from_entity(entity, name, value, value_pb,\n is_list=False):\n \"\"\"Add meaning information (from an entity) to a protobuf.\n \n- :type entity: :class:`gcloud.datastore.entity.Entity`\n+ :type entity: :class:`google.cloud.datastore.entity.Entity`\n :param entity: The entity to be turned into a protobuf.\n \n :type name: string\n@@ -168,7 +170,7 @@ def _set_pb_meaning_from_entity(entity, name, value, value_pb,\n :type value: object\n :param value: The current value stored as property ``name``.\n \n- :type value_pb: :class:`gcloud.datastore._generated.entity_pb2.Value`\n+ :type value_pb: :class:`google.cloud.datastore._generated.entity_pb2.Value`\n :param value_pb: The protobuf value to add meaning / meanings to.\n \n :type is_list: bool\n@@ -200,10 +202,10 @@ def _set_pb_meaning_from_entity(entity, name, value, value_pb,\n def entity_to_protobuf(entity):\n \"\"\"Converts an entity into a protobuf.\n \n- :type entity: :class:`gcloud.datastore.entity.Entity`\n+ :type entity: :class:`google.cloud.datastore.entity.Entity`\n :param entity: The entity to be turned into a protobuf.\n \n- :rtype: :class:`gcloud.datastore._generated.entity_pb2.Entity`\n+ :rtype: :class:`google.cloud.datastore._generated.entity_pb2.Entity`\n :returns: The protobuf representing the entity.\n \"\"\"\n entity_pb = _entity_pb2.Entity()\n@@ -241,10 +243,10 @@ def key_from_protobuf(pb):\n The protobuf should be one returned from the Cloud Datastore\n Protobuf API.\n \n- :type pb: :class:`gcloud.datastore._generated.entity_pb2.Key`\n+ :type pb: :class:`google.cloud.datastore._generated.entity_pb2.Key`\n :param pb: The Protobuf representing the key.\n \n- :rtype: :class:`gcloud.datastore.key.Key`\n+ :rtype: :class:`google.cloud.datastore.key.Key`\n :returns: a new `Key` instance\n \"\"\"\n path_args = []\n@@ -277,7 +279,7 @@ def _pb_attr_value(val):\n \n Certain value types need to be coerced into a different type (such\n as a `datetime.datetime` into an integer timestamp, or a\n- `gcloud.datastore.key.Key` into a Protobuf representation. This\n+ `google.cloud.datastore.key.Key` into a Protobuf representation. This\n function handles that for you.\n \n .. note::\n@@ -292,7 +294,7 @@ def _pb_attr_value(val):\n >>> _pb_attr_value('my_string')\n ('string_value', 'my_string')\n \n- :type val: `datetime.datetime`, :class:`gcloud.datastore.key.Key`,\n+ :type val: `datetime.datetime`, :class:`google.cloud.datastore.key.Key`,\n bool, float, integer, string\n :param val: The value to be scrutinized.\n \n@@ -339,7 +341,7 @@ def _get_value_from_value_pb(value_pb):\n Some work is done to coerce the return value into a more useful type\n (particularly in the case of a timestamp value, or a key value).\n \n- :type value_pb: :class:`gcloud.datastore._generated.entity_pb2.Value`\n+ :type value_pb: :class:`google.cloud.datastore._generated.entity_pb2.Value`\n :param value_pb: The Value Protobuf.\n \n :rtype: object\n@@ -399,12 +401,12 @@ def _set_protobuf_value(value_pb, val):\n Some value types (entities, keys, lists) cannot be directly\n assigned; this function handles them correctly.\n \n- :type value_pb: :class:`gcloud.datastore._generated.entity_pb2.Value`\n+ :type value_pb: :class:`google.cloud.datastore._generated.entity_pb2.Value`\n :param value_pb: The value protobuf to which the value is being assigned.\n \n :type val: :class:`datetime.datetime`, boolean, float, integer, string,\n- :class:`gcloud.datastore.key.Key`,\n- :class:`gcloud.datastore.entity.Entity`\n+ :class:`google.cloud.datastore.key.Key`,\n+ :class:`google.cloud.datastore.entity.Entity`\n :param val: The value to be assigned.\n \"\"\"\n attr, val = _pb_attr_value(val)\ndiff --git a/gcloud/datastore/key.py b/google/cloud/datastore/key.py\nsimilarity index 95%\nrename from gcloud/datastore/key.py\nrename to google/cloud/datastore/key.py\n--- a/gcloud/datastore/key.py\n+++ b/google/cloud/datastore/key.py\n@@ -12,12 +12,12 @@\n # See the License for the specific language governing permissions and\n # limitations under the License.\n \n-\"\"\"Create / interact with gcloud datastore keys.\"\"\"\n+\"\"\"Create / interact with Google Cloud Datastore keys.\"\"\"\n \n import copy\n import six\n \n-from gcloud.datastore._generated import entity_pb2 as _entity_pb2\n+from google.cloud.datastore._generated import entity_pb2 as _entity_pb2\n \n \n class Key(object):\n@@ -53,7 +53,7 @@ class Key(object):\n \n * namespace (string): A namespace identifier for the key.\n * project (string): The project associated with the key.\n- * parent (:class:`gcloud.datastore.key.Key`): The parent of the key.\n+ * parent (:class:`google.cloud.datastore.key.Key`): The parent of the key.\n \n The project argument is required unless it has been set implicitly.\n \"\"\"\n@@ -193,7 +193,7 @@ def _clone(self):\n Most attributes are simple types, so don't require copying. Other\n attributes like ``parent`` are long-lived and so we re-use them.\n \n- :rtype: :class:`gcloud.datastore.key.Key`\n+ :rtype: :class:`google.cloud.datastore.key.Key`\n :returns: A new ``Key`` instance with the same data as the current one.\n \"\"\"\n cloned_self = self.__class__(*self.flat_path,\n@@ -210,7 +210,7 @@ def completed_key(self, id_or_name):\n :type id_or_name: string or integer\n :param id_or_name: ID or name to be added to the key.\n \n- :rtype: :class:`gcloud.datastore.key.Key`\n+ :rtype: :class:`google.cloud.datastore.key.Key`\n :returns: A new ``Key`` instance with the same data as the current one\n and an extra ID or name added.\n :raises: :class:`ValueError` if the current key is not partial or if\n@@ -235,7 +235,7 @@ def completed_key(self, id_or_name):\n def to_protobuf(self):\n \"\"\"Return a protobuf corresponding to the key.\n \n- :rtype: :class:`gcloud.datastore._generated.entity_pb2.Key`\n+ :rtype: :class:`google.cloud.datastore._generated.entity_pb2.Key`\n :returns: The protobuf representing the key.\n \"\"\"\n key = _entity_pb2.Key()\n@@ -346,7 +346,7 @@ def _make_parent(self):\n Extracts all but the last element in the key path and creates a new\n key, while still matching the namespace and the project.\n \n- :rtype: :class:`gcloud.datastore.key.Key` or :class:`NoneType`\n+ :rtype: :class:`google.cloud.datastore.key.Key` or :class:`NoneType`\n :returns: A new ``Key`` instance, whose path consists of all but the\n last element of current path. If the current key has only\n one path element, returns ``None``.\n@@ -363,7 +363,7 @@ def _make_parent(self):\n def parent(self):\n \"\"\"The parent of the current key.\n \n- :rtype: :class:`gcloud.datastore.key.Key` or :class:`NoneType`\n+ :rtype: :class:`google.cloud.datastore.key.Key` or :class:`NoneType`\n :returns: A new ``Key`` instance, whose path consists of all but the\n last element of current path. If the current key has only\n one path element, returns ``None``.\n@@ -388,7 +388,7 @@ def _validate_project(project, parent):\n :type project: string\n :param project: A project.\n \n- :type parent: :class:`gcloud.datastore.key.Key` or ``NoneType``\n+ :type parent: :class:`google.cloud.datastore.key.Key` or ``NoneType``\n :param parent: The parent of the key or ``None``.\n \n :rtype: string\ndiff --git a/gcloud/datastore/query.py b/google/cloud/datastore/query.py\nsimilarity index 94%\nrename from gcloud/datastore/query.py\nrename to google/cloud/datastore/query.py\n--- a/gcloud/datastore/query.py\n+++ b/google/cloud/datastore/query.py\n@@ -12,14 +12,14 @@\n # See the License for the specific language governing permissions and\n # limitations under the License.\n \n-\"\"\"Create / interact with gcloud datastore queries.\"\"\"\n+\"\"\"Create / interact with Google Cloud Datastore queries.\"\"\"\n \n import base64\n \n-from gcloud._helpers import _ensure_tuple_or_list\n-from gcloud.datastore._generated import query_pb2 as _query_pb2\n-from gcloud.datastore import helpers\n-from gcloud.datastore.key import Key\n+from google.cloud._helpers import _ensure_tuple_or_list\n+from google.cloud.datastore._generated import query_pb2 as _query_pb2\n+from google.cloud.datastore import helpers\n+from google.cloud.datastore.key import Key\n \n \n class Query(object):\n@@ -28,7 +28,7 @@ class Query(object):\n This class serves as an abstraction for creating a query over data\n stored in the Cloud Datastore.\n \n- :type client: :class:`gcloud.datastore.client.Client`\n+ :type client: :class:`google.cloud.datastore.client.Client`\n :param client: The client used to connect to Datastore.\n \n :type kind: string\n@@ -42,7 +42,7 @@ class Query(object):\n :param namespace: The namespace to which to restrict results. If not\n passed, uses the client's value.\n \n- :type ancestor: :class:`gcloud.datastore.key.Key` or None\n+ :type ancestor: :class:`google.cloud.datastore.key.Key` or None\n :param ancestor: key of the ancestor to which this query's results are\n restricted.\n \n@@ -197,7 +197,7 @@ def add_filter(self, property_name, operator, value):\n and operator is one of ``OPERATORS``\n (ie, ``=``, ``<``, ``<=``, ``>``, ``>=``)::\n \n- >>> from gcloud import datastore\n+ >>> from google.cloud import datastore\n >>> client = datastore.Client()\n >>> query = client.query(kind='Person')\n >>> query.add_filter('name', '=', 'James')\n@@ -212,7 +212,7 @@ def add_filter(self, property_name, operator, value):\n :type value: :class:`int`, :class:`str`, :class:`bool`,\n :class:`float`, :class:`NoneType`,\n :class:`datetime.datetime`,\n- :class:`gcloud.datastore.key.Key`\n+ :class:`google.cloud.datastore.key.Key`\n :param value: The value to filter on.\n \n :raises: :class:`ValueError` if ``operation`` is not one of the\n@@ -257,7 +257,7 @@ def keys_only(self):\n def key_filter(self, key, operator='='):\n \"\"\"Filter on a key.\n \n- :type key: :class:`gcloud.datastore.key.Key`\n+ :type key: :class:`google.cloud.datastore.key.Key`\n :param key: The key to filter on.\n \n :type operator: string\n@@ -318,7 +318,7 @@ def fetch(self, limit=None, offset=0, start_cursor=None, end_cursor=None,\n \n For example::\n \n- >>> from gcloud import datastore\n+ >>> from google.cloud import datastore\n >>> client = datastore.Client()\n >>> query = client.query(kind='Person')\n >>> query.add_filter('name', '=', 'Sally')\n@@ -339,7 +339,7 @@ def fetch(self, limit=None, offset=0, start_cursor=None, end_cursor=None,\n :type end_cursor: bytes\n :param end_cursor: An optional cursor passed through to the iterator.\n \n- :type client: :class:`gcloud.datastore.client.Client`\n+ :type client: :class:`google.cloud.datastore.client.Client`\n :param client: client used to connect to datastore.\n If not supplied, uses the query's value.\n \n@@ -358,12 +358,12 @@ def fetch(self, limit=None, offset=0, start_cursor=None, end_cursor=None,\n class Iterator(object):\n \"\"\"Represent the state of a given execution of a Query.\n \n- :type query: :class:`gcloud.datastore.query.Query`\n+ :type query: :class:`google.cloud.datastore.query.Query`\n :param query: Query object holding permanent configuration (i.e.\n things that don't change on with each page in\n a results set).\n \n- :type client: :class:`gcloud.datastore.client.Client`\n+ :type client: :class:`google.cloud.datastore.client.Client`\n :param client: The client used to make a request.\n \n :type limit: integer\n@@ -457,7 +457,7 @@ def next_page(self):\n def __iter__(self):\n \"\"\"Generator yielding all results matching our query.\n \n- :rtype: sequence of :class:`gcloud.datastore.entity.Entity`\n+ :rtype: sequence of :class:`google.cloud.datastore.entity.Entity`\n \"\"\"\n while True:\n self.next_page()\n@@ -480,7 +480,7 @@ def _pb_from_query(query):\n :type query: :class:`Query`\n :param query: The source query.\n \n- :rtype: :class:`gcloud.datastore._generated.query_pb2.Query`\n+ :rtype: :class:`google.cloud.datastore._generated.query_pb2.Query`\n :returns: A protobuf that can be sent to the protobuf API. N.b. that\n it does not contain \"in-flight\" fields for ongoing query\n executions (cursors, offset, limit).\ndiff --git a/gcloud/datastore/transaction.py b/google/cloud/datastore/transaction.py\nsimilarity index 94%\nrename from gcloud/datastore/transaction.py\nrename to google/cloud/datastore/transaction.py\n--- a/gcloud/datastore/transaction.py\n+++ b/google/cloud/datastore/transaction.py\n@@ -12,9 +12,9 @@\n # See the License for the specific language governing permissions and\n # limitations under the License.\n \n-\"\"\"Create / interact with gcloud datastore transactions.\"\"\"\n+\"\"\"Create / interact with Google Cloud Datastore transactions.\"\"\"\n \n-from gcloud.datastore.batch import Batch\n+from google.cloud.datastore.batch import Batch\n \n \n class Transaction(Batch):\n@@ -27,7 +27,7 @@ class Transaction(Batch):\n operations (either ``insert`` or ``upsert``) into the same\n mutation, and execute those within a transaction::\n \n- >>> from gcloud import datastore\n+ >>> from google.cloud import datastore\n >>> client = datastore.Client()\n >>> with client.transaction():\n ... client.put_multi([entity1, entity2])\n@@ -86,7 +86,7 @@ class Transaction(Batch):\n ... else:\n ... transaction.commit()\n \n- :type client: :class:`gcloud.datastore.client.Client`\n+ :type client: :class:`google.cloud.datastore.client.Client`\n :param client: the client used to connect to datastore.\n \"\"\"\n \n@@ -111,7 +111,7 @@ def current(self):\n If the topmost element on the stack is not a transaction,\n returns None.\n \n- :rtype: :class:`gcloud.datastore.transaction.Transaction` or None\n+ :rtype: :class:`google.cloud.datastore.transaction.Transaction` or None\n :returns: The current transaction (if any are active).\n \"\"\"\n top = super(Transaction, self).current()\ndiff --git a/gcloud/dns/__init__.py b/google/cloud/dns/__init__.py\nsimilarity index 54%\nrename from gcloud/dns/__init__.py\nrename to google/cloud/dns/__init__.py\n--- a/gcloud/dns/__init__.py\n+++ b/google/cloud/dns/__init__.py\n@@ -16,18 +16,20 @@\n \n The main concepts with this API are:\n \n-- :class:`gcloud.DNS.zone.ManagedZone` represents an collection of tables.\n-- :class:`gcloud.DNS.resource_record_set.ResourceRecordSet` represents a\n- single resource definition within a zone.\n-- :class:`gcloud.DNS.changes.Changes` represents a set of changes (adding/\n- deleting resource record sets) to a zone.\n+- :class:`~google.cloud.DNS.zone.ManagedZone` represents an collection of\n+ tables.\n+- :class:`~google.cloud.DNS.resource_record_set.ResourceRecordSet` represents\n+ a single resource definition within a zone.\n+- :class:`~google.cloud.DNS.changes.Changes` represents a set of changes\n+ (adding/deleting resource record sets) to a zone.\n \"\"\"\n \n-from gcloud.dns.zone import Changes\n-from gcloud.dns.client import Client\n-from gcloud.dns.connection import Connection\n-from gcloud.dns.zone import ManagedZone\n-from gcloud.dns.resource_record_set import ResourceRecordSet\n+\n+from google.cloud.dns.zone import Changes\n+from google.cloud.dns.client import Client\n+from google.cloud.dns.connection import Connection\n+from google.cloud.dns.zone import ManagedZone\n+from google.cloud.dns.resource_record_set import ResourceRecordSet\n \n \n SCOPE = Connection.SCOPE\ndiff --git a/gcloud/dns/changes.py b/google/cloud/dns/changes.py\nsimilarity index 87%\nrename from gcloud/dns/changes.py\nrename to google/cloud/dns/changes.py\n--- a/gcloud/dns/changes.py\n+++ b/google/cloud/dns/changes.py\n@@ -16,20 +16,20 @@\n \n import six\n \n-from gcloud._helpers import _rfc3339_to_datetime\n-from gcloud.exceptions import NotFound\n-from gcloud.dns.resource_record_set import ResourceRecordSet\n+from google.cloud._helpers import _rfc3339_to_datetime\n+from google.cloud.exceptions import NotFound\n+from google.cloud.dns.resource_record_set import ResourceRecordSet\n \n \n class Changes(object):\n \"\"\"Changes are bundled additions / deletions of DNS resource records.\n \n- Changes are owned by a :class:`gcloud.dns.zone.ManagedZone` instance.\n+ Changes are owned by a :class:`google.cloud.dns.zone.ManagedZone` instance.\n \n See:\n https://cloud.google.com/dns/api/v1/changes\n \n- :type zone: :class:`gcloud.dns.zone.ManagedZone`\n+ :type zone: :class:`google.cloud.dns.zone.ManagedZone`\n :param zone: A zone which holds one or more record sets.\n \"\"\"\n \n@@ -45,10 +45,10 @@ def from_api_repr(cls, resource, zone):\n :type resource: dict\n :param resource: change set representation returned from the API.\n \n- :type zone: :class:`gcloud.dns.zone.ManagedZone`\n+ :type zone: :class:`google.cloud.dns.zone.ManagedZone`\n :param zone: A zone which holds zero or more change sets.\n \n- :rtype: :class:`gcloud.dns.changes.Changes`\n+ :rtype: :class:`google.cloud.dns.changes.Changes`\n :returns: RRS parsed from ``resource``.\n \"\"\"\n changes = cls(zone=zone)\n@@ -125,7 +125,7 @@ def additions(self):\n \"\"\"Resource record sets to be added to the zone.\n \n :rtype: sequence of\n- :class:`gcloud.dns.resource_record_set.ResourceRecordSet`.\n+ :class:`google.cloud.dns.resource_record_set.ResourceRecordSet`.\n :returns: record sets appended via :meth:`add_record_set`.\n \"\"\"\n return self._additions\n@@ -135,7 +135,7 @@ def deletions(self):\n \"\"\"Resource record sets to be deleted from the zone.\n \n :rtype: sequence of\n- :class:`gcloud.dns.resource_record_set.ResourceRecordSet`.\n+ :class:`google.cloud.dns.resource_record_set.ResourceRecordSet`.\n :returns: record sets appended via :meth:`delete_record_set`.\n \"\"\"\n return self._deletions\n@@ -144,7 +144,7 @@ def add_record_set(self, record_set):\n \"\"\"Append a record set to the 'additions' for the change set.\n \n :type record_set:\n- :class:`gcloud.dns.resource_record_set.ResourceRecordSet`\n+ :class:`google.cloud.dns.resource_record_set.ResourceRecordSet`\n :param record_set: the record set to append.\n \n :raises: ``ValueError`` if ``record_set`` is not of the required type.\n@@ -157,7 +157,7 @@ def delete_record_set(self, record_set):\n \"\"\"Append a record set to the 'deletions' for the change set.\n \n :type record_set:\n- :class:`gcloud.dns.resource_record_set.ResourceRecordSet`\n+ :class:`google.cloud.dns.resource_record_set.ResourceRecordSet`\n :param record_set: the record set to append.\n \n :raises: ``ValueError`` if ``record_set`` is not of the required type.\n@@ -169,11 +169,11 @@ def delete_record_set(self, record_set):\n def _require_client(self, client):\n \"\"\"Check client or verify over-ride.\n \n- :type client: :class:`gcloud.dns.client.Client` or ``NoneType``\n+ :type client: :class:`google.cloud.dns.client.Client` or ``NoneType``\n :param client: the client to use. If not passed, falls back to the\n ``client`` stored on the current zone.\n \n- :rtype: :class:`gcloud.dns.client.Client`\n+ :rtype: :class:`google.cloud.dns.client.Client`\n :returns: The client passed in or the currently bound client.\n \"\"\"\n if client is None:\n@@ -207,7 +207,7 @@ def create(self, client=None):\n See:\n https://cloud.google.com/dns/api/v1/changes/create\n \n- :type client: :class:`gcloud.dns.client.Client` or ``NoneType``\n+ :type client: :class:`google.cloud.dns.client.Client` or ``NoneType``\n :param client: the client to use. If not passed, falls back to the\n ``client`` stored on the current zone.\n \"\"\"\n@@ -226,7 +226,7 @@ def exists(self, client=None):\n See\n https://cloud.google.com/dns/api/v1/changes/get\n \n- :type client: :class:`gcloud.dns.client.Client` or ``NoneType``\n+ :type client: :class:`google.cloud.dns.client.Client` or ``NoneType``\n :param client: the client to use. If not passed, falls back to the\n ``client`` stored on the current zone.\n \n@@ -248,7 +248,7 @@ def reload(self, client=None):\n See\n https://cloud.google.com/dns/api/v1/changes/get\n \n- :type client: :class:`gcloud.dns.client.Client` or ``NoneType``\n+ :type client: :class:`google.cloud.dns.client.Client` or ``NoneType``\n :param client: the client to use. If not passed, falls back to the\n ``client`` stored on the current zone.\n \"\"\"\ndiff --git a/gcloud/dns/client.py b/google/cloud/dns/client.py\nsimilarity index 94%\nrename from gcloud/dns/client.py\nrename to google/cloud/dns/client.py\n--- a/gcloud/dns/client.py\n+++ b/google/cloud/dns/client.py\n@@ -15,9 +15,9 @@\n \"\"\"Client for interacting with the Google Cloud DNS API.\"\"\"\n \n \n-from gcloud.client import JSONClient\n-from gcloud.dns.connection import Connection\n-from gcloud.dns.zone import ManagedZone\n+from google.cloud.client import JSONClient\n+from google.cloud.dns.connection import Connection\n+from google.cloud.dns.zone import ManagedZone\n \n \n class Client(JSONClient):\n@@ -75,7 +75,7 @@ def list_zones(self, max_results=None, page_token=None):\n zones.\n \n :rtype: tuple, (list, str)\n- :returns: list of :class:`gcloud.dns.zone.ManagedZone`, plus a\n+ :returns: list of :class:`google.cloud.dns.zone.ManagedZone`, plus a\n \"next page token\" string: if the token is not None,\n indicates that more zones can be retrieved with another\n call (pass that value as ``page_token``).\n@@ -109,7 +109,7 @@ def zone(self, name, dns_name=None, description=None):\n :param description: the description for the zone. If not passed,\n defaults to the value of 'dns_name'.\n \n- :rtype: :class:`gcloud.dns.zone.ManagedZone`\n+ :rtype: :class:`google.cloud.dns.zone.ManagedZone`\n :returns: a new ``ManagedZone`` instance.\n \"\"\"\n return ManagedZone(name, dns_name, client=self,\ndiff --git a/gcloud/dns/connection.py b/google/cloud/dns/connection.py\nsimilarity index 91%\nrename from gcloud/dns/connection.py\nrename to google/cloud/dns/connection.py\n--- a/gcloud/dns/connection.py\n+++ b/google/cloud/dns/connection.py\n@@ -12,9 +12,9 @@\n # See the License for the specific language governing permissions and\n # limitations under the License.\n \n-\"\"\"Create / interact with gcloud dns connections.\"\"\"\n+\"\"\"Create / interact with Google Cloud DNS connections.\"\"\"\n \n-from gcloud import connection as base_connection\n+from google.cloud import connection as base_connection\n \n \n class Connection(base_connection.JSONConnection):\ndiff --git a/gcloud/dns/resource_record_set.py b/google/cloud/dns/resource_record_set.py\nsimilarity index 88%\nrename from gcloud/dns/resource_record_set.py\nrename to google/cloud/dns/resource_record_set.py\n--- a/gcloud/dns/resource_record_set.py\n+++ b/google/cloud/dns/resource_record_set.py\n@@ -18,7 +18,7 @@\n class ResourceRecordSet(object):\n \"\"\"ResourceRecordSets are DNS resource records.\n \n- RRS are owned by a :class:`gcloud.dns.zone.ManagedZone` instance.\n+ RRS are owned by a :class:`google.cloud.dns.zone.ManagedZone` instance.\n \n See:\n https://cloud.google.com/dns/api/v1/resourceRecordSets\n@@ -35,7 +35,7 @@ class ResourceRecordSet(object):\n :type rrdatas: list of string\n :param rrdatas: one or more lines containing the resource data.\n \n- :type zone: :class:`gcloud.dns.zone.ManagedZone`\n+ :type zone: :class:`google.cloud.dns.zone.ManagedZone`\n :param zone: A zone which holds one or more record sets.\n \"\"\"\n \n@@ -53,10 +53,10 @@ def from_api_repr(cls, resource, zone):\n :type resource: dict\n :param resource: record sets representation returned from the API\n \n- :type zone: :class:`gcloud.dns.zone.ManagedZone`\n+ :type zone: :class:`google.cloud.dns.zone.ManagedZone`\n :param zone: A zone which holds one or more record sets.\n \n- :rtype: :class:`gcloud.dns.zone.ResourceRecordSet`\n+ :rtype: :class:`google.cloud.dns.zone.ResourceRecordSet`\n :returns: RRS parsed from ``resource``.\n \"\"\"\n name = resource['name']\ndiff --git a/gcloud/dns/zone.py b/google/cloud/dns/zone.py\nsimilarity index 90%\nrename from gcloud/dns/zone.py\nrename to google/cloud/dns/zone.py\n--- a/gcloud/dns/zone.py\n+++ b/google/cloud/dns/zone.py\n@@ -15,10 +15,10 @@\n \"\"\"Define API ManagedZones.\"\"\"\n import six\n \n-from gcloud._helpers import _rfc3339_to_datetime\n-from gcloud.exceptions import NotFound\n-from gcloud.dns.changes import Changes\n-from gcloud.dns.resource_record_set import ResourceRecordSet\n+from google.cloud._helpers import _rfc3339_to_datetime\n+from google.cloud.exceptions import NotFound\n+from google.cloud.dns.changes import Changes\n+from google.cloud.dns.resource_record_set import ResourceRecordSet\n \n \n class ManagedZone(object):\n@@ -34,7 +34,7 @@ class ManagedZone(object):\n :param dns_name: the DNS name of the zone. If not passed, then calls\n to :meth:`create` will fail.\n \n- :type client: :class:`gcloud.dns.client.Client`\n+ :type client: :class:`google.cloud.dns.client.Client`\n :param client: A client which holds credentials and project configuration\n for the zone (which requires a project).\n \n@@ -59,11 +59,11 @@ def from_api_repr(cls, resource, client):\n :type resource: dict\n :param resource: zone resource representation returned from the API\n \n- :type client: :class:`gcloud.dns.client.Client`\n+ :type client: :class:`google.cloud.dns.client.Client`\n :param client: Client which holds credentials and project\n configuration for the zone.\n \n- :rtype: :class:`gcloud.dns.zone.ManagedZone`\n+ :rtype: :class:`google.cloud.dns.zone.ManagedZone`\n :returns: Zone parsed from ``resource``.\n \"\"\"\n name = resource.get('name')\n@@ -184,7 +184,7 @@ def resource_record_set(self, name, record_type, ttl, rrdatas):\n :type rrdatas: list of string\n :param rrdatas: resource data for the RR\n \n- :rtype: :class:`gcloud.dns.resource_record_set.ResourceRecordSet`\n+ :rtype: :class:`google.cloud.dns.resource_record_set.ResourceRecordSet`\n :returns: a new ``ResourceRecordSet`` instance\n \"\"\"\n return ResourceRecordSet(name, record_type, ttl, rrdatas, zone=self)\n@@ -192,7 +192,7 @@ def resource_record_set(self, name, record_type, ttl, rrdatas):\n def changes(self):\n \"\"\"Construct a change set bound to this zone.\n \n- :rtype: :class:`gcloud.dns.changes.Changes`\n+ :rtype: :class:`google.cloud.dns.changes.Changes`\n :returns: a new ``Changes`` instance\n \"\"\"\n return Changes(zone=self)\n@@ -200,11 +200,11 @@ def changes(self):\n def _require_client(self, client):\n \"\"\"Check client or verify over-ride.\n \n- :type client: :class:`gcloud.dns.client.Client` or ``NoneType``\n+ :type client: :class:`google.cloud.dns.client.Client` or ``NoneType``\n :param client: the client to use. If not passed, falls back to the\n ``client`` stored on the current zone.\n \n- :rtype: :class:`gcloud.dns.client.Client`\n+ :rtype: :class:`google.cloud.dns.client.Client`\n :returns: The client passed in or the currently bound client.\n \"\"\"\n if client is None:\n@@ -248,7 +248,7 @@ def create(self, client=None):\n See:\n https://cloud.google.com/dns/api/v1/managedZones/create\n \n- :type client: :class:`gcloud.dns.client.Client` or ``NoneType``\n+ :type client: :class:`google.cloud.dns.client.Client` or ``NoneType``\n :param client: the client to use. If not passed, falls back to the\n ``client`` stored on the current zone.\n \"\"\"\n@@ -264,7 +264,7 @@ def exists(self, client=None):\n See\n https://cloud.google.com/dns/api/v1/managedZones/get\n \n- :type client: :class:`gcloud.dns.client.Client` or ``NoneType``\n+ :type client: :class:`google.cloud.dns.client.Client` or ``NoneType``\n :param client: the client to use. If not passed, falls back to the\n ``client`` stored on the current zone.\n \n@@ -287,7 +287,7 @@ def reload(self, client=None):\n See\n https://cloud.google.com/dns/api/v1/managedZones/get\n \n- :type client: :class:`gcloud.dns.client.Client` or ``NoneType``\n+ :type client: :class:`google.cloud.dns.client.Client` or ``NoneType``\n :param client: the client to use. If not passed, falls back to the\n ``client`` stored on the current zone.\n \"\"\"\n@@ -303,7 +303,7 @@ def delete(self, client=None):\n See:\n https://cloud.google.com/dns/api/v1/managedZones/delete\n \n- :type client: :class:`gcloud.dns.client.Client` or ``NoneType``\n+ :type client: :class:`google.cloud.dns.client.Client` or ``NoneType``\n :param client: the client to use. If not passed, falls back to the\n ``client`` stored on the current zone.\n \"\"\"\n@@ -326,13 +326,13 @@ def list_resource_record_sets(self, max_results=None, page_token=None,\n not passed, the API will return the first page of\n zones.\n \n- :type client: :class:`gcloud.dns.client.Client` or ``NoneType``\n+ :type client: :class:`google.cloud.dns.client.Client` or ``NoneType``\n :param client: the client to use. If not passed, falls back to the\n ``client`` stored on the current zone.\n \n :rtype: tuple, (list, str)\n :returns: list of\n- :class:`gcloud.dns.resource_record_set.ResourceRecordSet`,\n+ :class:`google.cloud.dns.resource_record_set.ResourceRecordSet`,\n plus a \"next page token\" string: if the token is not None,\n indicates that more zones can be retrieved with another\n call (pass that value as ``page_token``).\n@@ -369,13 +369,13 @@ def list_changes(self, max_results=None, page_token=None, client=None):\n not passed, the API will return the first page of\n zones.\n \n- :type client: :class:`gcloud.dns.client.Client` or ``NoneType``\n+ :type client: :class:`google.cloud.dns.client.Client` or ``NoneType``\n :param client: the client to use. If not passed, falls back to the\n ``client`` stored on the current zone.\n \n :rtype: tuple, (list, str)\n :returns: list of\n- :class:`gcloud.dns.resource_record_set.ResourceRecordSet`,\n+ :class:`google.cloud.dns.resource_record_set.ResourceRecordSet`,\n plus a \"next page token\" string: if the token is not None,\n indicates that more zones can be retrieved with another\n call (pass that value as ``page_token``).\ndiff --git a/gcloud/environment_vars.py b/google/cloud/environment_vars.py\nsimilarity index 88%\nrename from gcloud/environment_vars.py\nrename to google/cloud/environment_vars.py\n--- a/gcloud/environment_vars.py\n+++ b/google/cloud/environment_vars.py\n@@ -12,16 +12,16 @@\n # See the License for the specific language governing permissions and\n # limitations under the License.\n \n-\"\"\"Comprehensive list of environment variables used in gcloud.\n+\"\"\"Comprehensive list of environment variables used in google-cloud.\n \n These enable many types of implicit behavior in both production\n and tests.\n \"\"\"\n \n-PROJECT = 'GCLOUD_PROJECT'\n+PROJECT = 'GOOGLE_CLOUD_PROJECT'\n \"\"\"Environment variable defining default project.\"\"\"\n \n-TESTS_PROJECT = 'GCLOUD_TESTS_PROJECT_ID'\n+TESTS_PROJECT = 'GOOGLE_CLOUD_TESTS_PROJECT_ID'\n \"\"\"Environment variable defining project for tests.\"\"\"\n \n GCD_DATASET = 'DATASTORE_DATASET'\ndiff --git a/gcloud/error_reporting/__init__.py b/google/cloud/error_reporting/__init__.py\nsimilarity index 92%\nrename from gcloud/error_reporting/__init__.py\nrename to google/cloud/error_reporting/__init__.py\n--- a/gcloud/error_reporting/__init__.py\n+++ b/google/cloud/error_reporting/__init__.py\n@@ -14,4 +14,5 @@\n \n \"\"\"Client library for Stackdriver Error Reporting\"\"\"\n \n-from gcloud.error_reporting.client import Client\n+\n+from google.cloud.error_reporting.client import Client\ndiff --git a/gcloud/error_reporting/client.py b/google/cloud/error_reporting/client.py\nsimilarity index 96%\nrename from gcloud/error_reporting/client.py\nrename to google/cloud/error_reporting/client.py\n--- a/gcloud/error_reporting/client.py\n+++ b/google/cloud/error_reporting/client.py\n@@ -16,7 +16,7 @@\n \n import traceback\n \n-import gcloud.logging.client\n+import google.cloud.logging.client\n import six\n \n \n@@ -105,7 +105,7 @@ def __init__(self, project=None,\n http=None,\n service=None,\n version=None):\n- self.logging_client = gcloud.logging.client.Client(\n+ self.logging_client = google.cloud.logging.client.Client(\n project, credentials, http)\n self.service = service if service else self.DEFAULT_SERVICE\n self.version = version\n@@ -139,7 +139,7 @@ def _send_error_report(self, message,\n This should be a Python dict that contains the keys 'filePath',\n 'lineNumber', and 'functionName'\n \n- :type http_context: :class`gcloud.error_reporting.HTTPContext`\n+ :type http_context: :class`google.cloud.error_reporting.HTTPContext`\n :param http_context: The HTTP request which was processed when the\n error was triggered.\n \n@@ -190,7 +190,7 @@ def report(self, message, http_context=None, user=None):\n :param message: A user-supplied message to report\n \n \n- :type http_context: :class`gcloud.error_reporting.HTTPContext`\n+ :type http_context: :class`google.cloud.error_reporting.HTTPContext`\n :param http_context: The HTTP request which was processed when the\n error was triggered.\n \n@@ -226,7 +226,7 @@ def report_exception(self, http_context=None, user=None):\n \"\"\" Reports the details of the latest exceptions to Stackdriver Error\n Reporting.\n \n- :type http_context: :class`gcloud.error_reporting.HTTPContext`\n+ :type http_context: :class`google.cloud.error_reporting.HTTPContext`\n :param http_context: The HTTP request which was processed when the\n error was triggered.\n \ndiff --git a/gcloud/exceptions.py b/google/cloud/exceptions.py\nsimilarity index 91%\nrename from gcloud/exceptions.py\nrename to google/cloud/exceptions.py\n--- a/gcloud/exceptions.py\n+++ b/google/cloud/exceptions.py\n@@ -12,7 +12,7 @@\n # See the License for the specific language governing permissions and\n # limitations under the License.\n \n-\"\"\"Custom exceptions for :mod:`gcloud` package.\n+\"\"\"Custom exceptions for :mod:`google.cloud` package.\n \n See: https://cloud.google.com/storage/docs/json_api/v1/status-codes\n \"\"\"\n@@ -24,8 +24,8 @@\n _HTTP_CODE_TO_EXCEPTION = {} # populated at end of module\n \n \n-class GCloudError(Exception):\n- \"\"\"Base error class for gcloud errors (abstract).\n+class GoogleCloudError(Exception):\n+ \"\"\"Base error class for Google Cloud errors (abstract).\n \n Each subclass represents a single type of HTTP error response.\n \"\"\"\n@@ -36,7 +36,7 @@ class GCloudError(Exception):\n \"\"\"\n \n def __init__(self, message, errors=()):\n- super(GCloudError, self).__init__(message)\n+ super(GoogleCloudError, self).__init__(message)\n self.message = message\n self._errors = errors\n \n@@ -53,7 +53,7 @@ def errors(self):\n return [copy.deepcopy(error) for error in self._errors]\n \n \n-class Redirection(GCloudError):\n+class Redirection(GoogleCloudError):\n \"\"\"Base for 3xx responses\n \n This class is abstract.\n@@ -80,7 +80,7 @@ class ResumeIncomplete(Redirection):\n code = 308\n \n \n-class ClientError(GCloudError):\n+class ClientError(GoogleCloudError):\n \"\"\"Base for 4xx responses\n \n This class is abstract\n@@ -137,7 +137,7 @@ class TooManyRequests(ClientError):\n code = 429\n \n \n-class ServerError(GCloudError):\n+class ServerError(GoogleCloudError):\n \"\"\"Base for 5xx responses: (abstract)\"\"\"\n \n \n@@ -178,7 +178,7 @@ def make_exception(response, content, error_info=None, use_json=True):\n :type use_json: bool\n :param use_json: Flag indicating if ``content`` is expected to be JSON.\n \n- :rtype: instance of :class:`GCloudError`, or a concrete subclass.\n+ :rtype: instance of :class:`GoogleCloudError`, or a concrete subclass.\n :returns: Exception specific to the error response.\n \"\"\"\n if isinstance(content, six.binary_type):\n@@ -206,7 +206,7 @@ def make_exception(response, content, error_info=None, use_json=True):\n try:\n klass = _HTTP_CODE_TO_EXCEPTION[response.status]\n except KeyError:\n- error = GCloudError(message, errors)\n+ error = GoogleCloudError(message, errors)\n error.code = response.status\n else:\n error = klass(message, errors)\n@@ -222,7 +222,7 @@ def _walk_subclasses(klass):\n \n \n # Build the code->exception class mapping.\n-for _eklass in _walk_subclasses(GCloudError):\n+for _eklass in _walk_subclasses(GoogleCloudError):\n code = getattr(_eklass, 'code', None)\n if code is not None:\n _HTTP_CODE_TO_EXCEPTION[code] = _eklass\ndiff --git a/gcloud/iterator.py b/google/cloud/iterator.py\nsimilarity index 99%\nrename from gcloud/iterator.py\nrename to google/cloud/iterator.py\n--- a/gcloud/iterator.py\n+++ b/google/cloud/iterator.py\n@@ -48,7 +48,7 @@ def get_items_from_response(self, response):\n class Iterator(object):\n \"\"\"A generic class for iterating through Cloud JSON APIs list responses.\n \n- :type client: :class:`gcloud.client.Client`\n+ :type client: :class:`google.cloud.client.Client`\n :param client: The client, which owns a connection to make requests.\n \n :type path: string\ndiff --git a/gcloud/language/__init__.py b/google/cloud/language/__init__.py\nsimilarity index 81%\nrename from gcloud/language/__init__.py\nrename to google/cloud/language/__init__.py\n--- a/gcloud/language/__init__.py\n+++ b/google/cloud/language/__init__.py\n@@ -14,6 +14,7 @@\n \n \"\"\"Client library for Google Cloud Natural Language API.\"\"\"\n \n-from gcloud.language.client import Client\n-from gcloud.language.document import Document\n-from gcloud.language.document import Encoding\n+\n+from google.cloud.language.client import Client\n+from google.cloud.language.document import Document\n+from google.cloud.language.document import Encoding\ndiff --git a/gcloud/language/client.py b/google/cloud/language/client.py\nsimilarity index 96%\nrename from gcloud/language/client.py\nrename to google/cloud/language/client.py\n--- a/gcloud/language/client.py\n+++ b/google/cloud/language/client.py\n@@ -15,9 +15,9 @@\n \"\"\"Basic client for Google Cloud Natural Language API.\"\"\"\n \n \n-from gcloud import client as client_module\n-from gcloud.language.connection import Connection\n-from gcloud.language.document import Document\n+from google.cloud import client as client_module\n+from google.cloud.language.connection import Connection\n+from google.cloud.language.document import Document\n \n \n class Client(client_module.Client):\ndiff --git a/gcloud/language/connection.py b/google/cloud/language/connection.py\nsimilarity index 95%\nrename from gcloud/language/connection.py\nrename to google/cloud/language/connection.py\n--- a/gcloud/language/connection.py\n+++ b/google/cloud/language/connection.py\n@@ -14,7 +14,7 @@\n \n \"\"\"Basic connection for Google Cloud Natural Language API.\"\"\"\n \n-from gcloud import connection as base_connection\n+from google.cloud import connection as base_connection\n \n \n class Connection(base_connection.JSONConnection):\ndiff --git a/gcloud/language/document.py b/google/cloud/language/document.py\nsimilarity index 97%\nrename from gcloud/language/document.py\nrename to google/cloud/language/document.py\n--- a/gcloud/language/document.py\n+++ b/google/cloud/language/document.py\n@@ -19,10 +19,10 @@\n \n import collections\n \n-from gcloud.language.entity import Entity\n-from gcloud.language.sentiment import Sentiment\n-from gcloud.language.syntax import Sentence\n-from gcloud.language.syntax import Token\n+from google.cloud.language.entity import Entity\n+from google.cloud.language.sentiment import Sentiment\n+from google.cloud.language.syntax import Sentence\n+from google.cloud.language.syntax import Token\n \n \n DEFAULT_LANGUAGE = 'en-US'\n@@ -72,7 +72,7 @@ class Document(object):\n stored on the document or referred to in a Google Cloud Storage\n object.\n \n- :type client: :class:`~gcloud.language.client.Client`\n+ :type client: :class:`~google.cloud.language.client.Client`\n :param client: A client which holds credentials and other\n configuration.\n \ndiff --git a/gcloud/language/entity.py b/google/cloud/language/entity.py\nsimilarity index 100%\nrename from gcloud/language/entity.py\nrename to google/cloud/language/entity.py\ndiff --git a/gcloud/language/sentiment.py b/google/cloud/language/sentiment.py\nsimilarity index 100%\nrename from gcloud/language/sentiment.py\nrename to google/cloud/language/sentiment.py\ndiff --git a/gcloud/language/syntax.py b/google/cloud/language/syntax.py\nsimilarity index 100%\nrename from gcloud/language/syntax.py\nrename to google/cloud/language/syntax.py\ndiff --git a/gcloud/logging/__init__.py b/google/cloud/logging/__init__.py\nsimilarity index 78%\nrename from gcloud/logging/__init__.py\nrename to google/cloud/logging/__init__.py\n--- a/gcloud/logging/__init__.py\n+++ b/google/cloud/logging/__init__.py\n@@ -14,10 +14,9 @@\n \n \"\"\"Google Stackdriver Logging API wrapper.\"\"\"\n \n-from gcloud.logging.client import Client\n-from gcloud.logging.connection import Connection\n-\n-\n-SCOPE = Connection.SCOPE\n-ASCENDING = 'timestamp asc'\n-DESCENDING = 'timestamp desc'\n+try:\n+ import pkg_resources\n+ pkg_resources.declare_namespace(__name__)\n+except ImportError:\n+ import pkgutil\n+ __path__ = pkgutil.extend_path(__path__, __name__)\ndiff --git a/gcloud/logging/_gax.py b/google/cloud/logging/_gax.py\nsimilarity index 97%\nrename from gcloud/logging/_gax.py\nrename to google/cloud/logging/_gax.py\n--- a/gcloud/logging/_gax.py\n+++ b/google/cloud/logging/_gax.py\n@@ -26,11 +26,13 @@\n from google.protobuf.json_format import Parse\n from grpc import StatusCode\n \n-from gcloud._helpers import _datetime_to_pb_timestamp\n-from gcloud._helpers import _pb_timestamp_to_rfc3339\n-from gcloud._helpers import exc_to_code\n-from gcloud.exceptions import Conflict\n-from gcloud.exceptions import NotFound\n+# pylint: disable=ungrouped-imports\n+from google.cloud._helpers import _datetime_to_pb_timestamp\n+from google.cloud._helpers import _pb_timestamp_to_rfc3339\n+from google.cloud._helpers import exc_to_code\n+from google.cloud.exceptions import Conflict\n+from google.cloud.exceptions import NotFound\n+# pylint: enable=ungrouped-imports\n \n \n class _LoggingAPI(object):\n@@ -56,8 +58,8 @@ def list_entries(self, projects, filter_='', order_by='',\n https://cloud.google.com/logging/docs/view/advanced_filters\n \n :type order_by: str\n- :param order_by: One of :data:`gcloud.logging.ASCENDING` or\n- :data:`gcloud.logging.DESCENDING`.\n+ :param order_by: One of :data:`~google.cloud.logging.client.ASCENDING`\n+ or :data:`~google.cloud.logging.client.DESCENDING`.\n \n :type page_size: int\n :param page_size: maximum number of entries to return, If not passed,\ndiff --git a/gcloud/logging/client.py b/google/cloud/logging/client.py\nsimilarity index 85%\nrename from gcloud/logging/client.py\nrename to google/cloud/logging/client.py\n--- a/gcloud/logging/client.py\n+++ b/google/cloud/logging/client.py\n@@ -23,9 +23,9 @@\n LoggingServiceV2Api as GeneratedLoggingAPI)\n from google.cloud.logging.v2.metrics_service_v2_api import (\n MetricsServiceV2Api as GeneratedMetricsAPI)\n- from gcloud.logging._gax import _LoggingAPI as GAXLoggingAPI\n- from gcloud.logging._gax import _MetricsAPI as GAXMetricsAPI\n- from gcloud.logging._gax import _SinksAPI as GAXSinksAPI\n+ from google.cloud.logging._gax import _LoggingAPI as GAXLoggingAPI\n+ from google.cloud.logging._gax import _MetricsAPI as GAXMetricsAPI\n+ from google.cloud.logging._gax import _SinksAPI as GAXSinksAPI\n except ImportError: # pragma: NO COVER\n _HAVE_GAX = False\n GeneratedLoggingAPI = GAXLoggingAPI = None\n@@ -34,21 +34,25 @@\n else:\n _HAVE_GAX = True\n \n-from gcloud.client import JSONClient\n-from gcloud.logging.connection import Connection\n-from gcloud.logging.connection import _LoggingAPI as JSONLoggingAPI\n-from gcloud.logging.connection import _MetricsAPI as JSONMetricsAPI\n-from gcloud.logging.connection import _SinksAPI as JSONSinksAPI\n-from gcloud.logging.entries import ProtobufEntry\n-from gcloud.logging.entries import StructEntry\n-from gcloud.logging.entries import TextEntry\n-from gcloud.logging.logger import Logger\n-from gcloud.logging.metric import Metric\n-from gcloud.logging.sink import Sink\n+from google.cloud.client import JSONClient\n+from google.cloud.logging.connection import Connection\n+from google.cloud.logging.connection import _LoggingAPI as JSONLoggingAPI\n+from google.cloud.logging.connection import _MetricsAPI as JSONMetricsAPI\n+from google.cloud.logging.connection import _SinksAPI as JSONSinksAPI\n+from google.cloud.logging.entries import ProtobufEntry\n+from google.cloud.logging.entries import StructEntry\n+from google.cloud.logging.entries import TextEntry\n+from google.cloud.logging.logger import Logger\n+from google.cloud.logging.metric import Metric\n+from google.cloud.logging.sink import Sink\n \n \n-_DISABLE_GAX = os.getenv('GCLOUD_DISABLE_GAX', False)\n+_DISABLE_GAX = os.getenv('GOOGLE_CLOUD_DISABLE_GAX', False)\n _USE_GAX = _HAVE_GAX and not _DISABLE_GAX\n+ASCENDING = 'timestamp asc'\n+\"\"\"Query string to order by ascending timestamps.\"\"\"\n+DESCENDING = 'timestamp desc'\n+\"\"\"Query string to order by decending timestamps.\"\"\"\n \n \n class Client(JSONClient):\n@@ -127,7 +131,7 @@ def logger(self, name):\n :type name: str\n :param name: the name of the logger to be constructed.\n \n- :rtype: :class:`gcloud.logging.logger.Logger`\n+ :rtype: :class:`google.cloud.logging.logger.Logger`\n :returns: Logger created with the current client.\n \"\"\"\n return Logger(name, client=self)\n@@ -143,9 +147,9 @@ def _entry_from_resource(self, resource, loggers):\n passed, the entry will have a newly-created logger.\n \n :rtype: One of:\n- :class:`gcloud.logging.entries.TextEntry`,\n- :class:`gcloud.logging.entries.StructEntry`,\n- :class:`gcloud.logging.entries.ProtobufEntry`\n+ :class:`google.cloud.logging.entries.TextEntry`,\n+ :class:`google.cloud.logging.entries.StructEntry`,\n+ :class:`google.cloud.logging.entries.ProtobufEntry`\n :returns: the entry instance, constructed via the resource\n \"\"\"\n if 'textPayload' in resource:\n@@ -172,8 +176,8 @@ def list_entries(self, projects=None, filter_=None, order_by=None,\n https://cloud.google.com/logging/docs/view/advanced_filters\n \n :type order_by: str\n- :param order_by: One of :data:`gcloud.logging.ASCENDING` or\n- :data:`gcloud.logging.DESCENDING`.\n+ :param order_by: One of :data:`~google.cloud.logging.client.ASCENDING`\n+ or :data:`~google.cloud.logging.client.DESCENDING`.\n \n :type page_size: int\n :param page_size: maximum number of entries to return, If not passed,\n@@ -185,7 +189,7 @@ def list_entries(self, projects=None, filter_=None, order_by=None,\n entries.\n \n :rtype: tuple, (list, str)\n- :returns: list of :class:`gcloud.logging.entry.TextEntry`, plus a\n+ :returns: list of :class:`google.cloud.logging.entry.TextEntry`, plus a\n \"next page token\" string: if not None, indicates that\n more entries can be retrieved with another call (pass that\n value as ``page_token``).\n@@ -219,7 +223,7 @@ def sink(self, name, filter_=None, destination=None):\n already exist, to be refreshed via\n :meth:`Sink.reload`.\n \n- :rtype: :class:`gcloud.logging.sink.Sink`\n+ :rtype: :class:`google.cloud.logging.sink.Sink`\n :returns: Sink created with the current client.\n \"\"\"\n return Sink(name, filter_, destination, client=self)\n@@ -240,7 +244,7 @@ def list_sinks(self, page_size=None, page_token=None):\n sinks.\n \n :rtype: tuple, (list, str)\n- :returns: list of :class:`gcloud.logging.sink.Sink`, plus a\n+ :returns: list of :class:`google.cloud.logging.sink.Sink`, plus a\n \"next page token\" string: if not None, indicates that\n more sinks can be retrieved with another call (pass that\n value as ``page_token``).\n@@ -268,7 +272,7 @@ def metric(self, name, filter_=None, description=''):\n If not passed, the instance should already exist,\n to be refreshed via :meth:`Metric.reload`.\n \n- :rtype: :class:`gcloud.logging.metric.Metric`\n+ :rtype: :class:`google.cloud.logging.metric.Metric`\n :returns: Metric created with the current client.\n \"\"\"\n return Metric(name, filter_, client=self, description=description)\n@@ -289,7 +293,7 @@ def list_metrics(self, page_size=None, page_token=None):\n metrics.\n \n :rtype: tuple, (list, str)\n- :returns: list of :class:`gcloud.logging.metric.Metric`, plus a\n+ :returns: list of :class:`google.cloud.logging.metric.Metric`, plus a\n \"next page token\" string: if not None, indicates that\n more metrics can be retrieved with another call (pass that\n value as ``page_token``).\ndiff --git a/gcloud/logging/connection.py b/google/cloud/logging/connection.py\nsimilarity index 97%\nrename from gcloud/logging/connection.py\nrename to google/cloud/logging/connection.py\n--- a/gcloud/logging/connection.py\n+++ b/google/cloud/logging/connection.py\n@@ -14,7 +14,7 @@\n \n \"\"\"Create / interact with Stackdriver Logging connections.\"\"\"\n \n-from gcloud import connection as base_connection\n+from google.cloud import connection as base_connection\n \n \n class Connection(base_connection.JSONConnection):\n@@ -55,7 +55,7 @@ class _LoggingAPI(object):\n https://cloud.google.com/logging/docs/api/ref_v2beta1/rest/v2beta1/entries\n https://cloud.google.com/logging/docs/api/ref_v2beta1/rest/v2beta1/projects.logs\n \n- :type connection: :class:`gcloud.logging.connection.Connection`\n+ :type connection: :class:`google.cloud.logging.connection.Connection`\n :param connection: the connection used to make API requests.\n \"\"\"\n def __init__(self, connection):\n@@ -77,8 +77,8 @@ def list_entries(self, projects, filter_=None, order_by=None,\n https://cloud.google.com/logging/docs/view/advanced_filters\n \n :type order_by: str\n- :param order_by: One of :data:`gcloud.logging.ASCENDING` or\n- :data:`gcloud.logging.DESCENDING`.\n+ :param order_by: One of :data:`~google.cloud.logging.client.ASCENDING`\n+ or :data:`~google.cloud.logging.client.DESCENDING`.\n \n :type page_size: int\n :param page_size: maximum number of entries to return, If not passed,\n@@ -171,7 +171,7 @@ class _SinksAPI(object):\n See:\n https://cloud.google.com/logging/docs/api/ref_v2beta1/rest/v2beta1/projects.sinks\n \n- :type connection: :class:`gcloud.logging.connection.Connection`\n+ :type connection: :class:`google.cloud.logging.connection.Connection`\n :param connection: the connection used to make API requests.\n \"\"\"\n def __init__(self, connection):\n@@ -310,7 +310,7 @@ class _MetricsAPI(object):\n See:\n https://cloud.google.com/logging/docs/api/ref_v2beta1/rest/v2beta1/projects.metrics\n \n- :type connection: :class:`gcloud.logging.connection.Connection`\n+ :type connection: :class:`google.cloud.logging.connection.Connection`\n :param connection: the connection used to make API requests.\n \"\"\"\n def __init__(self, connection):\ndiff --git a/gcloud/logging/entries.py b/google/cloud/logging/entries.py\nsimilarity index 94%\nrename from gcloud/logging/entries.py\nrename to google/cloud/logging/entries.py\n--- a/gcloud/logging/entries.py\n+++ b/google/cloud/logging/entries.py\n@@ -19,8 +19,8 @@\n \n from google.protobuf.json_format import Parse\n \n-from gcloud._helpers import _name_from_project_path\n-from gcloud._helpers import _rfc3339_nanos_to_datetime\n+from google.cloud._helpers import _name_from_project_path\n+from google.cloud._helpers import _rfc3339_nanos_to_datetime\n \n \n _LOGGER_TEMPLATE = re.compile(r\"\"\"\n@@ -53,7 +53,7 @@ class _BaseEntry(object):\n :param payload: The payload passed as ``textPayload``, ``jsonPayload``,\n or ``protoPayload``.\n \n- :type logger: :class:`gcloud.logging.logger.Logger`\n+ :type logger: :class:`google.cloud.logging.logger.Logger`\n :param logger: the logger used to write the entry.\n \n :type insert_id: text, or :class:`NoneType`\n@@ -90,7 +90,7 @@ def from_api_repr(cls, resource, client, loggers=None):\n :param resource: text entry resource representation returned from\n the API\n \n- :type client: :class:`gcloud.logging.client.Client`\n+ :type client: :class:`google.cloud.logging.client.Client`\n :param client: Client which holds credentials and project\n configuration.\n \n@@ -98,7 +98,7 @@ def from_api_repr(cls, resource, client, loggers=None):\n :param loggers: A mapping of logger fullnames -> loggers. If not\n passed, the entry will have a newly-created logger.\n \n- :rtype: :class:`gcloud.logging.entries.TextEntry`\n+ :rtype: :class:`google.cloud.logging.entries.TextEntry`\n :returns: Text entry parsed from ``resource``.\n \"\"\"\n if loggers is None:\ndiff --git a/gcloud/logging/handlers/__init__.py b/google/cloud/logging/handlers/__init__.py\nsimilarity index 83%\nrename from gcloud/logging/handlers/__init__.py\nrename to google/cloud/logging/handlers/__init__.py\n--- a/gcloud/logging/handlers/__init__.py\n+++ b/google/cloud/logging/handlers/__init__.py\n@@ -14,5 +14,5 @@\n \n \"\"\"Python :mod:`logging` handlers for Google Cloud Logging.\"\"\"\n \n-from gcloud.logging.handlers.handlers import CloudLoggingHandler\n-from gcloud.logging.handlers.handlers import setup_logging\n+from google.cloud.logging.handlers.handlers import CloudLoggingHandler\n+from google.cloud.logging.handlers.handlers import setup_logging\ndiff --git a/gcloud/logging/handlers/handlers.py b/google/cloud/logging/handlers/handlers.py\nsimilarity index 86%\nrename from gcloud/logging/handlers/handlers.py\nrename to google/cloud/logging/handlers/handlers.py\n--- a/gcloud/logging/handlers/handlers.py\n+++ b/google/cloud/logging/handlers/handlers.py\n@@ -16,11 +16,11 @@\n \n import logging\n \n-from gcloud.logging.handlers.transports import BackgroundThreadTransport\n+from google.cloud.logging.handlers.transports import BackgroundThreadTransport\n \n \n EXCLUDE_LOGGER_DEFAULTS = (\n- 'gcloud',\n+ 'google.cloud',\n 'oauth2client'\n )\n \n@@ -37,9 +37,9 @@ class CloudLoggingHandler(logging.StreamHandler):\n which means each logging statement that uses this handler will require\n an API call.\n \n- :type client: :class:`gcloud.logging.client`\n- :param client: the authenticated gcloud logging client for this handler\n- to use\n+ :type client: :class:`google.cloud.logging.client`\n+ :param client: the authenticated Google Cloud Logging client for this\n+ handler to use\n \n :type name: str\n :param name: the name of the custom log in Stackdriver Logging. Defaults\n@@ -57,10 +57,10 @@ class CloudLoggingHandler(logging.StreamHandler):\n \n .. doctest::\n \n- import gcloud.logging\n- from gcloud.logging.handlers import CloudLoggingHandler\n+ import google.cloud.logging\n+ from google.cloud.logging.handlers import CloudLoggingHandler\n \n- client = gcloud.logging.Client()\n+ client = google.cloud.logging.Client()\n handler = CloudLoggingHandler(client)\n \n cloud_logger = logging.getLogger('cloudLogger')\n@@ -112,12 +112,12 @@ def setup_logging(handler, excluded_loggers=EXCLUDE_LOGGER_DEFAULTS):\n .. doctest::\n \n import logging\n- import gcloud.logging\n- from gcloud.logging.handlers import CloudLoggingHandler\n+ import google.cloud.logging\n+ from google.cloud.logging.handlers import CloudLoggingHandler\n \n- client = gcloud.logging.Client()\n+ client = google.cloud.logging.Client()\n handler = CloudLoggingHandler(client)\n- gcloud.logging.setup_logging(handler)\n+ google.cloud.logging.setup_logging(handler)\n logging.getLogger().setLevel(logging.DEBUG)\n \n logging.error('bad news') # API call\ndiff --git a/gcloud/logging/handlers/transports/__init__.py b/google/cloud/logging/handlers/transports/__init__.py\nsimilarity index 74%\nrename from gcloud/logging/handlers/transports/__init__.py\nrename to google/cloud/logging/handlers/transports/__init__.py\n--- a/gcloud/logging/handlers/transports/__init__.py\n+++ b/google/cloud/logging/handlers/transports/__init__.py\n@@ -16,11 +16,11 @@\n \n Currently two options are provided, a synchronous transport that makes\n an API call for each log statement, and an asynchronous handler that\n-sends the API using a :class:`~gcloud.logging.logger.Batch` object in\n+sends the API using a :class:`~google.cloud.logging.logger.Batch` object in\n the background.\n \"\"\"\n \n-from gcloud.logging.handlers.transports.base import Transport\n-from gcloud.logging.handlers.transports.sync import SyncTransport\n-from gcloud.logging.handlers.transports.background_thread import (\n+from google.cloud.logging.handlers.transports.base import Transport\n+from google.cloud.logging.handlers.transports.sync import SyncTransport\n+from google.cloud.logging.handlers.transports.background_thread import (\n BackgroundThreadTransport)\ndiff --git a/gcloud/logging/handlers/transports/background_thread.py b/google/cloud/logging/handlers/transports/background_thread.py\nsimilarity index 96%\nrename from gcloud/logging/handlers/transports/background_thread.py\nrename to google/cloud/logging/handlers/transports/background_thread.py\n--- a/gcloud/logging/handlers/transports/background_thread.py\n+++ b/google/cloud/logging/handlers/transports/background_thread.py\n@@ -21,8 +21,8 @@\n import copy\n import threading\n \n-from gcloud.logging import Client\n-from gcloud.logging.handlers.transports.base import Transport\n+from google.cloud.logging.client import Client\n+from google.cloud.logging.handlers.transports.base import Transport\n \n \n class _Worker(object):\n@@ -97,7 +97,7 @@ def _start(self):\n self._entries_condition.acquire()\n self._thread = threading.Thread(\n target=self._run,\n- name='gcloud.logging.handlers.transport.Worker')\n+ name='google.cloud.logging.handlers.transport.Worker')\n self._thread.setDaemon(True)\n self._thread.start()\n finally:\ndiff --git a/gcloud/logging/handlers/transports/base.py b/google/cloud/logging/handlers/transports/base.py\nsimilarity index 95%\nrename from gcloud/logging/handlers/transports/base.py\nrename to google/cloud/logging/handlers/transports/base.py\n--- a/gcloud/logging/handlers/transports/base.py\n+++ b/google/cloud/logging/handlers/transports/base.py\n@@ -16,7 +16,7 @@\n \n \n class Transport(object):\n- \"\"\"Base class for ``gcloud`` logging handler transports.\n+ \"\"\"Base class for Google Cloud Logging handler transports.\n \n Subclasses of :class:`Transport` must have constructors that accept a\n client and name object, and must override :meth:`send`.\ndiff --git a/gcloud/logging/handlers/transports/sync.py b/google/cloud/logging/handlers/transports/sync.py\nsimilarity index 95%\nrename from gcloud/logging/handlers/transports/sync.py\nrename to google/cloud/logging/handlers/transports/sync.py\n--- a/gcloud/logging/handlers/transports/sync.py\n+++ b/google/cloud/logging/handlers/transports/sync.py\n@@ -17,7 +17,7 @@\n Logs directly to the the Stackdriver Logging API with a synchronous call.\n \"\"\"\n \n-from gcloud.logging.handlers.transports.base import Transport\n+from google.cloud.logging.handlers.transports.base import Transport\n \n \n class SyncTransport(Transport):\ndiff --git a/gcloud/logging/logger.py b/google/cloud/logging/logger.py\nsimilarity index 92%\nrename from gcloud/logging/logger.py\nrename to google/cloud/logging/logger.py\n--- a/gcloud/logging/logger.py\n+++ b/google/cloud/logging/logger.py\n@@ -28,7 +28,7 @@ class Logger(object):\n :type name: string\n :param name: the name of the logger\n \n- :type client: :class:`gcloud.logging.client.Client`\n+ :type client: :class:`google.cloud.logging.client.Client`\n :param client: A client which holds credentials and project configuration\n for the logger (which requires a project).\n \n@@ -64,11 +64,12 @@ def path(self):\n def _require_client(self, client):\n \"\"\"Check client or verify over-ride.\n \n- :type client: :class:`gcloud.logging.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.logging.client.Client` or\n+ ``NoneType``\n :param client: the client to use. If not passed, falls back to the\n ``client`` stored on the current logger.\n \n- :rtype: :class:`gcloud.logging.client.Client`\n+ :rtype: :class:`google.cloud.logging.client.Client`\n :returns: The client passed in or the currently bound client.\n \"\"\"\n if client is None:\n@@ -78,7 +79,8 @@ def _require_client(self, client):\n def batch(self, client=None):\n \"\"\"Return a batch to use as a context manager.\n \n- :type client: :class:`gcloud.logging.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.logging.client.Client` or\n+ ``NoneType``\n :param client: the client to use. If not passed, falls back to the\n ``client`` stored on the current topic.\n \n@@ -165,7 +167,8 @@ def log_text(self, text, client=None, labels=None, insert_id=None,\n :type text: text\n :param text: the log message.\n \n- :type client: :class:`gcloud.logging.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.logging.client.Client` or\n+ ``NoneType``\n :param client: the client to use. If not passed, falls back to the\n ``client`` stored on the current logger.\n \n@@ -198,7 +201,8 @@ def log_struct(self, info, client=None, labels=None, insert_id=None,\n :type info: dict\n :param info: the log entry information\n \n- :type client: :class:`gcloud.logging.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.logging.client.Client` or\n+ ``NoneType``\n :param client: the client to use. If not passed, falls back to the\n ``client`` stored on the current logger.\n \n@@ -231,7 +235,8 @@ def log_proto(self, message, client=None, labels=None, insert_id=None,\n :type message: Protobuf message\n :param message: the message to be logged\n \n- :type client: :class:`gcloud.logging.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.logging.client.Client` or\n+ ``NoneType``\n :param client: the client to use. If not passed, falls back to the\n ``client`` stored on the current logger.\n \n@@ -260,7 +265,8 @@ def delete(self, client=None):\n See:\n https://cloud.google.com/logging/docs/api/ref_v2beta1/rest/v2beta1/projects.logs/delete\n \n- :type client: :class:`gcloud.logging.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.logging.client.Client` or\n+ ``NoneType``\n :param client: the client to use. If not passed, falls back to the\n ``client`` stored on the current logger.\n \"\"\"\n@@ -283,8 +289,8 @@ def list_entries(self, projects=None, filter_=None, order_by=None,\n https://cloud.google.com/logging/docs/view/advanced_filters\n \n :type order_by: string\n- :param order_by: One of :data:`gcloud.logging.ASCENDING` or\n- :data:`gcloud.logging.DESCENDING`.\n+ :param order_by: One of :data:`~google.cloud.logging.client.ASCENDING`\n+ or :data:`~google.cloud.logging.client.DESCENDING`.\n \n :type page_size: int\n :param page_size: maximum number of entries to return, If not passed,\n@@ -296,7 +302,7 @@ def list_entries(self, projects=None, filter_=None, order_by=None,\n entries.\n \n :rtype: tuple, (list, str)\n- :returns: list of :class:`gcloud.logging.entry.TextEntry`, plus a\n+ :returns: list of :class:`google.cloud.logging.entry.TextEntry`, plus a\n \"next page token\" string: if not None, indicates that\n more entries can be retrieved with another call (pass that\n value as ``page_token``).\n@@ -316,10 +322,10 @@ class Batch(object):\n \n Helper returned by :meth:`Logger.batch`\n \n- :type logger: :class:`gcloud.logging.logger.Logger`\n+ :type logger: :class:`google.cloud.logging.logger.Logger`\n :param logger: the logger to which entries will be logged.\n \n- :type client: :class:`gcloud.logging.client.Client`\n+ :type client: :class:`google.cloud.logging.client.Client`\n :param client: The client to use.\n \"\"\"\n def __init__(self, logger, client):\n@@ -406,7 +412,8 @@ def log_proto(self, message, labels=None, insert_id=None, severity=None,\n def commit(self, client=None):\n \"\"\"Send saved log entries as a single API call.\n \n- :type client: :class:`gcloud.logging.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.logging.client.Client` or\n+ ``NoneType``\n :param client: the client to use. If not passed, falls back to the\n ``client`` stored on the current batch.\n \"\"\"\ndiff --git a/gcloud/logging/metric.py b/google/cloud/logging/metric.py\nsimilarity index 86%\nrename from gcloud/logging/metric.py\nrename to google/cloud/logging/metric.py\n--- a/gcloud/logging/metric.py\n+++ b/google/cloud/logging/metric.py\n@@ -14,7 +14,7 @@\n \n \"\"\"Define Stackdriver Logging API Metrics.\"\"\"\n \n-from gcloud.exceptions import NotFound\n+from google.cloud.exceptions import NotFound\n \n \n class Metric(object):\n@@ -31,7 +31,7 @@ class Metric(object):\n tracked by the metric. If not passed, the instance should\n already exist, to be refreshed via :meth:`reload`.\n \n- :type client: :class:`gcloud.logging.client.Client`\n+ :type client: :class:`google.cloud.logging.client.Client`\n :param client: A client which holds credentials and project configuration\n for the metric (which requires a project).\n \n@@ -71,11 +71,11 @@ def from_api_repr(cls, resource, client):\n :type resource: dict\n :param resource: metric resource representation returned from the API\n \n- :type client: :class:`gcloud.logging.client.Client`\n+ :type client: :class:`google.cloud.logging.client.Client`\n :param client: Client which holds credentials and project\n configuration for the metric.\n \n- :rtype: :class:`gcloud.logging.metric.Metric`\n+ :rtype: :class:`google.cloud.logging.metric.Metric`\n :returns: Metric parsed from ``resource``.\n \"\"\"\n metric_name = resource['name']\n@@ -87,11 +87,12 @@ def from_api_repr(cls, resource, client):\n def _require_client(self, client):\n \"\"\"Check client or verify over-ride.\n \n- :type client: :class:`gcloud.logging.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.logging.client.Client` or\n+ ``NoneType``\n :param client: the client to use. If not passed, falls back to the\n ``client`` stored on the current metric.\n \n- :rtype: :class:`gcloud.logging.client.Client`\n+ :rtype: :class:`google.cloud.logging.client.Client`\n :returns: The client passed in or the currently bound client.\n \"\"\"\n if client is None:\n@@ -104,7 +105,8 @@ def create(self, client=None):\n See:\n https://cloud.google.com/logging/docs/api/ref_v2beta1/rest/v2beta1/projects.metrics/create\n \n- :type client: :class:`gcloud.logging.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.logging.client.Client` or\n+ ``NoneType``\n :param client: the client to use. If not passed, falls back to the\n ``client`` stored on the current metric.\n \"\"\"\n@@ -118,7 +120,8 @@ def exists(self, client=None):\n See\n https://cloud.google.com/logging/docs/api/ref_v2beta1/rest/v2beta1/projects.metrics/get\n \n- :type client: :class:`gcloud.logging.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.logging.client.Client` or\n+ ``NoneType``\n :param client: the client to use. If not passed, falls back to the\n ``client`` stored on the current metric.\n \n@@ -140,7 +143,8 @@ def reload(self, client=None):\n See\n https://cloud.google.com/logging/docs/api/ref_v2beta1/rest/v2beta1/projects.metrics/get\n \n- :type client: :class:`gcloud.logging.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.logging.client.Client` or\n+ ``NoneType``\n :param client: the client to use. If not passed, falls back to the\n ``client`` stored on the current metric.\n \"\"\"\n@@ -155,7 +159,8 @@ def update(self, client=None):\n See\n https://cloud.google.com/logging/docs/api/ref_v2beta1/rest/v2beta1/projects.metrics/update\n \n- :type client: :class:`gcloud.logging.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.logging.client.Client` or\n+ ``NoneType``\n :param client: the client to use. If not passed, falls back to the\n ``client`` stored on the current metric.\n \"\"\"\n@@ -169,7 +174,8 @@ def delete(self, client=None):\n See\n https://cloud.google.com/logging/docs/api/ref_v2beta1/rest/v2beta1/projects.metrics/delete\n \n- :type client: :class:`gcloud.logging.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.logging.client.Client` or\n+ ``NoneType``\n :param client: the client to use. If not passed, falls back to the\n ``client`` stored on the current metric.\n \"\"\"\ndiff --git a/gcloud/logging/sink.py b/google/cloud/logging/sink.py\nsimilarity index 86%\nrename from gcloud/logging/sink.py\nrename to google/cloud/logging/sink.py\n--- a/gcloud/logging/sink.py\n+++ b/google/cloud/logging/sink.py\n@@ -14,7 +14,7 @@\n \n \"\"\"Define Stackdriver Logging API Sinks.\"\"\"\n \n-from gcloud.exceptions import NotFound\n+from google.cloud.exceptions import NotFound\n \n \n class Sink(object):\n@@ -36,7 +36,7 @@ class Sink(object):\n If not passed, the instance should already exist, to\n be refreshed via :meth:`reload`.\n \n- :type client: :class:`gcloud.logging.client.Client`\n+ :type client: :class:`google.cloud.logging.client.Client`\n :param client: A client which holds credentials and project configuration\n for the sink (which requires a project).\n \"\"\"\n@@ -73,11 +73,11 @@ def from_api_repr(cls, resource, client):\n :type resource: dict\n :param resource: sink resource representation returned from the API\n \n- :type client: :class:`gcloud.logging.client.Client`\n+ :type client: :class:`google.cloud.logging.client.Client`\n :param client: Client which holds credentials and project\n configuration for the sink.\n \n- :rtype: :class:`gcloud.logging.sink.Sink`\n+ :rtype: :class:`google.cloud.logging.sink.Sink`\n :returns: Sink parsed from ``resource``.\n :raises: :class:`ValueError` if ``client`` is not ``None`` and the\n project from the resource does not agree with the project\n@@ -91,11 +91,12 @@ def from_api_repr(cls, resource, client):\n def _require_client(self, client):\n \"\"\"Check client or verify over-ride.\n \n- :type client: :class:`gcloud.logging.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.logging.client.Client` or\n+ ``NoneType``\n :param client: the client to use. If not passed, falls back to the\n ``client`` stored on the current sink.\n \n- :rtype: :class:`gcloud.logging.client.Client`\n+ :rtype: :class:`google.cloud.logging.client.Client`\n :returns: The client passed in or the currently bound client.\n \"\"\"\n if client is None:\n@@ -108,7 +109,8 @@ def create(self, client=None):\n See:\n https://cloud.google.com/logging/docs/api/ref_v2beta1/rest/v2beta1/projects.sinks/create\n \n- :type client: :class:`gcloud.logging.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.logging.client.Client` or\n+ ``NoneType``\n :param client: the client to use. If not passed, falls back to the\n ``client`` stored on the current sink.\n \"\"\"\n@@ -122,7 +124,8 @@ def exists(self, client=None):\n See\n https://cloud.google.com/logging/docs/api/ref_v2beta1/rest/v2beta1/projects.sinks/get\n \n- :type client: :class:`gcloud.logging.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.logging.client.Client` or\n+ ``NoneType``\n :param client: the client to use. If not passed, falls back to the\n ``client`` stored on the current sink.\n \n@@ -144,7 +147,8 @@ def reload(self, client=None):\n See\n https://cloud.google.com/logging/docs/api/ref_v2beta1/rest/v2beta1/projects.sinks/get\n \n- :type client: :class:`gcloud.logging.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.logging.client.Client` or\n+ ``NoneType``\n :param client: the client to use. If not passed, falls back to the\n ``client`` stored on the current sink.\n \"\"\"\n@@ -159,7 +163,8 @@ def update(self, client=None):\n See\n https://cloud.google.com/logging/docs/api/ref_v2beta1/rest/v2beta1/projects.sinks/update\n \n- :type client: :class:`gcloud.logging.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.logging.client.Client` or\n+ ``NoneType``\n :param client: the client to use. If not passed, falls back to the\n ``client`` stored on the current sink.\n \"\"\"\n@@ -173,7 +178,8 @@ def delete(self, client=None):\n See\n https://cloud.google.com/logging/docs/api/ref_v2beta1/rest/v2beta1/projects.sinks/delete\n \n- :type client: :class:`gcloud.logging.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.logging.client.Client` or\n+ ``NoneType``\n :param client: the client to use. If not passed, falls back to the\n ``client`` stored on the current sink.\n \"\"\"\ndiff --git a/gcloud/monitoring/__init__.py b/google/cloud/monitoring/__init__.py\nsimilarity index 52%\nrename from gcloud/monitoring/__init__.py\nrename to google/cloud/monitoring/__init__.py\n--- a/gcloud/monitoring/__init__.py\n+++ b/google/cloud/monitoring/__init__.py\n@@ -14,22 +14,22 @@\n \n \"\"\"Google Stackdriver Monitoring API wrapper.\"\"\"\n \n-from gcloud.monitoring.client import Client\n-from gcloud.monitoring.connection import Connection\n-from gcloud.monitoring.group import Group\n-from gcloud.monitoring.label import LabelDescriptor\n-from gcloud.monitoring.label import LabelValueType\n-from gcloud.monitoring.metric import Metric\n-from gcloud.monitoring.metric import MetricDescriptor\n-from gcloud.monitoring.metric import MetricKind\n-from gcloud.monitoring.metric import ValueType\n-from gcloud.monitoring.query import Aligner\n-from gcloud.monitoring.query import Query\n-from gcloud.monitoring.query import Reducer\n-from gcloud.monitoring.resource import Resource\n-from gcloud.monitoring.resource import ResourceDescriptor\n-from gcloud.monitoring.timeseries import Point\n-from gcloud.monitoring.timeseries import TimeSeries\n+from google.cloud.monitoring.client import Client\n+from google.cloud.monitoring.connection import Connection\n+from google.cloud.monitoring.group import Group\n+from google.cloud.monitoring.label import LabelDescriptor\n+from google.cloud.monitoring.label import LabelValueType\n+from google.cloud.monitoring.metric import Metric\n+from google.cloud.monitoring.metric import MetricDescriptor\n+from google.cloud.monitoring.metric import MetricKind\n+from google.cloud.monitoring.metric import ValueType\n+from google.cloud.monitoring.query import Aligner\n+from google.cloud.monitoring.query import Query\n+from google.cloud.monitoring.query import Reducer\n+from google.cloud.monitoring.resource import Resource\n+from google.cloud.monitoring.resource import ResourceDescriptor\n+from google.cloud.monitoring.timeseries import Point\n+from google.cloud.monitoring.timeseries import TimeSeries\n \n __all__ = (\n 'Client',\ndiff --git a/gcloud/monitoring/_dataframe.py b/google/cloud/monitoring/_dataframe.py\nsimilarity index 98%\nrename from gcloud/monitoring/_dataframe.py\nrename to google/cloud/monitoring/_dataframe.py\n--- a/gcloud/monitoring/_dataframe.py\n+++ b/google/cloud/monitoring/_dataframe.py\n@@ -30,7 +30,7 @@ def _build_dataframe(time_series_iterable,\n \"\"\"Build a :mod:`pandas` dataframe out of time series.\n \n :type time_series_iterable:\n- iterable over :class:`~gcloud.monitoring.timeseries.TimeSeries`\n+ iterable over :class:`~google.cloud.monitoring.timeseries.TimeSeries`\n :param time_series_iterable:\n An iterable (e.g., a query object) yielding time series.\n \ndiff --git a/gcloud/monitoring/client.py b/google/cloud/monitoring/client.py\nsimilarity index 82%\nrename from gcloud/monitoring/client.py\nrename to google/cloud/monitoring/client.py\n--- a/gcloud/monitoring/client.py\n+++ b/google/cloud/monitoring/client.py\n@@ -16,7 +16,7 @@\n \n Example::\n \n- >>> from gcloud import monitoring\n+ >>> from google.cloud import monitoring\n >>> client = monitoring.Client()\n >>> query = client.query(minutes=5)\n >>> print(query.as_dataframe()) # Requires pandas.\n@@ -30,18 +30,18 @@\n \n import datetime\n \n-from gcloud.client import JSONClient\n-from gcloud.monitoring.connection import Connection\n-from gcloud.monitoring.group import Group\n-from gcloud.monitoring.metric import Metric\n-from gcloud.monitoring.metric import MetricDescriptor\n-from gcloud.monitoring.metric import MetricKind\n-from gcloud.monitoring.metric import ValueType\n-from gcloud.monitoring.query import Query\n-from gcloud.monitoring.resource import Resource\n-from gcloud.monitoring.resource import ResourceDescriptor\n-from gcloud.monitoring.timeseries import Point\n-from gcloud.monitoring.timeseries import TimeSeries\n+from google.cloud.client import JSONClient\n+from google.cloud.monitoring.connection import Connection\n+from google.cloud.monitoring.group import Group\n+from google.cloud.monitoring.metric import Metric\n+from google.cloud.monitoring.metric import MetricDescriptor\n+from google.cloud.monitoring.metric import MetricKind\n+from google.cloud.monitoring.metric import ValueType\n+from google.cloud.monitoring.query import Query\n+from google.cloud.monitoring.resource import Resource\n+from google.cloud.monitoring.resource import ResourceDescriptor\n+from google.cloud.monitoring.timeseries import Point\n+from google.cloud.monitoring.timeseries import TimeSeries\n \n _UTCNOW = datetime.datetime.utcnow # To be replaced by tests.\n \n@@ -82,7 +82,7 @@ def query(self,\n :type metric_type: string\n :param metric_type: The metric type name. The default value is\n :data:`Query.DEFAULT_METRIC_TYPE\n- `,\n+ `,\n but please note that this default value is provided only for\n demonstration purposes and is subject to change. See the\n `supported metrics`_.\n@@ -98,7 +98,7 @@ def query(self,\n \n It is also allowed to omit the end time and duration here,\n in which case\n- :meth:`~gcloud.monitoring.query.Query.select_interval`\n+ :meth:`~google.cloud.monitoring.query.Query.select_interval`\n must be called before the query is executed.\n \n :type days: integer\n@@ -110,13 +110,13 @@ def query(self,\n :type minutes: integer\n :param minutes: The number of minutes in the time interval.\n \n- :rtype: :class:`~gcloud.monitoring.query.Query`\n+ :rtype: :class:`~google.cloud.monitoring.query.Query`\n :returns: The query object.\n \n :raises: :exc:`ValueError` if ``end_time`` is specified but\n ``days``, ``hours``, and ``minutes`` are all zero.\n If you really want to specify a point in time, use\n- :meth:`~gcloud.monitoring.query.Query.select_interval`.\n+ :meth:`~google.cloud.monitoring.query.Query.select_interval`.\n \n .. _supported metrics: https://cloud.google.com/monitoring/api/metrics\n \"\"\"\n@@ -133,7 +133,8 @@ def metric_descriptor(self, type_,\n Metric descriptors specify the schema for a particular metric type.\n \n This factory method is used most often in conjunction with the metric\n- descriptor :meth:`~gcloud.monitoring.metric.MetricDescriptor.create`\n+ descriptor\n+ :meth:`~google.cloud.monitoring.metric.MetricDescriptor.create`\n method to define custom metrics::\n \n >>> descriptor = client.metric_descriptor(\n@@ -166,7 +167,7 @@ def metric_descriptor(self, type_,\n The kind of measurement. It must be one of\n :data:`MetricKind.GAUGE`, :data:`MetricKind.DELTA`,\n or :data:`MetricKind.CUMULATIVE`.\n- See :class:`~gcloud.monitoring.metric.MetricKind`.\n+ See :class:`~google.cloud.monitoring.metric.MetricKind`.\n \n :type value_type: string\n :param value_type:\n@@ -176,7 +177,8 @@ def metric_descriptor(self, type_,\n or :data:`ValueType.DISTRIBUTION`.\n See :class:`ValueType`.\n \n- :type labels: list of :class:`~gcloud.monitoring.label.LabelDescriptor`\n+ :type labels:\n+ list of :class:`~google.cloud.monitoring.label.LabelDescriptor`\n :param labels:\n A sequence of zero or more label descriptors specifying the labels\n used to identify a specific instance of this metric.\n@@ -207,10 +209,10 @@ def metric_descriptor(self, type_,\n def metric(type_, labels):\n \"\"\"Factory for constructing metric objects.\n \n- :class:`~gcloud.monitoring.metric.Metric` objects are typically\n+ :class:`~google.cloud.monitoring.metric.Metric` objects are typically\n created to write custom metric values. The type should match the\n metric type specified in the\n- :class:`~gcloud.monitoring.metric.MetricDescriptor` used to\n+ :class:`~google.cloud.monitoring.metric.MetricDescriptor` used to\n create the custom metric::\n \n >>> metric = client.metric('custom.googleapis.com/my_metric',\n@@ -224,9 +226,9 @@ def metric(type_, labels):\n :type labels: dict\n :param labels: A mapping from label names to values for all labels\n enumerated in the associated\n- :class:`~gcloud.monitoring.metric.MetricDescriptor`.\n+ :class:`~google.cloud.monitoring.metric.MetricDescriptor`.\n \n- :rtype: :class:`~gcloud.monitoring.metric.Metric`\n+ :rtype: :class:`~google.cloud.monitoring.metric.Metric`\n :returns: The metric object.\n \"\"\"\n return Metric(type=type_, labels=labels)\n@@ -236,9 +238,9 @@ def resource(type_, labels):\n \"\"\"Factory for constructing monitored resource objects.\n \n A monitored resource object (\n- :class:`~gcloud.monitoring.resource.Resource`) is\n+ :class:`~google.cloud.monitoring.resource.Resource`) is\n typically used to create a\n- :class:`~gcloud.monitoring.timeseries.TimeSeries` object.\n+ :class:`~google.cloud.monitoring.timeseries.TimeSeries` object.\n \n For a list of possible monitored resource types and their associated\n labels, see:\n@@ -251,11 +253,11 @@ def resource(type_, labels):\n :type labels: dict\n :param labels: A mapping from label names to values for all labels\n enumerated in the associated\n- :class:`~gcloud.monitoring.resource.ResourceDescriptor`,\n+ :class:`~google.cloud.monitoring.resource.ResourceDescriptor`,\n except that ``project_id`` can and should be omitted\n when writing time series data.\n \n- :rtype: :class:`~gcloud.monitoring.resource.Resource`\n+ :rtype: :class:`~google.cloud.monitoring.resource.Resource`\n :returns: A monitored resource object.\n \"\"\"\n return Resource(type_, labels)\n@@ -267,9 +269,9 @@ def time_series(metric, resource, value,\n \n .. note::\n \n- While :class:`~gcloud.monitoring.timeseries.TimeSeries` objects\n- returned by the API typically have multiple data points,\n- :class:`~gcloud.monitoring.timeseries.TimeSeries` objects\n+ While :class:`~google.cloud.monitoring.timeseries.TimeSeries`\n+ objects returned by the API typically have multiple data points,\n+ :class:`~google.cloud.monitoring.timeseries.TimeSeries` objects\n sent to the API must have at most one point.\n \n For example::\n@@ -281,17 +283,17 @@ def time_series(metric, resource, value,\n \n https://cloud.google.com/monitoring/api/ref_v3/rest/v3/TimeSeries\n \n- :type metric: :class:`~gcloud.monitoring.metric.Metric`\n- :param metric: A :class:`~gcloud.monitoring.metric.Metric` object.\n+ :type metric: :class:`~google.cloud.monitoring.metric.Metric`\n+ :param metric: A :class:`~google.cloud.monitoring.metric.Metric`.\n \n- :type resource: :class:`~gcloud.monitoring.resource.Resource`\n- :param resource: A :class:`~gcloud.monitoring.resource.Resource`\n+ :type resource: :class:`~google.cloud.monitoring.resource.Resource`\n+ :param resource: A :class:`~google.cloud.monitoring.resource.Resource`\n object.\n \n :type value: bool, int, string, or float\n :param value:\n The value of the data point to create for the\n- :class:`~gcloud.monitoring.timeseries.TimeSeries`.\n+ :class:`~google.cloud.monitoring.timeseries.TimeSeries`.\n \n .. note::\n \n@@ -314,7 +316,7 @@ def time_series(metric, resource, value,\n Defaults to None. If the start time is unspecified,\n the API interprets the start time to be the same as the end time.\n \n- :rtype: :class:`~gcloud.monitoring.timeseries.TimeSeries`\n+ :rtype: :class:`~google.cloud.monitoring.timeseries.TimeSeries`\n :returns: A time series object.\n \"\"\"\n if end_time is None:\n@@ -334,11 +336,11 @@ def fetch_metric_descriptor(self, metric_type):\n :type metric_type: string\n :param metric_type: The metric type name.\n \n- :rtype: :class:`~gcloud.monitoring.metric.MetricDescriptor`\n+ :rtype: :class:`~google.cloud.monitoring.metric.MetricDescriptor`\n :returns: The metric descriptor instance.\n \n- :raises: :class:`gcloud.exceptions.NotFound` if the metric descriptor\n- is not found.\n+ :raises: :class:`google.cloud.exceptions.NotFound` if the metric\n+ descriptor is not found.\n \"\"\"\n return MetricDescriptor._fetch(self, metric_type)\n \n@@ -364,7 +366,8 @@ def list_metric_descriptors(self, filter_string=None, type_prefix=None):\n metric types. This adds ``metric.type = starts_with(\"\")``\n to the filter.\n \n- :rtype: list of :class:`~gcloud.monitoring.metric.MetricDescriptor`\n+ :rtype:\n+ list of :class:`~google.cloud.monitoring.metric.MetricDescriptor`\n :returns: A list of metric descriptor instances.\n \n .. _filter documentation:\n@@ -383,11 +386,11 @@ def fetch_resource_descriptor(self, resource_type):\n :type resource_type: string\n :param resource_type: The resource type name.\n \n- :rtype: :class:`~gcloud.monitoring.resource.ResourceDescriptor`\n+ :rtype: :class:`~google.cloud.monitoring.resource.ResourceDescriptor`\n :returns: The resource descriptor instance.\n \n- :raises: :class:`gcloud.exceptions.NotFound` if the resource descriptor\n- is not found.\n+ :raises: :class:`google.cloud.exceptions.NotFound` if the resource\n+ descriptor is not found.\n \"\"\"\n return ResourceDescriptor._fetch(self, resource_type)\n \n@@ -404,7 +407,8 @@ def list_resource_descriptors(self, filter_string=None):\n An optional filter expression describing the resource descriptors\n to be returned. See the `filter documentation`_.\n \n- :rtype: list of :class:`~gcloud.monitoring.resource.ResourceDescriptor`\n+ :rtype: list of\n+ :class:`~google.cloud.monitoring.resource.ResourceDescriptor`\n :returns: A list of resource descriptor instances.\n \n .. _filter documentation:\n@@ -465,16 +469,17 @@ def fetch_group(self, group_id):\n \n >>> try:\n >>> group = client.fetch_group('1234')\n- >>> except gcloud.exceptions.NotFound:\n+ >>> except google.cloud.exceptions.NotFound:\n >>> print('That group does not exist!')\n \n :type group_id: string\n :param group_id: The ID of the group.\n \n- :rtype: :class:`~gcloud.monitoring.group.Group`\n+ :rtype: :class:`~google.cloud.monitoring.group.Group`\n :returns: The group instance.\n \n- :raises: :class:`gcloud.exceptions.NotFound` if the group is not found.\n+ :raises: :class:`google.cloud.exceptions.NotFound` if the group\n+ is not found.\n \"\"\"\n return Group._fetch(self, group_id)\n \n@@ -486,7 +491,7 @@ def list_groups(self):\n >>> for group in client.list_groups():\n ... print((group.display_name, group.name))\n \n- :rtype: list of :class:`~gcloud.monitoring.group.Group`\n+ :rtype: list of :class:`~google.cloud.monitoring.group.Group`\n :returns: A list of group instances.\n \"\"\"\n return Group._list(self)\ndiff --git a/gcloud/monitoring/connection.py b/google/cloud/monitoring/connection.py\nsimilarity index 97%\nrename from gcloud/monitoring/connection.py\nrename to google/cloud/monitoring/connection.py\n--- a/gcloud/monitoring/connection.py\n+++ b/google/cloud/monitoring/connection.py\n@@ -14,7 +14,7 @@\n \n \"\"\"Create / interact with Stackdriver Monitoring connections.\"\"\"\n \n-from gcloud import connection as base_connection\n+from google.cloud import connection as base_connection\n \n \n class Connection(base_connection.JSONConnection):\ndiff --git a/gcloud/monitoring/group.py b/google/cloud/monitoring/group.py\nsimilarity index 94%\nrename from gcloud/monitoring/group.py\nrename to google/cloud/monitoring/group.py\n--- a/gcloud/monitoring/group.py\n+++ b/google/cloud/monitoring/group.py\n@@ -21,10 +21,10 @@\n \n import re\n \n-from gcloud._helpers import _datetime_to_rfc3339\n-from gcloud._helpers import _name_from_project_path\n-from gcloud.exceptions import NotFound\n-from gcloud.monitoring.resource import Resource\n+from google.cloud._helpers import _datetime_to_rfc3339\n+from google.cloud._helpers import _name_from_project_path\n+from google.cloud.exceptions import NotFound\n+from google.cloud.monitoring.resource import Resource\n \n \n _GROUP_TEMPLATE = re.compile(r\"\"\"\n@@ -73,7 +73,7 @@ def _group_name_from_id(project, group_id):\n class Group(object):\n \"\"\"A dynamic collection of monitored resources.\n \n- :type client: :class:`gcloud.monitoring.client.Client`\n+ :type client: :class:`google.cloud.monitoring.client.Client`\n :param client: A client for operating on the metric descriptor.\n \n :type group_id: string or None\n@@ -242,7 +242,7 @@ def list_children(self):\n Returns groups whose parent_name field contains the group name. If no\n groups have this parent, the results are empty.\n \n- :rtype: list of :class:`~gcloud.monitoring.group.Group`\n+ :rtype: list of :class:`~google.cloud.monitoring.group.Group`\n :returns: A list of group instances.\n \"\"\"\n return self._list(self.client, children_of_group=self.name)\n@@ -254,7 +254,7 @@ def list_ancestors(self):\n and ending with the most distant ancestor. If the specified group has\n no immediate parent, the results are empty.\n \n- :rtype: list of :class:`~gcloud.monitoring.group.Group`\n+ :rtype: list of :class:`~google.cloud.monitoring.group.Group`\n :returns: A list of group instances.\n \"\"\"\n return self._list(self.client, ancestors_of_group=self.name)\n@@ -265,7 +265,7 @@ def list_descendants(self):\n This returns a superset of the results returned by the :meth:`children`\n method, and includes children-of-children, and so forth.\n \n- :rtype: list of :class:`~gcloud.monitoring.group.Group`\n+ :rtype: list of :class:`~google.cloud.monitoring.group.Group`\n :returns: A list of group instances.\n \"\"\"\n return self._list(self.client, descendants_of_group=self.name)\n@@ -313,7 +313,7 @@ def list_members(self, filter_string=None, end_time=None, start_time=None):\n The start time (exclusive) of the time interval for which results\n should be returned, as a datetime object.\n \n- :rtype: list of :class:`~gcloud.monitoring.resource.Resource`\n+ :rtype: list of :class:`~google.cloud.monitoring.resource.Resource`\n :returns: A list of resource instances.\n \n :raises:\n@@ -362,7 +362,7 @@ def list_members(self, filter_string=None, end_time=None, start_time=None):\n def _fetch(cls, client, group_id):\n \"\"\"Fetch a group from the API based on it's ID.\n \n- :type client: :class:`gcloud.monitoring.client.Client`\n+ :type client: :class:`google.cloud.monitoring.client.Client`\n :param client: The client to use.\n \n :type group_id: string\n@@ -371,7 +371,7 @@ def _fetch(cls, client, group_id):\n :rtype: :class:`Group`\n :returns: The group instance.\n \n- :raises: :class:`gcloud.exceptions.NotFound` if the group\n+ :raises: :class:`google.cloud.exceptions.NotFound` if the group\n is not found.\n \"\"\"\n new_group = cls(client, group_id)\n@@ -383,7 +383,7 @@ def _list(cls, client, children_of_group=None, ancestors_of_group=None,\n descendants_of_group=None):\n \"\"\"Lists all groups in the project.\n \n- :type client: :class:`gcloud.monitoring.client.Client`\n+ :type client: :class:`google.cloud.monitoring.client.Client`\n :param client: The client to use.\n \n :type children_of_group: string or None\n@@ -402,7 +402,7 @@ def _list(cls, client, children_of_group=None, ancestors_of_group=None,\n of the results returned by the children_of_group filter, and\n includes children-of-children, and so forth.\n \n- :rtype: list of :class:`~gcloud.monitoring.group.Group`\n+ :rtype: list of :class:`~google.cloud.monitoring.group.Group`\n :returns: A list of group instances.\n \"\"\"\n path = '/projects/%s/groups/' % (client.project,)\n@@ -438,7 +438,7 @@ def _list(cls, client, children_of_group=None, ancestors_of_group=None,\n def _from_dict(cls, client, info):\n \"\"\"Constructs a Group instance from the parsed JSON representation.\n \n- :type client: :class:`gcloud.monitoring.client.Client`\n+ :type client: :class:`google.cloud.monitoring.client.Client`\n :param client: A client to be included in the returned object.\n \n :type info: dict\ndiff --git a/gcloud/monitoring/label.py b/google/cloud/monitoring/label.py\nsimilarity index 100%\nrename from gcloud/monitoring/label.py\nrename to google/cloud/monitoring/label.py\ndiff --git a/gcloud/monitoring/metric.py b/google/cloud/monitoring/metric.py\nsimilarity index 93%\nrename from gcloud/monitoring/metric.py\nrename to google/cloud/monitoring/metric.py\n--- a/gcloud/monitoring/metric.py\n+++ b/google/cloud/monitoring/metric.py\n@@ -21,7 +21,7 @@\n \n import collections\n \n-from gcloud.monitoring.label import LabelDescriptor\n+from google.cloud.monitoring.label import LabelDescriptor\n \n \n class MetricKind(object):\n@@ -62,10 +62,10 @@ class MetricDescriptor(object):\n \"\"\"Specification of a metric type and its schema.\n \n The preferred way to construct a metric descriptor object is using the\n- :meth:`~gcloud.monitoring.client.Client.metric_descriptor` factory method\n- of the :class:`~gcloud.monitoring.client.Client` class.\n+ :meth:`~google.cloud.monitoring.client.Client.metric_descriptor` factory\n+ method of the :class:`~google.cloud.monitoring.client.Client` class.\n \n- :type client: :class:`gcloud.monitoring.client.Client`\n+ :type client: :class:`google.cloud.monitoring.client.Client`\n :param client: A client for operating on the metric descriptor.\n \n :type type_: string\n@@ -87,7 +87,8 @@ class MetricDescriptor(object):\n or :data:`ValueType.DISTRIBUTION`.\n See :class:`ValueType`.\n \n- :type labels: list of :class:`~gcloud.monitoring.label.LabelDescriptor`\n+ :type labels:\n+ list of :class:`~google.cloud.monitoring.label.LabelDescriptor`\n :param labels:\n A sequence of zero or more label descriptors specifying the labels\n used to identify a specific instance of this metric.\n@@ -172,7 +173,7 @@ def delete(self):\n def _fetch(cls, client, metric_type):\n \"\"\"Look up a metric descriptor by type.\n \n- :type client: :class:`gcloud.monitoring.client.Client`\n+ :type client: :class:`google.cloud.monitoring.client.Client`\n :param client: The client to use.\n \n :type metric_type: string\n@@ -181,8 +182,8 @@ def _fetch(cls, client, metric_type):\n :rtype: :class:`MetricDescriptor`\n :returns: The metric descriptor instance.\n \n- :raises: :class:`gcloud.exceptions.NotFound` if the metric descriptor\n- is not found.\n+ :raises: :class:`google.cloud.exceptions.NotFound` if the metric\n+ descriptor is not found.\n \"\"\"\n path = '/projects/{project}/metricDescriptors/{type}'.format(\n project=client.project,\n@@ -194,7 +195,7 @@ def _fetch(cls, client, metric_type):\n def _list(cls, client, filter_string=None, type_prefix=None):\n \"\"\"List all metric descriptors for the project.\n \n- :type client: :class:`gcloud.monitoring.client.Client`\n+ :type client: :class:`google.cloud.monitoring.client.Client`\n :param client: The client to use.\n \n :type filter_string: string or None\n@@ -250,7 +251,7 @@ def _list(cls, client, filter_string=None, type_prefix=None):\n def _from_dict(cls, client, info):\n \"\"\"Construct a metric descriptor from the parsed JSON representation.\n \n- :type client: :class:`gcloud.monitoring.client.Client`\n+ :type client: :class:`google.cloud.monitoring.client.Client`\n :param client: A client to be included in the returned object.\n \n :type info: dict\n@@ -320,8 +321,8 @@ class Metric(collections.namedtuple('Metric', 'type labels')):\n \"\"\"A specific metric identified by specifying values for all labels.\n \n The preferred way to construct a metric object is using the\n- :meth:`~gcloud.monitoring.client.Client.metric` factory method\n- of the :class:`~gcloud.monitoring.client.Client` class.\n+ :meth:`~google.cloud.monitoring.client.Client.metric` factory method\n+ of the :class:`~google.cloud.monitoring.client.Client` class.\n \n :type type: string\n :param type: The metric type name.\ndiff --git a/gcloud/monitoring/query.py b/google/cloud/monitoring/query.py\nsimilarity index 97%\nrename from gcloud/monitoring/query.py\nrename to google/cloud/monitoring/query.py\n--- a/gcloud/monitoring/query.py\n+++ b/google/cloud/monitoring/query.py\n@@ -25,9 +25,9 @@\n \n import six\n \n-from gcloud._helpers import _datetime_to_rfc3339\n-from gcloud.monitoring._dataframe import _build_dataframe\n-from gcloud.monitoring.timeseries import TimeSeries\n+from google.cloud._helpers import _datetime_to_rfc3339\n+from google.cloud.monitoring._dataframe import _build_dataframe\n+from google.cloud.monitoring.timeseries import TimeSeries\n \n _UTCNOW = datetime.datetime.utcnow # To be replaced by tests.\n \n@@ -72,16 +72,16 @@ class Query(object):\n \"\"\"Query object for retrieving metric data.\n \n The preferred way to construct a query object is using the\n- :meth:`~gcloud.monitoring.client.Client.query` method\n- of the :class:`~gcloud.monitoring.client.Client` class.\n+ :meth:`~google.cloud.monitoring.client.Client.query` method\n+ of the :class:`~google.cloud.monitoring.client.Client` class.\n \n- :type client: :class:`gcloud.monitoring.client.Client`\n+ :type client: :class:`google.cloud.monitoring.client.Client`\n :param client: The client to use.\n \n :type metric_type: string\n :param metric_type: The metric type name. The default value is\n :data:`Query.DEFAULT_METRIC_TYPE\n- `,\n+ `,\n but please note that this default value is provided only for\n demonstration purposes and is subject to change. See the\n `supported metrics`_.\n@@ -97,7 +97,7 @@ class Query(object):\n \n It is also allowed to omit the end time and duration here,\n in which case\n- :meth:`~gcloud.monitoring.query.Query.select_interval`\n+ :meth:`~google.cloud.monitoring.query.Query.select_interval`\n must be called before the query is executed.\n \n :type days: integer\n@@ -112,7 +112,7 @@ class Query(object):\n :raises: :exc:`ValueError` if ``end_time`` is specified but\n ``days``, ``hours``, and ``minutes`` are all zero.\n If you really want to specify a point in time, use\n- :meth:`~gcloud.monitoring.query.Query.select_interval`.\n+ :meth:`~google.cloud.monitoring.query.Query.select_interval`.\n \n .. _supported metrics: https://cloud.google.com/monitoring/api/metrics\n \"\"\"\n@@ -430,7 +430,7 @@ def iter(self, headers_only=False, page_size=None):\n \"\"\"Yield all time series objects selected by the query.\n \n The generator returned iterates over\n- :class:`~gcloud.monitoring.timeseries.TimeSeries` objects\n+ :class:`~google.cloud.monitoring.timeseries.TimeSeries` objects\n containing points ordered from oldest to newest.\n \n Note that the :class:`Query` object itself is an iterable, such that\n@@ -508,7 +508,7 @@ def _build_query_params(self, headers_only=False,\n :type headers_only: boolean\n :param headers_only:\n Whether to omit the point data from the\n- :class:`~gcloud.monitoring.timeseries.TimeSeries` objects.\n+ :class:`~google.cloud.monitoring.timeseries.TimeSeries` objects.\n \n :type page_size: integer or None\n :param page_size: A limit on the number of points to return per page.\ndiff --git a/gcloud/monitoring/resource.py b/google/cloud/monitoring/resource.py\nsimilarity index 91%\nrename from gcloud/monitoring/resource.py\nrename to google/cloud/monitoring/resource.py\n--- a/gcloud/monitoring/resource.py\n+++ b/google/cloud/monitoring/resource.py\n@@ -22,7 +22,7 @@\n \n import collections\n \n-from gcloud.monitoring.label import LabelDescriptor\n+from google.cloud.monitoring.label import LabelDescriptor\n \n \n class ResourceDescriptor(object):\n@@ -45,7 +45,8 @@ class ResourceDescriptor(object):\n :param description:\n A detailed description that might be used in documentation.\n \n- :type labels: list of :class:`~gcloud.monitoring.label.LabelDescriptor`\n+ :type labels:\n+ list of :class:`~google.cloud.monitoring.label.LabelDescriptor`\n :param labels:\n A sequence of label descriptors specifying the labels used\n to identify a specific instance of this monitored resource.\n@@ -62,7 +63,7 @@ def __init__(self, name, type_, display_name, description, labels):\n def _fetch(cls, client, resource_type):\n \"\"\"Look up a monitored resource descriptor by type.\n \n- :type client: :class:`gcloud.monitoring.client.Client`\n+ :type client: :class:`google.cloud.monitoring.client.Client`\n :param client: The client to use.\n \n :type resource_type: string\n@@ -71,8 +72,8 @@ def _fetch(cls, client, resource_type):\n :rtype: :class:`ResourceDescriptor`\n :returns: The resource descriptor instance.\n \n- :raises: :class:`gcloud.exceptions.NotFound` if the resource descriptor\n- is not found.\n+ :raises: :class:`google.cloud.exceptions.NotFound` if the resource\n+ descriptor is not found.\n \"\"\"\n path = ('/projects/{project}/monitoredResourceDescriptors/{type}'\n .format(project=client.project,\n@@ -84,7 +85,7 @@ def _fetch(cls, client, resource_type):\n def _list(cls, client, filter_string=None):\n \"\"\"List all monitored resource descriptors for the project.\n \n- :type client: :class:`gcloud.monitoring.client.Client`\n+ :type client: :class:`google.cloud.monitoring.client.Client`\n :param client: The client to use.\n \n :type filter_string: string or None\n@@ -159,8 +160,8 @@ class Resource(collections.namedtuple('Resource', 'type labels')):\n \"\"\"A monitored resource identified by specifying values for all labels.\n \n The preferred way to construct a resource object is using the\n- :meth:`~gcloud.monitoring.client.Client.resource` factory method\n- of the :class:`~gcloud.monitoring.client.Client` class.\n+ :meth:`~google.cloud.monitoring.client.Client.resource` factory method\n+ of the :class:`~google.cloud.monitoring.client.Client` class.\n \n :type type: string\n :param type: The resource type name.\ndiff --git a/gcloud/monitoring/timeseries.py b/google/cloud/monitoring/timeseries.py\nsimilarity index 88%\nrename from gcloud/monitoring/timeseries.py\nrename to google/cloud/monitoring/timeseries.py\n--- a/gcloud/monitoring/timeseries.py\n+++ b/google/cloud/monitoring/timeseries.py\n@@ -24,8 +24,8 @@\n \n import collections\n \n-from gcloud.monitoring.metric import Metric\n-from gcloud.monitoring.resource import Resource\n+from google.cloud.monitoring.metric import Metric\n+from google.cloud.monitoring.resource import Resource\n \n \n class TimeSeries(collections.namedtuple(\n@@ -33,28 +33,29 @@ class TimeSeries(collections.namedtuple(\n \"\"\"A single time series of metric values.\n \n The preferred way to construct a\n- :class:`~gcloud.monitoring.timeseries.TimeSeries` object is\n- using the :meth:`~gcloud.monitoring.client.Client.time_series` factory\n- method of the :class:`~gcloud.monitoring.client.Client` class.\n+ :class:`~google.cloud.monitoring.timeseries.TimeSeries` object is\n+ using the :meth:`~google.cloud.monitoring.client.Client.time_series`\n+ factory method of the :class:`~google.cloud.monitoring.client.Client`\n+ class.\n \n- :type metric: :class:`~gcloud.monitoring.metric.Metric`\n+ :type metric: :class:`~google.cloud.monitoring.metric.Metric`\n :param metric: A metric object.\n \n- :type resource: :class:`~gcloud.monitoring.resource.Resource`\n+ :type resource: :class:`~google.cloud.monitoring.resource.Resource`\n :param resource: A resource object.\n \n :type metric_kind: string\n :param metric_kind:\n The kind of measurement: :data:`MetricKind.GAUGE`,\n :data:`MetricKind.DELTA`, or :data:`MetricKind.CUMULATIVE`.\n- See :class:`~gcloud.monitoring.metric.MetricKind`.\n+ See :class:`~google.cloud.monitoring.metric.MetricKind`.\n \n :type value_type: string\n :param value_type:\n The value type of the metric: :data:`ValueType.BOOL`,\n :data:`ValueType.INT64`, :data:`ValueType.DOUBLE`,\n :data:`ValueType.STRING`, or :data:`ValueType.DISTRIBUTION`.\n- See :class:`~gcloud.monitoring.metric.ValueType`.\n+ See :class:`~google.cloud.monitoring.metric.ValueType`.\n \n :type points: list of :class:`Point`\n :param points: A list of point objects.\ndiff --git a/gcloud/operation.py b/google/cloud/operation.py\nsimilarity index 100%\nrename from gcloud/operation.py\nrename to google/cloud/operation.py\ndiff --git a/gcloud/pubsub/__init__.py b/google/cloud/pubsub/__init__.py\nsimilarity index 64%\nrename from gcloud/pubsub/__init__.py\nrename to google/cloud/pubsub/__init__.py\n--- a/gcloud/pubsub/__init__.py\n+++ b/google/cloud/pubsub/__init__.py\n@@ -16,17 +16,16 @@\n \n The main concepts with this API are:\n \n-- :class:`gcloud.pubsub.topic.Topic` represents an endpoint to which messages\n- can be published using the Cloud Storage Pubsub API.\n+- :class:`~google.cloud.pubsub.topic.Topic` represents an endpoint to which\n+ messages can be published using the Cloud Storage Pubsub API.\n \n-- :class:`gcloud.pubsub.subscription.Subscription` represents a named\n+- :class:`~google.cloud.pubsub.subscription.Subscription` represents a named\n subscription (either pull or push) to a topic.\n \"\"\"\n \n-from gcloud.pubsub.client import Client\n-from gcloud.pubsub.connection import Connection\n-from gcloud.pubsub.subscription import Subscription\n-from gcloud.pubsub.topic import Topic\n-\n-\n-SCOPE = Connection.SCOPE\n+try:\n+ import pkg_resources\n+ pkg_resources.declare_namespace(__name__)\n+except ImportError:\n+ import pkgutil\n+ __path__ = pkgutil.extend_path(__path__, __name__)\ndiff --git a/gcloud/pubsub/_gax.py b/google/cloud/pubsub/_gax.py\nsimilarity index 97%\nrename from gcloud/pubsub/_gax.py\nrename to google/cloud/pubsub/_gax.py\n--- a/gcloud/pubsub/_gax.py\n+++ b/google/cloud/pubsub/_gax.py\n@@ -21,10 +21,12 @@\n from google.pubsub.v1.pubsub_pb2 import PushConfig\n from grpc import StatusCode\n \n-from gcloud._helpers import _to_bytes\n-from gcloud._helpers import exc_to_code\n-from gcloud.exceptions import Conflict\n-from gcloud.exceptions import NotFound\n+# pylint: disable=ungrouped-imports\n+from google.cloud._helpers import _to_bytes\n+from google.cloud._helpers import exc_to_code\n+from google.cloud.exceptions import Conflict\n+from google.cloud.exceptions import NotFound\n+# pylint: enable=ungrouped-imports\n \n \n class _PublisherAPI(object):\n@@ -82,7 +84,7 @@ def topic_create(self, topic_path):\n \n :rtype: dict\n :returns: ``Topic`` resource returned from the API.\n- :raises: :exc:`gcloud.exceptions.Conflict` if the topic already\n+ :raises: :exc:`google.cloud.exceptions.Conflict` if the topic already\n exists\n \"\"\"\n try:\n@@ -105,7 +107,7 @@ def topic_get(self, topic_path):\n \n :rtype: dict\n :returns: ``Topic`` resource returned from the API.\n- :raises: :exc:`gcloud.exceptions.NotFound` if the topic does not\n+ :raises: :exc:`google.cloud.exceptions.NotFound` if the topic does not\n exist\n \"\"\"\n try:\n@@ -148,7 +150,7 @@ def topic_publish(self, topic_path, messages):\n \n :rtype: list of string\n :returns: list of opaque IDs for published messages.\n- :raises: :exc:`gcloud.exceptions.NotFound` if the topic does not\n+ :raises: :exc:`google.cloud.exceptions.NotFound` if the topic does not\n exist\n \"\"\"\n options = CallOptions(is_bundling=False)\n@@ -186,7 +188,7 @@ def topic_list_subscriptions(self, topic_path, page_size=0,\n :rtype: list of strings\n :returns: fully-qualified names of subscriptions for the supplied\n topic.\n- :raises: :exc:`gcloud.exceptions.NotFound` if the topic does not\n+ :raises: :exc:`google.cloud.exceptions.NotFound` if the topic does not\n exist\n \"\"\"\n if page_token is None:\ndiff --git a/gcloud/pubsub/_helpers.py b/google/cloud/pubsub/_helpers.py\nsimilarity index 97%\nrename from gcloud/pubsub/_helpers.py\nrename to google/cloud/pubsub/_helpers.py\n--- a/gcloud/pubsub/_helpers.py\n+++ b/google/cloud/pubsub/_helpers.py\n@@ -16,7 +16,7 @@\n \n import re\n \n-from gcloud._helpers import _name_from_project_path\n+from google.cloud._helpers import _name_from_project_path\n \n \n _TOPIC_TEMPLATE = re.compile(r\"\"\"\ndiff --git a/gcloud/pubsub/client.py b/google/cloud/pubsub/client.py\nsimilarity index 89%\nrename from gcloud/pubsub/client.py\nrename to google/cloud/pubsub/client.py\n--- a/gcloud/pubsub/client.py\n+++ b/google/cloud/pubsub/client.py\n@@ -16,13 +16,13 @@\n \n import os\n \n-from gcloud.client import JSONClient\n-from gcloud.pubsub.connection import Connection\n-from gcloud.pubsub.connection import _PublisherAPI as JSONPublisherAPI\n-from gcloud.pubsub.connection import _SubscriberAPI as JSONSubscriberAPI\n-from gcloud.pubsub.connection import _IAMPolicyAPI\n-from gcloud.pubsub.subscription import Subscription\n-from gcloud.pubsub.topic import Topic\n+from google.cloud.client import JSONClient\n+from google.cloud.pubsub.connection import Connection\n+from google.cloud.pubsub.connection import _PublisherAPI as JSONPublisherAPI\n+from google.cloud.pubsub.connection import _SubscriberAPI as JSONSubscriberAPI\n+from google.cloud.pubsub.connection import _IAMPolicyAPI\n+from google.cloud.pubsub.subscription import Subscription\n+from google.cloud.pubsub.topic import Topic\n \n # pylint: disable=ungrouped-imports\n try:\n@@ -30,8 +30,8 @@\n PublisherApi as GeneratedPublisherAPI)\n from google.cloud.pubsub.v1.subscriber_api import (\n SubscriberApi as GeneratedSubscriberAPI)\n- from gcloud.pubsub._gax import _PublisherAPI as GAXPublisherAPI\n- from gcloud.pubsub._gax import _SubscriberAPI as GAXSubscriberAPI\n+ from google.cloud.pubsub._gax import _PublisherAPI as GAXPublisherAPI\n+ from google.cloud.pubsub._gax import _SubscriberAPI as GAXSubscriberAPI\n except ImportError: # pragma: NO COVER\n _HAVE_GAX = False\n GeneratedPublisherAPI = GAXPublisherAPI = None\n@@ -41,7 +41,7 @@\n # pylint: enable=ungrouped-imports\n \n \n-_DISABLE_GAX = os.getenv('GCLOUD_DISABLE_GAX', False)\n+_DISABLE_GAX = os.getenv('GOOGLE_CLOUD_DISABLE_GAX', False)\n _USE_GAX = _HAVE_GAX and not _DISABLE_GAX\n \n \n@@ -120,7 +120,7 @@ def list_topics(self, page_size=None, page_token=None):\n topics.\n \n :rtype: tuple, (list, str)\n- :returns: list of :class:`gcloud.pubsub.topic.Topic`, plus a\n+ :returns: list of :class:`google.cloud.pubsub.topic.Topic`, plus a\n \"next page token\" string: if not None, indicates that\n more topics can be retrieved with another call (pass that\n value as ``page_token``).\n@@ -154,7 +154,7 @@ def list_subscriptions(self, page_size=None, page_token=None):\n topics.\n \n :rtype: tuple, (list, str)\n- :returns: list of :class:`gcloud.pubsub.subscription.Subscription`,\n+ :returns: list of :class:`~.pubsub.subscription.Subscription`,\n plus a \"next page token\" string: if not None, indicates that\n more topics can be retrieved with another call (pass that\n value as ``page_token``).\n@@ -183,7 +183,7 @@ def topic(self, name, timestamp_messages=False):\n :type timestamp_messages: boolean\n :param timestamp_messages: To be passed to ``Topic`` constructor.\n \n- :rtype: :class:`gcloud.pubsub.topic.Topic`\n+ :rtype: :class:`google.cloud.pubsub.topic.Topic`\n :returns: Topic created with the current client.\n \"\"\"\n return Topic(name, client=self, timestamp_messages=timestamp_messages)\ndiff --git a/gcloud/pubsub/connection.py b/google/cloud/pubsub/connection.py\nsimilarity index 99%\nrename from gcloud/pubsub/connection.py\nrename to google/cloud/pubsub/connection.py\n--- a/gcloud/pubsub/connection.py\n+++ b/google/cloud/pubsub/connection.py\n@@ -12,12 +12,12 @@\n # See the License for the specific language governing permissions and\n # limitations under the License.\n \n-\"\"\"Create / interact with gcloud pubsub connections.\"\"\"\n+\"\"\"Create / interact with Google Cloud Pub/Sub connections.\"\"\"\n \n import os\n \n-from gcloud import connection as base_connection\n-from gcloud.environment_vars import PUBSUB_EMULATOR\n+from google.cloud import connection as base_connection\n+from google.cloud.environment_vars import PUBSUB_EMULATOR\n \n \n class Connection(base_connection.JSONConnection):\ndiff --git a/gcloud/pubsub/iam.py b/google/cloud/pubsub/iam.py\nsimilarity index 100%\nrename from gcloud/pubsub/iam.py\nrename to google/cloud/pubsub/iam.py\ndiff --git a/gcloud/pubsub/message.py b/google/cloud/pubsub/message.py\nsimilarity index 98%\nrename from gcloud/pubsub/message.py\nrename to google/cloud/pubsub/message.py\n--- a/gcloud/pubsub/message.py\n+++ b/google/cloud/pubsub/message.py\n@@ -16,7 +16,7 @@\n \n import base64\n \n-from gcloud._helpers import _rfc3339_to_datetime\n+from google.cloud._helpers import _rfc3339_to_datetime\n \n \n class Message(object):\ndiff --git a/gcloud/pubsub/subscription.py b/google/cloud/pubsub/subscription.py\nsimilarity index 88%\nrename from gcloud/pubsub/subscription.py\nrename to google/cloud/pubsub/subscription.py\n--- a/gcloud/pubsub/subscription.py\n+++ b/google/cloud/pubsub/subscription.py\n@@ -14,10 +14,10 @@\n \n \"\"\"Define API Subscriptions.\"\"\"\n \n-from gcloud.exceptions import NotFound\n-from gcloud.pubsub._helpers import topic_name_from_path\n-from gcloud.pubsub.iam import Policy\n-from gcloud.pubsub.message import Message\n+from google.cloud.exceptions import NotFound\n+from google.cloud.pubsub._helpers import topic_name_from_path\n+from google.cloud.pubsub.iam import Policy\n+from google.cloud.pubsub.message import Message\n \n \n class Subscription(object):\n@@ -29,7 +29,7 @@ class Subscription(object):\n :type name: string\n :param name: the name of the subscription.\n \n- :type topic: :class:`gcloud.pubsub.topic.Topic` or ``NoneType``\n+ :type topic: :class:`google.cloud.pubsub.topic.Topic` or ``NoneType``\n :param topic: the topic to which the subscription belongs; if ``None``,\n the subscription's topic has been deleted.\n \n@@ -41,7 +41,8 @@ class Subscription(object):\n :param push_endpoint: URL to which messages will be pushed by the back-end.\n If not set, the application must pull messages.\n \n- :type client: :class:`gcloud.pubsub.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.pubsub.client.Client` or\n+ ``NoneType``\n :param client: the client to use. If not passed, falls back to the\n ``client`` stored on the topic.\n \"\"\"\n@@ -76,7 +77,7 @@ def from_api_repr(cls, resource, client, topics=None):\n :type resource: dict\n :param resource: topic resource representation returned from the API.\n \n- :type client: :class:`gcloud.pubsub.client.Client`\n+ :type client: :class:`google.cloud.pubsub.client.Client`\n :param client: Client which holds credentials and project\n configuration for a topic.\n \n@@ -84,7 +85,7 @@ def from_api_repr(cls, resource, client, topics=None):\n :param topics: A mapping of topic names -> topics. If not passed,\n the subscription will have a newly-created topic.\n \n- :rtype: :class:`gcloud.pubsub.subscription.Subscription`\n+ :rtype: :class:`google.cloud.pubsub.subscription.Subscription`\n :returns: Subscription parsed from ``resource``.\n \"\"\"\n if topics is None:\n@@ -132,7 +133,8 @@ def auto_ack(self, return_immediately=False, max_messages=1, client=None):\n :type max_messages: int\n :param max_messages: passed through to :meth:`Subscription.pull`\n \n- :type client: :class:`gcloud.pubsub.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.pubsub.client.Client` or\n+ ``NoneType``\n :param client: passed through to :meth:`Subscription.pull` and\n :meth:`Subscription.acknowledge`.\n \n@@ -144,12 +146,13 @@ def auto_ack(self, return_immediately=False, max_messages=1, client=None):\n def _require_client(self, client):\n \"\"\"Check client or verify over-ride.\n \n- :type client: :class:`gcloud.pubsub.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.pubsub.client.Client` or\n+ ``NoneType``\n :param client: the client to use. If not passed, falls back to the\n ``client`` stored on the topic of the\n current subscription.\n \n- :rtype: :class:`gcloud.pubsub.client.Client`\n+ :rtype: :class:`google.cloud.pubsub.client.Client`\n :returns: The client passed in or the currently bound client.\n \"\"\"\n if client is None:\n@@ -168,7 +171,8 @@ def create(self, client=None):\n :start-after: [START subscription_create]\n :end-before: [END subscription_create]\n \n- :type client: :class:`gcloud.pubsub.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.pubsub.client.Client` or\n+ ``NoneType``\n :param client: the client to use. If not passed, falls back to the\n ``client`` stored on the current subscription's topic.\n \"\"\"\n@@ -190,7 +194,8 @@ def exists(self, client=None):\n :start-after: [START subscription_exists]\n :end-before: [END subscription_exists]\n \n- :type client: :class:`gcloud.pubsub.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.pubsub.client.Client` or\n+ ``NoneType``\n :param client: the client to use. If not passed, falls back to the\n ``client`` stored on the current subscription's topic.\n \n@@ -218,7 +223,8 @@ def reload(self, client=None):\n :start-after: [START subscription_reload]\n :end-before: [END subscription_reload]\n \n- :type client: :class:`gcloud.pubsub.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.pubsub.client.Client` or\n+ ``NoneType``\n :param client: the client to use. If not passed, falls back to the\n ``client`` stored on the current subscription's topic.\n \"\"\"\n@@ -241,7 +247,8 @@ def delete(self, client=None):\n :start-after: [START subscription_delete]\n :end-before: [END subscription_delete]\n \n- :type client: :class:`gcloud.pubsub.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.pubsub.client.Client` or\n+ ``NoneType``\n :param client: the client to use. If not passed, falls back to the\n ``client`` stored on the current subscription's topic.\n \"\"\"\n@@ -270,7 +277,8 @@ def modify_push_configuration(self, push_endpoint, client=None):\n back-end. If None, the application must pull\n messages.\n \n- :type client: :class:`gcloud.pubsub.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.pubsub.client.Client` or\n+ ``NoneType``\n :param client: the client to use. If not passed, falls back to the\n ``client`` stored on the current subscription's topic.\n \"\"\"\n@@ -300,14 +308,16 @@ def pull(self, return_immediately=False, max_messages=1, client=None):\n :type max_messages: int\n :param max_messages: the maximum number of messages to return.\n \n- :type client: :class:`gcloud.pubsub.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.pubsub.client.Client` or\n+ ``NoneType``\n :param client: the client to use. If not passed, falls back to the\n ``client`` stored on the current subscription's topic.\n \n :rtype: list of (ack_id, message) tuples\n :returns: sequence of tuples: ``ack_id`` is the ID to be used in a\n subsequent call to :meth:`acknowledge`, and ``message``\n- is an instance of :class:`gcloud.pubsub.message.Message`.\n+ is an instance of\n+ :class:`~google.cloud.pubsub.message.Message`.\n \"\"\"\n client = self._require_client(client)\n api = client.subscriber_api\n@@ -331,7 +341,8 @@ def acknowledge(self, ack_ids, client=None):\n :type ack_ids: list of string\n :param ack_ids: ack IDs of messages being acknowledged\n \n- :type client: :class:`gcloud.pubsub.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.pubsub.client.Client` or\n+ ``NoneType``\n :param client: the client to use. If not passed, falls back to the\n ``client`` stored on the current subscription's topic.\n \"\"\"\n@@ -351,7 +362,8 @@ def modify_ack_deadline(self, ack_ids, ack_deadline, client=None):\n :type ack_deadline: int\n :param ack_deadline: new deadline for the message, in seconds\n \n- :type client: :class:`gcloud.pubsub.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.pubsub.client.Client` or\n+ ``NoneType``\n :param client: the client to use. If not passed, falls back to the\n ``client`` stored on the current subscription's topic.\n \"\"\"\n@@ -372,11 +384,12 @@ def get_iam_policy(self, client=None):\n :start-after: [START subscription_get_iam_policy]\n :end-before: [END subscription_get_iam_policy]\n \n- :type client: :class:`gcloud.pubsub.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.pubsub.client.Client` or\n+ ``NoneType``\n :param client: the client to use. If not passed, falls back to the\n ``client`` stored on the current subscription's topic.\n \n- :rtype: :class:`gcloud.pubsub.iam.Policy`\n+ :rtype: :class:`google.cloud.pubsub.iam.Policy`\n :returns: policy created from the resource returned by the\n ``getIamPolicy`` API request.\n \"\"\"\n@@ -397,15 +410,16 @@ def set_iam_policy(self, policy, client=None):\n :start-after: [START subscription_set_iam_policy]\n :end-before: [END subscription_set_iam_policy]\n \n- :type policy: :class:`gcloud.pubsub.iam.Policy`\n+ :type policy: :class:`google.cloud.pubsub.iam.Policy`\n :param policy: the new policy, typically fetched via\n :meth:`get_iam_policy` and updated in place.\n \n- :type client: :class:`gcloud.pubsub.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.pubsub.client.Client` or\n+ ``NoneType``\n :param client: the client to use. If not passed, falls back to the\n ``client`` stored on the current subscription's topic.\n \n- :rtype: :class:`gcloud.pubsub.iam.Policy`\n+ :rtype: :class:`google.cloud.pubsub.iam.Policy`\n :returns: updated policy created from the resource returned by the\n ``setIamPolicy`` API request.\n \"\"\"\n@@ -430,7 +444,8 @@ def check_iam_permissions(self, permissions, client=None):\n :type permissions: list of string\n :param permissions: list of permissions to be tested\n \n- :type client: :class:`gcloud.pubsub.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.pubsub.client.Client` or\n+ ``NoneType``\n :param client: the client to use. If not passed, falls back to the\n ``client`` stored on the current subscription's topic.\n \n@@ -471,7 +486,8 @@ class AutoAck(dict):\n :type max_messages: int\n :param max_messages: passed through to :meth:`Subscription.pull`\n \n- :type client: :class:`gcloud.pubsub.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.pubsub.client.Client` or\n+ ``NoneType``\n :param client: passed through to :meth:`Subscription.pull` and\n :meth:`Subscription.acknowledge`.\n \"\"\"\ndiff --git a/gcloud/pubsub/topic.py b/google/cloud/pubsub/topic.py\nsimilarity index 87%\nrename from gcloud/pubsub/topic.py\nrename to google/cloud/pubsub/topic.py\n--- a/gcloud/pubsub/topic.py\n+++ b/google/cloud/pubsub/topic.py\n@@ -16,13 +16,13 @@\n \n import base64\n \n-from gcloud._helpers import _datetime_to_rfc3339\n-from gcloud._helpers import _NOW\n-from gcloud.exceptions import NotFound\n-from gcloud.pubsub._helpers import subscription_name_from_path\n-from gcloud.pubsub._helpers import topic_name_from_path\n-from gcloud.pubsub.iam import Policy\n-from gcloud.pubsub.subscription import Subscription\n+from google.cloud._helpers import _datetime_to_rfc3339\n+from google.cloud._helpers import _NOW\n+from google.cloud.exceptions import NotFound\n+from google.cloud.pubsub._helpers import subscription_name_from_path\n+from google.cloud.pubsub._helpers import topic_name_from_path\n+from google.cloud.pubsub.iam import Policy\n+from google.cloud.pubsub.subscription import Subscription\n \n \n class Topic(object):\n@@ -36,7 +36,7 @@ class Topic(object):\n :type name: string\n :param name: the name of the topic\n \n- :type client: :class:`gcloud.pubsub.client.Client`\n+ :type client: :class:`google.cloud.pubsub.client.Client`\n :param client: A client which holds credentials and project configuration\n for the topic (which requires a project).\n \n@@ -96,11 +96,11 @@ def from_api_repr(cls, resource, client):\n :type resource: dict\n :param resource: topic resource representation returned from the API\n \n- :type client: :class:`gcloud.pubsub.client.Client`\n+ :type client: :class:`google.cloud.pubsub.client.Client`\n :param client: Client which holds credentials and project\n configuration for the topic.\n \n- :rtype: :class:`gcloud.pubsub.topic.Topic`\n+ :rtype: :class:`google.cloud.pubsub.topic.Topic`\n :returns: Topic parsed from ``resource``.\n :raises: :class:`ValueError` if ``client`` is not ``None`` and the\n project from the resource does not agree with the project\n@@ -122,11 +122,12 @@ def full_name(self):\n def _require_client(self, client):\n \"\"\"Check client or verify over-ride.\n \n- :type client: :class:`gcloud.pubsub.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.pubsub.client.Client` or\n+ ``NoneType``\n :param client: the client to use. If not passed, falls back to the\n ``client`` stored on the current topic.\n \n- :rtype: :class:`gcloud.pubsub.client.Client`\n+ :rtype: :class:`google.cloud.pubsub.client.Client`\n :returns: The client passed in or the currently bound client.\n \"\"\"\n if client is None:\n@@ -145,7 +146,8 @@ def create(self, client=None):\n :start-after: [START topic_create]\n :end-before: [END topic_create]\n \n- :type client: :class:`gcloud.pubsub.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.pubsub.client.Client` or\n+ ``NoneType``\n :param client: the client to use. If not passed, falls back to the\n ``client`` stored on the current topic.\n \"\"\"\n@@ -165,7 +167,8 @@ def exists(self, client=None):\n :start-after: [START topic_exists]\n :end-before: [END topic_exists]\n \n- :type client: :class:`gcloud.pubsub.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.pubsub.client.Client` or\n+ ``NoneType``\n :param client: the client to use. If not passed, falls back to the\n ``client`` stored on the current topic.\n \n@@ -194,7 +197,8 @@ def delete(self, client=None):\n :start-after: [START topic_delete]\n :end-before: [END topic_delete]\n \n- :type client: :class:`gcloud.pubsub.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.pubsub.client.Client` or\n+ ``NoneType``\n :param client: the client to use. If not passed, falls back to the\n ``client`` stored on the current topic.\n \"\"\"\n@@ -233,7 +237,8 @@ def publish(self, message, client=None, **attrs):\n :type message: bytes\n :param message: the message payload\n \n- :type client: :class:`gcloud.pubsub.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.pubsub.client.Client` or\n+ ``NoneType``\n :param client: the client to use. If not passed, falls back to the\n ``client`` stored on the current topic.\n \n@@ -267,7 +272,8 @@ def batch(self, client=None):\n used as a context manager, and only if the block exits without\n raising an exception.\n \n- :type client: :class:`gcloud.pubsub.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.pubsub.client.Client` or\n+ ``NoneType``\n :param client: the client to use. If not passed, falls back to the\n ``client`` stored on the current topic.\n \n@@ -298,12 +304,13 @@ def list_subscriptions(self, page_size=None, page_token=None, client=None):\n passed, the API will return the first page of\n topics.\n \n- :type client: :class:`gcloud.pubsub.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.pubsub.client.Client` or\n+ ``NoneType``\n :param client: the client to use. If not passed, falls back to the\n ``client`` stored on the current topic.\n \n :rtype: tuple, (list, str)\n- :returns: list of :class:`gcloud.pubsub.subscription.Subscription`,\n+ :returns: list of :class:`~.pubsub.subscription.Subscription`,\n plus a \"next page token\" string: if not None, indicates that\n more topics can be retrieved with another call (pass that\n value as ``page_token``).\n@@ -330,11 +337,12 @@ def get_iam_policy(self, client=None):\n :start-after: [START topic_get_iam_policy]\n :end-before: [END topic_get_iam_policy]\n \n- :type client: :class:`gcloud.pubsub.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.pubsub.client.Client` or\n+ ``NoneType``\n :param client: the client to use. If not passed, falls back to the\n ``client`` stored on the current batch.\n \n- :rtype: :class:`gcloud.pubsub.iam.Policy`\n+ :rtype: :class:`google.cloud.pubsub.iam.Policy`\n :returns: policy created from the resource returned by the\n ``getIamPolicy`` API request.\n \"\"\"\n@@ -355,15 +363,16 @@ def set_iam_policy(self, policy, client=None):\n :start-after: [START topic_set_iam_policy]\n :end-before: [END topic_set_iam_policy]\n \n- :type policy: :class:`gcloud.pubsub.iam.Policy`\n+ :type policy: :class:`google.cloud.pubsub.iam.Policy`\n :param policy: the new policy, typically fetched via\n :meth:`get_iam_policy` and updated in place.\n \n- :type client: :class:`gcloud.pubsub.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.pubsub.client.Client` or\n+ ``NoneType``\n :param client: the client to use. If not passed, falls back to the\n ``client`` stored on the current batch.\n \n- :rtype: :class:`gcloud.pubsub.iam.Policy`\n+ :rtype: :class:`google.cloud.pubsub.iam.Policy`\n :returns: updated policy created from the resource returned by the\n ``setIamPolicy`` API request.\n \"\"\"\n@@ -388,7 +397,8 @@ def check_iam_permissions(self, permissions, client=None):\n :type permissions: list of string\n :param permissions: list of permissions to be tested\n \n- :type client: :class:`gcloud.pubsub.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.pubsub.client.Client` or\n+ ``NoneType``\n :param client: the client to use. If not passed, falls back to the\n ``client`` stored on the current batch.\n \n@@ -406,10 +416,10 @@ class Batch(object):\n \n Helper returned by :meth:Topic.batch\n \n- :type topic: :class:`gcloud.pubsub.topic.Topic`\n+ :type topic: :class:`google.cloud.pubsub.topic.Topic`\n :param topic: the topic being published\n \n- :type client: :class:`gcloud.pubsub.client.Client`\n+ :type client: :class:`google.cloud.pubsub.client.Client`\n :param client: The client to use.\n \"\"\"\n def __init__(self, topic, client):\n@@ -445,7 +455,8 @@ def publish(self, message, **attrs):\n def commit(self, client=None):\n \"\"\"Send saved messages as a single API call.\n \n- :type client: :class:`gcloud.pubsub.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.pubsub.client.Client` or\n+ ``NoneType``\n :param client: the client to use. If not passed, falls back to the\n ``client`` stored on the current batch.\n \"\"\"\ndiff --git a/gcloud/resource_manager/__init__.py b/google/cloud/resource_manager/__init__.py\nsimilarity index 79%\nrename from gcloud/resource_manager/__init__.py\nrename to google/cloud/resource_manager/__init__.py\n--- a/gcloud/resource_manager/__init__.py\n+++ b/google/cloud/resource_manager/__init__.py\n@@ -14,9 +14,10 @@\n \n \"\"\"Google Cloud Resource Manager API wrapper.\"\"\"\n \n-from gcloud.resource_manager.client import Client\n-from gcloud.resource_manager.connection import Connection\n-from gcloud.resource_manager.project import Project\n+\n+from google.cloud.resource_manager.client import Client\n+from google.cloud.resource_manager.connection import Connection\n+from google.cloud.resource_manager.project import Project\n \n \n SCOPE = Connection.SCOPE\ndiff --git a/gcloud/resource_manager/client.py b/google/cloud/resource_manager/client.py\nsimilarity index 90%\nrename from gcloud/resource_manager/client.py\nrename to google/cloud/resource_manager/client.py\n--- a/gcloud/resource_manager/client.py\n+++ b/google/cloud/resource_manager/client.py\n@@ -15,10 +15,10 @@\n \"\"\"A Client for interacting with the Resource Manager API.\"\"\"\n \n \n-from gcloud.client import Client as BaseClient\n-from gcloud.iterator import Iterator\n-from gcloud.resource_manager.connection import Connection\n-from gcloud.resource_manager.project import Project\n+from google.cloud.client import Client as BaseClient\n+from google.cloud.iterator import Iterator\n+from google.cloud.resource_manager.connection import Connection\n+from google.cloud.resource_manager.project import Project\n \n \n class Client(BaseClient):\n@@ -30,7 +30,7 @@ class Client(BaseClient):\n \n Automatically get credentials::\n \n- >>> from gcloud import resource_manager\n+ >>> from google.cloud import resource_manager\n >>> client = resource_manager.Client()\n \n :type credentials: :class:`oauth2client.client.OAuth2Credentials` or\n@@ -52,7 +52,7 @@ def new_project(self, project_id, name=None, labels=None):\n \"\"\"Create a :class:`.Project` bound to the current client.\n \n Use :meth:`Project.reload() \\\n- ` to retrieve\n+ ` to retrieve\n project metadata after creating a :class:`.Project` instance.\n \n .. note:\n@@ -81,7 +81,7 @@ def fetch_project(self, project_id):\n .. note::\n \n If the project does not exist, this will raise a\n- :class:`NotFound ` error.\n+ :class:`NotFound ` error.\n \n :type project_id: str\n :param project_id: The ID for this project.\n@@ -98,7 +98,7 @@ def list_projects(self, filter_params=None, page_size=None):\n \n Example::\n \n- >>> from gcloud import resource_manager\n+ >>> from google.cloud import resource_manager\n >>> client = resource_manager.Client()\n >>> for project in client.list_projects():\n ... print project.project_id\n@@ -106,7 +106,7 @@ def list_projects(self, filter_params=None, page_size=None):\n List all projects with label ``'environment'`` set to ``'prod'``\n (filtering by labels)::\n \n- >>> from gcloud import resource_manager\n+ >>> from google.cloud import resource_manager\n >>> client = resource_manager.Client()\n >>> env_filter = {'labels.environment': 'prod'}\n >>> for project in client.list_projects(env_filter):\n@@ -159,10 +159,10 @@ class _ProjectIterator(Iterator):\n \"\"\"An iterator over a list of Project resources.\n \n You shouldn't have to use this directly, but instead should use the\n- helper methods on :class:`gcloud.resource_manager.client.Client`\n+ helper methods on :class:`google.cloud.resource_manager.client.Client`\n objects.\n \n- :type client: :class:`gcloud.resource_manager.client.Client`\n+ :type client: :class:`google.cloud.resource_manager.client.Client`\n :param client: The client to use for making connections.\n \n :type extra_params: dict\ndiff --git a/gcloud/resource_manager/connection.py b/google/cloud/resource_manager/connection.py\nsimilarity index 92%\nrename from gcloud/resource_manager/connection.py\nrename to google/cloud/resource_manager/connection.py\n--- a/gcloud/resource_manager/connection.py\n+++ b/google/cloud/resource_manager/connection.py\n@@ -12,10 +12,10 @@\n # See the License for the specific language governing permissions and\n # limitations under the License.\n \n-\"\"\"Create / interact with gcloud.resource_manager connections.\"\"\"\n+\"\"\"Create / interact with Google Cloud Resource Manager connections.\"\"\"\n \n \n-from gcloud import connection as base_connection\n+from google.cloud import connection as base_connection\n \n \n class Connection(base_connection.JSONConnection):\ndiff --git a/gcloud/resource_manager/project.py b/google/cloud/resource_manager/project.py\nsimilarity index 90%\nrename from gcloud/resource_manager/project.py\nrename to google/cloud/resource_manager/project.py\n--- a/gcloud/resource_manager/project.py\n+++ b/google/cloud/resource_manager/project.py\n@@ -15,7 +15,7 @@\n \"\"\"Utility for managing projects via the Cloud Resource Manager API.\"\"\"\n \n \n-from gcloud.exceptions import NotFound\n+from google.cloud.exceptions import NotFound\n \n \n class Project(object):\n@@ -25,11 +25,11 @@ class Project(object):\n \n A :class:`Project` can also be created via\n :meth:`Client.new_project() \\\n- `\n+ `\n \n To manage labels on a :class:`Project`::\n \n- >>> from gcloud import resource_manager\n+ >>> from google.cloud import resource_manager\n >>> client = resource_manager.Client()\n >>> project = client.new_project('purple-spaceship-123')\n >>> project.labels = {'color': 'purple'}\n@@ -42,7 +42,7 @@ class Project(object):\n :type project_id: string\n :param project_id: The globally unique ID of the project.\n \n- :type client: :class:`gcloud.resource_manager.client.Client`\n+ :type client: :class:`google.cloud.resource_manager.client.Client`\n :param client: The Client used with this project.\n \n :type name: string\n@@ -69,10 +69,10 @@ def from_api_repr(cls, resource, client):\n :type resource: dict\n :param resource: project resource representation returned from the API\n \n- :type client: :class:`gcloud.resource_manager.client.Client`\n+ :type client: :class:`google.cloud.resource_manager.client.Client`\n :param client: The Client used with this project.\n \n- :rtype: :class:`gcloud.resource_manager.project.Project`\n+ :rtype: :class:`google.cloud.resource_manager.project.Project`\n :returns: The project created.\n \"\"\"\n project = cls(project_id=resource['projectId'], client=client)\n@@ -101,12 +101,12 @@ def path(self):\n def _require_client(self, client):\n \"\"\"Check client or verify over-ride.\n \n- :type client: :class:`gcloud.resource_manager.client.Client` or\n+ :type client: :class:`google.cloud.resource_manager.client.Client` or\n ``NoneType``\n :param client: the client to use. If not passed, falls back to the\n ``client`` stored on the current project.\n \n- :rtype: :class:`gcloud.resource_manager.client.Client`\n+ :rtype: :class:`google.cloud.resource_manager.client.Client`\n :returns: The client passed in or the currently bound client.\n \"\"\"\n if client is None:\n@@ -119,7 +119,7 @@ def create(self, client=None):\n See\n https://cloud.google.com/resource-manager/reference/rest/v1beta1/projects/create\n \n- :type client: :class:`gcloud.resource_manager.client.Client` or\n+ :type client: :class:`google.cloud.resource_manager.client.Client` or\n :data:`NoneType `\n :param client: the client to use. If not passed, falls back to\n the client stored on the current project.\n@@ -141,7 +141,7 @@ def reload(self, client=None):\n This method will reload the newest metadata for the project. If you've\n created a new :class:`Project` instance via\n :meth:`Client.new_project() \\\n- `,\n+ `,\n this method will retrieve project metadata.\n \n .. warning::\n@@ -152,7 +152,7 @@ def reload(self, client=None):\n See\n https://cloud.google.com/resource-manager/reference/rest/v1beta1/projects/get\n \n- :type client: :class:`gcloud.resource_manager.client.Client` or\n+ :type client: :class:`google.cloud.resource_manager.client.Client` or\n :data:`NoneType `\n :param client: the client to use. If not passed, falls back to\n the client stored on the current project.\n@@ -170,7 +170,7 @@ def exists(self, client=None):\n See\n https://cloud.google.com/resource-manager/reference/rest/v1beta1/projects/get\n \n- :type client: :class:`gcloud.resource_manager.client.Client` or\n+ :type client: :class:`google.cloud.resource_manager.client.Client` or\n :data:`NoneType `\n :param client: the client to use. If not passed, falls back to\n the client stored on the current project.\n@@ -195,7 +195,7 @@ def update(self, client=None):\n See\n https://cloud.google.com/resource-manager/reference/rest/v1beta1/projects/update\n \n- :type client: :class:`gcloud.resource_manager.client.Client` or\n+ :type client: :class:`google.cloud.resource_manager.client.Client` or\n :data:`NoneType `\n :param client: the client to use. If not passed, falls back to\n the client stored on the current project.\n@@ -219,7 +219,7 @@ def delete(self, client=None, reload_data=False):\n ``DELETE_IN_PROGRESS`` state, which means the deleting has actually\n begun.\n \n- :type client: :class:`gcloud.resource_manager.client.Client` or\n+ :type client: :class:`google.cloud.resource_manager.client.Client` or\n :data:`NoneType `\n :param client: the client to use. If not passed, falls back to\n the client stored on the current project.\n@@ -249,7 +249,7 @@ def undelete(self, client=None, reload_data=False):\n If the project has already reached a status of ``DELETE_IN_PROGRESS``,\n this request will fail and the project cannot be restored.\n \n- :type client: :class:`gcloud.resource_manager.client.Client` or\n+ :type client: :class:`google.cloud.resource_manager.client.Client` or\n :data:`NoneType `\n :param client: the client to use. If not passed, falls back to\n the client stored on the current project.\ndiff --git a/gcloud/storage/__init__.py b/google/cloud/storage/__init__.py\nsimilarity index 72%\nrename from gcloud/storage/__init__.py\nrename to google/cloud/storage/__init__.py\n--- a/gcloud/storage/__init__.py\n+++ b/google/cloud/storage/__init__.py\n@@ -16,7 +16,7 @@\n \n You'll typically use these to get started with the API:\n \n->>> from gcloud import storage\n+>>> from google.cloud import storage\n >>> client = storage.Client()\n >>> bucket = client.get_bucket('bucket-id-here')\n >>> # Then do other things...\n@@ -28,22 +28,22 @@\n \n The main concepts with this API are:\n \n-- :class:`gcloud.storage.connection.Connection` which represents a\n+- :class:`google.cloud.storage.connection.Connection` which represents a\n connection between your machine and the Cloud Storage API.\n \n-- :class:`gcloud.storage.bucket.Bucket` which represents a particular\n+- :class:`google.cloud.storage.bucket.Bucket` which represents a particular\n bucket (akin to a mounted disk on a computer).\n \n-- :class:`gcloud.storage.blob.Blob` which represents a pointer to a\n+- :class:`google.cloud.storage.blob.Blob` which represents a pointer to a\n particular entity in Cloud Storage (akin to a file path on a remote\n machine).\n \"\"\"\n \n-from gcloud.storage.batch import Batch\n-from gcloud.storage.blob import Blob\n-from gcloud.storage.bucket import Bucket\n-from gcloud.storage.client import Client\n-from gcloud.storage.connection import Connection\n+from google.cloud.storage.batch import Batch\n+from google.cloud.storage.blob import Blob\n+from google.cloud.storage.bucket import Bucket\n+from google.cloud.storage.client import Client\n+from google.cloud.storage.connection import Connection\n \n \n SCOPE = Connection.SCOPE\ndiff --git a/gcloud/storage/_helpers.py b/google/cloud/storage/_helpers.py\nsimilarity index 92%\nrename from gcloud/storage/_helpers.py\nrename to google/cloud/storage/_helpers.py\n--- a/gcloud/storage/_helpers.py\n+++ b/google/cloud/storage/_helpers.py\n@@ -50,11 +50,12 @@ def client(self):\n def _require_client(self, client):\n \"\"\"Check client or verify over-ride.\n \n- :type client: :class:`gcloud.storage.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.storage.client.Client` or\n+ ``NoneType``\n :param client: the client to use. If not passed, falls back to the\n ``client`` stored on the current object.\n \n- :rtype: :class:`gcloud.storage.client.Client`\n+ :rtype: :class:`google.cloud.storage.client.Client`\n :returns: The client passed in or the currently bound client.\n \"\"\"\n if client is None:\n@@ -64,7 +65,8 @@ def _require_client(self, client):\n def reload(self, client=None):\n \"\"\"Reload properties from Cloud Storage.\n \n- :type client: :class:`gcloud.storage.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.storage.client.Client` or\n+ ``NoneType``\n :param client: the client to use. If not passed, falls back to the\n ``client`` stored on the current object.\n \"\"\"\n@@ -98,7 +100,7 @@ def _patch_property(self, name, value):\n def _set_properties(self, value):\n \"\"\"Set the properties for the current object.\n \n- :type value: dict or :class:`gcloud.storage.batch._FutureDict`\n+ :type value: dict or :class:`google.cloud.storage.batch._FutureDict`\n :param value: The properties to be set.\n \"\"\"\n self._properties = value\n@@ -110,7 +112,8 @@ def patch(self, client=None):\n \n Updates the ``_properties`` with the response from the backend.\n \n- :type client: :class:`gcloud.storage.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.storage.client.Client` or\n+ ``NoneType``\n :param client: the client to use. If not passed, falls back to the\n ``client`` stored on the current object.\n \"\"\"\ndiff --git a/gcloud/storage/acl.py b/google/cloud/storage/acl.py\nsimilarity index 93%\nrename from gcloud/storage/acl.py\nrename to google/cloud/storage/acl.py\n--- a/gcloud/storage/acl.py\n+++ b/google/cloud/storage/acl.py\n@@ -14,11 +14,11 @@\n \n \"\"\"Manipulate access control lists that Cloud Storage provides.\n \n-:class:`gcloud.storage.bucket.Bucket` has a getting method that creates\n+:class:`google.cloud.storage.bucket.Bucket` has a getting method that creates\n an ACL object under the hood, and you can interact with that using\n-:func:`gcloud.storage.bucket.Bucket.acl`::\n+:func:`google.cloud.storage.bucket.Bucket.acl`::\n \n- >>> from gcloud import storage\n+ >>> from google.cloud import storage\n >>> client = storage.Client()\n >>> bucket = client.get_bucket(bucket_name)\n >>> acl = bucket.acl\n@@ -58,13 +58,13 @@\n >>> acl.all().grant_read().revoke_write()\n \n After that, you can save any changes you make with the\n-:func:`gcloud.storage.acl.ACL.save` method::\n+:func:`google.cloud.storage.acl.ACL.save` method::\n \n >>> acl.save()\n \n-You can alternatively save any existing :class:`gcloud.storage.acl.ACL`\n+You can alternatively save any existing :class:`google.cloud.storage.acl.ACL`\n object (whether it was created by a factory method or not) from a\n-:class:`gcloud.storage.bucket.Bucket`::\n+:class:`google.cloud.storage.bucket.Bucket`::\n \n >>> bucket.acl.save(acl=acl)\n \n@@ -382,11 +382,12 @@ def client(self):\n def _require_client(self, client):\n \"\"\"Check client or verify over-ride.\n \n- :type client: :class:`gcloud.storage.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.storage.client.Client` or\n+ ``NoneType``\n :param client: the client to use. If not passed, falls back to the\n ``client`` stored on the current ACL.\n \n- :rtype: :class:`gcloud.storage.client.Client`\n+ :rtype: :class:`google.cloud.storage.client.Client`\n :returns: The client passed in or the currently bound client.\n \"\"\"\n if client is None:\n@@ -396,7 +397,8 @@ def _require_client(self, client):\n def reload(self, client=None):\n \"\"\"Reload the ACL data from Cloud Storage.\n \n- :type client: :class:`gcloud.storage.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.storage.client.Client` or\n+ ``NoneType``\n :param client: Optional. The client to use. If not passed, falls back\n to the ``client`` stored on the ACL's parent.\n \"\"\"\n@@ -413,7 +415,7 @@ def reload(self, client=None):\n def _save(self, acl, predefined, client):\n \"\"\"Helper for :meth:`save` and :meth:`save_predefined`.\n \n- :type acl: :class:`gcloud.storage.acl.ACL`, or a compatible list.\n+ :type acl: :class:`google.cloud.storage.acl.ACL`, or a compatible list.\n :param acl: The ACL object to save. If left blank, this will save\n current entries.\n \n@@ -422,7 +424,8 @@ def _save(self, acl, predefined, client):\n of the keys in :attr:`PREDEFINED_JSON_ACLS`\n If passed, `acl` must be None.\n \n- :type client: :class:`gcloud.storage.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.storage.client.Client` or\n+ ``NoneType``\n :param client: Optional. The client to use. If not passed, falls back\n to the ``client`` stored on the ACL's parent.\n \"\"\"\n@@ -446,11 +449,12 @@ def _save(self, acl, predefined, client):\n def save(self, acl=None, client=None):\n \"\"\"Save this ACL for the current bucket.\n \n- :type acl: :class:`gcloud.storage.acl.ACL`, or a compatible list.\n+ :type acl: :class:`google.cloud.storage.acl.ACL`, or a compatible list.\n :param acl: The ACL object to save. If left blank, this will save\n current entries.\n \n- :type client: :class:`gcloud.storage.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.storage.client.Client` or\n+ ``NoneType``\n :param client: Optional. The client to use. If not passed, falls back\n to the ``client`` stored on the ACL's parent.\n \"\"\"\n@@ -473,7 +477,8 @@ def save_predefined(self, predefined, client=None):\n aliased to the corresponding JSON name).\n If passed, `acl` must be None.\n \n- :type client: :class:`gcloud.storage.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.storage.client.Client` or\n+ ``NoneType``\n :param client: Optional. The client to use. If not passed, falls back\n to the ``client`` stored on the ACL's parent.\n \"\"\"\n@@ -492,7 +497,8 @@ def clear(self, client=None):\n have access to a bucket that you created even after you clear\n ACL rules with this method.\n \n- :type client: :class:`gcloud.storage.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.storage.client.Client` or\n+ ``NoneType``\n :param client: Optional. The client to use. If not passed, falls back\n to the ``client`` stored on the ACL's parent.\n \"\"\"\n@@ -502,7 +508,7 @@ def clear(self, client=None):\n class BucketACL(ACL):\n \"\"\"An ACL specifically for a bucket.\n \n- :type bucket: :class:`gcloud.storage.bucket.Bucket`\n+ :type bucket: :class:`google.cloud.storage.bucket.Bucket`\n :param bucket: The bucket to which this ACL relates.\n \"\"\"\n \n@@ -536,7 +542,7 @@ class DefaultObjectACL(BucketACL):\n class ObjectACL(ACL):\n \"\"\"An ACL specifically for a Cloud Storage object / blob.\n \n- :type blob: :class:`gcloud.storage.blob.Blob`\n+ :type blob: :class:`google.cloud.storage.blob.Blob`\n :param blob: The blob that this ACL corresponds to.\n \"\"\"\n \ndiff --git a/gcloud/storage/batch.py b/google/cloud/storage/batch.py\nsimilarity index 98%\nrename from gcloud/storage/batch.py\nrename to google/cloud/storage/batch.py\n--- a/gcloud/storage/batch.py\n+++ b/google/cloud/storage/batch.py\n@@ -26,8 +26,8 @@\n import httplib2\n import six\n \n-from gcloud.exceptions import make_exception\n-from gcloud.storage.connection import Connection\n+from google.cloud.exceptions import make_exception\n+from google.cloud.storage.connection import Connection\n \n \n class MIMEApplicationHTTP(MIMEApplication):\n@@ -126,7 +126,7 @@ def __setitem__(self, key, value):\n class Batch(Connection):\n \"\"\"Proxy an underlying connection, batching up change operations.\n \n- :type client: :class:`gcloud.storage.client.Client`\n+ :type client: :class:`google.cloud.storage.client.Client`\n :param client: The client to use for making connections.\n \"\"\"\n _MAX_BATCH_SIZE = 1000\ndiff --git a/gcloud/storage/blob.py b/google/cloud/storage/blob.py\nsimilarity index 92%\nrename from gcloud/storage/blob.py\nrename to google/cloud/storage/blob.py\n--- a/gcloud/storage/blob.py\n+++ b/google/cloud/storage/blob.py\n@@ -28,20 +28,20 @@\n import six\n from six.moves.urllib.parse import quote\n \n-from gcloud._helpers import _rfc3339_to_datetime\n-from gcloud._helpers import _to_bytes\n-from gcloud._helpers import _bytes_to_unicode\n-from gcloud.credentials import generate_signed_url\n-from gcloud.exceptions import NotFound\n-from gcloud.exceptions import make_exception\n-from gcloud.storage._helpers import _PropertyMixin\n-from gcloud.storage._helpers import _scalar_property\n-from gcloud.storage.acl import ObjectACL\n-from gcloud.streaming.http_wrapper import Request\n-from gcloud.streaming.http_wrapper import make_api_request\n-from gcloud.streaming.transfer import Download\n-from gcloud.streaming.transfer import RESUMABLE_UPLOAD\n-from gcloud.streaming.transfer import Upload\n+from google.cloud._helpers import _rfc3339_to_datetime\n+from google.cloud._helpers import _to_bytes\n+from google.cloud._helpers import _bytes_to_unicode\n+from google.cloud.credentials import generate_signed_url\n+from google.cloud.exceptions import NotFound\n+from google.cloud.exceptions import make_exception\n+from google.cloud.storage._helpers import _PropertyMixin\n+from google.cloud.storage._helpers import _scalar_property\n+from google.cloud.storage.acl import ObjectACL\n+from google.cloud.streaming.http_wrapper import Request\n+from google.cloud.streaming.http_wrapper import make_api_request\n+from google.cloud.streaming.transfer import Download\n+from google.cloud.streaming.transfer import RESUMABLE_UPLOAD\n+from google.cloud.streaming.transfer import Upload\n \n \n _API_ACCESS_ENDPOINT = 'https://storage.googleapis.com'\n@@ -54,7 +54,7 @@ class Blob(_PropertyMixin):\n :param name: The name of the blob. This corresponds to the\n unique path of the object in the bucket.\n \n- :type bucket: :class:`gcloud.storage.bucket.Bucket`\n+ :type bucket: :class:`google.cloud.storage.bucket.Bucket`\n :param bucket: The bucket to which this blob belongs.\n \n :type chunk_size: integer\n@@ -170,7 +170,7 @@ def generate_signed_url(self, expiration, method='GET',\n service account from a JSON file rather than a GCE service account.\n \n .. _Issue 922: https://github.com/GoogleCloudPlatform/\\\n- gcloud-python/issues/922\n+ google-cloud-python/issues/922\n \n If you have a blob that you want to allow access to for a set\n amount of time, you can use this method to generate a URL that\n@@ -207,7 +207,8 @@ def generate_signed_url(self, expiration, method='GET',\n for the signed URL. Used to over-ride the content\n type of the underlying blob/object.\n \n- :type client: :class:`gcloud.storage.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.storage.client.Client` or\n+ ``NoneType``\n :param client: (Optional) The client to use. If not passed, falls back\n to the ``client`` stored on the blob's bucket.\n \n@@ -242,7 +243,8 @@ def generate_signed_url(self, expiration, method='GET',\n def exists(self, client=None):\n \"\"\"Determines whether or not this blob exists.\n \n- :type client: :class:`gcloud.storage.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.storage.client.Client` or\n+ ``NoneType``\n :param client: Optional. The client to use. If not passed, falls back\n to the ``client`` stored on the blob's bucket.\n \n@@ -269,15 +271,16 @@ def exists(self, client=None):\n def delete(self, client=None):\n \"\"\"Deletes a blob from Cloud Storage.\n \n- :type client: :class:`gcloud.storage.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.storage.client.Client` or\n+ ``NoneType``\n :param client: Optional. The client to use. If not passed, falls back\n to the ``client`` stored on the blob's bucket.\n \n :rtype: :class:`Blob`\n :returns: The blob that was just deleted.\n- :raises: :class:`gcloud.exceptions.NotFound`\n+ :raises: :class:`google.cloud.exceptions.NotFound`\n (propagated from\n- :meth:`gcloud.storage.bucket.Bucket.delete_blob`).\n+ :meth:`google.cloud.storage.bucket.Bucket.delete_blob`).\n \"\"\"\n return self.bucket.delete_blob(self.name, client=client)\n \n@@ -292,8 +295,8 @@ def download_to_file(self, file_obj, encryption_key=None, client=None):\n Downloading a file that has been encrypted with a `customer-supplied`_\n encryption key::\n \n- >>> from gcloud import storage\n- >>> from gcloud.storage import Blob\n+ >>> from google.cloud import storage\n+ >>> from google.cloud.storage import Blob\n \n >>> client = storage.Client(project='my-project')\n >>> bucket = client.get_bucket('my-bucket')\n@@ -316,11 +319,12 @@ def download_to_file(self, file_obj, encryption_key=None, client=None):\n :param encryption_key: Optional 32 byte encryption key for\n customer-supplied encryption.\n \n- :type client: :class:`gcloud.storage.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.storage.client.Client` or\n+ ``NoneType``\n :param client: Optional. The client to use. If not passed, falls back\n to the ``client`` stored on the blob's bucket.\n \n- :raises: :class:`gcloud.exceptions.NotFound`\n+ :raises: :class:`google.cloud.exceptions.NotFound`\n \"\"\"\n client = self._require_client(client)\n if self.media_link is None: # not yet loaded\n@@ -358,11 +362,12 @@ def download_to_filename(self, filename, encryption_key=None, client=None):\n :param encryption_key: Optional 32 byte encryption key for\n customer-supplied encryption.\n \n- :type client: :class:`gcloud.storage.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.storage.client.Client` or\n+ ``NoneType``\n :param client: Optional. The client to use. If not passed, falls back\n to the ``client`` stored on the blob's bucket.\n \n- :raises: :class:`gcloud.exceptions.NotFound`\n+ :raises: :class:`google.cloud.exceptions.NotFound`\n \"\"\"\n with open(filename, 'wb') as file_obj:\n self.download_to_file(file_obj, encryption_key=encryption_key,\n@@ -378,13 +383,14 @@ def download_as_string(self, encryption_key=None, client=None):\n :param encryption_key: Optional 32 byte encryption key for\n customer-supplied encryption.\n \n- :type client: :class:`gcloud.storage.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.storage.client.Client` or\n+ ``NoneType``\n :param client: Optional. The client to use. If not passed, falls back\n to the ``client`` stored on the blob's bucket.\n \n :rtype: bytes\n :returns: The data stored in this blob.\n- :raises: :class:`gcloud.exceptions.NotFound`\n+ :raises: :class:`google.cloud.exceptions.NotFound`\n \"\"\"\n string_buffer = BytesIO()\n self.download_to_file(string_buffer, encryption_key=encryption_key,\n@@ -425,8 +431,8 @@ def upload_from_file(self, file_obj, rewind=False, size=None,\n \n Uploading a file with a `customer-supplied`_ encryption key::\n \n- >>> from gcloud import storage\n- >>> from gcloud.storage import Blob\n+ >>> from google.cloud import storage\n+ >>> from google.cloud.storage import Blob\n \n >>> client = storage.Client(project='my-project')\n >>> bucket = client.get_bucket('my-bucket')\n@@ -465,13 +471,14 @@ def upload_from_file(self, file_obj, rewind=False, size=None,\n :type num_retries: integer\n :param num_retries: Number of upload retries. Defaults to 6.\n \n- :type client: :class:`gcloud.storage.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.storage.client.Client` or\n+ ``NoneType``\n :param client: Optional. The client to use. If not passed, falls back\n to the ``client`` stored on the blob's bucket.\n \n :raises: :class:`ValueError` if size is not passed in and can not be\n- determined; :class:`gcloud.exceptions.GCloudError` if the\n- upload response returns an error status.\n+ determined; :class:`google.cloud.exceptions.GoogleCloudError`\n+ if the upload response returns an error status.\n \"\"\"\n client = self._require_client(client)\n # Use the private ``_connection`` rather than the public\n@@ -584,7 +591,8 @@ def upload_from_filename(self, filename, content_type=None,\n :param encryption_key: Optional 32 byte encryption key for\n customer-supplied encryption.\n \n- :type client: :class:`gcloud.storage.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.storage.client.Client` or\n+ ``NoneType``\n :param client: Optional. The client to use. If not passed, falls back\n to the ``client`` stored on the blob's bucket.\n \"\"\"\n@@ -623,7 +631,8 @@ def upload_from_string(self, data, content_type='text/plain',\n :param encryption_key: Optional 32 byte encryption key for\n customer-supplied encryption.\n \n- :type client: :class:`gcloud.storage.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.storage.client.Client` or\n+ ``NoneType``\n :param client: Optional. The client to use. If not passed, falls back\n to the ``client`` stored on the blob's bucket.\n \"\"\"\n@@ -638,7 +647,8 @@ def upload_from_string(self, data, content_type='text/plain',\n def make_public(self, client=None):\n \"\"\"Make this blob public giving all users read access.\n \n- :type client: :class:`gcloud.storage.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.storage.client.Client` or\n+ ``NoneType``\n :param client: Optional. The client to use. If not passed, falls back\n to the ``client`` stored on the blob's bucket.\n \"\"\"\ndiff --git a/gcloud/storage/bucket.py b/google/cloud/storage/bucket.py\nsimilarity index 90%\nrename from gcloud/storage/bucket.py\nrename to google/cloud/storage/bucket.py\n--- a/gcloud/storage/bucket.py\n+++ b/google/cloud/storage/bucket.py\n@@ -12,35 +12,35 @@\n # See the License for the specific language governing permissions and\n # limitations under the License.\n \n-\"\"\"Create / interact with gcloud storage buckets.\"\"\"\n+\"\"\"Create / interact with Google Cloud Storage buckets.\"\"\"\n \n import copy\n \n import six\n \n-from gcloud._helpers import _rfc3339_to_datetime\n-from gcloud.exceptions import NotFound\n-from gcloud.iterator import Iterator\n-from gcloud.storage._helpers import _PropertyMixin\n-from gcloud.storage._helpers import _scalar_property\n-from gcloud.storage.acl import BucketACL\n-from gcloud.storage.acl import DefaultObjectACL\n-from gcloud.storage.blob import Blob\n+from google.cloud._helpers import _rfc3339_to_datetime\n+from google.cloud.exceptions import NotFound\n+from google.cloud.iterator import Iterator\n+from google.cloud.storage._helpers import _PropertyMixin\n+from google.cloud.storage._helpers import _scalar_property\n+from google.cloud.storage.acl import BucketACL\n+from google.cloud.storage.acl import DefaultObjectACL\n+from google.cloud.storage.blob import Blob\n \n \n class _BlobIterator(Iterator):\n \"\"\"An iterator listing blobs in a bucket\n \n You shouldn't have to use this directly, but instead should use the\n- :class:`gcloud.storage.blob.Bucket.list_blobs` method.\n+ :class:`google.cloud.storage.blob.Bucket.list_blobs` method.\n \n- :type bucket: :class:`gcloud.storage.bucket.Bucket`\n+ :type bucket: :class:`google.cloud.storage.bucket.Bucket`\n :param bucket: The bucket from which to list blobs.\n \n :type extra_params: dict or None\n :param extra_params: Extra query string parameters for the API call.\n \n- :type client: :class:`gcloud.storage.client.Client`\n+ :type client: :class:`google.cloud.storage.client.Client`\n :param client: Optional. The client to use for making connections.\n Defaults to the bucket's client.\n \"\"\"\n@@ -72,7 +72,7 @@ def get_items_from_response(self, response):\n class Bucket(_PropertyMixin):\n \"\"\"A class representing a Bucket on Cloud Storage.\n \n- :type client: :class:`gcloud.storage.client.Client`\n+ :type client: :class:`google.cloud.storage.client.Client`\n :param client: A client which holds credentials and project configuration\n for the bucket (which requires a project).\n \n@@ -118,7 +118,7 @@ def blob(self, blob_name, chunk_size=None):\n (1 MB). This must be a multiple of 256 KB per the\n API specification.\n \n- :rtype: :class:`gcloud.storage.blob.Blob`\n+ :rtype: :class:`google.cloud.storage.blob.Blob`\n :returns: The blob object created.\n \"\"\"\n return Blob(name=blob_name, bucket=self, chunk_size=chunk_size)\n@@ -126,7 +126,8 @@ def blob(self, blob_name, chunk_size=None):\n def exists(self, client=None):\n \"\"\"Determines whether or not this bucket exists.\n \n- :type client: :class:`gcloud.storage.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.storage.client.Client` or\n+ ``NoneType``\n :param client: Optional. The client to use. If not passed, falls back\n to the ``client`` stored on the current bucket.\n \n@@ -154,11 +155,12 @@ def create(self, client=None):\n \"\"\"Creates current bucket.\n \n If the bucket already exists, will raise\n- :class:`gcloud.exceptions.Conflict`.\n+ :class:`google.cloud.exceptions.Conflict`.\n \n This implements \"storage.buckets.insert\".\n \n- :type client: :class:`gcloud.storage.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.storage.client.Client` or\n+ ``NoneType``\n :param client: Optional. The client to use. If not passed, falls back\n to the ``client`` stored on the current bucket.\n \"\"\"\n@@ -207,7 +209,7 @@ def get_blob(self, blob_name, client=None):\n \n This will return None if the blob doesn't exist::\n \n- >>> from gcloud import storage\n+ >>> from google.cloud import storage\n >>> client = storage.Client()\n >>> bucket = client.get_bucket('my-bucket')\n >>> print bucket.get_blob('/path/to/blob.txt')\n@@ -218,11 +220,12 @@ def get_blob(self, blob_name, client=None):\n :type blob_name: string\n :param blob_name: The name of the blob to retrieve.\n \n- :type client: :class:`gcloud.storage.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.storage.client.Client` or\n+ ``NoneType``\n :param client: Optional. The client to use. If not passed, falls back\n to the ``client`` stored on the current bucket.\n \n- :rtype: :class:`gcloud.storage.blob.Blob` or None\n+ :rtype: :class:`google.cloud.storage.blob.Blob` or None\n :returns: The blob object if it exists, otherwise None.\n \"\"\"\n client = self._require_client(client)\n@@ -273,7 +276,8 @@ def list_blobs(self, max_results=None, page_token=None, prefix=None,\n and the language of each blob returned:\n 'items/contentLanguage,nextPageToken'\n \n- :type client: :class:`gcloud.storage.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.storage.client.Client` or\n+ ``NoneType``\n :param client: Optional. The client to use. If not passed, falls back\n to the ``client`` stored on the current bucket.\n \n@@ -304,7 +308,9 @@ def list_blobs(self, max_results=None, page_token=None, prefix=None,\n # Page token must be handled specially since the base `Iterator`\n # class has it as a reserved property.\n if page_token is not None:\n+ # pylint: disable=attribute-defined-outside-init\n result.next_page_token = page_token\n+ # pylint: enable=attribute-defined-outside-init\n return result\n \n def delete(self, force=False, client=None):\n@@ -315,8 +321,9 @@ def delete(self, force=False, client=None):\n objects / blobs in the bucket (i.e. try to empty the bucket).\n \n If the bucket doesn't exist, this will raise\n- :class:`gcloud.exceptions.NotFound`. If the bucket is not empty\n- (and ``force=False``), will raise :class:`gcloud.exceptions.Conflict`.\n+ :class:`google.cloud.exceptions.NotFound`. If the bucket is not empty\n+ (and ``force=False``), will raise\n+ :class:`google.cloud.exceptions.Conflict`.\n \n If ``force=True`` and the bucket contains more than 256 objects / blobs\n this will cowardly refuse to delete the objects (or the bucket). This\n@@ -326,7 +333,8 @@ def delete(self, force=False, client=None):\n :type force: boolean\n :param force: If True, empties the bucket's objects then deletes it.\n \n- :type client: :class:`gcloud.storage.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.storage.client.Client` or\n+ ``NoneType``\n :param client: Optional. The client to use. If not passed, falls back\n to the ``client`` stored on the current bucket.\n \n@@ -361,12 +369,12 @@ def delete_blob(self, blob_name, client=None):\n \"\"\"Deletes a blob from the current bucket.\n \n If the blob isn't found (backend 404), raises a\n- :class:`gcloud.exceptions.NotFound`.\n+ :class:`google.cloud.exceptions.NotFound`.\n \n For example::\n \n- >>> from gcloud.exceptions import NotFound\n- >>> from gcloud import storage\n+ >>> from google.cloud.exceptions import NotFound\n+ >>> from google.cloud import storage\n >>> client = storage.Client()\n >>> bucket = client.get_bucket('my-bucket')\n >>> print bucket.list_blobs()\n@@ -380,11 +388,12 @@ def delete_blob(self, blob_name, client=None):\n :type blob_name: string\n :param blob_name: A blob name to delete.\n \n- :type client: :class:`gcloud.storage.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.storage.client.Client` or\n+ ``NoneType``\n :param client: Optional. The client to use. If not passed, falls back\n to the ``client`` stored on the current bucket.\n \n- :raises: :class:`gcloud.exceptions.NotFound` (to suppress\n+ :raises: :class:`google.cloud.exceptions.NotFound` (to suppress\n the exception, call ``delete_blobs``, passing a no-op\n ``on_error`` callback, e.g.::\n \n@@ -403,19 +412,20 @@ def delete_blobs(self, blobs, on_error=None, client=None):\n \n Uses :func:`Bucket.delete_blob` to delete each individual blob.\n \n- :type blobs: list of string or :class:`gcloud.storage.blob.Blob`\n+ :type blobs: list of string or :class:`google.cloud.storage.blob.Blob`\n :param blobs: A list of blob names or Blob objects to delete.\n \n :type on_error: a callable taking (blob)\n :param on_error: If not ``None``, called once for each blob raising\n- :class:`gcloud.exceptions.NotFound`;\n+ :class:`google.cloud.exceptions.NotFound`;\n otherwise, the exception is propagated.\n \n- :type client: :class:`gcloud.storage.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.storage.client.Client` or\n+ ``NoneType``\n :param client: Optional. The client to use. If not passed, falls back\n to the ``client`` stored on the current bucket.\n \n- :raises: :class:`gcloud.exceptions.NotFound` (if\n+ :raises: :class:`google.cloud.exceptions.NotFound` (if\n `on_error` is not passed).\n \"\"\"\n for blob in blobs:\n@@ -434,21 +444,22 @@ def copy_blob(self, blob, destination_bucket, new_name=None,\n client=None):\n \"\"\"Copy the given blob to the given bucket, optionally with a new name.\n \n- :type blob: :class:`gcloud.storage.blob.Blob`\n+ :type blob: :class:`google.cloud.storage.blob.Blob`\n :param blob: The blob to be copied.\n \n- :type destination_bucket: :class:`gcloud.storage.bucket.Bucket`\n+ :type destination_bucket: :class:`google.cloud.storage.bucket.Bucket`\n :param destination_bucket: The bucket into which the blob should be\n copied.\n \n :type new_name: string\n :param new_name: (optional) the new name for the copied file.\n \n- :type client: :class:`gcloud.storage.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.storage.client.Client` or\n+ ``NoneType``\n :param client: Optional. The client to use. If not passed, falls back\n to the ``client`` stored on the current bucket.\n \n- :rtype: :class:`gcloud.storage.blob.Blob`\n+ :rtype: :class:`google.cloud.storage.blob.Blob`\n :returns: The new Blob.\n \"\"\"\n client = self._require_client(client)\n@@ -473,13 +484,14 @@ def rename_blob(self, blob, new_name, client=None):\n old blob. This means that with very large objects renaming\n could be a very (temporarily) costly or a very slow operation.\n \n- :type blob: :class:`gcloud.storage.blob.Blob`\n+ :type blob: :class:`google.cloud.storage.blob.Blob`\n :param blob: The blob to be renamed.\n \n :type new_name: string\n :param new_name: The new name for this blob.\n \n- :type client: :class:`gcloud.storage.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.storage.client.Client` or\n+ ``NoneType``\n :param client: Optional. The client to use. If not passed, falls back\n to the ``client`` stored on the current bucket.\n \n@@ -735,7 +747,7 @@ def configure_website(self, main_page_suffix=None, not_found_page=None):\n If you want this bucket to host a website, just provide the name\n of an index page and a page to use when a blob isn't found::\n \n- >>> from gcloud import storage\n+ >>> from google.cloud import storage\n >>> client = storage.Client()\n >>> bucket = client.get_bucket(bucket_name)\n >>> bucket.configure_website('index.html', '404.html')\n@@ -785,7 +797,8 @@ def make_public(self, recursive=False, future=False, client=None):\n :param future: If True, this will make all objects created in the\n future public as well.\n \n- :type client: :class:`gcloud.storage.client.Client` or ``NoneType``\n+ :type client: :class:`~google.cloud.storage.client.Client` or\n+ ``NoneType``\n :param client: Optional. The client to use. If not passed, falls back\n to the ``client`` stored on the current bucket.\n \"\"\"\ndiff --git a/gcloud/storage/client.py b/google/cloud/storage/client.py\nsimilarity index 86%\nrename from gcloud/storage/client.py\nrename to google/cloud/storage/client.py\n--- a/gcloud/storage/client.py\n+++ b/google/cloud/storage/client.py\n@@ -15,13 +15,13 @@\n \"\"\"Client for interacting with the Google Cloud Storage API.\"\"\"\n \n \n-from gcloud._helpers import _LocalStack\n-from gcloud.client import JSONClient\n-from gcloud.exceptions import NotFound\n-from gcloud.iterator import Iterator\n-from gcloud.storage.batch import Batch\n-from gcloud.storage.bucket import Bucket\n-from gcloud.storage.connection import Connection\n+from google.cloud._helpers import _LocalStack\n+from google.cloud.client import JSONClient\n+from google.cloud.exceptions import NotFound\n+from google.cloud.iterator import Iterator\n+from google.cloud.storage.batch import Batch\n+from google.cloud.storage.bucket import Bucket\n+from google.cloud.storage.connection import Connection\n \n \n class Client(JSONClient):\n@@ -57,7 +57,7 @@ def __init__(self, project=None, credentials=None, http=None):\n def connection(self):\n \"\"\"Get connection or batch on the client.\n \n- :rtype: :class:`gcloud.storage.connection.Connection`\n+ :rtype: :class:`google.cloud.storage.connection.Connection`\n :returns: The connection set on the client, or the batch\n if one is set.\n \"\"\"\n@@ -74,7 +74,7 @@ def connection(self, value):\n self.connection = connection\n Will raise if the connection is set more than once.\n \n- :type value: :class:`gcloud.storage.connection.Connection`\n+ :type value: :class:`google.cloud.storage.connection.Connection`\n :param value: The connection set on the client.\n \n :raises: :class:`ValueError` if connection has already been set.\n@@ -88,7 +88,7 @@ def _push_batch(self, batch):\n \n \"Protected\", intended for use by batch context mgrs.\n \n- :type batch: :class:`gcloud.storage.batch.Batch`\n+ :type batch: :class:`google.cloud.storage.batch.Batch`\n :param batch: newly-active batch\n \"\"\"\n self._batch_stack.push(batch)\n@@ -99,7 +99,7 @@ def _pop_batch(self):\n \"Protected\", intended for use by batch context mgrs.\n \n :raises: IndexError if the stack is empty.\n- :rtype: :class:`gcloud.storage.batch.Batch`\n+ :rtype: :class:`google.cloud.storage.batch.Batch`\n :returns: the top-most batch/transaction, after removing it.\n \"\"\"\n return self._batch_stack.pop()\n@@ -108,7 +108,7 @@ def _pop_batch(self):\n def current_batch(self):\n \"\"\"Currently-active batch.\n \n- :rtype: :class:`gcloud.storage.batch.Batch` or ``NoneType`` (if\n+ :rtype: :class:`google.cloud.storage.batch.Batch` or ``NoneType`` (if\n no batch is active).\n :returns: The batch at the top of the batch stack.\n \"\"\"\n@@ -124,7 +124,7 @@ def bucket(self, bucket_name):\n :type bucket_name: string\n :param bucket_name: The name of the bucket to be instantiated.\n \n- :rtype: :class:`gcloud.storage.bucket.Bucket`\n+ :rtype: :class:`google.cloud.storage.bucket.Bucket`\n :returns: The bucket object created.\n \"\"\"\n return Bucket(client=self, name=bucket_name)\n@@ -136,7 +136,7 @@ def batch(self):\n This will not make an HTTP request; it simply instantiates\n a batch object owned by this client.\n \n- :rtype: :class:`gcloud.storage.batch.Batch`\n+ :rtype: :class:`google.cloud.storage.batch.Batch`\n :returns: The batch object created.\n \"\"\"\n return Batch(client=self)\n@@ -145,13 +145,13 @@ def get_bucket(self, bucket_name):\n \"\"\"Get a bucket by name.\n \n If the bucket isn't found, this will raise a\n- :class:`gcloud.storage.exceptions.NotFound`.\n+ :class:`google.cloud.storage.exceptions.NotFound`.\n \n For example::\n \n >>> try:\n >>> bucket = client.get_bucket('my-bucket')\n- >>> except gcloud.exceptions.NotFound:\n+ >>> except google.cloud.exceptions.NotFound:\n >>> print 'Sorry, that bucket does not exist!'\n \n This implements \"storage.buckets.get\".\n@@ -159,9 +159,9 @@ def get_bucket(self, bucket_name):\n :type bucket_name: string\n :param bucket_name: The name of the bucket to get.\n \n- :rtype: :class:`gcloud.storage.bucket.Bucket`\n+ :rtype: :class:`google.cloud.storage.bucket.Bucket`\n :returns: The bucket matching the name provided.\n- :raises: :class:`gcloud.exceptions.NotFound`\n+ :raises: :class:`google.cloud.exceptions.NotFound`\n \"\"\"\n bucket = Bucket(self, name=bucket_name)\n bucket.reload(client=self)\n@@ -183,7 +183,7 @@ def lookup_bucket(self, bucket_name):\n :type bucket_name: string\n :param bucket_name: The name of the bucket to get.\n \n- :rtype: :class:`gcloud.storage.bucket.Bucket`\n+ :rtype: :class:`google.cloud.storage.bucket.Bucket`\n :returns: The bucket matching the name provided or None if not found.\n \"\"\"\n try:\n@@ -203,12 +203,12 @@ def create_bucket(self, bucket_name):\n This implements \"storage.buckets.insert\".\n \n If the bucket already exists, will raise\n- :class:`gcloud.exceptions.Conflict`.\n+ :class:`google.cloud.exceptions.Conflict`.\n \n :type bucket_name: string\n :param bucket_name: The bucket name to create.\n \n- :rtype: :class:`gcloud.storage.bucket.Bucket`\n+ :rtype: :class:`google.cloud.storage.bucket.Bucket`\n :returns: The newly created bucket.\n \"\"\"\n bucket = Bucket(self, name=bucket_name)\n@@ -250,7 +250,8 @@ def list_buckets(self, max_results=None, page_token=None, prefix=None,\n and the language of each bucket returned:\n 'items/id,nextPageToken'\n \n- :rtype: iterable of :class:`gcloud.storage.bucket.Bucket` objects.\n+ :rtype: iterable of :class:`google.cloud.storage.bucket.Bucket`\n+ objects.\n :returns: All buckets belonging to this project.\n \"\"\"\n extra_params = {'project': self.project}\n@@ -271,7 +272,9 @@ def list_buckets(self, max_results=None, page_token=None, prefix=None,\n # Page token must be handled specially since the base `Iterator`\n # class has it as a reserved property.\n if page_token is not None:\n+ # pylint: disable=attribute-defined-outside-init\n result.next_page_token = page_token\n+ # pylint: enable=attribute-defined-outside-init\n return result\n \n \n@@ -279,10 +282,10 @@ class _BucketIterator(Iterator):\n \"\"\"An iterator listing all buckets.\n \n You shouldn't have to use this directly, but instead should use the\n- helper methods on :class:`gcloud.storage.connection.Connection`\n+ helper methods on :class:`google.cloud.storage.connection.Connection`\n objects.\n \n- :type client: :class:`gcloud.storage.client.Client`\n+ :type client: :class:`google.cloud.storage.client.Client`\n :param client: The client to use for making connections.\n \n :type extra_params: dict or ``NoneType``\ndiff --git a/gcloud/storage/connection.py b/google/cloud/storage/connection.py\nsimilarity index 93%\nrename from gcloud/storage/connection.py\nrename to google/cloud/storage/connection.py\n--- a/gcloud/storage/connection.py\n+++ b/google/cloud/storage/connection.py\n@@ -12,9 +12,9 @@\n # See the License for the specific language governing permissions and\n # limitations under the License.\n \n-\"\"\"Create / interact with gcloud storage connections.\"\"\"\n+\"\"\"Create / interact with Google Cloud Storage connections.\"\"\"\n \n-from gcloud import connection as base_connection\n+from google.cloud import connection as base_connection\n \n \n class Connection(base_connection.JSONConnection):\ndiff --git a/gcloud/streaming/__init__.py b/google/cloud/streaming/__init__.py\nsimilarity index 93%\nrename from gcloud/streaming/__init__.py\nrename to google/cloud/streaming/__init__.py\n--- a/gcloud/streaming/__init__.py\n+++ b/google/cloud/streaming/__init__.py\n@@ -13,3 +13,5 @@\n # limitations under the License.\n \n # Vendored-in from google-apitools 0.4.11\n+\n+\"\"\"Base ``google.cloud.streaming`` package.\"\"\"\ndiff --git a/gcloud/streaming/buffered_stream.py b/google/cloud/streaming/buffered_stream.py\nsimilarity index 100%\nrename from gcloud/streaming/buffered_stream.py\nrename to google/cloud/streaming/buffered_stream.py\ndiff --git a/gcloud/streaming/exceptions.py b/google/cloud/streaming/exceptions.py\nsimilarity index 95%\nrename from gcloud/streaming/exceptions.py\nrename to google/cloud/streaming/exceptions.py\n--- a/gcloud/streaming/exceptions.py\n+++ b/google/cloud/streaming/exceptions.py\n@@ -59,7 +59,7 @@ def status_code(self):\n def from_response(cls, http_response):\n \"\"\"Factory: construct an exception from a response.\n \n- :type http_response: :class:`gcloud.streaming.http_wrapper.Response`\n+ :type http_response: :class:`~.streaming.http_wrapper.Response`\n :param http_response: the response which returned the error\n \n :rtype: :class:`HttpError`\n@@ -108,7 +108,7 @@ def __init__(self, response, content, url, retry_after):\n def from_response(cls, http_response):\n \"\"\"Factory: construct an exception from a response.\n \n- :type http_response: :class:`gcloud.streaming.http_wrapper.Response`\n+ :type http_response: :class:`~.streaming.http_wrapper.Response`\n :param http_response: the response which returned the error.\n \n :rtype: :class:`RetryAfterError`\ndiff --git a/gcloud/streaming/http_wrapper.py b/google/cloud/streaming/http_wrapper.py\nsimilarity index 94%\nrename from gcloud/streaming/http_wrapper.py\nrename to google/cloud/streaming/http_wrapper.py\n--- a/gcloud/streaming/http_wrapper.py\n+++ b/google/cloud/streaming/http_wrapper.py\n@@ -29,10 +29,10 @@\n from six.moves import http_client\n from six.moves.urllib import parse\n \n-from gcloud.streaming.exceptions import BadStatusCodeError\n-from gcloud.streaming.exceptions import RequestError\n-from gcloud.streaming.exceptions import RetryAfterError\n-from gcloud.streaming.util import calculate_wait_for_retry\n+from google.cloud.streaming.exceptions import BadStatusCodeError\n+from google.cloud.streaming.exceptions import RequestError\n+from google.cloud.streaming.exceptions import RetryAfterError\n+from google.cloud.streaming.util import calculate_wait_for_retry\n \n \n _REDIRECTIONS = 5\n@@ -272,10 +272,10 @@ def _check_response(response):\n :type response: :class:`Response`\n :param response: the response to validate\n \n- :raises: :exc:`gcloud.streaming.exceptions.RequestError` if response is\n- None, :exc:`gcloud.streaming.exceptions.BadStatusCodeError` if\n+ :raises: :exc:`google.cloud.streaming.exceptions.RequestError` if response\n+ is None, :exc:`~.streaming.exceptions.BadStatusCodeError` if\n response status code indicates an error, or\n- :exc:`gcloud.streaming.exceptions.RetryAfterError` if response\n+ :exc:`~.streaming.exceptions.RetryAfterError` if response\n indicates a retry interval.\n \"\"\"\n if response is None:\n@@ -326,8 +326,8 @@ def _make_api_request_no_retry(http, http_request, redirections=_REDIRECTIONS):\n :rtype: :class:`Response`\n :returns: an object representing the server's response\n \n- :raises: :exc:`gcloud.streaming.exceptions.RequestError` if no response\n- could be parsed.\n+ :raises: :exc:`google.cloud.streaming.exceptions.RequestError` if no\n+ response could be parsed.\n \"\"\"\n connection_type = None\n # Handle overrides for connection types. This is used if the caller\n@@ -374,8 +374,8 @@ def make_api_request(http, http_request, retries=7,\n :rtype: :class:`Response`\n :returns: an object representing the server's response.\n \n- :raises: :exc:`gcloud.streaming.exceptions.RequestError` if no response\n- could be parsed.\n+ :raises: :exc:`google.cloud.streaming.exceptions.RequestError` if no\n+ response could be parsed.\n \"\"\"\n retry = 0\n while True:\ndiff --git a/gcloud/streaming/stream_slice.py b/google/cloud/streaming/stream_slice.py\nsimilarity index 100%\nrename from gcloud/streaming/stream_slice.py\nrename to google/cloud/streaming/stream_slice.py\ndiff --git a/gcloud/streaming/transfer.py b/google/cloud/streaming/transfer.py\nsimilarity index 94%\nrename from gcloud/streaming/transfer.py\nrename to google/cloud/streaming/transfer.py\n--- a/gcloud/streaming/transfer.py\n+++ b/google/cloud/streaming/transfer.py\n@@ -26,17 +26,17 @@\n import six\n from six.moves import http_client\n \n-from gcloud._helpers import _to_bytes\n-from gcloud.streaming.buffered_stream import BufferedStream\n-from gcloud.streaming.exceptions import CommunicationError\n-from gcloud.streaming.exceptions import HttpError\n-from gcloud.streaming.exceptions import TransferInvalidError\n-from gcloud.streaming.exceptions import TransferRetryError\n-from gcloud.streaming.http_wrapper import make_api_request\n-from gcloud.streaming.http_wrapper import Request\n-from gcloud.streaming.http_wrapper import RESUME_INCOMPLETE\n-from gcloud.streaming.stream_slice import StreamSlice\n-from gcloud.streaming.util import acceptable_mime_type\n+from google.cloud._helpers import _to_bytes\n+from google.cloud.streaming.buffered_stream import BufferedStream\n+from google.cloud.streaming.exceptions import CommunicationError\n+from google.cloud.streaming.exceptions import HttpError\n+from google.cloud.streaming.exceptions import TransferInvalidError\n+from google.cloud.streaming.exceptions import TransferRetryError\n+from google.cloud.streaming.http_wrapper import make_api_request\n+from google.cloud.streaming.http_wrapper import Request\n+from google.cloud.streaming.http_wrapper import RESUME_INCOMPLETE\n+from google.cloud.streaming.stream_slice import StreamSlice\n+from google.cloud.streaming.util import acceptable_mime_type\n \n \n RESUMABLE_UPLOAD_THRESHOLD = 5 << 20\n@@ -200,7 +200,7 @@ def initialized(self):\n def _ensure_initialized(self):\n \"\"\"Helper: assert that the instance is initialized.\n \n- :raises: :exc:`gcloud.streaming.exceptions.TransferInvalidError`\n+ :raises: :exc:`google.cloud.streaming.exceptions.TransferInvalidError`\n if the instance is not initialized.\n \"\"\"\n if not self.initialized:\n@@ -210,7 +210,7 @@ def _ensure_initialized(self):\n def _ensure_uninitialized(self):\n \"\"\"Helper: assert that the instance is not initialized.\n \n- :raises: :exc:`gcloud.streaming.exceptions.TransferInvalidError`\n+ :raises: :exc:`google.cloud.streaming.exceptions.TransferInvalidError`\n if the instance is already initialized.\n \"\"\"\n if self.initialized:\n@@ -334,7 +334,7 @@ def __repr__(self):\n def configure_request(self, http_request, url_builder):\n \"\"\"Update http_request/url_builder with download-appropriate values.\n \n- :type http_request: :class:`gcloud.streaming.http_wrapper.Request`\n+ :type http_request: :class:`~.streaming.http_wrapper.Request`\n :param http_request: the request to be updated\n \n :type url_builder: instance with settable 'query_params' attribute.\n@@ -366,7 +366,7 @@ def initialize_download(self, http_request, http):\n If the instance has :attr:`auto_transfer` enabled, begins the\n download immediately.\n \n- :type http_request: :class:`gcloud.streaming.http_wrapper.Request`\n+ :type http_request: :class:`~.streaming.http_wrapper.Request`\n :param http_request: the request to use to initialize this download.\n \n :type http: :class:`httplib2.Http` (or workalike)\n@@ -402,7 +402,7 @@ def _normalize_start_end(self, start, end=None):\n \n :rtype: tuple, (start, end)\n :returns: the normalized start, end pair.\n- :raises: :exc:`gcloud.streaming.exceptions.TransferInvalidError`\n+ :raises: :exc:`google.cloud.streaming.exceptions.TransferInvalidError`\n for invalid combinations of start, end.\n \"\"\"\n if end is not None:\n@@ -426,7 +426,7 @@ def _normalize_start_end(self, start, end=None):\n def _set_range_header(request, start, end=None):\n \"\"\"Update the 'Range' header in a request to match a byte range.\n \n- :type request: :class:`gcloud.streaming.http_wrapper.Request`\n+ :type request: :class:`google.cloud.streaming.http_wrapper.Request`\n :param request: the request to update\n \n :type start: integer\n@@ -497,7 +497,7 @@ def _get_chunk(self, start, end):\n :type end: integer or None\n :param end: end byte of the range.\n \n- :rtype: :class:`gcloud.streaming.http_wrapper.Response`\n+ :rtype: :class:`google.cloud.streaming.http_wrapper.Response`\n :returns: response from the chunk request.\n \"\"\"\n self._ensure_initialized()\n@@ -509,14 +509,14 @@ def _get_chunk(self, start, end):\n def _process_response(self, response):\n \"\"\"Update attribtes and writing stream, based on response.\n \n- :type response: :class:`gcloud.streaming.http_wrapper.Response`\n+ :type response: :class:`google.cloud.streaming.http_wrapper.Response`\n :param response: response from a download request.\n \n- :rtype: :class:`gcloud.streaming.http_wrapper.Response`\n+ :rtype: :class:`google.cloud.streaming.http_wrapper.Response`\n :returns: the response\n- :raises: :exc:`gcloud.streaming.exceptions.HttpError` for\n+ :raises: :exc:`google.cloud.streaming.exceptions.HttpError` for\n missing / unauthorized responses;\n- :exc:`gcloud.streaming.exceptions.TransferRetryError`\n+ :exc:`google.cloud.streaming.exceptions.TransferRetryError`\n for other error responses.\n \"\"\"\n if response.status_code not in self._ACCEPTABLE_STATUSES:\n@@ -564,7 +564,7 @@ def get_range(self, start, end=None, use_chunks=True):\n and fetch this range in a single request.\n If True, streams via chunks.\n \n- :raises: :exc:`gcloud.streaming.exceptions.TransferRetryError`\n+ :raises: :exc:`google.cloud.streaming.exceptions.TransferRetryError`\n if a request returns an empty response.\n \"\"\"\n self._ensure_initialized()\n@@ -810,7 +810,7 @@ def _set_default_strategy(self, upload_config, http_request):\n attributes\n :param upload_config: Configuration for the upload endpoint.\n \n- :type http_request: :class:`gcloud.streaming.http_wrapper.Request`\n+ :type http_request: :class:`~.streaming.http_wrapper.Request`\n :param http_request: The associated http request.\n \"\"\"\n if upload_config.resumable_path is None:\n@@ -834,7 +834,7 @@ def configure_request(self, upload_config, http_request, url_builder):\n attributes\n :param upload_config: transfer policy object to be queried\n \n- :type http_request: :class:`gcloud.streaming.http_wrapper.Request`\n+ :type http_request: :class:`~.streaming.http_wrapper.Request`\n :param http_request: the request to be updated\n \n :type url_builder: instance with settable 'relative_path' and\n@@ -970,7 +970,7 @@ def refresh_upload_state(self):\n def _get_range_header(response):\n \"\"\"Return a 'Range' header from a response.\n \n- :type response: :class:`gcloud.streaming.http_wrapper.Response`\n+ :type response: :class:`google.cloud.streaming.http_wrapper.Response`\n :param response: response to be queried\n \n :rtype: string\n@@ -994,7 +994,7 @@ def _get_range_header(response):\n def initialize_upload(self, http_request, http):\n \"\"\"Initialize this upload from the given http_request.\n \n- :type http_request: :class:`gcloud.streaming.http_wrapper.Request`\n+ :type http_request: :class:`~.streaming.http_wrapper.Request`\n :param http_request: the request to be used\n \n :type http: :class:`httplib2.Http` (or workalike)\n@@ -1002,7 +1002,7 @@ def initialize_upload(self, http_request, http):\n \n :raises: :exc:`ValueError` if the instance has not been configured\n with a strategy.\n- :rtype: :class:`~gcloud.streaming.http_wrapper.Response`\n+ :rtype: :class:`~google.cloud.streaming.http_wrapper.Response`\n :returns: The response if the upload is resumable and auto transfer\n is not used.\n \"\"\"\n@@ -1070,7 +1070,7 @@ def stream_file(self, use_chunks=True):\n :param use_chunks: If False, send the stream in a single request.\n Otherwise, send it in chunks.\n \n- :rtype: :class:`gcloud.streaming.http_wrapper.Response`\n+ :rtype: :class:`google.cloud.streaming.http_wrapper.Response`\n :returns: The response for the final request made.\n \"\"\"\n if self.strategy != RESUMABLE_UPLOAD:\n@@ -1110,15 +1110,15 @@ def _send_media_request(self, request, end):\n \n Helper for _send_media_body & _send_chunk:\n \n- :type request: :class:`gcloud.streaming.http_wrapper.Request`\n+ :type request: :class:`google.cloud.streaming.http_wrapper.Request`\n :param request: the request to upload\n \n :type end: integer\n :param end: end byte of the to be uploaded\n \n- :rtype: :class:`gcloud.streaming.http_wrapper.Response`\n+ :rtype: :class:`google.cloud.streaming.http_wrapper.Response`\n :returns: the response\n- :raises: :exc:`gcloud.streaming.exceptions.HttpError` if the status\n+ :raises: :exc:`~.streaming.exceptions.HttpError` if the status\n code from the response indicates an error.\n \"\"\"\n response = make_api_request(\n@@ -1144,7 +1144,7 @@ def _send_media_body(self, start):\n :type start: integer\n :param start: start byte of the range.\n \n- :rtype: :class:`gcloud.streaming.http_wrapper.Response`\n+ :rtype: :class:`google.cloud.streaming.http_wrapper.Response`\n :returns: The response from the media upload request.\n \"\"\"\n self._ensure_initialized()\n@@ -1174,7 +1174,7 @@ def _send_chunk(self, start):\n :type start: integer\n :param start: start byte of the range.\n \n- :rtype: :class:`gcloud.streaming.http_wrapper.Response`\n+ :rtype: :class:`google.cloud.streaming.http_wrapper.Response`\n :returns: The response from the chunked upload request.\n \"\"\"\n self._ensure_initialized()\ndiff --git a/gcloud/streaming/util.py b/google/cloud/streaming/util.py\nsimilarity index 100%\nrename from gcloud/streaming/util.py\nrename to google/cloud/streaming/util.py\ndiff --git a/gcloud/translate/__init__.py b/google/cloud/translate/__init__.py\nsimilarity index 85%\nrename from gcloud/translate/__init__.py\nrename to google/cloud/translate/__init__.py\n--- a/gcloud/translate/__init__.py\n+++ b/google/cloud/translate/__init__.py\n@@ -14,5 +14,5 @@\n \n \"\"\"Google Cloud Translate API wrapper.\"\"\"\n \n-from gcloud.translate.client import Client\n-from gcloud.translate.connection import Connection\n+from google.cloud.translate.client import Client\n+from google.cloud.translate.connection import Connection\ndiff --git a/gcloud/translate/client.py b/google/cloud/translate/client.py\nsimilarity index 98%\nrename from gcloud/translate/client.py\nrename to google/cloud/translate/client.py\n--- a/gcloud/translate/client.py\n+++ b/google/cloud/translate/client.py\n@@ -18,8 +18,8 @@\n import httplib2\n import six\n \n-from gcloud._helpers import _to_bytes\n-from gcloud.translate.connection import Connection\n+from google.cloud._helpers import _to_bytes\n+from google.cloud.translate.connection import Connection\n \n \n ENGLISH_ISO_639 = 'en'\ndiff --git a/gcloud/translate/connection.py b/google/cloud/translate/connection.py\nsimilarity index 95%\nrename from gcloud/translate/connection.py\nrename to google/cloud/translate/connection.py\n--- a/gcloud/translate/connection.py\n+++ b/google/cloud/translate/connection.py\n@@ -14,7 +14,7 @@\n \n \"\"\"Create / interact with Google Cloud Translate connections.\"\"\"\n \n-from gcloud import connection as base_connection\n+from google.cloud import connection as base_connection\n \n \n class Connection(base_connection.JSONConnection):\ndiff --git a/gcloud/vision/__init__.py b/google/cloud/vision/__init__.py\nsimilarity index 100%\nrename from gcloud/vision/__init__.py\nrename to google/cloud/vision/__init__.py\ndiff --git a/gcloud/vision/_fixtures.py b/google/cloud/vision/_fixtures.py\nsimilarity index 100%\nrename from gcloud/vision/_fixtures.py\nrename to google/cloud/vision/_fixtures.py\ndiff --git a/gcloud/vision/client.py b/google/cloud/vision/client.py\nsimilarity index 93%\nrename from gcloud/vision/client.py\nrename to google/cloud/vision/client.py\n--- a/gcloud/vision/client.py\n+++ b/google/cloud/vision/client.py\n@@ -15,10 +15,10 @@\n \"\"\"Client for interacting with the Google Cloud Vision API.\"\"\"\n \n \n-from gcloud.client import JSONClient\n-from gcloud.vision.connection import Connection\n-from gcloud.vision.feature import Feature\n-from gcloud.vision.image import Image\n+from google.cloud.client import JSONClient\n+from google.cloud.vision.connection import Connection\n+from google.cloud.vision.feature import Feature\n+from google.cloud.vision.image import Image\n \n \n class VisionRequest(object):\n@@ -89,7 +89,7 @@ def annotate(self, image, features):\n :param image: A string which can be a URL, a Google Cloud Storage path,\n or a byte stream of the image.\n \n- :type features: list of :class:`gcloud.vision.feature.Feature`\n+ :type features: list of :class:`google.cloud.vision.feature.Feature`\n :param features: The type of detection that the Vision API should\n use to determine image attributes. Pricing is\n based on the number of Feature Types.\ndiff --git a/gcloud/vision/connection.py b/google/cloud/vision/connection.py\nsimilarity index 93%\nrename from gcloud/vision/connection.py\nrename to google/cloud/vision/connection.py\n--- a/gcloud/vision/connection.py\n+++ b/google/cloud/vision/connection.py\n@@ -13,10 +13,10 @@\n # limitations under the License.\n \n \n-\"\"\"Create / interact with gcloud Vision connections.\"\"\"\n+\"\"\"Create / interact with Google Cloud Vision connections.\"\"\"\n \n \n-from gcloud import connection as base_connection\n+from google.cloud import connection as base_connection\n \n \n class Connection(base_connection.JSONConnection):\ndiff --git a/gcloud/vision/feature.py b/google/cloud/vision/feature.py\nsimilarity index 96%\nrename from gcloud/vision/feature.py\nrename to google/cloud/vision/feature.py\n--- a/gcloud/vision/feature.py\n+++ b/google/cloud/vision/feature.py\n@@ -34,7 +34,7 @@ class Feature(object):\n \n :type feature_type: str\n :param feature_type: String representation of\n- :class:`gcloud.vision.feature.FeatureType`.\n+ :class:`google.cloud.vision.feature.FeatureType`.\n \n :type max_results: int\n :param max_results: Number of results to return for the specified\ndiff --git a/gcloud/vision/image.py b/google/cloud/vision/image.py\nsimilarity index 95%\nrename from gcloud/vision/image.py\nrename to google/cloud/vision/image.py\n--- a/gcloud/vision/image.py\n+++ b/google/cloud/vision/image.py\n@@ -17,8 +17,8 @@\n \n from base64 import b64encode\n \n-from gcloud._helpers import _to_bytes\n-from gcloud._helpers import _bytes_to_unicode\n+from google.cloud._helpers import _to_bytes\n+from google.cloud._helpers import _bytes_to_unicode\n \n \n class Image(object):\ndiff --git a/scripts/generate_json_docs.py b/scripts/generate_json_docs.py\n--- a/scripts/generate_json_docs.py\n+++ b/scripts/generate_json_docs.py\n@@ -244,7 +244,7 @@ def build_link_from_list_of_types(type_names, object_type=None):\n def build_link_from_type(type_name, object_type=None):\n type_name = clean_type_name(type_name)\n \n- if not type_name.startswith('gcloud'):\n+ if not type_name.startswith('google.cloud'):\n return type_name\n doc_path = type_name\n \n@@ -425,13 +425,13 @@ def write_docs_file(path, contents):\n \n def generate_doc_types_json(modules, types_file_path):\n doc_types_list = [{\n- 'id': 'gcloud',\n+ 'id': 'google-cloud',\n 'contents': 'index.json',\n- 'title': 'gcloud'\n+ 'title': 'google-cloud'\n }]\n \n for module_name in modules:\n- if module_name == 'gcloud.__init__':\n+ if module_name == 'google.cloud.__init__':\n continue\n \n module_title = module_name.replace('.__init__', '').split('.')\n@@ -485,7 +485,7 @@ def generate_module_docs(modules, docs_path, real_base_path, toc):\n module_docs_path = os.path.join(docs_path, module_path) + '.json'\n \n if pdoc_module.functions():\n- toc_key = module_name.replace('gcloud.', '').split('.')[0]\n+ toc_key = module_name.replace('google.cloud.', '').split('.')[0]\n toc_entry = build_toc_entry(module.name, module_path)\n toc['services'][toc_key].append(toc_entry)\n \n@@ -501,7 +501,7 @@ def generate_class_docs(module, klass, base_path, toc):\n .replace('__init__', 'index'))\n module_docs_path = os.path.join(base_path, module_path,\n klass.name.lower()) + '.json'\n- toc_key = module.name.replace('gcloud.', '').split('.')[0]\n+ toc_key = module.name.replace('google.cloud.', '').split('.')[0]\n \n toc_entry = build_toc_entry(klass.name,\n os.path.join(module_path,\n@@ -590,8 +590,8 @@ def package_files(generated_json_dir, docs_build_dir, static_json_dir,\n shutil.rmtree(package_path, ignore_errors=True)\n \n shutil.copytree(static_json_dir, package_path)\n- shutil.copytree(os.path.join(generated_json_dir, 'gcloud'),\n- os.path.join(package_path, 'json', tag, 'gcloud'))\n+ shutil.copytree(os.path.join(generated_json_dir, 'google', 'cloud'),\n+ os.path.join(package_path, 'json', tag, 'google', 'cloud'))\n shutil.copyfile(os.path.join(generated_json_dir, 'types.json'),\n os.path.join(package_path, 'json', tag, 'types.json'))\n \n@@ -609,7 +609,8 @@ def main():\n toc = {\n 'services': {\n '__init__': [],\n- 'gcloud': [],\n+ 'google': [],\n+ 'cloud': [],\n 'bigquery': [],\n 'bigtable': [],\n 'client': [],\n@@ -641,8 +642,9 @@ def main():\n JSON_DOCS_DIR = os.path.join(DOCS_BUILD_DIR, 'json', args.tag)\n LIB_DIR = os.path.abspath(args.basepath)\n \n- library_dir = os.path.join(LIB_DIR, 'gcloud')\n- public_mods = get_public_modules(library_dir, base_package='gcloud')\n+ library_dir = os.path.join(LIB_DIR, 'google', 'cloud')\n+ public_mods = get_public_modules(library_dir,\n+ base_package='google.cloud')\n \n generate_module_docs(public_mods, JSON_DOCS_DIR, BASE_DIR, toc)\n generate_doc_types_json(public_mods,\ndiff --git a/scripts/make_datastore_grpc.py b/scripts/make_datastore_grpc.py\n--- a/scripts/make_datastore_grpc.py\n+++ b/scripts/make_datastore_grpc.py\n@@ -26,7 +26,7 @@\n PROTOS_DIR = os.path.join(ROOT_DIR, 'googleapis-pb')\n PROTO_PATH = os.path.join(PROTOS_DIR, 'google', 'datastore',\n 'v1', 'datastore.proto')\n-GRPC_ONLY_FILE = os.path.join(ROOT_DIR, 'gcloud', 'datastore',\n+GRPC_ONLY_FILE = os.path.join(ROOT_DIR, 'google', 'cloud', 'datastore',\n '_generated', 'datastore_grpc_pb2.py')\n GRPCIO_VIRTUALENV = os.getenv('GRPCIO_VIRTUALENV')\n if GRPCIO_VIRTUALENV is None:\n@@ -34,7 +34,8 @@\n else:\n PYTHON_EXECUTABLE = os.path.join(GRPCIO_VIRTUALENV, 'bin', 'python')\n MESSAGE_SNIPPET = ' = _reflection.GeneratedProtocolMessageType('\n-IMPORT_TEMPLATE = 'from gcloud.datastore._generated.datastore_pb2 import %s\\n'\n+IMPORT_TEMPLATE = (\n+ 'from google.cloud.datastore._generated.datastore_pb2 import %s\\n')\n \n \n def get_pb2_contents_with_grpc():\ndiff --git a/scripts/make_operations_grpc.py b/scripts/make_operations_grpc.py\n--- a/scripts/make_operations_grpc.py\n+++ b/scripts/make_operations_grpc.py\n@@ -27,7 +27,7 @@\n PROTO_PATH = os.path.join(PROTOS_DIR, 'google', 'longrunning',\n 'operations.proto')\n GENERATED_SUBDIR = os.getenv('GENERATED_SUBDIR', '_generated')\n-GRPC_ONLY_FILE = os.path.join(ROOT_DIR, 'gcloud', 'bigtable',\n+GRPC_ONLY_FILE = os.path.join(ROOT_DIR, 'google', 'cloud', 'bigtable',\n GENERATED_SUBDIR, 'operations_grpc_pb2.py')\n GRPCIO_VIRTUALENV = os.getenv('GRPCIO_VIRTUALENV')\n if GRPCIO_VIRTUALENV is None:\ndiff --git a/scripts/pycodestyle_on_repo.py b/scripts/pycodestyle_on_repo.py\n--- a/scripts/pycodestyle_on_repo.py\n+++ b/scripts/pycodestyle_on_repo.py\n@@ -12,7 +12,7 @@\n # See the License for the specific language governing permissions and\n # limitations under the License.\n \n-\"\"\"Custom script to run pycodestyle on gcloud codebase.\n+\"\"\"Custom script to run pycodestyle on google-cloud codebase.\n \n This runs pycodestyle as a script via subprocess but only runs it on the\n .py files that are checked in to the repository.\ndiff --git a/scripts/rewrite_imports.py b/scripts/rewrite_imports.py\n--- a/scripts/rewrite_imports.py\n+++ b/scripts/rewrite_imports.py\n@@ -25,10 +25,10 @@\n IMPORT_FROM_TEMPLATE = 'from %s import '\n REPLACEMENTS = {\n # Bigtable v2\n- 'google.bigtable.v2': 'gcloud.bigtable._generated',\n- 'google.bigtable.admin.v2': 'gcloud.bigtable._generated',\n+ 'google.bigtable.v2': 'google.cloud.bigtable._generated',\n+ 'google.bigtable.admin.v2': 'google.cloud.bigtable._generated',\n # Datastore v1\n- 'google.datastore.v1': 'gcloud.datastore._generated',\n+ 'google.datastore.v1': 'google.cloud.datastore._generated',\n }\n \n \ndiff --git a/scripts/run_pylint.py b/scripts/run_pylint.py\n--- a/scripts/run_pylint.py\n+++ b/scripts/run_pylint.py\n@@ -12,7 +12,7 @@\n # See the License for the specific language governing permissions and\n # limitations under the License.\n \n-\"\"\"Custom script to run PyLint on gcloud codebase.\n+\"\"\"Custom script to run PyLint on google-cloud codebase.\n \n This runs pylint as a script via subprocess in two different\n subprocesses. The first lints the production/library code\n@@ -31,14 +31,14 @@\n \n \n IGNORED_DIRECTORIES = [\n- os.path.join('gcloud', 'bigtable', '_generated'),\n- os.path.join('gcloud', 'datastore', '_generated'),\n+ os.path.join('google', 'cloud', 'bigtable', '_generated'),\n+ os.path.join('google', 'cloud', 'datastore', '_generated'),\n 'scripts/verify_included_modules.py',\n ]\n IGNORED_FILES = [\n os.path.join('docs', 'conf.py'),\n 'setup.py',\n- os.path.join('gcloud', 'vision', '_fixtures.py'),\n+ os.path.join('google', 'cloud', 'vision', '_fixtures.py'),\n ]\n SCRIPTS_DIR = os.path.abspath(os.path.dirname(__file__))\n PRODUCTION_RC = os.path.join(SCRIPTS_DIR, 'pylintrc_default')\n@@ -144,8 +144,8 @@ def get_files_for_linting(allow_limited=True):\n this value is not dependable.\n \n To allow faster local ``tox`` runs, the environment variables\n- ``GCLOUD_REMOTE_FOR_LINT`` and ``GCLOUD_BRANCH_FOR_LINT`` can be set to\n- specify a remote branch to diff against.\n+ ``GOOGLE_CLOUD_REMOTE_FOR_LINT`` and ``GOOGLE_CLOUD_BRANCH_FOR_LINT`` can\n+ be set to specify a remote branch to diff against.\n \n :type allow_limited: bool\n :param allow_limited: Boolean indicating if a reduced set of files can\n@@ -163,8 +163,8 @@ def get_files_for_linting(allow_limited=True):\n diff_base = 'origin/master'\n elif os.getenv('TRAVIS') is None:\n # Only allow specified remote and branch in local dev.\n- remote = os.getenv('GCLOUD_REMOTE_FOR_LINT')\n- branch = os.getenv('GCLOUD_BRANCH_FOR_LINT')\n+ remote = os.getenv('GOOGLE_CLOUD_REMOTE_FOR_LINT')\n+ branch = os.getenv('GOOGLE_CLOUD_BRANCH_FOR_LINT')\n if remote is not None and branch is not None:\n diff_base = '%s/%s' % (remote, branch)\n \ndiff --git a/scripts/verify_included_modules.py b/scripts/verify_included_modules.py\n--- a/scripts/verify_included_modules.py\n+++ b/scripts/verify_included_modules.py\n@@ -30,31 +30,31 @@\n DOCS_DIR = os.path.join(BASE_DIR, 'docs')\n IGNORED_PREFIXES = ('test_', '_')\n IGNORED_MODULES = frozenset([\n- 'gcloud.__init__',\n- 'gcloud.bigquery.__init__',\n- 'gcloud.bigtable.__init__',\n- 'gcloud.datastore.__init__',\n- 'gcloud.dns.__init__',\n- 'gcloud.error_reporting.__init__',\n- 'gcloud.iterator',\n- 'gcloud.language.__init__',\n- 'gcloud.logging.__init__',\n- 'gcloud.logging.handlers.__init__',\n- 'gcloud.logging.handlers.transports.__init__',\n- 'gcloud.monitoring.__init__',\n- 'gcloud.pubsub.__init__',\n- 'gcloud.resource_manager.__init__',\n- 'gcloud.storage.__init__',\n- 'gcloud.streaming.__init__',\n- 'gcloud.streaming.buffered_stream',\n- 'gcloud.streaming.exceptions',\n- 'gcloud.streaming.http_wrapper',\n- 'gcloud.streaming.stream_slice',\n- 'gcloud.streaming.transfer',\n- 'gcloud.streaming.util',\n- 'gcloud.translate.__init__',\n- 'gcloud.vision.__init__',\n- 'gcloud.vision.fixtures',\n+ 'google.cloud.__init__',\n+ 'google.cloud.bigquery.__init__',\n+ 'google.cloud.bigtable.__init__',\n+ 'google.cloud.datastore.__init__',\n+ 'google.cloud.dns.__init__',\n+ 'google.cloud.error_reporting.__init__',\n+ 'google.cloud.iterator',\n+ 'google.cloud.language.__init__',\n+ 'google.cloud.logging.__init__',\n+ 'google.cloud.logging.handlers.__init__',\n+ 'google.cloud.logging.handlers.transports.__init__',\n+ 'google.cloud.monitoring.__init__',\n+ 'google.cloud.pubsub.__init__',\n+ 'google.cloud.resource_manager.__init__',\n+ 'google.cloud.storage.__init__',\n+ 'google.cloud.streaming.__init__',\n+ 'google.cloud.streaming.buffered_stream',\n+ 'google.cloud.streaming.exceptions',\n+ 'google.cloud.streaming.http_wrapper',\n+ 'google.cloud.streaming.stream_slice',\n+ 'google.cloud.streaming.transfer',\n+ 'google.cloud.streaming.util',\n+ 'google.cloud.translate.__init__',\n+ 'google.cloud.vision.__init__',\n+ 'google.cloud.vision.fixtures',\n ])\n \n \n@@ -133,9 +133,9 @@ def main(build_root='_build'):\n object_inventory_relpath)\n sphinx_mods = set(inventory['py:module'].keys())\n \n- library_dir = os.path.join(BASE_DIR, 'gcloud')\n+ library_dir = os.path.join(BASE_DIR, 'google', 'cloud')\n public_mods = get_public_modules(library_dir,\n- base_package='gcloud')\n+ base_package='google.cloud')\n public_mods = set(public_mods)\n \n if not sphinx_mods <= public_mods:\n@@ -163,8 +163,9 @@ def get_parser():\n :rtype: :class:`argparse.ArgumentParser`\n :returns: The parser for this script.\n \"\"\"\n- parser = argparse.ArgumentParser(\n- description='Run check that all GCloud modules are included in docs.')\n+ description = ('Run check that all google-cloud '\n+ 'modules are included in docs.')\n+ parser = argparse.ArgumentParser(description=description)\n parser.add_argument('--build-root', dest='build_root',\n help='The root directory where docs are located.')\n return parser\ndiff --git a/setup.py b/setup.py\n--- a/setup.py\n+++ b/setup.py\n@@ -32,14 +32,20 @@\n REQUIREMENTS.extend(GRPC_EXTRAS)\n \n setup(\n- name='gcloud',\n+ name='google-cloud',\n version='0.18.0',\n description='API Client library for Google Cloud',\n author='Google Cloud Platform',\n- author_email='jjg+gcloud-python@google.com',\n+ author_email='jjg+google-cloud-python@google.com',\n long_description=README,\n scripts=[],\n- url='https://github.com/GoogleCloudPlatform/gcloud-python',\n+ url='https://github.com/GoogleCloudPlatform/google-cloud-python',\n+ namespace_packages=[\n+ 'google',\n+ 'google.cloud',\n+ 'google.cloud.logging',\n+ 'google.cloud.pubsub',\n+ ],\n packages=find_packages(),\n license='Apache 2.0',\n platforms='Posix; MacOS X; Windows',\n", "test_patch": "diff --git a/system_tests/bigquery.py b/system_tests/bigquery.py\n--- a/system_tests/bigquery.py\n+++ b/system_tests/bigquery.py\n@@ -16,10 +16,10 @@\n \n import unittest\n \n-from gcloud import _helpers\n-from gcloud.environment_vars import TESTS_PROJECT\n-from gcloud import bigquery\n-from gcloud.exceptions import Forbidden\n+from google.cloud import _helpers\n+from google.cloud.environment_vars import TESTS_PROJECT\n+from google.cloud import bigquery\n+from google.cloud.exceptions import Forbidden\n \n from retry import RetryErrors\n from retry import RetryInstanceState\n@@ -244,7 +244,7 @@ def test_update_table(self):\n \n def test_load_table_then_dump_table(self):\n import datetime\n- from gcloud._helpers import UTC\n+ from google.cloud._helpers import UTC\n \n NOW_SECONDS = 1448911495.484366\n NOW = datetime.datetime.utcfromtimestamp(\n@@ -294,7 +294,7 @@ def _has_rows(result):\n def test_load_table_from_storage_then_dump_table(self):\n import csv\n import tempfile\n- from gcloud.storage import Client as StorageClient\n+ from google.cloud.storage import Client as StorageClient\n local_id = unique_resource_id()\n BUCKET_NAME = 'bq_load_test' + local_id\n BLOB_NAME = 'person_ages.csv'\ndiff --git a/system_tests/bigtable.py b/system_tests/bigtable.py\n--- a/system_tests/bigtable.py\n+++ b/system_tests/bigtable.py\n@@ -17,19 +17,19 @@\n \n import unittest\n \n-from gcloud import _helpers\n-from gcloud._helpers import _datetime_from_microseconds\n-from gcloud._helpers import _microseconds_from_datetime\n-from gcloud._helpers import UTC\n-from gcloud.bigtable.client import Client\n-from gcloud.bigtable.column_family import MaxVersionsGCRule\n-from gcloud.bigtable.row_filters import ApplyLabelFilter\n-from gcloud.bigtable.row_filters import ColumnQualifierRegexFilter\n-from gcloud.bigtable.row_filters import RowFilterChain\n-from gcloud.bigtable.row_filters import RowFilterUnion\n-from gcloud.bigtable.row_data import Cell\n-from gcloud.bigtable.row_data import PartialRowData\n-from gcloud.environment_vars import TESTS_PROJECT\n+from google.cloud import _helpers\n+from google.cloud._helpers import _datetime_from_microseconds\n+from google.cloud._helpers import _microseconds_from_datetime\n+from google.cloud._helpers import UTC\n+from google.cloud.bigtable.client import Client\n+from google.cloud.bigtable.column_family import MaxVersionsGCRule\n+from google.cloud.bigtable.row_filters import ApplyLabelFilter\n+from google.cloud.bigtable.row_filters import ColumnQualifierRegexFilter\n+from google.cloud.bigtable.row_filters import RowFilterChain\n+from google.cloud.bigtable.row_filters import RowFilterUnion\n+from google.cloud.bigtable.row_data import Cell\n+from google.cloud.bigtable.row_data import PartialRowData\n+from google.cloud.environment_vars import TESTS_PROJECT\n \n from retry import RetryErrors\n from retry import RetryResult\n@@ -37,8 +37,8 @@\n \n \n LOCATION_ID = 'us-central1-c'\n-INSTANCE_ID = 'gcloud' + unique_resource_id('-')\n-TABLE_ID = 'gcloud-python-test-table'\n+INSTANCE_ID = 'google-cloud' + unique_resource_id('-')\n+TABLE_ID = 'google-cloud-python-test-table'\n COLUMN_FAMILY_ID1 = u'col-fam-id1'\n COLUMN_FAMILY_ID2 = u'col-fam-id2'\n COL_NAME1 = b'col-name1'\n@@ -66,7 +66,7 @@ class Config(object):\n def _wait_until_complete(operation, max_attempts=5):\n \"\"\"Wait until an operation has completed.\n \n- :type operation: :class:`gcloud.bigtable.instance.Operation`\n+ :type operation: :class:`google.cloud.bigtable.instance.Operation`\n :param operation: Operation that has not complete.\n \n :type max_attempts: int\ndiff --git a/system_tests/clear_datastore.py b/system_tests/clear_datastore.py\n--- a/system_tests/clear_datastore.py\n+++ b/system_tests/clear_datastore.py\n@@ -20,8 +20,8 @@\n \n import six\n \n-from gcloud import datastore\n-from gcloud.environment_vars import TESTS_PROJECT\n+from google.cloud import datastore\n+from google.cloud.environment_vars import TESTS_PROJECT\n \n \n FETCH_MAX = 20\n@@ -36,7 +36,7 @@\n \n \n def print_func(message):\n- if os.getenv('GCLOUD_NO_PRINT') != 'true':\n+ if os.getenv('GOOGLE_CLOUD_NO_PRINT') != 'true':\n print(message)\n \n \ndiff --git a/system_tests/datastore.py b/system_tests/datastore.py\n--- a/system_tests/datastore.py\n+++ b/system_tests/datastore.py\n@@ -18,13 +18,13 @@\n \n import httplib2\n \n-from gcloud import _helpers\n-from gcloud._helpers import UTC\n-from gcloud import datastore\n-from gcloud.datastore.helpers import GeoPoint\n-from gcloud.environment_vars import GCD_DATASET\n-from gcloud.environment_vars import TESTS_PROJECT\n-from gcloud.exceptions import Conflict\n+from google.cloud import _helpers\n+from google.cloud._helpers import UTC\n+from google.cloud import datastore\n+from google.cloud.datastore.helpers import GeoPoint\n+from google.cloud.environment_vars import GCD_DATASET\n+from google.cloud.environment_vars import TESTS_PROJECT\n+from google.cloud.exceptions import Conflict\n \n import clear_datastore\n import populate_datastore\n@@ -433,7 +433,8 @@ def test_transaction_via_with_statement(self):\n self.assertEqual(retrieved_entity, entity)\n \n def test_transaction_via_explicit_begin_get_commit(self):\n- # See https://github.com/GoogleCloudPlatform/gcloud-python/issues/1859\n+ # See\n+ # github.com/GoogleCloudPlatform/google-cloud-python/issues/1859\n # Note that this example lacks the threading which provokes the race\n # condition in that issue: we are basically just exercising the\n # \"explict\" path for using transactions.\ndiff --git a/system_tests/language.py b/system_tests/language.py\n--- a/system_tests/language.py\n+++ b/system_tests/language.py\n@@ -14,11 +14,11 @@\n \n import unittest\n \n-from gcloud import _helpers\n-from gcloud.environment_vars import TESTS_PROJECT\n-from gcloud import exceptions\n-from gcloud import language\n-from gcloud import storage\n+from google.cloud import _helpers\n+from google.cloud.environment_vars import TESTS_PROJECT\n+from google.cloud import exceptions\n+from google.cloud import language\n+from google.cloud import storage\n \n from system_test_utils import unique_resource_id\n from retry import RetryErrors\n@@ -67,7 +67,7 @@ def tearDown(self):\n value.delete()\n \n def _check_analyze_entities_result(self, entities):\n- from gcloud.language.entity import EntityType\n+ from google.cloud.language.entity import EntityType\n \n self.assertEqual(len(entities), 3)\n entity1, entity2, entity3 = entities\ndiff --git a/system_tests/local_test_setup.sample b/system_tests/local_test_setup.sample\n--- a/system_tests/local_test_setup.sample\n+++ b/system_tests/local_test_setup.sample\n@@ -1,4 +1,4 @@\n export GOOGLE_APPLICATION_CREDENTIALS=\"app_credentials.json.sample\"\n-export GCLOUD_TESTS_PROJECT_ID=\"my-project\"\n-export GCLOUD_REMOTE_FOR_LINT=\"upstream\"\n-export GCLOUD_BRANCH_FOR_LINT=\"master\"\n+export GOOGLE_CLOUD_TESTS_PROJECT_ID=\"my-project\"\n+export GOOGLE_CLOUD_REMOTE_FOR_LINT=\"upstream\"\n+export GOOGLE_CLOUD_BRANCH_FOR_LINT=\"master\"\ndiff --git a/system_tests/logging_.py b/system_tests/logging_.py\n--- a/system_tests/logging_.py\n+++ b/system_tests/logging_.py\n@@ -15,12 +15,12 @@\n import logging\n import unittest\n \n-import gcloud.logging\n-import gcloud.logging.handlers.handlers\n-from gcloud.logging.handlers.handlers import CloudLoggingHandler\n-from gcloud.logging.handlers.transports import SyncTransport\n-from gcloud import _helpers\n-from gcloud.environment_vars import TESTS_PROJECT\n+import google.cloud.logging\n+import google.cloud.logging.handlers.handlers\n+from google.cloud.logging.handlers.handlers import CloudLoggingHandler\n+from google.cloud.logging.handlers.transports import SyncTransport\n+from google.cloud import _helpers\n+from google.cloud.environment_vars import TESTS_PROJECT\n \n from retry import RetryErrors\n from retry import RetryResult\n@@ -31,9 +31,9 @@\n DEFAULT_SINK_NAME = 'system-tests-sink%s' % (_RESOURCE_ID,)\n DEFAULT_FILTER = 'logName:syslog AND severity>=INFO'\n DEFAULT_DESCRIPTION = 'System testing'\n-BUCKET_NAME = 'gcloud-python-system-testing%s' % (_RESOURCE_ID,)\n+BUCKET_NAME = 'google-cloud-python-system-testing%s' % (_RESOURCE_ID,)\n DATASET_NAME = ('system_testing_dataset' + _RESOURCE_ID).replace('-', '_')\n-TOPIC_NAME = 'gcloud-python-system-testing%s' % (_RESOURCE_ID,)\n+TOPIC_NAME = 'google-cloud-python-system-testing%s' % (_RESOURCE_ID,)\n \n \n def _retry_on_unavailable(exc):\n@@ -57,7 +57,7 @@ class Config(object):\n \n def setUpModule():\n _helpers.PROJECT = TESTS_PROJECT\n- Config.CLIENT = gcloud.logging.Client()\n+ Config.CLIENT = google.cloud.logging.Client()\n \n \n class TestLogging(unittest.TestCase):\n@@ -67,7 +67,7 @@ def setUp(self):\n self._handlers_cache = logging.getLogger().handlers[:]\n \n def tearDown(self):\n- from gcloud.exceptions import NotFound\n+ from google.cloud.exceptions import NotFound\n retry = RetryErrors(NotFound)\n for doomed in self.to_delete:\n retry(doomed.delete)()\n@@ -188,7 +188,7 @@ def test_log_root_handler(self):\n logger = Config.CLIENT.logger(handler.name)\n self.to_delete.append(logger)\n \n- gcloud.logging.handlers.handlers.setup_logging(handler)\n+ google.cloud.logging.handlers.handlers.setup_logging(handler)\n logging.warn(LOG_MESSAGE)\n \n entries, _ = self._list_entries(logger)\n@@ -254,7 +254,7 @@ def test_list_metrics(self):\n set([DEFAULT_METRIC_NAME]))\n \n def test_reload_metric(self):\n- from gcloud.exceptions import Conflict\n+ from google.cloud.exceptions import Conflict\n retry = RetryErrors(Conflict)\n metric = Config.CLIENT.metric(\n DEFAULT_METRIC_NAME, DEFAULT_FILTER, DEFAULT_DESCRIPTION)\n@@ -268,7 +268,7 @@ def test_reload_metric(self):\n self.assertEqual(metric.description, DEFAULT_DESCRIPTION)\n \n def test_update_metric(self):\n- from gcloud.exceptions import Conflict\n+ from google.cloud.exceptions import Conflict\n retry = RetryErrors(Conflict)\n NEW_FILTER = 'logName:other'\n NEW_DESCRIPTION = 'updated'\n@@ -287,7 +287,7 @@ def test_update_metric(self):\n self.assertEqual(after.description, NEW_DESCRIPTION)\n \n def _init_storage_bucket(self):\n- from gcloud import storage\n+ from google.cloud import storage\n BUCKET_URI = 'storage.googleapis.com/%s' % (BUCKET_NAME,)\n \n # Create the destination bucket, and set up the ACL to allow\n@@ -313,7 +313,7 @@ def test_create_sink_storage_bucket(self):\n self.assertTrue(sink.exists())\n \n def test_create_sink_pubsub_topic(self):\n- from gcloud import pubsub\n+ from google.cloud import pubsub\n \n # Create the destination topic, and set up the IAM policy to allow\n # Stackdriver Logging to write into it.\n@@ -335,8 +335,8 @@ def test_create_sink_pubsub_topic(self):\n self.assertTrue(sink.exists())\n \n def _init_bigquery_dataset(self):\n- from gcloud import bigquery\n- from gcloud.bigquery.dataset import AccessGrant\n+ from google.cloud import bigquery\n+ from google.cloud.bigquery.dataset import AccessGrant\n DATASET_URI = 'bigquery.googleapis.com/projects/%s/datasets/%s' % (\n Config.CLIENT.project, DATASET_NAME,)\n \n@@ -377,7 +377,7 @@ def test_list_sinks(self):\n set([DEFAULT_SINK_NAME]))\n \n def test_reload_sink(self):\n- from gcloud.exceptions import Conflict\n+ from google.cloud.exceptions import Conflict\n retry = RetryErrors(Conflict)\n uri = self._init_bigquery_dataset()\n sink = Config.CLIENT.sink(DEFAULT_SINK_NAME, DEFAULT_FILTER, uri)\n@@ -391,7 +391,7 @@ def test_reload_sink(self):\n self.assertEqual(sink.destination, uri)\n \n def test_update_sink(self):\n- from gcloud.exceptions import Conflict\n+ from google.cloud.exceptions import Conflict\n retry = RetryErrors(Conflict)\n bucket_uri = self._init_storage_bucket()\n dataset_uri = self._init_bigquery_dataset()\ndiff --git a/system_tests/monitoring.py b/system_tests/monitoring.py\n--- a/system_tests/monitoring.py\n+++ b/system_tests/monitoring.py\n@@ -14,12 +14,12 @@\n \n import unittest\n \n-from gcloud import _helpers\n-from gcloud.environment_vars import TESTS_PROJECT\n-from gcloud.exceptions import InternalServerError\n-from gcloud.exceptions import NotFound\n-from gcloud.exceptions import ServiceUnavailable\n-from gcloud import monitoring\n+from google.cloud import _helpers\n+from google.cloud.environment_vars import TESTS_PROJECT\n+from google.cloud.exceptions import InternalServerError\n+from google.cloud.exceptions import NotFound\n+from google.cloud.exceptions import ServiceUnavailable\n+from google.cloud import monitoring\n \n from retry import RetryErrors\n from system_test_utils import unique_resource_id\ndiff --git a/system_tests/populate_datastore.py b/system_tests/populate_datastore.py\n--- a/system_tests/populate_datastore.py\n+++ b/system_tests/populate_datastore.py\n@@ -21,8 +21,8 @@\n \n import six\n \n-from gcloud import datastore\n-from gcloud.environment_vars import TESTS_PROJECT\n+from google.cloud import datastore\n+from google.cloud.environment_vars import TESTS_PROJECT\n \n \n ANCESTOR = ('Book', 'GoT')\n@@ -84,7 +84,7 @@\n \n \n def print_func(message):\n- if os.getenv('GCLOUD_NO_PRINT') != 'true':\n+ if os.getenv('GOOGLE_CLOUD_NO_PRINT') != 'true':\n print(message)\n \n \ndiff --git a/system_tests/pubsub.py b/system_tests/pubsub.py\n--- a/system_tests/pubsub.py\n+++ b/system_tests/pubsub.py\n@@ -20,10 +20,12 @@\n from grpc._channel import _Rendezvous\n import httplib2\n \n-from gcloud import _helpers\n-from gcloud.environment_vars import PUBSUB_EMULATOR\n-from gcloud.environment_vars import TESTS_PROJECT\n-from gcloud import pubsub\n+# pylint: disable=ungrouped-imports\n+from google.cloud import _helpers\n+from google.cloud.environment_vars import PUBSUB_EMULATOR\n+from google.cloud.environment_vars import TESTS_PROJECT\n+from google.cloud import pubsub\n+# pylint: enable=ungrouped-imports\n \n from retry import RetryInstanceState\n from retry import RetryResult\n@@ -213,7 +215,7 @@ def _maybe_emulator_skip(self):\n self.skipTest('IAM not supported by Pub/Sub emulator')\n \n def test_topic_iam_policy(self):\n- from gcloud.pubsub.iam import PUBSUB_TOPICS_GET_IAM_POLICY\n+ from google.cloud.pubsub.iam import PUBSUB_TOPICS_GET_IAM_POLICY\n self._maybe_emulator_skip()\n topic_name = 'test-topic-iam-policy-topic' + unique_resource_id('-')\n topic = Config.CLIENT.topic(topic_name)\n@@ -233,7 +235,7 @@ def test_topic_iam_policy(self):\n self.assertEqual(new_policy.viewers, policy.viewers)\n \n def test_subscription_iam_policy(self):\n- from gcloud.pubsub.iam import PUBSUB_SUBSCRIPTIONS_GET_IAM_POLICY\n+ from google.cloud.pubsub.iam import PUBSUB_SUBSCRIPTIONS_GET_IAM_POLICY\n self._maybe_emulator_skip()\n topic_name = 'test-sub-iam-policy-topic' + unique_resource_id('-')\n topic = Config.CLIENT.topic(topic_name)\n@@ -264,10 +266,10 @@ def test_subscription_iam_policy(self):\n self.assertEqual(new_policy.viewers, policy.viewers)\n \n # This test is ultra-flaky. See:\n- # https://github.com/GoogleCloudPlatform/gcloud-python/issues/2080\n+ # https://github.com/GoogleCloudPlatform/google-cloud-python/issues/2080\n @unittest.expectedFailure\n def test_fetch_delete_subscription_w_deleted_topic(self):\n- from gcloud.iterator import MethodIterator\n+ from google.cloud.iterator import MethodIterator\n TO_DELETE = 'delete-me' + unique_resource_id('-')\n ORPHANED = 'orphaned' + unique_resource_id('-')\n topic = Config.CLIENT.topic(TO_DELETE)\ndiff --git a/system_tests/run_emulator.py b/system_tests/run_emulator.py\n--- a/system_tests/run_emulator.py\n+++ b/system_tests/run_emulator.py\n@@ -25,9 +25,9 @@\n \n import psutil\n \n-from gcloud.environment_vars import GCD_DATASET\n-from gcloud.environment_vars import GCD_HOST\n-from gcloud.environment_vars import PUBSUB_EMULATOR\n+from google.cloud.environment_vars import GCD_DATASET\n+from google.cloud.environment_vars import GCD_HOST\n+from google.cloud.environment_vars import PUBSUB_EMULATOR\n from run_system_test import run_module_tests\n \n \n@@ -46,7 +46,7 @@ def get_parser():\n :returns: The parser for this script.\n \"\"\"\n parser = argparse.ArgumentParser(\n- description='Run GCloud system tests against local emulator.')\n+ description='Run google-cloud system tests against local emulator.')\n parser.add_argument('--package', dest='package',\n choices=('datastore', 'pubsub'),\n default='datastore', help='Package to be tested.')\ndiff --git a/system_tests/run_system_test.py b/system_tests/run_system_test.py\n--- a/system_tests/run_system_test.py\n+++ b/system_tests/run_system_test.py\n@@ -47,7 +47,7 @@ class FailedSystemTestModule(Exception):\n \n def get_parser():\n parser = argparse.ArgumentParser(\n- description='GCloud test runner against actual project.')\n+ description='google-cloud test runner against actual project.')\n parser.add_argument('--package', dest='package',\n choices=TEST_MODULES.keys(),\n default='datastore', help='Package to be tested.')\ndiff --git a/system_tests/storage.py b/system_tests/storage.py\n--- a/system_tests/storage.py\n+++ b/system_tests/storage.py\n@@ -20,11 +20,11 @@\n import httplib2\n import six\n \n-from gcloud import _helpers\n-from gcloud.environment_vars import TESTS_PROJECT\n-from gcloud import exceptions\n-from gcloud import storage\n-from gcloud.storage._helpers import _base64_md5hash\n+from google.cloud import _helpers\n+from google.cloud.environment_vars import TESTS_PROJECT\n+from google.cloud import exceptions\n+from google.cloud import storage\n+from google.cloud.storage._helpers import _base64_md5hash\n \n from system_test_utils import unique_resource_id\n from retry import RetryErrors\ndiff --git a/system_tests/system_test_utils.py b/system_tests/system_test_utils.py\n--- a/system_tests/system_test_utils.py\n+++ b/system_tests/system_test_utils.py\n@@ -17,8 +17,8 @@\n import sys\n import time\n \n-from gcloud.environment_vars import CREDENTIALS as TEST_CREDENTIALS\n-from gcloud.environment_vars import TESTS_PROJECT\n+from google.cloud.environment_vars import CREDENTIALS as TEST_CREDENTIALS\n+from google.cloud.environment_vars import TESTS_PROJECT\n \n \n # From shell environ. May be None.\ndiff --git a/system_tests/translate.py b/system_tests/translate.py\n--- a/system_tests/translate.py\n+++ b/system_tests/translate.py\n@@ -17,10 +17,10 @@\n \n import unittest\n \n-from gcloud import translate\n+from google.cloud import translate\n \n \n-ENV_VAR = 'GCLOUD_TESTS_API_KEY'\n+ENV_VAR = 'GOOGLE_CLOUD_TESTS_API_KEY'\n \n \n class Config(object):\ndiff --git a/unit_tests/bigquery/test__helpers.py b/unit_tests/bigquery/test__helpers.py\n--- a/unit_tests/bigquery/test__helpers.py\n+++ b/unit_tests/bigquery/test__helpers.py\n@@ -18,7 +18,7 @@\n class Test_ConfigurationProperty(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.bigquery._helpers import _ConfigurationProperty\n+ from google.cloud.bigquery._helpers import _ConfigurationProperty\n return _ConfigurationProperty\n \n def _makeOne(self, *args, **kw):\n@@ -53,7 +53,7 @@ def __init__(self):\n class Test_TypedProperty(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.bigquery._helpers import _TypedProperty\n+ from google.cloud.bigquery._helpers import _TypedProperty\n return _TypedProperty\n \n def _makeOne(self, *args, **kw):\n@@ -86,7 +86,7 @@ def __init__(self):\n class Test_EnumProperty(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.bigquery._helpers import _EnumProperty\n+ from google.cloud.bigquery._helpers import _EnumProperty\n return _EnumProperty\n \n def test_it(self):\ndiff --git a/unit_tests/bigquery/test_client.py b/unit_tests/bigquery/test_client.py\n--- a/unit_tests/bigquery/test_client.py\n+++ b/unit_tests/bigquery/test_client.py\n@@ -18,14 +18,14 @@\n class TestClient(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.bigquery.client import Client\n+ from google.cloud.bigquery.client import Client\n return Client\n \n def _makeOne(self, *args, **kw):\n return self._getTargetClass()(*args, **kw)\n \n def test_ctor(self):\n- from gcloud.bigquery.connection import Connection\n+ from google.cloud.bigquery.connection import Connection\n PROJECT = 'PROJECT'\n creds = _Credentials()\n http = object()\n@@ -35,7 +35,7 @@ def test_ctor(self):\n self.assertTrue(client.connection.http is http)\n \n def test_list_datasets_defaults(self):\n- from gcloud.bigquery.dataset import Dataset\n+ from google.cloud.bigquery.dataset import Dataset\n PROJECT = 'PROJECT'\n DATASET_1 = 'dataset_one'\n DATASET_2 = 'dataset_two'\n@@ -97,7 +97,7 @@ def test_list_datasets_explicit_response_missing_datasets_key(self):\n {'all': True, 'maxResults': 3, 'pageToken': TOKEN})\n \n def test_dataset(self):\n- from gcloud.bigquery.dataset import Dataset\n+ from google.cloud.bigquery.dataset import Dataset\n PROJECT = 'PROJECT'\n DATASET = 'dataset_name'\n creds = _Credentials()\n@@ -116,10 +116,10 @@ def test_job_from_resource_unknown_type(self):\n client.job_from_resource({'configuration': {'nonesuch': {}}})\n \n def test_list_jobs_defaults(self):\n- from gcloud.bigquery.job import LoadTableFromStorageJob\n- from gcloud.bigquery.job import CopyJob\n- from gcloud.bigquery.job import ExtractTableToStorageJob\n- from gcloud.bigquery.job import QueryJob\n+ from google.cloud.bigquery.job import LoadTableFromStorageJob\n+ from google.cloud.bigquery.job import CopyJob\n+ from google.cloud.bigquery.job import ExtractTableToStorageJob\n+ from google.cloud.bigquery.job import QueryJob\n PROJECT = 'PROJECT'\n DATASET = 'test_dataset'\n SOURCE_TABLE = 'source_table'\n@@ -243,7 +243,7 @@ def test_list_jobs_defaults(self):\n self.assertEqual(req['query_params'], {'projection': 'full'})\n \n def test_list_jobs_load_job_wo_sourceUris(self):\n- from gcloud.bigquery.job import LoadTableFromStorageJob\n+ from google.cloud.bigquery.job import LoadTableFromStorageJob\n PROJECT = 'PROJECT'\n DATASET = 'test_dataset'\n SOURCE_TABLE = 'source_table'\n@@ -321,7 +321,7 @@ def test_list_jobs_explicit_missing(self):\n 'stateFilter': 'done'})\n \n def test_load_table_from_storage(self):\n- from gcloud.bigquery.job import LoadTableFromStorageJob\n+ from google.cloud.bigquery.job import LoadTableFromStorageJob\n PROJECT = 'PROJECT'\n JOB = 'job_name'\n DATASET = 'dataset_name'\n@@ -340,7 +340,7 @@ def test_load_table_from_storage(self):\n self.assertTrue(job.destination is destination)\n \n def test_copy_table(self):\n- from gcloud.bigquery.job import CopyJob\n+ from google.cloud.bigquery.job import CopyJob\n PROJECT = 'PROJECT'\n JOB = 'job_name'\n DATASET = 'dataset_name'\n@@ -360,7 +360,7 @@ def test_copy_table(self):\n self.assertTrue(job.destination is destination)\n \n def test_extract_table_to_storage(self):\n- from gcloud.bigquery.job import ExtractTableToStorageJob\n+ from google.cloud.bigquery.job import ExtractTableToStorageJob\n PROJECT = 'PROJECT'\n JOB = 'job_name'\n DATASET = 'dataset_name'\n@@ -379,7 +379,7 @@ def test_extract_table_to_storage(self):\n self.assertEqual(list(job.destination_uris), [DESTINATION])\n \n def test_run_async_query(self):\n- from gcloud.bigquery.job import QueryJob\n+ from google.cloud.bigquery.job import QueryJob\n PROJECT = 'PROJECT'\n JOB = 'job_name'\n QUERY = 'select count(*) from persons'\n@@ -393,7 +393,7 @@ def test_run_async_query(self):\n self.assertEqual(job.query, QUERY)\n \n def test_run_sync_query(self):\n- from gcloud.bigquery.query import QueryResults\n+ from google.cloud.bigquery.query import QueryResults\n PROJECT = 'PROJECT'\n QUERY = 'select count(*) from persons'\n creds = _Credentials()\ndiff --git a/unit_tests/bigquery/test_connection.py b/unit_tests/bigquery/test_connection.py\n--- a/unit_tests/bigquery/test_connection.py\n+++ b/unit_tests/bigquery/test_connection.py\n@@ -18,7 +18,7 @@\n class TestConnection(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.bigquery.connection import Connection\n+ from google.cloud.bigquery.connection import Connection\n return Connection\n \n def _makeOne(self, *args, **kw):\ndiff --git a/unit_tests/bigquery/test_dataset.py b/unit_tests/bigquery/test_dataset.py\n--- a/unit_tests/bigquery/test_dataset.py\n+++ b/unit_tests/bigquery/test_dataset.py\n@@ -18,7 +18,7 @@\n class TestAccessGrant(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.bigquery.dataset import AccessGrant\n+ from google.cloud.bigquery.dataset import AccessGrant\n return AccessGrant\n \n def _makeOne(self, *args, **kw):\n@@ -81,7 +81,7 @@ class TestDataset(unittest.TestCase):\n DS_NAME = 'dataset-name'\n \n def _getTargetClass(self):\n- from gcloud.bigquery.dataset import Dataset\n+ from google.cloud.bigquery.dataset import Dataset\n return Dataset\n \n def _makeOne(self, *args, **kw):\n@@ -89,7 +89,7 @@ def _makeOne(self, *args, **kw):\n \n def _setUpConstants(self):\n import datetime\n- from gcloud._helpers import UTC\n+ from google.cloud._helpers import UTC\n \n self.WHEN_TS = 1437767599.006\n self.WHEN = datetime.datetime.utcfromtimestamp(self.WHEN_TS).replace(\n@@ -201,7 +201,7 @@ def test_access_roles_setter_non_list(self):\n dataset.access_grants = object()\n \n def test_access_roles_setter_invalid_field(self):\n- from gcloud.bigquery.dataset import AccessGrant\n+ from google.cloud.bigquery.dataset import AccessGrant\n client = _Client(self.PROJECT)\n dataset = self._makeOne(self.DS_NAME, client)\n phred = AccessGrant('OWNER', 'userByEmail', 'phred@example.com')\n@@ -209,7 +209,7 @@ def test_access_roles_setter_invalid_field(self):\n dataset.access_grants = [phred, object()]\n \n def test_access_roles_setter(self):\n- from gcloud.bigquery.dataset import AccessGrant\n+ from google.cloud.bigquery.dataset import AccessGrant\n client = _Client(self.PROJECT)\n dataset = self._makeOne(self.DS_NAME, client)\n phred = AccessGrant('OWNER', 'userByEmail', 'phred@example.com')\n@@ -340,7 +340,7 @@ def test_create_w_bound_client(self):\n self._verifyResourceProperties(dataset, RESOURCE)\n \n def test_create_w_alternate_client(self):\n- from gcloud.bigquery.dataset import AccessGrant\n+ from google.cloud.bigquery.dataset import AccessGrant\n PATH = 'projects/%s/datasets' % self.PROJECT\n USER_EMAIL = 'phred@example.com'\n GROUP_EMAIL = 'group-name@lists.example.com'\n@@ -649,7 +649,7 @@ def test_list_tables_empty(self):\n self.assertEqual(req['path'], '/%s' % PATH)\n \n def test_list_tables_defaults(self):\n- from gcloud.bigquery.table import Table\n+ from google.cloud.bigquery.table import Table\n \n TABLE_1 = 'table_one'\n TABLE_2 = 'table_two'\n@@ -692,7 +692,7 @@ def test_list_tables_defaults(self):\n self.assertEqual(req['path'], '/%s' % PATH)\n \n def test_list_tables_explicit(self):\n- from gcloud.bigquery.table import Table\n+ from google.cloud.bigquery.table import Table\n \n TABLE_1 = 'table_one'\n TABLE_2 = 'table_two'\n@@ -736,7 +736,7 @@ def test_list_tables_explicit(self):\n {'maxResults': 3, 'pageToken': TOKEN})\n \n def test_table_wo_schema(self):\n- from gcloud.bigquery.table import Table\n+ from google.cloud.bigquery.table import Table\n conn = _Connection({})\n client = _Client(project=self.PROJECT, connection=conn)\n dataset = self._makeOne(self.DS_NAME, client=client)\n@@ -747,8 +747,8 @@ def test_table_wo_schema(self):\n self.assertEqual(table.schema, [])\n \n def test_table_w_schema(self):\n- from gcloud.bigquery.schema import SchemaField\n- from gcloud.bigquery.table import Table\n+ from google.cloud.bigquery.schema import SchemaField\n+ from google.cloud.bigquery.table import Table\n conn = _Connection({})\n client = _Client(project=self.PROJECT, connection=conn)\n dataset = self._makeOne(self.DS_NAME, client=client)\n@@ -775,7 +775,7 @@ def __init__(self, *responses):\n self._requested = []\n \n def api_request(self, **kw):\n- from gcloud.exceptions import NotFound\n+ from google.cloud.exceptions import NotFound\n self._requested.append(kw)\n \n try:\ndiff --git a/unit_tests/bigquery/test_job.py b/unit_tests/bigquery/test_job.py\n--- a/unit_tests/bigquery/test_job.py\n+++ b/unit_tests/bigquery/test_job.py\n@@ -18,7 +18,7 @@\n class Test_UDFResourcesProperty(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.bigquery.job import UDFResourcesProperty\n+ from google.cloud.bigquery.job import UDFResourcesProperty\n return UDFResourcesProperty\n \n def _makeOne(self, *args, **kw):\n@@ -43,7 +43,7 @@ def test_instance_getter_empty(self):\n self.assertEqual(instance.udf_resources, [])\n \n def test_instance_getter_w_non_empty_list(self):\n- from gcloud.bigquery.job import UDFResource\n+ from google.cloud.bigquery.job import UDFResource\n RESOURCE_URI = 'gs://some-bucket/js/lib.js'\n udf_resources = [UDFResource(\"resourceUri\", RESOURCE_URI)]\n _, klass = self._descriptor_and_klass()\n@@ -53,7 +53,7 @@ def test_instance_getter_w_non_empty_list(self):\n self.assertEqual(instance.udf_resources, udf_resources)\n \n def test_instance_setter_w_empty_list(self):\n- from gcloud.bigquery.job import UDFResource\n+ from google.cloud.bigquery.job import UDFResource\n RESOURCE_URI = 'gs://some-bucket/js/lib.js'\n udf_resources = [UDFResource(\"resourceUri\", RESOURCE_URI)]\n _, klass = self._descriptor_and_klass()\n@@ -65,7 +65,7 @@ def test_instance_setter_w_empty_list(self):\n self.assertEqual(instance.udf_resources, [])\n \n def test_instance_setter_w_valid_udf(self):\n- from gcloud.bigquery.job import UDFResource\n+ from google.cloud.bigquery.job import UDFResource\n RESOURCE_URI = 'gs://some-bucket/js/lib.js'\n udf_resources = [UDFResource(\"resourceUri\", RESOURCE_URI)]\n _, klass = self._descriptor_and_klass()\n@@ -97,7 +97,7 @@ def _makeOne(self, *args, **kw):\n \n def _setUpConstants(self):\n import datetime\n- from gcloud._helpers import UTC\n+ from google.cloud._helpers import UTC\n \n self.WHEN_TS = 1437767599.006\n self.WHEN = datetime.datetime.utcfromtimestamp(self.WHEN_TS).replace(\n@@ -193,7 +193,7 @@ class TestLoadTableFromStorageJob(unittest.TestCase, _Base):\n JOB_TYPE = 'load'\n \n def _getTargetClass(self):\n- from gcloud.bigquery.job import LoadTableFromStorageJob\n+ from google.cloud.bigquery.job import LoadTableFromStorageJob\n return LoadTableFromStorageJob\n \n def _setUpConstants(self):\n@@ -332,7 +332,7 @@ def test_ctor(self):\n self.assertTrue(job.write_disposition is None)\n \n def test_ctor_w_schema(self):\n- from gcloud.bigquery.schema import SchemaField\n+ from google.cloud.bigquery.schema import SchemaField\n client = _Client(self.PROJECT)\n table = _Table()\n full_name = SchemaField('full_name', 'STRING', mode='REQUIRED')\n@@ -349,7 +349,7 @@ def test_schema_setter_non_list(self):\n job.schema = object()\n \n def test_schema_setter_invalid_field(self):\n- from gcloud.bigquery.schema import SchemaField\n+ from google.cloud.bigquery.schema import SchemaField\n client = _Client(self.PROJECT)\n table = _Table()\n job = self._makeOne(self.JOB_NAME, table, [self.SOURCE1], client)\n@@ -358,7 +358,7 @@ def test_schema_setter_invalid_field(self):\n job.schema = [full_name, object()]\n \n def test_schema_setter(self):\n- from gcloud.bigquery.schema import SchemaField\n+ from google.cloud.bigquery.schema import SchemaField\n client = _Client(self.PROJECT)\n table = _Table()\n job = self._makeOne(self.JOB_NAME, table, [self.SOURCE1], client)\n@@ -369,8 +369,8 @@ def test_schema_setter(self):\n \n def test_props_set_by_server(self):\n import datetime\n- from gcloud._helpers import UTC\n- from gcloud._helpers import _millis\n+ from google.cloud._helpers import UTC\n+ from google.cloud._helpers import _millis\n \n CREATED = datetime.datetime(2015, 8, 11, 12, 13, 22, tzinfo=UTC)\n STARTED = datetime.datetime(2015, 8, 11, 13, 47, 15, tzinfo=UTC)\n@@ -523,7 +523,7 @@ def test_begin_w_bound_client(self):\n self._verifyResourceProperties(job, RESOURCE)\n \n def test_begin_w_alternate_client(self):\n- from gcloud.bigquery.schema import SchemaField\n+ from google.cloud.bigquery.schema import SchemaField\n PATH = 'projects/%s/jobs' % self.PROJECT\n RESOURCE = self._makeResource(ended=True)\n LOAD_CONFIGURATION = {\n@@ -701,7 +701,7 @@ class TestCopyJob(unittest.TestCase, _Base):\n DESTINATION_TABLE = 'destination_table'\n \n def _getTargetClass(self):\n- from gcloud.bigquery.job import CopyJob\n+ from google.cloud.bigquery.job import CopyJob\n return CopyJob\n \n def _makeResource(self, started=False, ended=False):\n@@ -998,7 +998,7 @@ class TestExtractTableToStorageJob(unittest.TestCase, _Base):\n DESTINATION_URI = 'gs://bucket_name/object_name'\n \n def _getTargetClass(self):\n- from gcloud.bigquery.job import ExtractTableToStorageJob\n+ from google.cloud.bigquery.job import ExtractTableToStorageJob\n return ExtractTableToStorageJob\n \n def _makeResource(self, started=False, ended=False):\n@@ -1291,7 +1291,7 @@ class TestQueryJob(unittest.TestCase, _Base):\n DESTINATION_TABLE = 'destination_table'\n \n def _getTargetClass(self):\n- from gcloud.bigquery.job import QueryJob\n+ from google.cloud.bigquery.job import QueryJob\n return QueryJob\n \n def _makeResource(self, started=False, ended=False):\n@@ -1474,8 +1474,8 @@ def test_begin_w_bound_client(self):\n self._verifyResourceProperties(job, RESOURCE)\n \n def test_begin_w_alternate_client(self):\n- from gcloud.bigquery.dataset import Dataset\n- from gcloud.bigquery.dataset import Table\n+ from google.cloud.bigquery.dataset import Dataset\n+ from google.cloud.bigquery.dataset import Table\n PATH = 'projects/%s/jobs' % self.PROJECT\n TABLE = 'TABLE'\n DS_NAME = 'DATASET'\n@@ -1539,7 +1539,7 @@ def test_begin_w_alternate_client(self):\n self._verifyResourceProperties(job, RESOURCE)\n \n def test_begin_w_bound_client_and_udf(self):\n- from gcloud.bigquery.job import UDFResource\n+ from google.cloud.bigquery.job import UDFResource\n RESOURCE_URI = 'gs://some-bucket/js/lib.js'\n PATH = 'projects/%s/jobs' % self.PROJECT\n RESOURCE = self._makeResource()\n@@ -1611,8 +1611,8 @@ def test_exists_hit_w_alternate_client(self):\n self.assertEqual(req['query_params'], {'fields': 'id'})\n \n def test_reload_w_bound_client(self):\n- from gcloud.bigquery.dataset import Dataset\n- from gcloud.bigquery.dataset import Table\n+ from google.cloud.bigquery.dataset import Dataset\n+ from google.cloud.bigquery.dataset import Table\n PATH = 'projects/%s/jobs/%s' % (self.PROJECT, self.JOB_NAME)\n DS_NAME = 'DATASET'\n DEST_TABLE = 'dest_table'\n@@ -1669,7 +1669,7 @@ def __init__(self, project='project', connection=None):\n self.connection = connection\n \n def dataset(self, name):\n- from gcloud.bigquery.dataset import Dataset\n+ from google.cloud.bigquery.dataset import Dataset\n return Dataset(name, client=self)\n \n \n@@ -1700,7 +1700,7 @@ def __init__(self, *responses):\n self._requested = []\n \n def api_request(self, **kw):\n- from gcloud.exceptions import NotFound\n+ from google.cloud.exceptions import NotFound\n self._requested.append(kw)\n \n try:\ndiff --git a/unit_tests/bigquery/test_query.py b/unit_tests/bigquery/test_query.py\n--- a/unit_tests/bigquery/test_query.py\n+++ b/unit_tests/bigquery/test_query.py\n@@ -24,7 +24,7 @@ class TestQueryResults(unittest.TestCase):\n TOKEN = 'TOKEN'\n \n def _getTargetClass(self):\n- from gcloud.bigquery.query import QueryResults\n+ from google.cloud.bigquery.query import QueryResults\n return QueryResults\n \n def _makeOne(self, *args, **kw):\n@@ -73,7 +73,7 @@ def _makeResource(self, complete=False):\n return resource\n \n def _verifySchema(self, query, resource):\n- from gcloud.bigquery.schema import SchemaField\n+ from google.cloud.bigquery.schema import SchemaField\n if 'schema' in resource:\n fields = resource['schema']['fields']\n self.assertEqual(len(query.schema), len(fields))\n@@ -144,7 +144,7 @@ def test_job_wo_jobid(self):\n self.assertTrue(query.job is None)\n \n def test_job_w_jobid(self):\n- from gcloud.bigquery.job import QueryJob\n+ from google.cloud.bigquery.job import QueryJob\n SERVER_GENERATED = 'SERVER_GENERATED'\n client = _Client(self.PROJECT)\n query = self._makeOne(self.QUERY, client)\n@@ -234,7 +234,7 @@ def test_run_w_alternate_client(self):\n self._verifyResourceProperties(query, RESOURCE)\n \n def test_run_w_inline_udf(self):\n- from gcloud.bigquery.job import UDFResource\n+ from google.cloud.bigquery.job import UDFResource\n INLINE_UDF_CODE = 'var someCode = \"here\";'\n PATH = 'projects/%s/queries' % self.PROJECT\n RESOURCE = self._makeResource(complete=False)\n@@ -256,7 +256,7 @@ def test_run_w_inline_udf(self):\n self._verifyResourceProperties(query, RESOURCE)\n \n def test_run_w_udf_resource_uri(self):\n- from gcloud.bigquery.job import UDFResource\n+ from google.cloud.bigquery.job import UDFResource\n RESOURCE_URI = 'gs://some-bucket/js/lib.js'\n PATH = 'projects/%s/queries' % self.PROJECT\n RESOURCE = self._makeResource(complete=False)\n@@ -278,7 +278,7 @@ def test_run_w_udf_resource_uri(self):\n self._verifyResourceProperties(query, RESOURCE)\n \n def test_run_w_mixed_udfs(self):\n- from gcloud.bigquery.job import UDFResource\n+ from google.cloud.bigquery.job import UDFResource\n RESOURCE_URI = 'gs://some-bucket/js/lib.js'\n INLINE_UDF_CODE = 'var someCode = \"here\";'\n PATH = 'projects/%s/queries' % self.PROJECT\n@@ -388,7 +388,7 @@ def __init__(self, project='project', connection=None):\n self.connection = connection\n \n def dataset(self, name):\n- from gcloud.bigquery.dataset import Dataset\n+ from google.cloud.bigquery.dataset import Dataset\n return Dataset(name, client=self)\n \n \ndiff --git a/unit_tests/bigquery/test_schema.py b/unit_tests/bigquery/test_schema.py\n--- a/unit_tests/bigquery/test_schema.py\n+++ b/unit_tests/bigquery/test_schema.py\n@@ -18,7 +18,7 @@\n class TestSchemaField(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.bigquery.schema import SchemaField\n+ from google.cloud.bigquery.schema import SchemaField\n return SchemaField\n \n def _makeOne(self, *args, **kw):\ndiff --git a/unit_tests/bigquery/test_table.py b/unit_tests/bigquery/test_table.py\n--- a/unit_tests/bigquery/test_table.py\n+++ b/unit_tests/bigquery/test_table.py\n@@ -36,7 +36,7 @@ class TestTable(unittest.TestCase, _SchemaBase):\n TABLE_NAME = 'table-name'\n \n def _getTargetClass(self):\n- from gcloud.bigquery.table import Table\n+ from google.cloud.bigquery.table import Table\n return Table\n \n def _makeOne(self, *args, **kw):\n@@ -44,7 +44,7 @@ def _makeOne(self, *args, **kw):\n \n def _setUpConstants(self):\n import datetime\n- from gcloud._helpers import UTC\n+ from google.cloud._helpers import UTC\n \n self.WHEN_TS = 1437767599.006\n self.WHEN = datetime.datetime.utcfromtimestamp(self.WHEN_TS).replace(\n@@ -160,7 +160,7 @@ def test_ctor(self):\n self.assertEqual(table.view_query, None)\n \n def test_ctor_w_schema(self):\n- from gcloud.bigquery.table import SchemaField\n+ from google.cloud.bigquery.table import SchemaField\n client = _Client(self.PROJECT)\n dataset = _Dataset(client)\n full_name = SchemaField('full_name', 'STRING', mode='REQUIRED')\n@@ -221,7 +221,7 @@ def test_schema_setter_non_list(self):\n table.schema = object()\n \n def test_schema_setter_invalid_field(self):\n- from gcloud.bigquery.table import SchemaField\n+ from google.cloud.bigquery.table import SchemaField\n client = _Client(self.PROJECT)\n dataset = _Dataset(client)\n table = self._makeOne(self.TABLE_NAME, dataset)\n@@ -230,7 +230,7 @@ def test_schema_setter_invalid_field(self):\n table.schema = [full_name, object()]\n \n def test_schema_setter(self):\n- from gcloud.bigquery.table import SchemaField\n+ from google.cloud.bigquery.table import SchemaField\n client = _Client(self.PROJECT)\n dataset = _Dataset(client)\n table = self._makeOne(self.TABLE_NAME, dataset)\n@@ -241,8 +241,8 @@ def test_schema_setter(self):\n \n def test_props_set_by_server(self):\n import datetime\n- from gcloud._helpers import UTC\n- from gcloud._helpers import _millis\n+ from google.cloud._helpers import UTC\n+ from google.cloud._helpers import _millis\n \n CREATED = datetime.datetime(2015, 7, 29, 12, 13, 22, tzinfo=UTC)\n MODIFIED = datetime.datetime(2015, 7, 29, 14, 47, 15, tzinfo=UTC)\n@@ -294,7 +294,7 @@ def test_expires_setter_bad_value(self):\n \n def test_expires_setter(self):\n import datetime\n- from gcloud._helpers import UTC\n+ from google.cloud._helpers import UTC\n \n WHEN = datetime.datetime(2015, 7, 28, 16, 39, tzinfo=UTC)\n client = _Client(self.PROJECT)\n@@ -400,7 +400,7 @@ def test_create_no_view_query_no_schema(self):\n table.create()\n \n def test_create_w_bound_client(self):\n- from gcloud.bigquery.table import SchemaField\n+ from google.cloud.bigquery.table import SchemaField\n PATH = 'projects/%s/datasets/%s/tables' % (self.PROJECT, self.DS_NAME)\n RESOURCE = self._makeResource()\n conn = _Connection(RESOURCE)\n@@ -430,7 +430,7 @@ def test_create_w_bound_client(self):\n self._verifyResourceProperties(table, RESOURCE)\n \n def test_create_w_partition_no_expire(self):\n- from gcloud.bigquery.table import SchemaField\n+ from google.cloud.bigquery.table import SchemaField\n PATH = 'projects/%s/datasets/%s/tables' % (self.PROJECT, self.DS_NAME)\n RESOURCE = self._makeResource()\n conn = _Connection(RESOURCE)\n@@ -464,7 +464,7 @@ def test_create_w_partition_no_expire(self):\n self._verifyResourceProperties(table, RESOURCE)\n \n def test_create_w_partition_and_expire(self):\n- from gcloud.bigquery.table import SchemaField\n+ from google.cloud.bigquery.table import SchemaField\n PATH = 'projects/%s/datasets/%s/tables' % (self.PROJECT, self.DS_NAME)\n RESOURCE = self._makeResource()\n conn = _Connection(RESOURCE)\n@@ -498,7 +498,7 @@ def test_create_w_partition_and_expire(self):\n self._verifyResourceProperties(table, RESOURCE)\n \n def test_partition_type_setter_bad_type(self):\n- from gcloud.bigquery.table import SchemaField\n+ from google.cloud.bigquery.table import SchemaField\n RESOURCE = self._makeResource()\n conn = _Connection(RESOURCE)\n client = _Client(project=self.PROJECT, connection=conn)\n@@ -511,7 +511,7 @@ def test_partition_type_setter_bad_type(self):\n table.partitioning_type = 123\n \n def test_partition_type_setter_unknown_value(self):\n- from gcloud.bigquery.table import SchemaField\n+ from google.cloud.bigquery.table import SchemaField\n RESOURCE = self._makeResource()\n conn = _Connection(RESOURCE)\n client = _Client(project=self.PROJECT, connection=conn)\n@@ -524,7 +524,7 @@ def test_partition_type_setter_unknown_value(self):\n table.partitioning_type = \"HASH\"\n \n def test_partition_type_setter_w_known_value(self):\n- from gcloud.bigquery.table import SchemaField\n+ from google.cloud.bigquery.table import SchemaField\n RESOURCE = self._makeResource()\n conn = _Connection(RESOURCE)\n client = _Client(project=self.PROJECT, connection=conn)\n@@ -538,7 +538,7 @@ def test_partition_type_setter_w_known_value(self):\n self.assertEqual(table.partitioning_type, 'DAY')\n \n def test_partition_type_setter_w_none(self):\n- from gcloud.bigquery.table import SchemaField\n+ from google.cloud.bigquery.table import SchemaField\n RESOURCE = self._makeResource()\n conn = _Connection(RESOURCE)\n client = _Client(project=self.PROJECT, connection=conn)\n@@ -553,7 +553,7 @@ def test_partition_type_setter_w_none(self):\n self.assertFalse('timePartitioning' in table._properties)\n \n def test_partition_experation_bad_type(self):\n- from gcloud.bigquery.table import SchemaField\n+ from google.cloud.bigquery.table import SchemaField\n RESOURCE = self._makeResource()\n conn = _Connection(RESOURCE)\n client = _Client(project=self.PROJECT, connection=conn)\n@@ -566,7 +566,7 @@ def test_partition_experation_bad_type(self):\n table.partition_expiration = \"NEVER\"\n \n def test_partition_expiration_w_integer(self):\n- from gcloud.bigquery.table import SchemaField\n+ from google.cloud.bigquery.table import SchemaField\n RESOURCE = self._makeResource()\n conn = _Connection(RESOURCE)\n client = _Client(project=self.PROJECT, connection=conn)\n@@ -581,7 +581,7 @@ def test_partition_expiration_w_integer(self):\n self.assertEqual(table.partition_expiration, 100)\n \n def test_partition_expiration_w_none(self):\n- from gcloud.bigquery.table import SchemaField\n+ from google.cloud.bigquery.table import SchemaField\n RESOURCE = self._makeResource()\n conn = _Connection(RESOURCE)\n client = _Client(project=self.PROJECT, connection=conn)\n@@ -600,7 +600,7 @@ def test_partition_expiration_w_none(self):\n self.assertEqual(table.partition_expiration, None)\n \n def test_partition_expiration_w_none_no_partition_set(self):\n- from gcloud.bigquery.table import SchemaField\n+ from google.cloud.bigquery.table import SchemaField\n RESOURCE = self._makeResource()\n conn = _Connection(RESOURCE)\n client = _Client(project=self.PROJECT, connection=conn)\n@@ -615,7 +615,7 @@ def test_partition_expiration_w_none_no_partition_set(self):\n self.assertEqual(table.partition_expiration, None)\n \n def test_list_partitions(self):\n- from gcloud.bigquery.table import SchemaField\n+ from google.cloud.bigquery.table import SchemaField\n conn = _Connection()\n client = _Client(project=self.PROJECT, connection=conn)\n client._query_results = [(20160804, None), (20160805, None)]\n@@ -628,9 +628,9 @@ def test_list_partitions(self):\n \n def test_create_w_alternate_client(self):\n import datetime\n- from gcloud._helpers import UTC\n- from gcloud._helpers import _millis\n- from gcloud.bigquery.table import SchemaField\n+ from google.cloud._helpers import UTC\n+ from google.cloud._helpers import _millis\n+ from google.cloud.bigquery.table import SchemaField\n \n PATH = 'projects/%s/datasets/%s/tables' % (self.PROJECT, self.DS_NAME)\n DESCRIPTION = 'DESCRIPTION'\n@@ -680,7 +680,7 @@ def test_create_w_alternate_client(self):\n def test_create_w_missing_output_properties(self):\n # In the wild, the resource returned from 'dataset.create' sometimes\n # lacks 'creationTime' / 'lastModifiedTime'\n- from gcloud.bigquery.table import SchemaField\n+ from google.cloud.bigquery.table import SchemaField\n PATH = 'projects/%s/datasets/%s/tables' % (self.PROJECT, self.DS_NAME)\n RESOURCE = self._makeResource()\n del RESOURCE['creationTime']\n@@ -825,9 +825,9 @@ def test_patch_w_bound_client(self):\n \n def test_patch_w_alternate_client(self):\n import datetime\n- from gcloud._helpers import UTC\n- from gcloud._helpers import _millis\n- from gcloud.bigquery.table import SchemaField\n+ from google.cloud._helpers import UTC\n+ from google.cloud._helpers import _millis\n+ from google.cloud.bigquery.table import SchemaField\n \n PATH = 'projects/%s/datasets/%s/tables/%s' % (\n self.PROJECT, self.DS_NAME, self.TABLE_NAME)\n@@ -894,7 +894,7 @@ def test_patch_w_schema_None(self):\n self._verifyResourceProperties(table, RESOURCE)\n \n def test_update_w_bound_client(self):\n- from gcloud.bigquery.table import SchemaField\n+ from google.cloud.bigquery.table import SchemaField\n PATH = 'projects/%s/datasets/%s/tables/%s' % (\n self.PROJECT, self.DS_NAME, self.TABLE_NAME)\n DESCRIPTION = 'DESCRIPTION'\n@@ -934,8 +934,8 @@ def test_update_w_bound_client(self):\n \n def test_update_w_alternate_client(self):\n import datetime\n- from gcloud._helpers import UTC\n- from gcloud._helpers import _millis\n+ from google.cloud._helpers import UTC\n+ from google.cloud._helpers import _millis\n \n PATH = 'projects/%s/datasets/%s/tables/%s' % (\n self.PROJECT, self.DS_NAME, self.TABLE_NAME)\n@@ -1015,8 +1015,8 @@ def test_delete_w_alternate_client(self):\n \n def test_fetch_data_w_bound_client(self):\n import datetime\n- from gcloud._helpers import UTC\n- from gcloud.bigquery.table import SchemaField\n+ from google.cloud._helpers import UTC\n+ from google.cloud.bigquery.table import SchemaField\n \n PATH = 'projects/%s/datasets/%s/tables/%s/data' % (\n self.PROJECT, self.DS_NAME, self.TABLE_NAME)\n@@ -1084,7 +1084,7 @@ def _bigquery_timestamp_float_repr(ts_float):\n self.assertEqual(req['path'], '/%s' % PATH)\n \n def test_fetch_data_w_alternate_client(self):\n- from gcloud.bigquery.table import SchemaField\n+ from google.cloud.bigquery.table import SchemaField\n PATH = 'projects/%s/datasets/%s/tables/%s/data' % (\n self.PROJECT, self.DS_NAME, self.TABLE_NAME)\n MAX = 10\n@@ -1150,7 +1150,7 @@ def test_fetch_data_w_alternate_client(self):\n {'maxResults': MAX, 'pageToken': TOKEN})\n \n def test_fetch_data_w_repeated_fields(self):\n- from gcloud.bigquery.table import SchemaField\n+ from google.cloud.bigquery.table import SchemaField\n PATH = 'projects/%s/datasets/%s/tables/%s/data' % (\n self.PROJECT, self.DS_NAME, self.TABLE_NAME)\n ROWS = 1234\n@@ -1192,7 +1192,7 @@ def test_fetch_data_w_repeated_fields(self):\n self.assertEqual(req['path'], '/%s' % PATH)\n \n def test_fetch_data_w_record_schema(self):\n- from gcloud.bigquery.table import SchemaField\n+ from google.cloud.bigquery.table import SchemaField\n PATH = 'projects/%s/datasets/%s/tables/%s/data' % (\n self.PROJECT, self.DS_NAME, self.TABLE_NAME)\n ROWS = 1234\n@@ -1250,9 +1250,9 @@ def test_fetch_data_w_record_schema(self):\n \n def test_insert_data_w_bound_client(self):\n import datetime\n- from gcloud._helpers import UTC\n- from gcloud._helpers import _microseconds_from_datetime\n- from gcloud.bigquery.table import SchemaField\n+ from google.cloud._helpers import UTC\n+ from google.cloud._helpers import _microseconds_from_datetime\n+ from google.cloud.bigquery.table import SchemaField\n \n WHEN_TS = 1437767599.006\n WHEN = datetime.datetime.utcfromtimestamp(WHEN_TS).replace(\n@@ -1296,7 +1296,7 @@ def _row_data(row):\n self.assertEqual(req['data'], SENT)\n \n def test_insert_data_w_alternate_client(self):\n- from gcloud.bigquery.table import SchemaField\n+ from google.cloud.bigquery.table import SchemaField\n PATH = 'projects/%s/datasets/%s/tables/%s/insertAll' % (\n self.PROJECT, self.DS_NAME, self.TABLE_NAME)\n RESPONSE = {\n@@ -1360,7 +1360,7 @@ def _row_data(row):\n self.assertEqual(req['data'], SENT)\n \n def test_insert_data_w_repeated_fields(self):\n- from gcloud.bigquery.table import SchemaField\n+ from google.cloud.bigquery.table import SchemaField\n PATH = 'projects/%s/datasets/%s/tables/%s/insertAll' % (\n self.PROJECT, self.DS_NAME, self.TABLE_NAME)\n conn = _Connection({})\n@@ -1395,7 +1395,7 @@ def _row_data(row):\n self.assertEqual(req['data'], SENT)\n \n def test_insert_data_w_record_schema(self):\n- from gcloud.bigquery.table import SchemaField\n+ from google.cloud.bigquery.table import SchemaField\n PATH = 'projects/%s/datasets/%s/tables/%s/insertAll' % (\n self.PROJECT, self.DS_NAME, self.TABLE_NAME)\n conn = _Connection({})\n@@ -1462,9 +1462,9 @@ def _upload_from_file_helper(self, **kw):\n import csv\n import datetime\n from six.moves.http_client import OK\n- from gcloud._helpers import UTC\n+ from google.cloud._helpers import UTC\n from unit_tests._testing import _NamedTemporaryFile\n- from gcloud.bigquery.table import SchemaField\n+ from google.cloud.bigquery.table import SchemaField\n \n WHEN_TS = 1437767599.006\n WHEN = datetime.datetime.utcfromtimestamp(WHEN_TS).replace(\n@@ -1514,7 +1514,7 @@ def test_upload_from_file_w_bound_client_multipart(self):\n import json\n from six.moves.urllib.parse import parse_qsl\n from six.moves.urllib.parse import urlsplit\n- from gcloud._helpers import _to_bytes\n+ from google.cloud._helpers import _to_bytes\n \n requested, PATH, BODY = self._upload_from_file_helper()\n parse_chunk = _email_chunk_parser()\n@@ -1571,7 +1571,7 @@ def test_upload_from_file_w_explicit_client_resumable(self):\n from six.moves.urllib.parse import parse_qsl\n from six.moves.urllib.parse import urlsplit\n from unit_tests._testing import _Monkey\n- from gcloud.bigquery import table as MUT\n+ from google.cloud.bigquery import table as MUT\n \n UPLOAD_PATH = 'https://example.com/upload/test'\n initial_response = {'status': OK, 'location': UPLOAD_PATH}\n@@ -1656,7 +1656,7 @@ class _UploadConfig(object):\n class Test_parse_schema_resource(unittest.TestCase, _SchemaBase):\n \n def _callFUT(self, resource):\n- from gcloud.bigquery.table import _parse_schema_resource\n+ from google.cloud.bigquery.table import _parse_schema_resource\n return _parse_schema_resource(resource)\n \n def _makeResource(self):\n@@ -1700,11 +1700,11 @@ def test__parse_schema_resource_fields_without_mode(self):\n class Test_build_schema_resource(unittest.TestCase, _SchemaBase):\n \n def _callFUT(self, resource):\n- from gcloud.bigquery.table import _build_schema_resource\n+ from google.cloud.bigquery.table import _build_schema_resource\n return _build_schema_resource(resource)\n \n def test_defaults(self):\n- from gcloud.bigquery.table import SchemaField\n+ from google.cloud.bigquery.table import SchemaField\n full_name = SchemaField('full_name', 'STRING', mode='REQUIRED')\n age = SchemaField('age', 'INTEGER', mode='REQUIRED')\n resource = self._callFUT([full_name, age])\n@@ -1719,7 +1719,7 @@ def test_defaults(self):\n 'mode': 'REQUIRED'})\n \n def test_w_description(self):\n- from gcloud.bigquery.table import SchemaField\n+ from google.cloud.bigquery.table import SchemaField\n DESCRIPTION = 'DESCRIPTION'\n full_name = SchemaField('full_name', 'STRING', mode='REQUIRED',\n description=DESCRIPTION)\n@@ -1737,7 +1737,7 @@ def test_w_description(self):\n 'mode': 'REQUIRED'})\n \n def test_w_subfields(self):\n- from gcloud.bigquery.table import SchemaField\n+ from google.cloud.bigquery.table import SchemaField\n full_name = SchemaField('full_name', 'STRING', mode='REQUIRED')\n ph_type = SchemaField('type', 'STRING', 'REQUIRED')\n ph_num = SchemaField('number', 'STRING', 'REQUIRED')\n@@ -1836,7 +1836,7 @@ def __init__(self, *responses):\n self.http = _HTTP(*responses)\n \n def api_request(self, **kw):\n- from gcloud.exceptions import NotFound\n+ from google.cloud.exceptions import NotFound\n self._requested.append(kw)\n \n try:\ndiff --git a/unit_tests/bigtable/test_client.py b/unit_tests/bigtable/test_client.py\n--- a/unit_tests/bigtable/test_client.py\n+++ b/unit_tests/bigtable/test_client.py\n@@ -19,12 +19,12 @@\n class Test__make_data_stub(unittest.TestCase):\n \n def _callFUT(self, client):\n- from gcloud.bigtable.client import _make_data_stub\n+ from google.cloud.bigtable.client import _make_data_stub\n return _make_data_stub(client)\n \n def test_it(self):\n from unit_tests._testing import _Monkey\n- from gcloud.bigtable import client as MUT\n+ from google.cloud.bigtable import client as MUT\n \n credentials = _Credentials()\n user_agent = 'you-sir-age-int'\n@@ -55,12 +55,12 @@ def mock_make_stub(*args):\n class Test__make_instance_stub(unittest.TestCase):\n \n def _callFUT(self, client):\n- from gcloud.bigtable.client import _make_instance_stub\n+ from google.cloud.bigtable.client import _make_instance_stub\n return _make_instance_stub(client)\n \n def test_it(self):\n from unit_tests._testing import _Monkey\n- from gcloud.bigtable import client as MUT\n+ from google.cloud.bigtable import client as MUT\n \n credentials = _Credentials()\n user_agent = 'you-sir-age-int'\n@@ -91,12 +91,12 @@ def mock_make_stub(*args):\n class Test__make_operations_stub(unittest.TestCase):\n \n def _callFUT(self, client):\n- from gcloud.bigtable.client import _make_operations_stub\n+ from google.cloud.bigtable.client import _make_operations_stub\n return _make_operations_stub(client)\n \n def test_it(self):\n from unit_tests._testing import _Monkey\n- from gcloud.bigtable import client as MUT\n+ from google.cloud.bigtable import client as MUT\n \n credentials = _Credentials()\n user_agent = 'you-sir-age-int'\n@@ -127,12 +127,12 @@ def mock_make_stub(*args):\n class Test__make_table_stub(unittest.TestCase):\n \n def _callFUT(self, client):\n- from gcloud.bigtable.client import _make_table_stub\n+ from google.cloud.bigtable.client import _make_table_stub\n return _make_table_stub(client)\n \n def test_it(self):\n from unit_tests._testing import _Monkey\n- from gcloud.bigtable import client as MUT\n+ from google.cloud.bigtable import client as MUT\n \n credentials = _Credentials()\n user_agent = 'you-sir-age-int'\n@@ -168,7 +168,7 @@ class TestClient(unittest.TestCase):\n USER_AGENT = 'you-sir-age-int'\n \n def _getTargetClass(self):\n- from gcloud.bigtable.client import Client\n+ from google.cloud.bigtable.client import Client\n return Client\n \n def _makeOne(self, *args, **kwargs):\n@@ -176,7 +176,7 @@ def _makeOne(self, *args, **kwargs):\n \n def _makeOneWithMocks(self, *args, **kwargs):\n from unit_tests._testing import _Monkey\n- from gcloud.bigtable import client as MUT\n+ from google.cloud.bigtable import client as MUT\n \n mock_make_data_stub = _MakeStubMock()\n mock_make_instance_stub = _MakeStubMock()\n@@ -192,7 +192,7 @@ def _constructor_test_helper(self, expected_scopes, creds,\n read_only=False, admin=False,\n user_agent=None, expected_creds=None):\n from unit_tests._testing import _Monkey\n- from gcloud.bigtable import client as MUT\n+ from google.cloud.bigtable import client as MUT\n \n user_agent = user_agent or MUT.DEFAULT_USER_AGENT\n \n@@ -241,14 +241,14 @@ def _constructor_test_helper(self, expected_scopes, creds,\n self.assertIsNone(client._table_stub_internal)\n \n def test_constructor_default_scopes(self):\n- from gcloud.bigtable import client as MUT\n+ from google.cloud.bigtable import client as MUT\n \n expected_scopes = [MUT.DATA_SCOPE]\n creds = _Credentials()\n self._constructor_test_helper(expected_scopes, creds)\n \n def test_constructor_custom_user_agent(self):\n- from gcloud.bigtable import client as MUT\n+ from google.cloud.bigtable import client as MUT\n \n CUSTOM_USER_AGENT = 'custom-application'\n expected_scopes = [MUT.DATA_SCOPE]\n@@ -257,14 +257,14 @@ def test_constructor_custom_user_agent(self):\n user_agent=CUSTOM_USER_AGENT)\n \n def test_constructor_with_admin(self):\n- from gcloud.bigtable import client as MUT\n+ from google.cloud.bigtable import client as MUT\n \n expected_scopes = [MUT.DATA_SCOPE, MUT.ADMIN_SCOPE]\n creds = _Credentials()\n self._constructor_test_helper(expected_scopes, creds, admin=True)\n \n def test_constructor_with_read_only(self):\n- from gcloud.bigtable import client as MUT\n+ from google.cloud.bigtable import client as MUT\n \n expected_scopes = [MUT.READ_ONLY_SCOPE]\n creds = _Credentials()\n@@ -278,7 +278,7 @@ def test_constructor_both_admin_and_read_only(self):\n \n def test_constructor_implicit_credentials(self):\n from unit_tests._testing import _Monkey\n- from gcloud.bigtable import client as MUT\n+ from google.cloud.bigtable import client as MUT\n \n creds = _Credentials()\n expected_scopes = [MUT.DATA_SCOPE]\n@@ -297,7 +297,7 @@ def test_constructor_credentials_wo_create_scoped(self):\n \n def _copy_test_helper(self, read_only=False, admin=False):\n from unit_tests._testing import _Monkey\n- from gcloud.bigtable import client as MUT\n+ from google.cloud.bigtable import client as MUT\n \n credentials = _Credentials('value')\n client = self._makeOneWithMocks(\n@@ -407,9 +407,10 @@ def test_table_stub_non_admin_failure(self):\n getattr(client, '_table_stub')\n \n def test_instance_factory_defaults(self):\n- from gcloud.bigtable.cluster import DEFAULT_SERVE_NODES\n- from gcloud.bigtable.instance import Instance\n- from gcloud.bigtable.instance import _EXISTING_INSTANCE_LOCATION_ID\n+ from google.cloud.bigtable.cluster import DEFAULT_SERVE_NODES\n+ from google.cloud.bigtable.instance import Instance\n+ from google.cloud.bigtable.instance import (\n+ _EXISTING_INSTANCE_LOCATION_ID)\n \n PROJECT = 'PROJECT'\n INSTANCE_ID = 'instance-id'\n@@ -429,7 +430,7 @@ def test_instance_factory_defaults(self):\n self.assertTrue(instance._client is client)\n \n def test_instance_factory_w_explicit_serve_nodes(self):\n- from gcloud.bigtable.instance import Instance\n+ from google.cloud.bigtable.instance import Instance\n \n PROJECT = 'PROJECT'\n INSTANCE_ID = 'instance-id'\n@@ -452,9 +453,9 @@ def test_instance_factory_w_explicit_serve_nodes(self):\n self.assertTrue(instance._client is client)\n \n def test_list_instances(self):\n- from gcloud.bigtable._generated import (\n+ from google.cloud.bigtable._generated import (\n instance_pb2 as data_v2_pb2)\n- from gcloud.bigtable._generated import (\n+ from google.cloud.bigtable._generated import (\n bigtable_instance_admin_pb2 as messages_v2_pb2)\n from unit_tests.bigtable._testing import _FakeStub\n \ndiff --git a/unit_tests/bigtable/test_cluster.py b/unit_tests/bigtable/test_cluster.py\n--- a/unit_tests/bigtable/test_cluster.py\n+++ b/unit_tests/bigtable/test_cluster.py\n@@ -26,14 +26,14 @@ class TestCluster(unittest.TestCase):\n '/clusters/' + CLUSTER_ID)\n \n def _getTargetClass(self):\n- from gcloud.bigtable.cluster import Cluster\n+ from google.cloud.bigtable.cluster import Cluster\n return Cluster\n \n def _makeOne(self, *args, **kwargs):\n return self._getTargetClass()(*args, **kwargs)\n \n def test_constructor_defaults(self):\n- from gcloud.bigtable.cluster import DEFAULT_SERVE_NODES\n+ from google.cloud.bigtable.cluster import DEFAULT_SERVE_NODES\n client = _Client(self.PROJECT)\n instance = _Instance(self.INSTANCE_ID, client)\n \n@@ -70,7 +70,7 @@ def test_copy(self):\n self.assertEqual(cluster, new_cluster)\n \n def test__update_from_pb_success(self):\n- from gcloud.bigtable.cluster import DEFAULT_SERVE_NODES\n+ from google.cloud.bigtable.cluster import DEFAULT_SERVE_NODES\n \n SERVE_NODES = 8\n cluster_pb = _ClusterPB(\n@@ -85,7 +85,7 @@ def test__update_from_pb_success(self):\n self.assertEqual(cluster.serve_nodes, SERVE_NODES)\n \n def test__update_from_pb_no_serve_nodes(self):\n- from gcloud.bigtable.cluster import DEFAULT_SERVE_NODES\n+ from google.cloud.bigtable.cluster import DEFAULT_SERVE_NODES\n \n cluster_pb = _ClusterPB()\n client = _Client(self.PROJECT)\n@@ -188,7 +188,7 @@ def test___ne__(self):\n \n def test_reload(self):\n from unit_tests.bigtable._testing import _FakeStub\n- from gcloud.bigtable.cluster import DEFAULT_SERVE_NODES\n+ from google.cloud.bigtable.cluster import DEFAULT_SERVE_NODES\n \n SERVE_NODES = 31\n LOCATION = 'LOCATION'\n@@ -229,8 +229,8 @@ def test_reload(self):\n \n def test_create(self):\n from google.longrunning import operations_pb2\n- from gcloud.operation import Operation\n- from gcloud.bigtable._generated import (\n+ from google.cloud.operation import Operation\n+ from google.cloud.bigtable._generated import (\n bigtable_instance_admin_pb2 as messages_v2_pb2)\n from unit_tests.bigtable._testing import _FakeStub\n \n@@ -274,15 +274,15 @@ def test_create(self):\n def test_update(self):\n import datetime\n from google.longrunning import operations_pb2\n- from gcloud.operation import Operation\n+ from google.cloud.operation import Operation\n from google.protobuf.any_pb2 import Any\n- from gcloud._helpers import _datetime_to_pb_timestamp\n- from gcloud.bigtable._generated import (\n+ from google.cloud._helpers import _datetime_to_pb_timestamp\n+ from google.cloud.bigtable._generated import (\n instance_pb2 as data_v2_pb2)\n- from gcloud.bigtable._generated import (\n+ from google.cloud.bigtable._generated import (\n bigtable_instance_admin_pb2 as messages_v2_pb2)\n from unit_tests.bigtable._testing import _FakeStub\n- from gcloud.bigtable.cluster import _UPDATE_CLUSTER_METADATA_URL\n+ from google.cloud.bigtable.cluster import _UPDATE_CLUSTER_METADATA_URL\n \n NOW = datetime.datetime.utcnow()\n NOW_PB = _datetime_to_pb_timestamp(NOW)\n@@ -372,11 +372,11 @@ def test_delete(self):\n class Test__prepare_create_request(unittest.TestCase):\n \n def _callFUT(self, cluster):\n- from gcloud.bigtable.cluster import _prepare_create_request\n+ from google.cloud.bigtable.cluster import _prepare_create_request\n return _prepare_create_request(cluster)\n \n def test_it(self):\n- from gcloud.bigtable.cluster import Cluster\n+ from google.cloud.bigtable.cluster import Cluster\n \n PROJECT = 'PROJECT'\n INSTANCE_ID = 'instance-id'\n@@ -396,19 +396,19 @@ def test_it(self):\n \n \n def _ClusterPB(*args, **kw):\n- from gcloud.bigtable._generated import (\n+ from google.cloud.bigtable._generated import (\n instance_pb2 as instance_v2_pb2)\n return instance_v2_pb2.Cluster(*args, **kw)\n \n \n def _DeleteClusterRequestPB(*args, **kw):\n- from gcloud.bigtable._generated import (\n+ from google.cloud.bigtable._generated import (\n bigtable_instance_admin_pb2 as messages_v2_pb2)\n return messages_v2_pb2.DeleteClusterRequest(*args, **kw)\n \n \n def _GetClusterRequestPB(*args, **kw):\n- from gcloud.bigtable._generated import (\n+ from google.cloud.bigtable._generated import (\n bigtable_instance_admin_pb2 as messages_v2_pb2)\n return messages_v2_pb2.GetClusterRequest(*args, **kw)\n \ndiff --git a/unit_tests/bigtable/test_column_family.py b/unit_tests/bigtable/test_column_family.py\n--- a/unit_tests/bigtable/test_column_family.py\n+++ b/unit_tests/bigtable/test_column_family.py\n@@ -19,7 +19,8 @@\n class Test__timedelta_to_duration_pb(unittest.TestCase):\n \n def _callFUT(self, *args, **kwargs):\n- from gcloud.bigtable.column_family import _timedelta_to_duration_pb\n+ from google.cloud.bigtable.column_family import (\n+ _timedelta_to_duration_pb)\n return _timedelta_to_duration_pb(*args, **kwargs)\n \n def test_it(self):\n@@ -64,7 +65,8 @@ def test_with_negative_seconds(self):\n class Test__duration_pb_to_timedelta(unittest.TestCase):\n \n def _callFUT(self, *args, **kwargs):\n- from gcloud.bigtable.column_family import _duration_pb_to_timedelta\n+ from google.cloud.bigtable.column_family import (\n+ _duration_pb_to_timedelta)\n return _duration_pb_to_timedelta(*args, **kwargs)\n \n def test_it(self):\n@@ -84,7 +86,7 @@ def test_it(self):\n class TestMaxVersionsGCRule(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.bigtable.column_family import MaxVersionsGCRule\n+ from google.cloud.bigtable.column_family import MaxVersionsGCRule\n return MaxVersionsGCRule\n \n def _makeOne(self, *args, **kwargs):\n@@ -117,7 +119,7 @@ def test_to_pb(self):\n class TestMaxAgeGCRule(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.bigtable.column_family import MaxAgeGCRule\n+ from google.cloud.bigtable.column_family import MaxAgeGCRule\n return MaxAgeGCRule\n \n def _makeOne(self, *args, **kwargs):\n@@ -156,7 +158,7 @@ def test_to_pb(self):\n class TestGCRuleUnion(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.bigtable.column_family import GCRuleUnion\n+ from google.cloud.bigtable.column_family import GCRuleUnion\n return GCRuleUnion\n \n def _makeOne(self, *args, **kwargs):\n@@ -189,8 +191,8 @@ def test___ne__same_value(self):\n def test_to_pb(self):\n import datetime\n from google.protobuf import duration_pb2\n- from gcloud.bigtable.column_family import MaxAgeGCRule\n- from gcloud.bigtable.column_family import MaxVersionsGCRule\n+ from google.cloud.bigtable.column_family import MaxAgeGCRule\n+ from google.cloud.bigtable.column_family import MaxVersionsGCRule\n \n max_num_versions = 42\n rule1 = MaxVersionsGCRule(max_num_versions)\n@@ -211,8 +213,8 @@ def test_to_pb(self):\n def test_to_pb_nested(self):\n import datetime\n from google.protobuf import duration_pb2\n- from gcloud.bigtable.column_family import MaxAgeGCRule\n- from gcloud.bigtable.column_family import MaxVersionsGCRule\n+ from google.cloud.bigtable.column_family import MaxAgeGCRule\n+ from google.cloud.bigtable.column_family import MaxVersionsGCRule\n \n max_num_versions1 = 42\n rule1 = MaxVersionsGCRule(max_num_versions1)\n@@ -242,7 +244,7 @@ def test_to_pb_nested(self):\n class TestGCRuleIntersection(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.bigtable.column_family import GCRuleIntersection\n+ from google.cloud.bigtable.column_family import GCRuleIntersection\n return GCRuleIntersection\n \n def _makeOne(self, *args, **kwargs):\n@@ -275,8 +277,8 @@ def test___ne__same_value(self):\n def test_to_pb(self):\n import datetime\n from google.protobuf import duration_pb2\n- from gcloud.bigtable.column_family import MaxAgeGCRule\n- from gcloud.bigtable.column_family import MaxVersionsGCRule\n+ from google.cloud.bigtable.column_family import MaxAgeGCRule\n+ from google.cloud.bigtable.column_family import MaxVersionsGCRule\n \n max_num_versions = 42\n rule1 = MaxVersionsGCRule(max_num_versions)\n@@ -298,8 +300,8 @@ def test_to_pb(self):\n def test_to_pb_nested(self):\n import datetime\n from google.protobuf import duration_pb2\n- from gcloud.bigtable.column_family import MaxAgeGCRule\n- from gcloud.bigtable.column_family import MaxVersionsGCRule\n+ from google.cloud.bigtable.column_family import MaxAgeGCRule\n+ from google.cloud.bigtable.column_family import MaxVersionsGCRule\n \n max_num_versions1 = 42\n rule1 = MaxVersionsGCRule(max_num_versions1)\n@@ -331,7 +333,7 @@ def test_to_pb_nested(self):\n class TestColumnFamily(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.bigtable.column_family import ColumnFamily\n+ from google.cloud.bigtable.column_family import ColumnFamily\n return ColumnFamily\n \n def _makeOne(self, *args, **kwargs):\n@@ -395,7 +397,7 @@ def test_to_pb_no_rules(self):\n self.assertEqual(pb_val, expected)\n \n def test_to_pb_with_rule(self):\n- from gcloud.bigtable.column_family import MaxVersionsGCRule\n+ from google.cloud.bigtable.column_family import MaxVersionsGCRule\n \n gc_rule = MaxVersionsGCRule(1)\n column_family = self._makeOne('column_family_id', None,\n@@ -405,7 +407,7 @@ def test_to_pb_with_rule(self):\n self.assertEqual(pb_val, expected)\n \n def _create_test_helper(self, gc_rule=None):\n- from gcloud.bigtable._generated import (\n+ from google.cloud.bigtable._generated import (\n bigtable_table_admin_pb2 as table_admin_v2_pb2)\n from unit_tests.bigtable._testing import _FakeStub\n \n@@ -458,13 +460,13 @@ def test_create(self):\n self._create_test_helper(gc_rule=None)\n \n def test_create_with_gc_rule(self):\n- from gcloud.bigtable.column_family import MaxVersionsGCRule\n+ from google.cloud.bigtable.column_family import MaxVersionsGCRule\n gc_rule = MaxVersionsGCRule(1337)\n self._create_test_helper(gc_rule=gc_rule)\n \n def _update_test_helper(self, gc_rule=None):\n from unit_tests.bigtable._testing import _FakeStub\n- from gcloud.bigtable._generated import (\n+ from google.cloud.bigtable._generated import (\n bigtable_table_admin_pb2 as table_admin_v2_pb2)\n \n project_id = 'project-id'\n@@ -516,13 +518,13 @@ def test_update(self):\n self._update_test_helper(gc_rule=None)\n \n def test_update_with_gc_rule(self):\n- from gcloud.bigtable.column_family import MaxVersionsGCRule\n+ from google.cloud.bigtable.column_family import MaxVersionsGCRule\n gc_rule = MaxVersionsGCRule(1337)\n self._update_test_helper(gc_rule=gc_rule)\n \n def test_delete(self):\n from google.protobuf import empty_pb2\n- from gcloud.bigtable._generated import (\n+ from google.cloud.bigtable._generated import (\n bigtable_table_admin_pb2 as table_admin_v2_pb2)\n from unit_tests.bigtable._testing import _FakeStub\n \n@@ -569,7 +571,7 @@ def test_delete(self):\n class Test__gc_rule_from_pb(unittest.TestCase):\n \n def _callFUT(self, *args, **kwargs):\n- from gcloud.bigtable.column_family import _gc_rule_from_pb\n+ from google.cloud.bigtable.column_family import _gc_rule_from_pb\n return _gc_rule_from_pb(*args, **kwargs)\n \n def test_empty(self):\n@@ -578,7 +580,7 @@ def test_empty(self):\n self.assertEqual(self._callFUT(gc_rule_pb), None)\n \n def test_max_num_versions(self):\n- from gcloud.bigtable.column_family import MaxVersionsGCRule\n+ from google.cloud.bigtable.column_family import MaxVersionsGCRule\n \n orig_rule = MaxVersionsGCRule(1)\n gc_rule_pb = orig_rule.to_pb()\n@@ -588,7 +590,7 @@ def test_max_num_versions(self):\n \n def test_max_age(self):\n import datetime\n- from gcloud.bigtable.column_family import MaxAgeGCRule\n+ from google.cloud.bigtable.column_family import MaxAgeGCRule\n \n orig_rule = MaxAgeGCRule(datetime.timedelta(seconds=1))\n gc_rule_pb = orig_rule.to_pb()\n@@ -598,9 +600,9 @@ def test_max_age(self):\n \n def test_union(self):\n import datetime\n- from gcloud.bigtable.column_family import GCRuleUnion\n- from gcloud.bigtable.column_family import MaxAgeGCRule\n- from gcloud.bigtable.column_family import MaxVersionsGCRule\n+ from google.cloud.bigtable.column_family import GCRuleUnion\n+ from google.cloud.bigtable.column_family import MaxAgeGCRule\n+ from google.cloud.bigtable.column_family import MaxVersionsGCRule\n \n rule1 = MaxVersionsGCRule(1)\n rule2 = MaxAgeGCRule(datetime.timedelta(seconds=1))\n@@ -612,9 +614,9 @@ def test_union(self):\n \n def test_intersection(self):\n import datetime\n- from gcloud.bigtable.column_family import GCRuleIntersection\n- from gcloud.bigtable.column_family import MaxAgeGCRule\n- from gcloud.bigtable.column_family import MaxVersionsGCRule\n+ from google.cloud.bigtable.column_family import GCRuleIntersection\n+ from google.cloud.bigtable.column_family import MaxAgeGCRule\n+ from google.cloud.bigtable.column_family import MaxVersionsGCRule\n \n rule1 = MaxVersionsGCRule(1)\n rule2 = MaxAgeGCRule(datetime.timedelta(seconds=1))\n@@ -640,25 +642,25 @@ def WhichOneof(cls, name):\n \n \n def _GcRulePB(*args, **kw):\n- from gcloud.bigtable._generated import (\n+ from google.cloud.bigtable._generated import (\n table_pb2 as table_v2_pb2)\n return table_v2_pb2.GcRule(*args, **kw)\n \n \n def _GcRuleIntersectionPB(*args, **kw):\n- from gcloud.bigtable._generated import (\n+ from google.cloud.bigtable._generated import (\n table_pb2 as table_v2_pb2)\n return table_v2_pb2.GcRule.Intersection(*args, **kw)\n \n \n def _GcRuleUnionPB(*args, **kw):\n- from gcloud.bigtable._generated import (\n+ from google.cloud.bigtable._generated import (\n table_pb2 as table_v2_pb2)\n return table_v2_pb2.GcRule.Union(*args, **kw)\n \n \n def _ColumnFamilyPB(*args, **kw):\n- from gcloud.bigtable._generated import (\n+ from google.cloud.bigtable._generated import (\n table_pb2 as table_v2_pb2)\n return table_v2_pb2.ColumnFamily(*args, **kw)\n \ndiff --git a/unit_tests/bigtable/test_instance.py b/unit_tests/bigtable/test_instance.py\n--- a/unit_tests/bigtable/test_instance.py\n+++ b/unit_tests/bigtable/test_instance.py\n@@ -31,14 +31,14 @@ class TestInstance(unittest.TestCase):\n TABLE_NAME = INSTANCE_NAME + '/tables/' + TABLE_ID\n \n def _getTargetClass(self):\n- from gcloud.bigtable.instance import Instance\n+ from google.cloud.bigtable.instance import Instance\n return Instance\n \n def _makeOne(self, *args, **kwargs):\n return self._getTargetClass()(*args, **kwargs)\n \n def test_constructor_defaults(self):\n- from gcloud.bigtable.cluster import DEFAULT_SERVE_NODES\n+ from google.cloud.bigtable.cluster import DEFAULT_SERVE_NODES\n \n client = object()\n instance = self._makeOne(self.INSTANCE_ID, client, self.LOCATION_ID)\n@@ -74,7 +74,7 @@ def test_copy(self):\n self.assertEqual(instance, new_instance)\n \n def test_table_factory(self):\n- from gcloud.bigtable.table import Table\n+ from google.cloud.bigtable.table import Table\n \n instance = self._makeOne(self.INSTANCE_ID, None, self.LOCATION_ID)\n \n@@ -84,7 +84,7 @@ def test_table_factory(self):\n self.assertEqual(table._instance, instance)\n \n def test__update_from_pb_success(self):\n- from gcloud.bigtable._generated import (\n+ from google.cloud.bigtable._generated import (\n instance_pb2 as data_v2_pb2)\n \n display_name = 'display_name'\n@@ -98,7 +98,7 @@ def test__update_from_pb_success(self):\n self.assertEqual(instance.display_name, display_name)\n \n def test__update_from_pb_no_display_name(self):\n- from gcloud.bigtable._generated import (\n+ from google.cloud.bigtable._generated import (\n instance_pb2 as data_v2_pb2)\n \n instance_pb = data_v2_pb2.Instance()\n@@ -109,8 +109,9 @@ def test__update_from_pb_no_display_name(self):\n self.assertEqual(instance.display_name, None)\n \n def test_from_pb_success(self):\n- from gcloud.bigtable.instance import _EXISTING_INSTANCE_LOCATION_ID\n- from gcloud.bigtable._generated import (\n+ from google.cloud.bigtable.instance import (\n+ _EXISTING_INSTANCE_LOCATION_ID)\n+ from google.cloud.bigtable._generated import (\n instance_pb2 as data_v2_pb2)\n \n client = _Client(project=self.PROJECT)\n@@ -129,7 +130,7 @@ def test_from_pb_success(self):\n _EXISTING_INSTANCE_LOCATION_ID)\n \n def test_from_pb_bad_instance_name(self):\n- from gcloud.bigtable._generated import (\n+ from google.cloud.bigtable._generated import (\n instance_pb2 as data_v2_pb2)\n \n instance_name = 'INCORRECT_FORMAT'\n@@ -140,7 +141,7 @@ def test_from_pb_bad_instance_name(self):\n klass.from_pb(instance_pb, None)\n \n def test_from_pb_project_mistmatch(self):\n- from gcloud.bigtable._generated import (\n+ from google.cloud.bigtable._generated import (\n instance_pb2 as data_v2_pb2)\n \n ALT_PROJECT = 'ALT_PROJECT'\n@@ -185,9 +186,9 @@ def test___ne__(self):\n self.assertNotEqual(instance1, instance2)\n \n def test_reload(self):\n- from gcloud.bigtable._generated import (\n+ from google.cloud.bigtable._generated import (\n instance_pb2 as data_v2_pb2)\n- from gcloud.bigtable._generated import (\n+ from google.cloud.bigtable._generated import (\n bigtable_instance_admin_pb2 as messages_v2_pb)\n from unit_tests.bigtable._testing import _FakeStub\n \n@@ -229,13 +230,14 @@ def test_create(self):\n import datetime\n from google.longrunning import operations_pb2\n from google.protobuf.any_pb2 import Any\n- from gcloud.bigtable._generated import (\n+ from google.cloud.bigtable._generated import (\n bigtable_instance_admin_pb2 as messages_v2_pb2)\n- from gcloud._helpers import _datetime_to_pb_timestamp\n+ from google.cloud._helpers import _datetime_to_pb_timestamp\n from unit_tests.bigtable._testing import _FakeStub\n- from gcloud.operation import Operation\n- from gcloud.bigtable.cluster import DEFAULT_SERVE_NODES\n- from gcloud.bigtable.instance import _CREATE_INSTANCE_METADATA_URL\n+ from google.cloud.operation import Operation\n+ from google.cloud.bigtable.cluster import DEFAULT_SERVE_NODES\n+ from google.cloud.bigtable.instance import (\n+ _CREATE_INSTANCE_METADATA_URL)\n \n NOW = datetime.datetime.utcnow()\n NOW_PB = _datetime_to_pb_timestamp(NOW)\n@@ -283,10 +285,10 @@ def test_create(self):\n \n def test_create_w_explicit_serve_nodes(self):\n from google.longrunning import operations_pb2\n- from gcloud.bigtable._generated import (\n+ from google.cloud.bigtable._generated import (\n bigtable_instance_admin_pb2 as messages_v2_pb2)\n from unit_tests.bigtable._testing import _FakeStub\n- from gcloud.operation import Operation\n+ from google.cloud.operation import Operation\n \n SERVE_NODES = 5\n \n@@ -322,7 +324,7 @@ def test_create_w_explicit_serve_nodes(self):\n self.assertEqual(kwargs, {})\n \n def test_update(self):\n- from gcloud.bigtable._generated import (\n+ from google.cloud.bigtable._generated import (\n instance_pb2 as data_v2_pb2)\n from unit_tests.bigtable._testing import _FakeStub\n \n@@ -357,7 +359,7 @@ def test_update(self):\n \n def test_delete(self):\n from google.protobuf import empty_pb2\n- from gcloud.bigtable._generated import (\n+ from google.cloud.bigtable._generated import (\n bigtable_instance_admin_pb2 as messages_v2_pb)\n from unit_tests.bigtable._testing import _FakeStub\n \n@@ -388,9 +390,9 @@ def test_delete(self):\n )])\n \n def test_list_clusters(self):\n- from gcloud.bigtable._generated import (\n+ from google.cloud.bigtable._generated import (\n instance_pb2 as instance_v2_pb2)\n- from gcloud.bigtable._generated import (\n+ from google.cloud.bigtable._generated import (\n bigtable_instance_admin_pb2 as messages_v2_pb2)\n from unit_tests.bigtable._testing import _FakeStub\n \n@@ -445,9 +447,9 @@ def test_list_clusters(self):\n )])\n \n def _list_tables_helper(self, table_name=None):\n- from gcloud.bigtable._generated import (\n+ from google.cloud.bigtable._generated import (\n table_pb2 as table_data_v2_pb2)\n- from gcloud.bigtable._generated import (\n+ from google.cloud.bigtable._generated import (\n bigtable_table_admin_pb2 as table_messages_v1_pb2)\n from unit_tests.bigtable._testing import _FakeStub\n \n@@ -511,16 +513,16 @@ class Test__prepare_create_request(unittest.TestCase):\n CLUSTER_NAME = INSTANCE_NAME + '/clusters/' + INSTANCE_ID\n \n def _callFUT(self, instance, **kw):\n- from gcloud.bigtable.instance import _prepare_create_request\n+ from google.cloud.bigtable.instance import _prepare_create_request\n return _prepare_create_request(instance, **kw)\n \n def test_w_defaults(self):\n- from gcloud.bigtable.cluster import DEFAULT_SERVE_NODES\n- from gcloud.bigtable._generated import (\n+ from google.cloud.bigtable.cluster import DEFAULT_SERVE_NODES\n+ from google.cloud.bigtable._generated import (\n instance_pb2 as data_v2_pb2)\n- from gcloud.bigtable._generated import (\n+ from google.cloud.bigtable._generated import (\n bigtable_instance_admin_pb2 as messages_v2_pb)\n- from gcloud.bigtable.instance import Instance\n+ from google.cloud.bigtable.instance import Instance\n \n client = _Client(self.PROJECT)\n \n@@ -542,11 +544,11 @@ def test_w_defaults(self):\n self.assertEqual(cluster.serve_nodes, DEFAULT_SERVE_NODES)\n \n def test_w_explicit_serve_nodes(self):\n- from gcloud.bigtable._generated import (\n+ from google.cloud.bigtable._generated import (\n instance_pb2 as data_v2_pb2)\n- from gcloud.bigtable._generated import (\n+ from google.cloud.bigtable._generated import (\n bigtable_instance_admin_pb2 as messages_v2_pb)\n- from gcloud.bigtable.instance import Instance\n+ from google.cloud.bigtable.instance import Instance\n DISPLAY_NAME = u'DISPLAY_NAME'\n SERVE_NODES = 5\n client = _Client(self.PROJECT)\ndiff --git a/unit_tests/bigtable/test_row.py b/unit_tests/bigtable/test_row.py\n--- a/unit_tests/bigtable/test_row.py\n+++ b/unit_tests/bigtable/test_row.py\n@@ -19,7 +19,7 @@\n class Test_SetDeleteRow(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.bigtable.row import _SetDeleteRow\n+ from google.cloud.bigtable.row import _SetDeleteRow\n return _SetDeleteRow\n \n def _makeOne(self, *args, **kwargs):\n@@ -34,7 +34,7 @@ def test__get_mutations_virtual(self):\n class TestDirectRow(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.bigtable.row import DirectRow\n+ from google.cloud.bigtable.row import DirectRow\n return DirectRow\n \n def _makeOne(self, *args, **kwargs):\n@@ -123,7 +123,7 @@ def test_set_cell_with_non_bytes_value(self):\n \n def test_set_cell_with_non_null_timestamp(self):\n import datetime\n- from gcloud._helpers import _EPOCH\n+ from google.cloud._helpers import _EPOCH\n \n microseconds = 898294371\n millis_granularity = microseconds - (microseconds % 1000)\n@@ -243,8 +243,8 @@ def test_delete_cells_no_time_range(self):\n \n def test_delete_cells_with_time_range(self):\n import datetime\n- from gcloud._helpers import _EPOCH\n- from gcloud.bigtable.row_filters import TimestampRange\n+ from google.cloud._helpers import _EPOCH\n+ from google.cloud.bigtable.row_filters import TimestampRange\n \n microseconds = 30871000 # Makes sure already milliseconds granularity\n start = _EPOCH + datetime.timedelta(microseconds=microseconds)\n@@ -344,7 +344,7 @@ def test_commit(self):\n \n def test_commit_too_many_mutations(self):\n from unit_tests._testing import _Monkey\n- from gcloud.bigtable import row as MUT\n+ from google.cloud.bigtable import row as MUT\n \n row_key = b'row_key'\n table = object()\n@@ -377,7 +377,7 @@ def test_commit_no_mutations(self):\n class TestConditionalRow(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.bigtable.row import ConditionalRow\n+ from google.cloud.bigtable.row import ConditionalRow\n return ConditionalRow\n \n def _makeOne(self, *args, **kwargs):\n@@ -408,7 +408,7 @@ def test__get_mutations(self):\n \n def test_commit(self):\n from unit_tests.bigtable._testing import _FakeStub\n- from gcloud.bigtable.row_filters import RowSampleFilter\n+ from google.cloud.bigtable.row_filters import RowSampleFilter\n \n row_key = b'row_key'\n table_name = 'projects/more-stuff'\n@@ -482,7 +482,7 @@ def test_commit(self):\n \n def test_commit_too_many_mutations(self):\n from unit_tests._testing import _Monkey\n- from gcloud.bigtable import row as MUT\n+ from google.cloud.bigtable import row as MUT\n \n row_key = b'row_key'\n table = object()\n@@ -518,7 +518,7 @@ def test_commit_no_mutations(self):\n class TestAppendRow(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.bigtable.row import AppendRow\n+ from google.cloud.bigtable.row import AppendRow\n return AppendRow\n \n def _makeOne(self, *args, **kwargs):\n@@ -574,7 +574,7 @@ def test_increment_cell_value(self):\n def test_commit(self):\n from unit_tests._testing import _Monkey\n from unit_tests.bigtable._testing import _FakeStub\n- from gcloud.bigtable import row as MUT\n+ from google.cloud.bigtable import row as MUT\n \n row_key = b'row_key'\n table_name = 'projects/more-stuff'\n@@ -647,7 +647,7 @@ def test_commit_no_rules(self):\n \n def test_commit_too_many_mutations(self):\n from unit_tests._testing import _Monkey\n- from gcloud.bigtable import row as MUT\n+ from google.cloud.bigtable import row as MUT\n \n row_key = b'row_key'\n table = object()\n@@ -662,11 +662,11 @@ def test_commit_too_many_mutations(self):\n class Test__parse_rmw_row_response(unittest.TestCase):\n \n def _callFUT(self, row_response):\n- from gcloud.bigtable.row import _parse_rmw_row_response\n+ from google.cloud.bigtable.row import _parse_rmw_row_response\n return _parse_rmw_row_response(row_response)\n \n def test_it(self):\n- from gcloud._helpers import _datetime_from_microseconds\n+ from google.cloud._helpers import _datetime_from_microseconds\n col_fam1 = u'col-fam-id'\n col_fam2 = u'col-fam-id2'\n col_name1 = b'col-name1'\n@@ -747,11 +747,11 @@ def test_it(self):\n class Test__parse_family_pb(unittest.TestCase):\n \n def _callFUT(self, family_pb):\n- from gcloud.bigtable.row import _parse_family_pb\n+ from google.cloud.bigtable.row import _parse_family_pb\n return _parse_family_pb(family_pb)\n \n def test_it(self):\n- from gcloud._helpers import _datetime_from_microseconds\n+ from google.cloud._helpers import _datetime_from_microseconds\n col_fam1 = u'col-fam-id'\n col_name1 = b'col-name1'\n col_name2 = b'col-name2'\n@@ -802,91 +802,91 @@ def test_it(self):\n \n \n def _CheckAndMutateRowRequestPB(*args, **kw):\n- from gcloud.bigtable._generated import (\n+ from google.cloud.bigtable._generated import (\n bigtable_pb2 as messages_v2_pb2)\n return messages_v2_pb2.CheckAndMutateRowRequest(*args, **kw)\n \n \n def _CheckAndMutateRowResponsePB(*args, **kw):\n- from gcloud.bigtable._generated import (\n+ from google.cloud.bigtable._generated import (\n bigtable_pb2 as messages_v2_pb2)\n return messages_v2_pb2.CheckAndMutateRowResponse(*args, **kw)\n \n \n def _MutateRowRequestPB(*args, **kw):\n- from gcloud.bigtable._generated import (\n+ from google.cloud.bigtable._generated import (\n bigtable_pb2 as messages_v2_pb2)\n return messages_v2_pb2.MutateRowRequest(*args, **kw)\n \n \n def _ReadModifyWriteRowRequestPB(*args, **kw):\n- from gcloud.bigtable._generated import (\n+ from google.cloud.bigtable._generated import (\n bigtable_pb2 as messages_v2_pb2)\n return messages_v2_pb2.ReadModifyWriteRowRequest(*args, **kw)\n \n \n def _ReadModifyWriteRowResponsePB(*args, **kw):\n- from gcloud.bigtable._generated import (\n+ from google.cloud.bigtable._generated import (\n bigtable_pb2 as messages_v2_pb2)\n return messages_v2_pb2.ReadModifyWriteRowResponse(*args, **kw)\n \n \n def _CellPB(*args, **kw):\n- from gcloud.bigtable._generated import (\n+ from google.cloud.bigtable._generated import (\n data_pb2 as data_v2_pb2)\n return data_v2_pb2.Cell(*args, **kw)\n \n \n def _ColumnPB(*args, **kw):\n- from gcloud.bigtable._generated import (\n+ from google.cloud.bigtable._generated import (\n data_pb2 as data_v2_pb2)\n return data_v2_pb2.Column(*args, **kw)\n \n \n def _FamilyPB(*args, **kw):\n- from gcloud.bigtable._generated import (\n+ from google.cloud.bigtable._generated import (\n data_pb2 as data_v2_pb2)\n return data_v2_pb2.Family(*args, **kw)\n \n \n def _MutationPB(*args, **kw):\n- from gcloud.bigtable._generated import (\n+ from google.cloud.bigtable._generated import (\n data_pb2 as data_v2_pb2)\n return data_v2_pb2.Mutation(*args, **kw)\n \n \n def _MutationSetCellPB(*args, **kw):\n- from gcloud.bigtable._generated import (\n+ from google.cloud.bigtable._generated import (\n data_pb2 as data_v2_pb2)\n return data_v2_pb2.Mutation.SetCell(*args, **kw)\n \n \n def _MutationDeleteFromColumnPB(*args, **kw):\n- from gcloud.bigtable._generated import (\n+ from google.cloud.bigtable._generated import (\n data_pb2 as data_v2_pb2)\n return data_v2_pb2.Mutation.DeleteFromColumn(*args, **kw)\n \n \n def _MutationDeleteFromFamilyPB(*args, **kw):\n- from gcloud.bigtable._generated import (\n+ from google.cloud.bigtable._generated import (\n data_pb2 as data_v2_pb2)\n return data_v2_pb2.Mutation.DeleteFromFamily(*args, **kw)\n \n \n def _MutationDeleteFromRowPB(*args, **kw):\n- from gcloud.bigtable._generated import (\n+ from google.cloud.bigtable._generated import (\n data_pb2 as data_v2_pb2)\n return data_v2_pb2.Mutation.DeleteFromRow(*args, **kw)\n \n \n def _RowPB(*args, **kw):\n- from gcloud.bigtable._generated import (\n+ from google.cloud.bigtable._generated import (\n data_pb2 as data_v2_pb2)\n return data_v2_pb2.Row(*args, **kw)\n \n \n def _ReadModifyWriteRulePB(*args, **kw):\n- from gcloud.bigtable._generated import (\n+ from google.cloud.bigtable._generated import (\n data_pb2 as data_v2_pb2)\n return data_v2_pb2.ReadModifyWriteRule(*args, **kw)\n \ndiff --git a/unit_tests/bigtable/test_row_data.py b/unit_tests/bigtable/test_row_data.py\n--- a/unit_tests/bigtable/test_row_data.py\n+++ b/unit_tests/bigtable/test_row_data.py\n@@ -19,7 +19,7 @@\n class TestCell(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.bigtable.row_data import Cell\n+ from google.cloud.bigtable.row_data import Cell\n return Cell\n \n def _makeOne(self, *args, **kwargs):\n@@ -27,8 +27,8 @@ def _makeOne(self, *args, **kwargs):\n \n def _from_pb_test_helper(self, labels=None):\n import datetime\n- from gcloud._helpers import _EPOCH\n- from gcloud.bigtable._generated import (\n+ from google.cloud._helpers import _EPOCH\n+ from google.cloud.bigtable._generated import (\n data_pb2 as data_v2_pb2)\n \n timestamp_micros = 18738724000 # Make sure millis granularity\n@@ -94,7 +94,7 @@ def test___ne__(self):\n class TestPartialRowData(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.bigtable.row_data import PartialRowData\n+ from google.cloud.bigtable.row_data import PartialRowData\n return PartialRowData\n \n def _makeOne(self, *args, **kwargs):\n@@ -185,7 +185,7 @@ def test_row_key_getter(self):\n class TestPartialRowsData(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.bigtable.row_data import PartialRowsData\n+ from google.cloud.bigtable.row_data import PartialRowsData\n return PartialRowsData\n \n def _getDoNothingClass(self):\n@@ -383,7 +383,7 @@ def test__save_row_no_cell(self):\n self.assertTrue(prd._rows[ROW_KEY] is row)\n \n def test_invalid_last_scanned_row_key_on_start(self):\n- from gcloud.bigtable.row_data import InvalidReadRowsResponse\n+ from google.cloud.bigtable.row_data import InvalidReadRowsResponse\n response = _ReadRowsResponseV2(chunks=(), last_scanned_row_key='ABC')\n iterator = _MockCancellableIterator(response)\n prd = self._makeOne(iterator)\n@@ -400,7 +400,7 @@ def test_valid_last_scanned_row_key_on_start(self):\n self.assertEqual(prd._last_scanned_row_key, 'AFTER')\n \n def test_invalid_empty_chunk(self):\n- from gcloud.bigtable.row_data import InvalidChunk\n+ from google.cloud.bigtable.row_data import InvalidChunk\n chunks = _generate_cell_chunks([''])\n response = _ReadRowsResponseV2(chunks)\n iterator = _MockCancellableIterator(response)\n@@ -409,7 +409,7 @@ def test_invalid_empty_chunk(self):\n prd.consume_next()\n \n def test_invalid_empty_second_chunk(self):\n- from gcloud.bigtable.row_data import InvalidChunk\n+ from google.cloud.bigtable.row_data import InvalidChunk\n chunks = _generate_cell_chunks(['', ''])\n first = chunks[0]\n first.row_key = b'RK'\n@@ -427,7 +427,7 @@ class TestPartialRowsData_JSON_acceptance_tests(unittest.TestCase):\n _json_tests = None\n \n def _getTargetClass(self):\n- from gcloud.bigtable.row_data import PartialRowsData\n+ from google.cloud.bigtable.row_data import PartialRowsData\n return PartialRowsData\n \n def _makeOne(self, *args, **kwargs):\n@@ -447,7 +447,7 @@ def _load_json_test(self, test_name):\n # JSON Error cases: invalid chunks\n \n def _fail_during_consume(self, testcase_name):\n- from gcloud.bigtable.row_data import InvalidChunk\n+ from google.cloud.bigtable.row_data import InvalidChunk\n chunks, results = self._load_json_test(testcase_name)\n response = _ReadRowsResponseV2(chunks)\n iterator = _MockCancellableIterator(response)\n@@ -637,8 +637,8 @@ def test_empty_cell_chunk(self):\n def _flatten_cells(prd):\n # Match results format from JSON testcases.\n # Doesn't handle error cases.\n- from gcloud._helpers import _bytes_to_unicode\n- from gcloud._helpers import _microseconds_from_datetime\n+ from google.cloud._helpers import _bytes_to_unicode\n+ from google.cloud._helpers import _microseconds_from_datetime\n for row_key, row in prd.rows.items():\n for family_name, family in row.cells.items():\n for qualifier, column in family.items():\n@@ -698,7 +698,7 @@ def __init__(self, chunks, last_scanned_row_key=''):\n \n def _generate_cell_chunks(chunk_text_pbs):\n from google.protobuf.text_format import Merge\n- from gcloud.bigtable._generated.bigtable_pb2 import ReadRowsResponse\n+ from google.cloud.bigtable._generated.bigtable_pb2 import ReadRowsResponse\n \n chunks = []\n \ndiff --git a/unit_tests/bigtable/test_row_filters.py b/unit_tests/bigtable/test_row_filters.py\n--- a/unit_tests/bigtable/test_row_filters.py\n+++ b/unit_tests/bigtable/test_row_filters.py\n@@ -19,7 +19,7 @@\n class Test_BoolFilter(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.bigtable.row_filters import _BoolFilter\n+ from google.cloud.bigtable.row_filters import _BoolFilter\n return _BoolFilter\n \n def _makeOne(self, *args, **kwargs):\n@@ -53,7 +53,7 @@ def test___ne__same_value(self):\n class TestSinkFilter(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.bigtable.row_filters import SinkFilter\n+ from google.cloud.bigtable.row_filters import SinkFilter\n return SinkFilter\n \n def _makeOne(self, *args, **kwargs):\n@@ -70,7 +70,7 @@ def test_to_pb(self):\n class TestPassAllFilter(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.bigtable.row_filters import PassAllFilter\n+ from google.cloud.bigtable.row_filters import PassAllFilter\n return PassAllFilter\n \n def _makeOne(self, *args, **kwargs):\n@@ -87,7 +87,7 @@ def test_to_pb(self):\n class TestBlockAllFilter(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.bigtable.row_filters import BlockAllFilter\n+ from google.cloud.bigtable.row_filters import BlockAllFilter\n return BlockAllFilter\n \n def _makeOne(self, *args, **kwargs):\n@@ -104,7 +104,7 @@ def test_to_pb(self):\n class Test_RegexFilter(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.bigtable.row_filters import _RegexFilter\n+ from google.cloud.bigtable.row_filters import _RegexFilter\n return _RegexFilter\n \n def _makeOne(self, *args, **kwargs):\n@@ -143,7 +143,7 @@ def test___ne__same_value(self):\n class TestRowKeyRegexFilter(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.bigtable.row_filters import RowKeyRegexFilter\n+ from google.cloud.bigtable.row_filters import RowKeyRegexFilter\n return RowKeyRegexFilter\n \n def _makeOne(self, *args, **kwargs):\n@@ -160,7 +160,7 @@ def test_to_pb(self):\n class TestRowSampleFilter(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.bigtable.row_filters import RowSampleFilter\n+ from google.cloud.bigtable.row_filters import RowSampleFilter\n return RowSampleFilter\n \n def _makeOne(self, *args, **kwargs):\n@@ -194,7 +194,7 @@ def test_to_pb(self):\n class TestFamilyNameRegexFilter(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.bigtable.row_filters import FamilyNameRegexFilter\n+ from google.cloud.bigtable.row_filters import FamilyNameRegexFilter\n return FamilyNameRegexFilter\n \n def _makeOne(self, *args, **kwargs):\n@@ -211,7 +211,8 @@ def test_to_pb(self):\n class TestColumnQualifierRegexFilter(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.bigtable.row_filters import ColumnQualifierRegexFilter\n+ from google.cloud.bigtable.row_filters import (\n+ ColumnQualifierRegexFilter)\n return ColumnQualifierRegexFilter\n \n def _makeOne(self, *args, **kwargs):\n@@ -229,7 +230,7 @@ def test_to_pb(self):\n class TestTimestampRange(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.bigtable.row_filters import TimestampRange\n+ from google.cloud.bigtable.row_filters import TimestampRange\n return TimestampRange\n \n def _makeOne(self, *args, **kwargs):\n@@ -266,7 +267,7 @@ def test___ne__same_value(self):\n \n def _to_pb_helper(self, start_micros=None, end_micros=None):\n import datetime\n- from gcloud._helpers import _EPOCH\n+ from google.cloud._helpers import _EPOCH\n pb_kwargs = {}\n \n start = None\n@@ -303,7 +304,7 @@ def test_to_pb_end_only(self):\n class TestTimestampRangeFilter(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.bigtable.row_filters import TimestampRangeFilter\n+ from google.cloud.bigtable.row_filters import TimestampRangeFilter\n return TimestampRangeFilter\n \n def _makeOne(self, *args, **kwargs):\n@@ -327,7 +328,7 @@ def test___eq__same_value(self):\n self.assertEqual(row_filter1, row_filter2)\n \n def test_to_pb(self):\n- from gcloud.bigtable.row_filters import TimestampRange\n+ from google.cloud.bigtable.row_filters import TimestampRange\n \n range_ = TimestampRange()\n row_filter = self._makeOne(range_)\n@@ -340,7 +341,7 @@ def test_to_pb(self):\n class TestColumnRangeFilter(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.bigtable.row_filters import ColumnRangeFilter\n+ from google.cloud.bigtable.row_filters import ColumnRangeFilter\n return ColumnRangeFilter\n \n def _makeOne(self, *args, **kwargs):\n@@ -464,7 +465,7 @@ def test_to_pb_exclusive_end(self):\n class TestValueRegexFilter(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.bigtable.row_filters import ValueRegexFilter\n+ from google.cloud.bigtable.row_filters import ValueRegexFilter\n return ValueRegexFilter\n \n def _makeOne(self, *args, **kwargs):\n@@ -481,7 +482,7 @@ def test_to_pb(self):\n class TestValueRangeFilter(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.bigtable.row_filters import ValueRangeFilter\n+ from google.cloud.bigtable.row_filters import ValueRangeFilter\n return ValueRangeFilter\n \n def _makeOne(self, *args, **kwargs):\n@@ -572,7 +573,7 @@ def test_to_pb_exclusive_end(self):\n class Test_CellCountFilter(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.bigtable.row_filters import _CellCountFilter\n+ from google.cloud.bigtable.row_filters import _CellCountFilter\n return _CellCountFilter\n \n def _makeOne(self, *args, **kwargs):\n@@ -606,7 +607,7 @@ def test___ne__same_value(self):\n class TestCellsRowOffsetFilter(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.bigtable.row_filters import CellsRowOffsetFilter\n+ from google.cloud.bigtable.row_filters import CellsRowOffsetFilter\n return CellsRowOffsetFilter\n \n def _makeOne(self, *args, **kwargs):\n@@ -624,7 +625,7 @@ def test_to_pb(self):\n class TestCellsRowLimitFilter(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.bigtable.row_filters import CellsRowLimitFilter\n+ from google.cloud.bigtable.row_filters import CellsRowLimitFilter\n return CellsRowLimitFilter\n \n def _makeOne(self, *args, **kwargs):\n@@ -642,7 +643,7 @@ def test_to_pb(self):\n class TestCellsColumnLimitFilter(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.bigtable.row_filters import CellsColumnLimitFilter\n+ from google.cloud.bigtable.row_filters import CellsColumnLimitFilter\n return CellsColumnLimitFilter\n \n def _makeOne(self, *args, **kwargs):\n@@ -660,7 +661,8 @@ def test_to_pb(self):\n class TestStripValueTransformerFilter(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.bigtable.row_filters import StripValueTransformerFilter\n+ from google.cloud.bigtable.row_filters import (\n+ StripValueTransformerFilter)\n return StripValueTransformerFilter\n \n def _makeOne(self, *args, **kwargs):\n@@ -677,7 +679,7 @@ def test_to_pb(self):\n class TestApplyLabelFilter(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.bigtable.row_filters import ApplyLabelFilter\n+ from google.cloud.bigtable.row_filters import ApplyLabelFilter\n return ApplyLabelFilter\n \n def _makeOne(self, *args, **kwargs):\n@@ -711,7 +713,7 @@ def test_to_pb(self):\n class Test_FilterCombination(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.bigtable.row_filters import _FilterCombination\n+ from google.cloud.bigtable.row_filters import _FilterCombination\n return _FilterCombination\n \n def _makeOne(self, *args, **kwargs):\n@@ -742,15 +744,16 @@ def test___eq__type_differ(self):\n class TestRowFilterChain(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.bigtable.row_filters import RowFilterChain\n+ from google.cloud.bigtable.row_filters import RowFilterChain\n return RowFilterChain\n \n def _makeOne(self, *args, **kwargs):\n return self._getTargetClass()(*args, **kwargs)\n \n def test_to_pb(self):\n- from gcloud.bigtable.row_filters import RowSampleFilter\n- from gcloud.bigtable.row_filters import StripValueTransformerFilter\n+ from google.cloud.bigtable.row_filters import RowSampleFilter\n+ from google.cloud.bigtable.row_filters import (\n+ StripValueTransformerFilter)\n \n row_filter1 = StripValueTransformerFilter(True)\n row_filter1_pb = row_filter1.to_pb()\n@@ -769,9 +772,10 @@ def test_to_pb(self):\n self.assertEqual(filter_pb, expected_pb)\n \n def test_to_pb_nested(self):\n- from gcloud.bigtable.row_filters import CellsRowLimitFilter\n- from gcloud.bigtable.row_filters import RowSampleFilter\n- from gcloud.bigtable.row_filters import StripValueTransformerFilter\n+ from google.cloud.bigtable.row_filters import CellsRowLimitFilter\n+ from google.cloud.bigtable.row_filters import RowSampleFilter\n+ from google.cloud.bigtable.row_filters import (\n+ StripValueTransformerFilter)\n \n row_filter1 = StripValueTransformerFilter(True)\n row_filter2 = RowSampleFilter(0.25)\n@@ -796,15 +800,16 @@ def test_to_pb_nested(self):\n class TestRowFilterUnion(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.bigtable.row_filters import RowFilterUnion\n+ from google.cloud.bigtable.row_filters import RowFilterUnion\n return RowFilterUnion\n \n def _makeOne(self, *args, **kwargs):\n return self._getTargetClass()(*args, **kwargs)\n \n def test_to_pb(self):\n- from gcloud.bigtable.row_filters import RowSampleFilter\n- from gcloud.bigtable.row_filters import StripValueTransformerFilter\n+ from google.cloud.bigtable.row_filters import RowSampleFilter\n+ from google.cloud.bigtable.row_filters import (\n+ StripValueTransformerFilter)\n \n row_filter1 = StripValueTransformerFilter(True)\n row_filter1_pb = row_filter1.to_pb()\n@@ -823,9 +828,10 @@ def test_to_pb(self):\n self.assertEqual(filter_pb, expected_pb)\n \n def test_to_pb_nested(self):\n- from gcloud.bigtable.row_filters import CellsRowLimitFilter\n- from gcloud.bigtable.row_filters import RowSampleFilter\n- from gcloud.bigtable.row_filters import StripValueTransformerFilter\n+ from google.cloud.bigtable.row_filters import CellsRowLimitFilter\n+ from google.cloud.bigtable.row_filters import RowSampleFilter\n+ from google.cloud.bigtable.row_filters import (\n+ StripValueTransformerFilter)\n \n row_filter1 = StripValueTransformerFilter(True)\n row_filter2 = RowSampleFilter(0.25)\n@@ -850,7 +856,7 @@ def test_to_pb_nested(self):\n class TestConditionalRowFilter(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.bigtable.row_filters import ConditionalRowFilter\n+ from google.cloud.bigtable.row_filters import ConditionalRowFilter\n return ConditionalRowFilter\n \n def _makeOne(self, *args, **kwargs):\n@@ -890,9 +896,10 @@ def test___eq__type_differ(self):\n self.assertNotEqual(cond_filter1, cond_filter2)\n \n def test_to_pb(self):\n- from gcloud.bigtable.row_filters import CellsRowOffsetFilter\n- from gcloud.bigtable.row_filters import RowSampleFilter\n- from gcloud.bigtable.row_filters import StripValueTransformerFilter\n+ from google.cloud.bigtable.row_filters import CellsRowOffsetFilter\n+ from google.cloud.bigtable.row_filters import RowSampleFilter\n+ from google.cloud.bigtable.row_filters import (\n+ StripValueTransformerFilter)\n \n row_filter1 = StripValueTransformerFilter(True)\n row_filter1_pb = row_filter1.to_pb()\n@@ -917,8 +924,9 @@ def test_to_pb(self):\n self.assertEqual(filter_pb, expected_pb)\n \n def test_to_pb_true_only(self):\n- from gcloud.bigtable.row_filters import RowSampleFilter\n- from gcloud.bigtable.row_filters import StripValueTransformerFilter\n+ from google.cloud.bigtable.row_filters import RowSampleFilter\n+ from google.cloud.bigtable.row_filters import (\n+ StripValueTransformerFilter)\n \n row_filter1 = StripValueTransformerFilter(True)\n row_filter1_pb = row_filter1.to_pb()\n@@ -938,8 +946,9 @@ def test_to_pb_true_only(self):\n self.assertEqual(filter_pb, expected_pb)\n \n def test_to_pb_false_only(self):\n- from gcloud.bigtable.row_filters import RowSampleFilter\n- from gcloud.bigtable.row_filters import StripValueTransformerFilter\n+ from google.cloud.bigtable.row_filters import RowSampleFilter\n+ from google.cloud.bigtable.row_filters import (\n+ StripValueTransformerFilter)\n \n row_filter1 = StripValueTransformerFilter(True)\n row_filter1_pb = row_filter1.to_pb()\n@@ -960,42 +969,42 @@ def test_to_pb_false_only(self):\n \n \n def _ColumnRangePB(*args, **kw):\n- from gcloud.bigtable._generated import (\n+ from google.cloud.bigtable._generated import (\n data_pb2 as data_v2_pb2)\n return data_v2_pb2.ColumnRange(*args, **kw)\n \n \n def _RowFilterPB(*args, **kw):\n- from gcloud.bigtable._generated import (\n+ from google.cloud.bigtable._generated import (\n data_pb2 as data_v2_pb2)\n return data_v2_pb2.RowFilter(*args, **kw)\n \n \n def _RowFilterChainPB(*args, **kw):\n- from gcloud.bigtable._generated import (\n+ from google.cloud.bigtable._generated import (\n data_pb2 as data_v2_pb2)\n return data_v2_pb2.RowFilter.Chain(*args, **kw)\n \n \n def _RowFilterConditionPB(*args, **kw):\n- from gcloud.bigtable._generated import (\n+ from google.cloud.bigtable._generated import (\n data_pb2 as data_v2_pb2)\n return data_v2_pb2.RowFilter.Condition(*args, **kw)\n \n \n def _RowFilterInterleavePB(*args, **kw):\n- from gcloud.bigtable._generated import (\n+ from google.cloud.bigtable._generated import (\n data_pb2 as data_v2_pb2)\n return data_v2_pb2.RowFilter.Interleave(*args, **kw)\n \n \n def _TimestampRangePB(*args, **kw):\n- from gcloud.bigtable._generated import (\n+ from google.cloud.bigtable._generated import (\n data_pb2 as data_v2_pb2)\n return data_v2_pb2.TimestampRange(*args, **kw)\n \n \n def _ValueRangePB(*args, **kw):\n- from gcloud.bigtable._generated import (\n+ from google.cloud.bigtable._generated import (\n data_pb2 as data_v2_pb2)\n return data_v2_pb2.ValueRange(*args, **kw)\ndiff --git a/unit_tests/bigtable/test_table.py b/unit_tests/bigtable/test_table.py\n--- a/unit_tests/bigtable/test_table.py\n+++ b/unit_tests/bigtable/test_table.py\n@@ -30,7 +30,7 @@ class TestTable(unittest.TestCase):\n VALUE = b'value'\n \n def _getTargetClass(self):\n- from gcloud.bigtable.table import Table\n+ from google.cloud.bigtable.table import Table\n return Table\n \n def _makeOne(self, *args, **kwargs):\n@@ -54,7 +54,7 @@ def test_name_property(self):\n self.assertEqual(table.name, expected_name)\n \n def test_column_family_factory(self):\n- from gcloud.bigtable.column_family import ColumnFamily\n+ from google.cloud.bigtable.column_family import ColumnFamily\n \n table_id = 'table-id'\n gc_rule = object()\n@@ -68,7 +68,7 @@ def test_column_family_factory(self):\n self.assertEqual(column_family._table, table)\n \n def test_row_factory_direct(self):\n- from gcloud.bigtable.row import DirectRow\n+ from google.cloud.bigtable.row import DirectRow\n \n table_id = 'table-id'\n table = self._makeOne(table_id, None)\n@@ -80,7 +80,7 @@ def test_row_factory_direct(self):\n self.assertEqual(row._table, table)\n \n def test_row_factory_conditional(self):\n- from gcloud.bigtable.row import ConditionalRow\n+ from google.cloud.bigtable.row import ConditionalRow\n \n table_id = 'table-id'\n table = self._makeOne(table_id, None)\n@@ -93,7 +93,7 @@ def test_row_factory_conditional(self):\n self.assertEqual(row._table, table)\n \n def test_row_factory_append(self):\n- from gcloud.bigtable.row import AppendRow\n+ from google.cloud.bigtable.row import AppendRow\n \n table_id = 'table-id'\n table = self._makeOne(table_id, None)\n@@ -133,7 +133,7 @@ def test___ne__(self):\n self.assertNotEqual(table1, table2)\n \n def _create_test_helper(self, initial_split_keys, column_families=()):\n- from gcloud._helpers import _to_bytes\n+ from google.cloud._helpers import _to_bytes\n from unit_tests.bigtable._testing import _FakeStub\n \n client = _Client()\n@@ -186,8 +186,8 @@ def test_create_with_split_keys(self):\n self._create_test_helper(initial_split_keys)\n \n def test_create_with_column_families(self):\n- from gcloud.bigtable.column_family import ColumnFamily\n- from gcloud.bigtable.column_family import MaxVersionsGCRule\n+ from google.cloud.bigtable.column_family import ColumnFamily\n+ from google.cloud.bigtable.column_family import MaxVersionsGCRule\n \n cf_id1 = 'col-fam-id1'\n cf1 = ColumnFamily(cf_id1, None)\n@@ -269,7 +269,7 @@ def test_delete(self):\n def _read_row_helper(self, chunks, expected_result):\n from unit_tests._testing import _Monkey\n from unit_tests.bigtable._testing import _FakeStub\n- from gcloud.bigtable import table as MUT\n+ from google.cloud.bigtable import table as MUT\n \n client = _Client()\n instance = _Instance(self.INSTANCE_NAME, client=client)\n@@ -315,8 +315,8 @@ def test_read_row_miss_no_chunks_in_response(self):\n self._read_row_helper(chunks, None)\n \n def test_read_row_complete(self):\n- from gcloud.bigtable.row_data import Cell\n- from gcloud.bigtable.row_data import PartialRowData\n+ from google.cloud.bigtable.row_data import Cell\n+ from google.cloud.bigtable.row_data import PartialRowData\n \n chunk = _ReadRowsResponseCellChunkPB(\n row_key=self.ROW_KEY,\n@@ -349,8 +349,8 @@ def test_read_row_still_partial(self):\n def test_read_rows(self):\n from unit_tests._testing import _Monkey\n from unit_tests.bigtable._testing import _FakeStub\n- from gcloud.bigtable.row_data import PartialRowsData\n- from gcloud.bigtable import table as MUT\n+ from google.cloud.bigtable.row_data import PartialRowsData\n+ from google.cloud.bigtable import table as MUT\n \n client = _Client()\n instance = _Instance(self.INSTANCE_NAME, client=client)\n@@ -430,7 +430,7 @@ class Test__create_row_request(unittest.TestCase):\n \n def _callFUT(self, table_name, row_key=None, start_key=None, end_key=None,\n filter_=None, limit=None):\n- from gcloud.bigtable.table import _create_row_request\n+ from google.cloud.bigtable.table import _create_row_request\n return _create_row_request(\n table_name, row_key=row_key, start_key=start_key, end_key=end_key,\n filter_=filter_, limit=limit)\n@@ -484,7 +484,7 @@ def test_row_range_both_keys(self):\n self.assertEqual(result, expected_result)\n \n def test_with_filter(self):\n- from gcloud.bigtable.row_filters import RowSampleFilter\n+ from google.cloud.bigtable.row_filters import RowSampleFilter\n table_name = 'table_name'\n row_filter = RowSampleFilter(0.33)\n result = self._callFUT(table_name, filter_=row_filter)\n@@ -506,37 +506,37 @@ def test_with_limit(self):\n \n \n def _CreateTableRequestPB(*args, **kw):\n- from gcloud.bigtable._generated import (\n+ from google.cloud.bigtable._generated import (\n bigtable_table_admin_pb2 as table_admin_v2_pb2)\n return table_admin_v2_pb2.CreateTableRequest(*args, **kw)\n \n \n def _CreateTableRequestSplitPB(*args, **kw):\n- from gcloud.bigtable._generated import (\n+ from google.cloud.bigtable._generated import (\n bigtable_table_admin_pb2 as table_admin_v2_pb2)\n return table_admin_v2_pb2.CreateTableRequest.Split(*args, **kw)\n \n \n def _DeleteTableRequestPB(*args, **kw):\n- from gcloud.bigtable._generated import (\n+ from google.cloud.bigtable._generated import (\n bigtable_table_admin_pb2 as table_admin_v2_pb2)\n return table_admin_v2_pb2.DeleteTableRequest(*args, **kw)\n \n \n def _GetTableRequestPB(*args, **kw):\n- from gcloud.bigtable._generated import (\n+ from google.cloud.bigtable._generated import (\n bigtable_table_admin_pb2 as table_admin_v2_pb2)\n return table_admin_v2_pb2.GetTableRequest(*args, **kw)\n \n \n def _ReadRowsRequestPB(*args, **kw):\n- from gcloud.bigtable._generated import (\n+ from google.cloud.bigtable._generated import (\n bigtable_pb2 as messages_v2_pb2)\n return messages_v2_pb2.ReadRowsRequest(*args, **kw)\n \n \n def _ReadRowsResponseCellChunkPB(*args, **kw):\n- from gcloud.bigtable._generated import (\n+ from google.cloud.bigtable._generated import (\n bigtable_pb2 as messages_v2_pb2)\n family_name = kw.pop('family_name')\n qualifier = kw.pop('qualifier')\n@@ -547,25 +547,25 @@ def _ReadRowsResponseCellChunkPB(*args, **kw):\n \n \n def _ReadRowsResponsePB(*args, **kw):\n- from gcloud.bigtable._generated import (\n+ from google.cloud.bigtable._generated import (\n bigtable_pb2 as messages_v2_pb2)\n return messages_v2_pb2.ReadRowsResponse(*args, **kw)\n \n \n def _SampleRowKeysRequestPB(*args, **kw):\n- from gcloud.bigtable._generated import (\n+ from google.cloud.bigtable._generated import (\n bigtable_pb2 as messages_v2_pb2)\n return messages_v2_pb2.SampleRowKeysRequest(*args, **kw)\n \n \n def _TablePB(*args, **kw):\n- from gcloud.bigtable._generated import (\n+ from google.cloud.bigtable._generated import (\n table_pb2 as table_v2_pb2)\n return table_v2_pb2.Table(*args, **kw)\n \n \n def _ColumnFamilyPB(*args, **kw):\n- from gcloud.bigtable._generated import (\n+ from google.cloud.bigtable._generated import (\n table_pb2 as table_v2_pb2)\n return table_v2_pb2.ColumnFamily(*args, **kw)\n \ndiff --git a/unit_tests/datastore/test_batch.py b/unit_tests/datastore/test_batch.py\n--- a/unit_tests/datastore/test_batch.py\n+++ b/unit_tests/datastore/test_batch.py\n@@ -18,7 +18,7 @@\n class TestBatch(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.datastore.batch import Batch\n+ from google.cloud.datastore.batch import Batch\n \n return Batch\n \n@@ -26,7 +26,7 @@ def _makeOne(self, client):\n return self._getTargetClass()(client)\n \n def test_ctor(self):\n- from gcloud.datastore._generated import datastore_pb2\n+ from google.cloud.datastore._generated import datastore_pb2\n _PROJECT = 'PROJECT'\n _NAMESPACE = 'NAMESPACE'\n connection = _Connection()\n@@ -97,7 +97,7 @@ def test_put_entity_w_partial_key(self):\n self.assertEqual(batch._partial_key_entities, [entity])\n \n def test_put_entity_w_completed_key(self):\n- from gcloud.datastore.helpers import _property_tuples\n+ from google.cloud.datastore.helpers import _property_tuples\n \n _PROJECT = 'PROJECT'\n _PROPERTIES = {\n@@ -344,7 +344,7 @@ def is_partial(self):\n return self._id is None\n \n def to_protobuf(self):\n- from gcloud.datastore._generated import entity_pb2\n+ from google.cloud.datastore._generated import entity_pb2\n key = self._key = entity_pb2.Key()\n # Don't assign it, because it will just get ripped out\n # key.partition_id.project_id = self.project\ndiff --git a/unit_tests/datastore/test_client.py b/unit_tests/datastore/test_client.py\n--- a/unit_tests/datastore/test_client.py\n+++ b/unit_tests/datastore/test_client.py\n@@ -16,8 +16,8 @@\n \n \n def _make_entity_pb(project, kind, integer_id, name=None, str_val=None):\n- from gcloud.datastore._generated import entity_pb2\n- from gcloud.datastore.helpers import _new_value_pb\n+ from google.cloud.datastore._generated import entity_pb2\n+ from google.cloud.datastore.helpers import _new_value_pb\n \n entity_pb = entity_pb2.Entity()\n entity_pb.key.partition_id.project_id = project\n@@ -34,7 +34,7 @@ def _make_entity_pb(project, kind, integer_id, name=None, str_val=None):\n class Test__get_gcd_project(unittest.TestCase):\n \n def _callFUT(self):\n- from gcloud.datastore.client import _get_gcd_project\n+ from google.cloud.datastore.client import _get_gcd_project\n return _get_gcd_project()\n \n def test_no_value(self):\n@@ -49,7 +49,7 @@ def test_no_value(self):\n def test_value_set(self):\n import os\n from unit_tests._testing import _Monkey\n- from gcloud.datastore.client import GCD_DATASET\n+ from google.cloud.datastore.client import GCD_DATASET\n \n MOCK_PROJECT = object()\n environ = {GCD_DATASET: MOCK_PROJECT}\n@@ -61,14 +61,14 @@ def test_value_set(self):\n class Test__determine_default_project(unittest.TestCase):\n \n def _callFUT(self, project=None):\n- from gcloud.datastore.client import (\n+ from google.cloud.datastore.client import (\n _determine_default_project)\n return _determine_default_project(project=project)\n \n def _determine_default_helper(self, gcd=None, fallback=None,\n project_called=None):\n from unit_tests._testing import _Monkey\n- from gcloud.datastore import client\n+ from google.cloud.datastore import client\n \n _callers = []\n \n@@ -129,7 +129,7 @@ def tearDown(self):\n KLASS._connection_class = self.original_cnxn_class\n \n def _getTargetClass(self):\n- from gcloud.datastore.client import Client\n+ from google.cloud.datastore.client import Client\n return Client\n \n def _makeOne(self, project=PROJECT, namespace=None,\n@@ -141,7 +141,7 @@ def _makeOne(self, project=PROJECT, namespace=None,\n \n def test_ctor_w_project_no_environ(self):\n from unit_tests._testing import _Monkey\n- from gcloud.datastore import client as _MUT\n+ from google.cloud.datastore import client as _MUT\n \n # Some environments (e.g. AppVeyor CI) run in GCE, so\n # this test would fail artificially.\n@@ -150,8 +150,8 @@ def test_ctor_w_project_no_environ(self):\n \n def test_ctor_w_implicit_inputs(self):\n from unit_tests._testing import _Monkey\n- from gcloud.datastore import client as _MUT\n- from gcloud import client as _base_client\n+ from google.cloud.datastore import client as _MUT\n+ from google.cloud import client as _base_client\n \n OTHER = 'other'\n creds = object()\n@@ -263,7 +263,7 @@ def test_get_multi_no_keys(self):\n self.assertEqual(results, [])\n \n def test_get_multi_miss(self):\n- from gcloud.datastore.key import Key\n+ from google.cloud.datastore.key import Key\n \n creds = object()\n client = self._makeOne(credentials=creds)\n@@ -273,8 +273,8 @@ def test_get_multi_miss(self):\n self.assertEqual(results, [])\n \n def test_get_multi_miss_w_missing(self):\n- from gcloud.datastore._generated import entity_pb2\n- from gcloud.datastore.key import Key\n+ from google.cloud.datastore._generated import entity_pb2\n+ from google.cloud.datastore.key import Key\n \n KIND = 'Kind'\n ID = 1234\n@@ -299,7 +299,7 @@ def test_get_multi_miss_w_missing(self):\n [key.to_protobuf()])\n \n def test_get_multi_w_missing_non_empty(self):\n- from gcloud.datastore.key import Key\n+ from google.cloud.datastore.key import Key\n \n creds = object()\n client = self._makeOne(credentials=creds)\n@@ -310,7 +310,7 @@ def test_get_multi_w_missing_non_empty(self):\n [key], missing=missing)\n \n def test_get_multi_w_deferred_non_empty(self):\n- from gcloud.datastore.key import Key\n+ from google.cloud.datastore.key import Key\n \n creds = object()\n client = self._makeOne(credentials=creds)\n@@ -321,7 +321,7 @@ def test_get_multi_w_deferred_non_empty(self):\n [key], deferred=deferred)\n \n def test_get_multi_miss_w_deferred(self):\n- from gcloud.datastore.key import Key\n+ from google.cloud.datastore.key import Key\n \n key = Key('Kind', 1234, project=self.PROJECT)\n \n@@ -337,9 +337,9 @@ def test_get_multi_miss_w_deferred(self):\n [key.to_protobuf()])\n \n def test_get_multi_w_deferred_from_backend_but_not_passed(self):\n- from gcloud.datastore._generated import entity_pb2\n- from gcloud.datastore.entity import Entity\n- from gcloud.datastore.key import Key\n+ from google.cloud.datastore._generated import entity_pb2\n+ from google.cloud.datastore.entity import Entity\n+ from google.cloud.datastore.key import Key\n \n key1 = Key('Kind', project=self.PROJECT)\n key1_pb = key1.to_protobuf()\n@@ -390,7 +390,7 @@ def test_get_multi_w_deferred_from_backend_but_not_passed(self):\n self.assertTrue(tid is None)\n \n def test_get_multi_hit(self):\n- from gcloud.datastore.key import Key\n+ from google.cloud.datastore.key import Key\n \n KIND = 'Kind'\n ID = 1234\n@@ -416,7 +416,7 @@ def test_get_multi_hit(self):\n self.assertEqual(result['foo'], 'Foo')\n \n def test_get_multi_hit_w_transaction(self):\n- from gcloud.datastore.key import Key\n+ from google.cloud.datastore.key import Key\n \n TXN_ID = '123'\n KIND = 'Kind'\n@@ -450,7 +450,7 @@ def test_get_multi_hit_w_transaction(self):\n self.assertEqual(transaction_id, TXN_ID)\n \n def test_get_multi_hit_multiple_keys_same_project(self):\n- from gcloud.datastore.key import Key\n+ from google.cloud.datastore.key import Key\n \n KIND = 'Kind'\n ID1 = 1234\n@@ -476,7 +476,7 @@ def test_get_multi_hit_multiple_keys_same_project(self):\n self.assertEqual(dict(retrieved2), {})\n \n def test_get_multi_hit_multiple_keys_different_project(self):\n- from gcloud.datastore.key import Key\n+ from google.cloud.datastore.key import Key\n \n PROJECT1 = 'PROJECT'\n PROJECT2 = 'PROJECT-ALT'\n@@ -495,8 +495,8 @@ def test_get_multi_hit_multiple_keys_different_project(self):\n \n def test_get_multi_max_loops(self):\n from unit_tests._testing import _Monkey\n- from gcloud.datastore import client as _MUT\n- from gcloud.datastore.key import Key\n+ from google.cloud.datastore import client as _MUT\n+ from google.cloud.datastore.key import Key\n \n KIND = 'Kind'\n ID = 1234\n@@ -544,15 +544,15 @@ def test_put_multi_no_entities(self):\n self.assertEqual(client.put_multi([]), None)\n \n def test_put_multi_w_single_empty_entity(self):\n- # https://github.com/GoogleCloudPlatform/gcloud-python/issues/649\n- from gcloud.datastore.entity import Entity\n+ # https://github.com/GoogleCloudPlatform/google-cloud-python/issues/649\n+ from google.cloud.datastore.entity import Entity\n \n creds = object()\n client = self._makeOne(credentials=creds)\n self.assertRaises(ValueError, client.put_multi, Entity())\n \n def test_put_multi_no_batch_w_partial_key(self):\n- from gcloud.datastore.helpers import _property_tuples\n+ from google.cloud.datastore.helpers import _property_tuples\n \n entity = _Entity(foo=u'bar')\n key = entity.key = _Key(self.PROJECT)\n@@ -582,7 +582,7 @@ def test_put_multi_no_batch_w_partial_key(self):\n self.assertTrue(transaction_id is None)\n \n def test_put_multi_existing_batch_w_completed_key(self):\n- from gcloud.datastore.helpers import _property_tuples\n+ from google.cloud.datastore.helpers import _property_tuples\n \n creds = object()\n client = self._makeOne(credentials=creds)\n@@ -701,7 +701,7 @@ def test_key_w_project(self):\n client.key, KIND, ID, project=self.PROJECT)\n \n def test_key_wo_project(self):\n- from gcloud.datastore import client as MUT\n+ from google.cloud.datastore import client as MUT\n from unit_tests._testing import _Monkey\n \n KIND = 'KIND'\n@@ -722,7 +722,7 @@ def test_key_wo_project(self):\n self.assertEqual(key.kwargs, expected_kwargs)\n \n def test_key_w_namespace(self):\n- from gcloud.datastore import client as MUT\n+ from google.cloud.datastore import client as MUT\n from unit_tests._testing import _Monkey\n \n KIND = 'KIND'\n@@ -743,7 +743,7 @@ def test_key_w_namespace(self):\n self.assertEqual(key.kwargs, expected_kwargs)\n \n def test_key_w_namespace_collision(self):\n- from gcloud.datastore import client as MUT\n+ from google.cloud.datastore import client as MUT\n from unit_tests._testing import _Monkey\n \n KIND = 'KIND'\n@@ -765,7 +765,7 @@ def test_key_w_namespace_collision(self):\n self.assertEqual(key.kwargs, expected_kwargs)\n \n def test_batch(self):\n- from gcloud.datastore import client as MUT\n+ from google.cloud.datastore import client as MUT\n from unit_tests._testing import _Monkey\n \n creds = object()\n@@ -779,7 +779,7 @@ def test_batch(self):\n self.assertEqual(batch.kwargs, {})\n \n def test_transaction_defaults(self):\n- from gcloud.datastore import client as MUT\n+ from google.cloud.datastore import client as MUT\n from unit_tests._testing import _Monkey\n \n creds = object()\n@@ -811,7 +811,7 @@ def test_query_w_project(self):\n client.query, kind=KIND, project=self.PROJECT)\n \n def test_query_w_defaults(self):\n- from gcloud.datastore import client as MUT\n+ from google.cloud.datastore import client as MUT\n from unit_tests._testing import _Monkey\n \n creds = object()\n@@ -829,7 +829,7 @@ def test_query_w_defaults(self):\n self.assertEqual(query.kwargs, expected_kwargs)\n \n def test_query_explicit(self):\n- from gcloud.datastore import client as MUT\n+ from google.cloud.datastore import client as MUT\n from unit_tests._testing import _Monkey\n \n KIND = 'KIND'\n@@ -869,7 +869,7 @@ def test_query_explicit(self):\n self.assertEqual(query.kwargs, kwargs)\n \n def test_query_w_namespace(self):\n- from gcloud.datastore import client as MUT\n+ from google.cloud.datastore import client as MUT\n from unit_tests._testing import _Monkey\n \n KIND = 'KIND'\n@@ -891,7 +891,7 @@ def test_query_w_namespace(self):\n self.assertEqual(query.kwargs, expected_kwargs)\n \n def test_query_w_namespace_collision(self):\n- from gcloud.datastore import client as MUT\n+ from google.cloud.datastore import client as MUT\n from unit_tests._testing import _Monkey\n \n KIND = 'KIND'\n@@ -957,7 +957,7 @@ def allocate_ids(self, project, key_pbs):\n class _NoCommitBatch(object):\n \n def __init__(self, client):\n- from gcloud.datastore.batch import Batch\n+ from google.cloud.datastore.batch import Batch\n self._client = client\n self._batch = Batch(client)\n \n@@ -972,7 +972,7 @@ def __exit__(self, *args):\n class _NoCommitTransaction(object):\n \n def __init__(self, client, transaction_id='TRANSACTION'):\n- from gcloud.datastore.transaction import Transaction\n+ from google.cloud.datastore.transaction import Transaction\n self._client = client\n xact = self._transaction = Transaction(client)\n xact._id = transaction_id\n@@ -1007,7 +1007,7 @@ def is_partial(self):\n return self._id is None\n \n def to_protobuf(self):\n- from gcloud.datastore._generated import entity_pb2\n+ from google.cloud.datastore._generated import entity_pb2\n key = self._key = entity_pb2.Key()\n # Don't assign it, because it will just get ripped out\n # key.partition_id.project_id = self.project\ndiff --git a/unit_tests/datastore/test_connection.py b/unit_tests/datastore/test_connection.py\n--- a/unit_tests/datastore/test_connection.py\n+++ b/unit_tests/datastore/test_connection.py\n@@ -14,13 +14,13 @@\n \n import unittest\n \n-from gcloud.datastore.connection import _HAVE_GRPC\n+from google.cloud.datastore.connection import _HAVE_GRPC\n \n \n class Test_DatastoreAPIOverHttp(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.datastore.connection import _DatastoreAPIOverHttp\n+ from google.cloud.datastore.connection import _DatastoreAPIOverHttp\n return _DatastoreAPIOverHttp\n \n def _makeOne(self, *args, **kw):\n@@ -84,7 +84,7 @@ def test__request_w_200(self):\n [{'method': METHOD, 'project': PROJECT}])\n \n def test__request_not_200(self):\n- from gcloud.exceptions import BadRequest\n+ from google.cloud.exceptions import BadRequest\n from google.rpc import status_pb2\n \n error = status_pb2.Status()\n@@ -109,12 +109,12 @@ def test__request_not_200(self):\n class Test_DatastoreAPIOverGRPC(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.datastore.connection import _DatastoreAPIOverGRPC\n+ from google.cloud.datastore.connection import _DatastoreAPIOverGRPC\n return _DatastoreAPIOverGRPC\n \n def _makeOne(self, stub, connection=None, mock_args=None):\n from unit_tests._testing import _Monkey\n- from gcloud.datastore import connection as MUT\n+ from google.cloud.datastore import connection as MUT\n \n if connection is None:\n connection = _Connection(None)\n@@ -131,7 +131,7 @@ def mock_make_stub(*args):\n return self._getTargetClass()(connection)\n \n def test_constructor(self):\n- from gcloud.datastore import connection as MUT\n+ from google.cloud.datastore import connection as MUT\n \n conn = _Connection(None)\n conn.credentials = object()\n@@ -221,7 +221,7 @@ def test_commit_failure_aborted(self):\n from grpc import StatusCode\n from grpc._channel import _Rendezvous\n from grpc._channel import _RPCState\n- from gcloud.exceptions import Conflict\n+ from google.cloud.exceptions import Conflict\n \n details = 'Bad things.'\n exc_state = _RPCState((), None, None, StatusCode.ABORTED, details)\n@@ -274,19 +274,19 @@ def test_allocate_ids(self):\n class TestConnection(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.datastore.connection import Connection\n+ from google.cloud.datastore.connection import Connection\n \n return Connection\n \n def _make_key_pb(self, project, id_=1234):\n- from gcloud.datastore.key import Key\n+ from google.cloud.datastore.key import Key\n path_args = ('Kind',)\n if id_ is not None:\n path_args += (id_,)\n return Key(*path_args, project=project).to_protobuf()\n \n def _make_query_pb(self, kind):\n- from gcloud.datastore._generated import query_pb2\n+ from google.cloud.datastore._generated import query_pb2\n pb = query_pb2.Query()\n pb.kind.add().name = kind\n return pb\n@@ -294,7 +294,7 @@ def _make_query_pb(self, kind):\n def _makeOne(self, credentials=None, http=None,\n api_base_url=None, have_grpc=False):\n from unit_tests._testing import _Monkey\n- from gcloud.datastore import connection as MUT\n+ from google.cloud.datastore import connection as MUT\n with _Monkey(MUT, _HAVE_GRPC=have_grpc):\n return self._getTargetClass()(credentials=credentials, http=http,\n api_base_url=api_base_url)\n@@ -315,8 +315,8 @@ def test_default_url(self):\n def test_custom_url_from_env(self):\n import os\n from unit_tests._testing import _Monkey\n- from gcloud.connection import API_BASE_URL\n- from gcloud.environment_vars import GCD_HOST\n+ from google.cloud.connection import API_BASE_URL\n+ from google.cloud.environment_vars import GCD_HOST\n \n HOST = 'CURR_HOST'\n fake_environ = {GCD_HOST: HOST}\n@@ -328,7 +328,7 @@ def test_custom_url_from_env(self):\n self.assertEqual(conn.api_base_url, HOST + '/datastore')\n \n def test_custom_url_from_constructor(self):\n- from gcloud.connection import API_BASE_URL\n+ from google.cloud.connection import API_BASE_URL\n \n HOST = object()\n conn = self._makeOne(api_base_url=HOST)\n@@ -338,8 +338,8 @@ def test_custom_url_from_constructor(self):\n def test_custom_url_constructor_and_env(self):\n import os\n from unit_tests._testing import _Monkey\n- from gcloud.connection import API_BASE_URL\n- from gcloud.environment_vars import GCD_HOST\n+ from google.cloud.connection import API_BASE_URL\n+ from google.cloud.environment_vars import GCD_HOST\n \n HOST1 = object()\n HOST2 = object()\n@@ -358,7 +358,7 @@ def test_ctor_defaults(self):\n \n def test_ctor_without_grpc(self):\n from unit_tests._testing import _Monkey\n- from gcloud.datastore import connection as MUT\n+ from google.cloud.datastore import connection as MUT\n \n connections = []\n return_val = object()\n@@ -376,7 +376,7 @@ def mock_api(connection):\n \n def test_ctor_with_grpc(self):\n from unit_tests._testing import _Monkey\n- from gcloud.datastore import connection as MUT\n+ from google.cloud.datastore import connection as MUT\n \n connections = []\n return_val = object()\n@@ -460,7 +460,7 @@ def test_build_api_url_w_explicit_base_version(self):\n URI)\n \n def test_lookup_single_key_empty_response(self):\n- from gcloud.datastore._generated import datastore_pb2\n+ from google.cloud.datastore._generated import datastore_pb2\n \n PROJECT = 'PROJECT'\n key_pb = self._make_key_pb(PROJECT)\n@@ -487,7 +487,7 @@ def test_lookup_single_key_empty_response(self):\n self.assertEqual(key_pb, keys[0])\n \n def test_lookup_single_key_empty_response_w_eventual(self):\n- from gcloud.datastore._generated import datastore_pb2\n+ from google.cloud.datastore._generated import datastore_pb2\n \n PROJECT = 'PROJECT'\n key_pb = self._make_key_pb(PROJECT)\n@@ -526,7 +526,7 @@ def test_lookup_single_key_empty_response_w_eventual_and_transaction(self):\n eventual=True, transaction_id=TRANSACTION)\n \n def test_lookup_single_key_empty_response_w_transaction(self):\n- from gcloud.datastore._generated import datastore_pb2\n+ from google.cloud.datastore._generated import datastore_pb2\n \n PROJECT = 'PROJECT'\n TRANSACTION = b'TRANSACTION'\n@@ -556,8 +556,8 @@ def test_lookup_single_key_empty_response_w_transaction(self):\n self.assertEqual(request.read_options.transaction, TRANSACTION)\n \n def test_lookup_single_key_nonempty_response(self):\n- from gcloud.datastore._generated import datastore_pb2\n- from gcloud.datastore._generated import entity_pb2\n+ from google.cloud.datastore._generated import datastore_pb2\n+ from google.cloud.datastore._generated import entity_pb2\n \n PROJECT = 'PROJECT'\n key_pb = self._make_key_pb(PROJECT)\n@@ -588,7 +588,7 @@ def test_lookup_single_key_nonempty_response(self):\n self.assertEqual(key_pb, keys[0])\n \n def test_lookup_multiple_keys_empty_response(self):\n- from gcloud.datastore._generated import datastore_pb2\n+ from google.cloud.datastore._generated import datastore_pb2\n \n PROJECT = 'PROJECT'\n key_pb1 = self._make_key_pb(PROJECT)\n@@ -617,7 +617,7 @@ def test_lookup_multiple_keys_empty_response(self):\n self.assertEqual(key_pb2, keys[1])\n \n def test_lookup_multiple_keys_w_missing(self):\n- from gcloud.datastore._generated import datastore_pb2\n+ from google.cloud.datastore._generated import datastore_pb2\n \n PROJECT = 'PROJECT'\n key_pb1 = self._make_key_pb(PROJECT)\n@@ -651,7 +651,7 @@ def test_lookup_multiple_keys_w_missing(self):\n self.assertEqual(key_pb2, keys[1])\n \n def test_lookup_multiple_keys_w_deferred(self):\n- from gcloud.datastore._generated import datastore_pb2\n+ from google.cloud.datastore._generated import datastore_pb2\n \n PROJECT = 'PROJECT'\n key_pb1 = self._make_key_pb(PROJECT)\n@@ -687,8 +687,8 @@ def test_lookup_multiple_keys_w_deferred(self):\n self.assertEqual(key_pb2, keys[1])\n \n def test_run_query_w_eventual_no_transaction(self):\n- from gcloud.datastore._generated import datastore_pb2\n- from gcloud.datastore._generated import query_pb2\n+ from google.cloud.datastore._generated import datastore_pb2\n+ from google.cloud.datastore._generated import query_pb2\n \n PROJECT = 'PROJECT'\n KIND = 'Nonesuch'\n@@ -725,8 +725,8 @@ def test_run_query_w_eventual_no_transaction(self):\n self.assertEqual(request.read_options.transaction, b'')\n \n def test_run_query_wo_eventual_w_transaction(self):\n- from gcloud.datastore._generated import datastore_pb2\n- from gcloud.datastore._generated import query_pb2\n+ from google.cloud.datastore._generated import datastore_pb2\n+ from google.cloud.datastore._generated import query_pb2\n \n PROJECT = 'PROJECT'\n KIND = 'Nonesuch'\n@@ -765,8 +765,8 @@ def test_run_query_wo_eventual_w_transaction(self):\n self.assertEqual(request.read_options.transaction, TRANSACTION)\n \n def test_run_query_w_eventual_and_transaction(self):\n- from gcloud.datastore._generated import datastore_pb2\n- from gcloud.datastore._generated import query_pb2\n+ from google.cloud.datastore._generated import datastore_pb2\n+ from google.cloud.datastore._generated import query_pb2\n \n PROJECT = 'PROJECT'\n KIND = 'Nonesuch'\n@@ -783,8 +783,8 @@ def test_run_query_w_eventual_and_transaction(self):\n eventual=True, transaction_id=TRANSACTION)\n \n def test_run_query_wo_namespace_empty_result(self):\n- from gcloud.datastore._generated import datastore_pb2\n- from gcloud.datastore._generated import query_pb2\n+ from google.cloud.datastore._generated import datastore_pb2\n+ from google.cloud.datastore._generated import query_pb2\n \n PROJECT = 'PROJECT'\n KIND = 'Nonesuch'\n@@ -817,8 +817,8 @@ def test_run_query_wo_namespace_empty_result(self):\n self.assertEqual(request.query, q_pb)\n \n def test_run_query_w_namespace_nonempty_result(self):\n- from gcloud.datastore._generated import datastore_pb2\n- from gcloud.datastore._generated import entity_pb2\n+ from google.cloud.datastore._generated import datastore_pb2\n+ from google.cloud.datastore._generated import entity_pb2\n \n PROJECT = 'PROJECT'\n KIND = 'Kind'\n@@ -847,7 +847,7 @@ def test_run_query_w_namespace_nonempty_result(self):\n self.assertEqual(request.query, q_pb)\n \n def test_begin_transaction(self):\n- from gcloud.datastore._generated import datastore_pb2\n+ from google.cloud.datastore._generated import datastore_pb2\n \n PROJECT = 'PROJECT'\n TRANSACTION = b'TRANSACTION'\n@@ -870,9 +870,9 @@ def test_begin_transaction(self):\n \n def test_commit_wo_transaction(self):\n from unit_tests._testing import _Monkey\n- from gcloud.datastore._generated import datastore_pb2\n- from gcloud.datastore import connection as MUT\n- from gcloud.datastore.helpers import _new_value_pb\n+ from google.cloud.datastore._generated import datastore_pb2\n+ from google.cloud.datastore import connection as MUT\n+ from google.cloud.datastore.helpers import _new_value_pb\n \n PROJECT = 'PROJECT'\n key_pb = self._make_key_pb(PROJECT)\n@@ -916,9 +916,9 @@ def mock_parse(response):\n \n def test_commit_w_transaction(self):\n from unit_tests._testing import _Monkey\n- from gcloud.datastore._generated import datastore_pb2\n- from gcloud.datastore import connection as MUT\n- from gcloud.datastore.helpers import _new_value_pb\n+ from google.cloud.datastore._generated import datastore_pb2\n+ from google.cloud.datastore import connection as MUT\n+ from google.cloud.datastore.helpers import _new_value_pb\n \n PROJECT = 'PROJECT'\n key_pb = self._make_key_pb(PROJECT)\n@@ -961,7 +961,7 @@ def mock_parse(response):\n self.assertEqual(_parsed, [rsp_pb])\n \n def test_rollback_ok(self):\n- from gcloud.datastore._generated import datastore_pb2\n+ from google.cloud.datastore._generated import datastore_pb2\n PROJECT = 'PROJECT'\n TRANSACTION = b'xact'\n \n@@ -983,7 +983,7 @@ def test_rollback_ok(self):\n self.assertEqual(request.transaction, TRANSACTION)\n \n def test_allocate_ids_empty(self):\n- from gcloud.datastore._generated import datastore_pb2\n+ from google.cloud.datastore._generated import datastore_pb2\n \n PROJECT = 'PROJECT'\n rsp_pb = datastore_pb2.AllocateIdsResponse()\n@@ -1004,7 +1004,7 @@ def test_allocate_ids_empty(self):\n self.assertEqual(list(request.keys), [])\n \n def test_allocate_ids_non_empty(self):\n- from gcloud.datastore._generated import datastore_pb2\n+ from google.cloud.datastore._generated import datastore_pb2\n \n PROJECT = 'PROJECT'\n before_key_pbs = [\n@@ -1041,12 +1041,12 @@ def test_allocate_ids_non_empty(self):\n class Test__parse_commit_response(unittest.TestCase):\n \n def _callFUT(self, commit_response_pb):\n- from gcloud.datastore.connection import _parse_commit_response\n+ from google.cloud.datastore.connection import _parse_commit_response\n return _parse_commit_response(commit_response_pb)\n \n def test_it(self):\n- from gcloud.datastore._generated import datastore_pb2\n- from gcloud.datastore._generated import entity_pb2\n+ from google.cloud.datastore._generated import datastore_pb2\n+ from google.cloud.datastore._generated import entity_pb2\n \n index_updates = 1337\n keys = [\ndiff --git a/unit_tests/datastore/test_entity.py b/unit_tests/datastore/test_entity.py\n--- a/unit_tests/datastore/test_entity.py\n+++ b/unit_tests/datastore/test_entity.py\n@@ -22,7 +22,7 @@\n class TestEntity(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.datastore.entity import Entity\n+ from google.cloud.datastore.entity import Entity\n return Entity\n \n def _makeOne(self, key=None, exclude_from_indexes=()):\n@@ -51,14 +51,14 @@ def test_ctor_bad_exclude_from_indexes(self):\n exclude_from_indexes=BAD_EXCLUDE_FROM_INDEXES)\n \n def test___eq_____ne___w_non_entity(self):\n- from gcloud.datastore.key import Key\n+ from google.cloud.datastore.key import Key\n key = Key(_KIND, _ID, project=_PROJECT)\n entity = self._makeOne(key=key)\n self.assertFalse(entity == object())\n self.assertTrue(entity != object())\n \n def test___eq_____ne___w_different_keys(self):\n- from gcloud.datastore.key import Key\n+ from google.cloud.datastore.key import Key\n _ID1 = 1234\n _ID2 = 2345\n key1 = Key(_KIND, _ID1, project=_PROJECT)\n@@ -69,7 +69,7 @@ def test___eq_____ne___w_different_keys(self):\n self.assertTrue(entity1 != entity2)\n \n def test___eq_____ne___w_same_keys(self):\n- from gcloud.datastore.key import Key\n+ from google.cloud.datastore.key import Key\n \n name = 'foo'\n value = 42\n@@ -89,7 +89,7 @@ def test___eq_____ne___w_same_keys(self):\n self.assertFalse(entity1 != entity2)\n \n def test___eq_____ne___w_same_keys_different_props(self):\n- from gcloud.datastore.key import Key\n+ from google.cloud.datastore.key import Key\n key1 = Key(_KIND, _ID, project=_PROJECT)\n entity1 = self._makeOne(key=key1)\n entity1['foo'] = 'Foo'\n@@ -100,7 +100,7 @@ def test___eq_____ne___w_same_keys_different_props(self):\n self.assertTrue(entity1 != entity2)\n \n def test___eq_____ne___w_same_keys_props_w_equiv_keys_as_value(self):\n- from gcloud.datastore.key import Key\n+ from google.cloud.datastore.key import Key\n key1 = Key(_KIND, _ID, project=_PROJECT)\n key2 = Key(_KIND, _ID, project=_PROJECT)\n entity1 = self._makeOne(key=key1)\n@@ -111,7 +111,7 @@ def test___eq_____ne___w_same_keys_props_w_equiv_keys_as_value(self):\n self.assertFalse(entity1 != entity2)\n \n def test___eq_____ne___w_same_keys_props_w_diff_keys_as_value(self):\n- from gcloud.datastore.key import Key\n+ from google.cloud.datastore.key import Key\n _ID1 = 1234\n _ID2 = 2345\n key1 = Key(_KIND, _ID1, project=_PROJECT)\n@@ -124,7 +124,7 @@ def test___eq_____ne___w_same_keys_props_w_diff_keys_as_value(self):\n self.assertTrue(entity1 != entity2)\n \n def test___eq_____ne___w_same_keys_props_w_equiv_entities_as_value(self):\n- from gcloud.datastore.key import Key\n+ from google.cloud.datastore.key import Key\n key = Key(_KIND, _ID, project=_PROJECT)\n entity1 = self._makeOne(key=key)\n sub1 = self._makeOne()\n@@ -138,7 +138,7 @@ def test___eq_____ne___w_same_keys_props_w_equiv_entities_as_value(self):\n self.assertFalse(entity1 != entity2)\n \n def test___eq_____ne___w_same_keys_props_w_diff_entities_as_value(self):\n- from gcloud.datastore.key import Key\n+ from google.cloud.datastore.key import Key\n key = Key(_KIND, _ID, project=_PROJECT)\n entity1 = self._makeOne(key=key)\n sub1 = self._makeOne()\n@@ -152,7 +152,7 @@ def test___eq_____ne___w_same_keys_props_w_diff_entities_as_value(self):\n self.assertTrue(entity1 != entity2)\n \n def test__eq__same_value_different_exclude(self):\n- from gcloud.datastore.key import Key\n+ from google.cloud.datastore.key import Key\n \n name = 'foo'\n value = 42\n@@ -167,7 +167,7 @@ def test__eq__same_value_different_exclude(self):\n self.assertFalse(entity1 == entity2)\n \n def test__eq__same_value_different_meanings(self):\n- from gcloud.datastore.key import Key\n+ from google.cloud.datastore.key import Key\n \n name = 'foo'\n value = 42\ndiff --git a/unit_tests/datastore/test_helpers.py b/unit_tests/datastore/test_helpers.py\n--- a/unit_tests/datastore/test_helpers.py\n+++ b/unit_tests/datastore/test_helpers.py\n@@ -18,11 +18,11 @@\n class Test__new_value_pb(unittest.TestCase):\n \n def _callFUT(self, entity_pb, name):\n- from gcloud.datastore.helpers import _new_value_pb\n+ from google.cloud.datastore.helpers import _new_value_pb\n return _new_value_pb(entity_pb, name)\n \n def test_it(self):\n- from gcloud.datastore._generated import entity_pb2\n+ from google.cloud.datastore._generated import entity_pb2\n \n entity_pb = entity_pb2.Entity()\n name = 'foo'\n@@ -36,13 +36,13 @@ def test_it(self):\n class Test__property_tuples(unittest.TestCase):\n \n def _callFUT(self, entity_pb):\n- from gcloud.datastore.helpers import _property_tuples\n+ from google.cloud.datastore.helpers import _property_tuples\n return _property_tuples(entity_pb)\n \n def test_it(self):\n import types\n- from gcloud.datastore._generated import entity_pb2\n- from gcloud.datastore.helpers import _new_value_pb\n+ from google.cloud.datastore._generated import entity_pb2\n+ from google.cloud.datastore.helpers import _new_value_pb\n \n entity_pb = entity_pb2.Entity()\n name1 = 'foo'\n@@ -59,12 +59,12 @@ def test_it(self):\n class Test_entity_from_protobuf(unittest.TestCase):\n \n def _callFUT(self, val):\n- from gcloud.datastore.helpers import entity_from_protobuf\n+ from google.cloud.datastore.helpers import entity_from_protobuf\n return entity_from_protobuf(val)\n \n def test_it(self):\n- from gcloud.datastore._generated import entity_pb2\n- from gcloud.datastore.helpers import _new_value_pb\n+ from google.cloud.datastore._generated import entity_pb2\n+ from google.cloud.datastore.helpers import _new_value_pb\n \n _PROJECT = 'PROJECT'\n _KIND = 'KIND'\n@@ -109,8 +109,8 @@ def test_it(self):\n self.assertEqual(key.id, _ID)\n \n def test_mismatched_value_indexed(self):\n- from gcloud.datastore._generated import entity_pb2\n- from gcloud.datastore.helpers import _new_value_pb\n+ from google.cloud.datastore._generated import entity_pb2\n+ from google.cloud.datastore.helpers import _new_value_pb\n \n _PROJECT = 'PROJECT'\n _KIND = 'KIND'\n@@ -133,7 +133,7 @@ def test_mismatched_value_indexed(self):\n self._callFUT(entity_pb)\n \n def test_entity_no_key(self):\n- from gcloud.datastore._generated import entity_pb2\n+ from google.cloud.datastore._generated import entity_pb2\n \n entity_pb = entity_pb2.Entity()\n entity = self._callFUT(entity_pb)\n@@ -142,8 +142,8 @@ def test_entity_no_key(self):\n self.assertEqual(dict(entity), {})\n \n def test_entity_with_meaning(self):\n- from gcloud.datastore._generated import entity_pb2\n- from gcloud.datastore.helpers import _new_value_pb\n+ from google.cloud.datastore._generated import entity_pb2\n+ from google.cloud.datastore.helpers import _new_value_pb\n \n entity_pb = entity_pb2.Entity()\n name = 'hello'\n@@ -157,8 +157,8 @@ def test_entity_with_meaning(self):\n self.assertEqual(entity._meanings, {name: (meaning, val)})\n \n def test_nested_entity_no_key(self):\n- from gcloud.datastore._generated import entity_pb2\n- from gcloud.datastore.helpers import _new_value_pb\n+ from google.cloud.datastore._generated import entity_pb2\n+ from google.cloud.datastore.helpers import _new_value_pb\n \n PROJECT = 'FOO'\n KIND = 'KIND'\n@@ -192,11 +192,11 @@ def test_nested_entity_no_key(self):\n class Test_entity_to_protobuf(unittest.TestCase):\n \n def _callFUT(self, entity):\n- from gcloud.datastore.helpers import entity_to_protobuf\n+ from google.cloud.datastore.helpers import entity_to_protobuf\n return entity_to_protobuf(entity)\n \n def _compareEntityProto(self, entity_pb1, entity_pb2):\n- from gcloud.datastore.helpers import _property_tuples\n+ from google.cloud.datastore.helpers import _property_tuples\n \n self.assertEqual(entity_pb1.key, entity_pb2.key)\n value_list1 = sorted(_property_tuples(entity_pb1))\n@@ -214,17 +214,17 @@ def _compareEntityProto(self, entity_pb1, entity_pb2):\n self.assertEqual(val1, val2)\n \n def test_empty(self):\n- from gcloud.datastore._generated import entity_pb2\n- from gcloud.datastore.entity import Entity\n+ from google.cloud.datastore._generated import entity_pb2\n+ from google.cloud.datastore.entity import Entity\n \n entity = Entity()\n entity_pb = self._callFUT(entity)\n self._compareEntityProto(entity_pb, entity_pb2.Entity())\n \n def test_key_only(self):\n- from gcloud.datastore._generated import entity_pb2\n- from gcloud.datastore.entity import Entity\n- from gcloud.datastore.key import Key\n+ from google.cloud.datastore._generated import entity_pb2\n+ from google.cloud.datastore.entity import Entity\n+ from google.cloud.datastore.key import Key\n \n kind, name = 'PATH', 'NAME'\n project = 'PROJECT'\n@@ -241,9 +241,9 @@ def test_key_only(self):\n self._compareEntityProto(entity_pb, expected_pb)\n \n def test_simple_fields(self):\n- from gcloud.datastore._generated import entity_pb2\n- from gcloud.datastore.entity import Entity\n- from gcloud.datastore.helpers import _new_value_pb\n+ from google.cloud.datastore._generated import entity_pb2\n+ from google.cloud.datastore.entity import Entity\n+ from google.cloud.datastore.helpers import _new_value_pb\n \n entity = Entity()\n name1 = 'foo'\n@@ -261,8 +261,8 @@ def test_simple_fields(self):\n self._compareEntityProto(entity_pb, expected_pb)\n \n def test_with_empty_list(self):\n- from gcloud.datastore._generated import entity_pb2\n- from gcloud.datastore.entity import Entity\n+ from google.cloud.datastore._generated import entity_pb2\n+ from google.cloud.datastore.entity import Entity\n \n entity = Entity()\n entity['foo'] = []\n@@ -271,9 +271,9 @@ def test_with_empty_list(self):\n self._compareEntityProto(entity_pb, entity_pb2.Entity())\n \n def test_inverts_to_protobuf(self):\n- from gcloud.datastore._generated import entity_pb2\n- from gcloud.datastore.helpers import _new_value_pb\n- from gcloud.datastore.helpers import entity_from_protobuf\n+ from google.cloud.datastore._generated import entity_pb2\n+ from google.cloud.datastore.helpers import _new_value_pb\n+ from google.cloud.datastore.helpers import entity_from_protobuf\n \n original_pb = entity_pb2.Entity()\n # Add a key.\n@@ -324,9 +324,9 @@ def test_inverts_to_protobuf(self):\n self._compareEntityProto(original_pb, new_pb)\n \n def test_meaning_with_change(self):\n- from gcloud.datastore._generated import entity_pb2\n- from gcloud.datastore.entity import Entity\n- from gcloud.datastore.helpers import _new_value_pb\n+ from google.cloud.datastore._generated import entity_pb2\n+ from google.cloud.datastore.entity import Entity\n+ from google.cloud.datastore.helpers import _new_value_pb\n \n entity = Entity()\n name = 'foo'\n@@ -342,9 +342,9 @@ def test_meaning_with_change(self):\n self._compareEntityProto(entity_pb, expected_pb)\n \n def test_variable_meanings(self):\n- from gcloud.datastore._generated import entity_pb2\n- from gcloud.datastore.entity import Entity\n- from gcloud.datastore.helpers import _new_value_pb\n+ from google.cloud.datastore._generated import entity_pb2\n+ from google.cloud.datastore.entity import Entity\n+ from google.cloud.datastore.helpers import _new_value_pb\n \n entity = Entity()\n name = 'quux'\n@@ -371,12 +371,12 @@ def test_variable_meanings(self):\n class Test_key_from_protobuf(unittest.TestCase):\n \n def _callFUT(self, val):\n- from gcloud.datastore.helpers import key_from_protobuf\n+ from google.cloud.datastore.helpers import key_from_protobuf\n \n return key_from_protobuf(val)\n \n def _makePB(self, project=None, namespace=None, path=()):\n- from gcloud.datastore._generated import entity_pb2\n+ from google.cloud.datastore._generated import entity_pb2\n pb = entity_pb2.Key()\n if project is not None:\n pb.partition_id.project_id = project\n@@ -425,14 +425,14 @@ def test_w_nothing_in_pb(self):\n class Test__pb_attr_value(unittest.TestCase):\n \n def _callFUT(self, val):\n- from gcloud.datastore.helpers import _pb_attr_value\n+ from google.cloud.datastore.helpers import _pb_attr_value\n \n return _pb_attr_value(val)\n \n def test_datetime_naive(self):\n import calendar\n import datetime\n- from gcloud._helpers import UTC\n+ from google.cloud._helpers import UTC\n \n micros = 4375\n naive = datetime.datetime(2014, 9, 16, 10, 19, 32, micros) # No zone.\n@@ -445,7 +445,7 @@ def test_datetime_naive(self):\n def test_datetime_w_zone(self):\n import calendar\n import datetime\n- from gcloud._helpers import UTC\n+ from google.cloud._helpers import UTC\n \n micros = 4375\n utc = datetime.datetime(2014, 9, 16, 10, 19, 32, micros, UTC)\n@@ -455,7 +455,7 @@ def test_datetime_w_zone(self):\n self.assertEqual(value.nanos, 1000 * micros)\n \n def test_key(self):\n- from gcloud.datastore.key import Key\n+ from google.cloud.datastore.key import Key\n \n key = Key('PATH', 1234, project='PROJECT')\n name, value = self._callFUT(key)\n@@ -503,7 +503,7 @@ def test_unicode(self):\n self.assertEqual(value, u'str')\n \n def test_entity(self):\n- from gcloud.datastore.entity import Entity\n+ from google.cloud.datastore.entity import Entity\n entity = Entity()\n name, value = self._callFUT(entity)\n self.assertEqual(name, 'entity_value')\n@@ -517,7 +517,7 @@ def test_array(self):\n \n def test_geo_point(self):\n from google.type import latlng_pb2\n- from gcloud.datastore.helpers import GeoPoint\n+ from google.cloud.datastore.helpers import GeoPoint\n \n lat = 42.42\n lng = 99.0007\n@@ -541,12 +541,12 @@ def test_object(self):\n class Test__get_value_from_value_pb(unittest.TestCase):\n \n def _callFUT(self, pb):\n- from gcloud.datastore.helpers import _get_value_from_value_pb\n+ from google.cloud.datastore.helpers import _get_value_from_value_pb\n \n return _get_value_from_value_pb(pb)\n \n def _makePB(self, attr_name, value):\n- from gcloud.datastore._generated import entity_pb2\n+ from google.cloud.datastore._generated import entity_pb2\n \n pb = entity_pb2.Value()\n setattr(pb, attr_name, value)\n@@ -555,8 +555,8 @@ def _makePB(self, attr_name, value):\n def test_datetime(self):\n import calendar\n import datetime\n- from gcloud._helpers import UTC\n- from gcloud.datastore._generated import entity_pb2\n+ from google.cloud._helpers import UTC\n+ from google.cloud.datastore._generated import entity_pb2\n \n micros = 4375\n utc = datetime.datetime(2014, 9, 16, 10, 19, 32, micros, UTC)\n@@ -566,8 +566,8 @@ def test_datetime(self):\n self.assertEqual(self._callFUT(pb), utc)\n \n def test_key(self):\n- from gcloud.datastore._generated import entity_pb2\n- from gcloud.datastore.key import Key\n+ from google.cloud.datastore._generated import entity_pb2\n+ from google.cloud.datastore.key import Key\n \n pb = entity_pb2.Value()\n expected = Key('KIND', 1234, project='PROJECT').to_protobuf()\n@@ -596,9 +596,9 @@ def test_unicode(self):\n self.assertEqual(self._callFUT(pb), u'str')\n \n def test_entity(self):\n- from gcloud.datastore._generated import entity_pb2\n- from gcloud.datastore.entity import Entity\n- from gcloud.datastore.helpers import _new_value_pb\n+ from google.cloud.datastore._generated import entity_pb2\n+ from google.cloud.datastore.entity import Entity\n+ from google.cloud.datastore.helpers import _new_value_pb\n \n pb = entity_pb2.Value()\n entity_pb = pb.entity_value\n@@ -612,7 +612,7 @@ def test_entity(self):\n self.assertEqual(entity['foo'], 'Foo')\n \n def test_array(self):\n- from gcloud.datastore._generated import entity_pb2\n+ from google.cloud.datastore._generated import entity_pb2\n \n pb = entity_pb2.Value()\n array_pb = pb.array_value.values\n@@ -625,8 +625,8 @@ def test_array(self):\n \n def test_geo_point(self):\n from google.type import latlng_pb2\n- from gcloud.datastore._generated import entity_pb2\n- from gcloud.datastore.helpers import GeoPoint\n+ from google.cloud.datastore._generated import entity_pb2\n+ from google.cloud.datastore.helpers import GeoPoint\n \n lat = -3.14\n lng = 13.37\n@@ -639,14 +639,14 @@ def test_geo_point(self):\n \n def test_null(self):\n from google.protobuf import struct_pb2\n- from gcloud.datastore._generated import entity_pb2\n+ from google.cloud.datastore._generated import entity_pb2\n \n pb = entity_pb2.Value(null_value=struct_pb2.NULL_VALUE)\n result = self._callFUT(pb)\n self.assertIsNone(result)\n \n def test_unknown(self):\n- from gcloud.datastore._generated import entity_pb2\n+ from google.cloud.datastore._generated import entity_pb2\n \n pb = entity_pb2.Value()\n with self.assertRaises(ValueError):\n@@ -656,18 +656,18 @@ def test_unknown(self):\n class Test_set_protobuf_value(unittest.TestCase):\n \n def _callFUT(self, value_pb, val):\n- from gcloud.datastore.helpers import _set_protobuf_value\n+ from google.cloud.datastore.helpers import _set_protobuf_value\n \n return _set_protobuf_value(value_pb, val)\n \n def _makePB(self):\n- from gcloud.datastore._generated import entity_pb2\n+ from google.cloud.datastore._generated import entity_pb2\n return entity_pb2.Value()\n \n def test_datetime(self):\n import calendar\n import datetime\n- from gcloud._helpers import UTC\n+ from google.cloud._helpers import UTC\n \n pb = self._makePB()\n micros = 4375\n@@ -678,7 +678,7 @@ def test_datetime(self):\n self.assertEqual(value.nanos, 1000 * micros)\n \n def test_key(self):\n- from gcloud.datastore.key import Key\n+ from google.cloud.datastore.key import Key\n \n pb = self._makePB()\n key = Key('KIND', 1234, project='PROJECT')\n@@ -739,8 +739,8 @@ def test_unicode(self):\n self.assertEqual(value, u'str')\n \n def test_entity_empty_wo_key(self):\n- from gcloud.datastore.entity import Entity\n- from gcloud.datastore.helpers import _property_tuples\n+ from google.cloud.datastore.entity import Entity\n+ from google.cloud.datastore.helpers import _property_tuples\n \n pb = self._makePB()\n entity = Entity()\n@@ -750,9 +750,9 @@ def test_entity_empty_wo_key(self):\n self.assertEqual(len(list(_property_tuples(value))), 0)\n \n def test_entity_w_key(self):\n- from gcloud.datastore.entity import Entity\n- from gcloud.datastore.helpers import _property_tuples\n- from gcloud.datastore.key import Key\n+ from google.cloud.datastore.entity import Entity\n+ from google.cloud.datastore.helpers import _property_tuples\n+ from google.cloud.datastore.key import Key\n \n name = 'foo'\n value = u'Foo'\n@@ -781,7 +781,7 @@ def test_array(self):\n \n def test_geo_point(self):\n from google.type import latlng_pb2\n- from gcloud.datastore.helpers import GeoPoint\n+ from google.cloud.datastore.helpers import GeoPoint\n \n pb = self._makePB()\n lat = 9.11\n@@ -795,18 +795,18 @@ def test_geo_point(self):\n class Test__get_meaning(unittest.TestCase):\n \n def _callFUT(self, *args, **kwargs):\n- from gcloud.datastore.helpers import _get_meaning\n+ from google.cloud.datastore.helpers import _get_meaning\n return _get_meaning(*args, **kwargs)\n \n def test_no_meaning(self):\n- from gcloud.datastore._generated import entity_pb2\n+ from google.cloud.datastore._generated import entity_pb2\n \n value_pb = entity_pb2.Value()\n result = self._callFUT(value_pb)\n self.assertEqual(result, None)\n \n def test_single(self):\n- from gcloud.datastore._generated import entity_pb2\n+ from google.cloud.datastore._generated import entity_pb2\n \n value_pb = entity_pb2.Value()\n value_pb.meaning = meaning = 22\n@@ -815,7 +815,7 @@ def test_single(self):\n self.assertEqual(meaning, result)\n \n def test_empty_array_value(self):\n- from gcloud.datastore._generated import entity_pb2\n+ from google.cloud.datastore._generated import entity_pb2\n \n value_pb = entity_pb2.Value()\n value_pb.array_value.values.add()\n@@ -825,7 +825,7 @@ def test_empty_array_value(self):\n self.assertEqual(None, result)\n \n def test_array_value(self):\n- from gcloud.datastore._generated import entity_pb2\n+ from google.cloud.datastore._generated import entity_pb2\n \n value_pb = entity_pb2.Value()\n meaning = 9\n@@ -840,7 +840,7 @@ def test_array_value(self):\n self.assertEqual(meaning, result)\n \n def test_array_value_multiple_meanings(self):\n- from gcloud.datastore._generated import entity_pb2\n+ from google.cloud.datastore._generated import entity_pb2\n \n value_pb = entity_pb2.Value()\n meaning1 = 9\n@@ -857,7 +857,7 @@ def test_array_value_multiple_meanings(self):\n self.assertEqual(result, [meaning1, meaning2])\n \n def test_array_value_meaning_partially_unset(self):\n- from gcloud.datastore._generated import entity_pb2\n+ from google.cloud.datastore._generated import entity_pb2\n \n value_pb = entity_pb2.Value()\n meaning1 = 9\n@@ -875,7 +875,7 @@ def test_array_value_meaning_partially_unset(self):\n class TestGeoPoint(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.datastore.helpers import GeoPoint\n+ from google.cloud.datastore.helpers import GeoPoint\n return GeoPoint\n \n def _makeOne(self, *args, **kwargs):\ndiff --git a/unit_tests/datastore/test_key.py b/unit_tests/datastore/test_key.py\n--- a/unit_tests/datastore/test_key.py\n+++ b/unit_tests/datastore/test_key.py\n@@ -20,7 +20,7 @@ class TestKey(unittest.TestCase):\n _DEFAULT_PROJECT = 'PROJECT'\n \n def _getTargetClass(self):\n- from gcloud.datastore.key import Key\n+ from google.cloud.datastore.key import Key\n return Key\n \n def _makeOne(self, *args, **kwargs):\n@@ -313,7 +313,7 @@ def test_completed_key_on_complete(self):\n self.assertRaises(ValueError, key.completed_key, 5678)\n \n def test_to_protobuf_defaults(self):\n- from gcloud.datastore._generated import entity_pb2\n+ from google.cloud.datastore._generated import entity_pb2\n \n _KIND = 'KIND'\n key = self._makeOne(_KIND, project=self._DEFAULT_PROJECT)\ndiff --git a/unit_tests/datastore/test_query.py b/unit_tests/datastore/test_query.py\n--- a/unit_tests/datastore/test_query.py\n+++ b/unit_tests/datastore/test_query.py\n@@ -20,7 +20,7 @@ class TestQuery(unittest.TestCase):\n _PROJECT = 'PROJECT'\n \n def _getTargetClass(self):\n- from gcloud.datastore.query import Query\n+ from google.cloud.datastore.query import Query\n return Query\n \n def _makeOne(self, *args, **kw):\n@@ -45,7 +45,7 @@ def test_ctor_defaults(self):\n self.assertEqual(query.distinct_on, [])\n \n def test_ctor_explicit(self):\n- from gcloud.datastore.key import Key\n+ from google.cloud.datastore.key import Key\n _PROJECT = 'OTHER_PROJECT'\n _KIND = 'KIND'\n _NAMESPACE = 'OTHER_NAMESPACE'\n@@ -143,7 +143,7 @@ def _assign(val):\n self.assertRaises(TypeError, _assign, ['KIND', 'NAME'])\n \n def test_ancestor_setter_w_key(self):\n- from gcloud.datastore.key import Key\n+ from google.cloud.datastore.key import Key\n _NAME = u'NAME'\n key = Key('KIND', 123, project=self._PROJECT)\n query = self._makeOne(self._makeClient())\n@@ -152,7 +152,7 @@ def test_ancestor_setter_w_key(self):\n self.assertEqual(query.ancestor.path, key.path)\n \n def test_ancestor_deleter_w_key(self):\n- from gcloud.datastore.key import Key\n+ from google.cloud.datastore.key import Key\n key = Key('KIND', 123, project=self._PROJECT)\n query = self._makeOne(client=self._makeClient(), ancestor=key)\n del query.ancestor\n@@ -183,7 +183,7 @@ def test_add_filter_w_all_operators(self):\n self.assertEqual(query.filters[4], ('eq_prop', '=', u'val5'))\n \n def test_add_filter_w_known_operator_and_entity(self):\n- from gcloud.datastore.entity import Entity\n+ from google.cloud.datastore.entity import Entity\n query = self._makeOne(self._makeClient())\n other = Entity()\n other['firstname'] = u'John'\n@@ -198,14 +198,14 @@ def test_add_filter_w_whitespace_property_name(self):\n self.assertEqual(query.filters, [(PROPERTY_NAME, '=', u'John')])\n \n def test_add_filter___key__valid_key(self):\n- from gcloud.datastore.key import Key\n+ from google.cloud.datastore.key import Key\n query = self._makeOne(self._makeClient())\n key = Key('Foo', project=self._PROJECT)\n query.add_filter('__key__', '=', key)\n self.assertEqual(query.filters, [('__key__', '=', key)])\n \n def test_filter___key__not_equal_operator(self):\n- from gcloud.datastore.key import Key\n+ from google.cloud.datastore.key import Key\n key = Key('Foo', project=self._PROJECT)\n query = self._makeOne(self._makeClient())\n query.add_filter('__key__', '<', key)\n@@ -245,7 +245,7 @@ def test_keys_only(self):\n self.assertEqual(query.projection, ['__key__'])\n \n def test_key_filter_defaults(self):\n- from gcloud.datastore.key import Key\n+ from google.cloud.datastore.key import Key\n \n client = self._makeClient()\n query = self._makeOne(client)\n@@ -255,7 +255,7 @@ def test_key_filter_defaults(self):\n self.assertEqual(query.filters, [('__key__', '=', key)])\n \n def test_key_filter_explicit(self):\n- from gcloud.datastore.key import Key\n+ from google.cloud.datastore.key import Key\n \n client = self._makeClient()\n query = self._makeOne(client)\n@@ -339,7 +339,7 @@ class TestIterator(unittest.TestCase):\n _END = b'\\xFF'\n \n def _getTargetClass(self):\n- from gcloud.datastore.query import Iterator\n+ from google.cloud.datastore.query import Iterator\n return Iterator\n \n def _makeOne(self, *args, **kw):\n@@ -347,9 +347,9 @@ def _makeOne(self, *args, **kw):\n \n def _addQueryResults(self, connection, cursor=_END, more=False,\n skipped_results=None, no_entity=False):\n- from gcloud.datastore._generated import entity_pb2\n- from gcloud.datastore._generated import query_pb2\n- from gcloud.datastore.helpers import _new_value_pb\n+ from google.cloud.datastore._generated import entity_pb2\n+ from google.cloud.datastore._generated import query_pb2\n+ from google.cloud.datastore.helpers import _new_value_pb\n \n if more:\n more_enum = query_pb2.QueryResultBatch.NOT_FINISHED\n@@ -394,7 +394,7 @@ def test_ctor_explicit(self):\n self.assertEqual(iterator._offset, 29)\n \n def test_next_page_no_cursors_no_more(self):\n- from gcloud.datastore.query import _pb_from_query\n+ from google.cloud.datastore.query import _pb_from_query\n connection = _Connection()\n client = self._makeClient(connection)\n query = _Query(client, self._KIND, self._PROJECT, self._NAMESPACE)\n@@ -421,7 +421,7 @@ def test_next_page_no_cursors_no_more(self):\n self.assertEqual(connection._called_with, [EXPECTED])\n \n def test_next_page_no_cursors_no_more_w_offset_and_limit(self):\n- from gcloud.datastore.query import _pb_from_query\n+ from google.cloud.datastore.query import _pb_from_query\n connection = _Connection()\n client = self._makeClient(connection)\n query = _Query(client, self._KIND, self._PROJECT, self._NAMESPACE)\n@@ -453,7 +453,7 @@ def test_next_page_no_cursors_no_more_w_offset_and_limit(self):\n def test_next_page_w_cursors_w_more(self):\n from base64 import urlsafe_b64decode\n from base64 import urlsafe_b64encode\n- from gcloud.datastore.query import _pb_from_query\n+ from google.cloud.datastore.query import _pb_from_query\n connection = _Connection()\n client = self._makeClient(connection)\n query = _Query(client, self._KIND, self._PROJECT, self._NAMESPACE)\n@@ -496,7 +496,7 @@ def test_next_page_w_cursors_w_bogus_more(self):\n self.assertRaises(ValueError, iterator.next_page)\n \n def test___iter___no_more(self):\n- from gcloud.datastore.query import _pb_from_query\n+ from google.cloud.datastore.query import _pb_from_query\n connection = _Connection()\n client = self._makeClient(connection)\n query = _Query(client, self._KIND, self._PROJECT, self._NAMESPACE)\n@@ -520,7 +520,7 @@ def test___iter___no_more(self):\n self.assertEqual(connection._called_with, [EXPECTED])\n \n def test___iter___w_more(self):\n- from gcloud.datastore.query import _pb_from_query\n+ from google.cloud.datastore.query import _pb_from_query\n connection = _Connection()\n client = self._makeClient(connection)\n query = _Query(client, self._KIND, self._PROJECT, self._NAMESPACE)\n@@ -556,7 +556,7 @@ def test___iter___w_more(self):\n self.assertEqual(connection._called_with[1], EXPECTED2)\n \n def test___iter___w_limit(self):\n- from gcloud.datastore.query import _pb_from_query\n+ from google.cloud.datastore.query import _pb_from_query\n \n connection = _Connection()\n client = self._makeClient(connection)\n@@ -614,11 +614,11 @@ def test___iter___w_limit(self):\n class Test__pb_from_query(unittest.TestCase):\n \n def _callFUT(self, query):\n- from gcloud.datastore.query import _pb_from_query\n+ from google.cloud.datastore.query import _pb_from_query\n return _pb_from_query(query)\n \n def test_empty(self):\n- from gcloud.datastore._generated import query_pb2\n+ from google.cloud.datastore._generated import query_pb2\n \n pb = self._callFUT(_Query())\n self.assertEqual(list(pb.projection), [])\n@@ -645,8 +645,8 @@ def test_kind(self):\n self.assertEqual([item.name for item in pb.kind], ['KIND'])\n \n def test_ancestor(self):\n- from gcloud.datastore.key import Key\n- from gcloud.datastore._generated import query_pb2\n+ from google.cloud.datastore.key import Key\n+ from google.cloud.datastore._generated import query_pb2\n \n ancestor = Key('Ancestor', 123, project='PROJECT')\n pb = self._callFUT(_Query(ancestor=ancestor))\n@@ -659,7 +659,7 @@ def test_ancestor(self):\n self.assertEqual(pfilter.value.key_value, ancestor_pb)\n \n def test_filter(self):\n- from gcloud.datastore._generated import query_pb2\n+ from google.cloud.datastore._generated import query_pb2\n \n query = _Query(filters=[('name', '=', u'John')])\n query.OPERATORS = {\n@@ -674,8 +674,8 @@ def test_filter(self):\n self.assertEqual(pfilter.value.string_value, u'John')\n \n def test_filter_key(self):\n- from gcloud.datastore.key import Key\n- from gcloud.datastore._generated import query_pb2\n+ from google.cloud.datastore.key import Key\n+ from google.cloud.datastore._generated import query_pb2\n \n key = Key('Kind', 123, project='PROJECT')\n query = _Query(filters=[('__key__', '=', key)])\n@@ -692,7 +692,7 @@ def test_filter_key(self):\n self.assertEqual(pfilter.value.key_value, key_pb)\n \n def test_order(self):\n- from gcloud.datastore._generated import query_pb2\n+ from google.cloud.datastore._generated import query_pb2\n \n pb = self._callFUT(_Query(order=['a', '-b', 'c']))\n self.assertEqual([item.property.name for item in pb.order],\ndiff --git a/unit_tests/datastore/test_transaction.py b/unit_tests/datastore/test_transaction.py\n--- a/unit_tests/datastore/test_transaction.py\n+++ b/unit_tests/datastore/test_transaction.py\n@@ -18,14 +18,14 @@\n class TestTransaction(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.datastore.transaction import Transaction\n+ from google.cloud.datastore.transaction import Transaction\n return Transaction\n \n def _makeOne(self, client, **kw):\n return self._getTargetClass()(client, **kw)\n \n def test_ctor_defaults(self):\n- from gcloud.datastore._generated import datastore_pb2\n+ from google.cloud.datastore._generated import datastore_pb2\n \n _PROJECT = 'PROJECT'\n connection = _Connection()\n@@ -164,7 +164,7 @@ class Foo(Exception):\n \n \n def _make_key(kind, id_, project):\n- from gcloud.datastore._generated import entity_pb2\n+ from google.cloud.datastore._generated import entity_pb2\n \n key = entity_pb2.Key()\n key.partition_id.project_id = project\n@@ -199,7 +199,7 @@ class _Entity(dict):\n \n def __init__(self):\n super(_Entity, self).__init__()\n- from gcloud.datastore.key import Key\n+ from google.cloud.datastore.key import Key\n self.key = Key('KIND', project='PROJECT')\n \n \n@@ -225,7 +225,7 @@ def current_batch(self):\n class _NoCommitBatch(object):\n \n def __init__(self, client):\n- from gcloud.datastore.batch import Batch\n+ from google.cloud.datastore.batch import Batch\n self._client = client\n self._batch = Batch(client)\n \ndiff --git a/unit_tests/dns/test_changes.py b/unit_tests/dns/test_changes.py\n--- a/unit_tests/dns/test_changes.py\n+++ b/unit_tests/dns/test_changes.py\n@@ -21,19 +21,19 @@ class TestChanges(unittest.TestCase):\n CHANGES_NAME = 'changeset_id'\n \n def _getTargetClass(self):\n- from gcloud.dns.changes import Changes\n+ from google.cloud.dns.changes import Changes\n return Changes\n \n def _makeOne(self, *args, **kw):\n return self._getTargetClass()(*args, **kw)\n \n def _setUpConstants(self):\n- from gcloud._helpers import UTC\n- from gcloud._helpers import _NOW\n+ from google.cloud._helpers import UTC\n+ from google.cloud._helpers import _NOW\n self.WHEN = _NOW().replace(tzinfo=UTC)\n \n def _makeResource(self):\n- from gcloud._helpers import _datetime_to_rfc3339\n+ from google.cloud._helpers import _datetime_to_rfc3339\n when_str = _datetime_to_rfc3339(self.WHEN)\n return {\n 'kind': 'dns#change',\n@@ -55,7 +55,7 @@ def _makeResource(self):\n }\n \n def _verifyResourceProperties(self, changes, resource, zone):\n- from gcloud._helpers import _rfc3339_to_datetime\n+ from google.cloud._helpers import _rfc3339_to_datetime\n self.assertEqual(changes.name, resource['id'])\n started = _rfc3339_to_datetime(resource['startTime'])\n self.assertEqual(changes.started, started)\n@@ -133,7 +133,7 @@ def test_add_record_set_invalid_value(self):\n changes.add_record_set(object())\n \n def test_add_record_set(self):\n- from gcloud.dns.resource_record_set import ResourceRecordSet\n+ from google.cloud.dns.resource_record_set import ResourceRecordSet\n zone = _Zone()\n changes = self._makeOne(zone)\n rrs = ResourceRecordSet('test.example.com', 'CNAME', 3600,\n@@ -149,7 +149,7 @@ def test_delete_record_set_invalid_value(self):\n changes.delete_record_set(object())\n \n def test_delete_record_set(self):\n- from gcloud.dns.resource_record_set import ResourceRecordSet\n+ from google.cloud.dns.resource_record_set import ResourceRecordSet\n zone = _Zone()\n changes = self._makeOne(zone)\n rrs = ResourceRecordSet('test.example.com', 'CNAME', 3600,\n@@ -171,7 +171,7 @@ def test_create_wo_additions_or_deletions(self):\n self.assertEqual(len(conn._requested), 0)\n \n def test_create_w_bound_client(self):\n- from gcloud.dns.resource_record_set import ResourceRecordSet\n+ from google.cloud.dns.resource_record_set import ResourceRecordSet\n self._setUpConstants()\n RESOURCE = self._makeResource()\n PATH = 'projects/%s/managedZones/%s/changes' % (\n@@ -199,7 +199,7 @@ def test_create_w_bound_client(self):\n self._verifyResourceProperties(changes, RESOURCE, zone)\n \n def test_create_w_alternate_client(self):\n- from gcloud.dns.resource_record_set import ResourceRecordSet\n+ from google.cloud.dns.resource_record_set import ResourceRecordSet\n self._setUpConstants()\n RESOURCE = self._makeResource()\n PATH = 'projects/%s/managedZones/%s/changes' % (\n@@ -332,7 +332,7 @@ def __init__(self, *responses):\n self._requested = []\n \n def api_request(self, **kw):\n- from gcloud.exceptions import NotFound\n+ from google.cloud.exceptions import NotFound\n self._requested.append(kw)\n \n try:\ndiff --git a/unit_tests/dns/test_client.py b/unit_tests/dns/test_client.py\n--- a/unit_tests/dns/test_client.py\n+++ b/unit_tests/dns/test_client.py\n@@ -21,14 +21,14 @@ class TestClient(unittest.TestCase):\n ZONE_NAME = 'zone-name'\n \n def _getTargetClass(self):\n- from gcloud.dns.client import Client\n+ from google.cloud.dns.client import Client\n return Client\n \n def _makeOne(self, *args, **kw):\n return self._getTargetClass()(*args, **kw)\n \n def test_ctor(self):\n- from gcloud.dns.connection import Connection\n+ from google.cloud.dns.connection import Connection\n creds = _Credentials()\n http = object()\n client = self._makeOne(project=self.PROJECT, credentials=creds,\n@@ -106,7 +106,7 @@ def test_quotas_w_kind_key(self):\n self.assertEqual(req['path'], '/%s' % PATH)\n \n def test_list_zones_defaults(self):\n- from gcloud.dns.zone import ManagedZone\n+ from google.cloud.dns.zone import ManagedZone\n ID_1 = '123'\n ZONE_1 = 'zone_one'\n DNS_1 = 'one.example.com'\n@@ -148,7 +148,7 @@ def test_list_zones_defaults(self):\n self.assertEqual(req['path'], '/%s' % PATH)\n \n def test_list_zones_explicit(self):\n- from gcloud.dns.zone import ManagedZone\n+ from google.cloud.dns.zone import ManagedZone\n ID_1 = '123'\n ZONE_1 = 'zone_one'\n DNS_1 = 'one.example.com'\n@@ -191,7 +191,7 @@ def test_list_zones_explicit(self):\n {'maxResults': 3, 'pageToken': TOKEN})\n \n def test_zone_explicit(self):\n- from gcloud.dns.zone import ManagedZone\n+ from google.cloud.dns.zone import ManagedZone\n DESCRIPTION = 'DESCRIPTION'\n DNS_NAME = 'test.example.com'\n creds = _Credentials()\n@@ -204,7 +204,7 @@ def test_zone_explicit(self):\n self.assertTrue(zone._client is client)\n \n def test_zone_w_dns_name_wo_description(self):\n- from gcloud.dns.zone import ManagedZone\n+ from google.cloud.dns.zone import ManagedZone\n DNS_NAME = 'test.example.com'\n creds = _Credentials()\n client = self._makeOne(self.PROJECT, creds)\n@@ -216,7 +216,7 @@ def test_zone_w_dns_name_wo_description(self):\n self.assertTrue(zone._client is client)\n \n def test_zone_wo_dns_name(self):\n- from gcloud.dns.zone import ManagedZone\n+ from google.cloud.dns.zone import ManagedZone\n creds = _Credentials()\n client = self._makeOne(self.PROJECT, creds)\n zone = client.zone(self.ZONE_NAME)\ndiff --git a/unit_tests/dns/test_connection.py b/unit_tests/dns/test_connection.py\n--- a/unit_tests/dns/test_connection.py\n+++ b/unit_tests/dns/test_connection.py\n@@ -18,7 +18,7 @@\n class TestConnection(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.dns.connection import Connection\n+ from google.cloud.dns.connection import Connection\n return Connection\n \n def _makeOne(self, *args, **kw):\ndiff --git a/unit_tests/dns/test_resource_record_set.py b/unit_tests/dns/test_resource_record_set.py\n--- a/unit_tests/dns/test_resource_record_set.py\n+++ b/unit_tests/dns/test_resource_record_set.py\n@@ -18,7 +18,7 @@\n class TestResourceRecordSet(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.dns.resource_record_set import ResourceRecordSet\n+ from google.cloud.dns.resource_record_set import ResourceRecordSet\n return ResourceRecordSet\n \n def _makeOne(self, *args, **kw):\ndiff --git a/unit_tests/dns/test_zone.py b/unit_tests/dns/test_zone.py\n--- a/unit_tests/dns/test_zone.py\n+++ b/unit_tests/dns/test_zone.py\n@@ -22,7 +22,7 @@ class TestManagedZone(unittest.TestCase):\n DNS_NAME = 'test.example.com'\n \n def _getTargetClass(self):\n- from gcloud.dns.zone import ManagedZone\n+ from google.cloud.dns.zone import ManagedZone\n return ManagedZone\n \n def _makeOne(self, *args, **kw):\n@@ -30,7 +30,7 @@ def _makeOne(self, *args, **kw):\n \n def _setUpConstants(self):\n import datetime\n- from gcloud._helpers import UTC\n+ from google.cloud._helpers import UTC\n \n year = 2015\n month = 7\n@@ -184,7 +184,7 @@ def test_name_server_set_setter(self):\n self.assertEqual(zone.name_server_set, 'NAME_SERVER_SET')\n \n def test_resource_record_set(self):\n- from gcloud.dns.resource_record_set import ResourceRecordSet\n+ from google.cloud.dns.resource_record_set import ResourceRecordSet\n RRS_NAME = 'other.example.com'\n RRS_TYPE = 'CNAME'\n TTL = 3600\n@@ -200,7 +200,7 @@ def test_resource_record_set(self):\n self.assertTrue(rrs.zone is zone)\n \n def test_changes(self):\n- from gcloud.dns.changes import Changes\n+ from google.cloud.dns.changes import Changes\n client = _Client(self.PROJECT)\n zone = self._makeOne(self.ZONE_NAME, self.DNS_NAME, client)\n changes = zone.changes()\n@@ -260,7 +260,7 @@ def test_create_w_alternate_client(self):\n self._verifyResourceProperties(zone, RESOURCE)\n \n def test_create_wo_dns_name_or_description(self):\n- from gcloud.exceptions import BadRequest\n+ from google.cloud.exceptions import BadRequest\n PATH = 'projects/%s/managedZones' % self.PROJECT\n \n _requested = []\n@@ -409,7 +409,7 @@ def test_delete_w_alternate_client(self):\n self.assertEqual(req['path'], '/%s' % PATH)\n \n def test_list_resource_record_sets_defaults(self):\n- from gcloud.dns.resource_record_set import ResourceRecordSet\n+ from google.cloud.dns.resource_record_set import ResourceRecordSet\n PATH = 'projects/%s/managedZones/%s/rrsets' % (\n self.PROJECT, self.ZONE_NAME)\n TOKEN = 'TOKEN'\n@@ -457,7 +457,7 @@ def test_list_resource_record_sets_defaults(self):\n self.assertEqual(req['path'], '/%s' % PATH)\n \n def test_list_resource_record_sets_explicit(self):\n- from gcloud.dns.resource_record_set import ResourceRecordSet\n+ from google.cloud.dns.resource_record_set import ResourceRecordSet\n PATH = 'projects/%s/managedZones/%s/rrsets' % (\n self.PROJECT, self.ZONE_NAME)\n TOKEN = 'TOKEN'\n@@ -510,9 +510,9 @@ def test_list_resource_record_sets_explicit(self):\n {'maxResults': 3, 'pageToken': TOKEN})\n \n def test_list_changes_defaults(self):\n- from gcloud._helpers import _datetime_to_rfc3339\n- from gcloud.dns.changes import Changes\n- from gcloud.dns.resource_record_set import ResourceRecordSet\n+ from google.cloud._helpers import _datetime_to_rfc3339\n+ from google.cloud.dns.changes import Changes\n+ from google.cloud.dns.resource_record_set import ResourceRecordSet\n self._setUpConstants()\n PATH = 'projects/%s/managedZones/%s/changes' % (\n self.PROJECT, self.ZONE_NAME)\n@@ -586,9 +586,9 @@ def test_list_changes_defaults(self):\n self.assertEqual(req['path'], '/%s' % PATH)\n \n def test_list_changes_explicit(self):\n- from gcloud._helpers import _datetime_to_rfc3339\n- from gcloud.dns.changes import Changes\n- from gcloud.dns.resource_record_set import ResourceRecordSet\n+ from google.cloud._helpers import _datetime_to_rfc3339\n+ from google.cloud.dns.changes import Changes\n+ from google.cloud.dns.resource_record_set import ResourceRecordSet\n self._setUpConstants()\n PATH = 'projects/%s/managedZones/%s/changes' % (\n self.PROJECT, self.ZONE_NAME)\n@@ -681,7 +681,7 @@ def __init__(self, *responses):\n self._requested = []\n \n def api_request(self, **kw):\n- from gcloud.exceptions import NotFound\n+ from google.cloud.exceptions import NotFound\n self._requested.append(kw)\n \n try:\ndiff --git a/unit_tests/error_reporting/test_client.py b/unit_tests/error_reporting/test_client.py\n--- a/unit_tests/error_reporting/test_client.py\n+++ b/unit_tests/error_reporting/test_client.py\n@@ -19,11 +19,11 @@\n class TestClient(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.error_reporting.client import Client\n+ from google.cloud.error_reporting.client import Client\n return Client\n \n def _getHttpContext(self):\n- from gcloud.error_reporting.client import HTTPContext\n+ from google.cloud.error_reporting.client import HTTPContext\n return HTTPContext\n \n def _makeOne(self, *args, **kw):\ndiff --git a/unit_tests/language/test_client.py b/unit_tests/language/test_client.py\n--- a/unit_tests/language/test_client.py\n+++ b/unit_tests/language/test_client.py\n@@ -18,14 +18,14 @@\n class TestClient(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.language.client import Client\n+ from google.cloud.language.client import Client\n return Client\n \n def _makeOne(self, *args, **kw):\n return self._getTargetClass()(*args, **kw)\n \n def test_ctor(self):\n- from gcloud.language.connection import Connection\n+ from google.cloud.language.connection import Connection\n \n creds = _Credentials()\n http = object()\n@@ -35,7 +35,7 @@ def test_ctor(self):\n self.assertTrue(client.connection.http is http)\n \n def test_document_from_text_factory(self):\n- from gcloud.language.document import Document\n+ from google.cloud.language.document import Document\n \n creds = _Credentials()\n client = self._makeOne(credentials=creds, http=object())\n@@ -59,7 +59,7 @@ def test_document_from_text_factory_failure(self):\n client.document_from_text('abc', doc_type='foo')\n \n def test_document_from_html_factory(self):\n- from gcloud.language.document import Document\n+ from google.cloud.language.document import Document\n \n creds = _Credentials()\n client = self._makeOne(credentials=creds, http=object())\n@@ -83,7 +83,7 @@ def test_document_from_html_factory_failure(self):\n client.document_from_html('abc', doc_type='foo')\n \n def test_document_from_url_factory(self):\n- from gcloud.language.document import Document\n+ from google.cloud.language.document import Document\n \n creds = _Credentials()\n client = self._makeOne(credentials=creds, http=object())\n@@ -97,8 +97,8 @@ def test_document_from_url_factory(self):\n self.assertEqual(document.doc_type, Document.PLAIN_TEXT)\n \n def test_document_from_url_factory_explicit(self):\n- from gcloud.language.document import Document\n- from gcloud.language.document import Encoding\n+ from google.cloud.language.document import Document\n+ from google.cloud.language.document import Encoding\n \n creds = _Credentials()\n client = self._makeOne(credentials=creds, http=object())\ndiff --git a/unit_tests/language/test_connection.py b/unit_tests/language/test_connection.py\n--- a/unit_tests/language/test_connection.py\n+++ b/unit_tests/language/test_connection.py\n@@ -18,7 +18,7 @@\n class TestConnection(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.language.connection import Connection\n+ from google.cloud.language.connection import Connection\n return Connection\n \n def _makeOne(self, *args, **kw):\ndiff --git a/unit_tests/language/test_document.py b/unit_tests/language/test_document.py\n--- a/unit_tests/language/test_document.py\n+++ b/unit_tests/language/test_document.py\n@@ -40,7 +40,7 @@ def _make_token_json(name, part_of_speech, head, edge_label):\n \n \n def _get_token_and_sentences(include_syntax):\n- from gcloud.language.syntax import PartOfSpeech\n+ from google.cloud.language.syntax import PartOfSpeech\n \n if include_syntax:\n token_info = [\n@@ -68,7 +68,7 @@ def _get_token_and_sentences(include_syntax):\n \n \n def _get_entities(include_entities):\n- from gcloud.language.entity import EntityType\n+ from google.cloud.language.entity import EntityType\n \n if include_entities:\n entities = [\n@@ -98,14 +98,14 @@ def _get_entities(include_entities):\n class TestDocument(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.language.document import Document\n+ from google.cloud.language.document import Document\n return Document\n \n def _makeOne(self, *args, **kw):\n return self._getTargetClass()(*args, **kw)\n \n def test_constructor_defaults(self):\n- import gcloud.language.document as MUT\n+ import google.cloud.language.document as MUT\n \n client = object()\n content = 'abc'\n@@ -118,7 +118,7 @@ def test_constructor_defaults(self):\n self.assertEqual(document.encoding, MUT.Encoding.UTF8)\n \n def test_constructor_explicit(self):\n- import gcloud.language.document as MUT\n+ import google.cloud.language.document as MUT\n \n client = object()\n gcs_url = 'gs://some-bucket/some-obj.html'\n@@ -176,7 +176,7 @@ def test__to_dict_with_no_content(self):\n })\n \n def _verify_entity(self, entity, name, entity_type, wiki_url, salience):\n- from gcloud.language.entity import Entity\n+ from google.cloud.language.entity import Entity\n \n self.assertIsInstance(entity, Entity)\n self.assertEqual(entity.name, name)\n@@ -187,7 +187,7 @@ def _verify_entity(self, entity, name, entity_type, wiki_url, salience):\n self.assertEqual(entity.mentions, [name])\n \n def test_analyze_entities(self):\n- from gcloud.language.entity import EntityType\n+ from google.cloud.language.entity import EntityType\n \n name1 = 'R-O-C-K'\n name2 = 'USA'\n@@ -248,7 +248,7 @@ def test_analyze_entities(self):\n self.assertEqual(req['method'], 'POST')\n \n def _verify_sentiment(self, sentiment, polarity, magnitude):\n- from gcloud.language.sentiment import Sentiment\n+ from google.cloud.language.sentiment import Sentiment\n \n self.assertIsInstance(sentiment, Sentiment)\n self.assertEqual(sentiment.polarity, polarity)\n@@ -279,7 +279,7 @@ def test_analyze_sentiment(self):\n self.assertEqual(req['method'], 'POST')\n \n def _verify_sentences(self, include_syntax, annotations):\n- from gcloud.language.syntax import Sentence\n+ from google.cloud.language.syntax import Sentence\n \n if include_syntax:\n self.assertEqual(len(annotations.sentences), 1)\n@@ -291,7 +291,7 @@ def _verify_sentences(self, include_syntax, annotations):\n self.assertEqual(annotations.sentences, [])\n \n def _verify_tokens(self, annotations, token_info):\n- from gcloud.language.syntax import Token\n+ from google.cloud.language.syntax import Token\n \n self.assertEqual(len(annotations.tokens), len(token_info))\n for token, info in zip(annotations.tokens, token_info):\n@@ -305,8 +305,8 @@ def _verify_tokens(self, annotations, token_info):\n \n def _annotate_text_helper(self, include_sentiment,\n include_entities, include_syntax):\n- from gcloud.language.document import Annotations\n- from gcloud.language.entity import EntityType\n+ from google.cloud.language.document import Annotations\n+ from google.cloud.language.entity import EntityType\n \n token_info, sentences = _get_token_and_sentences(include_syntax)\n entities = _get_entities(include_entities)\ndiff --git a/unit_tests/language/test_entity.py b/unit_tests/language/test_entity.py\n--- a/unit_tests/language/test_entity.py\n+++ b/unit_tests/language/test_entity.py\n@@ -18,7 +18,7 @@\n class TestEntity(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.language.entity import Entity\n+ from google.cloud.language.entity import Entity\n return Entity\n \n def _makeOne(self, *args, **kw):\ndiff --git a/unit_tests/language/test_sentiment.py b/unit_tests/language/test_sentiment.py\n--- a/unit_tests/language/test_sentiment.py\n+++ b/unit_tests/language/test_sentiment.py\n@@ -18,7 +18,7 @@\n class TestSentiment(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.language.sentiment import Sentiment\n+ from google.cloud.language.sentiment import Sentiment\n return Sentiment\n \n def _makeOne(self, *args, **kw):\ndiff --git a/unit_tests/language/test_syntax.py b/unit_tests/language/test_syntax.py\n--- a/unit_tests/language/test_syntax.py\n+++ b/unit_tests/language/test_syntax.py\n@@ -18,7 +18,7 @@\n class TestPartOfSpeech(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.language.syntax import PartOfSpeech\n+ from google.cloud.language.syntax import PartOfSpeech\n return PartOfSpeech\n \n def test_reverse(self):\n@@ -36,14 +36,14 @@ def test_reverse(self):\n class TestToken(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.language.syntax import Token\n+ from google.cloud.language.syntax import Token\n return Token\n \n def _makeOne(self, *args, **kw):\n return self._getTargetClass()(*args, **kw)\n \n def test_constructor(self):\n- from gcloud.language.syntax import PartOfSpeech\n+ from google.cloud.language.syntax import PartOfSpeech\n \n text_content = 'All'\n text_begin = -1\n@@ -61,7 +61,7 @@ def test_constructor(self):\n self.assertEqual(token.lemma, lemma)\n \n def test_from_api_repr(self):\n- from gcloud.language.syntax import PartOfSpeech\n+ from google.cloud.language.syntax import PartOfSpeech\n \n klass = self._getTargetClass()\n text_content = 'pretty'\n@@ -96,7 +96,7 @@ def test_from_api_repr(self):\n class TestSentence(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.language.syntax import Sentence\n+ from google.cloud.language.syntax import Sentence\n return Sentence\n \n def _makeOne(self, *args, **kw):\ndiff --git a/unit_tests/logging/handlers/test_handlers.py b/unit_tests/logging/handlers/test_handlers.py\n--- a/unit_tests/logging/handlers/test_handlers.py\n+++ b/unit_tests/logging/handlers/test_handlers.py\n@@ -21,7 +21,7 @@ class TestCloudLoggingHandler(unittest.TestCase):\n PROJECT = 'PROJECT'\n \n def _getTargetClass(self):\n- from gcloud.logging.handlers.handlers import CloudLoggingHandler\n+ from google.cloud.logging.handlers.handlers import CloudLoggingHandler\n return CloudLoggingHandler\n \n def _makeOne(self, *args, **kw):\n@@ -46,7 +46,7 @@ def test_emit(self):\n class TestSetupLogging(unittest.TestCase):\n \n def _callFUT(self, handler, excludes=None):\n- from gcloud.logging.handlers.handlers import setup_logging\n+ from google.cloud.logging.handlers.handlers import setup_logging\n if excludes:\n return setup_logging(handler, excluded_loggers=excludes)\n else:\ndiff --git a/unit_tests/logging/handlers/transports/test_background_thread.py b/unit_tests/logging/handlers/transports/test_background_thread.py\n--- a/unit_tests/logging/handlers/transports/test_background_thread.py\n+++ b/unit_tests/logging/handlers/transports/test_background_thread.py\n@@ -22,7 +22,7 @@ class TestBackgroundThreadHandler(unittest.TestCase):\n PROJECT = 'PROJECT'\n \n def _getTargetClass(self):\n- from gcloud.logging.handlers.transports import (\n+ from google.cloud.logging.handlers.transports import (\n BackgroundThreadTransport)\n return BackgroundThreadTransport\n \n@@ -58,9 +58,8 @@ def test_send(self):\n class TestWorker(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.logging.handlers.transports.background_thread import (\n- _Worker)\n- return _Worker\n+ from google.cloud.logging.handlers.transports import background_thread\n+ return background_thread._Worker\n \n def _makeOne(self, *args, **kw):\n return self._getTargetClass()(*args, **kw)\ndiff --git a/unit_tests/logging/handlers/transports/test_base.py b/unit_tests/logging/handlers/transports/test_base.py\n--- a/unit_tests/logging/handlers/transports/test_base.py\n+++ b/unit_tests/logging/handlers/transports/test_base.py\n@@ -20,7 +20,7 @@ class TestBaseHandler(unittest.TestCase):\n PROJECT = 'PROJECT'\n \n def _getTargetClass(self):\n- from gcloud.logging.handlers.transports import Transport\n+ from google.cloud.logging.handlers.transports import Transport\n return Transport\n \n def _makeOne(self, *args, **kw):\ndiff --git a/unit_tests/logging/handlers/transports/test_sync.py b/unit_tests/logging/handlers/transports/test_sync.py\n--- a/unit_tests/logging/handlers/transports/test_sync.py\n+++ b/unit_tests/logging/handlers/transports/test_sync.py\n@@ -21,7 +21,7 @@ class TestSyncHandler(unittest.TestCase):\n PROJECT = 'PROJECT'\n \n def _getTargetClass(self):\n- from gcloud.logging.handlers.transports import SyncTransport\n+ from google.cloud.logging.handlers.transports import SyncTransport\n return SyncTransport\n \n def _makeOne(self, *args, **kw):\ndiff --git a/unit_tests/logging/test__gax.py b/unit_tests/logging/test__gax.py\n--- a/unit_tests/logging/test__gax.py\n+++ b/unit_tests/logging/test__gax.py\n@@ -17,7 +17,7 @@\n \n try:\n # pylint: disable=unused-import\n- import gcloud.logging._gax\n+ import google.cloud.logging._gax\n # pylint: enable=unused-import\n except ImportError: # pragma: NO COVER\n _HAVE_GAX = False\n@@ -41,7 +41,7 @@ class Test_LoggingAPI(_Base, unittest.TestCase):\n LOG_NAME = 'log_name'\n \n def _getTargetClass(self):\n- from gcloud.logging._gax import _LoggingAPI\n+ from google.cloud.logging._gax import _LoggingAPI\n return _LoggingAPI\n \n def test_ctor(self):\n@@ -51,7 +51,7 @@ def test_ctor(self):\n \n def test_list_entries_no_paging(self):\n from google.gax import INITIAL_PAGE\n- from gcloud.logging import DESCENDING\n+ from google.cloud.logging.client import DESCENDING\n from unit_tests._testing import _GAXPageIterator\n TOKEN = 'TOKEN'\n TEXT = 'TEXT'\n@@ -116,9 +116,9 @@ def test_list_entries_with_extra_properties(self):\n from datetime import datetime\n from google.logging.type.log_severity_pb2 import WARNING\n from unit_tests._testing import _GAXPageIterator\n- from gcloud._helpers import UTC\n- from gcloud._helpers import _datetime_to_rfc3339\n- from gcloud._helpers import _datetime_to_pb_timestamp\n+ from google.cloud._helpers import UTC\n+ from google.cloud._helpers import _datetime_to_rfc3339\n+ from google.cloud._helpers import _datetime_to_pb_timestamp\n NOW = datetime.utcnow().replace(tzinfo=UTC)\n SIZE = 23\n TOKEN = 'TOKEN'\n@@ -224,7 +224,7 @@ def test_write_entries_w_extra_properties(self):\n from datetime import datetime\n from google.logging.type.log_severity_pb2 import WARNING\n from google.logging.v2.log_entry_pb2 import LogEntry\n- from gcloud._helpers import UTC, _pb_timestamp_to_datetime\n+ from google.cloud._helpers import UTC, _pb_timestamp_to_datetime\n NOW = datetime.utcnow().replace(tzinfo=UTC)\n TEXT = 'TEXT'\n LOG_PATH = 'projects/%s/logs/%s' % (self.PROJECT, self.LOG_NAME)\n@@ -322,7 +322,7 @@ def test_write_entries_multiple(self):\n from google.logging.v2.log_entry_pb2 import LogEntry\n from google.protobuf.any_pb2 import Any\n from google.protobuf.struct_pb2 import Struct\n- from gcloud._helpers import _datetime_to_rfc3339, UTC\n+ from google.cloud._helpers import _datetime_to_rfc3339, UTC\n TEXT = 'TEXT'\n NOW = datetime.datetime.utcnow().replace(tzinfo=UTC)\n TIMESTAMP_TYPE_URL = 'type.googleapis.com/google.protobuf.Timestamp'\n@@ -409,7 +409,7 @@ def test_logger_delete(self):\n self.assertEqual(options, None)\n \n def test_logger_delete_not_found(self):\n- from gcloud.exceptions import NotFound\n+ from google.cloud.exceptions import NotFound\n LOG_PATH = 'projects/%s/logs/%s' % (self.PROJECT, self.LOG_NAME)\n gax_api = _GAXLoggingAPI(_delete_not_found=True)\n api = self._makeOne(gax_api)\n@@ -442,7 +442,7 @@ class Test_SinksAPI(_Base, unittest.TestCase):\n DESTINATION_URI = 'faux.googleapis.com/destination'\n \n def _getTargetClass(self):\n- from gcloud.logging._gax import _SinksAPI\n+ from google.cloud.logging._gax import _SinksAPI\n return _SinksAPI\n \n def test_ctor(self):\n@@ -512,7 +512,7 @@ def test_sink_create_error(self):\n self.DESTINATION_URI)\n \n def test_sink_create_conflict(self):\n- from gcloud.exceptions import Conflict\n+ from google.cloud.exceptions import Conflict\n gax_api = _GAXSinksAPI(_create_sink_conflict=True)\n api = self._makeOne(gax_api)\n \n@@ -539,7 +539,7 @@ def test_sink_create_ok(self):\n self.assertEqual(options, None)\n \n def test_sink_get_error(self):\n- from gcloud.exceptions import NotFound\n+ from google.cloud.exceptions import NotFound\n gax_api = _GAXSinksAPI()\n api = self._makeOne(gax_api)\n \n@@ -584,7 +584,7 @@ def test_sink_update_error(self):\n self.DESTINATION_URI)\n \n def test_sink_update_miss(self):\n- from gcloud.exceptions import NotFound\n+ from google.cloud.exceptions import NotFound\n gax_api = _GAXSinksAPI()\n api = self._makeOne(gax_api)\n \n@@ -621,7 +621,7 @@ def test_sink_delete_error(self):\n api.sink_delete(self.PROJECT, self.SINK_NAME)\n \n def test_sink_delete_miss(self):\n- from gcloud.exceptions import NotFound\n+ from google.cloud.exceptions import NotFound\n gax_api = _GAXSinksAPI(_sink_not_found=True)\n api = self._makeOne(gax_api)\n \n@@ -646,7 +646,7 @@ class Test_MetricsAPI(_Base, unittest.TestCase):\n DESCRIPTION = 'Description'\n \n def _getTargetClass(self):\n- from gcloud.logging._gax import _MetricsAPI\n+ from google.cloud.logging._gax import _MetricsAPI\n return _MetricsAPI\n \n def test_ctor(self):\n@@ -716,7 +716,7 @@ def test_metric_create_error(self):\n self.DESCRIPTION)\n \n def test_metric_create_conflict(self):\n- from gcloud.exceptions import Conflict\n+ from google.cloud.exceptions import Conflict\n gax_api = _GAXMetricsAPI(_create_log_metric_conflict=True)\n api = self._makeOne(gax_api)\n \n@@ -743,7 +743,7 @@ def test_metric_create_ok(self):\n self.assertEqual(options, None)\n \n def test_metric_get_error(self):\n- from gcloud.exceptions import NotFound\n+ from google.cloud.exceptions import NotFound\n gax_api = _GAXMetricsAPI()\n api = self._makeOne(gax_api)\n \n@@ -788,7 +788,7 @@ def test_metric_update_error(self):\n self.DESCRIPTION)\n \n def test_metric_update_miss(self):\n- from gcloud.exceptions import NotFound\n+ from google.cloud.exceptions import NotFound\n gax_api = _GAXMetricsAPI()\n api = self._makeOne(gax_api)\n \n@@ -825,7 +825,7 @@ def test_metric_delete_error(self):\n api.metric_delete(self.PROJECT, self.METRIC_NAME)\n \n def test_metric_delete_miss(self):\n- from gcloud.exceptions import NotFound\n+ from google.cloud.exceptions import NotFound\n gax_api = _GAXMetricsAPI(_log_metric_not_found=True)\n api = self._makeOne(gax_api)\n \n@@ -847,7 +847,7 @@ def test_metric_delete_hit(self):\n class Test_value_pb_to_value(_Base, unittest.TestCase):\n \n def _callFUT(self, value_pb):\n- from gcloud.logging._gax import _value_pb_to_value\n+ from google.cloud.logging._gax import _value_pb_to_value\n return _value_pb_to_value(value_pb)\n \n def test_w_null_values(self):\n@@ -1080,8 +1080,8 @@ def HasField(self, field_name):\n @staticmethod\n def _make_timestamp():\n from datetime import datetime\n- from gcloud._helpers import UTC\n- from gcloud._helpers import _datetime_to_pb_timestamp\n+ from google.cloud._helpers import UTC\n+ from google.cloud._helpers import _datetime_to_pb_timestamp\n NOW = datetime.utcnow().replace(tzinfo=UTC)\n return _datetime_to_pb_timestamp(NOW)\n \ndiff --git a/unit_tests/logging/test_client.py b/unit_tests/logging/test_client.py\n--- a/unit_tests/logging/test_client.py\n+++ b/unit_tests/logging/test_client.py\n@@ -27,7 +27,7 @@ class TestClient(unittest.TestCase):\n DESCRIPTION = 'DESCRIPTION'\n \n def _getTargetClass(self):\n- from gcloud.logging.client import Client\n+ from google.cloud.logging.client import Client\n return Client\n \n def _makeOne(self, *args, **kw):\n@@ -39,8 +39,8 @@ def test_ctor(self):\n self.assertEqual(client.project, self.PROJECT)\n \n def test_logging_api_wo_gax(self):\n- from gcloud.logging.connection import _LoggingAPI\n- from gcloud.logging import client as MUT\n+ from google.cloud.logging.connection import _LoggingAPI\n+ from google.cloud.logging import client as MUT\n from unit_tests._testing import _Monkey\n client = self._makeOne(self.PROJECT, credentials=_Credentials())\n conn = client.connection = object()\n@@ -55,7 +55,7 @@ def test_logging_api_wo_gax(self):\n self.assertTrue(again is api)\n \n def test_logging_api_w_gax(self):\n- from gcloud.logging import client as MUT\n+ from google.cloud.logging import client as MUT\n from unit_tests._testing import _Monkey\n \n wrapped = object()\n@@ -86,8 +86,8 @@ def __init__(self, _wrapped):\n self.assertTrue(again is api)\n \n def test_sinks_api_wo_gax(self):\n- from gcloud.logging.connection import _SinksAPI\n- from gcloud.logging import client as MUT\n+ from google.cloud.logging.connection import _SinksAPI\n+ from google.cloud.logging import client as MUT\n from unit_tests._testing import _Monkey\n client = self._makeOne(self.PROJECT, credentials=_Credentials())\n conn = client.connection = object()\n@@ -102,7 +102,7 @@ def test_sinks_api_wo_gax(self):\n self.assertTrue(again is api)\n \n def test_sinks_api_w_gax(self):\n- from gcloud.logging import client as MUT\n+ from google.cloud.logging import client as MUT\n from unit_tests._testing import _Monkey\n \n wrapped = object()\n@@ -133,8 +133,8 @@ def __init__(self, _wrapped):\n self.assertTrue(again is api)\n \n def test_metrics_api_wo_gax(self):\n- from gcloud.logging.connection import _MetricsAPI\n- from gcloud.logging import client as MUT\n+ from google.cloud.logging.connection import _MetricsAPI\n+ from google.cloud.logging import client as MUT\n from unit_tests._testing import _Monkey\n client = self._makeOne(self.PROJECT, credentials=_Credentials())\n conn = client.connection = object()\n@@ -149,7 +149,7 @@ def test_metrics_api_wo_gax(self):\n self.assertTrue(again is api)\n \n def test_metrics_api_w_gax(self):\n- from gcloud.logging import client as MUT\n+ from google.cloud.logging import client as MUT\n from unit_tests._testing import _Monkey\n \n wrapped = object()\n@@ -180,7 +180,7 @@ def __init__(self, _wrapped):\n self.assertTrue(again is api)\n \n def test_logger(self):\n- from gcloud.logging.logger import Logger\n+ from google.cloud.logging.logger import Logger\n creds = _Credentials()\n client = self._makeOne(project=self.PROJECT, credentials=creds)\n logger = client.logger(self.LOGGER_NAME)\n@@ -198,7 +198,7 @@ def test__entry_from_resource_unknown_type(self):\n client._entry_from_resource({'unknownPayload': {}}, loggers)\n \n def test_list_entries_defaults(self):\n- from gcloud.logging.entries import TextEntry\n+ from google.cloud.logging.entries import TextEntry\n IID = 'IID'\n TEXT = 'TEXT'\n TOKEN = 'TOKEN'\n@@ -234,10 +234,10 @@ def test_list_entries_defaults(self):\n ([self.PROJECT], None, None, None, None))\n \n def test_list_entries_explicit(self):\n- from gcloud.logging import DESCENDING\n- from gcloud.logging.entries import ProtobufEntry\n- from gcloud.logging.entries import StructEntry\n- from gcloud.logging.logger import Logger\n+ from google.cloud.logging.client import DESCENDING\n+ from google.cloud.logging.entries import ProtobufEntry\n+ from google.cloud.logging.entries import StructEntry\n+ from google.cloud.logging.logger import Logger\n PROJECT1 = 'PROJECT1'\n PROJECT2 = 'PROJECT2'\n FILTER = 'logName:LOGNAME'\n@@ -301,7 +301,7 @@ def test_list_entries_explicit(self):\n ([PROJECT1, PROJECT2], FILTER, DESCENDING, PAGE_SIZE, TOKEN))\n \n def test_sink_defaults(self):\n- from gcloud.logging.sink import Sink\n+ from google.cloud.logging.sink import Sink\n creds = _Credentials()\n client = self._makeOne(project=self.PROJECT, credentials=creds)\n sink = client.sink(self.SINK_NAME)\n@@ -313,7 +313,7 @@ def test_sink_defaults(self):\n self.assertEqual(sink.project, self.PROJECT)\n \n def test_sink_explicit(self):\n- from gcloud.logging.sink import Sink\n+ from google.cloud.logging.sink import Sink\n creds = _Credentials()\n client = self._makeOne(project=self.PROJECT, credentials=creds)\n sink = client.sink(self.SINK_NAME, self.FILTER, self.DESTINATION_URI)\n@@ -325,7 +325,7 @@ def test_sink_explicit(self):\n self.assertEqual(sink.project, self.PROJECT)\n \n def test_list_sinks_no_paging(self):\n- from gcloud.logging.sink import Sink\n+ from google.cloud.logging.sink import Sink\n PROJECT = 'PROJECT'\n TOKEN = 'TOKEN'\n SINK_NAME = 'sink_name'\n@@ -353,7 +353,7 @@ def test_list_sinks_no_paging(self):\n (PROJECT, None, None))\n \n def test_list_sinks_with_paging(self):\n- from gcloud.logging.sink import Sink\n+ from google.cloud.logging.sink import Sink\n PROJECT = 'PROJECT'\n SINK_NAME = 'sink_name'\n FILTER = 'logName:syslog AND severity>=ERROR'\n@@ -381,7 +381,7 @@ def test_list_sinks_with_paging(self):\n (PROJECT, PAGE_SIZE, TOKEN))\n \n def test_metric_defaults(self):\n- from gcloud.logging.metric import Metric\n+ from google.cloud.logging.metric import Metric\n creds = _Credentials()\n \n client_obj = self._makeOne(project=self.PROJECT, credentials=creds)\n@@ -394,7 +394,7 @@ def test_metric_defaults(self):\n self.assertEqual(metric.project, self.PROJECT)\n \n def test_metric_explicit(self):\n- from gcloud.logging.metric import Metric\n+ from google.cloud.logging.metric import Metric\n creds = _Credentials()\n \n client_obj = self._makeOne(project=self.PROJECT, credentials=creds)\n@@ -408,7 +408,7 @@ def test_metric_explicit(self):\n self.assertEqual(metric.project, self.PROJECT)\n \n def test_list_metrics_no_paging(self):\n- from gcloud.logging.metric import Metric\n+ from google.cloud.logging.metric import Metric\n PROJECT = 'PROJECT'\n TOKEN = 'TOKEN'\n METRICS = [{\n@@ -433,7 +433,7 @@ def test_list_metrics_no_paging(self):\n (PROJECT, None, None))\n \n def test_list_metrics_with_paging(self):\n- from gcloud.logging.metric import Metric\n+ from google.cloud.logging.metric import Metric\n PROJECT = 'PROJECT'\n TOKEN = 'TOKEN'\n PAGE_SIZE = 42\ndiff --git a/unit_tests/logging/test_connection.py b/unit_tests/logging/test_connection.py\n--- a/unit_tests/logging/test_connection.py\n+++ b/unit_tests/logging/test_connection.py\n@@ -21,7 +21,7 @@ class TestConnection(unittest.TestCase):\n FILTER = 'logName:syslog AND severity>=ERROR'\n \n def _getTargetClass(self):\n- from gcloud.logging.connection import Connection\n+ from google.cloud.logging.connection import Connection\n return Connection\n \n def _makeOne(self, *args, **kw):\n@@ -43,7 +43,7 @@ class Test_LoggingAPI(unittest.TestCase):\n FILTER = 'logName:syslog AND severity>=ERROR'\n \n def _getTargetClass(self):\n- from gcloud.logging.connection import _LoggingAPI\n+ from google.cloud.logging.connection import _LoggingAPI\n return _LoggingAPI\n \n def _makeOne(self, *args, **kw):\n@@ -57,7 +57,7 @@ def test_ctor(self):\n @staticmethod\n def _make_timestamp():\n from datetime import datetime\n- from gcloud._helpers import UTC\n+ from google.cloud._helpers import UTC\n \n NOW = datetime.utcnow().replace(tzinfo=UTC)\n return _datetime_to_rfc3339_w_nanos(NOW)\n@@ -97,7 +97,7 @@ def test_list_entries_no_paging(self):\n self.assertEqual(conn._called_with['data'], SENT)\n \n def test_list_entries_w_paging(self):\n- from gcloud.logging import DESCENDING\n+ from google.cloud.logging.client import DESCENDING\n PROJECT1 = 'PROJECT1'\n PROJECT2 = 'PROJECT2'\n TIMESTAMP = self._make_timestamp()\n@@ -227,7 +227,7 @@ class Test_SinksAPI(unittest.TestCase):\n DESTINATION_URI = 'faux.googleapis.com/destination'\n \n def _getTargetClass(self):\n- from gcloud.logging.connection import _SinksAPI\n+ from google.cloud.logging.connection import _SinksAPI\n return _SinksAPI\n \n def _makeOne(self, *args, **kw):\n@@ -287,7 +287,7 @@ def test_list_sinks_w_paging(self):\n {'pageSize': PAGE_SIZE, 'pageToken': TOKEN})\n \n def test_sink_create_conflict(self):\n- from gcloud.exceptions import Conflict\n+ from google.cloud.exceptions import Conflict\n SENT = {\n 'name': self.SINK_NAME,\n 'filter': self.FILTER,\n@@ -325,7 +325,7 @@ def test_sink_create_ok(self):\n self.assertEqual(conn._called_with['data'], SENT)\n \n def test_sink_get_miss(self):\n- from gcloud.exceptions import NotFound\n+ from google.cloud.exceptions import NotFound\n conn = _Connection()\n api = self._makeOne(conn)\n \n@@ -353,7 +353,7 @@ def test_sink_get_hit(self):\n self.assertEqual(conn._called_with['path'], path)\n \n def test_sink_update_miss(self):\n- from gcloud.exceptions import NotFound\n+ from google.cloud.exceptions import NotFound\n SENT = {\n 'name': self.SINK_NAME,\n 'filter': self.FILTER,\n@@ -390,7 +390,7 @@ def test_sink_update_hit(self):\n self.assertEqual(conn._called_with['data'], SENT)\n \n def test_sink_delete_miss(self):\n- from gcloud.exceptions import NotFound\n+ from google.cloud.exceptions import NotFound\n conn = _Connection()\n api = self._makeOne(conn)\n \n@@ -422,7 +422,7 @@ class Test_MetricsAPI(unittest.TestCase):\n DESCRIPTION = 'DESCRIPTION'\n \n def _getTargetClass(self):\n- from gcloud.logging.connection import _MetricsAPI\n+ from google.cloud.logging.connection import _MetricsAPI\n return _MetricsAPI\n \n def _makeOne(self, *args, **kw):\n@@ -474,7 +474,7 @@ def test_list_metrics_w_paging(self):\n {'pageSize': PAGE_SIZE, 'pageToken': TOKEN})\n \n def test_metric_create_conflict(self):\n- from gcloud.exceptions import Conflict\n+ from google.cloud.exceptions import Conflict\n SENT = {\n 'name': self.METRIC_NAME,\n 'filter': self.FILTER,\n@@ -512,7 +512,7 @@ def test_metric_create_ok(self):\n self.assertEqual(conn._called_with['data'], SENT)\n \n def test_metric_get_miss(self):\n- from gcloud.exceptions import NotFound\n+ from google.cloud.exceptions import NotFound\n conn = _Connection()\n api = self._makeOne(conn)\n \n@@ -540,7 +540,7 @@ def test_metric_get_hit(self):\n self.assertEqual(conn._called_with['path'], path)\n \n def test_metric_update_miss(self):\n- from gcloud.exceptions import NotFound\n+ from google.cloud.exceptions import NotFound\n SENT = {\n 'name': self.METRIC_NAME,\n 'filter': self.FILTER,\n@@ -577,7 +577,7 @@ def test_metric_update_hit(self):\n self.assertEqual(conn._called_with['data'], SENT)\n \n def test_metric_delete_miss(self):\n- from gcloud.exceptions import NotFound\n+ from google.cloud.exceptions import NotFound\n conn = _Connection()\n api = self._makeOne(conn)\n \n@@ -621,8 +621,8 @@ def __init__(self, *responses):\n self._responses = responses\n \n def api_request(self, **kw):\n- from gcloud.exceptions import Conflict\n- from gcloud.exceptions import NotFound\n+ from google.cloud.exceptions import Conflict\n+ from google.cloud.exceptions import NotFound\n self._called_with = kw\n if self._raise_conflict:\n raise Conflict('oops')\n@@ -634,6 +634,6 @@ def api_request(self, **kw):\n \n \n def _datetime_to_rfc3339_w_nanos(value):\n- from gcloud._helpers import _RFC3339_NO_FRACTION\n+ from google.cloud._helpers import _RFC3339_NO_FRACTION\n no_fraction = value.strftime(_RFC3339_NO_FRACTION)\n return '%s.%09dZ' % (no_fraction, value.microsecond * 1000)\ndiff --git a/unit_tests/logging/test_entries.py b/unit_tests/logging/test_entries.py\n--- a/unit_tests/logging/test_entries.py\n+++ b/unit_tests/logging/test_entries.py\n@@ -18,7 +18,7 @@\n class Test_logger_name_from_path(unittest.TestCase):\n \n def _callFUT(self, path):\n- from gcloud.logging.entries import logger_name_from_path\n+ from google.cloud.logging.entries import logger_name_from_path\n return logger_name_from_path(path)\n \n def test_w_simple_name(self):\n@@ -42,7 +42,7 @@ class Test_BaseEntry(unittest.TestCase):\n LOGGER_NAME = 'LOGGER_NAME'\n \n def _getTargetClass(self):\n- from gcloud.logging.entries import _BaseEntry\n+ from google.cloud.logging.entries import _BaseEntry\n \n class _Dummy(_BaseEntry):\n _PAYLOAD_KEY = 'dummyPayload'\n@@ -118,7 +118,7 @@ def test_from_api_repr_missing_data_no_loggers(self):\n \n def test_from_api_repr_w_loggers_no_logger_match(self):\n from datetime import datetime\n- from gcloud._helpers import UTC\n+ from google.cloud._helpers import UTC\n klass = self._getTargetClass()\n client = _Client(self.PROJECT)\n PAYLOAD = 'PAYLOAD'\n@@ -162,7 +162,7 @@ def test_from_api_repr_w_loggers_no_logger_match(self):\n \n def test_from_api_repr_w_loggers_w_logger_match(self):\n from datetime import datetime\n- from gcloud._helpers import UTC\n+ from google.cloud._helpers import UTC\n client = _Client(self.PROJECT)\n PAYLOAD = 'PAYLOAD'\n IID = 'IID'\n@@ -194,7 +194,7 @@ class TestProtobufEntry(unittest.TestCase):\n LOGGER_NAME = 'LOGGER_NAME'\n \n def _getTargetClass(self):\n- from gcloud.logging.entries import ProtobufEntry\n+ from google.cloud.logging.entries import ProtobufEntry\n return ProtobufEntry\n \n def _makeOne(self, *args, **kw):\n@@ -214,7 +214,7 @@ def test_parse_message(self):\n \n \n def _datetime_to_rfc3339_w_nanos(value):\n- from gcloud._helpers import _RFC3339_NO_FRACTION\n+ from google.cloud._helpers import _RFC3339_NO_FRACTION\n no_fraction = value.strftime(_RFC3339_NO_FRACTION)\n return '%s.%09dZ' % (no_fraction, value.microsecond * 1000)\n \ndiff --git a/unit_tests/logging/test_logger.py b/unit_tests/logging/test_logger.py\n--- a/unit_tests/logging/test_logger.py\n+++ b/unit_tests/logging/test_logger.py\n@@ -21,7 +21,7 @@ class TestLogger(unittest.TestCase):\n LOGGER_NAME = 'logger-name'\n \n def _getTargetClass(self):\n- from gcloud.logging.logger import Logger\n+ from google.cloud.logging.logger import Logger\n return Logger\n \n def _makeOne(self, *args, **kw):\n@@ -55,7 +55,7 @@ def test_ctor_explicit(self):\n self.assertEqual(logger.labels, LABELS)\n \n def test_batch_w_bound_client(self):\n- from gcloud.logging.logger import Batch\n+ from google.cloud.logging.logger import Batch\n conn = object()\n client = _Client(self.PROJECT, conn)\n logger = self._makeOne(self.LOGGER_NAME, client=client)\n@@ -65,7 +65,7 @@ def test_batch_w_bound_client(self):\n self.assertTrue(batch.client is client)\n \n def test_batch_w_alternate_client(self):\n- from gcloud.logging.logger import Batch\n+ from google.cloud.logging.logger import Batch\n conn1 = object()\n conn2 = object()\n client1 = _Client(self.PROJECT, conn1)\n@@ -364,7 +364,7 @@ def test_list_entries_defaults(self):\n self.assertEqual(client._listed, LISTED)\n \n def test_list_entries_explicit(self):\n- from gcloud.logging import DESCENDING\n+ from google.cloud.logging.client import DESCENDING\n PROJECT1 = 'PROJECT1'\n PROJECT2 = 'PROJECT2'\n FILTER = 'resource.type:global'\n@@ -393,7 +393,7 @@ class TestBatch(unittest.TestCase):\n PROJECT = 'test-project'\n \n def _getTargetClass(self):\n- from gcloud.logging.logger import Batch\n+ from google.cloud.logging.logger import Batch\n return Batch\n \n def _makeOne(self, *args, **kwargs):\n@@ -544,7 +544,7 @@ def test_commit_w_alternate_client(self):\n import json\n from google.protobuf.json_format import MessageToJson\n from google.protobuf.struct_pb2 import Struct, Value\n- from gcloud.logging.logger import Logger\n+ from google.cloud.logging.logger import Logger\n TEXT = 'This is the entry text'\n STRUCT = {'message': TEXT, 'weather': 'partly cloudy'}\n message = Struct(fields={'foo': Value(bool_value=True)})\n@@ -588,7 +588,7 @@ def test_context_mgr_success(self):\n import json\n from google.protobuf.json_format import MessageToJson\n from google.protobuf.struct_pb2 import Struct, Value\n- from gcloud.logging.logger import Logger\n+ from google.cloud.logging.logger import Logger\n TEXT = 'This is the entry text'\n STRUCT = {'message': TEXT, 'weather': 'partly cloudy'}\n message = Struct(fields={'foo': Value(bool_value=True)})\ndiff --git a/unit_tests/logging/test_metric.py b/unit_tests/logging/test_metric.py\n--- a/unit_tests/logging/test_metric.py\n+++ b/unit_tests/logging/test_metric.py\n@@ -23,7 +23,7 @@ class TestMetric(unittest.TestCase):\n DESCRIPTION = 'DESCRIPTION'\n \n def _getTargetClass(self):\n- from gcloud.logging.metric import Metric\n+ from google.cloud.logging.metric import Metric\n return Metric\n \n def _makeOne(self, *args, **kw):\n@@ -236,7 +236,7 @@ def metric_create(self, project, metric_name, filter_, description):\n project, metric_name, filter_, description)\n \n def metric_get(self, project, metric_name):\n- from gcloud.exceptions import NotFound\n+ from google.cloud.exceptions import NotFound\n self._metric_get_called_with = (project, metric_name)\n try:\n return self._metric_get_response\ndiff --git a/unit_tests/logging/test_sink.py b/unit_tests/logging/test_sink.py\n--- a/unit_tests/logging/test_sink.py\n+++ b/unit_tests/logging/test_sink.py\n@@ -23,7 +23,7 @@ class TestSink(unittest.TestCase):\n DESTINATION_URI = 'faux.googleapis.com/destination'\n \n def _getTargetClass(self):\n- from gcloud.logging.sink import Sink\n+ from google.cloud.logging.sink import Sink\n return Sink\n \n def _makeOne(self, *args, **kw):\n@@ -247,7 +247,7 @@ def sink_create(self, project, sink_name, filter_, destination):\n project, sink_name, filter_, destination)\n \n def sink_get(self, project, sink_name):\n- from gcloud.exceptions import NotFound\n+ from google.cloud.exceptions import NotFound\n self._sink_get_called_with = (project, sink_name)\n try:\n return self._sink_get_response\ndiff --git a/unit_tests/monitoring/test__dataframe.py b/unit_tests/monitoring/test__dataframe.py\n--- a/unit_tests/monitoring/test__dataframe.py\n+++ b/unit_tests/monitoring/test__dataframe.py\n@@ -54,16 +54,16 @@\n \n def parse_timestamps(): # pragma: NO COVER\n import datetime\n- from gcloud._helpers import _RFC3339_MICROS\n+ from google.cloud._helpers import _RFC3339_MICROS\n return [datetime.datetime.strptime(t, _RFC3339_MICROS)\n for t in TIMESTAMPS]\n \n \n def generate_query_results(): # pragma: NO COVER\n- from gcloud.monitoring.metric import Metric\n- from gcloud.monitoring.resource import Resource\n- from gcloud.monitoring.timeseries import Point\n- from gcloud.monitoring.timeseries import TimeSeries\n+ from google.cloud.monitoring.metric import Metric\n+ from google.cloud.monitoring.resource import Resource\n+ from google.cloud.monitoring.timeseries import Point\n+ from google.cloud.monitoring.timeseries import TimeSeries\n \n def P(timestamp, value):\n return Point(\n@@ -87,7 +87,7 @@ def P(timestamp, value):\n class Test__build_dataframe(unittest.TestCase): # pragma: NO COVER\n \n def _callFUT(self, *args, **kwargs):\n- from gcloud.monitoring._dataframe import _build_dataframe\n+ from google.cloud.monitoring._dataframe import _build_dataframe\n return _build_dataframe(*args, **kwargs)\n \n def test_both_label_and_labels_illegal(self):\n@@ -208,19 +208,19 @@ def test_empty_table_smart_labels(self):\n class Test__sorted_resource_labels(unittest.TestCase):\n \n def _callFUT(self, labels):\n- from gcloud.monitoring._dataframe import _sorted_resource_labels\n+ from google.cloud.monitoring._dataframe import _sorted_resource_labels\n return _sorted_resource_labels(labels)\n \n def test_empty(self):\n self.assertEqual(self._callFUT([]), [])\n \n def test_sorted(self):\n- from gcloud.monitoring._dataframe import TOP_RESOURCE_LABELS\n+ from google.cloud.monitoring._dataframe import TOP_RESOURCE_LABELS\n EXPECTED = TOP_RESOURCE_LABELS + ('other-1', 'other-2')\n self.assertSequenceEqual(self._callFUT(EXPECTED), EXPECTED)\n \n def test_reversed(self):\n- from gcloud.monitoring._dataframe import TOP_RESOURCE_LABELS\n+ from google.cloud.monitoring._dataframe import TOP_RESOURCE_LABELS\n EXPECTED = TOP_RESOURCE_LABELS + ('other-1', 'other-2')\n INPUT = list(reversed(EXPECTED))\n self.assertSequenceEqual(self._callFUT(INPUT), EXPECTED)\ndiff --git a/unit_tests/monitoring/test_client.py b/unit_tests/monitoring/test_client.py\n--- a/unit_tests/monitoring/test_client.py\n+++ b/unit_tests/monitoring/test_client.py\n@@ -20,7 +20,7 @@\n class TestClient(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.monitoring.client import Client\n+ from google.cloud.monitoring.client import Client\n return Client\n \n def _makeOne(self, *args, **kwargs):\n@@ -28,8 +28,8 @@ def _makeOne(self, *args, **kwargs):\n \n def test_query(self):\n import datetime\n- from gcloud._helpers import _datetime_to_rfc3339\n- from gcloud.exceptions import NotFound\n+ from google.cloud._helpers import _datetime_to_rfc3339\n+ from google.cloud.exceptions import NotFound\n \n START_TIME = datetime.datetime(2016, 4, 6, 22, 5, 0)\n END_TIME = datetime.datetime(2016, 4, 6, 22, 10, 0)\n@@ -185,7 +185,7 @@ def test_resource_factory(self):\n def test_timeseries_factory_gauge(self):\n import datetime\n from unit_tests._testing import _Monkey\n- import gcloud.monitoring.client\n+ import google.cloud.monitoring.client\n METRIC_TYPE = 'custom.googleapis.com/my_metric'\n METRIC_LABELS = {\n 'status': 'successful'\n@@ -218,7 +218,7 @@ def test_timeseries_factory_gauge(self):\n TIME2 = datetime.datetime.utcnow()\n # Construct a time series assuming a gauge metric using the current\n # time\n- with _Monkey(gcloud.monitoring.client, _UTCNOW=lambda: TIME2):\n+ with _Monkey(google.cloud.monitoring.client, _UTCNOW=lambda: TIME2):\n timeseries_no_end = client.time_series(metric, resource, VALUE)\n \n self.assertEqual(timeseries_no_end.points[0].end_time, TIME2)\n@@ -562,7 +562,7 @@ def __init__(self, *responses):\n self._requested = []\n \n def api_request(self, **kwargs):\n- from gcloud.exceptions import NotFound\n+ from google.cloud.exceptions import NotFound\n self._requested.append(kwargs)\n try:\n return self._responses.pop(0)\ndiff --git a/unit_tests/monitoring/test_connection.py b/unit_tests/monitoring/test_connection.py\n--- a/unit_tests/monitoring/test_connection.py\n+++ b/unit_tests/monitoring/test_connection.py\n@@ -18,7 +18,7 @@\n class TestConnection(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.monitoring.connection import Connection\n+ from google.cloud.monitoring.connection import Connection\n return Connection\n \n def _makeOne(self, *args, **kwargs):\ndiff --git a/unit_tests/monitoring/test_group.py b/unit_tests/monitoring/test_group.py\n--- a/unit_tests/monitoring/test_group.py\n+++ b/unit_tests/monitoring/test_group.py\n@@ -18,7 +18,7 @@\n class Test_group_id_from_name(unittest.TestCase):\n \n def _callFUT(self, path, project):\n- from gcloud.monitoring.group import _group_id_from_name\n+ from google.cloud.monitoring.group import _group_id_from_name\n return _group_id_from_name(path, project)\n \n def test_w_empty_name(self):\n@@ -86,7 +86,7 @@ def setUp(self):\n }\n \n def _setUpResources(self):\n- from gcloud.monitoring.resource import Resource\n+ from google.cloud.monitoring.resource import Resource\n info1 = {\n 'type': 'gce_instance',\n 'labels': {\n@@ -108,7 +108,7 @@ def _setUpResources(self):\n self.MEMBERS = [info1, info2]\n \n def _getTargetClass(self):\n- from gcloud.monitoring.group import Group\n+ from google.cloud.monitoring.group import Group\n return Group\n \n def _makeOne(self, *args, **kwargs):\n@@ -350,7 +350,7 @@ def test_list(self):\n self.assertEqual(request, expected_request)\n \n def test_list_paged(self):\n- from gcloud.exceptions import NotFound\n+ from google.cloud.exceptions import NotFound\n \n LIST_OF_GROUPS = [self.JSON_GROUP, self.JSON_PARENT]\n TOKEN = 'second-page-please'\n@@ -483,7 +483,7 @@ def test_list_members_paged(self):\n \n def test_list_members_w_all_arguments(self):\n import datetime\n- from gcloud._helpers import _datetime_to_rfc3339\n+ from google.cloud._helpers import _datetime_to_rfc3339\n \n self._setUpResources()\n \n@@ -532,7 +532,7 @@ def __init__(self, *responses):\n self._requested = []\n \n def api_request(self, **kwargs):\n- from gcloud.exceptions import NotFound\n+ from google.cloud.exceptions import NotFound\n self._requested.append(kwargs)\n try:\n return self._responses.pop(0)\ndiff --git a/unit_tests/monitoring/test_label.py b/unit_tests/monitoring/test_label.py\n--- a/unit_tests/monitoring/test_label.py\n+++ b/unit_tests/monitoring/test_label.py\n@@ -18,7 +18,7 @@\n class TestLabelValueType(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.monitoring.label import LabelValueType\n+ from google.cloud.monitoring.label import LabelValueType\n return LabelValueType\n \n def test_one(self):\n@@ -33,7 +33,7 @@ def test_names(self):\n class TestLabelDescriptor(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.monitoring.label import LabelDescriptor\n+ from google.cloud.monitoring.label import LabelDescriptor\n return LabelDescriptor\n \n def _makeOne(self, *args, **kwargs):\ndiff --git a/unit_tests/monitoring/test_metric.py b/unit_tests/monitoring/test_metric.py\n--- a/unit_tests/monitoring/test_metric.py\n+++ b/unit_tests/monitoring/test_metric.py\n@@ -18,7 +18,7 @@\n class TestMetricKind(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.monitoring.metric import MetricKind\n+ from google.cloud.monitoring.metric import MetricKind\n return MetricKind\n \n def test_one(self):\n@@ -33,7 +33,7 @@ def test_names(self):\n class TestValueType(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.monitoring.metric import ValueType\n+ from google.cloud.monitoring.metric import ValueType\n return ValueType\n \n def test_one(self):\n@@ -48,14 +48,14 @@ def test_names(self):\n class TestMetricDescriptor(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.monitoring.metric import MetricDescriptor\n+ from google.cloud.monitoring.metric import MetricDescriptor\n return MetricDescriptor\n \n def _makeOne(self, *args, **kwargs):\n return self._getTargetClass()(*args, **kwargs)\n \n def test_constructor(self):\n- from gcloud.monitoring.label import LabelDescriptor\n+ from google.cloud.monitoring.label import LabelDescriptor\n \n TYPE = 'appengine.googleapis.com/http/server/response_count'\n NAME = 'projects/my-project/metricDescriptors/' + TYPE\n@@ -384,7 +384,7 @@ def test_list(self):\n self.assertEqual(request, expected_request)\n \n def test_list_paged(self):\n- from gcloud.exceptions import NotFound\n+ from google.cloud.exceptions import NotFound\n \n PROJECT = 'my-project'\n PATH = 'projects/{project}/metricDescriptors/'.format(project=PROJECT)\n@@ -493,7 +493,7 @@ def test_list_filtered_by_type_prefix(self):\n class TestMetric(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.monitoring.metric import Metric\n+ from google.cloud.monitoring.metric import Metric\n return Metric\n \n def _makeOne(self, *args, **kwargs):\n@@ -538,7 +538,7 @@ def __init__(self, *responses):\n self._requested = []\n \n def api_request(self, **kwargs):\n- from gcloud.exceptions import NotFound\n+ from google.cloud.exceptions import NotFound\n self._requested.append(kwargs)\n try:\n return self._responses.pop(0)\ndiff --git a/unit_tests/monitoring/test_query.py b/unit_tests/monitoring/test_query.py\n--- a/unit_tests/monitoring/test_query.py\n+++ b/unit_tests/monitoring/test_query.py\n@@ -43,7 +43,7 @@\n class TestAligner(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.monitoring.query import Aligner\n+ from google.cloud.monitoring.query import Aligner\n return Aligner\n \n def test_one(self):\n@@ -58,7 +58,7 @@ def test_names(self):\n class TestReducer(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.monitoring.query import Reducer\n+ from google.cloud.monitoring.query import Reducer\n return Reducer\n \n def test_one(self):\n@@ -74,7 +74,7 @@ def test_names(self):\n class TestQuery(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.monitoring.query import Query\n+ from google.cloud.monitoring.query import Query\n return Query\n \n def _makeOne(self, *args, **kwargs):\n@@ -82,7 +82,7 @@ def _makeOne(self, *args, **kwargs):\n \n @staticmethod\n def _make_timestamp(value):\n- from gcloud._helpers import _datetime_to_rfc3339\n+ from google.cloud._helpers import _datetime_to_rfc3339\n return _datetime_to_rfc3339(value)\n \n def test_constructor_minimal(self):\n@@ -127,7 +127,7 @@ def test_constructor_maximal(self):\n def test_constructor_default_end_time(self):\n import datetime\n from unit_tests._testing import _Monkey\n- from gcloud.monitoring import query as MUT\n+ from google.cloud.monitoring import query as MUT\n \n MINUTES = 5\n NOW, T0, T1 = [\n@@ -334,7 +334,7 @@ def test_iteration(self):\n def test_iteration_paged(self):\n import copy\n import datetime\n- from gcloud.exceptions import NotFound\n+ from google.cloud.exceptions import NotFound\n \n T0 = datetime.datetime(2016, 4, 6, 22, 5, 0)\n T1 = datetime.datetime(2016, 4, 6, 22, 10, 0)\n@@ -500,7 +500,7 @@ def test_iteration_headers_only(self):\n class Test_Filter(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.monitoring.query import _Filter\n+ from google.cloud.monitoring.query import _Filter\n return _Filter\n \n def _makeOne(self, metric_type):\n@@ -534,7 +534,7 @@ def test_maximal(self):\n class Test__build_label_filter(unittest.TestCase):\n \n def _callFUT(self, *args, **kwargs):\n- from gcloud.monitoring.query import _build_label_filter\n+ from google.cloud.monitoring.query import _build_label_filter\n return _build_label_filter(*args, **kwargs)\n \n def test_no_labels(self):\n@@ -630,7 +630,7 @@ def __init__(self, *responses):\n self._requested = []\n \n def api_request(self, **kwargs):\n- from gcloud.exceptions import NotFound\n+ from google.cloud.exceptions import NotFound\n self._requested.append(kwargs)\n try:\n return self._responses.pop(0)\ndiff --git a/unit_tests/monitoring/test_resource.py b/unit_tests/monitoring/test_resource.py\n--- a/unit_tests/monitoring/test_resource.py\n+++ b/unit_tests/monitoring/test_resource.py\n@@ -18,14 +18,14 @@\n class TestResourceDescriptor(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.monitoring.resource import ResourceDescriptor\n+ from google.cloud.monitoring.resource import ResourceDescriptor\n return ResourceDescriptor\n \n def _makeOne(self, *args, **kwargs):\n return self._getTargetClass()(*args, **kwargs)\n \n def test_constructor(self):\n- from gcloud.monitoring.label import LabelDescriptor\n+ from google.cloud.monitoring.label import LabelDescriptor\n \n TYPE = 'gce_instance'\n NAME = 'projects/my-project/monitoredResourceDescriptors/' + TYPE\n@@ -192,7 +192,7 @@ def test_list(self):\n self.assertEqual(request, expected_request)\n \n def test_list_paged(self):\n- from gcloud.exceptions import NotFound\n+ from google.cloud.exceptions import NotFound\n \n PROJECT = 'my-project'\n PATH = 'projects/{project}/monitoredResourceDescriptors/'.format(\n@@ -277,7 +277,7 @@ def test_list_filtered(self):\n class TestResource(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.monitoring.resource import Resource\n+ from google.cloud.monitoring.resource import Resource\n return Resource\n \n def _makeOne(self, *args, **kwargs):\n@@ -324,7 +324,7 @@ def __init__(self, *responses):\n self._requested = []\n \n def api_request(self, **kwargs):\n- from gcloud.exceptions import NotFound\n+ from google.cloud.exceptions import NotFound\n self._requested.append(kwargs)\n try:\n return self._responses.pop(0)\ndiff --git a/unit_tests/monitoring/test_timeseries.py b/unit_tests/monitoring/test_timeseries.py\n--- a/unit_tests/monitoring/test_timeseries.py\n+++ b/unit_tests/monitoring/test_timeseries.py\n@@ -35,16 +35,16 @@\n class TestTimeSeries(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.monitoring.timeseries import TimeSeries\n+ from google.cloud.monitoring.timeseries import TimeSeries\n return TimeSeries\n \n def _makeOne(self, *args, **kwargs):\n return self._getTargetClass()(*args, **kwargs)\n \n def test_constructor(self):\n- from gcloud.monitoring.metric import Metric\n- from gcloud.monitoring.resource import Resource\n- from gcloud.monitoring.timeseries import Point\n+ from google.cloud.monitoring.metric import Metric\n+ from google.cloud.monitoring.resource import Resource\n+ from google.cloud.monitoring.timeseries import Point\n \n VALUE = 60 # seconds\n \n@@ -151,7 +151,7 @@ def test_labels(self):\n class TestPoint(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.monitoring.timeseries import Point\n+ from google.cloud.monitoring.timeseries import Point\n return Point\n \n def _makeOne(self, *args, **kwargs):\ndiff --git a/unit_tests/pubsub/test__gax.py b/unit_tests/pubsub/test__gax.py\n--- a/unit_tests/pubsub/test__gax.py\n+++ b/unit_tests/pubsub/test__gax.py\n@@ -17,7 +17,7 @@\n \n try:\n # pylint: disable=unused-import\n- import gcloud.pubsub._gax\n+ import google.cloud.pubsub._gax\n # pylint: enable=unused-import\n except ImportError: # pragma: NO COVER\n _HAVE_GAX = False\n@@ -45,7 +45,7 @@ def _makeOne(self, *args, **kw):\n class Test_PublisherAPI(_Base, unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.pubsub._gax import _PublisherAPI\n+ from google.cloud.pubsub._gax import _PublisherAPI\n return _PublisherAPI\n \n def test_ctor(self):\n@@ -111,7 +111,7 @@ def test_topic_create(self):\n self.assertEqual(options, None)\n \n def test_topic_create_already_exists(self):\n- from gcloud.exceptions import Conflict\n+ from google.cloud.exceptions import Conflict\n gax_api = _GAXPublisherAPI(_create_topic_conflict=True)\n api = self._makeOne(gax_api)\n \n@@ -147,7 +147,7 @@ def test_topic_get_hit(self):\n self.assertEqual(options, None)\n \n def test_topic_get_miss(self):\n- from gcloud.exceptions import NotFound\n+ from google.cloud.exceptions import NotFound\n gax_api = _GAXPublisherAPI()\n api = self._makeOne(gax_api)\n \n@@ -181,7 +181,7 @@ def test_topic_delete_hit(self):\n self.assertEqual(options, None)\n \n def test_topic_delete_miss(self):\n- from gcloud.exceptions import NotFound\n+ from google.cloud.exceptions import NotFound\n gax_api = _GAXPublisherAPI(_delete_topic_ok=False)\n api = self._makeOne(gax_api)\n \n@@ -226,7 +226,7 @@ def test_topic_publish_hit(self):\n \n def test_topic_publish_miss_w_attrs_w_bytes_payload(self):\n import base64\n- from gcloud.exceptions import NotFound\n+ from google.cloud.exceptions import NotFound\n PAYLOAD = b'This is the message text'\n B64 = base64.b64encode(PAYLOAD)\n MESSAGE = {'data': B64, 'attributes': {'foo': 'bar'}}\n@@ -314,7 +314,7 @@ def test_topic_list_subscriptions_with_paging(self):\n \n def test_topic_list_subscriptions_miss(self):\n from google.gax import INITIAL_PAGE\n- from gcloud.exceptions import NotFound\n+ from google.cloud.exceptions import NotFound\n gax_api = _GAXPublisherAPI()\n api = self._makeOne(gax_api)\n \n@@ -349,7 +349,7 @@ class Test_SubscriberAPI(_Base, unittest.TestCase):\n PUSH_ENDPOINT = 'https://api.example.com/push'\n \n def _getTargetClass(self):\n- from gcloud.pubsub._gax import _SubscriberAPI\n+ from google.cloud.pubsub._gax import _SubscriberAPI\n return _SubscriberAPI\n \n def test_ctor(self):\n@@ -432,7 +432,7 @@ def test_subscription_create(self):\n self.assertEqual(options, None)\n \n def test_subscription_create_already_exists(self):\n- from gcloud.exceptions import Conflict\n+ from google.cloud.exceptions import Conflict\n DEADLINE = 600\n gax_api = _GAXSubscriberAPI(_create_subscription_conflict=True)\n api = self._makeOne(gax_api)\n@@ -487,7 +487,7 @@ def test_subscription_get_hit(self):\n self.assertEqual(options, None)\n \n def test_subscription_get_miss(self):\n- from gcloud.exceptions import NotFound\n+ from google.cloud.exceptions import NotFound\n gax_api = _GAXSubscriberAPI()\n api = self._makeOne(gax_api)\n \n@@ -521,7 +521,7 @@ def test_subscription_delete_hit(self):\n self.assertEqual(options, None)\n \n def test_subscription_delete_miss(self):\n- from gcloud.exceptions import NotFound\n+ from google.cloud.exceptions import NotFound\n gax_api = _GAXSubscriberAPI(_delete_subscription_ok=False)\n api = self._makeOne(gax_api)\n \n@@ -556,7 +556,7 @@ def test_subscription_modify_push_config_hit(self):\n self.assertEqual(options, None)\n \n def test_subscription_modify_push_config_miss(self):\n- from gcloud.exceptions import NotFound\n+ from google.cloud.exceptions import NotFound\n gax_api = _GAXSubscriberAPI()\n api = self._makeOne(gax_api)\n \n@@ -609,7 +609,7 @@ def test_subscription_pull_explicit(self):\n self.assertEqual(options, None)\n \n def test_subscription_pull_defaults_miss(self):\n- from gcloud.exceptions import NotFound\n+ from google.cloud.exceptions import NotFound\n gax_api = _GAXSubscriberAPI()\n api = self._makeOne(gax_api)\n \n@@ -652,7 +652,7 @@ def test_subscription_acknowledge_hit(self):\n self.assertEqual(options, None)\n \n def test_subscription_acknowledge_miss(self):\n- from gcloud.exceptions import NotFound\n+ from google.cloud.exceptions import NotFound\n ACK_ID1 = 'DEADBEEF'\n ACK_ID2 = 'BEADCAFE'\n gax_api = _GAXSubscriberAPI()\n@@ -699,7 +699,7 @@ def test_subscription_modify_ack_deadline_hit(self):\n self.assertEqual(options, None)\n \n def test_subscription_modify_ack_deadline_miss(self):\n- from gcloud.exceptions import NotFound\n+ from google.cloud.exceptions import NotFound\n ACK_ID1 = 'DEADBEEF'\n ACK_ID2 = 'BEADCAFE'\n NEW_DEADLINE = 90\ndiff --git a/unit_tests/pubsub/test__helpers.py b/unit_tests/pubsub/test__helpers.py\n--- a/unit_tests/pubsub/test__helpers.py\n+++ b/unit_tests/pubsub/test__helpers.py\n@@ -18,7 +18,7 @@\n class Test_topic_name_from_path(unittest.TestCase):\n \n def _callFUT(self, path, project):\n- from gcloud.pubsub._helpers import topic_name_from_path\n+ from google.cloud.pubsub._helpers import topic_name_from_path\n return topic_name_from_path(path, project)\n \n def test_w_simple_name(self):\n@@ -39,7 +39,7 @@ def test_w_name_w_all_extras(self):\n class Test_subscription_name_from_path(unittest.TestCase):\n \n def _callFUT(self, path, project):\n- from gcloud.pubsub._helpers import subscription_name_from_path\n+ from google.cloud.pubsub._helpers import subscription_name_from_path\n return subscription_name_from_path(path, project)\n \n def test_w_simple_name(self):\ndiff --git a/unit_tests/pubsub/test_client.py b/unit_tests/pubsub/test_client.py\n--- a/unit_tests/pubsub/test_client.py\n+++ b/unit_tests/pubsub/test_client.py\n@@ -23,15 +23,15 @@ class TestClient(unittest.TestCase):\n SUB_PATH = 'projects/%s/subscriptions/%s' % (PROJECT, SUB_NAME)\n \n def _getTargetClass(self):\n- from gcloud.pubsub.client import Client\n+ from google.cloud.pubsub.client import Client\n return Client\n \n def _makeOne(self, *args, **kw):\n return self._getTargetClass()(*args, **kw)\n \n def test_publisher_api_wo_gax(self):\n- from gcloud.pubsub.connection import _PublisherAPI\n- from gcloud.pubsub import client as MUT\n+ from google.cloud.pubsub.connection import _PublisherAPI\n+ from google.cloud.pubsub import client as MUT\n from unit_tests._testing import _Monkey\n creds = _Credentials()\n client = self._makeOne(project=self.PROJECT, credentials=creds)\n@@ -47,7 +47,7 @@ def test_publisher_api_wo_gax(self):\n self.assertTrue(again is api)\n \n def test_publisher_api_w_gax(self):\n- from gcloud.pubsub import client as MUT\n+ from google.cloud.pubsub import client as MUT\n from unit_tests._testing import _Monkey\n \n wrapped = object()\n@@ -78,8 +78,8 @@ def __init__(self, _wrapped):\n self.assertTrue(again is api)\n \n def test_subscriber_api_wo_gax(self):\n- from gcloud.pubsub.connection import _SubscriberAPI\n- from gcloud.pubsub import client as MUT\n+ from google.cloud.pubsub.connection import _SubscriberAPI\n+ from google.cloud.pubsub import client as MUT\n from unit_tests._testing import _Monkey\n creds = _Credentials()\n client = self._makeOne(project=self.PROJECT, credentials=creds)\n@@ -95,7 +95,7 @@ def test_subscriber_api_wo_gax(self):\n self.assertTrue(again is api)\n \n def test_subscriber_api_w_gax(self):\n- from gcloud.pubsub import client as MUT\n+ from google.cloud.pubsub import client as MUT\n from unit_tests._testing import _Monkey\n \n wrapped = object()\n@@ -126,7 +126,7 @@ def __init__(self, _wrapped):\n self.assertTrue(again is api)\n \n def test_iam_policy_api(self):\n- from gcloud.pubsub.connection import _IAMPolicyAPI\n+ from google.cloud.pubsub.connection import _IAMPolicyAPI\n creds = _Credentials()\n client = self._makeOne(project=self.PROJECT, credentials=creds)\n conn = client.connection = object()\n@@ -138,7 +138,7 @@ def test_iam_policy_api(self):\n self.assertTrue(again is api)\n \n def test_list_topics_no_paging(self):\n- from gcloud.pubsub.topic import Topic\n+ from google.cloud.pubsub.topic import Topic\n creds = _Credentials()\n client = self._makeOne(project=self.PROJECT, credentials=creds)\n client.connection = object()\n@@ -155,7 +155,7 @@ def test_list_topics_no_paging(self):\n self.assertEqual(api._listed_topics, (self.PROJECT, None, None))\n \n def test_list_topics_with_paging(self):\n- from gcloud.pubsub.topic import Topic\n+ from google.cloud.pubsub.topic import Topic\n TOKEN1 = 'TOKEN1'\n TOKEN2 = 'TOKEN2'\n SIZE = 1\n@@ -189,7 +189,7 @@ def test_list_topics_missing_key(self):\n self.assertEqual(api._listed_topics, (self.PROJECT, None, None))\n \n def test_list_subscriptions_no_paging(self):\n- from gcloud.pubsub.subscription import Subscription\n+ from google.cloud.pubsub.subscription import Subscription\n SUB_INFO = {'name': self.SUB_PATH, 'topic': self.TOPIC_PATH}\n creds = _Credentials()\n client = self._makeOne(project=self.PROJECT, credentials=creds)\n@@ -209,7 +209,7 @@ def test_list_subscriptions_no_paging(self):\n (self.PROJECT, None, None))\n \n def test_list_subscriptions_with_paging(self):\n- from gcloud.pubsub.subscription import Subscription\n+ from google.cloud.pubsub.subscription import Subscription\n SUB_INFO = {'name': self.SUB_PATH, 'topic': self.TOPIC_PATH}\n creds = _Credentials()\n client = self._makeOne(project=self.PROJECT, credentials=creds)\ndiff --git a/unit_tests/pubsub/test_connection.py b/unit_tests/pubsub/test_connection.py\n--- a/unit_tests/pubsub/test_connection.py\n+++ b/unit_tests/pubsub/test_connection.py\n@@ -32,7 +32,7 @@ def _makeOne(self, *args, **kw):\n class TestConnection(_Base):\n \n def _getTargetClass(self):\n- from gcloud.pubsub.connection import Connection\n+ from google.cloud.pubsub.connection import Connection\n return Connection\n \n def test_default_url(self):\n@@ -43,7 +43,7 @@ def test_default_url(self):\n def test_custom_url_from_env(self):\n import os\n from unit_tests._testing import _Monkey\n- from gcloud.environment_vars import PUBSUB_EMULATOR\n+ from google.cloud.environment_vars import PUBSUB_EMULATOR\n \n HOST = 'localhost:8187'\n fake_environ = {PUBSUB_EMULATOR: HOST}\n@@ -66,7 +66,7 @@ def test_custom_url_from_constructor(self):\n def test_custom_url_constructor_and_env(self):\n import os\n from unit_tests._testing import _Monkey\n- from gcloud.environment_vars import PUBSUB_EMULATOR\n+ from google.cloud.environment_vars import PUBSUB_EMULATOR\n \n HOST1 = object()\n HOST2 = object()\n@@ -117,7 +117,7 @@ def test_build_api_url_w_base_url_override(self):\n class Test_PublisherAPI(_Base):\n \n def _getTargetClass(self):\n- from gcloud.pubsub.connection import _PublisherAPI\n+ from google.cloud.pubsub.connection import _PublisherAPI\n return _PublisherAPI\n \n def _makeOne(self, *args, **kw):\n@@ -200,7 +200,7 @@ def test_topic_create(self):\n self.assertEqual(connection._called_with['path'], path)\n \n def test_topic_create_already_exists(self):\n- from gcloud.exceptions import Conflict\n+ from google.cloud.exceptions import Conflict\n connection = _Connection()\n connection._no_response_error = Conflict\n api = self._makeOne(connection)\n@@ -225,7 +225,7 @@ def test_topic_get_hit(self):\n self.assertEqual(connection._called_with['path'], path)\n \n def test_topic_get_miss(self):\n- from gcloud.exceptions import NotFound\n+ from google.cloud.exceptions import NotFound\n connection = _Connection()\n api = self._makeOne(connection)\n \n@@ -248,7 +248,7 @@ def test_topic_delete_hit(self):\n self.assertEqual(connection._called_with['path'], path)\n \n def test_topic_delete_miss(self):\n- from gcloud.exceptions import NotFound\n+ from google.cloud.exceptions import NotFound\n connection = _Connection()\n api = self._makeOne(connection)\n \n@@ -280,7 +280,7 @@ def test_topic_publish_hit(self):\n \n def test_topic_publish_miss(self):\n import base64\n- from gcloud.exceptions import NotFound\n+ from google.cloud.exceptions import NotFound\n PAYLOAD = b'This is the message text'\n B64 = base64.b64encode(PAYLOAD).decode('ascii')\n MESSAGE = {'data': B64, 'attributes': {}}\n@@ -362,7 +362,7 @@ def test_topic_list_subscriptions_missing_key(self):\n self.assertEqual(connection._called_with['query_params'], {})\n \n def test_topic_list_subscriptions_miss(self):\n- from gcloud.exceptions import NotFound\n+ from google.cloud.exceptions import NotFound\n connection = _Connection()\n api = self._makeOne(connection)\n \n@@ -378,7 +378,7 @@ def test_topic_list_subscriptions_miss(self):\n class Test_SubscriberAPI(_Base):\n \n def _getTargetClass(self):\n- from gcloud.pubsub.connection import _SubscriberAPI\n+ from google.cloud.pubsub.connection import _SubscriberAPI\n return _SubscriberAPI\n \n def _makeOne(self, *args, **kw):\n@@ -631,7 +631,7 @@ def test_subscription_modify_ack_deadline(self):\n class Test_IAMPolicyAPI(_Base):\n \n def _getTargetClass(self):\n- from gcloud.pubsub.connection import _IAMPolicyAPI\n+ from google.cloud.pubsub.connection import _IAMPolicyAPI\n return _IAMPolicyAPI\n \n def test_ctor(self):\n@@ -640,7 +640,10 @@ def test_ctor(self):\n self.assertTrue(api._connection is connection)\n \n def test_get_iam_policy(self):\n- from gcloud.pubsub.iam import OWNER_ROLE, EDITOR_ROLE, VIEWER_ROLE\n+ from google.cloud.pubsub.iam import OWNER_ROLE\n+ from google.cloud.pubsub.iam import EDITOR_ROLE\n+ from google.cloud.pubsub.iam import VIEWER_ROLE\n+\n OWNER1 = 'user:phred@example.com'\n OWNER2 = 'group:cloud-logs@google.com'\n EDITOR1 = 'domain:google.com'\n@@ -667,7 +670,10 @@ def test_get_iam_policy(self):\n self.assertEqual(connection._called_with['path'], path)\n \n def test_set_iam_policy(self):\n- from gcloud.pubsub.iam import OWNER_ROLE, EDITOR_ROLE, VIEWER_ROLE\n+ from google.cloud.pubsub.iam import OWNER_ROLE\n+ from google.cloud.pubsub.iam import EDITOR_ROLE\n+ from google.cloud.pubsub.iam import VIEWER_ROLE\n+\n OWNER1 = 'user:phred@example.com'\n OWNER2 = 'group:cloud-logs@google.com'\n EDITOR1 = 'domain:google.com'\n@@ -697,7 +703,10 @@ def test_set_iam_policy(self):\n {'policy': POLICY})\n \n def test_test_iam_permissions(self):\n- from gcloud.pubsub.iam import OWNER_ROLE, EDITOR_ROLE, VIEWER_ROLE\n+ from google.cloud.pubsub.iam import OWNER_ROLE\n+ from google.cloud.pubsub.iam import EDITOR_ROLE\n+ from google.cloud.pubsub.iam import VIEWER_ROLE\n+\n ALL_ROLES = [OWNER_ROLE, EDITOR_ROLE, VIEWER_ROLE]\n ALLOWED = ALL_ROLES[1:]\n RETURNED = {'permissions': ALLOWED}\n@@ -714,7 +723,10 @@ def test_test_iam_permissions(self):\n {'permissions': ALL_ROLES})\n \n def test_test_iam_permissions_missing_key(self):\n- from gcloud.pubsub.iam import OWNER_ROLE, EDITOR_ROLE, VIEWER_ROLE\n+ from google.cloud.pubsub.iam import OWNER_ROLE\n+ from google.cloud.pubsub.iam import EDITOR_ROLE\n+ from google.cloud.pubsub.iam import VIEWER_ROLE\n+\n ALL_ROLES = [OWNER_ROLE, EDITOR_ROLE, VIEWER_ROLE]\n RETURNED = {}\n connection = _Connection(RETURNED)\n@@ -739,7 +751,7 @@ def __init__(self, *responses):\n self._responses = responses\n \n def api_request(self, **kw):\n- from gcloud.exceptions import NotFound\n+ from google.cloud.exceptions import NotFound\n self._called_with = kw\n try:\n response, self._responses = self._responses[0], self._responses[1:]\ndiff --git a/unit_tests/pubsub/test_iam.py b/unit_tests/pubsub/test_iam.py\n--- a/unit_tests/pubsub/test_iam.py\n+++ b/unit_tests/pubsub/test_iam.py\n@@ -18,7 +18,7 @@\n class TestPolicy(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.pubsub.iam import Policy\n+ from google.cloud.pubsub.iam import Policy\n return Policy\n \n def _makeOne(self, *args, **kw):\n@@ -91,7 +91,7 @@ def test_from_api_repr_only_etag(self):\n self.assertEqual(list(policy.viewers), [])\n \n def test_from_api_repr_complete(self):\n- from gcloud.pubsub.iam import (\n+ from google.cloud.pubsub.iam import (\n OWNER_ROLE,\n EDITOR_ROLE,\n VIEWER_ROLE,\n@@ -150,7 +150,7 @@ def test_to_api_repr_only_etag(self):\n self.assertEqual(policy.to_api_repr(), {'etag': 'DEADBEEF'})\n \n def test_to_api_repr_full(self):\n- from gcloud.pubsub.iam import (\n+ from google.cloud.pubsub.iam import (\n PUBSUB_ADMIN_ROLE,\n PUBSUB_EDITOR_ROLE,\n PUBSUB_VIEWER_ROLE,\ndiff --git a/unit_tests/pubsub/test_message.py b/unit_tests/pubsub/test_message.py\n--- a/unit_tests/pubsub/test_message.py\n+++ b/unit_tests/pubsub/test_message.py\n@@ -18,7 +18,7 @@\n class TestMessage(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.pubsub.message import Message\n+ from google.cloud.pubsub.message import Message\n return Message\n \n def _makeOne(self, *args, **kw):\n@@ -68,8 +68,8 @@ def _to_fail():\n \n def test_timestamp_w_timestamp_in_attributes(self):\n from datetime import datetime\n- from gcloud._helpers import _RFC3339_MICROS\n- from gcloud._helpers import UTC\n+ from google.cloud._helpers import _RFC3339_MICROS\n+ from google.cloud._helpers import UTC\n DATA = b'DEADBEEF'\n MESSAGE_ID = b'12345'\n TIMESTAMP = '2015-04-10T18:42:27.131956Z'\ndiff --git a/unit_tests/pubsub/test_subscription.py b/unit_tests/pubsub/test_subscription.py\n--- a/unit_tests/pubsub/test_subscription.py\n+++ b/unit_tests/pubsub/test_subscription.py\n@@ -25,7 +25,7 @@ class TestSubscription(unittest.TestCase):\n ENDPOINT = 'https://api.example.com/push'\n \n def _getTargetClass(self):\n- from gcloud.pubsub.subscription import Subscription\n+ from google.cloud.pubsub.subscription import Subscription\n return Subscription\n \n def _makeOne(self, *args, **kw):\n@@ -68,7 +68,7 @@ def test_ctor_w_neither_topic_nor_client(self):\n self._makeOne(self.SUB_NAME)\n \n def test_from_api_repr_no_topics(self):\n- from gcloud.pubsub.topic import Topic\n+ from google.cloud.pubsub.topic import Topic\n resource = {'topic': self.TOPIC_PATH,\n 'name': self.SUB_PATH,\n 'ackDeadlineSeconds': self.DEADLINE,\n@@ -99,7 +99,7 @@ def test_from_api_repr_w_deleted_topic(self):\n self.assertEqual(subscription.push_endpoint, self.ENDPOINT)\n \n def test_from_api_repr_w_topics_no_topic_match(self):\n- from gcloud.pubsub.topic import Topic\n+ from google.cloud.pubsub.topic import Topic\n resource = {'topic': self.TOPIC_PATH,\n 'name': self.SUB_PATH,\n 'ackDeadlineSeconds': self.DEADLINE,\n@@ -144,7 +144,7 @@ def test_full_name_and_path(self):\n self.assertEqual(subscription.path, SUB_PATH)\n \n def test_autoack_defaults(self):\n- from gcloud.pubsub.subscription import AutoAck\n+ from google.cloud.pubsub.subscription import AutoAck\n client = _Client(project=self.PROJECT)\n topic = _Topic(self.TOPIC_NAME, client=client)\n subscription = self._makeOne(self.SUB_NAME, topic)\n@@ -156,7 +156,7 @@ def test_autoack_defaults(self):\n self.assertTrue(auto_ack._client is None)\n \n def test_autoack_explicit(self):\n- from gcloud.pubsub.subscription import AutoAck\n+ from google.cloud.pubsub.subscription import AutoAck\n client1 = _Client(project=self.PROJECT)\n client2 = _Client(project=self.PROJECT)\n topic = _Topic(self.TOPIC_NAME, client=client1)\n@@ -323,7 +323,7 @@ def test_modify_push_config_wo_endpoint_w_alternate_client(self):\n \n def test_pull_wo_return_immediately_max_messages_w_bound_client(self):\n import base64\n- from gcloud.pubsub.message import Message\n+ from google.cloud.pubsub.message import Message\n ACK_ID = 'DEADBEEF'\n MSG_ID = 'BEADCAFE'\n PAYLOAD = b'This is the message text'\n@@ -350,7 +350,7 @@ def test_pull_wo_return_immediately_max_messages_w_bound_client(self):\n \n def test_pull_w_return_immediately_w_max_messages_w_alt_client(self):\n import base64\n- from gcloud.pubsub.message import Message\n+ from google.cloud.pubsub.message import Message\n ACK_ID = 'DEADBEEF'\n MSG_ID = 'BEADCAFE'\n PAYLOAD = b'This is the message text'\n@@ -450,7 +450,7 @@ def test_modify_ack_deadline_w_alternate_client(self):\n (self.SUB_PATH, [ACK_ID1, ACK_ID2], self.DEADLINE))\n \n def test_get_iam_policy_w_bound_client(self):\n- from gcloud.pubsub.iam import (\n+ from google.cloud.pubsub.iam import (\n PUBSUB_ADMIN_ROLE,\n PUBSUB_EDITOR_ROLE,\n PUBSUB_VIEWER_ROLE,\n@@ -515,8 +515,8 @@ def test_get_iam_policy_w_alternate_client(self):\n self.assertEqual(api._got_iam_policy, self.SUB_PATH)\n \n def test_set_iam_policy_w_bound_client(self):\n- from gcloud.pubsub.iam import Policy\n- from gcloud.pubsub.iam import (\n+ from google.cloud.pubsub.iam import Policy\n+ from google.cloud.pubsub.iam import (\n PUBSUB_ADMIN_ROLE,\n PUBSUB_EDITOR_ROLE,\n PUBSUB_VIEWER_ROLE,\n@@ -572,7 +572,7 @@ def test_set_iam_policy_w_bound_client(self):\n self.assertEqual(api._set_iam_policy, (self.SUB_PATH, POLICY))\n \n def test_set_iam_policy_w_alternate_client(self):\n- from gcloud.pubsub.iam import Policy\n+ from google.cloud.pubsub.iam import Policy\n RESPONSE = {'etag': 'ACAB'}\n client1 = _Client(project=self.PROJECT)\n client2 = _Client(project=self.PROJECT)\n@@ -592,7 +592,10 @@ def test_set_iam_policy_w_alternate_client(self):\n self.assertEqual(api._set_iam_policy, (self.SUB_PATH, {}))\n \n def test_check_iam_permissions_w_bound_client(self):\n- from gcloud.pubsub.iam import OWNER_ROLE, EDITOR_ROLE, VIEWER_ROLE\n+ from google.cloud.pubsub.iam import OWNER_ROLE\n+ from google.cloud.pubsub.iam import EDITOR_ROLE\n+ from google.cloud.pubsub.iam import VIEWER_ROLE\n+\n ROLES = [VIEWER_ROLE, EDITOR_ROLE, OWNER_ROLE]\n client = _Client(project=self.PROJECT)\n api = client.iam_policy_api = _FauxIAMPolicy()\n@@ -607,7 +610,10 @@ def test_check_iam_permissions_w_bound_client(self):\n (self.SUB_PATH, ROLES))\n \n def test_check_iam_permissions_w_alternate_client(self):\n- from gcloud.pubsub.iam import OWNER_ROLE, EDITOR_ROLE, VIEWER_ROLE\n+ from google.cloud.pubsub.iam import OWNER_ROLE\n+ from google.cloud.pubsub.iam import EDITOR_ROLE\n+ from google.cloud.pubsub.iam import VIEWER_ROLE\n+\n ROLES = [VIEWER_ROLE, EDITOR_ROLE, OWNER_ROLE]\n client1 = _Client(project=self.PROJECT)\n client2 = _Client(project=self.PROJECT)\n@@ -632,7 +638,7 @@ def subscription_create(self, subscription_path, topic_path,\n return self._subscription_create_response\n \n def subscription_get(self, subscription_path):\n- from gcloud.exceptions import NotFound\n+ from google.cloud.exceptions import NotFound\n self._subscription_got = subscription_path\n try:\n return self._subscription_get_response\n@@ -669,7 +675,7 @@ def subscription_modify_ack_deadline(self, subscription_path, ack_ids,\n class TestAutoAck(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.pubsub.subscription import AutoAck\n+ from google.cloud.pubsub.subscription import AutoAck\n return AutoAck\n \n def _makeOne(self, *args, **kw):\n@@ -774,7 +780,7 @@ def __init__(self, project):\n self.project = project\n \n def topic(self, name, timestamp_messages=False):\n- from gcloud.pubsub.topic import Topic\n+ from google.cloud.pubsub.topic import Topic\n return Topic(name, client=self, timestamp_messages=timestamp_messages)\n \n \ndiff --git a/unit_tests/pubsub/test_topic.py b/unit_tests/pubsub/test_topic.py\n--- a/unit_tests/pubsub/test_topic.py\n+++ b/unit_tests/pubsub/test_topic.py\n@@ -21,7 +21,7 @@ class TestTopic(unittest.TestCase):\n TOPIC_PATH = 'projects/%s/topics/%s' % (PROJECT, TOPIC_NAME)\n \n def _getTargetClass(self):\n- from gcloud.pubsub.topic import Topic\n+ from google.cloud.pubsub.topic import Topic\n return Topic\n \n def _makeOne(self, *args, **kw):\n@@ -138,8 +138,8 @@ def test_publish_single_bytes_wo_attrs_w_bound_client(self):\n def test_publish_single_bytes_wo_attrs_w_add_timestamp_alt_client(self):\n import base64\n import datetime\n- from gcloud.pubsub import topic as MUT\n- from gcloud._helpers import _RFC3339_MICROS\n+ from google.cloud.pubsub import topic as MUT\n+ from google.cloud._helpers import _RFC3339_MICROS\n from unit_tests._testing import _Monkey\n NOW = datetime.datetime.utcnow()\n \n@@ -287,7 +287,7 @@ def test_publish_multiple_error(self):\n self.assertEqual(getattr(api, '_topic_published', self), self)\n \n def test_subscription(self):\n- from gcloud.pubsub.subscription import Subscription\n+ from google.cloud.pubsub.subscription import Subscription\n client = _Client(project=self.PROJECT)\n topic = self._makeOne(self.TOPIC_NAME, client=client)\n \n@@ -298,7 +298,7 @@ def test_subscription(self):\n self.assertTrue(subscription.topic is topic)\n \n def test_list_subscriptions_no_paging(self):\n- from gcloud.pubsub.subscription import Subscription\n+ from google.cloud.pubsub.subscription import Subscription\n SUB_NAME_1 = 'subscription_1'\n SUB_PATH_1 = 'projects/%s/subscriptions/%s' % (\n self.PROJECT, SUB_NAME_1)\n@@ -332,7 +332,7 @@ def test_list_subscriptions_no_paging(self):\n (self.TOPIC_PATH, None, None))\n \n def test_list_subscriptions_with_paging(self):\n- from gcloud.pubsub.subscription import Subscription\n+ from google.cloud.pubsub.subscription import Subscription\n SUB_NAME_1 = 'subscription_1'\n SUB_PATH_1 = 'projects/%s/subscriptions/%s' % (\n self.PROJECT, SUB_NAME_1)\n@@ -382,7 +382,7 @@ def test_list_subscriptions_missing_key(self):\n (self.TOPIC_PATH, None, None))\n \n def test_get_iam_policy_w_bound_client(self):\n- from gcloud.pubsub.iam import (\n+ from google.cloud.pubsub.iam import (\n PUBSUB_ADMIN_ROLE,\n PUBSUB_EDITOR_ROLE,\n PUBSUB_VIEWER_ROLE,\n@@ -447,8 +447,8 @@ def test_get_iam_policy_w_alternate_client(self):\n self.assertEqual(api._got_iam_policy, self.TOPIC_PATH)\n \n def test_set_iam_policy_w_bound_client(self):\n- from gcloud.pubsub.iam import Policy\n- from gcloud.pubsub.iam import (\n+ from google.cloud.pubsub.iam import Policy\n+ from google.cloud.pubsub.iam import (\n PUBSUB_ADMIN_ROLE,\n PUBSUB_EDITOR_ROLE,\n PUBSUB_VIEWER_ROLE,\n@@ -509,7 +509,7 @@ def test_set_iam_policy_w_bound_client(self):\n self.assertEqual(api._set_iam_policy, (self.TOPIC_PATH, POLICY))\n \n def test_set_iam_policy_w_alternate_client(self):\n- from gcloud.pubsub.iam import Policy\n+ from google.cloud.pubsub.iam import Policy\n RESPONSE = {'etag': 'ACAB'}\n \n client1 = _Client(project=self.PROJECT)\n@@ -530,7 +530,10 @@ def test_set_iam_policy_w_alternate_client(self):\n self.assertEqual(api._set_iam_policy, (self.TOPIC_PATH, {}))\n \n def test_check_iam_permissions_w_bound_client(self):\n- from gcloud.pubsub.iam import OWNER_ROLE, EDITOR_ROLE, VIEWER_ROLE\n+ from google.cloud.pubsub.iam import OWNER_ROLE\n+ from google.cloud.pubsub.iam import EDITOR_ROLE\n+ from google.cloud.pubsub.iam import VIEWER_ROLE\n+\n ROLES = [VIEWER_ROLE, EDITOR_ROLE, OWNER_ROLE]\n client = _Client(project=self.PROJECT)\n api = client.iam_policy_api = _FauxIAMPolicy()\n@@ -544,7 +547,10 @@ def test_check_iam_permissions_w_bound_client(self):\n (self.TOPIC_PATH, ROLES))\n \n def test_check_iam_permissions_w_alternate_client(self):\n- from gcloud.pubsub.iam import OWNER_ROLE, EDITOR_ROLE, VIEWER_ROLE\n+ from google.cloud.pubsub.iam import OWNER_ROLE\n+ from google.cloud.pubsub.iam import EDITOR_ROLE\n+ from google.cloud.pubsub.iam import VIEWER_ROLE\n+\n ROLES = [VIEWER_ROLE, EDITOR_ROLE, OWNER_ROLE]\n client1 = _Client(project=self.PROJECT)\n client2 = _Client(project=self.PROJECT)\n@@ -563,7 +569,7 @@ class TestBatch(unittest.TestCase):\n PROJECT = 'PROJECT'\n \n def _getTargetClass(self):\n- from gcloud.pubsub.topic import Batch\n+ from google.cloud.pubsub.topic import Batch\n return Batch\n \n def _makeOne(self, *args, **kwargs):\n@@ -735,7 +741,7 @@ def topic_create(self, topic_path):\n return self._topic_create_response\n \n def topic_get(self, topic_path):\n- from gcloud.exceptions import NotFound\n+ from google.cloud.exceptions import NotFound\n self._topic_got = topic_path\n try:\n return self._topic_get_response\ndiff --git a/unit_tests/resource_manager/test_client.py b/unit_tests/resource_manager/test_client.py\n--- a/unit_tests/resource_manager/test_client.py\n+++ b/unit_tests/resource_manager/test_client.py\n@@ -18,7 +18,7 @@\n class Test__ProjectIterator(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.resource_manager.client import _ProjectIterator\n+ from google.cloud.resource_manager.client import _ProjectIterator\n return _ProjectIterator\n \n def _makeOne(self, *args, **kw):\n@@ -39,7 +39,7 @@ def test_get_items_from_response_empty(self):\n self.assertEqual(list(iterator.get_items_from_response({})), [])\n \n def test_get_items_from_response_non_empty(self):\n- from gcloud.resource_manager.project import Project\n+ from google.cloud.resource_manager.project import Project\n \n PROJECT_ID = 'project-id'\n PROJECT_NAME = 'My Project Name'\n@@ -72,14 +72,14 @@ def test_get_items_from_response_non_empty(self):\n class TestClient(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.resource_manager.client import Client\n+ from google.cloud.resource_manager.client import Client\n return Client\n \n def _makeOne(self, *args, **kw):\n return self._getTargetClass()(*args, **kw)\n \n def test_constructor(self):\n- from gcloud.resource_manager.connection import Connection\n+ from google.cloud.resource_manager.connection import Connection\n \n http = object()\n credentials = _Credentials()\n@@ -89,7 +89,7 @@ def test_constructor(self):\n self.assertEqual(client.connection._http, http)\n \n def test_new_project_factory(self):\n- from gcloud.resource_manager.project import Project\n+ from google.cloud.resource_manager.project import Project\n \n credentials = _Credentials()\n client = self._makeOne(credentials=credentials)\n@@ -105,7 +105,7 @@ def test_new_project_factory(self):\n self.assertEqual(project.labels, labels)\n \n def test_fetch_project(self):\n- from gcloud.resource_manager.project import Project\n+ from google.cloud.resource_manager.project import Project\n \n project_id = 'project-id'\n project_number = 123\n@@ -132,7 +132,7 @@ def test_fetch_project(self):\n self.assertEqual(project.labels, labels)\n \n def test_list_projects_return_type(self):\n- from gcloud.resource_manager.client import _ProjectIterator\n+ from google.cloud.resource_manager.client import _ProjectIterator\n \n credentials = _Credentials()\n client = self._makeOne(credentials=credentials)\ndiff --git a/unit_tests/resource_manager/test_connection.py b/unit_tests/resource_manager/test_connection.py\n--- a/unit_tests/resource_manager/test_connection.py\n+++ b/unit_tests/resource_manager/test_connection.py\n@@ -18,7 +18,7 @@\n class TestConnection(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.resource_manager.connection import Connection\n+ from google.cloud.resource_manager.connection import Connection\n return Connection\n \n def _makeOne(self, *args, **kw):\ndiff --git a/unit_tests/resource_manager/test_project.py b/unit_tests/resource_manager/test_project.py\n--- a/unit_tests/resource_manager/test_project.py\n+++ b/unit_tests/resource_manager/test_project.py\n@@ -18,7 +18,7 @@\n class TestProject(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.resource_manager.project import Project\n+ from google.cloud.resource_manager.project import Project\n return Project\n \n def _makeOne(self, *args, **kw):\n@@ -323,7 +323,7 @@ def __init__(self, *responses):\n self._requested = []\n \n def api_request(self, **kw):\n- from gcloud.exceptions import NotFound\n+ from google.cloud.exceptions import NotFound\n self._requested.append(kw)\n \n try:\ndiff --git a/unit_tests/storage/test__helpers.py b/unit_tests/storage/test__helpers.py\n--- a/unit_tests/storage/test__helpers.py\n+++ b/unit_tests/storage/test__helpers.py\n@@ -18,7 +18,7 @@\n class Test_PropertyMixin(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.storage._helpers import _PropertyMixin\n+ from google.cloud.storage._helpers import _PropertyMixin\n return _PropertyMixin\n \n def _makeOne(self, *args, **kw):\n@@ -97,7 +97,7 @@ def test_patch(self):\n class Test__scalar_property(unittest.TestCase):\n \n def _callFUT(self, fieldName):\n- from gcloud.storage._helpers import _scalar_property\n+ from google.cloud.storage._helpers import _scalar_property\n return _scalar_property(fieldName)\n \n def test_getter(self):\n@@ -125,7 +125,7 @@ def _patch_property(self, name, value):\n class Test__base64_md5hash(unittest.TestCase):\n \n def _callFUT(self, bytes_to_sign):\n- from gcloud.storage._helpers import _base64_md5hash\n+ from google.cloud.storage._helpers import _base64_md5hash\n return _base64_md5hash(bytes_to_sign)\n \n def test_it(self):\n@@ -140,7 +140,7 @@ def test_it(self):\n \n def test_it_with_stubs(self):\n from unit_tests._testing import _Monkey\n- from gcloud.storage import _helpers as MUT\n+ from google.cloud.storage import _helpers as MUT\n \n class _Buffer(object):\n \ndiff --git a/unit_tests/storage/test_acl.py b/unit_tests/storage/test_acl.py\n--- a/unit_tests/storage/test_acl.py\n+++ b/unit_tests/storage/test_acl.py\n@@ -18,7 +18,7 @@\n class Test_ACLEntity(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.storage.acl import _ACLEntity\n+ from google.cloud.storage.acl import _ACLEntity\n return _ACLEntity\n \n def _makeOne(self, *args, **kw):\n@@ -127,7 +127,7 @@ def test_revoke_owner(self):\n class Test_ACL(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.storage.acl import ACL\n+ from google.cloud.storage.acl import ACL\n return ACL\n \n def _makeOne(self, *args, **kw):\n@@ -267,7 +267,7 @@ def _reload():\n self.assertTrue(acl.loaded)\n \n def test_has_entity_miss_entity(self):\n- from gcloud.storage.acl import _ACLEntity\n+ from google.cloud.storage.acl import _ACLEntity\n TYPE = 'type'\n ID = 'id'\n entity = _ACLEntity(TYPE, ID)\n@@ -307,7 +307,7 @@ def _reload():\n self.assertTrue(acl.loaded)\n \n def test_get_entity_miss_entity_no_default(self):\n- from gcloud.storage.acl import _ACLEntity\n+ from google.cloud.storage.acl import _ACLEntity\n TYPE = 'type'\n ID = 'id'\n entity = _ACLEntity(TYPE, ID)\n@@ -322,7 +322,7 @@ def test_get_entity_miss_str_w_default(self):\n self.assertTrue(acl.get_entity('nonesuch', DEFAULT) is DEFAULT)\n \n def test_get_entity_miss_entity_w_default(self):\n- from gcloud.storage.acl import _ACLEntity\n+ from google.cloud.storage.acl import _ACLEntity\n DEFAULT = object()\n TYPE = 'type'\n ID = 'id'\n@@ -348,7 +348,7 @@ def test_get_entity_hit_entity(self):\n self.assertTrue(acl.has_entity(entity))\n \n def test_add_entity_miss_eager(self):\n- from gcloud.storage.acl import _ACLEntity\n+ from google.cloud.storage.acl import _ACLEntity\n TYPE = 'type'\n ID = 'id'\n ROLE = 'role'\n@@ -363,7 +363,7 @@ def test_add_entity_miss_eager(self):\n self.assertEqual(list(acl.get_entities()), [entity])\n \n def test_add_entity_miss_lazy(self):\n- from gcloud.storage.acl import _ACLEntity\n+ from google.cloud.storage.acl import _ACLEntity\n TYPE = 'type'\n ID = 'id'\n ROLE = 'role'\n@@ -383,7 +383,7 @@ def _reload():\n self.assertTrue(acl.loaded)\n \n def test_add_entity_hit(self):\n- from gcloud.storage.acl import _ACLEntity\n+ from google.cloud.storage.acl import _ACLEntity\n TYPE = 'type'\n ID = 'id'\n ENTITY_VAL = '%s-%s' % (TYPE, ID)\n@@ -510,7 +510,7 @@ def test_get_entities_nonempty(self):\n self.assertEqual(acl.get_entities(), [entity])\n \n def test_reload_missing(self):\n- # https://github.com/GoogleCloudPlatform/gcloud-python/issues/652\n+ # https://github.com/GoogleCloudPlatform/google-cloud-python/issues/652\n ROLE = 'role'\n connection = _Connection({})\n client = _Client(connection)\n@@ -711,7 +711,7 @@ def test_clear(self):\n class Test_BucketACL(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.storage.acl import BucketACL\n+ from google.cloud.storage.acl import BucketACL\n return BucketACL\n \n def _makeOne(self, *args, **kw):\n@@ -731,7 +731,7 @@ def test_ctor(self):\n class Test_DefaultObjectACL(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.storage.acl import DefaultObjectACL\n+ from google.cloud.storage.acl import DefaultObjectACL\n return DefaultObjectACL\n \n def _makeOne(self, *args, **kw):\n@@ -751,7 +751,7 @@ def test_ctor(self):\n class Test_ObjectACL(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.storage.acl import ObjectACL\n+ from google.cloud.storage.acl import ObjectACL\n return ObjectACL\n \n def _makeOne(self, *args, **kw):\ndiff --git a/unit_tests/storage/test_batch.py b/unit_tests/storage/test_batch.py\n--- a/unit_tests/storage/test_batch.py\n+++ b/unit_tests/storage/test_batch.py\n@@ -18,7 +18,7 @@\n class TestMIMEApplicationHTTP(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.storage.batch import MIMEApplicationHTTP\n+ from google.cloud.storage.batch import MIMEApplicationHTTP\n return MIMEApplicationHTTP\n \n def _makeOne(self, *args, **kw):\n@@ -69,7 +69,7 @@ def test_ctor_body_dict(self):\n class TestBatch(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.storage.batch import Batch\n+ from google.cloud.storage.batch import Batch\n return Batch\n \n def _makeOne(self, *args, **kw):\n@@ -85,7 +85,7 @@ def test_ctor(self):\n self.assertEqual(len(batch._target_objects), 0)\n \n def test_current(self):\n- from gcloud.storage.client import Client\n+ from google.cloud.storage.client import Client\n project = 'PROJECT'\n credentials = _Credentials()\n client = Client(project=project, credentials=credentials)\n@@ -100,7 +100,7 @@ def test_current(self):\n self.assertTrue(batch1.current() is batch2)\n \n def test__make_request_GET_normal(self):\n- from gcloud.storage.batch import _FutureDict\n+ from google.cloud.storage.batch import _FutureDict\n URL = 'http://example.com/api'\n expected = _Response()\n http = _HTTP((expected, ''))\n@@ -126,7 +126,7 @@ def test__make_request_GET_normal(self):\n self.assertEqual(solo_request[3], None)\n \n def test__make_request_POST_normal(self):\n- from gcloud.storage.batch import _FutureDict\n+ from google.cloud.storage.batch import _FutureDict\n URL = 'http://example.com/api'\n http = _HTTP() # no requests expected\n connection = _Connection(http=http)\n@@ -151,7 +151,7 @@ def test__make_request_POST_normal(self):\n self.assertEqual(solo_request[3], {'foo': 1})\n \n def test__make_request_PATCH_normal(self):\n- from gcloud.storage.batch import _FutureDict\n+ from google.cloud.storage.batch import _FutureDict\n URL = 'http://example.com/api'\n http = _HTTP() # no requests expected\n connection = _Connection(http=http)\n@@ -176,7 +176,7 @@ def test__make_request_PATCH_normal(self):\n self.assertEqual(solo_request[3], {'foo': 1})\n \n def test__make_request_DELETE_normal(self):\n- from gcloud.storage.batch import _FutureDict\n+ from google.cloud.storage.batch import _FutureDict\n URL = 'http://example.com/api'\n http = _HTTP() # no requests expected\n connection = _Connection(http=http)\n@@ -315,7 +315,7 @@ def test_finish_responses_mismatch(self):\n self.assertRaises(ValueError, batch.finish)\n \n def test_finish_nonempty_with_status_failure(self):\n- from gcloud.exceptions import NotFound\n+ from google.cloud.exceptions import NotFound\n URL = 'http://api.example.com/other_api'\n expected = _Response()\n expected['content-type'] = 'multipart/mixed; boundary=\"DEADBEEF=\"'\n@@ -370,7 +370,7 @@ def test_finish_nonempty_non_multipart_response(self):\n self.assertRaises(ValueError, batch.finish)\n \n def test_as_context_mgr_wo_error(self):\n- from gcloud.storage.client import Client\n+ from google.cloud.storage.client import Client\n URL = 'http://example.com/api'\n expected = _Response()\n expected['content-type'] = 'multipart/mixed; boundary=\"DEADBEEF=\"'\n@@ -406,8 +406,8 @@ def test_as_context_mgr_wo_error(self):\n self.assertEqual(target3._properties, '')\n \n def test_as_context_mgr_w_error(self):\n- from gcloud.storage.batch import _FutureDict\n- from gcloud.storage.client import Client\n+ from google.cloud.storage.batch import _FutureDict\n+ from google.cloud.storage.client import Client\n URL = 'http://example.com/api'\n http = _HTTP()\n connection = _Connection(http=http)\n@@ -447,7 +447,7 @@ def test_as_context_mgr_w_error(self):\n class Test__unpack_batch_response(unittest.TestCase):\n \n def _callFUT(self, response, content):\n- from gcloud.storage.batch import _unpack_batch_response\n+ from google.cloud.storage.batch import _unpack_batch_response\n return _unpack_batch_response(response, content)\n \n def _unpack_helper(self, response, content):\n@@ -538,7 +538,7 @@ def test_unicode(self):\n class Test__FutureDict(unittest.TestCase):\n \n def _makeOne(self, *args, **kw):\n- from gcloud.storage.batch import _FutureDict\n+ from google.cloud.storage.batch import _FutureDict\n return _FutureDict(*args, **kw)\n \n def test_get(self):\ndiff --git a/unit_tests/storage/test_blob.py b/unit_tests/storage/test_blob.py\n--- a/unit_tests/storage/test_blob.py\n+++ b/unit_tests/storage/test_blob.py\n@@ -18,7 +18,7 @@\n class Test_Blob(unittest.TestCase):\n \n def _makeOne(self, *args, **kw):\n- from gcloud.storage.blob import Blob\n+ from google.cloud.storage.blob import Blob\n properties = kw.pop('properties', None)\n blob = Blob(*args, **kw)\n blob._properties = properties or {}\n@@ -36,7 +36,7 @@ def test_ctor(self):\n self.assertTrue(blob._acl.blob is blob)\n \n def test_chunk_size_ctor(self):\n- from gcloud.storage.blob import Blob\n+ from google.cloud.storage.blob import Blob\n BLOB_NAME = 'blob-name'\n BUCKET = object()\n chunk_size = 10 * Blob._CHUNK_SIZE_MULTIPLE\n@@ -71,7 +71,7 @@ def test_chunk_size_setter_bad_value(self):\n blob.chunk_size = 11\n \n def test_acl_property(self):\n- from gcloud.storage.acl import ObjectACL\n+ from google.cloud.storage.acl import ObjectACL\n FAKE_BUCKET = _Bucket()\n blob = self._makeOne(None, bucket=FAKE_BUCKET)\n acl = blob.acl\n@@ -119,7 +119,7 @@ def test_public_url_w_slash_in_name(self):\n \n def _basic_generate_signed_url_helper(self, credentials=None):\n from unit_tests._testing import _Monkey\n- from gcloud.storage import blob as MUT\n+ from google.cloud.storage import blob as MUT\n \n BLOB_NAME = 'blob-name'\n EXPIRATION = '2014-10-16T20:34:37.000Z'\n@@ -158,7 +158,7 @@ def test_generate_signed_url_w_default_method(self):\n \n def test_generate_signed_url_w_content_type(self):\n from unit_tests._testing import _Monkey\n- from gcloud.storage import blob as MUT\n+ from google.cloud.storage import blob as MUT\n \n BLOB_NAME = 'blob-name'\n EXPIRATION = '2014-10-16T20:34:37.000Z'\n@@ -196,7 +196,7 @@ def test_generate_signed_url_w_credentials(self):\n \n def test_generate_signed_url_w_slash_in_name(self):\n from unit_tests._testing import _Monkey\n- from gcloud.storage import blob as MUT\n+ from google.cloud.storage import blob as MUT\n \n BLOB_NAME = 'parent/child'\n EXPIRATION = '2014-10-16T20:34:37.000Z'\n@@ -227,7 +227,7 @@ def test_generate_signed_url_w_slash_in_name(self):\n \n def test_generate_signed_url_w_method_arg(self):\n from unit_tests._testing import _Monkey\n- from gcloud.storage import blob as MUT\n+ from google.cloud.storage import blob as MUT\n \n BLOB_NAME = 'blob-name'\n EXPIRATION = '2014-10-16T20:34:37.000Z'\n@@ -511,7 +511,7 @@ def test_upload_from_file_stream(self):\n from six.moves.http_client import OK\n from six.moves.urllib.parse import parse_qsl\n from six.moves.urllib.parse import urlsplit\n- from gcloud.streaming import http_wrapper\n+ from google.cloud.streaming import http_wrapper\n \n BLOB_NAME = 'blob-name'\n UPLOAD_URL = 'http://example.com/upload/name/key'\n@@ -608,7 +608,7 @@ def test_upload_from_file_simple(self):\n \n def test_upload_from_file_simple_not_found(self):\n from six.moves.http_client import NOT_FOUND\n- from gcloud.exceptions import NotFound\n+ from google.cloud.exceptions import NotFound\n with self.assertRaises(NotFound):\n self._upload_from_file_simple_test_helper(status=NOT_FOUND)\n \n@@ -643,8 +643,8 @@ def test_upload_from_file_resumable(self):\n from six.moves.urllib.parse import urlsplit\n from unit_tests._testing import _Monkey\n from unit_tests._testing import _NamedTemporaryFile\n- from gcloud.streaming import http_wrapper\n- from gcloud.streaming import transfer\n+ from google.cloud.streaming import http_wrapper\n+ from google.cloud.streaming import transfer\n \n BLOB_NAME = 'blob-name'\n UPLOAD_URL = 'http://example.com/upload/name/key'\n@@ -727,8 +727,8 @@ def test_upload_from_file_resumable_w_error(self):\n from six.moves.urllib.parse import urlsplit\n from unit_tests._testing import _Monkey\n from unit_tests._testing import _NamedTemporaryFile\n- from gcloud.streaming import transfer\n- from gcloud.streaming.exceptions import HttpError\n+ from google.cloud.streaming import transfer\n+ from google.cloud.streaming.exceptions import HttpError\n \n BLOB_NAME = 'blob-name'\n DATA = b'ABCDEF'\n@@ -780,7 +780,7 @@ def test_upload_from_file_w_slash_in_name(self):\n from six.moves.urllib.parse import parse_qsl\n from six.moves.urllib.parse import urlsplit\n from unit_tests._testing import _NamedTemporaryFile\n- from gcloud.streaming import http_wrapper\n+ from google.cloud.streaming import http_wrapper\n \n BLOB_NAME = 'parent/child'\n UPLOAD_URL = 'http://example.com/upload/name/parent%2Fchild'\n@@ -830,7 +830,7 @@ def test_upload_from_filename_w_key(self):\n from six.moves.urllib.parse import parse_qsl\n from six.moves.urllib.parse import urlsplit\n from unit_tests._testing import _NamedTemporaryFile\n- from gcloud.streaming import http_wrapper\n+ from google.cloud.streaming import http_wrapper\n \n BLOB_NAME = 'blob-name'\n UPLOAD_URL = 'http://example.com/upload/name/key'\n@@ -889,7 +889,7 @@ def _upload_from_filename_test_helper(self, properties=None,\n from six.moves.urllib.parse import parse_qsl\n from six.moves.urllib.parse import urlsplit\n from unit_tests._testing import _NamedTemporaryFile\n- from gcloud.streaming import http_wrapper\n+ from google.cloud.streaming import http_wrapper\n \n BLOB_NAME = 'blob-name'\n UPLOAD_URL = 'http://example.com/upload/name/key'\n@@ -959,7 +959,7 @@ def test_upload_from_string_w_bytes(self):\n from six.moves.http_client import OK\n from six.moves.urllib.parse import parse_qsl\n from six.moves.urllib.parse import urlsplit\n- from gcloud.streaming import http_wrapper\n+ from google.cloud.streaming import http_wrapper\n BLOB_NAME = 'blob-name'\n UPLOAD_URL = 'http://example.com/upload/name/key'\n DATA = b'ABCDEF'\n@@ -998,7 +998,7 @@ def test_upload_from_string_w_text(self):\n from six.moves.http_client import OK\n from six.moves.urllib.parse import parse_qsl\n from six.moves.urllib.parse import urlsplit\n- from gcloud.streaming import http_wrapper\n+ from google.cloud.streaming import http_wrapper\n BLOB_NAME = 'blob-name'\n UPLOAD_URL = 'http://example.com/upload/name/key'\n DATA = u'ABCDEF\\u1234'\n@@ -1038,7 +1038,7 @@ def test_upload_from_string_text_w_key(self):\n from six.moves.http_client import OK\n from six.moves.urllib.parse import parse_qsl\n from six.moves.urllib.parse import urlsplit\n- from gcloud.streaming import http_wrapper\n+ from google.cloud.streaming import http_wrapper\n BLOB_NAME = 'blob-name'\n KEY = 'aa426195405adee2c8081bb9e7e74b19'\n HEADER_KEY_VALUE = 'YWE0MjYxOTU0MDVhZGVlMmM4MDgxYmI5ZTdlNzRiMTk='\n@@ -1084,7 +1084,7 @@ def test_upload_from_string_text_w_key(self):\n \n def test_make_public(self):\n from six.moves.http_client import OK\n- from gcloud.storage.acl import _ACLEntity\n+ from google.cloud.storage.acl import _ACLEntity\n BLOB_NAME = 'blob-name'\n permissive = [{'entity': 'allUsers', 'role': _ACLEntity.READER_ROLE}]\n after = ({'status': OK}, {'acl': permissive})\n@@ -1368,8 +1368,8 @@ def test_storage_class(self):\n \n def test_time_deleted(self):\n import datetime\n- from gcloud._helpers import _RFC3339_MICROS\n- from gcloud._helpers import UTC\n+ from google.cloud._helpers import _RFC3339_MICROS\n+ from google.cloud._helpers import UTC\n BLOB_NAME = 'blob-name'\n bucket = _Bucket()\n TIMESTAMP = datetime.datetime(2014, 11, 5, 20, 34, 37, tzinfo=UTC)\n@@ -1385,8 +1385,8 @@ def test_time_deleted_unset(self):\n \n def test_updated(self):\n import datetime\n- from gcloud._helpers import _RFC3339_MICROS\n- from gcloud._helpers import UTC\n+ from google.cloud._helpers import _RFC3339_MICROS\n+ from google.cloud._helpers import UTC\n BLOB_NAME = 'blob-name'\n bucket = _Bucket()\n TIMESTAMP = datetime.datetime(2014, 11, 5, 20, 34, 37, tzinfo=UTC)\n@@ -1426,7 +1426,7 @@ def __init__(self, *responses):\n \n def api_request(self, **kw):\n from six.moves.http_client import NOT_FOUND\n- from gcloud.exceptions import NotFound\n+ from google.cloud.exceptions import NotFound\n info, content = self._respond(**kw)\n if info.get('status') == NOT_FOUND:\n raise NotFound(info)\ndiff --git a/unit_tests/storage/test_bucket.py b/unit_tests/storage/test_bucket.py\n--- a/unit_tests/storage/test_bucket.py\n+++ b/unit_tests/storage/test_bucket.py\n@@ -18,7 +18,7 @@\n class Test__BlobIterator(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.storage.bucket import _BlobIterator\n+ from google.cloud.storage.bucket import _BlobIterator\n return _BlobIterator\n \n def _makeOne(self, *args, **kw):\n@@ -46,7 +46,7 @@ def test_get_items_from_response_empty(self):\n self.assertEqual(iterator.prefixes, set())\n \n def test_get_items_from_response_non_empty(self):\n- from gcloud.storage.blob import Blob\n+ from google.cloud.storage.blob import Blob\n BLOB_NAME = 'blob-name'\n response = {'items': [{'name': BLOB_NAME}], 'prefixes': ['foo']}\n connection = _Connection()\n@@ -61,7 +61,7 @@ def test_get_items_from_response_non_empty(self):\n self.assertEqual(iterator.prefixes, set(['foo']))\n \n def test_get_items_from_response_cumulative_prefixes(self):\n- from gcloud.storage.blob import Blob\n+ from google.cloud.storage.blob import Blob\n BLOB_NAME = 'blob-name1'\n response1 = {'items': [{'name': BLOB_NAME}], 'prefixes': ['foo']}\n response2 = {\n@@ -88,7 +88,7 @@ def test_get_items_from_response_cumulative_prefixes(self):\n class Test_Bucket(unittest.TestCase):\n \n def _makeOne(self, client=None, name=None, properties=None):\n- from gcloud.storage.bucket import Bucket\n+ from google.cloud.storage.bucket import Bucket\n if client is None:\n connection = _Connection()\n client = _Client(connection)\n@@ -108,7 +108,7 @@ def test_ctor(self):\n self.assertTrue(bucket._default_object_acl.bucket is bucket)\n \n def test_blob(self):\n- from gcloud.storage.blob import Blob\n+ from google.cloud.storage.blob import Blob\n \n BUCKET_NAME = 'BUCKET_NAME'\n BLOB_NAME = 'BLOB_NAME'\n@@ -123,7 +123,7 @@ def test_blob(self):\n self.assertEqual(blob.chunk_size, CHUNK_SIZE)\n \n def test_exists_miss(self):\n- from gcloud.exceptions import NotFound\n+ from google.cloud.exceptions import NotFound\n \n class _FakeConnection(object):\n \n@@ -230,14 +230,14 @@ def test_create_w_extra_properties(self):\n self.assertEqual(kw['data'], DATA)\n \n def test_acl_property(self):\n- from gcloud.storage.acl import BucketACL\n+ from google.cloud.storage.acl import BucketACL\n bucket = self._makeOne()\n acl = bucket.acl\n self.assertTrue(isinstance(acl, BucketACL))\n self.assertTrue(acl is bucket._acl)\n \n def test_default_object_acl_property(self):\n- from gcloud.storage.acl import DefaultObjectACL\n+ from google.cloud.storage.acl import DefaultObjectACL\n bucket = self._makeOne()\n acl = bucket.default_object_acl\n self.assertTrue(isinstance(acl, DefaultObjectACL))\n@@ -342,7 +342,7 @@ def test_list_blobs(self):\n self.assertEqual(kw['query_params'], {'projection': 'noAcl'})\n \n def test_delete_miss(self):\n- from gcloud.exceptions import NotFound\n+ from google.cloud.exceptions import NotFound\n NAME = 'name'\n connection = _Connection()\n client = _Client(connection)\n@@ -435,7 +435,7 @@ def test_delete_too_many(self):\n self.assertEqual(connection._deleted_buckets, [])\n \n def test_delete_blob_miss(self):\n- from gcloud.exceptions import NotFound\n+ from google.cloud.exceptions import NotFound\n NAME = 'name'\n NONESUCH = 'nonesuch'\n connection = _Connection()\n@@ -479,7 +479,7 @@ def test_delete_blobs_hit(self):\n self.assertEqual(kw[0]['path'], '/b/%s/o/%s' % (NAME, BLOB_NAME))\n \n def test_delete_blobs_miss_no_on_error(self):\n- from gcloud.exceptions import NotFound\n+ from google.cloud.exceptions import NotFound\n NAME = 'name'\n BLOB_NAME = 'blob-name'\n NONESUCH = 'nonesuch'\n@@ -790,8 +790,8 @@ def test_storage_class_setter_DURABLE_REDUCED_AVAILABILITY(self):\n \n def test_time_created(self):\n import datetime\n- from gcloud._helpers import _RFC3339_MICROS\n- from gcloud._helpers import UTC\n+ from google.cloud._helpers import _RFC3339_MICROS\n+ from google.cloud._helpers import UTC\n TIMESTAMP = datetime.datetime(2014, 11, 5, 20, 34, 37, tzinfo=UTC)\n TIME_CREATED = TIMESTAMP.strftime(_RFC3339_MICROS)\n properties = {'timeCreated': TIME_CREATED}\n@@ -845,7 +845,7 @@ def test_disable_website(self):\n self.assertEqual(bucket._properties, UNSET)\n \n def test_make_public_defaults(self):\n- from gcloud.storage.acl import _ACLEntity\n+ from google.cloud.storage.acl import _ACLEntity\n NAME = 'name'\n permissive = [{'entity': 'allUsers', 'role': _ACLEntity.READER_ROLE}]\n after = {'acl': permissive, 'defaultObjectAcl': []}\n@@ -865,7 +865,7 @@ def test_make_public_defaults(self):\n self.assertEqual(kw[0]['query_params'], {'projection': 'full'})\n \n def _make_public_w_future_helper(self, default_object_acl_loaded=True):\n- from gcloud.storage.acl import _ACLEntity\n+ from google.cloud.storage.acl import _ACLEntity\n NAME = 'name'\n permissive = [{'entity': 'allUsers', 'role': _ACLEntity.READER_ROLE}]\n after1 = {'acl': permissive, 'defaultObjectAcl': []}\n@@ -907,8 +907,8 @@ def test_make_public_w_future_reload_default(self):\n self._make_public_w_future_helper(default_object_acl_loaded=False)\n \n def test_make_public_recursive(self):\n- from gcloud.storage.acl import _ACLEntity\n- from gcloud.storage.bucket import _BlobIterator\n+ from google.cloud.storage.acl import _ACLEntity\n+ from google.cloud.storage.bucket import _BlobIterator\n _saved = []\n \n class _Blob(object):\n@@ -965,7 +965,7 @@ def get_items_from_response(self, response):\n {'maxResults': max_results, 'projection': 'full'})\n \n def test_make_public_recursive_too_many(self):\n- from gcloud.storage.acl import _ACLEntity\n+ from google.cloud.storage.acl import _ACLEntity\n \n PERMISSIVE = [{'entity': 'allUsers', 'role': _ACLEntity.READER_ROLE}]\n AFTER = {'acl': PERMISSIVE, 'defaultObjectAcl': []}\n@@ -1004,7 +1004,7 @@ def _is_bucket_path(path):\n return path.startswith('/b/') and path.count('/') == 2\n \n def api_request(self, **kw):\n- from gcloud.exceptions import NotFound\n+ from google.cloud.exceptions import NotFound\n self._requested.append(kw)\n \n method = kw.get('method')\ndiff --git a/unit_tests/storage/test_client.py b/unit_tests/storage/test_client.py\n--- a/unit_tests/storage/test_client.py\n+++ b/unit_tests/storage/test_client.py\n@@ -18,14 +18,14 @@\n class TestClient(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.storage.client import Client\n+ from google.cloud.storage.client import Client\n return Client\n \n def _makeOne(self, *args, **kw):\n return self._getTargetClass()(*args, **kw)\n \n def test_ctor_connection_type(self):\n- from gcloud.storage.connection import Connection\n+ from google.cloud.storage.connection import Connection\n \n PROJECT = 'PROJECT'\n CREDENTIALS = _Credentials()\n@@ -38,7 +38,7 @@ def test_ctor_connection_type(self):\n self.assertEqual(list(client._batch_stack), [])\n \n def test__push_batch_and__pop_batch(self):\n- from gcloud.storage.batch import Batch\n+ from google.cloud.storage.batch import Batch\n \n PROJECT = 'PROJECT'\n CREDENTIALS = _Credentials()\n@@ -80,7 +80,7 @@ def test_connection_getter_no_batch(self):\n self.assertTrue(client.current_batch is None)\n \n def test_connection_getter_with_batch(self):\n- from gcloud.storage.batch import Batch\n+ from google.cloud.storage.batch import Batch\n PROJECT = 'PROJECT'\n CREDENTIALS = _Credentials()\n client = self._makeOne(project=PROJECT, credentials=CREDENTIALS)\n@@ -91,7 +91,7 @@ def test_connection_getter_with_batch(self):\n self.assertTrue(client.current_batch is batch)\n \n def test_bucket(self):\n- from gcloud.storage.bucket import Bucket\n+ from google.cloud.storage.bucket import Bucket\n \n PROJECT = 'PROJECT'\n CREDENTIALS = _Credentials()\n@@ -104,7 +104,7 @@ def test_bucket(self):\n self.assertEqual(bucket.name, BUCKET_NAME)\n \n def test_batch(self):\n- from gcloud.storage.batch import Batch\n+ from google.cloud.storage.batch import Batch\n \n PROJECT = 'PROJECT'\n CREDENTIALS = _Credentials()\n@@ -115,7 +115,7 @@ def test_batch(self):\n self.assertTrue(batch._client is client)\n \n def test_get_bucket_miss(self):\n- from gcloud.exceptions import NotFound\n+ from google.cloud.exceptions import NotFound\n \n PROJECT = 'PROJECT'\n CREDENTIALS = _Credentials()\n@@ -138,7 +138,7 @@ def test_get_bucket_miss(self):\n self.assertEqual(http._called_with['uri'], URI)\n \n def test_get_bucket_hit(self):\n- from gcloud.storage.bucket import Bucket\n+ from google.cloud.storage.bucket import Bucket\n \n PROJECT = 'PROJECT'\n CREDENTIALS = _Credentials()\n@@ -186,7 +186,7 @@ def test_lookup_bucket_miss(self):\n self.assertEqual(http._called_with['uri'], URI)\n \n def test_lookup_bucket_hit(self):\n- from gcloud.storage.bucket import Bucket\n+ from google.cloud.storage.bucket import Bucket\n \n PROJECT = 'PROJECT'\n CREDENTIALS = _Credentials()\n@@ -212,7 +212,7 @@ def test_lookup_bucket_hit(self):\n self.assertEqual(http._called_with['uri'], URI)\n \n def test_create_bucket_conflict(self):\n- from gcloud.exceptions import Conflict\n+ from google.cloud.exceptions import Conflict\n \n PROJECT = 'PROJECT'\n CREDENTIALS = _Credentials()\n@@ -235,7 +235,7 @@ def test_create_bucket_conflict(self):\n self.assertEqual(http._called_with['uri'], URI)\n \n def test_create_bucket_success(self):\n- from gcloud.storage.bucket import Bucket\n+ from google.cloud.storage.bucket import Bucket\n \n PROJECT = 'PROJECT'\n CREDENTIALS = _Credentials()\n@@ -373,7 +373,7 @@ def test_list_buckets_all_arguments(self):\n class Test__BucketIterator(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.storage.client import _BucketIterator\n+ from google.cloud.storage.client import _BucketIterator\n return _BucketIterator\n \n def _makeOne(self, *args, **kw):\n@@ -395,7 +395,7 @@ def test_get_items_from_response_empty(self):\n self.assertEqual(list(iterator.get_items_from_response({})), [])\n \n def test_get_items_from_response_non_empty(self):\n- from gcloud.storage.bucket import Bucket\n+ from google.cloud.storage.bucket import Bucket\n BLOB_NAME = 'blob-name'\n response = {'items': [{'name': BLOB_NAME}]}\n connection = object()\ndiff --git a/unit_tests/storage/test_connection.py b/unit_tests/storage/test_connection.py\n--- a/unit_tests/storage/test_connection.py\n+++ b/unit_tests/storage/test_connection.py\n@@ -18,7 +18,7 @@\n class TestConnection(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.storage.connection import Connection\n+ from google.cloud.storage.connection import Connection\n return Connection\n \n def _makeOne(self, *args, **kw):\ndiff --git a/unit_tests/streaming/test_buffered_stream.py b/unit_tests/streaming/test_buffered_stream.py\n--- a/unit_tests/streaming/test_buffered_stream.py\n+++ b/unit_tests/streaming/test_buffered_stream.py\n@@ -4,7 +4,7 @@\n class Test_BufferedStream(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.streaming.buffered_stream import BufferedStream\n+ from google.cloud.streaming.buffered_stream import BufferedStream\n return BufferedStream\n \n def _makeOne(self, *args, **kw):\ndiff --git a/unit_tests/streaming/test_exceptions.py b/unit_tests/streaming/test_exceptions.py\n--- a/unit_tests/streaming/test_exceptions.py\n+++ b/unit_tests/streaming/test_exceptions.py\n@@ -4,7 +4,7 @@\n class Test_HttpError(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.streaming.exceptions import HttpError\n+ from google.cloud.streaming.exceptions import HttpError\n return HttpError\n \n def _makeOne(self, *args, **kw):\n@@ -45,7 +45,7 @@ class _Response(object):\n class Test_RetryAfterError(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.streaming.exceptions import RetryAfterError\n+ from google.cloud.streaming.exceptions import RetryAfterError\n return RetryAfterError\n \n def _makeOne(self, *args, **kw):\ndiff --git a/unit_tests/streaming/test_http_wrapper.py b/unit_tests/streaming/test_http_wrapper.py\n--- a/unit_tests/streaming/test_http_wrapper.py\n+++ b/unit_tests/streaming/test_http_wrapper.py\n@@ -4,7 +4,7 @@\n class Test__httplib2_debug_level(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.streaming.http_wrapper import _httplib2_debug_level\n+ from google.cloud.streaming.http_wrapper import _httplib2_debug_level\n return _httplib2_debug_level\n \n def _makeOne(self, *args, **kw):\n@@ -12,7 +12,7 @@ def _makeOne(self, *args, **kw):\n \n def test_wo_loggable_body_wo_http(self):\n from unit_tests._testing import _Monkey\n- from gcloud.streaming import http_wrapper as MUT\n+ from google.cloud.streaming import http_wrapper as MUT\n \n request = _Request()\n LEVEL = 1\n@@ -23,7 +23,7 @@ def test_wo_loggable_body_wo_http(self):\n \n def test_w_loggable_body_wo_http(self):\n from unit_tests._testing import _Monkey\n- from gcloud.streaming import http_wrapper as MUT\n+ from google.cloud.streaming import http_wrapper as MUT\n \n request = _Request(loggable_body=object())\n LEVEL = 1\n@@ -35,7 +35,7 @@ def test_w_loggable_body_wo_http(self):\n \n def test_w_loggable_body_w_http(self):\n from unit_tests._testing import _Monkey\n- from gcloud.streaming import http_wrapper as MUT\n+ from google.cloud.streaming import http_wrapper as MUT\n \n class _Connection(object):\n debuglevel = 0\n@@ -63,7 +63,7 @@ def set_debuglevel(self, value):\n class Test_Request(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.streaming.http_wrapper import Request\n+ from google.cloud.streaming.http_wrapper import Request\n return Request\n \n def _makeOne(self, *args, **kw):\n@@ -78,7 +78,7 @@ def test_ctor_defaults(self):\n self.assertEqual(request.loggable_body, None)\n \n def test_loggable_body_setter_w_body_None(self):\n- from gcloud.streaming.exceptions import RequestError\n+ from google.cloud.streaming.exceptions import RequestError\n request = self._makeOne(body=None)\n with self.assertRaises(RequestError):\n request.loggable_body = 'abc'\n@@ -103,7 +103,7 @@ def test_body_setter_w_non_string(self):\n class Test_Response(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.streaming.http_wrapper import Response\n+ from google.cloud.streaming.http_wrapper import Response\n return Response\n \n def _makeOne(self, *args, **kw):\n@@ -199,23 +199,23 @@ def test_is_redirect_w_code_w_location(self):\n class Test__check_response(unittest.TestCase):\n \n def _callFUT(self, *args, **kw):\n- from gcloud.streaming.http_wrapper import _check_response\n+ from google.cloud.streaming.http_wrapper import _check_response\n return _check_response(*args, **kw)\n \n def test_w_none(self):\n- from gcloud.streaming.exceptions import RequestError\n+ from google.cloud.streaming.exceptions import RequestError\n with self.assertRaises(RequestError):\n self._callFUT(None)\n \n def test_w_TOO_MANY_REQUESTS(self):\n- from gcloud.streaming.exceptions import BadStatusCodeError\n- from gcloud.streaming.http_wrapper import TOO_MANY_REQUESTS\n+ from google.cloud.streaming.exceptions import BadStatusCodeError\n+ from google.cloud.streaming.http_wrapper import TOO_MANY_REQUESTS\n \n with self.assertRaises(BadStatusCodeError):\n self._callFUT(_Response(TOO_MANY_REQUESTS))\n \n def test_w_50x(self):\n- from gcloud.streaming.exceptions import BadStatusCodeError\n+ from google.cloud.streaming.exceptions import BadStatusCodeError\n \n with self.assertRaises(BadStatusCodeError):\n self._callFUT(_Response(500))\n@@ -224,7 +224,7 @@ def test_w_50x(self):\n self._callFUT(_Response(503))\n \n def test_w_retry_after(self):\n- from gcloud.streaming.exceptions import RetryAfterError\n+ from google.cloud.streaming.exceptions import RetryAfterError\n \n with self.assertRaises(RetryAfterError):\n self._callFUT(_Response(200, 20))\n@@ -236,7 +236,7 @@ def test_pass(self):\n class Test__reset_http_connections(unittest.TestCase):\n \n def _callFUT(self, *args, **kw):\n- from gcloud.streaming.http_wrapper import _reset_http_connections\n+ from google.cloud.streaming.http_wrapper import _reset_http_connections\n return _reset_http_connections(*args, **kw)\n \n def test_wo_connections(self):\n@@ -254,7 +254,8 @@ def test_w_connections(self):\n class Test___make_api_request_no_retry(unittest.TestCase):\n \n def _callFUT(self, *args, **kw):\n- from gcloud.streaming.http_wrapper import _make_api_request_no_retry\n+ from google.cloud.streaming.http_wrapper import (\n+ _make_api_request_no_retry)\n return _make_api_request_no_retry(*args, **kw)\n \n def _verify_requested(self, http, request,\n@@ -270,7 +271,7 @@ def _verify_requested(self, http, request,\n \n def test_defaults_wo_connections(self):\n from unit_tests._testing import _Monkey\n- from gcloud.streaming import http_wrapper as MUT\n+ from google.cloud.streaming import http_wrapper as MUT\n INFO = {'status': '200'}\n CONTENT = 'CONTENT'\n _http = _Http((INFO, CONTENT))\n@@ -290,7 +291,7 @@ def test_defaults_wo_connections(self):\n \n def test_w_http_connections_miss(self):\n from unit_tests._testing import _Monkey\n- from gcloud.streaming import http_wrapper as MUT\n+ from google.cloud.streaming import http_wrapper as MUT\n INFO = {'status': '200'}\n CONTENT = 'CONTENT'\n CONN_TYPE = object()\n@@ -312,7 +313,7 @@ def test_w_http_connections_miss(self):\n \n def test_w_http_connections_hit(self):\n from unit_tests._testing import _Monkey\n- from gcloud.streaming import http_wrapper as MUT\n+ from google.cloud.streaming import http_wrapper as MUT\n INFO = {'status': '200'}\n CONTENT = 'CONTENT'\n CONN_TYPE = object()\n@@ -334,8 +335,8 @@ def test_w_http_connections_hit(self):\n \n def test_w_request_returning_None(self):\n from unit_tests._testing import _Monkey\n- from gcloud.streaming import http_wrapper as MUT\n- from gcloud.streaming.exceptions import RequestError\n+ from google.cloud.streaming import http_wrapper as MUT\n+ from google.cloud.streaming.exceptions import RequestError\n INFO = None\n CONTENT = None\n CONN_TYPE = object()\n@@ -352,11 +353,11 @@ def test_w_request_returning_None(self):\n class Test_make_api_request(unittest.TestCase):\n \n def _callFUT(self, *args, **kw):\n- from gcloud.streaming.http_wrapper import make_api_request\n+ from google.cloud.streaming.http_wrapper import make_api_request\n return make_api_request(*args, **kw)\n \n def test_wo_exception(self):\n- from gcloud.streaming import http_wrapper as MUT\n+ from google.cloud.streaming import http_wrapper as MUT\n from unit_tests._testing import _Monkey\n \n HTTP, REQUEST, RESPONSE = object(), object(), object()\n@@ -376,8 +377,8 @@ def _wo_exception(*args, **kw):\n self.assertEqual(_checked, []) # not called by '_wo_exception'\n \n def test_w_exceptions_lt_max_retries(self):\n- from gcloud.streaming.exceptions import RetryAfterError\n- from gcloud.streaming import http_wrapper as MUT\n+ from google.cloud.streaming.exceptions import RetryAfterError\n+ from google.cloud.streaming import http_wrapper as MUT\n from unit_tests._testing import _Monkey\n \n HTTP, RESPONSE = object(), object()\n@@ -405,7 +406,7 @@ def _wo_exception(*args, **kw):\n \n def test_w_exceptions_gt_max_retries(self):\n from unit_tests._testing import _Monkey\n- from gcloud.streaming import http_wrapper as MUT\n+ from google.cloud.streaming import http_wrapper as MUT\n HTTP = object()\n REQUEST = _Request()\n _created, _checked = [], []\ndiff --git a/unit_tests/streaming/test_stream_slice.py b/unit_tests/streaming/test_stream_slice.py\n--- a/unit_tests/streaming/test_stream_slice.py\n+++ b/unit_tests/streaming/test_stream_slice.py\n@@ -4,7 +4,7 @@\n class Test_StreamSlice(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.streaming.stream_slice import StreamSlice\n+ from google.cloud.streaming.stream_slice import StreamSlice\n return StreamSlice\n \n def _makeOne(self, *args, **kw):\ndiff --git a/unit_tests/streaming/test_transfer.py b/unit_tests/streaming/test_transfer.py\n--- a/unit_tests/streaming/test_transfer.py\n+++ b/unit_tests/streaming/test_transfer.py\n@@ -5,14 +5,14 @@ class Test__Transfer(unittest.TestCase):\n URL = 'http://example.com/api'\n \n def _getTargetClass(self):\n- from gcloud.streaming.transfer import _Transfer\n+ from google.cloud.streaming.transfer import _Transfer\n return _Transfer\n \n def _makeOne(self, *args, **kw):\n return self._getTargetClass()(*args, **kw)\n \n def test_ctor_defaults(self):\n- from gcloud.streaming.transfer import _DEFAULT_CHUNKSIZE\n+ from google.cloud.streaming.transfer import _DEFAULT_CHUNKSIZE\n stream = _Stream()\n xfer = self._makeOne(stream)\n self.assertTrue(xfer.stream is stream)\n@@ -98,7 +98,7 @@ def test__initialize_w_existing_http(self):\n self.assertTrue(xfer.url is self.URL)\n \n def test__initialize_already_initialized(self):\n- from gcloud.streaming.exceptions import TransferInvalidError\n+ from google.cloud.streaming.exceptions import TransferInvalidError\n URL_2 = 'http://example.com/other'\n HTTP_1, HTTP_2 = object(), object()\n stream = _Stream()\n@@ -115,7 +115,7 @@ def test__ensure_initialized_hit(self):\n xfer._ensure_initialized() # no raise\n \n def test__ensure_initialized_miss(self):\n- from gcloud.streaming.exceptions import TransferInvalidError\n+ from google.cloud.streaming.exceptions import TransferInvalidError\n stream = _Stream()\n xfer = self._makeOne(stream)\n with self.assertRaises(TransferInvalidError):\n@@ -127,7 +127,7 @@ def test__ensure_uninitialized_hit(self):\n xfer._ensure_uninitialized() # no raise\n \n def test__ensure_uninitialized_miss(self):\n- from gcloud.streaming.exceptions import TransferInvalidError\n+ from google.cloud.streaming.exceptions import TransferInvalidError\n stream = _Stream()\n HTTP = object()\n xfer = self._makeOne(stream)\n@@ -149,7 +149,7 @@ class Test_Download(unittest.TestCase):\n URL = \"http://example.com/api\"\n \n def _getTargetClass(self):\n- from gcloud.streaming.transfer import Download\n+ from google.cloud.streaming.transfer import Download\n return Download\n \n def _makeOne(self, *args, **kw):\n@@ -258,7 +258,7 @@ def test__set_total_w_content_range_w_asterisk_total(self):\n self.assertEqual(download.total_size, 0)\n \n def test_initialize_download_already_initialized(self):\n- from gcloud.streaming.exceptions import TransferInvalidError\n+ from google.cloud.streaming.exceptions import TransferInvalidError\n request = _Request()\n download = self._makeOne(_Stream())\n download._initialize(None, self.URL)\n@@ -276,8 +276,8 @@ def test_initialize_download_wo_autotransfer(self):\n def test_initialize_download_w_autotransfer_failing(self):\n from six.moves import http_client\n from unit_tests._testing import _Monkey\n- from gcloud.streaming import transfer as MUT\n- from gcloud.streaming.exceptions import HttpError\n+ from google.cloud.streaming import transfer as MUT\n+ from google.cloud.streaming.exceptions import HttpError\n request = _Request()\n http = object()\n download = self._makeOne(_Stream(), auto_transfer=True)\n@@ -295,7 +295,7 @@ def test_initialize_download_w_autotransfer_failing(self):\n def test_initialize_download_w_autotransfer_w_content_location(self):\n from six.moves import http_client\n from unit_tests._testing import _Monkey\n- from gcloud.streaming import transfer as MUT\n+ from google.cloud.streaming import transfer as MUT\n REDIRECT_URL = 'http://example.com/other'\n request = _Request()\n http = object()\n@@ -316,14 +316,14 @@ def test_initialize_download_w_autotransfer_w_content_location(self):\n self.assertTrue(requester._requested[0][0] is request)\n \n def test__normalize_start_end_w_end_w_start_lt_0(self):\n- from gcloud.streaming.exceptions import TransferInvalidError\n+ from google.cloud.streaming.exceptions import TransferInvalidError\n download = self._makeOne(_Stream())\n \n with self.assertRaises(TransferInvalidError):\n download._normalize_start_end(-1, 0)\n \n def test__normalize_start_end_w_end_w_start_gt_total(self):\n- from gcloud.streaming.exceptions import TransferInvalidError\n+ from google.cloud.streaming.exceptions import TransferInvalidError\n download = self._makeOne(_Stream())\n download._set_total({'content-range': 'bytes 0-1/2'})\n \n@@ -331,7 +331,7 @@ def test__normalize_start_end_w_end_w_start_gt_total(self):\n download._normalize_start_end(3, 0)\n \n def test__normalize_start_end_w_end_lt_start(self):\n- from gcloud.streaming.exceptions import TransferInvalidError\n+ from google.cloud.streaming.exceptions import TransferInvalidError\n download = self._makeOne(_Stream())\n download._set_total({'content-range': 'bytes 0-1/2'})\n \n@@ -403,7 +403,7 @@ def test__compute_end_byte_w_start_ge_0_wo_end_w_total_size(self):\n self.assertEqual(download._compute_end_byte(0, use_chunks=False), 9)\n \n def test__get_chunk_not_initialized(self):\n- from gcloud.streaming.exceptions import TransferInvalidError\n+ from google.cloud.streaming.exceptions import TransferInvalidError\n download = self._makeOne(_Stream())\n \n with self.assertRaises(TransferInvalidError):\n@@ -412,7 +412,7 @@ def test__get_chunk_not_initialized(self):\n def test__get_chunk(self):\n from six.moves import http_client\n from unit_tests._testing import _Monkey\n- from gcloud.streaming import transfer as MUT\n+ from google.cloud.streaming import transfer as MUT\n http = object()\n download = self._makeOne(_Stream())\n download._initialize(http, self.URL)\n@@ -430,7 +430,7 @@ def test__get_chunk(self):\n self.assertEqual(request.headers['range'], 'bytes=0-10')\n \n def test__process_response_w_FORBIDDEN(self):\n- from gcloud.streaming.exceptions import HttpError\n+ from google.cloud.streaming.exceptions import HttpError\n from six.moves import http_client\n download = self._makeOne(_Stream())\n response = _makeResponse(http_client.FORBIDDEN)\n@@ -438,7 +438,7 @@ def test__process_response_w_FORBIDDEN(self):\n download._process_response(response)\n \n def test__process_response_w_NOT_FOUND(self):\n- from gcloud.streaming.exceptions import HttpError\n+ from google.cloud.streaming.exceptions import HttpError\n from six.moves import http_client\n download = self._makeOne(_Stream())\n response = _makeResponse(http_client.NOT_FOUND)\n@@ -446,7 +446,7 @@ def test__process_response_w_NOT_FOUND(self):\n download._process_response(response)\n \n def test__process_response_w_other_error(self):\n- from gcloud.streaming.exceptions import TransferRetryError\n+ from google.cloud.streaming.exceptions import TransferRetryError\n from six.moves import http_client\n download = self._makeOne(_Stream())\n response = _makeResponse(http_client.BAD_REQUEST)\n@@ -500,7 +500,7 @@ def test__process_response_w_NO_CONTENT(self):\n self.assertEqual(download.encoding, None)\n \n def test_get_range_not_initialized(self):\n- from gcloud.streaming.exceptions import TransferInvalidError\n+ from google.cloud.streaming.exceptions import TransferInvalidError\n download = self._makeOne(_Stream())\n with self.assertRaises(TransferInvalidError):\n download.get_range(0, 10)\n@@ -508,7 +508,7 @@ def test_get_range_not_initialized(self):\n def test_get_range_wo_total_size_complete(self):\n from six.moves import http_client\n from unit_tests._testing import _Monkey\n- from gcloud.streaming import transfer as MUT\n+ from google.cloud.streaming import transfer as MUT\n CONTENT = b'ABCDEFGHIJ'\n LEN = len(CONTENT)\n REQ_RANGE = 'bytes=0-%d' % (LEN,)\n@@ -535,7 +535,7 @@ def test_get_range_wo_total_size_complete(self):\n def test_get_range_wo_total_size_wo_end(self):\n from six.moves import http_client\n from unit_tests._testing import _Monkey\n- from gcloud.streaming import transfer as MUT\n+ from google.cloud.streaming import transfer as MUT\n CONTENT = b'ABCDEFGHIJ'\n LEN = len(CONTENT)\n START = 5\n@@ -564,7 +564,7 @@ def test_get_range_wo_total_size_wo_end(self):\n def test_get_range_w_total_size_partial(self):\n from six.moves import http_client\n from unit_tests._testing import _Monkey\n- from gcloud.streaming import transfer as MUT\n+ from google.cloud.streaming import transfer as MUT\n CONTENT = b'ABCDEFGHIJ'\n LEN = len(CONTENT)\n PARTIAL_LEN = 5\n@@ -593,8 +593,8 @@ def test_get_range_w_total_size_partial(self):\n def test_get_range_w_empty_chunk(self):\n from six.moves import http_client\n from unit_tests._testing import _Monkey\n- from gcloud.streaming import transfer as MUT\n- from gcloud.streaming.exceptions import TransferRetryError\n+ from google.cloud.streaming import transfer as MUT\n+ from google.cloud.streaming.exceptions import TransferRetryError\n CONTENT = b'ABCDEFGHIJ'\n LEN = len(CONTENT)\n START = 5\n@@ -624,7 +624,7 @@ def test_get_range_w_empty_chunk(self):\n def test_get_range_w_total_size_wo_use_chunks(self):\n from six.moves import http_client\n from unit_tests._testing import _Monkey\n- from gcloud.streaming import transfer as MUT\n+ from google.cloud.streaming import transfer as MUT\n CONTENT = b'ABCDEFGHIJ'\n LEN = len(CONTENT)\n CHUNK_SIZE = 3\n@@ -652,7 +652,7 @@ def test_get_range_w_total_size_wo_use_chunks(self):\n def test_get_range_w_multiple_chunks(self):\n from six.moves import http_client\n from unit_tests._testing import _Monkey\n- from gcloud.streaming import transfer as MUT\n+ from google.cloud.streaming import transfer as MUT\n CONTENT = b'ABCDE'\n LEN = len(CONTENT)\n CHUNK_SIZE = 3\n@@ -686,7 +686,7 @@ def test_get_range_w_multiple_chunks(self):\n self.assertEqual(download.total_size, LEN)\n \n def test_stream_file_not_initialized(self):\n- from gcloud.streaming.exceptions import TransferInvalidError\n+ from google.cloud.streaming.exceptions import TransferInvalidError\n download = self._makeOne(_Stream())\n \n with self.assertRaises(TransferInvalidError):\n@@ -713,7 +713,7 @@ def test_stream_file_w_initial_response_complete(self):\n def test_stream_file_w_initial_response_incomplete(self):\n from six.moves import http_client\n from unit_tests._testing import _Monkey\n- from gcloud.streaming import transfer as MUT\n+ from google.cloud.streaming import transfer as MUT\n CHUNK_SIZE = 3\n CONTENT = b'ABCDEF'\n LEN = len(CONTENT)\n@@ -750,7 +750,7 @@ def test_stream_file_w_initial_response_incomplete(self):\n def test_stream_file_wo_initial_response_wo_total_size(self):\n from six.moves import http_client\n from unit_tests._testing import _Monkey\n- from gcloud.streaming import transfer as MUT\n+ from google.cloud.streaming import transfer as MUT\n CONTENT = b'ABCDEFGHIJ'\n LEN = len(CONTENT)\n CHUNK_SIZE = 123\n@@ -784,14 +784,14 @@ class Test_Upload(unittest.TestCase):\n UPLOAD_URL = 'http://example.com/upload/id=foobar'\n \n def _getTargetClass(self):\n- from gcloud.streaming.transfer import Upload\n+ from google.cloud.streaming.transfer import Upload\n return Upload\n \n def _makeOne(self, stream, mime_type=MIME_TYPE, *args, **kw):\n return self._getTargetClass()(stream, mime_type, *args, **kw)\n \n def test_ctor_defaults(self):\n- from gcloud.streaming.transfer import _DEFAULT_CHUNKSIZE\n+ from google.cloud.streaming.transfer import _DEFAULT_CHUNKSIZE\n stream = _Stream()\n upload = self._makeOne(stream)\n self.assertTrue(upload.stream is stream)\n@@ -901,19 +901,19 @@ def test_strategy_setter_invalid(self):\n upload.strategy = 'unknown'\n \n def test_strategy_setter_SIMPLE_UPLOAD(self):\n- from gcloud.streaming.transfer import SIMPLE_UPLOAD\n+ from google.cloud.streaming.transfer import SIMPLE_UPLOAD\n upload = self._makeOne(_Stream())\n upload.strategy = SIMPLE_UPLOAD\n self.assertEqual(upload.strategy, SIMPLE_UPLOAD)\n \n def test_strategy_setter_RESUMABLE_UPLOAD(self):\n- from gcloud.streaming.transfer import RESUMABLE_UPLOAD\n+ from google.cloud.streaming.transfer import RESUMABLE_UPLOAD\n upload = self._makeOne(_Stream())\n upload.strategy = RESUMABLE_UPLOAD\n self.assertEqual(upload.strategy, RESUMABLE_UPLOAD)\n \n def test_total_size_setter_initialized(self):\n- from gcloud.streaming.exceptions import TransferInvalidError\n+ from google.cloud.streaming.exceptions import TransferInvalidError\n SIZE = 123\n upload = self._makeOne(_Stream)\n http = object()\n@@ -928,7 +928,7 @@ def test_total_size_setter_not_initialized(self):\n self.assertEqual(upload.total_size, SIZE)\n \n def test__set_default_strategy_w_existing_strategy(self):\n- from gcloud.streaming.transfer import RESUMABLE_UPLOAD\n+ from google.cloud.streaming.transfer import RESUMABLE_UPLOAD\n config = _Dummy(\n resumable_path='/resumable/endpoint',\n simple_multipart=True,\n@@ -941,7 +941,7 @@ def test__set_default_strategy_w_existing_strategy(self):\n self.assertEqual(upload.strategy, RESUMABLE_UPLOAD)\n \n def test__set_default_strategy_wo_resumable_path(self):\n- from gcloud.streaming.transfer import SIMPLE_UPLOAD\n+ from google.cloud.streaming.transfer import SIMPLE_UPLOAD\n config = _Dummy(\n resumable_path=None,\n simple_multipart=True,\n@@ -953,8 +953,8 @@ def test__set_default_strategy_wo_resumable_path(self):\n self.assertEqual(upload.strategy, SIMPLE_UPLOAD)\n \n def test__set_default_strategy_w_total_size_gt_threshhold(self):\n- from gcloud.streaming.transfer import RESUMABLE_UPLOAD_THRESHOLD\n- from gcloud.streaming.transfer import RESUMABLE_UPLOAD\n+ from google.cloud.streaming.transfer import RESUMABLE_UPLOAD_THRESHOLD\n+ from google.cloud.streaming.transfer import RESUMABLE_UPLOAD\n config = _UploadConfig()\n request = _Request()\n upload = self._makeOne(\n@@ -963,7 +963,7 @@ def test__set_default_strategy_w_total_size_gt_threshhold(self):\n self.assertEqual(upload.strategy, RESUMABLE_UPLOAD)\n \n def test__set_default_strategy_w_body_wo_multipart(self):\n- from gcloud.streaming.transfer import RESUMABLE_UPLOAD\n+ from google.cloud.streaming.transfer import RESUMABLE_UPLOAD\n CONTENT = b'ABCDEFGHIJ'\n config = _UploadConfig()\n config.simple_multipart = False\n@@ -973,7 +973,7 @@ def test__set_default_strategy_w_body_wo_multipart(self):\n self.assertEqual(upload.strategy, RESUMABLE_UPLOAD)\n \n def test__set_default_strategy_w_body_w_multipart_wo_simple_path(self):\n- from gcloud.streaming.transfer import RESUMABLE_UPLOAD\n+ from google.cloud.streaming.transfer import RESUMABLE_UPLOAD\n CONTENT = b'ABCDEFGHIJ'\n config = _UploadConfig()\n config.simple_path = None\n@@ -983,7 +983,7 @@ def test__set_default_strategy_w_body_w_multipart_wo_simple_path(self):\n self.assertEqual(upload.strategy, RESUMABLE_UPLOAD)\n \n def test__set_default_strategy_w_body_w_multipart_w_simple_path(self):\n- from gcloud.streaming.transfer import SIMPLE_UPLOAD\n+ from google.cloud.streaming.transfer import SIMPLE_UPLOAD\n CONTENT = b'ABCDEFGHIJ'\n config = _UploadConfig()\n request = _Request(body=CONTENT)\n@@ -1011,7 +1011,7 @@ def test_configure_request_w_invalid_mimetype(self):\n upload.configure_request(config, request, url_builder)\n \n def test_configure_request_w_simple_wo_body(self):\n- from gcloud.streaming.transfer import SIMPLE_UPLOAD\n+ from google.cloud.streaming.transfer import SIMPLE_UPLOAD\n CONTENT = b'CONTENT'\n config = _UploadConfig()\n request = _Request()\n@@ -1029,8 +1029,8 @@ def test_configure_request_w_simple_wo_body(self):\n self.assertEqual(request.loggable_body, '')\n \n def test_configure_request_w_simple_w_body(self):\n- from gcloud._helpers import _to_bytes\n- from gcloud.streaming.transfer import SIMPLE_UPLOAD\n+ from google.cloud._helpers import _to_bytes\n+ from google.cloud.streaming.transfer import SIMPLE_UPLOAD\n CONTENT = b'CONTENT'\n BODY = b'BODY'\n config = _UploadConfig()\n@@ -1072,7 +1072,7 @@ def test_configure_request_w_simple_w_body(self):\n self.assertTrue(b'' in request.loggable_body)\n \n def test_configure_request_w_resumable_wo_total_size(self):\n- from gcloud.streaming.transfer import RESUMABLE_UPLOAD\n+ from google.cloud.streaming.transfer import RESUMABLE_UPLOAD\n CONTENT = b'CONTENT'\n config = _UploadConfig()\n request = _Request()\n@@ -1089,7 +1089,7 @@ def test_configure_request_w_resumable_wo_total_size(self):\n {'X-Upload-Content-Type': self.MIME_TYPE})\n \n def test_configure_request_w_resumable_w_total_size(self):\n- from gcloud.streaming.transfer import RESUMABLE_UPLOAD\n+ from google.cloud.streaming.transfer import RESUMABLE_UPLOAD\n CONTENT = b'CONTENT'\n LEN = len(CONTENT)\n config = _UploadConfig()\n@@ -1109,14 +1109,14 @@ def test_configure_request_w_resumable_w_total_size(self):\n 'X-Upload-Content-Length': '%d' % (LEN,)})\n \n def test_refresh_upload_state_w_simple_strategy(self):\n- from gcloud.streaming.transfer import SIMPLE_UPLOAD\n+ from google.cloud.streaming.transfer import SIMPLE_UPLOAD\n upload = self._makeOne(_Stream())\n upload.strategy = SIMPLE_UPLOAD\n upload.refresh_upload_state() # no-op\n \n def test_refresh_upload_state_not_initialized(self):\n- from gcloud.streaming.exceptions import TransferInvalidError\n- from gcloud.streaming.transfer import RESUMABLE_UPLOAD\n+ from google.cloud.streaming.exceptions import TransferInvalidError\n+ from google.cloud.streaming.transfer import RESUMABLE_UPLOAD\n upload = self._makeOne(_Stream())\n upload.strategy = RESUMABLE_UPLOAD\n with self.assertRaises(TransferInvalidError):\n@@ -1125,8 +1125,8 @@ def test_refresh_upload_state_not_initialized(self):\n def test_refresh_upload_state_w_OK(self):\n from six.moves import http_client\n from unit_tests._testing import _Monkey\n- from gcloud.streaming import transfer as MUT\n- from gcloud.streaming.transfer import RESUMABLE_UPLOAD\n+ from google.cloud.streaming import transfer as MUT\n+ from google.cloud.streaming.transfer import RESUMABLE_UPLOAD\n CONTENT = b'ABCDEFGHIJ'\n LEN = len(CONTENT)\n RESP_RANGE = 'bytes 0-%d/%d' % (LEN - 1, LEN,)\n@@ -1152,8 +1152,8 @@ def test_refresh_upload_state_w_OK(self):\n def test_refresh_upload_state_w_CREATED(self):\n from six.moves import http_client\n from unit_tests._testing import _Monkey\n- from gcloud.streaming import transfer as MUT\n- from gcloud.streaming.transfer import RESUMABLE_UPLOAD\n+ from google.cloud.streaming import transfer as MUT\n+ from google.cloud.streaming.transfer import RESUMABLE_UPLOAD\n CONTENT = b'ABCDEFGHIJ'\n LEN = len(CONTENT)\n RESP_RANGE = 'bytes 0-%d/%d' % (LEN - 1, LEN,)\n@@ -1177,10 +1177,10 @@ def test_refresh_upload_state_w_CREATED(self):\n self.assertTrue(upload._final_response is response)\n \n def test_refresh_upload_state_w_RESUME_INCOMPLETE_w_range(self):\n- from gcloud.streaming import transfer as MUT\n- from gcloud.streaming.http_wrapper import RESUME_INCOMPLETE\n+ from google.cloud.streaming import transfer as MUT\n+ from google.cloud.streaming.http_wrapper import RESUME_INCOMPLETE\n from unit_tests._testing import _Monkey\n- from gcloud.streaming.transfer import RESUMABLE_UPLOAD\n+ from google.cloud.streaming.transfer import RESUMABLE_UPLOAD\n CONTENT = b'ABCDEFGHIJ'\n LEN = len(CONTENT)\n LAST = 5\n@@ -1204,10 +1204,10 @@ def test_refresh_upload_state_w_RESUME_INCOMPLETE_w_range(self):\n self.assertFalse(upload._final_response is response)\n \n def test_refresh_upload_state_w_RESUME_INCOMPLETE_wo_range(self):\n- from gcloud.streaming import transfer as MUT\n- from gcloud.streaming.http_wrapper import RESUME_INCOMPLETE\n+ from google.cloud.streaming import transfer as MUT\n+ from google.cloud.streaming.http_wrapper import RESUME_INCOMPLETE\n from unit_tests._testing import _Monkey\n- from gcloud.streaming.transfer import RESUMABLE_UPLOAD\n+ from google.cloud.streaming.transfer import RESUMABLE_UPLOAD\n CONTENT = b'ABCDEFGHIJ'\n LEN = len(CONTENT)\n http = object()\n@@ -1231,9 +1231,9 @@ def test_refresh_upload_state_w_RESUME_INCOMPLETE_wo_range(self):\n def test_refresh_upload_state_w_error(self):\n from six.moves import http_client\n from unit_tests._testing import _Monkey\n- from gcloud.streaming import transfer as MUT\n- from gcloud.streaming.exceptions import HttpError\n- from gcloud.streaming.transfer import RESUMABLE_UPLOAD\n+ from google.cloud.streaming import transfer as MUT\n+ from google.cloud.streaming.exceptions import HttpError\n+ from google.cloud.streaming.transfer import RESUMABLE_UPLOAD\n CONTENT = b'ABCDEFGHIJ'\n LEN = len(CONTENT)\n http = object()\n@@ -1272,15 +1272,15 @@ def test_initialize_upload_no_strategy(self):\n upload.initialize_upload(request, http=object())\n \n def test_initialize_upload_simple_w_http(self):\n- from gcloud.streaming.transfer import SIMPLE_UPLOAD\n+ from google.cloud.streaming.transfer import SIMPLE_UPLOAD\n request = _Request()\n upload = self._makeOne(_Stream())\n upload.strategy = SIMPLE_UPLOAD\n upload.initialize_upload(request, http=object()) # no-op\n \n def test_initialize_upload_resumable_already_initialized(self):\n- from gcloud.streaming.exceptions import TransferInvalidError\n- from gcloud.streaming.transfer import RESUMABLE_UPLOAD\n+ from google.cloud.streaming.exceptions import TransferInvalidError\n+ from google.cloud.streaming.transfer import RESUMABLE_UPLOAD\n request = _Request()\n upload = self._makeOne(_Stream())\n upload.strategy = RESUMABLE_UPLOAD\n@@ -1291,9 +1291,9 @@ def test_initialize_upload_resumable_already_initialized(self):\n def test_initialize_upload_w_http_resumable_not_initialized_w_error(self):\n from six.moves import http_client\n from unit_tests._testing import _Monkey\n- from gcloud.streaming import transfer as MUT\n- from gcloud.streaming.exceptions import HttpError\n- from gcloud.streaming.transfer import RESUMABLE_UPLOAD\n+ from google.cloud.streaming import transfer as MUT\n+ from google.cloud.streaming.exceptions import HttpError\n+ from google.cloud.streaming.transfer import RESUMABLE_UPLOAD\n request = _Request()\n upload = self._makeOne(_Stream())\n upload.strategy = RESUMABLE_UPLOAD\n@@ -1307,8 +1307,8 @@ def test_initialize_upload_w_http_resumable_not_initialized_w_error(self):\n def test_initialize_upload_w_http_wo_auto_transfer_w_OK(self):\n from six.moves import http_client\n from unit_tests._testing import _Monkey\n- from gcloud.streaming import transfer as MUT\n- from gcloud.streaming.transfer import RESUMABLE_UPLOAD\n+ from google.cloud.streaming import transfer as MUT\n+ from google.cloud.streaming.transfer import RESUMABLE_UPLOAD\n request = _Request()\n upload = self._makeOne(_Stream(), auto_transfer=False)\n upload.strategy = RESUMABLE_UPLOAD\n@@ -1328,8 +1328,8 @@ def test_initialize_upload_w_http_wo_auto_transfer_w_OK(self):\n def test_initialize_upload_w_granularity_w_auto_transfer_w_OK(self):\n from six.moves import http_client\n from unit_tests._testing import _Monkey\n- from gcloud.streaming import transfer as MUT\n- from gcloud.streaming.transfer import RESUMABLE_UPLOAD\n+ from google.cloud.streaming import transfer as MUT\n+ from google.cloud.streaming.transfer import RESUMABLE_UPLOAD\n CONTENT = b'ABCDEFGHIJ'\n http = object()\n request = _Request()\n@@ -1377,14 +1377,14 @@ def test__validate_chunksize_w__server_chunk_granularity_hit(self):\n upload._validate_chunksize(400)\n \n def test_stream_file_w_simple_strategy(self):\n- from gcloud.streaming.transfer import SIMPLE_UPLOAD\n+ from google.cloud.streaming.transfer import SIMPLE_UPLOAD\n upload = self._makeOne(_Stream())\n upload.strategy = SIMPLE_UPLOAD\n with self.assertRaises(ValueError):\n upload.stream_file()\n \n def test_stream_file_w_use_chunks_invalid_chunk_size(self):\n- from gcloud.streaming.transfer import RESUMABLE_UPLOAD\n+ from google.cloud.streaming.transfer import RESUMABLE_UPLOAD\n upload = self._makeOne(_Stream(), chunksize=1024)\n upload.strategy = RESUMABLE_UPLOAD\n upload._server_chunk_granularity = 100\n@@ -1392,8 +1392,8 @@ def test_stream_file_w_use_chunks_invalid_chunk_size(self):\n upload.stream_file(use_chunks=True)\n \n def test_stream_file_not_initialized(self):\n- from gcloud.streaming.exceptions import TransferInvalidError\n- from gcloud.streaming.transfer import RESUMABLE_UPLOAD\n+ from google.cloud.streaming.exceptions import TransferInvalidError\n+ from google.cloud.streaming.transfer import RESUMABLE_UPLOAD\n upload = self._makeOne(_Stream(), chunksize=1024)\n upload.strategy = RESUMABLE_UPLOAD\n upload._server_chunk_granularity = 128\n@@ -1401,7 +1401,7 @@ def test_stream_file_not_initialized(self):\n upload.stream_file()\n \n def test_stream_file_already_complete_w_unseekable_stream(self):\n- from gcloud.streaming.transfer import RESUMABLE_UPLOAD\n+ from google.cloud.streaming.transfer import RESUMABLE_UPLOAD\n http = object()\n stream = object()\n response = object()\n@@ -1414,8 +1414,8 @@ def test_stream_file_already_complete_w_unseekable_stream(self):\n self.assertTrue(upload.stream_file() is response)\n \n def test_stream_file_already_complete_w_seekable_stream_unsynced(self):\n- from gcloud.streaming.exceptions import CommunicationError\n- from gcloud.streaming.transfer import RESUMABLE_UPLOAD\n+ from google.cloud.streaming.exceptions import CommunicationError\n+ from google.cloud.streaming.transfer import RESUMABLE_UPLOAD\n CONTENT = b'ABCDEFGHIJ'\n http = object()\n stream = _Stream(CONTENT)\n@@ -1431,7 +1431,7 @@ def test_stream_file_already_complete_w_seekable_stream_unsynced(self):\n \n def test_stream_file_already_complete_wo_seekable_method_synced(self):\n import os\n- from gcloud.streaming.transfer import RESUMABLE_UPLOAD\n+ from google.cloud.streaming.transfer import RESUMABLE_UPLOAD\n CONTENT = b'ABCDEFGHIJ'\n http = object()\n stream = _Stream(CONTENT)\n@@ -1447,7 +1447,7 @@ def test_stream_file_already_complete_wo_seekable_method_synced(self):\n \n def test_stream_file_already_complete_w_seekable_method_true_synced(self):\n import os\n- from gcloud.streaming.transfer import RESUMABLE_UPLOAD\n+ from google.cloud.streaming.transfer import RESUMABLE_UPLOAD\n CONTENT = b'ABCDEFGHIJ'\n http = object()\n stream = _StreamWithSeekableMethod(CONTENT, True)\n@@ -1463,7 +1463,7 @@ def test_stream_file_already_complete_w_seekable_method_true_synced(self):\n \n def test_stream_file_already_complete_w_seekable_method_false(self):\n import os\n- from gcloud.streaming.transfer import RESUMABLE_UPLOAD\n+ from google.cloud.streaming.transfer import RESUMABLE_UPLOAD\n CONTENT = b'ABCDEFGHIJ'\n http = object()\n stream = _StreamWithSeekableMethod(CONTENT, False)\n@@ -1480,9 +1480,9 @@ def test_stream_file_already_complete_w_seekable_method_false(self):\n def test_stream_file_incomplete(self):\n from six.moves import http_client\n from unit_tests._testing import _Monkey\n- from gcloud.streaming import transfer as MUT\n- from gcloud.streaming.http_wrapper import RESUME_INCOMPLETE\n- from gcloud.streaming.transfer import RESUMABLE_UPLOAD\n+ from google.cloud.streaming import transfer as MUT\n+ from google.cloud.streaming.http_wrapper import RESUME_INCOMPLETE\n+ from google.cloud.streaming.transfer import RESUMABLE_UPLOAD\n CONTENT = b'ABCDEFGHIJ'\n http = object()\n stream = _Stream(CONTENT)\n@@ -1524,10 +1524,10 @@ def test_stream_file_incomplete(self):\n \n def test_stream_file_incomplete_w_transfer_error(self):\n from unit_tests._testing import _Monkey\n- from gcloud.streaming import transfer as MUT\n- from gcloud.streaming.exceptions import CommunicationError\n- from gcloud.streaming.http_wrapper import RESUME_INCOMPLETE\n- from gcloud.streaming.transfer import RESUMABLE_UPLOAD\n+ from google.cloud.streaming import transfer as MUT\n+ from google.cloud.streaming.exceptions import CommunicationError\n+ from google.cloud.streaming.http_wrapper import RESUME_INCOMPLETE\n+ from google.cloud.streaming.transfer import RESUMABLE_UPLOAD\n CONTENT = b'ABCDEFGHIJ'\n http = object()\n stream = _Stream(CONTENT)\n@@ -1562,8 +1562,8 @@ def test_stream_file_incomplete_w_transfer_error(self):\n \n def test__send_media_request_wo_error(self):\n from unit_tests._testing import _Monkey\n- from gcloud.streaming import transfer as MUT\n- from gcloud.streaming.http_wrapper import RESUME_INCOMPLETE\n+ from google.cloud.streaming import transfer as MUT\n+ from google.cloud.streaming.http_wrapper import RESUME_INCOMPLETE\n CONTENT = b'ABCDEFGHIJ'\n bytes_http = object()\n stream = _Stream(CONTENT)\n@@ -1590,10 +1590,10 @@ def test__send_media_request_wo_error(self):\n def test__send_media_request_w_error(self):\n from six.moves import http_client\n from unit_tests._testing import _Monkey\n- from gcloud.streaming import transfer as MUT\n- from gcloud.streaming.exceptions import HttpError\n- from gcloud.streaming.http_wrapper import RESUME_INCOMPLETE\n- from gcloud.streaming.transfer import RESUMABLE_UPLOAD\n+ from google.cloud.streaming import transfer as MUT\n+ from google.cloud.streaming.exceptions import HttpError\n+ from google.cloud.streaming.http_wrapper import RESUME_INCOMPLETE\n+ from google.cloud.streaming.transfer import RESUMABLE_UPLOAD\n CONTENT = b'ABCDEFGHIJ'\n bytes_http = object()\n http = object()\n@@ -1629,13 +1629,13 @@ def test__send_media_request_w_error(self):\n self.assertTrue(second_http is http)\n \n def test__send_media_body_not_initialized(self):\n- from gcloud.streaming.exceptions import TransferInvalidError\n+ from google.cloud.streaming.exceptions import TransferInvalidError\n upload = self._makeOne(_Stream())\n with self.assertRaises(TransferInvalidError):\n upload._send_media_body(0)\n \n def test__send_media_body_wo_total_size(self):\n- from gcloud.streaming.exceptions import TransferInvalidError\n+ from google.cloud.streaming.exceptions import TransferInvalidError\n http = object()\n upload = self._makeOne(_Stream())\n upload._initialize(http, _Request.URL)\n@@ -1643,7 +1643,7 @@ def test__send_media_body_wo_total_size(self):\n upload._send_media_body(0)\n \n def test__send_media_body_start_lt_total_size(self):\n- from gcloud.streaming.stream_slice import StreamSlice\n+ from google.cloud.streaming.stream_slice import StreamSlice\n SIZE = 1234\n http = object()\n stream = _Stream()\n@@ -1670,7 +1670,7 @@ def test__send_media_body_start_lt_total_size(self):\n self.assertEqual(end, SIZE)\n \n def test__send_media_body_start_eq_total_size(self):\n- from gcloud.streaming.stream_slice import StreamSlice\n+ from google.cloud.streaming.stream_slice import StreamSlice\n SIZE = 1234\n http = object()\n stream = _Stream()\n@@ -1697,7 +1697,7 @@ def test__send_media_body_start_eq_total_size(self):\n self.assertEqual(end, SIZE)\n \n def test__send_chunk_not_initialized(self):\n- from gcloud.streaming.exceptions import TransferInvalidError\n+ from google.cloud.streaming.exceptions import TransferInvalidError\n upload = self._makeOne(_Stream())\n with self.assertRaises(TransferInvalidError):\n upload._send_chunk(0)\n@@ -1756,7 +1756,7 @@ def test__send_chunk_wo_total_size_stream_not_exhausted(self):\n self.assertEqual(end, CHUNK_SIZE)\n \n def test__send_chunk_w_total_size_stream_not_exhausted(self):\n- from gcloud.streaming.stream_slice import StreamSlice\n+ from google.cloud.streaming.stream_slice import StreamSlice\n CONTENT = b'ABCDEFGHIJ'\n SIZE = len(CONTENT)\n CHUNK_SIZE = SIZE - 5\n@@ -1787,7 +1787,7 @@ def test__send_chunk_w_total_size_stream_not_exhausted(self):\n self.assertEqual(end, CHUNK_SIZE)\n \n def test__send_chunk_w_total_size_stream_exhausted(self):\n- from gcloud.streaming.stream_slice import StreamSlice\n+ from google.cloud.streaming.stream_slice import StreamSlice\n CONTENT = b'ABCDEFGHIJ'\n SIZE = len(CONTENT)\n CHUNK_SIZE = 1000\ndiff --git a/unit_tests/streaming/test_util.py b/unit_tests/streaming/test_util.py\n--- a/unit_tests/streaming/test_util.py\n+++ b/unit_tests/streaming/test_util.py\n@@ -4,7 +4,7 @@\n class Test_calculate_wait_for_retry(unittest.TestCase):\n \n def _callFUT(self, *args, **kw):\n- from gcloud.streaming.util import calculate_wait_for_retry\n+ from google.cloud.streaming.util import calculate_wait_for_retry\n return calculate_wait_for_retry(*args, **kw)\n \n def test_w_negative_jitter_lt_max_wait(self):\n@@ -23,7 +23,7 @@ def test_w_positive_jitter_gt_max_wait(self):\n class Test_acceptable_mime_type(unittest.TestCase):\n \n def _callFUT(self, *args, **kw):\n- from gcloud.streaming.util import acceptable_mime_type\n+ from google.cloud.streaming.util import acceptable_mime_type\n return acceptable_mime_type(*args, **kw)\n \n def test_pattern_wo_slash(self):\ndiff --git a/unit_tests/test__helpers.py b/unit_tests/test__helpers.py\n--- a/unit_tests/test__helpers.py\n+++ b/unit_tests/test__helpers.py\n@@ -19,7 +19,7 @@\n class Test__LocalStack(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud._helpers import _LocalStack\n+ from google.cloud._helpers import _LocalStack\n \n return _LocalStack\n \n@@ -47,14 +47,14 @@ def test_it(self):\n class Test__UTC(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud._helpers import _UTC\n+ from google.cloud._helpers import _UTC\n return _UTC\n \n def _makeOne(self):\n return self._getTargetClass()()\n \n def test_module_property(self):\n- from gcloud import _helpers as MUT\n+ from google.cloud import _helpers as MUT\n klass = self._getTargetClass()\n try:\n import pytz\n@@ -100,7 +100,7 @@ def test___str__(self):\n class Test__ensure_tuple_or_list(unittest.TestCase):\n \n def _callFUT(self, arg_name, tuple_or_list):\n- from gcloud._helpers import _ensure_tuple_or_list\n+ from google.cloud._helpers import _ensure_tuple_or_list\n return _ensure_tuple_or_list(arg_name, tuple_or_list)\n \n def test_valid_tuple(self):\n@@ -127,12 +127,12 @@ def test_invalid_iterable(self):\n class Test__app_engine_id(unittest.TestCase):\n \n def _callFUT(self):\n- from gcloud._helpers import _app_engine_id\n+ from google.cloud._helpers import _app_engine_id\n return _app_engine_id()\n \n def test_no_value(self):\n from unit_tests._testing import _Monkey\n- from gcloud import _helpers\n+ from google.cloud import _helpers\n \n with _Monkey(_helpers, app_identity=None):\n dataset_id = self._callFUT()\n@@ -140,7 +140,7 @@ def test_no_value(self):\n \n def test_value_set(self):\n from unit_tests._testing import _Monkey\n- from gcloud import _helpers\n+ from google.cloud import _helpers\n \n APP_ENGINE_ID = object()\n APP_IDENTITY = _AppIdentity(APP_ENGINE_ID)\n@@ -152,11 +152,11 @@ def test_value_set(self):\n class Test__file_project_id(unittest.TestCase):\n \n def _callFUT(self):\n- from gcloud._helpers import _file_project_id\n+ from google.cloud._helpers import _file_project_id\n return _file_project_id()\n \n def test_success(self):\n- from gcloud.environment_vars import CREDENTIALS\n+ from google.cloud.environment_vars import CREDENTIALS\n from unit_tests._testing import _Monkey\n from unit_tests._testing import _NamedTemporaryFile\n \n@@ -185,11 +185,11 @@ def test_no_environment_variable_set(self):\n class Test__get_nix_config_path(unittest.TestCase):\n \n def _callFUT(self):\n- from gcloud._helpers import _get_nix_config_path\n+ from google.cloud._helpers import _get_nix_config_path\n return _get_nix_config_path()\n \n def test_it(self):\n- from gcloud import _helpers as MUT\n+ from google.cloud import _helpers as MUT\n from unit_tests._testing import _Monkey\n \n user_root = 'a'\n@@ -205,11 +205,11 @@ def test_it(self):\n class Test__get_windows_config_path(unittest.TestCase):\n \n def _callFUT(self):\n- from gcloud._helpers import _get_windows_config_path\n+ from google.cloud._helpers import _get_windows_config_path\n return _get_windows_config_path()\n \n def test_it(self):\n- from gcloud import _helpers as MUT\n+ from google.cloud import _helpers as MUT\n from unit_tests._testing import _Monkey\n \n appdata_dir = 'a'\n@@ -228,11 +228,11 @@ class Test__default_service_project_id(unittest.TestCase):\n CONFIG_TEMPLATE = '[%s]\\n%s = %s\\n'\n \n def _callFUT(self):\n- from gcloud._helpers import _default_service_project_id\n+ from google.cloud._helpers import _default_service_project_id\n return _default_service_project_id()\n \n def test_nix(self):\n- from gcloud import _helpers as MUT\n+ from google.cloud import _helpers as MUT\n from unit_tests._testing import _Monkey\n from unit_tests._testing import _NamedTemporaryFile\n \n@@ -255,7 +255,7 @@ def mock_get_path():\n self.assertEqual(result, project_id)\n \n def test_windows(self):\n- from gcloud import _helpers as MUT\n+ from google.cloud import _helpers as MUT\n from unit_tests._testing import _Monkey\n from unit_tests._testing import _NamedTemporaryFile\n \n@@ -278,7 +278,7 @@ def mock_get_path():\n self.assertEqual(result, project_id)\n \n def test_gae(self):\n- from gcloud import _helpers as MUT\n+ from google.cloud import _helpers as MUT\n from unit_tests._testing import _Monkey\n \n with _Monkey(os, name='not-nt'):\n@@ -291,12 +291,12 @@ def test_gae(self):\n class Test__compute_engine_id(unittest.TestCase):\n \n def _callFUT(self):\n- from gcloud._helpers import _compute_engine_id\n+ from google.cloud._helpers import _compute_engine_id\n return _compute_engine_id()\n \n def _monkeyConnection(self, connection):\n from unit_tests._testing import _Monkey\n- from gcloud import _helpers\n+ from google.cloud import _helpers\n \n def _connection_factory(host, timeout):\n connection.host = host\n@@ -328,7 +328,7 @@ def test_socket_raises(self):\n class Test__get_production_project(unittest.TestCase):\n \n def _callFUT(self):\n- from gcloud._helpers import _get_production_project\n+ from google.cloud._helpers import _get_production_project\n return _get_production_project()\n \n def test_no_value(self):\n@@ -341,7 +341,7 @@ def test_no_value(self):\n \n def test_value_set(self):\n from unit_tests._testing import _Monkey\n- from gcloud._helpers import PROJECT\n+ from google.cloud._helpers import PROJECT\n \n MOCK_PROJECT = object()\n environ = {PROJECT: MOCK_PROJECT}\n@@ -353,13 +353,13 @@ def test_value_set(self):\n class Test__determine_default_project(unittest.TestCase):\n \n def _callFUT(self, project=None):\n- from gcloud._helpers import _determine_default_project\n+ from google.cloud._helpers import _determine_default_project\n return _determine_default_project(project=project)\n \n def _determine_default_helper(self, prod=None, gae=None, gce=None,\n file_id=None, srv_id=None, project=None):\n from unit_tests._testing import _Monkey\n- from gcloud import _helpers\n+ from google.cloud import _helpers\n \n _callers = []\n \n@@ -432,12 +432,12 @@ def test_gce(self):\n class Test__millis(unittest.TestCase):\n \n def _callFUT(self, value):\n- from gcloud._helpers import _millis\n+ from google.cloud._helpers import _millis\n return _millis(value)\n \n def test_one_second_from_epoch(self):\n import datetime\n- from gcloud._helpers import UTC\n+ from google.cloud._helpers import UTC\n \n WHEN = datetime.datetime(1970, 1, 1, 0, 0, 1, tzinfo=UTC)\n self.assertEqual(self._callFUT(WHEN), 1000)\n@@ -446,7 +446,7 @@ def test_one_second_from_epoch(self):\n class Test__microseconds_from_datetime(unittest.TestCase):\n \n def _callFUT(self, value):\n- from gcloud._helpers import _microseconds_from_datetime\n+ from google.cloud._helpers import _microseconds_from_datetime\n return _microseconds_from_datetime(value)\n \n def test_it(self):\n@@ -463,7 +463,7 @@ def test_it(self):\n class Test__millis_from_datetime(unittest.TestCase):\n \n def _callFUT(self, value):\n- from gcloud._helpers import _millis_from_datetime\n+ from google.cloud._helpers import _millis_from_datetime\n return _millis_from_datetime(value)\n \n def test_w_none(self):\n@@ -472,8 +472,8 @@ def test_w_none(self):\n def test_w_utc_datetime(self):\n import datetime\n import six\n- from gcloud._helpers import UTC\n- from gcloud._helpers import _microseconds_from_datetime\n+ from google.cloud._helpers import UTC\n+ from google.cloud._helpers import _microseconds_from_datetime\n \n NOW = datetime.datetime.utcnow().replace(tzinfo=UTC)\n NOW_MICROS = _microseconds_from_datetime(NOW)\n@@ -485,8 +485,8 @@ def test_w_utc_datetime(self):\n def test_w_non_utc_datetime(self):\n import datetime\n import six\n- from gcloud._helpers import _UTC\n- from gcloud._helpers import _microseconds_from_datetime\n+ from google.cloud._helpers import _UTC\n+ from google.cloud._helpers import _microseconds_from_datetime\n \n class CET(_UTC):\n _tzname = 'CET'\n@@ -503,8 +503,8 @@ class CET(_UTC):\n def test_w_naive_datetime(self):\n import datetime\n import six\n- from gcloud._helpers import UTC\n- from gcloud._helpers import _microseconds_from_datetime\n+ from google.cloud._helpers import UTC\n+ from google.cloud._helpers import _microseconds_from_datetime\n \n NOW = datetime.datetime.utcnow()\n UTC_NOW = NOW.replace(tzinfo=UTC)\n@@ -518,13 +518,13 @@ def test_w_naive_datetime(self):\n class Test__datetime_from_microseconds(unittest.TestCase):\n \n def _callFUT(self, value):\n- from gcloud._helpers import _datetime_from_microseconds\n+ from google.cloud._helpers import _datetime_from_microseconds\n return _datetime_from_microseconds(value)\n \n def test_it(self):\n import datetime\n- from gcloud._helpers import UTC\n- from gcloud._helpers import _microseconds_from_datetime\n+ from google.cloud._helpers import UTC\n+ from google.cloud._helpers import _microseconds_from_datetime\n \n NOW = datetime.datetime(2015, 7, 29, 17, 45, 21, 123456,\n tzinfo=UTC)\n@@ -535,7 +535,7 @@ def test_it(self):\n class Test__rfc3339_to_datetime(unittest.TestCase):\n \n def _callFUT(self, dt_str):\n- from gcloud._helpers import _rfc3339_to_datetime\n+ from google.cloud._helpers import _rfc3339_to_datetime\n return _rfc3339_to_datetime(dt_str)\n \n def test_w_bogus_zone(self):\n@@ -554,7 +554,7 @@ def test_w_bogus_zone(self):\n \n def test_w_microseconds(self):\n import datetime\n- from gcloud._helpers import UTC\n+ from google.cloud._helpers import UTC\n \n year = 2009\n month = 12\n@@ -589,7 +589,7 @@ def test_w_naonseconds(self):\n class Test__rfc3339_nanos_to_datetime(unittest.TestCase):\n \n def _callFUT(self, dt_str):\n- from gcloud._helpers import _rfc3339_nanos_to_datetime\n+ from google.cloud._helpers import _rfc3339_nanos_to_datetime\n return _rfc3339_nanos_to_datetime(dt_str)\n \n def test_w_bogus_zone(self):\n@@ -608,7 +608,7 @@ def test_w_bogus_zone(self):\n \n def test_w_truncated_nanos(self):\n import datetime\n- from gcloud._helpers import UTC\n+ from google.cloud._helpers import UTC\n \n year = 2009\n month = 12\n@@ -637,7 +637,7 @@ def test_w_truncated_nanos(self):\n \n def test_w_naonseconds(self):\n import datetime\n- from gcloud._helpers import UTC\n+ from google.cloud._helpers import UTC\n \n year = 2009\n month = 12\n@@ -659,12 +659,12 @@ def test_w_naonseconds(self):\n class Test__datetime_to_rfc3339(unittest.TestCase):\n \n def _callFUT(self, *args, **kwargs):\n- from gcloud._helpers import _datetime_to_rfc3339\n+ from google.cloud._helpers import _datetime_to_rfc3339\n return _datetime_to_rfc3339(*args, **kwargs)\n \n @staticmethod\n def _make_timezone(offset):\n- from gcloud._helpers import _UTC\n+ from google.cloud._helpers import _UTC\n \n class CET(_UTC):\n _tzname = 'CET'\n@@ -674,7 +674,7 @@ class CET(_UTC):\n \n def test_w_utc_datetime(self):\n import datetime\n- from gcloud._helpers import UTC\n+ from google.cloud._helpers import UTC\n \n TIMESTAMP = datetime.datetime(2016, 4, 5, 13, 30, 0, tzinfo=UTC)\n result = self._callFUT(TIMESTAMP, ignore_zone=False)\n@@ -682,7 +682,7 @@ def test_w_utc_datetime(self):\n \n def test_w_non_utc_datetime(self):\n import datetime\n- from gcloud._helpers import _UTC\n+ from google.cloud._helpers import _UTC\n \n zone = self._make_timezone(offset=datetime.timedelta(hours=-1))\n TIMESTAMP = datetime.datetime(2016, 4, 5, 13, 30, 0, tzinfo=zone)\n@@ -691,7 +691,7 @@ def test_w_non_utc_datetime(self):\n \n def test_w_non_utc_datetime_and_ignore_zone(self):\n import datetime\n- from gcloud._helpers import _UTC\n+ from google.cloud._helpers import _UTC\n \n zone = self._make_timezone(offset=datetime.timedelta(hours=-1))\n TIMESTAMP = datetime.datetime(2016, 4, 5, 13, 30, 0, tzinfo=zone)\n@@ -709,7 +709,7 @@ def test_w_naive_datetime(self):\n class Test__to_bytes(unittest.TestCase):\n \n def _callFUT(self, *args, **kwargs):\n- from gcloud._helpers import _to_bytes\n+ from google.cloud._helpers import _to_bytes\n return _to_bytes(*args, **kwargs)\n \n def test_with_bytes(self):\n@@ -736,7 +736,7 @@ def test_with_nonstring_type(self):\n class Test__bytes_to_unicode(unittest.TestCase):\n \n def _callFUT(self, *args, **kwargs):\n- from gcloud._helpers import _bytes_to_unicode\n+ from google.cloud._helpers import _bytes_to_unicode\n return _bytes_to_unicode(*args, **kwargs)\n \n def test_with_bytes(self):\n@@ -757,13 +757,13 @@ def test_with_nonstring_type(self):\n class Test__pb_timestamp_to_datetime(unittest.TestCase):\n \n def _callFUT(self, timestamp):\n- from gcloud._helpers import _pb_timestamp_to_datetime\n+ from google.cloud._helpers import _pb_timestamp_to_datetime\n return _pb_timestamp_to_datetime(timestamp)\n \n def test_it(self):\n import datetime\n from google.protobuf.timestamp_pb2 import Timestamp\n- from gcloud._helpers import UTC\n+ from google.cloud._helpers import UTC\n \n # Epoch is midnight on January 1, 1970 ...\n dt_stamp = datetime.datetime(1970, month=1, day=1, hour=0,\n@@ -778,7 +778,7 @@ def test_it(self):\n class Test__pb_timestamp_to_rfc3339(unittest.TestCase):\n \n def _callFUT(self, timestamp):\n- from gcloud._helpers import _pb_timestamp_to_rfc3339\n+ from google.cloud._helpers import _pb_timestamp_to_rfc3339\n return _pb_timestamp_to_rfc3339(timestamp)\n \n def test_it(self):\n@@ -795,13 +795,13 @@ def test_it(self):\n class Test__datetime_to_pb_timestamp(unittest.TestCase):\n \n def _callFUT(self, when):\n- from gcloud._helpers import _datetime_to_pb_timestamp\n+ from google.cloud._helpers import _datetime_to_pb_timestamp\n return _datetime_to_pb_timestamp(when)\n \n def test_it(self):\n import datetime\n from google.protobuf.timestamp_pb2 import Timestamp\n- from gcloud._helpers import UTC\n+ from google.cloud._helpers import UTC\n \n # Epoch is midnight on January 1, 1970 ...\n dt_stamp = datetime.datetime(1970, month=1, day=1, hour=0,\n@@ -820,7 +820,7 @@ class Test__name_from_project_path(unittest.TestCase):\n TEMPLATE = r'projects/(?P\\w+)/things/(?P\\w+)'\n \n def _callFUT(self, path, project, template):\n- from gcloud._helpers import _name_from_project_path\n+ from google.cloud._helpers import _name_from_project_path\n return _name_from_project_path(path, project, template)\n \n def test_w_invalid_path_length(self):\n@@ -858,7 +858,7 @@ def test_w_project_passed_as_none(self):\n class TestMetadataPlugin(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud._helpers import MetadataPlugin\n+ from google.cloud._helpers import MetadataPlugin\n return MetadataPlugin\n \n def _makeOne(self, *args, **kwargs):\n@@ -894,12 +894,12 @@ def callback(*args):\n class Test_make_stub(unittest.TestCase):\n \n def _callFUT(self, *args, **kwargs):\n- from gcloud._helpers import make_stub\n+ from google.cloud._helpers import make_stub\n return make_stub(*args, **kwargs)\n \n def test_it(self):\n from unit_tests._testing import _Monkey\n- from gcloud import _helpers as MUT\n+ from google.cloud import _helpers as MUT\n \n mock_result = object()\n stub_inputs = []\n@@ -972,7 +972,7 @@ def mock_plugin(*args):\n class Test_exc_to_code(unittest.TestCase):\n \n def _callFUT(self, exc):\n- from gcloud._helpers import exc_to_code\n+ from google.cloud._helpers import exc_to_code\n return exc_to_code(exc)\n \n def test_with_stable(self):\ndiff --git a/unit_tests/test_client.py b/unit_tests/test_client.py\n--- a/unit_tests/test_client.py\n+++ b/unit_tests/test_client.py\n@@ -18,7 +18,7 @@\n class Test_ClientFactoryMixin(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.client import _ClientFactoryMixin\n+ from google.cloud.client import _ClientFactoryMixin\n return _ClientFactoryMixin\n \n def test_virtual(self):\n@@ -38,7 +38,7 @@ def tearDown(self):\n KLASS._connection_class = self.original_cnxn_class\n \n def _getTargetClass(self):\n- from gcloud.client import Client\n+ from google.cloud.client import Client\n return Client\n \n def _makeOne(self, *args, **kw):\n@@ -46,7 +46,7 @@ def _makeOne(self, *args, **kw):\n \n def test_ctor_defaults(self):\n from unit_tests._testing import _Monkey\n- from gcloud import client\n+ from google.cloud import client\n \n CREDENTIALS = object()\n FUNC_CALLS = []\n@@ -73,7 +73,7 @@ def test_ctor_explicit(self):\n \n def test_from_service_account_json(self):\n from unit_tests._testing import _Monkey\n- from gcloud import client\n+ from google.cloud import client\n \n KLASS = self._getTargetClass()\n MOCK_FILENAME = 'foo.path'\n@@ -93,7 +93,7 @@ def test_from_service_account_json_fail(self):\n \n def test_from_service_account_p12(self):\n from unit_tests._testing import _Monkey\n- from gcloud import client\n+ from google.cloud import client\n \n KLASS = self._getTargetClass()\n CLIENT_EMAIL = 'phred@example.com'\n@@ -127,7 +127,7 @@ def tearDown(self):\n KLASS._connection_class = self.original_cnxn_class\n \n def _getTargetClass(self):\n- from gcloud.client import JSONClient\n+ from google.cloud.client import JSONClient\n return JSONClient\n \n def _makeOne(self, *args, **kw):\n@@ -135,7 +135,7 @@ def _makeOne(self, *args, **kw):\n \n def test_ctor_defaults(self):\n from unit_tests._testing import _Monkey\n- from gcloud import client\n+ from google.cloud import client\n \n PROJECT = 'PROJECT'\n CREDENTIALS = object()\n@@ -162,7 +162,7 @@ def mock_get_credentials():\n \n def test_ctor_missing_project(self):\n from unit_tests._testing import _Monkey\n- from gcloud import client\n+ from google.cloud import client\n \n FUNC_CALLS = []\n \ndiff --git a/unit_tests/test_connection.py b/unit_tests/test_connection.py\n--- a/unit_tests/test_connection.py\n+++ b/unit_tests/test_connection.py\n@@ -18,7 +18,7 @@\n class TestConnection(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.connection import Connection\n+ from google.cloud.connection import Connection\n return Connection\n \n def _makeOne(self, *args, **kw):\n@@ -69,8 +69,8 @@ def test_http_w_creds(self):\n \n def test_user_agent_format(self):\n from pkg_resources import get_distribution\n- expected_ua = 'gcloud-python/{0}'.format(\n- get_distribution('gcloud').version)\n+ expected_ua = 'google-cloud-python/{0}'.format(\n+ get_distribution('google-cloud').version)\n conn = self._makeOne()\n self.assertEqual(conn.USER_AGENT, expected_ua)\n \n@@ -78,7 +78,7 @@ def test_user_agent_format(self):\n class TestJSONConnection(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.connection import JSONConnection\n+ from google.cloud.connection import JSONConnection\n return JSONConnection\n \n def _makeOne(self, *args, **kw):\n@@ -316,7 +316,7 @@ def test_api_request_w_data(self):\n self.assertEqual(http._called_with['headers'], expected_headers)\n \n def test_api_request_w_404(self):\n- from gcloud.exceptions import NotFound\n+ from google.cloud.exceptions import NotFound\n conn = self._makeMockOne()\n conn._http = _Http(\n {'status': '404', 'content-type': 'text/plain'},\n@@ -325,7 +325,7 @@ def test_api_request_w_404(self):\n self.assertRaises(NotFound, conn.api_request, 'GET', '/')\n \n def test_api_request_w_500(self):\n- from gcloud.exceptions import InternalServerError\n+ from google.cloud.exceptions import InternalServerError\n conn = self._makeMockOne()\n conn._http = _Http(\n {'status': '500', 'content-type': 'text/plain'},\ndiff --git a/unit_tests/test_credentials.py b/unit_tests/test_credentials.py\n--- a/unit_tests/test_credentials.py\n+++ b/unit_tests/test_credentials.py\n@@ -18,12 +18,12 @@\n class Test_get_credentials(unittest.TestCase):\n \n def _callFUT(self):\n- from gcloud import credentials\n+ from google.cloud import credentials\n return credentials.get_credentials()\n \n def test_it(self):\n from unit_tests._testing import _Monkey\n- from gcloud import credentials as MUT\n+ from google.cloud import credentials as MUT\n \n client = _Client()\n with _Monkey(MUT, client=client):\n@@ -36,7 +36,7 @@ def test_it(self):\n class Test_generate_signed_url(unittest.TestCase):\n \n def _callFUT(self, *args, **kwargs):\n- from gcloud.credentials import generate_signed_url\n+ from google.cloud.credentials import generate_signed_url\n return generate_signed_url(*args, **kwargs)\n \n def _generate_helper(self, response_type=None, response_disposition=None,\n@@ -45,7 +45,7 @@ def _generate_helper(self, response_type=None, response_disposition=None,\n from six.moves.urllib.parse import parse_qs\n from six.moves.urllib.parse import urlsplit\n from unit_tests._testing import _Monkey\n- from gcloud import credentials as MUT\n+ from google.cloud import credentials as MUT\n \n ENDPOINT = 'http://api.example.com'\n RESOURCE = '/name/path'\n@@ -104,7 +104,7 @@ def test_w_custom_fields(self):\n class Test_generate_signed_url_exception(unittest.TestCase):\n def test_with_google_credentials(self):\n import time\n- from gcloud.credentials import generate_signed_url\n+ from google.cloud.credentials import generate_signed_url\n RESOURCE = '/name/path'\n \n credentials = _GoogleCredentials()\n@@ -116,7 +116,7 @@ def test_with_google_credentials(self):\n class Test__get_signed_query_params(unittest.TestCase):\n \n def _callFUT(self, credentials, expiration, string_to_sign):\n- from gcloud.credentials import _get_signed_query_params\n+ from google.cloud.credentials import _get_signed_query_params\n return _get_signed_query_params(credentials, expiration,\n string_to_sign)\n \n@@ -143,7 +143,7 @@ def test_it(self):\n class Test__get_expiration_seconds(unittest.TestCase):\n \n def _callFUT(self, expiration):\n- from gcloud.credentials import _get_expiration_seconds\n+ from google.cloud.credentials import _get_expiration_seconds\n return _get_expiration_seconds(expiration)\n \n def _utc_seconds(self, when):\n@@ -174,7 +174,7 @@ def test_w_naive_datetime(self):\n \n def test_w_utc_datetime(self):\n import datetime\n- from gcloud._helpers import UTC\n+ from google.cloud._helpers import UTC\n \n expiration_utc = datetime.datetime(2004, 8, 19, 0, 0, 0, 0, UTC)\n utc_seconds = self._utc_seconds(expiration_utc)\n@@ -182,7 +182,7 @@ def test_w_utc_datetime(self):\n \n def test_w_other_zone_datetime(self):\n import datetime\n- from gcloud._helpers import _UTC\n+ from google.cloud._helpers import _UTC\n \n class CET(_UTC):\n _tzname = 'CET'\n@@ -197,7 +197,7 @@ class CET(_UTC):\n def test_w_timedelta_seconds(self):\n import datetime\n from unit_tests._testing import _Monkey\n- from gcloud import credentials as MUT\n+ from google.cloud import credentials as MUT\n \n dummy_utcnow = datetime.datetime(2004, 8, 19, 0, 0, 0, 0)\n utc_seconds = self._utc_seconds(dummy_utcnow)\n@@ -211,7 +211,7 @@ def test_w_timedelta_seconds(self):\n def test_w_timedelta_days(self):\n import datetime\n from unit_tests._testing import _Monkey\n- from gcloud import credentials as MUT\n+ from google.cloud import credentials as MUT\n \n dummy_utcnow = datetime.datetime(2004, 8, 19, 0, 0, 0, 0)\n utc_seconds = self._utc_seconds(dummy_utcnow)\ndiff --git a/unit_tests/test_exceptions.py b/unit_tests/test_exceptions.py\n--- a/unit_tests/test_exceptions.py\n+++ b/unit_tests/test_exceptions.py\n@@ -15,11 +15,11 @@\n import unittest\n \n \n-class Test_GCloudError(unittest.TestCase):\n+class Test_GoogleCloudError(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.exceptions import GCloudError\n- return GCloudError\n+ from google.cloud.exceptions import GoogleCloudError\n+ return GoogleCloudError\n \n def _makeOne(self, message, errors=()):\n return self._getTargetClass()(message, errors=errors)\n@@ -49,12 +49,12 @@ def test_ctor_explicit(self):\n class Test_make_exception(unittest.TestCase):\n \n def _callFUT(self, response, content, error_info=None, use_json=True):\n- from gcloud.exceptions import make_exception\n+ from google.cloud.exceptions import make_exception\n return make_exception(response, content, error_info=error_info,\n use_json=use_json)\n \n def test_hit_w_content_as_str(self):\n- from gcloud.exceptions import NotFound\n+ from google.cloud.exceptions import NotFound\n response = _Response(404)\n content = b'{\"error\": {\"message\": \"Not Found\"}}'\n exception = self._callFUT(response, content)\n@@ -63,7 +63,7 @@ def test_hit_w_content_as_str(self):\n self.assertEqual(list(exception.errors), [])\n \n def test_miss_w_content_as_dict(self):\n- from gcloud.exceptions import GCloudError\n+ from google.cloud.exceptions import GoogleCloudError\n ERROR = {\n 'domain': 'global',\n 'location': 'test',\n@@ -74,12 +74,12 @@ def test_miss_w_content_as_dict(self):\n response = _Response(600)\n content = {\"error\": {\"message\": \"Unknown Error\", \"errors\": [ERROR]}}\n exception = self._callFUT(response, content)\n- self.assertTrue(isinstance(exception, GCloudError))\n+ self.assertTrue(isinstance(exception, GoogleCloudError))\n self.assertEqual(exception.message, 'Unknown Error')\n self.assertEqual(list(exception.errors), [ERROR])\n \n def test_html_when_json_expected(self):\n- from gcloud.exceptions import NotFound\n+ from google.cloud.exceptions import NotFound\n response = _Response(NotFound.code)\n content = '404 Not Found'\n exception = self._callFUT(response, content, use_json=True)\ndiff --git a/unit_tests/test_iterator.py b/unit_tests/test_iterator.py\n--- a/unit_tests/test_iterator.py\n+++ b/unit_tests/test_iterator.py\n@@ -18,7 +18,7 @@\n class TestIterator(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.iterator import Iterator\n+ from google.cloud.iterator import Iterator\n return Iterator\n \n def _makeOne(self, *args, **kw):\n@@ -175,7 +175,7 @@ def test_get_items_from_response_raises_NotImplementedError(self):\n class TestMethodIterator(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.iterator import MethodIterator\n+ from google.cloud.iterator import MethodIterator\n return MethodIterator\n \n def _makeOne(self, *args, **kw):\ndiff --git a/unit_tests/test_operation.py b/unit_tests/test_operation.py\n--- a/unit_tests/test_operation.py\n+++ b/unit_tests/test_operation.py\n@@ -18,14 +18,14 @@\n class Test__compute_type_url(unittest.TestCase):\n \n def _callFUT(self, klass, prefix=None):\n- from gcloud.operation import _compute_type_url\n+ from google.cloud.operation import _compute_type_url\n if prefix is None:\n return _compute_type_url(klass)\n return _compute_type_url(klass, prefix)\n \n def test_wo_prefix(self):\n from google.protobuf.struct_pb2 import Struct\n- from gcloud.operation import _GOOGLE_APIS_PREFIX\n+ from google.cloud.operation import _GOOGLE_APIS_PREFIX\n \n type_url = self._callFUT(Struct)\n \n@@ -35,7 +35,7 @@ def test_wo_prefix(self):\n \n def test_w_prefix(self):\n from google.protobuf.struct_pb2 import Struct\n- PREFIX = 'test.gcloud-python.com'\n+ PREFIX = 'test.google-cloud-python.com'\n \n type_url = self._callFUT(Struct, PREFIX)\n \n@@ -47,13 +47,13 @@ def test_w_prefix(self):\n class Test__register_type_url(unittest.TestCase):\n \n def _callFUT(self, type_url, klass):\n- from gcloud.operation import _register_type_url\n+ from google.cloud.operation import _register_type_url\n _register_type_url(type_url, klass)\n \n def test_simple(self):\n- from gcloud import operation as MUT\n+ from google.cloud import operation as MUT\n from unit_tests._testing import _Monkey\n- TYPE_URI = 'testing.gcloud-python.com/testing'\n+ TYPE_URI = 'testing.google-cloud-python.com/testing'\n klass = object()\n type_url_map = {}\n \n@@ -63,9 +63,9 @@ def test_simple(self):\n self.assertEqual(type_url_map, {TYPE_URI: klass})\n \n def test_w_same_class(self):\n- from gcloud import operation as MUT\n+ from google.cloud import operation as MUT\n from unit_tests._testing import _Monkey\n- TYPE_URI = 'testing.gcloud-python.com/testing'\n+ TYPE_URI = 'testing.google-cloud-python.com/testing'\n klass = object()\n type_url_map = {TYPE_URI: klass}\n \n@@ -75,9 +75,9 @@ def test_w_same_class(self):\n self.assertEqual(type_url_map, {TYPE_URI: klass})\n \n def test_w_conflict(self):\n- from gcloud import operation as MUT\n+ from google.cloud import operation as MUT\n from unit_tests._testing import _Monkey\n- TYPE_URI = 'testing.gcloud-python.com/testing'\n+ TYPE_URI = 'testing.google-cloud-python.com/testing'\n klass, other = object(), object()\n type_url_map = {TYPE_URI: other}\n \n@@ -93,7 +93,7 @@ class OperationTests(unittest.TestCase):\n OPERATION_NAME = 'operations/projects/foo/instances/bar/operations/123'\n \n def _getTargetClass(self):\n- from gcloud.operation import Operation\n+ from google.cloud.operation import Operation\n return Operation\n \n def _makeOne(self, *args, **kw):\n@@ -157,7 +157,7 @@ def test_from_pb_w_metadata_and_kwargs(self):\n from google.longrunning import operations_pb2\n from google.protobuf.any_pb2 import Any\n from google.protobuf.struct_pb2 import Struct, Value\n- from gcloud import operation as MUT\n+ from google.cloud import operation as MUT\n from unit_tests._testing import _Monkey\n TYPE_URI = 'type.googleapis.com/%s' % (Struct.DESCRIPTOR.full_name,)\n type_url_map = {TYPE_URI: Struct}\ndiff --git a/unit_tests/translate/test_client.py b/unit_tests/translate/test_client.py\n--- a/unit_tests/translate/test_client.py\n+++ b/unit_tests/translate/test_client.py\n@@ -20,15 +20,15 @@ class TestClient(unittest.TestCase):\n KEY = 'abc-123-my-key'\n \n def _getTargetClass(self):\n- from gcloud.translate.client import Client\n+ from google.cloud.translate.client import Client\n return Client\n \n def _makeOne(self, *args, **kw):\n return self._getTargetClass()(*args, **kw)\n \n def test_ctor(self):\n- from gcloud.translate.connection import Connection\n- from gcloud.translate.client import ENGLISH_ISO_639\n+ from google.cloud.translate.connection import Connection\n+ from google.cloud.translate.client import ENGLISH_ISO_639\n \n http = object()\n client = self._makeOne(self.KEY, http=http)\n@@ -38,7 +38,7 @@ def test_ctor(self):\n self.assertEqual(client.target_language, ENGLISH_ISO_639)\n \n def test_ctor_non_default(self):\n- from gcloud.translate.connection import Connection\n+ from google.cloud.translate.connection import Connection\n \n http = object()\n target = 'es'\n@@ -49,7 +49,7 @@ def test_ctor_non_default(self):\n self.assertEqual(client.target_language, target)\n \n def test_get_languages(self):\n- from gcloud.translate.client import ENGLISH_ISO_639\n+ from google.cloud.translate.client import ENGLISH_ISO_639\n \n client = self._makeOne(self.KEY)\n supported = [\ndiff --git a/unit_tests/translate/test_connection.py b/unit_tests/translate/test_connection.py\n--- a/unit_tests/translate/test_connection.py\n+++ b/unit_tests/translate/test_connection.py\n@@ -18,7 +18,7 @@\n class TestConnection(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.translate.connection import Connection\n+ from google.cloud.translate.connection import Connection\n return Connection\n \n def _makeOne(self, *args, **kw):\ndiff --git a/unit_tests/vision/test_client.py b/unit_tests/vision/test_client.py\n--- a/unit_tests/vision/test_client.py\n+++ b/unit_tests/vision/test_client.py\n@@ -16,7 +16,7 @@\n import base64\n import unittest\n \n-from gcloud._helpers import _to_bytes\n+from google.cloud._helpers import _to_bytes\n \n \n class TestClient(unittest.TestCase):\n@@ -26,7 +26,7 @@ class TestClient(unittest.TestCase):\n B64_IMAGE_CONTENT = base64.b64encode(IMAGE_CONTENT)\n \n def _getTargetClass(self):\n- from gcloud.vision.client import Client\n+ from google.cloud.vision.client import Client\n return Client\n \n def _makeOne(self, *args, **kw):\n@@ -39,9 +39,9 @@ def test_ctor(self):\n self.assertTrue('annotate' in dir(client))\n \n def test_face_annotation(self):\n+ from google.cloud.vision._fixtures import FACE_DETECTION_RESPONSE\n \n- from gcloud.vision._fixtures import FACE_DETECTION_RESPONSE as RETURNED\n-\n+ RETURNED = FACE_DETECTION_RESPONSE\n REQUEST = {\n \"requests\": [\n {\n@@ -61,7 +61,7 @@ def test_face_annotation(self):\n client = self._makeOne(project=self.PROJECT, credentials=credentials)\n client.connection = _Connection(RETURNED)\n \n- from gcloud.vision.feature import Feature, FeatureTypes\n+ from google.cloud.vision.feature import Feature, FeatureTypes\n \n features = [Feature(feature_type=FeatureTypes.FACE_DETECTION,\n max_results=3)]\n@@ -78,14 +78,14 @@ class TestVisionRequest(unittest.TestCase):\n _IMAGE_CONTENT = _to_bytes('/9j/4QNURXhpZgAASUkq')\n \n def _getTargetClass(self):\n- from gcloud.vision.client import VisionRequest\n+ from google.cloud.vision.client import VisionRequest\n return VisionRequest\n \n def _makeOne(self, *args, **kw):\n return self._getTargetClass()(*args, **kw)\n \n def test_make_vision_request(self):\n- from gcloud.vision.feature import Feature, FeatureTypes\n+ from google.cloud.vision.feature import Feature, FeatureTypes\n feature = Feature(feature_type=FeatureTypes.FACE_DETECTION,\n max_results=3)\n vision_request = self._makeOne(self._IMAGE_CONTENT, feature)\ndiff --git a/unit_tests/vision/test_connection.py b/unit_tests/vision/test_connection.py\n--- a/unit_tests/vision/test_connection.py\n+++ b/unit_tests/vision/test_connection.py\n@@ -18,7 +18,7 @@\n class TestConnection(unittest.TestCase):\n \n def _getTargetClass(self):\n- from gcloud.vision.connection import Connection\n+ from google.cloud.vision.connection import Connection\n return Connection\n \n def _makeOne(self, *args, **kw):\ndiff --git a/unit_tests/vision/test_feature.py b/unit_tests/vision/test_feature.py\n--- a/unit_tests/vision/test_feature.py\n+++ b/unit_tests/vision/test_feature.py\n@@ -17,14 +17,14 @@\n \n class TestFeature(unittest.TestCase):\n def _getTargetClass(self):\n- from gcloud.vision.feature import Feature\n+ from google.cloud.vision.feature import Feature\n return Feature\n \n def _makeOne(self, *args, **kw):\n return self._getTargetClass()(*args, **kw)\n \n def test_construct_feature(self):\n- from gcloud.vision.feature import FeatureTypes\n+ from google.cloud.vision.feature import FeatureTypes\n feature = self._makeOne(FeatureTypes.LABEL_DETECTION)\n self.assertEqual(1, feature.max_results)\n self.assertEqual('LABEL_DETECTION', feature.feature_type)\n@@ -34,7 +34,7 @@ def test_construct_feature(self):\n self.assertEqual('FACE_DETECTION', feature.feature_type)\n \n def test_feature_as_dict(self):\n- from gcloud.vision.feature import FeatureTypes\n+ from google.cloud.vision.feature import FeatureTypes\n feature = self._makeOne(FeatureTypes.FACE_DETECTION, max_results=5)\n EXPECTED = {\n 'type': 'FACE_DETECTION',\ndiff --git a/unit_tests/vision/test_image.py b/unit_tests/vision/test_image.py\n--- a/unit_tests/vision/test_image.py\n+++ b/unit_tests/vision/test_image.py\n@@ -15,7 +15,7 @@\n import unittest\n import base64\n \n-from gcloud._helpers import _to_bytes\n+from google.cloud._helpers import _to_bytes\n \n \n class TestVisionImage(unittest.TestCase):\n@@ -25,7 +25,7 @@ class TestVisionImage(unittest.TestCase):\n _CLIENT_MOCK = {'source': ''}\n \n def _getTargetClass(self):\n- from gcloud.vision.image import Image\n+ from google.cloud.vision.image import Image\n return Image\n \n def _makeOne(self, *args, **kw):\n@@ -42,7 +42,7 @@ def test_image_source_type_content(self):\n self.assertEqual(None, image.source)\n self.assertEqual(_AS_DICT, image.as_dict())\n \n- def test_image_source_type_gcloud_storage(self):\n+ def test_image_source_type_google_cloud_storage(self):\n image = self._makeOne(self._IMAGE_SOURCE, self._CLIENT_MOCK)\n \n _AS_DICT = {\n", "problem_statement": "", "hints_text": "", "created_at": "2016-08-30T18:42:57Z"}