Stay organized with collections Save and categorize content based on your preferences.
Client(
project=None,
credentials=None,
_http=None,
location=None,
default_query_job_config=None,
client_info=None,
client_options=None,
)
Client to bundle configuration needed for API requests.
Parameters Name Description projectOptional[str]
Project ID for the project which the client acts on behalf of. Will be passed when creating a dataset / job. If not passed, falls back to the default inferred from the environment.
credentialsOptional[google.auth.credentials.Credentials]
The OAuth2 Credentials to use for this client. If not passed (and if no _http
object is passed), falls back to the default inferred from the environment.
Optional[requests.Session]
HTTP object to make requests. Can be any object that defines request()
with the same interface as requests.Session.request
. If not passed, an _http
object is created that is bound to the credentials
for the current object. This parameter should be considered private, and could change in the future.
Optional[str]
Default location for jobs / datasets / tables.
default_query_job_configOptional[google.cloud.bigquery.job.QueryJobConfig]
Default QueryJobConfig
. Will be merged into job configs passed into the query
method.
Optional[google.api_core.client_info.ClientInfo]
The client info used to send a user-agent string along with API requests. If None
, then default info will be used. Generally, you only need to set this if you're developing your own library or partner tool.
Optional[Union[google.api_core.client_options.ClientOptions, Dict]]
Client options used to set user options on the client. API Endpoint should be set through client_options.
Inheritancebuiltins.object > google.cloud.client._ClientFactoryMixin > google.cloud.client.Client > builtins.object > google.cloud.client._ClientProjectMixin > google.cloud.client.ClientWithProject > Client Properties locationDefault location for jobs / datasets / tables.
Methods __getstate__Explicitly state that clients are not pickleable.
cancel_jobcancel_job(job_id, project=None, location=None, retry=<google.api_core.retry.Retry object>, timeout=None)
Parameter Name Description job_id str :keyword project: ID of the project which owns the job (defaults to the client's project). :kwtype project: Optional[str] :keyword location: Location where the job was run. :kwtype location: Optional[str] :keyword retry: How to retry the RPC. :kwtype retry: Optional[google.api_core.retry.Retry] :keyword timeout: The number of seconds to wait for the underlying HTTP transport before using ``retry``. :kwtype timeout: Optional[float]
Unique job identifier.
Returns Type Description Union[ google.cloud.bigquery.job.LoadJob, google.cloud.bigquery.job.CopyJob, google.cloud.bigquery.job.ExtractJob, google.cloud.bigquery.job.QueryJob, ] Job instance, based on the resource returned by the API. closeClose the underlying transport objects, releasing system resources.
Note: The client instance can be used for making additional requests even after closing, in which case the underlying connections are automatically re-created. copy_tablecopy_table(sources, destination, job_id=None, job_id_prefix=None, location=None, project=None, job_config=None, retry=<google.api_core.retry.Retry object>, timeout=None)
Parameters Name Description sources Union[ google.cloud.bigquery.table.Table, google.cloud.bigquery.table.TableReference, str, Sequence[ Union[ google.cloud.bigquery.table.Table, google.cloud.bigquery.table.TableReference, str, ] ], ]
Table or tables to be copied.
destinationUnion[ google.cloud.bigquery.table.Table, google.cloud.bigquery.table.TableReference, str, ] :keyword job_id: The ID of the job. :kwtype job_id: Optional[str] :keyword job_id_prefix: The user-provided prefix for a randomly generated job ID. This parameter will be ignored if a ``job_id`` is also given. :kwtype job_id_prefix: Optional[str] :keyword location: Location where to run the job. Must match the location of any source table as well as the destination table. :kwtype location: Optional[str] :keyword project: Project ID of the project of where to run the job. Defaults to the client's project. :kwtype project: Optional[str] :keyword job_config: Extra configuration options for the job. :kwtype job_config: Optional[google.cloud.bigquery.job.CopyJobConfig] :keyword retry: How to retry the RPC. :kwtype retry: Optional[google.api_core.retry.Retry] :keyword timeout: The number of seconds to wait for the underlying HTTP transport before using ``retry``. :kwtype timeout: Optional[float]
Table into which data is to be copied.
Exceptions Type Description TypeError If ``job_config`` is not an instance of CopyJobConfig class. Returns create_datasetcreate_dataset(dataset, exists_ok=False, retry=<google.api_core.retry.Retry object>, timeout=None)
Parameters Name Description dataset Union[ google.cloud.bigquery.dataset.Dataset, google.cloud.bigquery.dataset.DatasetReference, str, ]
A Dataset to create. If dataset
is a reference, an empty dataset is created with the specified ID and client's default location.
Optional[bool]
Defaults to False
. If True
, ignore "already exists" errors when creating the dataset.
Optional[google.api_core.retry.Retry]
How to retry the RPC.
timeoutOptional[float]
The number of seconds to wait for the underlying HTTP transport before using retry
.
create_job(job_config, retry=<google.api_core.retry.Retry object>, timeout=None)
Create a new job.
Parameter Name Description job_configdict :keyword retry: How to retry the RPC. :kwtype retry: Optional[google.api_core.retry.Retry] :keyword timeout: The number of seconds to wait for the underlying HTTP transport before using ``retry``. :kwtype timeout: Optional[float]
configuration job representation returned from the API.
Returns Type Description Union[ google.cloud.bigquery.job.LoadJob, google.cloud.bigquery.job.CopyJob, google.cloud.bigquery.job.ExtractJob, google.cloud.bigquery.job.QueryJob ] A new job instance. create_routinecreate_routine(routine, exists_ok=False, retry=<google.api_core.retry.Retry object>, timeout=None)
Parameters Name Description routine google.cloud.bigquery.routine.Routine
A Routine to create. The dataset that the routine belongs to must already exist.
exists_okOptional[bool]
Defaults to False
. If True
, ignore "already exists" errors when creating the routine.
Optional[google.api_core.retry.Retry]
How to retry the RPC.
timeoutOptional[float]
The number of seconds to wait for the underlying HTTP transport before using retry
.
create_table(table, exists_ok=False, retry=<google.api_core.retry.Retry object>, timeout=None)
Parameters Name Description table Union[ google.cloud.bigquery.table.Table, google.cloud.bigquery.table.TableReference, str, ]
A Table to create. If table
is a reference, an empty table is created with the specified ID. The dataset that the table belongs to must already exist.
Optional[bool]
Defaults to False
. If True
, ignore "already exists" errors when creating the table.
Optional[google.api_core.retry.Retry]
How to retry the RPC.
timeoutOptional[float]
The number of seconds to wait for the underlying HTTP transport before using retry
.
dataset(dataset_id, project=None)
Deprecated: Construct a reference to a dataset.
deprecated: Construct a xref_DatasetReference using its constructor or use a string where previously a reference object was used.As of
google-cloud-bigquery
version 1.7.0, all client methods that take a xref_DatasetReference or xref_TableReference also take a string in standard SQL format, e.g.
project.dataset_id
or
project.dataset_id.table_id
.
Parameters Name Description dataset_idstr
ID of the dataset.
projectOptional[str]
Project ID for the dataset (defaults to the project of the client).
Returns delete_datasetdelete_dataset(dataset, delete_contents=False, retry=<google.api_core.retry.Retry object>, timeout=None, not_found_ok=False)
Parameters Name Description dataset Union[ google.cloud.bigquery.dataset.Dataset, google.cloud.bigquery.dataset.DatasetReference, str, ]
A reference to the dataset to delete. If a string is passed in, this method attempts to create a dataset reference from a string using from_string.
delete_contentsOptional[bool]
If True, delete all the tables in the dataset. If False and the dataset contains tables, the request will fail. Default is False.
retryOptional[google.api_core.retry.Retry]
How to retry the RPC.
timeoutOptional[float]
The number of seconds to wait for the underlying HTTP transport before using retry
.
Optional[bool]
Defaults to False
. If True
, ignore "not found" errors when deleting the dataset.
delete_model(model, retry=<google.api_core.retry.Retry object>, timeout=None, not_found_ok=False)
Parameters Name Description model Union[ google.cloud.bigquery.model.Model, google.cloud.bigquery.model.ModelReference, str, ]
A reference to the model to delete. If a string is passed in, this method attempts to create a model reference from a string using from_string.
retryOptional[google.api_core.retry.Retry]
How to retry the RPC.
timeoutOptional[float]
The number of seconds to wait for the underlying HTTP transport before using retry
.
Optional[bool]
Defaults to False
. If True
, ignore "not found" errors when deleting the model.
delete_routine(routine, retry=<google.api_core.retry.Retry object>, timeout=None, not_found_ok=False)
Parameters Name Description routine Union[ google.cloud.bigquery.routine.Routine, google.cloud.bigquery.routine.RoutineReference, str, ]
A reference to the routine to delete. If a string is passed in, this method attempts to create a routine reference from a string using from_string.
retryOptional[google.api_core.retry.Retry]
How to retry the RPC.
timeoutOptional[float]
The number of seconds to wait for the underlying HTTP transport before using retry
.
Optional[bool]
Defaults to False
. If True
, ignore "not found" errors when deleting the routine.
delete_table(table, retry=<google.api_core.retry.Retry object>, timeout=None, not_found_ok=False)
Parameters Name Description table Union[ google.cloud.bigquery.table.Table, google.cloud.bigquery.table.TableReference, str, ]
A reference to the table to delete. If a string is passed in, this method attempts to create a table reference from a string using from_string.
retryOptional[google.api_core.retry.Retry]
How to retry the RPC.
timeoutOptional[float]
The number of seconds to wait for the underlying HTTP transport before using retry
.
Optional[bool]
Defaults to False
. If True
, ignore "not found" errors when deleting the table.
extract_table(source, destination_uris, job_id=None, job_id_prefix=None, location=None, project=None, job_config=None, retry=<google.api_core.retry.Retry object>, timeout=None, source_type='Table')
Parameters Name Description source Union[ google.cloud.bigquery.table.Table, google.cloud.bigquery.table.TableReference, google.cloud.bigquery.model.Model, google.cloud.bigquery.model.ModelReference, src, ]
Table or Model to be extracted.
destination_urisUnion[str, Sequence[str]] :keyword job_id: The ID of the job. :kwtype job_id: Optional[str] :keyword job_id_prefix: The user-provided prefix for a randomly generated job ID. This parameter will be ignored if a ``job_id`` is also given. :kwtype job_id_prefix: Optional[str] :keyword location: Location where to run the job. Must match the location of the source table. :kwtype location: Optional[str] :keyword project: Project ID of the project of where to run the job. Defaults to the client's project. :kwtype project: Optional[str] :keyword job_config: Extra configuration options for the job. :kwtype job_config: Optional[google.cloud.bigquery.job.ExtractJobConfig] :keyword retry: How to retry the RPC. :kwtype retry: Optional[google.api_core.retry.Retry] :keyword timeout: The number of seconds to wait for the underlying HTTP transport before using ``retry``. :kwtype timeout: Optional[float] :keyword source_type: Type of source to be extracted.``Table`` or ``Model``. Defaults to ``Table``. :kwtype source_type: Optional[str]
URIs of Cloud Storage file(s) into which table data is to be extracted; in format gs://<bucket_name>/<object_name_or_glob>
.
from_service_account_info(info, *args, **kwargs)
Factory to retrieve JSON credentials while creating client.
Parameters Name Description argstuple
Remaining positional arguments to pass to constructor.
infodict
The JSON object with a private key and other credentials information (downloaded from the Google APIs console).
Exceptions Type Description TypeError if there is a conflict with the kwargs and the credentials created by the factory. Returns Type Description `_ClientFactoryMixin` The client created with the retrieved JSON credentials. from_service_account_jsonfrom_service_account_json(json_credentials_path, *args, **kwargs)
Factory to retrieve JSON credentials while creating client.
Parameters Name Description argstuple
Remaining positional arguments to pass to constructor.
json_credentials_pathstr
The path to a private key file (this file was given to you when you created the service account). This file must contain a JSON object with a private key and other credentials information (downloaded from the Google APIs console).
Exceptions Type Description TypeError if there is a conflict with the kwargs and the credentials created by the factory. Returns Type Description `_ClientFactoryMixin` The client created with the retrieved JSON credentials. get_datasetget_dataset(dataset_ref, retry=<google.api_core.retry.Retry object>, timeout=None)
Fetch the dataset referenced by dataset_ref
Union[ google.cloud.bigquery.dataset.DatasetReference, str, ]
A reference to the dataset to fetch from the BigQuery API. If a string is passed in, this method attempts to create a dataset reference from a string using from_string.
retryOptional[google.api_core.retry.Retry]
How to retry the RPC.
timeoutOptional[float]
The number of seconds to wait for the underlying HTTP transport before using retry
.
get_job(job_id, project=None, location=None, retry=<google.api_core.retry.Retry object>, timeout=None)
Parameter Name Description job_id str :keyword project: ID of the project which owns the job (defaults to the client's project). :kwtype project: Optional[str] :keyword location: Location where the job was run. :kwtype location: Optional[str] :keyword retry: How to retry the RPC. :kwtype retry: Optional[google.api_core.retry.Retry] :keyword timeout: The number of seconds to wait for the underlying HTTP transport before using ``retry``. :kwtype timeout: Optional[float]
Unique job identifier.
Returns Type Description Union[ google.cloud.bigquery.job.LoadJob, google.cloud.bigquery.job.CopyJob, google.cloud.bigquery.job.ExtractJob, google.cloud.bigquery.job.QueryJob ] Job instance, based on the resource returned by the API. get_modelget_model(model_ref, retry=<google.api_core.retry.Retry object>, timeout=None)
[Beta] Fetch the model referenced by model_ref
.
Union[ google.cloud.bigquery.model.ModelReference, str, ]
A reference to the model to fetch from the BigQuery API. If a string is passed in, this method attempts to create a model reference from a string using from_string.
retryOptional[google.api_core.retry.Retry]
How to retry the RPC.
timeoutOptional[float]
The number of seconds to wait for the underlying HTTP transport before using retry
.
get_routine(routine_ref, retry=<google.api_core.retry.Retry object>, timeout=None)
[Beta] Get the routine referenced by routine_ref
.
Union[ google.cloud.bigquery.routine.Routine, google.cloud.bigquery.routine.RoutineReference, str, ]
A reference to the routine to fetch from the BigQuery API. If a string is passed in, this method attempts to create a reference from a string using from_string.
retryOptional[google.api_core.retry.Retry]
How to retry the API call.
timeoutOptional[float]
The number of seconds to wait for the underlying HTTP transport before using retry
.
get_service_account_email(project=None, retry=<google.api_core.retry.Retry object>, timeout=None)
Get the email address of the project's BigQuery service account
Note: This is the service account that BigQuery uses to manage tables encrypted by a key in KMS. Parameters Name Description projectOptional[str]
Project ID to use for retreiving service account email. Defaults to the client's project.
retryOptional[google.api_core.retry.Retry]
How to retry the RPC.
timeoutOptional[float]
The number of seconds to wait for the underlying HTTP transport before using retry
.
get_table(table, retry=<google.api_core.retry.Retry object>, timeout=None)
Fetch the table referenced by table
.
Union[ google.cloud.bigquery.table.Table, google.cloud.bigquery.table.TableReference, str, ]
A reference to the table to fetch from the BigQuery API. If a string is passed in, this method attempts to create a table reference from a string using from_string.
retryOptional[google.api_core.retry.Retry]
How to retry the RPC.
timeoutOptional[float]
The number of seconds to wait for the underlying HTTP transport before using retry
.
insert_rows(table, rows, selected_fields=None, **kwargs)
Parameters Name Description kwargs Dict
Keyword arguments to insert_rows_json.
tableUnion[ google.cloud.bigquery.table.Table, google.cloud.bigquery.table.TableReference, str, ]
The destination table for the row data, or a reference to it.
rowsUnion[Sequence[Tuple], Sequence[Dict]]
Row data to be inserted. If a list of tuples is given, each tuple should contain data for each schema field on the current table and in the same order as the schema fields. If a list of dictionaries is given, the keys must include all required fields in the schema. Keys which do not correspond to a field in the schema are ignored.
selected_fieldsSequence[google.cloud.bigquery.schema.SchemaField]
The fields to return. Required if table
is a TableReference.
insert_rows_from_dataframe(
table, dataframe, selected_fields=None, chunk_size=500, **kwargs
)
Insert rows into a table from a dataframe via the streaming API.
Parameters Name Description kwargsDict
Keyword arguments to insert_rows_json.
tableUnion[ google.cloud.bigquery.table.Table, google.cloud.bigquery.table.TableReference, str, ]
The destination table for the row data, or a reference to it.
dataframepandas.DataFrame
A pandas.DataFrame
containing the data to load. Any NaN
values present in the dataframe are omitted from the streaming API request(s).
Sequence[google.cloud.bigquery.schema.SchemaField]
The fields to return. Required if table
is a TableReference.
int
The number of rows to stream in a single chunk. Must be positive.
Exceptions Type Description ValueError if table's schema is not set Returns Type Description Sequence[Sequence[Mappings]] A list with insert errors for each insert chunk. Each element is a list containing one mapping per row with insert errors: the "index" key identifies the row, and the "errors" key contains a list of the mappings describing one or more problems with the row. insert_rows_jsoninsert_rows_json(table, json_rows, row_ids=None, skip_invalid_rows=None, ignore_unknown_values=None, template_suffix=None, retry=<google.api_core.retry.Retry object>, timeout=None)
Parameters Name Description table Union[ google.cloud.bigquery.table.Table google.cloud.bigquery.table.TableReference, str ]
The destination table for the row data, or a reference to it.
json_rowsSequence[Dict]
Row data to be inserted. Keys must match the table schema fields and values must be JSON-compatible representations.
row_idsOptional[Sequence[Optional[str]]]
Unique IDs, one per row being inserted. An ID can also be None
, indicating that an explicit insert ID should not be used for that row. If the argument is omitted altogether, unique IDs are created automatically.
Optional[bool]
Insert all valid rows of a request, even if invalid rows exist. The default value is False
, which causes the entire request to fail if any invalid rows exist.
Optional[bool]
Accept rows that contain values that do not match the schema. The unknown values are ignored. Default is False
, which treats unknown values as errors.
Optional[str]
Treat name
as a template table and provide a suffix. BigQuery will create the table <name> + <template_suffix>
based on the schema of the template table. See https://cloud.google.com/bigquery/streaming-data-into-bigquery#template-tables
Optional[google.api_core.retry.Retry]
How to retry the RPC.
timeoutOptional[float]
The number of seconds to wait for the underlying HTTP transport before using retry
.
job_from_resource(resource)
Detect correct job type from resource and instantiate.
Parameter Name Description resourceDict
one job resource from API response
Returns Type Description Union[ google.cloud.bigquery.job.LoadJob, google.cloud.bigquery.job.CopyJob, google.cloud.bigquery.job.ExtractJob, google.cloud.bigquery.job.QueryJob ] The job instance, constructed via the resource. list_datasetslist_datasets(project=None, include_all=False, filter=None, max_results=None, page_token=None, retry=<google.api_core.retry.Retry object>, timeout=None)
Parameters Name Description project Optional[str]
Project ID to use for retreiving datasets. Defaults to the client's project.
include_allOptional[bool]
True if results include hidden datasets. Defaults to False.
filterOptional[str]
An expression for filtering the results by label. For syntax, see https://cloud.google.com/bigquery/docs/reference/rest/v2/datasets/list#body.QUERY_PARAMETERS.filter
max_resultsOptional[int]
Maximum number of datasets to return.
page_tokenOptional[str]
Token representing a cursor into the datasets. If not passed, the API will return the first page of datasets. The token marks the beginning of the iterator to be returned and the value of the page_token
can be accessed at next_page_token
of the google.api_core.page_iterator.HTTPIterator
.
Optional[google.api_core.retry.Retry]
How to retry the RPC.
timeoutOptional[float]
The number of seconds to wait for the underlying HTTP transport before using retry
.
list_jobs(project=None, parent_job=None, max_results=None, page_token=None, all_users=None, state_filter=None, retry=<google.api_core.retry.Retry object>, timeout=None, min_creation_time=None, max_creation_time=None)
Parameters Name Description project Optional[str]
Project ID to use for retreiving datasets. Defaults to the client's project.
parent_jobOptional[Union[ google.cloud.bigquery.job._AsyncJob, str, ]]
If set, retrieve only child jobs of the specified parent.
max_resultsOptional[int]
Maximum number of jobs to return.
page_tokenOptional[str]
Opaque marker for the next "page" of jobs. If not passed, the API will return the first page of jobs. The token marks the beginning of the iterator to be returned and the value of the page_token
can be accessed at next_page_token
of google.api_core.page_iterator.HTTPIterator
.
Optional[bool]
If true, include jobs owned by all users in the project. Defaults to :data:False
.
Optional[str]
If set, include only jobs matching the given state. One of: * "done"
* "pending"
* "running"
Optional[google.api_core.retry.Retry]
How to retry the RPC.
timeoutOptional[float]
The number of seconds to wait for the underlying HTTP transport before using retry
.
Optional[datetime.datetime]
Min value for job creation time. If set, only jobs created after or at this timestamp are returned. If the datetime has no time zone assumes UTC time.
max_creation_timeOptional[datetime.datetime]
Max value for job creation time. If set, only jobs created before or at this timestamp are returned. If the datetime has no time zone assumes UTC time.
Returns Type Description google.api_core.page_iterator.Iterator Iterable of job instances. list_modelslist_models(dataset, max_results=None, page_token=None, retry=<google.api_core.retry.Retry object>, timeout=None)
Parameters Name Description dataset Union[ google.cloud.bigquery.dataset.Dataset, google.cloud.bigquery.dataset.DatasetReference, str, ]
A reference to the dataset whose models to list from the BigQuery API. If a string is passed in, this method attempts to create a dataset reference from a string using from_string.
max_resultsOptional[int]
Maximum number of models to return. If not passed, defaults to a value set by the API.
page_tokenOptional[str]
Token representing a cursor into the models. If not passed, the API will return the first page of models. The token marks the beginning of the iterator to be returned and the value of the page_token
can be accessed at next_page_token
of the google.api_core.page_iterator.HTTPIterator
.
Optional[google.api_core.retry.Retry]
How to retry the RPC.
timeoutOptional[float] Returns: google.api_core.page_iterator.Iterator: Iterator of Model contained within the requested dataset.
The number of seconds to wait for the underlying HTTP transport before using retry
.
list_partitions(table, retry=<google.api_core.retry.Retry object>, timeout=None)
List the partitions in a table.
Parameters Name Description tableUnion[ google.cloud.bigquery.table.Table, google.cloud.bigquery.table.TableReference, str, ]
The table or reference from which to get partition info
retryOptional[google.api_core.retry.Retry]
How to retry the RPC.
timeoutOptional[float]
The number of seconds to wait for the underlying HTTP transport before using retry
. If multiple requests are made under the hood, timeout
applies to each individual request.
list_projects(max_results=None, page_token=None, retry=<google.api_core.retry.Retry object>, timeout=None)
Parameters Name Description max_results Optional[int]
Maximum number of projects to return, If not passed, defaults to a value set by the API.
page_tokenOptional[str]
Token representing a cursor into the projects. If not passed, the API will return the first page of projects. The token marks the beginning of the iterator to be returned and the value of the page_token
can be accessed at next_page_token
of the google.api_core.page_iterator.HTTPIterator
.
Optional[google.api_core.retry.Retry]
How to retry the RPC.
timeoutOptional[float]
The number of seconds to wait for the underlying HTTP transport before using retry
.
list_routines(dataset, max_results=None, page_token=None, retry=<google.api_core.retry.Retry object>, timeout=None)
Parameters Name Description dataset Union[ google.cloud.bigquery.dataset.Dataset, google.cloud.bigquery.dataset.DatasetReference, str, ]
A reference to the dataset whose routines to list from the BigQuery API. If a string is passed in, this method attempts to create a dataset reference from a string using from_string.
max_resultsOptional[int]
Maximum number of routines to return. If not passed, defaults to a value set by the API.
page_tokenOptional[str]
Token representing a cursor into the routines. If not passed, the API will return the first page of routines. The token marks the beginning of the iterator to be returned and the value of the page_token
can be accessed at next_page_token
of the google.api_core.page_iterator.HTTPIterator
.
Optional[google.api_core.retry.Retry]
How to retry the RPC.
timeoutOptional[float] Returns: google.api_core.page_iterator.Iterator: Iterator of all Routines contained within the requested dataset, limited by ``max_results``.
The number of seconds to wait for the underlying HTTP transport before using retry
.
list_rows(table, selected_fields=None, max_results=None, page_token=None, start_index=None, page_size=None, retry=<google.api_core.retry.Retry object>, timeout=None)
List the rows of the table.
See https://cloud.google.com/bigquery/docs/reference/rest/v2/tabledata/list
Note: This method assumes that the provided schema is up-to-date with the schema as defined on the back-end: if the two schemas are not identical, the values returned may be incomplete. To ensure that the local copy of the schema is up-to-date, callclient.get_table
. Parameters Name Description table Union[ google.cloud.bigquery.table.Table, google.cloud.bigquery.table.TableListItem, google.cloud.bigquery.table.TableReference, str, ]
The table to list, or a reference to it. When the table object does not contain a schema and selected_fields
is not supplied, this method calls get_table
to fetch the table schema.
Sequence[google.cloud.bigquery.schema.SchemaField]
The fields to return. If not supplied, data for all columns are downloaded.
max_resultsOptional[int]
Maximum number of rows to return.
page_tokenOptional[str]
Token representing a cursor into the table's rows. If not passed, the API will return the first page of the rows. The token marks the beginning of the iterator to be returned and the value of the page_token
can be accessed at next_page_token
of the RowIterator.
Optional[int]
The zero-based index of the starting row to read.
page_sizeOptional[int]
The maximum number of rows in each page of results from this request. Non-positive values are ignored. Defaults to a sensible value set by the API.
retryOptional[google.api_core.retry.Retry]
How to retry the RPC.
timeoutOptional[float]
The number of seconds to wait for the underlying HTTP transport before using retry
. If multiple requests are made under the hood, timeout
applies to each individual request.
list_tables(dataset, max_results=None, page_token=None, retry=<google.api_core.retry.Retry object>, timeout=None)
Parameters Name Description dataset Union[ google.cloud.bigquery.dataset.Dataset, google.cloud.bigquery.dataset.DatasetReference, str, ]
A reference to the dataset whose tables to list from the BigQuery API. If a string is passed in, this method attempts to create a dataset reference from a string using from_string.
max_resultsOptional[int]
Maximum number of tables to return. If not passed, defaults to a value set by the API.
page_tokenOptional[str]
Token representing a cursor into the tables. If not passed, the API will return the first page of tables. The token marks the beginning of the iterator to be returned and the value of the page_token
can be accessed at next_page_token
of the google.api_core.page_iterator.HTTPIterator
.
Optional[google.api_core.retry.Retry]
How to retry the RPC.
timeoutOptional[float]
The number of seconds to wait for the underlying HTTP transport before using retry
.
load_table_from_dataframe(
dataframe,
destination,
num_retries=6,
job_id=None,
job_id_prefix=None,
location=None,
project=None,
job_config=None,
parquet_compression="snappy",
timeout=None,
)
Upload the contents of a table from a pandas DataFrame.
Similar to load_table_from_uri
, this method creates, starts and returns a xref_LoadJob.
parquet
file, a mismatch with the existing table schema can occur, so 100% compatibility cannot be guaranteed for REPEATED fields when using the parquet format. https://github.com/googleapis/python-bigquery/issues/17 Parameters Name Description dataframe pandas.DataFrame
A pandas.DataFrame
containing the data to load.
google.cloud.bigquery.table.TableReference :keyword num_retries: Number of upload retries. :kwtype num_retries: Optional[int] :keyword job_id: Name of the job. :kwtype job_id: Optional[str] :keyword job_id_prefix: The user-provided prefix for a randomly generated job ID. This parameter will be ignored if a ``job_id`` is also given. :kwtype job_id_prefix: Optional[str] :keyword location: Location where to run the job. Must match the location of the destination table. :kwtype location: Optional[str] :keyword project: Project ID of the project of where to run the job. Defaults to the client's project. :kwtype project: Optional[str] :keyword job_config: Extra configuration options for the job. To override the default pandas data type conversions, supply a value for schema with column names matching those of the dataframe. The BigQuery schema is used to determine the correct data type conversion. Indexes are not loaded. Requires the `pyarrow` library. By default, this method uses the parquet source format. To override this, supply a value for source_format with the format name. Currently only CSV and PARQUET are supported. :kwtype job_config: Optional[google.cloud.bigquery.job.LoadJobConfig] :keyword parquet_compression: [Beta] The compression method to use if intermittently serializing ``dataframe`` to a parquet file. The argument is directly passed as the ``compression`` argument to the underlying ``pyarrow.parquet.write_table()`` method (the default value "snappy" gets converted to uppercase). https://arrow.apache.org/docs/python/generated/pyarrow.parquet.write_table.html#pyarrow-parquet-write-table If the job config schema is missing, the argument is directly passed as the ``compression`` argument to the underlying ``DataFrame.to_parquet()`` method. https://pandas.pydata.org/pandas-docs/stable/reference/api/pandas.DataFrame.to_parquet.html#pandas.DataFrame.to_parquet :kwtype parquet_compression: Optional[str] :keyword timeout: The number of seconds to wait for the underlying HTTP transport before using ``retry``. :kwtype timeout: Optional[float]
The destination table to use for loading the data. If it is an existing table, the schema of the pandas.DataFrame
must match the schema of the destination table. If the table does not yet exist, the schema is inferred from the pandas.DataFrame
. If a string is passed in, this method attempts to create a table reference from a string using from_string.
load_table_from_file(
file_obj,
destination,
rewind=False,
size=None,
num_retries=6,
job_id=None,
job_id_prefix=None,
location=None,
project=None,
job_config=None,
timeout=None,
)
Upload the contents of this table from a file-like object.
Similar to load_table_from_uri
, this method creates, starts and returns a xref_LoadJob.
file
A file handle opened in binary mode for reading.
destinationUnion[ google.cloud.bigquery.table.Table, google.cloud.bigquery.table.TableReference, str, ] :keyword rewind: If True, seek to the beginning of the file handle before reading the file. :kwtype rewind: Optional[bool] :keyword size: The number of bytes to read from the file handle. If size is ``None`` or large, resumable upload will be used. Otherwise, multipart upload will be used. :kwtype size: Optional[int] :keyword num_retries: Number of upload retries. Defaults to 6. :kwtype num_retries: Optional[int] :keyword job_id: Name of the job. :kwtype job_id: Optional[str] :keyword job_id_prefix: The user-provided prefix for a randomly generated job ID. This parameter will be ignored if a ``job_id`` is also given. :kwtype job_id_prefix: Optional[str] :keyword location: Location where to run the job. Must match the location of the destination table. :kwtype location: Optional[str] :keyword project: Project ID of the project of where to run the job. Defaults to the client's project. :kwtype project: Optional[str] :keyword job_config: Extra configuration options for the job. :kwtype job_config: Optional[google.cloud.bigquery.job.LoadJobConfig] :keyword timeout: The number of seconds to wait for the underlying HTTP transport before using ``retry``. :kwtype timeout: Optional[float]
Table into which data is to be loaded. If a string is passed in, this method attempts to create a table reference from a string using from_string.
Exceptions Type Description ValueError If ``size`` is not passed in and can not be determined, or if the ``file_obj`` can be detected to be a file opened in text mode. TypeError If ``job_config`` is not an instance of LoadJobConfig class. Returns Type Description google.cloud.bigquery.job.LoadJob A new load job. load_table_from_jsonload_table_from_json(
json_rows,
destination,
num_retries=6,
job_id=None,
job_id_prefix=None,
location=None,
project=None,
job_config=None,
timeout=None,
)
Upload the contents of a table from a JSON string or dict.
Parameters Name Description json_rowsIterable[Dict[str, Any]]
Row data to be inserted. Keys must match the table schema fields and values must be JSON-compatible representations. .. note:: If your data is already a newline-delimited JSON string, it is best to wrap it into a file-like object and pass it to load_table_from_file:: import io from google.cloud import bigquery data = u'{"foo": "bar"}' data_as_file = io.StringIO(data) client = bigquery.Client() client.load_table_from_file(data_as_file, ...)
destinationUnion[ google.cloud.bigquery.table.Table, google.cloud.bigquery.table.TableReference, str, ] :keyword num_retries: Number of upload retries. :kwtype num_retries: Optional[int] :keyword job_id: Name of the job. :kwtype job_id: Optional[str] :keyword job_id_prefix: The user-provided prefix for a randomly generated job ID. This parameter will be ignored if a ``job_id`` is also given. :kwtype job_id_prefix: Optional[str] :keyword location: Location where to run the job. Must match the location of the destination table. :kwtype location: Optional[str] :keyword project: Project ID of the project of where to run the job. Defaults to the client's project. :kwtype project: Optional[str] :keyword job_config: Extra configuration options for the job. The ``source_format`` setting is always set to NEWLINE_DELIMITED_JSON. :kwtype job_config: Optional[google.cloud.bigquery.job.LoadJobConfig] :keyword timeout: The number of seconds to wait for the underlying HTTP transport before using ``retry``. :kwtype timeout: Optional[float]
Table into which data is to be loaded. If a string is passed in, this method attempts to create a table reference from a string using from_string.
Exceptions Type Description TypeError If ``job_config`` is not an instance of LoadJobConfig class. Returns Type Description google.cloud.bigquery.job.LoadJob A new load job. load_table_from_uriload_table_from_uri(source_uris, destination, job_id=None, job_id_prefix=None, location=None, project=None, job_config=None, retry=<google.api_core.retry.Retry object>, timeout=None)
Parameters Name Description source_uris Union[str, Sequence[str]]
URIs of data files to be loaded; in format gs://<bucket_name>/<object_name_or_glob>
.
Union[ google.cloud.bigquery.table.Table, google.cloud.bigquery.table.TableReference, str, ] :keyword job_id: Name of the job. :kwtype job_id: Optional[str] :keyword job_id_prefix: The user-provided prefix for a randomly generated job ID. This parameter will be ignored if a ``job_id`` is also given. :kwtype job_id_prefix: Optional[str] :keyword location: Location where to run the job. Must match the location of the destination table. :kwtype location: Optional[str] :keyword project: Project ID of the project of where to run the job. Defaults to the client's project. :kwtype project: Optional[str] :keyword job_config: Extra configuration options for the job. :kwtype job_config: Optional[google.cloud.bigquery.job.LoadJobConfig] :keyword retry: How to retry the RPC. :kwtype retry: Optional[google.api_core.retry.Retry] :keyword timeout: The number of seconds to wait for the underlying HTTP transport before using ``retry``. :kwtype timeout: Optional[float]
Table into which data is to be loaded. If a string is passed in, this method attempts to create a table reference from a string using from_string.
Exceptions Type Description TypeError If ``job_config`` is not an instance of LoadJobConfig class. Returns Type Description google.cloud.bigquery.job.LoadJob A new load job. queryquery(query, job_config=None, job_id=None, job_id_prefix=None, location=None, project=None, retry=<google.api_core.retry.Retry object>, timeout=None)
Parameter Name Description query str :keyword job_config: Extra configuration options for the job. To override any options that were previously set in the ``default_query_job_config`` given to the ``Client`` constructor, manually set those options to ``None``, or whatever value is preferred. :kwtype job_config: Optional[google.cloud.bigquery.job.QueryJobConfig] :keyword job_id: ID to use for the query job. :kwtype job_id: Optional[str] :keyword job_id_prefix: The prefix to use for a randomly generated job ID. This parameter will be ignored if a ``job_id`` is also given. :kwtype job_id_prefix: Optional[str] :keyword location: Location where to run the job. Must match the location of the any table used in the query as well as the destination table. :kwtype location: Optional[str] :keyword project: Project ID of the project of where to run the job. Defaults to the client's project. :kwtype project: Optional[str] :keyword retry: How to retry the RPC. :kwtype retry: Optional[google.api_core.retry.Retry] :keyword timeout: The number of seconds to wait for the underlying HTTP transport before using ``retry``. :kwtype timeout: Optional[float]
SQL query to be executed. Defaults to the standard SQL dialect. Use the job_config
parameter to change dialects.
schema_from_json(file_or_path)
Takes a file object or file path that contains json that describes a table schema.
schema_to_jsonschema_to_json(schema_list, destination)
Takes a list of schema field objects.
Serializes the list of schema field objects as json to a file.
Destination is a file path or a file object.
update_datasetupdate_dataset(dataset, fields, retry=<google.api_core.retry.Retry object>, timeout=None)
Change some fields of a dataset.
Use fields
to specify which fields to update. At least one field must be provided. If a field is listed in fields
and is None
in dataset
, it will be deleted.
If dataset.etag
is not None
, the update will only succeed if the dataset on the server has the same ETag. Thus reading a dataset with get_dataset
, changing its fields, and then passing it to update_dataset
will ensure that the changes will only be saved if no modifications to the dataset occurred since the read.
google.cloud.bigquery.dataset.Dataset
The dataset to update.
fieldsSequence[str]
The properties of dataset
to change. These are strings corresponding to the properties of Dataset. For example, to update the default expiration times, specify both properties in the fields
argument: .. code-block:: python bigquery_client.update_dataset( dataset, [ "default_partition_expiration_ms", "default_table_expiration_ms", ] )
Optional[google.api_core.retry.Retry]
How to retry the RPC.
timeoutOptional[float]
The number of seconds to wait for the underlying HTTP transport before using retry
.
update_model(model, fields, retry=<google.api_core.retry.Retry object>, timeout=None)
[Beta] Change some fields of a model.
Use fields
to specify which fields to update. At least one field must be provided. If a field is listed in fields
and is None
in model
, the field value will be deleted.
If model.etag
is not None
, the update will only succeed if the model on the server has the same ETag. Thus reading a model with get_model
, changing its fields, and then passing it to update_model
will ensure that the changes will only be saved if no modifications to the model occurred since the read.
google.cloud.bigquery.model.Model
The model to update.
fieldsSequence[str]
The properties of model
to change. These are strings corresponding to the properties of Model. For example, to update the descriptive properties of the model, specify them in the fields
argument: .. code-block:: python bigquery_client.update_model( model, ["description", "friendly_name"] )
Optional[google.api_core.retry.Retry]
A description of how to retry the API call.
timeoutOptional[float]
The number of seconds to wait for the underlying HTTP transport before using retry
.
update_routine(routine, fields, retry=<google.api_core.retry.Retry object>, timeout=None)
[Beta] Change some fields of a routine.
Use fields
to specify which fields to update. At least one field must be provided. If a field is listed in fields
and is None
in routine
, the field value will be deleted.
If xref_etag is not
None
, the update will only succeed if the resource on the server has the same ETag. Thus reading a routine with xref_get_routine, changing its fields, and then passing it to this method will ensure that the changes will only be saved if no modifications to the resource occurred since the read.
Parameters Name Description routinegoogle.cloud.bigquery.routine.Routine
The routine to update.
fieldsSequence[str]
The fields of routine
to change, spelled as the Routine properties. For example, to update the description property of the routine, specify it in the fields
argument: .. code-block:: python bigquery_client.update_routine( routine, ["description"] )
Optional[google.api_core.retry.Retry]
A description of how to retry the API call.
timeoutOptional[float]
The number of seconds to wait for the underlying HTTP transport before using retry
.
update_table(table, fields, retry=<google.api_core.retry.Retry object>, timeout=None)
Change some fields of a table.
Use fields
to specify which fields to update. At least one field must be provided. If a field is listed in fields
and is None
in table
, the field value will be deleted.
If table.etag
is not None
, the update will only succeed if the table on the server has the same ETag. Thus reading a table with get_table
, changing its fields, and then passing it to update_table
will ensure that the changes will only be saved if no modifications to the table occurred since the read.
google.cloud.bigquery.table.Table
The table to update.
fieldsSequence[str]
The fields of table
to change, spelled as the Table properties. For example, to update the descriptive properties of the table, specify them in the fields
argument: .. code-block:: python bigquery_client.update_table( table, ["description", "friendly_name"] )
Optional[google.api_core.retry.Retry]
A description of how to retry the API call.
timeoutOptional[float]
The number of seconds to wait for the underlying HTTP transport before using retry
.
__init__(
project=None,
credentials=None,
_http=None,
location=None,
default_query_job_config=None,
client_info=None,
client_options=None,
)
Initialize self. See help(type(self)) for accurate signature.
get_iam_policyget_iam_policy(table, requested_policy_version=1, retry=<google.api_core.retry.Retry object>, timeout=None)
API documentation for bigquery.client.Client.get_iam_policy
method.
set_iam_policy(table, policy, updateMask=None, retry=<google.api_core.retry.Retry object>, timeout=None)
API documentation for bigquery.client.Client.set_iam_policy
method.
test_iam_permissions(table, permissions, retry=<google.api_core.retry.Retry object>, timeout=None)
API documentation for bigquery.client.Client.test_iam_permissions
method.
Except as otherwise noted, the content of this page is licensed under the Creative Commons Attribution 4.0 License, and code samples are licensed under the Apache 2.0 License. For details, see the Google Developers Site Policies. Java is a registered trademark of Oracle and/or its affiliates.
Last updated 2025-08-07 UTC.
[[["Easy to understand","easyToUnderstand","thumb-up"],["Solved my problem","solvedMyProblem","thumb-up"],["Other","otherUp","thumb-up"]],[["Hard to understand","hardToUnderstand","thumb-down"],["Incorrect information or sample code","incorrectInformationOrSampleCode","thumb-down"],["Missing the information/samples I need","missingTheInformationSamplesINeed","thumb-down"],["Other","otherDown","thumb-down"]],["Last updated 2025-08-07 UTC."],[],[]]
RetroSearch is an open source project built by @garambo | Open a GitHub Issue
Search and Browse the WWW like it's 1997 | Search results from DuckDuckGo
HTML:
3.2
| Encoding:
UTF-8
| Version:
0.7.4