The support for python Bigquery API indicates that arrays are possible, however, when passing from a pandas dataframe to bigquery there is a pyarrow struct issue.
The only way round it seems its to drop columns then use JSON Normalise for a separate table.
from google.cloud import bigquery project = 'lake' client = bigquery.Client(credentials=credentials, project=project) dataset_ref = client.dataset('XXX') table_ref = dataset_ref.table('RAW_XXX') job_config = bigquery.LoadJobConfig() job_config.autodetect = True job_config.write_disposition = 'WRITE_TRUNCATE' client.load_table_from_dataframe(appended_data, table_ref,job_config=job_config).result()
This is the error recieved. NotImplementedError: struct
The reason I wanted to use this API as it indicates Nested Array support, which is perfect for our data lake in BQ but I assume this doesn't work?
RetroSearch is an open source project built by @garambo | Open a GitHub Issue
Search and Browse the WWW like it's 1997 | Search results from DuckDuckGo
HTML:
3.2
| Encoding:
UTF-8
| Version:
0.7.4