Connecting to Google BigQuery

Sisense’s Google BigQuery connector is a powerful tool that enables you to retrieve from BigQuery only the data that you require, based on timestamp, date range or other parameters defined by you. This reduces import times and the costs associated with querying a large number of tables in BigQuery.

This topic includes the following major sections:

  • Importing Into a Sisense ElastiCube - How to import data into ElastiCubes
  • Connecting Data to a Sisense Live Model - How to use Live Connect to connect data to a Sisense Live model

Note:

Make sure that you have the appropriate permissions to access your BigQuery dataset. For more information, see here .

Importing Into a Sisense ElastiCube

Sisense ’s Google BigQuery connector supports complex data types and partition types used in BigQuery. You can run queries on defined partitions, including tables partitioned by date field, sharded tables, and tables partitioned by ingestion time. Complex data types, including Nested and Nested Repeated, are ingested automatically without any need to write SQL or functions.

Authentication
When connecting to Google Big Query, you can choose the preferred authentication mechanism: OAuth or Service Account. A service account is a Google account associated with your GCP (Google Cloud Platform‎) project. Use a service account to access the BigQuery API if your application can run jobs associated with service credentials rather than end-user credentials. For more details, click here .

BigQuery Storage API

Optionally, the connector allows using the Big Query Storage API. BQ Storage API is a paid Big Query service that allows for additional parallelism among multiple consumers for a set of results. Thus it enables the driver to handle large result sets more efficiently. For details, see here .

Adding a Connection to Google BigQuery

Sisense enables easy and quick access to databases, tables and views contained within Google BigQuery.
To connect to your Google BigQuery database, you need to provide a Project ID. The Project ID is a unique identifier for your BigQuery project. You receive the Project ID when you create a project in Google BigQuery.

If you have already created a project and can’t find the Project ID, click here for help.

After you provide the Project ID, you can sign into your BigQuery account with your Google credentials. Once Google has authenticated your account, you can select what tables are to be imported into Sisense.

Note:

If you are connecting to a Google service remotely and the address of Sisense is something other than localhost, Google requires that you connect using the OAuth 2.0 protocol. See Google Authentication for more information.

To import Google BigQuery data:

  1. In the Data page, open an ElastiCube or click to create a new ElastiCube.

  2. In the Model Editor, click .The Add Data dialog box is displayed.

  3. In the Add Data dialog box, select Google BigQuery .

  4. If your preferred authentication method is OAuth, click Sign in with Google and enter your email, and then select Next to enter your password. If you have multiple accounts, select the account that has the Google BigQuery data you want to access and enter the password, if you're not already signed in. You must be signed out of all other Google accounts.

  5. Click Accept to allow your Sisense Server to access your Google BigQuery data.

  6. In Project ID for Authentication, enter your BigQuery Project ID. This project ID is only used for authentication. Once you log in, you will have access to all the projects to which you have Google Big Query permissions.

    If you have already created a project and can’t find the Project ID, click here for help.

  7. Close the browser window when notified to do so.

  8. In Record Fields Flattening Level, enter the number of nested levels you want to flatten. For example, in a table with nested data depth of seven, you may choose to flatten only the first two levels. The data on the other levels will not be imported into Sisense. See Connecting to Google BigQuery for more information.

  9. Select Record Repeated and Array Fields Unnesting if you want to unnest arrays and Nested Repeated fields, and to flatten and unnest arrays of Records into additional rows. If you do not select the checkbox, all these data types will be imported as strings.

  10. Click Allow Large Results if your BigQuery query result is larger than 128 MB. When importing query results larger than 128 MB, BigQuery creates a temporary table (see below, or click here for more information). For this table to be created, you must be granted the ‘Bigquery.tables.create’ permission in BigQuery (click here for more information). If you do not have this permission, the query results will not be imported into Sisense. The message that Sisense issues in this case is “Response too large to return. Consider setting allowLargeResults to true in your job configuration.”

Notes about Temporary Tables

  • The Sisense BigQuery connector creates temporary tables in BigQuery under a specified dataset. By default, temporary tables exist for 24 hours.

  • You can customize the behavior of the "allow large results" feature.

    • To change the location of the temporary tables, add the parameter LargeResultDataset=<datasetName> to the Additional parameter field in the connection settings page.

    • To change the expiration date of the dataset, add the parameter LargeResultsDatasetExpirationTime=<expirationTime> to the additional parameters field in the connection settings page. <expirationTime> is in milliseconds. This dataset must have an expiration date, otherwise the temporary tables will not be deleted.

  • You will need to manually delete temporary tables that were created prior to Sisense Version L8.2.6SP.

  1. If your preferred authentication method is Service Account, you will need to connect to BigQuery with a service account and load the credentials from a JSON file to Sisense :
    1. Create and download the Service Account Key JSON file. For instructions,click here.
    2. Upload the Service Account Key JSON file to the Sisense server (you can save it to any directory).
    3. In Sisense , select Authenticate With Service Account.
    4. In a field that opens, enter the full path to the Service Account Key JSON file on the Sisense server.
    /opt/sisense/storage/<prefix of the path on the Sisense server>

  2. Click Next . All tables and views associated with Google BigQuery are displayed.

  3. From the Tables list, select the relevant table or view you want to work with. You can click next to the relevant table or view to see a preview of the data inside it.

Note:

When selecting the tables or views, Connector Wizard displays all schemas from all projects accessed, not only the schemas from the project whose ID you provided for authentication.

  1. When you select the table or view, a new option is displayed at the bottom of the list, Add Import Query.

  2. (Optional) Click + to customize the data you want to import with SQL. See Custom Queries to Reduce the Amount of Imported Data above for more information.

  3. After you have selected all the relevant tables, click Done. The tables are added to your schema.

Temporary Tables: Known Issues

In Sisense versions below 8.2.1 (Windows), temporary tables do not expire by default. To change this behavior:

  1. Manually refresh the connection at the data source level by changing the connection in the Data tab.

    This applies an expiration date to temporary tables.

  2. In BigQuery, in the dataset of the building table, set Default table expiration to Other value than "Never". All newly created tables of any given dataset will inherit the dataset's expiration period.

Querying Partitions

Sisense’s Google BigQuery connector provides the ability to run queries on defined partitions:

  • Sharded tables - supports the ability to view the parent tables (grouped view, as wildcard Select statement) while running queries per the partitions as filters.
  • Tables partitioned by date field - querying by partitions as out-of-the-box functionality, via filters and custom queries.
  • Tables partitioned by ingestion time - using BigQuery SQL functions via custom queries.

Sisense provides a unique solution to rebuilding the parent table from its constituent sharded tables (grouped view). For this, Sisense provides an external file, where you can indicate the custom prefixes or regular expressions (regex) by which to group the shared tables into one parent table.

To group the shared tables into a single parent table:

  1. Open the prefixes.json file (you can open it in Notepad). This file is located in the following directory:

    ...ProgramData\Sisense\DataConnectors\JVMContainer\Connectors\GoogleBigQuery

    This file contains default prefix and regex parameters that group all Google Analytics daily tables of the format "ga_sessions_YYYYMMDD” into one parent table.

  2. To refine the grouping, modify the default parameters or add additional parameters as relevant to your data, and save the file. Make sure that the prefixes row is separated by a comma from the regex row.

  3. To update your data model based on the new parameters, do one of the following:

    • from the Connection Wizard, jump to a previous step and return to the step you were on.
    • from the data model, open the Connector Wizard to add another table.

In both cases, The Connector Wizard will display the parent table, grouped from the sharded tables based on the indicated parameters.

Controlling the Partitions to Import with Custom Queries

Note :

This feature is available in Live data models starting from Sisense version 8.0.1. On earlier versions it is available only in ElastiCube data models.

Google BigQuery supports partitions and sharded tables to improve performance, availability, and maintainability.

The Sisense BigQuery connector provides the following abilities to run queries on BigQuery's table partitions and sharded tables:

  • Partitioned tables : Tables that are partitioned based on a TIMESTAMP or DATE column. The Sisense BigQuery connector enables querying the partitions via filters and custom queries, simply by the where clause sent to BigQuery.
  • Tables partitioned by ingestion time : Tables partitioned based on the data's ingestion (load) date or arrival date. The Sisense BigQuery connector enables using BigQuery SQL functions via custom queries to retrieve the requested partitions. Click here for details about these two methods.
  • As an alternative to partitioned tables, Google BigQuery enables sharding tables using a time-based naming approach, such as [PREFIX]_YYYYMMDD. This is referred to as date-sharded tables . See here for details. The Sisense BigQuery connector supports the ability to view the parent tables (grouped view, as Wildcard Select statement) while running queries per the partitions as filters.

You can import subsets of your partitioned and sharded tables into Sisense using SQL.

Note:

Starting from Sisense version 8.0.1, you can create custom SQL queries in Live data models. With Live table queries, you can now add and transform data and build custom business logic for analytics from Live models using custom SQL. Using this feature you can add a WHERE clause on your specific partition. For details, see Creating Custom Live Table Queries.

Querying Partitioned Tables

Tables partitioned based on a TIMESTAMP or DATE column do not have pseudo columns in BigQuery. To limit the number of partitions scanned when querying partitioned tables, use a predicate filter (a WHERE clause).
A common use case of leveraging partitioned tables in BigQuery is the accumulate build. Defining the accumulate build on a TIMESTAMP or DATE field enables you to scan only specific partitions, thus reducing the import times and costs involved in querying in BigQuery.

For example, the following query prunes partitions:

SELECT
 t1.name,
 t2.category
FROM
 table1 t1
INNER JOIN
 table2 t2
ON t1.id_field = t2 field2
WHERE
 t1.ts = CURRENT_TIMESTAMP()

Querying Ingestion-time Partitioned Tables

When you create an ingestion-time partitioned table in BigQuery, two pseudo columns are added to the table: a _PARTITIONTIME pseudo column and a _PARTITIONDATE pseudo column. The _PARTITIONTIME pseudo column contains a date-based timestamp for data that is loaded into the table. The _PARTITIONDATE pseudo column contains a date representation.

You can use the _PARTITIONTIME and _PARTITIONDATE pseudo columns to limit the number of partitions scanned during a query. This reduces the on-demand analysis cost. For example, the following query scans only the partitions between the dates January 1, 2019 and January 2, 2019 from the partitioned table:

SELECT
 [COLUMN]
FROM
 [DATASET].[TABLE]
WHERE  
  _PARTITIONDATE BETWEEN '2019-01-01'
  AND '2019-01-02`

See here for a detailed explanation.

Querying Sharded Tables

BigQuery stores sharded tables in the format of table_name_SUFFIX (for example "ga_sessions_YYYYMMDD”). You can write a custom query to import only the sharded tables in the requested date range. This provides the benefit of faster and cheaper queries, since in BigQuery you pay for the amount of data you scan.

To do that, you can use wildcard tables combined with the _TABLE_SUFFIX pseudo column in the WHERE clause. The _TABLE_SUFFIX pseudo column contains the values matched by the table wildcard. For example, the following FROM clause uses the wildcard (* ) to match all tables in the ga_sessions dataset that match a specific date.

SELECT 
  date, 
SUM(totals.visits) AS visits, 
SUM(totals.pageviews) AS pageviews, 
FROM 
  `aerial-citron-207113.GA360.ga_sessions_*` 
WHERE 
 _TABLE SUFFIX BETWEEN '20170801'
  AND '20180801'
Group BY 
  date 
ORDER BY 
  date ASC

See here for a detailed explanation of the example.

Also, see here for additional examples of advanced custom queries.

Note :

Sisense uses the standard SQL dialect, and not legacy SQL (also known as the BigQuery SQL).

Querying Nested Objects

BigQuery supports nested records and arrays within tables. Nested records in BigQuery can be Single or Repeated records.

These record types are imported into Sisense as:

  • Single records - as additional columns
  • Repeated records - as additional columns and rows

For example: A BigQuery schema that has Single and Repeated records:

Would be imported into Sisense as:

Click here for more information.

Sisense ingests both nested records and arrays automatically without any need to write SQL and functions:

  • Nested data type : Flattens records/arrays to columns with the ability to choose how many levels to expand.
  • Nested Repeated data type : Flattens repeated records to rows and columns automatically, using the UNNEST function.

Nested Objects

BigQuery supports Nested data as objects of Record data type.

When importing data into Sisense, you need to indicate how many levels of nested data you want to flatten (see Connecting to Google BigQuery ). In Sisense, data on these levels will be flattened to columns using the dot operator (. ).

Data on the nested levels that you chose not to flatten is not imported into Sisense.

Array Objects

BigQuery supports Nested Repeated data as arrays of Repeated Record data type.

When importing data into Sisense, Nested Repeated data type is flattened to rows and columns automatically using the UNNEST function.

Note:

This feature is currently available only in ElastiCube data models. It will be available in Live data models starting from Sisense version 8.0. If you are on an earlier Sisense version and working with Live data models, you can use the UNNEST operator to retrieve nested and repeated data as flattened rows via custom query (see here for more information).

Data in Array (primitive) data type is imported into Sisense. Sisense converts this data to un-indexed strings.

Source Limitations

  • The following functions are not supported for Google BigQuery: Cot, Diff, Mod, Round
  • The Storage API cannot be used with federated tables or logical views
  • Maximum 15 nested levels (click here for details)
  • Maximum query length: 1 MB
  • Maximum response size: 128 MB compressed. This limitation can be overcome by clicking Allow Large Results Set as explained in Connecting to Google BigQuery .
  • Query execution time limit: 6 hours
  • Maximum number of tables referenced per query: 1,000
  • Concurrent queries - based on the account rate limits. Click here for information on BigQuery quotas.

Using Google BigQuery Storage API

The Sisense Big Query connector allows you to use the BigQuery Storage API. The BigQuery Storage API provides fast access to data stored in BigQuery. The BigQuery Storage API is paid BigQuery service. For details about this service, click here .

Once the Storage API is enabled in BigQuery, it becomes available in Sisense and helps speed up the build times. On each import data request from Sisense, the executed query will be automatically optimized to work with the Storage API to import the data from BigQuery. In some use cases, using Storage API for large imports, you can achieve up to x5 faster build time.

When enabled, the driver checks the number of rows in an incoming result set table and the number of pages needed to retrieve all the results. If the number of rows and pages exceeds the threshold defined by Google Big Query, the driver switches to using the BigQuery Storage API. The Storage API feature allows for additional parallelism among multiple consumers for a set of results, thus enabling the driver to handle large result sets more efficiently.

Enabling Storage API in Google BigQuery

Enable the below service in the BigQuery API Library:

Storage API is a beta feature by BigQuery.

To learn more:

  • For general information about the Storage API, click here .
  • For Storage API pricing details, click here .

Troubleshooting

Out-of-Memory Issues

When building an ElastiCube with this connector, you might receive an "out of memory" error.

Connecting Data to a Sisense Live Model

Sisense ’s Google BigQuery connector supports complex data types and partition types used in BigQuery. Complex data types, including Nested and Nested Repeated, are ingested automatically without any need to write SQL or functions.

In this topic:

Connecting to BigQuery

Sisense enables easy and quick access to databases, tables and views contained within Google BigQuery.
To connect to your Google BigQuery database, you need to provide a Project ID. The Project ID is a unique identifier for your BigQuery project. You receive the Project ID when you create a project in Google BigQuery.

If you have already created a project and can’t find the Project ID, click here for help.

After you provide the Project ID, you can sign into your BigQuery account with your Google credentials. Once Google has authenticated your account, you can select what tables are to be imported into Sisense.

Note:

If you are connecting to a Google service remotely and the address of Sisense is something other than localhost, Google requires that you connect using the OAuth 2.0 protocol. See Google Authentication for more information.

To add a Google BigQuery live connection:

  1. In the Data page, open a live model or click to create a new live model.

  2. In the Model Editor, click . The Add Data dialog box is displayed.

  3. In the Add Data dialog box, select Google BigQuery.

  4. In Project ID , enter your BigQuery Project ID. This project ID is only used for authentication. Once you log in, you will have access to all the projects to which you have Google Big Query permissions.

    If you have already created a project and can’t find the Project ID, click here for help.

  5. Click Sign in with Google and enter your email, and then select Next to enter your password. If you have multiple accounts, select the account that has the Google BigQuery data you want to access and enter the password, if you're not already signed in. You must be signed out of all other Google accounts.

  6. Click Accept to allow your Sisense Server to access your Google BigQuery data.

  7. Close the browser window when notified to do so.

  8. In Record Fields Flattening Level , enter the number of nested levels you want to flatten. For example, in a table with nested data depth of seven, you may choose to flatten only the first two levels. The data on the other levels will not be imported into Sisense. See Connecting to Google BigQuery for more information.

  9. Click Allow Large Results if your BigQuery table is larger than 128 MB. Important: This option is not usually recommended for Live connections because it is mainly for large import queries. However, for some use-cases with big data volumes, it could be relevant for Live models as well. When importing query results larger than 128 MB, BigQuery creates a temporary table (see above, or click here for more information). For this table to be created, you must be granted the ‘Bigquery.tables.create’ permission in BigQuery (click here for more information). If you do not have this permission, the query results will not be imported into Sisense. The message that Sisense issues in this case is “Response too large to return. Consider setting allowLargeResults to true in your job configuration.”

Notes about Temporary Tables

  • The Sisense BigQuery connector creates temporary tables in BigQuery under a specified dataset. By default, temporary tables exist for 24 hours.

  • Newly created temporary tables will be written to a hidden dataset named _simba_jdbc (hidden datasets are not visible in the BigQuery Web UI).

  • You can customize the behavior of the "allow large results" feature.

    • To change the location of the temporary tables, add the parameter LargeResultDataset=<datasetName> to the Additional parameter field in the connection settings page.

    • To change the expiration date of the dataset, add the parameter LargeResultsDatasetExpirationTime=<expirationTime> to the additional parameters field in the connection settings page. <expirationTime> is in milliseconds. This dataset must have an expiration date, otherwise the temporary tables will not be deleted.

  • You will need to manually delete temporary tables that were created prior to Sisense Version L8.2.6SP.

  1. Click Next . All tables and views associated with Google BigQuery are displayed.

  2. From the Tables list, select the relevant table or view you want to work with. You can click next to the relevant table or view to see a preview of the data inside it.

  3. After you have selected all the relevant tables, click Done. The tables are added to your schema.

Temporary Tables: Known Issues

In Sisense versions below 8.2.1 (Windows), temporary tables do not expire by default. To change this behavior:

  1. Manually refresh the connection at the data source level by changing the connection in the Data tab.

    This applies an expiration date to temporary tables.

  2. In BigQuery, in the dataset of the building table, set Default table expiration to Other value than "Never". All newly created tables of any given dataset will inherit the dataset's expiration period.

Sisense provides a unique solution to rebuilding the parent table from its constituent sharded tables (grouped view).

For this, Sisense provides an external file, where you can indicate the custom prefixes or regular expressions (regex) by which to group the shared tables into one parent table.

To group the shared tables into a single parent table:

  1. Open the prefixes.json file (you can open it in Notepad). This file is located in the following directory:

    ...ProgramData\Sisense\DataConnectors\JVMContainer\Connectors\GoogleBigQuery

    This file contains default prefix and regex parameters that group all Google Analytics daily tables of the format "ga_sessions_YYYYMMDD” into one parent table.

  2. To refine the grouping, modify the default parameters or add additional parameters as relevant to your data, and save the file. Make sure that the prefixes row is separated by a comma from the regex row.

  3. To update your data model based on the new parameters, do one of the following:

    • From the Connection Wizard, jump to a previous step and return to the step you were on.
    • From the data model, open the Connector Wizard to add another table.

In both cases, the Connector Wizard will display the parent table, grouped from the sharded tables based on the indicated parameters.

Controlling the Partitions to Import with Custom Queries

Note:

This feature is available in Live data models starting from Sisense version 8.0.1. On earlier versions it is available only in ElastiCube data models.

Google BigQuery supports partitions and sharded tables to improve performance, availability, and maintainability.

The Sisense BigQuery connector provides the following abilities to run queries on BigQuery's table partitions and sharded tables:

  • Partitioned tables : Tables that are partitioned based on a TIMESTAMP or DATE column. The Sisense BigQuery connector enables querying the partitions via filters and custom queries, simply by the where clause sent to BigQuery.
  • Tables partitioned by ingestion time : Tables partitioned based on the data's ingestion (load) date or arrival date. The Sisense BigQuery connector enables using BigQuery SQL functions via custom queries to retrieve the requested partitions. Click here for details about these two methods.
  • As an alternative to partitioned tables, Google BigQuery enables sharding tables using a time-based naming approach, such as [PREFIX]_YYYYMMDD. This is referred to as date-sharded tables . See here for details. The Sisense BigQuery connector supports the ability to view the parent tables (grouped view, as Wildcard Select statement) while running queries per the partitions as filters.

You can import subsets of your partitioned and sharded tables into Sisense using SQL.

Note:

Starting from Sisense version 8.0.1, you can create custom SQL queries in Live data models. With Live table queries, you can now add and transform data and build custom business logic for analytics from Live models using custom SQL. Using this feature you can add a WHERE clause on your specific partition. For details, see Creating Custom Live Table Queries.

Querying Nested Objects

BigQuery supports nested records and arrays within tables. Nested records in BigQuery can be Single or Repeated records.

These record types are imported into Sisense as:

  • Single records - as additional columns
  • Repeated records - as additional columns and rows

For example: A BigQuery schema that has Single and Repeated records:

Would be imported into Sisense as:

Click here for more information.
Sisense ingests both nested records and arrays automatically without any need to write SQL and functions:

  • Nested data type : Flattens records/arrays to columns with the ability to choose how many levels to expand.
  • Nested Repeated data type : Flattens repeated records to rows and columns automatically, using the UNNEST function.

Nested Objects

BigQuery supports Nested data as objects of Record data type.

When importing data into Sisense, you need to indicate how many levels of nested data you want to flatten (see Connecting to Google BigQuery). In Sisense, data on these levels will be flattened to columns using the dot operator (. ).

Data on the nested levels that you chose not to flatten is not imported into Sisense.

Array Objects

BigQuery supports Nested Repeated data as arrays of Repeated Record data type.

When importing data into Sisense, Nested Repeated data type is flattened to rows and columns automatically using the UNNEST function.

Data in Array (primitive) data type is imported into Sisense. Sisense converts this data to un-indexed strings.

Enabling Storage API in Google BigQuery

Note:

For Live models, Sisense recommends disabling the Storage API in Google BigQuery. This feature works best when importing large data volumes into ElastiCubes.

Enable the below service in the BigQuery API Library:

Storage API is a beta feature by BigQuery.

To learn more:

  • For general information about the Storage API, click here .
  • For Storage API pricing details, click here .

Source Limitations

  • The following functions are not supported for Google BigQuery: Cot, Diff, Mod, Round
  • Maximum 15 nested levels (click here for details)
  • Maximum query length: 1 MB
  • Maximum response size: 128 MB compressed. This limitation can be overcome by clicking Allow Large Results Set as explained in Connecting to Google BigQuery .
  • Query execution time limit: 6 hours
  • Maximum number of tables referenced per query: 1,000
  • Concurrent queries - based on the account rate limits. Click here for information on BigQuery quotas.

The Sisense Big Query connector allows you to use the BigQuery Storage API. The BigQuery Storage API provides fast access to data stored in BigQuery. The BigQuery Storage API is paid BigQuery service. For details about this service, click here .

Once the Storage API is enabled in BigQuery, it becomes available in Sisense and helps speed up the build times. On each import data request from Sisense, the executed query will be automatically optimized to work with the Storage API to import the data from BigQuery. In some use cases, using Storage API for large imports, you can achieve up to x5 faster build time.

When enabled, the driver checks the number of rows in an incoming result set table and the number of pages needed to retrieve all the results. If the number of rows and pages exceeds the threshold defined by Google Big Query, the driver switches to using the BigQuery Storage API. The Storage API feature allows for additional parallelism among multiple consumers for a set of results, thus enabling the driver to handle large result sets more efficiently.

.r.