New module: gcp_bigquery_table_facts (#50679)
This commit is contained in:
parent
13a2a8b5f6
commit
1d747649e6
2 changed files with 607 additions and 8 deletions
|
@ -19,6 +19,7 @@ except ImportError:
|
||||||
from ansible.module_utils.basic import AnsibleModule, env_fallback
|
from ansible.module_utils.basic import AnsibleModule, env_fallback
|
||||||
from ansible.module_utils.six import string_types
|
from ansible.module_utils.six import string_types
|
||||||
from ansible.module_utils._text import to_text
|
from ansible.module_utils._text import to_text
|
||||||
|
import ast
|
||||||
import os
|
import os
|
||||||
|
|
||||||
|
|
||||||
|
@ -60,10 +61,18 @@ def replace_resource_dict(item, value):
|
||||||
else:
|
else:
|
||||||
if not item:
|
if not item:
|
||||||
return item
|
return item
|
||||||
|
if isinstance(item, dict):
|
||||||
return item.get(value)
|
return item.get(value)
|
||||||
|
|
||||||
|
# Item could be a string or a string representing a dictionary.
|
||||||
|
try:
|
||||||
|
new_item = ast.literal_eval(item)
|
||||||
|
return replace_resource_dict(new_item, value)
|
||||||
|
except ValueError:
|
||||||
|
return new_item
|
||||||
|
|
||||||
# Handles all authentication and HTTP sessions for GCP API calls.
|
|
||||||
|
# Handles all authentation and HTTP sessions for GCP API calls.
|
||||||
class GcpSession(object):
|
class GcpSession(object):
|
||||||
def __init__(self, module, product):
|
def __init__(self, module, product):
|
||||||
self.module = module
|
self.module = module
|
||||||
|
@ -77,9 +86,25 @@ class GcpSession(object):
|
||||||
except getattr(requests.exceptions, 'RequestException') as inst:
|
except getattr(requests.exceptions, 'RequestException') as inst:
|
||||||
self.module.fail_json(msg=inst.message)
|
self.module.fail_json(msg=inst.message)
|
||||||
|
|
||||||
def post(self, url, body=None):
|
def post(self, url, body=None, headers=None, **kwargs):
|
||||||
|
if headers:
|
||||||
|
headers = self.merge_dictionaries(headers, self._headers())
|
||||||
|
else:
|
||||||
|
headers = self._headers()
|
||||||
|
|
||||||
try:
|
try:
|
||||||
return self.session().post(url, json=body, headers=self._headers())
|
return self.session().post(url, json=body, headers=headers)
|
||||||
|
except getattr(requests.exceptions, 'RequestException') as inst:
|
||||||
|
self.module.fail_json(msg=inst.message)
|
||||||
|
|
||||||
|
def post_contents(self, url, file_contents=None, headers=None, **kwargs):
|
||||||
|
if headers:
|
||||||
|
headers = self.merge_dictionaries(headers, self._headers())
|
||||||
|
else:
|
||||||
|
headers = self._headers()
|
||||||
|
|
||||||
|
try:
|
||||||
|
return self.session().post(url, data=file_contents, headers=headers)
|
||||||
except getattr(requests.exceptions, 'RequestException') as inst:
|
except getattr(requests.exceptions, 'RequestException') as inst:
|
||||||
self.module.fail_json(msg=inst.message)
|
self.module.fail_json(msg=inst.message)
|
||||||
|
|
||||||
|
@ -103,7 +128,8 @@ class GcpSession(object):
|
||||||
self.module.fail_json(msg=inst.message)
|
self.module.fail_json(msg=inst.message)
|
||||||
|
|
||||||
def session(self):
|
def session(self):
|
||||||
return AuthorizedSession(self._credentials())
|
return AuthorizedSession(
|
||||||
|
self._credentials())
|
||||||
|
|
||||||
def _validate(self):
|
def _validate(self):
|
||||||
if not HAS_REQUESTS:
|
if not HAS_REQUESTS:
|
||||||
|
@ -114,12 +140,12 @@ class GcpSession(object):
|
||||||
|
|
||||||
if self.module.params.get('service_account_email') is not None and self.module.params['auth_kind'] != 'machineaccount':
|
if self.module.params.get('service_account_email') is not None and self.module.params['auth_kind'] != 'machineaccount':
|
||||||
self.module.fail_json(
|
self.module.fail_json(
|
||||||
msg="Service Account Email only works with Machine Account-based authentication"
|
msg="Service Acccount Email only works with Machine Account-based authentication"
|
||||||
)
|
)
|
||||||
|
|
||||||
if self.module.params.get('service_account_file') is not None and self.module.params['auth_kind'] != 'serviceaccount':
|
if self.module.params.get('service_account_file') is not None and self.module.params['auth_kind'] != 'serviceaccount':
|
||||||
self.module.fail_json(
|
self.module.fail_json(
|
||||||
msg="Service Account File only works with Service Account-based authentication"
|
msg="Service Acccount File only works with Service Account-based authentication"
|
||||||
)
|
)
|
||||||
|
|
||||||
def _credentials(self):
|
def _credentials(self):
|
||||||
|
@ -134,13 +160,18 @@ class GcpSession(object):
|
||||||
return google.auth.compute_engine.Credentials(
|
return google.auth.compute_engine.Credentials(
|
||||||
self.module.params['service_account_email'])
|
self.module.params['service_account_email'])
|
||||||
else:
|
else:
|
||||||
self.module.fail_json(msg="Credential type '%s' not implemented" % cred_type)
|
self.module.fail_json(msg="Credential type '%s' not implmented" % cred_type)
|
||||||
|
|
||||||
def _headers(self):
|
def _headers(self):
|
||||||
return {
|
return {
|
||||||
'User-Agent': "Google-Ansible-MM-{0}".format(self.product)
|
'User-Agent': "Google-Ansible-MM-{0}".format(self.product)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
def _merge_dictionaries(self, a, b):
|
||||||
|
new = a.copy()
|
||||||
|
new.update(b)
|
||||||
|
return new
|
||||||
|
|
||||||
|
|
||||||
class GcpModule(AnsibleModule):
|
class GcpModule(AnsibleModule):
|
||||||
def __init__(self, *args, **kwargs):
|
def __init__(self, *args, **kwargs):
|
||||||
|
|
568
lib/ansible/modules/cloud/google/gcp_bigquery_table_facts.py
Normal file
568
lib/ansible/modules/cloud/google/gcp_bigquery_table_facts.py
Normal file
|
@ -0,0 +1,568 @@
|
||||||
|
#!/usr/bin/python
|
||||||
|
# -*- coding: utf-8 -*-
|
||||||
|
#
|
||||||
|
# Copyright (C) 2017 Google
|
||||||
|
# GNU General Public License v3.0+ (see COPYING or https://www.gnu.org/licenses/gpl-3.0.txt)
|
||||||
|
# ----------------------------------------------------------------------------
|
||||||
|
#
|
||||||
|
# *** AUTO GENERATED CODE *** AUTO GENERATED CODE ***
|
||||||
|
#
|
||||||
|
# ----------------------------------------------------------------------------
|
||||||
|
#
|
||||||
|
# This file is automatically generated by Magic Modules and manual
|
||||||
|
# changes will be clobbered when the file is regenerated.
|
||||||
|
#
|
||||||
|
# Please read more about how to change this file at
|
||||||
|
# https://www.github.com/GoogleCloudPlatform/magic-modules
|
||||||
|
#
|
||||||
|
# ----------------------------------------------------------------------------
|
||||||
|
|
||||||
|
from __future__ import absolute_import, division, print_function
|
||||||
|
__metaclass__ = type
|
||||||
|
|
||||||
|
################################################################################
|
||||||
|
# Documentation
|
||||||
|
################################################################################
|
||||||
|
|
||||||
|
ANSIBLE_METADATA = {'metadata_version': '1.1',
|
||||||
|
'status': ["preview"],
|
||||||
|
'supported_by': 'community'}
|
||||||
|
|
||||||
|
DOCUMENTATION = '''
|
||||||
|
---
|
||||||
|
module: gcp_bigquery_table_facts
|
||||||
|
description:
|
||||||
|
- Gather facts for GCP Table
|
||||||
|
short_description: Gather facts for GCP Table
|
||||||
|
version_added: 2.8
|
||||||
|
author: Google Inc. (@googlecloudplatform)
|
||||||
|
requirements:
|
||||||
|
- python >= 2.6
|
||||||
|
- requests >= 2.18.4
|
||||||
|
- google-auth >= 1.3.0
|
||||||
|
options:
|
||||||
|
dataset:
|
||||||
|
description:
|
||||||
|
- Name of the dataset.
|
||||||
|
required: false
|
||||||
|
extends_documentation_fragment: gcp
|
||||||
|
'''
|
||||||
|
|
||||||
|
EXAMPLES = '''
|
||||||
|
- name: a table facts
|
||||||
|
gcp_bigquery_table_facts:
|
||||||
|
dataset: example_dataset
|
||||||
|
project: test_project
|
||||||
|
auth_kind: serviceaccount
|
||||||
|
service_account_file: "/tmp/auth.pem"
|
||||||
|
'''
|
||||||
|
|
||||||
|
RETURN = '''
|
||||||
|
items:
|
||||||
|
description: List of items
|
||||||
|
returned: always
|
||||||
|
type: complex
|
||||||
|
contains:
|
||||||
|
tableReference:
|
||||||
|
description:
|
||||||
|
- Reference describing the ID of this table.
|
||||||
|
returned: success
|
||||||
|
type: complex
|
||||||
|
contains:
|
||||||
|
datasetId:
|
||||||
|
description:
|
||||||
|
- The ID of the dataset containing this table.
|
||||||
|
returned: success
|
||||||
|
type: str
|
||||||
|
projectId:
|
||||||
|
description:
|
||||||
|
- The ID of the project containing this table.
|
||||||
|
returned: success
|
||||||
|
type: str
|
||||||
|
tableId:
|
||||||
|
description:
|
||||||
|
- The ID of the the table.
|
||||||
|
returned: success
|
||||||
|
type: str
|
||||||
|
creationTime:
|
||||||
|
description:
|
||||||
|
- The time when this dataset was created, in milliseconds since the epoch.
|
||||||
|
returned: success
|
||||||
|
type: int
|
||||||
|
description:
|
||||||
|
description:
|
||||||
|
- A user-friendly description of the dataset.
|
||||||
|
returned: success
|
||||||
|
type: str
|
||||||
|
friendlyName:
|
||||||
|
description:
|
||||||
|
- A descriptive name for this table.
|
||||||
|
returned: success
|
||||||
|
type: str
|
||||||
|
id:
|
||||||
|
description:
|
||||||
|
- An opaque ID uniquely identifying the table.
|
||||||
|
returned: success
|
||||||
|
type: str
|
||||||
|
labels:
|
||||||
|
description:
|
||||||
|
- The labels associated with this dataset. You can use these to organize and
|
||||||
|
group your datasets .
|
||||||
|
returned: success
|
||||||
|
type: dict
|
||||||
|
lastModifiedTime:
|
||||||
|
description:
|
||||||
|
- The time when this table was last modified, in milliseconds since the epoch.
|
||||||
|
returned: success
|
||||||
|
type: int
|
||||||
|
location:
|
||||||
|
description:
|
||||||
|
- The geographic location where the table resides. This value is inherited from
|
||||||
|
the dataset.
|
||||||
|
returned: success
|
||||||
|
type: str
|
||||||
|
name:
|
||||||
|
description:
|
||||||
|
- Name of the table.
|
||||||
|
returned: success
|
||||||
|
type: str
|
||||||
|
numBytes:
|
||||||
|
description:
|
||||||
|
- The size of this table in bytes, excluding any data in the streaming buffer.
|
||||||
|
returned: success
|
||||||
|
type: int
|
||||||
|
numLongTermBytes:
|
||||||
|
description:
|
||||||
|
- The number of bytes in the table that are considered "long-term storage".
|
||||||
|
returned: success
|
||||||
|
type: int
|
||||||
|
numRows:
|
||||||
|
description:
|
||||||
|
- The number of rows of data in this table, excluding any data in the streaming
|
||||||
|
buffer.
|
||||||
|
returned: success
|
||||||
|
type: int
|
||||||
|
type:
|
||||||
|
description:
|
||||||
|
- Describes the table type.
|
||||||
|
returned: success
|
||||||
|
type: str
|
||||||
|
view:
|
||||||
|
description:
|
||||||
|
- The view definition.
|
||||||
|
returned: success
|
||||||
|
type: complex
|
||||||
|
contains:
|
||||||
|
useLegacySql:
|
||||||
|
description:
|
||||||
|
- Specifies whether to use BigQuery's legacy SQL for this view .
|
||||||
|
returned: success
|
||||||
|
type: bool
|
||||||
|
userDefinedFunctionResources:
|
||||||
|
description:
|
||||||
|
- Describes user-defined function resources used in the query.
|
||||||
|
returned: success
|
||||||
|
type: complex
|
||||||
|
contains:
|
||||||
|
inlineCode:
|
||||||
|
description:
|
||||||
|
- An inline resource that contains code for a user-defined function
|
||||||
|
(UDF). Providing a inline code resource is equivalent to providing
|
||||||
|
a URI for a file containing the same code.
|
||||||
|
returned: success
|
||||||
|
type: str
|
||||||
|
resourceUri:
|
||||||
|
description:
|
||||||
|
- A code resource to load from a Google Cloud Storage URI (gs://bucket/path).
|
||||||
|
returned: success
|
||||||
|
type: str
|
||||||
|
timePartitioning:
|
||||||
|
description:
|
||||||
|
- If specified, configures time-based partitioning for this table.
|
||||||
|
returned: success
|
||||||
|
type: complex
|
||||||
|
contains:
|
||||||
|
expirationMs:
|
||||||
|
description:
|
||||||
|
- Number of milliseconds for which to keep the storage for a partition.
|
||||||
|
returned: success
|
||||||
|
type: int
|
||||||
|
type:
|
||||||
|
description:
|
||||||
|
- The only type supported is DAY, which will generate one partition per
|
||||||
|
day.
|
||||||
|
returned: success
|
||||||
|
type: str
|
||||||
|
streamingBuffer:
|
||||||
|
description:
|
||||||
|
- Contains information regarding this table's streaming buffer, if one is present.
|
||||||
|
This field will be absent if the table is not being streamed to or if there
|
||||||
|
is no data in the streaming buffer.
|
||||||
|
returned: success
|
||||||
|
type: complex
|
||||||
|
contains:
|
||||||
|
estimatedBytes:
|
||||||
|
description:
|
||||||
|
- A lower-bound estimate of the number of bytes currently in the streaming
|
||||||
|
buffer.
|
||||||
|
returned: success
|
||||||
|
type: int
|
||||||
|
estimatedRows:
|
||||||
|
description:
|
||||||
|
- A lower-bound estimate of the number of rows currently in the streaming
|
||||||
|
buffer.
|
||||||
|
returned: success
|
||||||
|
type: int
|
||||||
|
oldestEntryTime:
|
||||||
|
description:
|
||||||
|
- Contains the timestamp of the oldest entry in the streaming buffer, in
|
||||||
|
milliseconds since the epoch, if the streaming buffer is available.
|
||||||
|
returned: success
|
||||||
|
type: int
|
||||||
|
schema:
|
||||||
|
description:
|
||||||
|
- Describes the schema of this table.
|
||||||
|
returned: success
|
||||||
|
type: complex
|
||||||
|
contains:
|
||||||
|
fields:
|
||||||
|
description:
|
||||||
|
- Describes the fields in a table.
|
||||||
|
returned: success
|
||||||
|
type: complex
|
||||||
|
contains:
|
||||||
|
description:
|
||||||
|
description:
|
||||||
|
- The field description. The maximum length is 1,024 characters.
|
||||||
|
returned: success
|
||||||
|
type: str
|
||||||
|
fields:
|
||||||
|
description:
|
||||||
|
- Describes the nested schema fields if the type property is set to
|
||||||
|
RECORD.
|
||||||
|
returned: success
|
||||||
|
type: list
|
||||||
|
mode:
|
||||||
|
description:
|
||||||
|
- The field mode.
|
||||||
|
returned: success
|
||||||
|
type: str
|
||||||
|
name:
|
||||||
|
description:
|
||||||
|
- The field name.
|
||||||
|
returned: success
|
||||||
|
type: str
|
||||||
|
type:
|
||||||
|
description:
|
||||||
|
- The field data type.
|
||||||
|
returned: success
|
||||||
|
type: str
|
||||||
|
encryptionConfiguration:
|
||||||
|
description:
|
||||||
|
- Custom encryption configuration.
|
||||||
|
returned: success
|
||||||
|
type: complex
|
||||||
|
contains:
|
||||||
|
kmsKeyName:
|
||||||
|
description:
|
||||||
|
- Describes the Cloud KMS encryption key that will be used to protect destination
|
||||||
|
BigQuery table. The BigQuery Service Account associated with your project
|
||||||
|
requires access to this encryption key.
|
||||||
|
returned: success
|
||||||
|
type: str
|
||||||
|
expirationTime:
|
||||||
|
description:
|
||||||
|
- The time when this table expires, in milliseconds since the epoch. If not
|
||||||
|
present, the table will persist indefinitely.
|
||||||
|
returned: success
|
||||||
|
type: int
|
||||||
|
externalDataConfiguration:
|
||||||
|
description:
|
||||||
|
- Describes the data format, location, and other properties of a table stored
|
||||||
|
outside of BigQuery. By defining these properties, the data source can then
|
||||||
|
be queried as if it were a standard BigQuery table.
|
||||||
|
returned: success
|
||||||
|
type: complex
|
||||||
|
contains:
|
||||||
|
autodetect:
|
||||||
|
description:
|
||||||
|
- Try to detect schema and format options automatically. Any option specified
|
||||||
|
explicitly will be honored.
|
||||||
|
returned: success
|
||||||
|
type: bool
|
||||||
|
compression:
|
||||||
|
description:
|
||||||
|
- The compression type of the data source.
|
||||||
|
returned: success
|
||||||
|
type: str
|
||||||
|
ignoreUnknownValues:
|
||||||
|
description:
|
||||||
|
- Indicates if BigQuery should allow extra values that are not represented
|
||||||
|
in the table schema .
|
||||||
|
returned: success
|
||||||
|
type: bool
|
||||||
|
maxBadRecords:
|
||||||
|
description:
|
||||||
|
- The maximum number of bad records that BigQuery can ignore when reading
|
||||||
|
data .
|
||||||
|
returned: success
|
||||||
|
type: int
|
||||||
|
sourceFormat:
|
||||||
|
description:
|
||||||
|
- The data format.
|
||||||
|
returned: success
|
||||||
|
type: str
|
||||||
|
sourceUris:
|
||||||
|
description:
|
||||||
|
- 'The fully-qualified URIs that point to your data in Google Cloud. For
|
||||||
|
Google Cloud Storage URIs: Each URI can contain one ''*'' wildcard character
|
||||||
|
and it must come after the ''bucket'' name. Size limits related to load
|
||||||
|
jobs apply to external data sources. For Google Cloud Bigtable URIs: Exactly
|
||||||
|
one URI can be specified and it has be a fully specified and valid HTTPS
|
||||||
|
URL for a Google Cloud Bigtable table. For Google Cloud Datastore backups,
|
||||||
|
exactly one URI can be specified. Also, the ''*'' wildcard character is
|
||||||
|
not allowed.'
|
||||||
|
returned: success
|
||||||
|
type: list
|
||||||
|
schema:
|
||||||
|
description:
|
||||||
|
- The schema for the data. Schema is required for CSV and JSON formats.
|
||||||
|
returned: success
|
||||||
|
type: complex
|
||||||
|
contains:
|
||||||
|
fields:
|
||||||
|
description:
|
||||||
|
- Describes the fields in a table.
|
||||||
|
returned: success
|
||||||
|
type: complex
|
||||||
|
contains:
|
||||||
|
description:
|
||||||
|
description:
|
||||||
|
- The field description.
|
||||||
|
returned: success
|
||||||
|
type: str
|
||||||
|
fields:
|
||||||
|
description:
|
||||||
|
- Describes the nested schema fields if the type property is set
|
||||||
|
to RECORD .
|
||||||
|
returned: success
|
||||||
|
type: list
|
||||||
|
mode:
|
||||||
|
description:
|
||||||
|
- Field mode.
|
||||||
|
returned: success
|
||||||
|
type: str
|
||||||
|
name:
|
||||||
|
description:
|
||||||
|
- Field name.
|
||||||
|
returned: success
|
||||||
|
type: str
|
||||||
|
type:
|
||||||
|
description:
|
||||||
|
- Field data type.
|
||||||
|
returned: success
|
||||||
|
type: str
|
||||||
|
googleSheetsOptions:
|
||||||
|
description:
|
||||||
|
- Additional options if sourceFormat is set to GOOGLE_SHEETS.
|
||||||
|
returned: success
|
||||||
|
type: complex
|
||||||
|
contains:
|
||||||
|
skipLeadingRows:
|
||||||
|
description:
|
||||||
|
- The number of rows at the top of a Google Sheet that BigQuery will
|
||||||
|
skip when reading the data.
|
||||||
|
returned: success
|
||||||
|
type: int
|
||||||
|
csvOptions:
|
||||||
|
description:
|
||||||
|
- Additional properties to set if sourceFormat is set to CSV.
|
||||||
|
returned: success
|
||||||
|
type: complex
|
||||||
|
contains:
|
||||||
|
allowJaggedRows:
|
||||||
|
description:
|
||||||
|
- Indicates if BigQuery should accept rows that are missing trailing
|
||||||
|
optional columns .
|
||||||
|
returned: success
|
||||||
|
type: bool
|
||||||
|
allowQuotedNewlines:
|
||||||
|
description:
|
||||||
|
- Indicates if BigQuery should allow quoted data sections that contain
|
||||||
|
newline characters in a CSV file .
|
||||||
|
returned: success
|
||||||
|
type: bool
|
||||||
|
encoding:
|
||||||
|
description:
|
||||||
|
- The character encoding of the data.
|
||||||
|
returned: success
|
||||||
|
type: str
|
||||||
|
fieldDelimiter:
|
||||||
|
description:
|
||||||
|
- The separator for fields in a CSV file.
|
||||||
|
returned: success
|
||||||
|
type: str
|
||||||
|
quote:
|
||||||
|
description:
|
||||||
|
- The value that is used to quote data sections in a CSV file.
|
||||||
|
returned: success
|
||||||
|
type: str
|
||||||
|
skipLeadingRows:
|
||||||
|
description:
|
||||||
|
- The number of rows at the top of a CSV file that BigQuery will skip
|
||||||
|
when reading the data.
|
||||||
|
returned: success
|
||||||
|
type: int
|
||||||
|
bigtableOptions:
|
||||||
|
description:
|
||||||
|
- Additional options if sourceFormat is set to BIGTABLE.
|
||||||
|
returned: success
|
||||||
|
type: complex
|
||||||
|
contains:
|
||||||
|
ignoreUnspecifiedColumnFamilies:
|
||||||
|
description:
|
||||||
|
- If field is true, then the column families that are not specified
|
||||||
|
in columnFamilies list are not exposed in the table schema .
|
||||||
|
returned: success
|
||||||
|
type: bool
|
||||||
|
readRowkeyAsString:
|
||||||
|
description:
|
||||||
|
- If field is true, then the rowkey column families will be read and
|
||||||
|
converted to string.
|
||||||
|
returned: success
|
||||||
|
type: bool
|
||||||
|
columnFamilies:
|
||||||
|
description:
|
||||||
|
- List of column families to expose in the table schema along with their
|
||||||
|
types.
|
||||||
|
returned: success
|
||||||
|
type: complex
|
||||||
|
contains:
|
||||||
|
columns:
|
||||||
|
description:
|
||||||
|
- Lists of columns that should be exposed as individual fields as
|
||||||
|
opposed to a list of (column name, value) pairs.
|
||||||
|
returned: success
|
||||||
|
type: complex
|
||||||
|
contains:
|
||||||
|
encoding:
|
||||||
|
description:
|
||||||
|
- The encoding of the values when the type is not STRING.
|
||||||
|
returned: success
|
||||||
|
type: str
|
||||||
|
fieldName:
|
||||||
|
description:
|
||||||
|
- If the qualifier is not a valid BigQuery field identifier,
|
||||||
|
a valid identifier must be provided as the column field name
|
||||||
|
and is used as field name in queries.
|
||||||
|
returned: success
|
||||||
|
type: str
|
||||||
|
onlyReadLatest:
|
||||||
|
description:
|
||||||
|
- If this is set, only the latest version of value in this column
|
||||||
|
are exposed .
|
||||||
|
returned: success
|
||||||
|
type: bool
|
||||||
|
qualifierString:
|
||||||
|
description:
|
||||||
|
- Qualifier of the column.
|
||||||
|
returned: success
|
||||||
|
type: str
|
||||||
|
type:
|
||||||
|
description:
|
||||||
|
- The type to convert the value in cells of this column.
|
||||||
|
returned: success
|
||||||
|
type: str
|
||||||
|
encoding:
|
||||||
|
description:
|
||||||
|
- The encoding of the values when the type is not STRING.
|
||||||
|
returned: success
|
||||||
|
type: str
|
||||||
|
familyId:
|
||||||
|
description:
|
||||||
|
- Identifier of the column family.
|
||||||
|
returned: success
|
||||||
|
type: str
|
||||||
|
onlyReadLatest:
|
||||||
|
description:
|
||||||
|
- If this is set only the latest version of value are exposed for
|
||||||
|
all columns in this column family .
|
||||||
|
returned: success
|
||||||
|
type: bool
|
||||||
|
type:
|
||||||
|
description:
|
||||||
|
- The type to convert the value in cells of this column family.
|
||||||
|
returned: success
|
||||||
|
type: str
|
||||||
|
dataset:
|
||||||
|
description:
|
||||||
|
- Name of the dataset.
|
||||||
|
returned: success
|
||||||
|
type: str
|
||||||
|
'''
|
||||||
|
|
||||||
|
################################################################################
|
||||||
|
# Imports
|
||||||
|
################################################################################
|
||||||
|
from ansible.module_utils.gcp_utils import navigate_hash, GcpSession, GcpModule, GcpRequest
|
||||||
|
import json
|
||||||
|
|
||||||
|
################################################################################
|
||||||
|
# Main
|
||||||
|
################################################################################
|
||||||
|
|
||||||
|
|
||||||
|
def main():
|
||||||
|
module = GcpModule(
|
||||||
|
argument_spec=dict(
|
||||||
|
dataset=dict(type='str')
|
||||||
|
)
|
||||||
|
)
|
||||||
|
|
||||||
|
if not module.params['scopes']:
|
||||||
|
module.params['scopes'] = ['https://www.googleapis.com/auth/bigquery']
|
||||||
|
|
||||||
|
items = fetch_list(module, collection(module))
|
||||||
|
if items.get('tables'):
|
||||||
|
items = items.get('tables')
|
||||||
|
else:
|
||||||
|
items = []
|
||||||
|
return_value = {
|
||||||
|
'items': items
|
||||||
|
}
|
||||||
|
module.exit_json(**return_value)
|
||||||
|
|
||||||
|
|
||||||
|
def collection(module):
|
||||||
|
return "https://www.googleapis.com/bigquery/v2/projects/{project}/datasets/{dataset}/tables".format(**module.params)
|
||||||
|
|
||||||
|
|
||||||
|
def fetch_list(module, link):
|
||||||
|
auth = GcpSession(module, 'bigquery')
|
||||||
|
response = auth.get(link)
|
||||||
|
return return_if_object(module, response)
|
||||||
|
|
||||||
|
|
||||||
|
def return_if_object(module, response):
|
||||||
|
# If not found, return nothing.
|
||||||
|
if response.status_code == 404:
|
||||||
|
return None
|
||||||
|
|
||||||
|
# If no content, return nothing.
|
||||||
|
if response.status_code == 204:
|
||||||
|
return None
|
||||||
|
|
||||||
|
try:
|
||||||
|
module.raise_for_status(response)
|
||||||
|
result = response.json()
|
||||||
|
except getattr(json.decoder, 'JSONDecodeError', ValueError) as inst:
|
||||||
|
module.fail_json(msg="Invalid JSON response with error: %s" % inst)
|
||||||
|
|
||||||
|
if navigate_hash(result, ['error', 'errors']):
|
||||||
|
module.fail_json(msg=navigate_hash(result, ['error', 'errors']))
|
||||||
|
|
||||||
|
return result
|
||||||
|
|
||||||
|
|
||||||
|
if __name__ == "__main__":
|
||||||
|
main()
|
Loading…
Reference in a new issue