code stringlengths 22 1.05M | apis listlengths 1 3.31k | extract_api stringlengths 75 3.25M |
|---|---|---|
import typer
import deciphon_cli.data as data
__all__ = ["app"]
app = typer.Typer()
@app.command()
def default():
typer.echo(data.env_example_content(), nl=False)
| [
"deciphon_cli.data.env_example_content",
"typer.Typer"
] | [((73, 86), 'typer.Typer', 'typer.Typer', ([], {}), '()\n', (84, 86), False, 'import typer\n'), ((134, 160), 'deciphon_cli.data.env_example_content', 'data.env_example_content', ([], {}), '()\n', (158, 160), True, 'import deciphon_cli.data as data\n')] |
import olll
import numpy as np
test1 = [[1,0,0,1,1,0,1],[0,1,0,5,0,0,0],[0,0,1,0,5,0,5]]
test2 = [[1,0,0,2,-1,1],[0,1,0,3,-4,-2],[0,0,1,5,-10,-8]]
test3 = [[1,0,0,1,1,0,1], [0,1,0,4,-1,0,-1], [0,0,1,1,1,0,1]]
test4 = [[1,0,0,2,5,3],[0,1,0,1,1,1,],[0,0,1,4,-2,0]]
test5 = [[1,0,0,0,0,0,2,1,1,2],[0,1,0,0,0,0,1,1,-1,-1],[0,0,1,0,0,0,-1,0,-2,-3],[0,0,0,1,0,0,1,-1,1,-1],[0,0,0,0,1,0,-1,2,-4,-3],[0,0,0,0,0,1,1,0,0,1]]
test6 = [[1, 0, 0, 5, 0, 0, 0],[0, 1, 0, 0, 5, 0, 5],[0, 0, 1, 1, 1, 0, 1]]
test7 = [[1, 0, 0, 20, 0, 0, 0],[0, 1, 0, 0, 20, 0, 20],[0, 0, 1, 4, 4, 0, 4]]
test8 = [[1, 0, 0, 10, 0, 0, 0],[0, 1, 0, 0, 10, 0, 10],[0, 0, 1, 2, 2, 0, 2]]
n = input("Please enter n: \n")
n = int(n)
k = input("Please enter k: \n")
k = int(k)
p = input("Please enter p: \n")
p = int(p)
id = np.identity(k)
A = [[]] * k
print("Please enter the generating set:\n")
for i in range(k):
print("\nEnter the generator a[",i,"]: ")
a = list(map(int,input().strip().split()))[:n]
#print(i, a)
a = [x * (2**p) for x in a]
y = list(id[i])
print(y[i],type(y[i]))
A[i] = y+a
print(A[i], type(A[i]))
print(A, type(A))
print(test7, type(test7))
rb = olll.reduction(test7,0.75)
print("Basis: ", rb)
| [
"numpy.identity",
"olll.reduction"
] | [((787, 801), 'numpy.identity', 'np.identity', (['k'], {}), '(k)\n', (798, 801), True, 'import numpy as np\n'), ((1169, 1196), 'olll.reduction', 'olll.reduction', (['test7', '(0.75)'], {}), '(test7, 0.75)\n', (1183, 1196), False, 'import olll\n')] |
from __future__ import annotations
import logging
from typing import Any
from homeassistant.components.integration.sensor import (
TRAPEZOIDAL_METHOD,
IntegrationSensor,
)
from homeassistant.components.sensor import DOMAIN as SENSOR_DOMAIN
from homeassistant.const import CONF_NAME, TIME_HOURS
from homeassistant.helpers.entity import async_generate_entity_id
from homeassistant.helpers.typing import HomeAssistantType
from custom_components.powercalc.common import SourceEntity
from custom_components.powercalc.const import (
ATTR_SOURCE_DOMAIN,
ATTR_SOURCE_ENTITY,
CONF_ENERGY_SENSOR_NAMING,
)
from custom_components.powercalc.migrate import async_migrate_entity_id
from .power import VirtualPowerSensor
ENERGY_ICON = "mdi:lightning-bolt"
ENTITY_ID_FORMAT = SENSOR_DOMAIN + ".{}"
_LOGGER = logging.getLogger(__name__)
async def create_energy_sensor(
hass: HomeAssistantType,
sensor_config: dict,
power_sensor: VirtualPowerSensor,
source_entity: SourceEntity,
) -> VirtualEnergySensor:
"""Create the energy sensor entity"""
name_pattern = sensor_config.get(CONF_ENERGY_SENSOR_NAMING)
name = sensor_config.get(CONF_NAME) or source_entity.name
name = name_pattern.format(name)
object_id = sensor_config.get(CONF_NAME) or source_entity.object_id
entity_id = async_generate_entity_id(
ENTITY_ID_FORMAT, name_pattern.format(object_id), hass=hass
)
unique_id = None
if source_entity.unique_id:
unique_id = f"{source_entity.unique_id}_energy"
async_migrate_entity_id(hass, "sensor", unique_id, entity_id)
_LOGGER.debug("Creating energy sensor: %s", name)
return VirtualEnergySensor(
source_entity=power_sensor.entity_id,
unique_id=unique_id,
entity_id=entity_id,
name=name,
round_digits=4,
unit_prefix="k",
unit_of_measurement=None,
unit_time=TIME_HOURS,
integration_method=TRAPEZOIDAL_METHOD,
powercalc_source_entity=source_entity.entity_id,
powercalc_source_domain=source_entity.domain,
)
class VirtualEnergySensor(IntegrationSensor):
"""Virtual energy sensor, totalling kWh"""
def __init__(
self,
source_entity,
unique_id,
entity_id,
name,
round_digits,
unit_prefix,
unit_time,
unit_of_measurement,
integration_method,
powercalc_source_entity: str,
powercalc_source_domain: str,
):
super().__init__(
source_entity,
name,
round_digits,
unit_prefix,
unit_time,
unit_of_measurement,
integration_method,
)
self._powercalc_source_entity = powercalc_source_entity
self._powercalc_source_domain = powercalc_source_domain
self.entity_id = entity_id
if unique_id:
self._attr_unique_id = unique_id
@property
def extra_state_attributes(self) -> dict[str, Any]:
"""Return the state attributes of the acceleration sensor."""
state_attr = super().extra_state_attributes
state_attr[ATTR_SOURCE_ENTITY] = self._powercalc_source_entity
state_attr[ATTR_SOURCE_DOMAIN] = self._powercalc_source_domain
return state_attr
@property
def icon(self):
return ENERGY_ICON
| [
"logging.getLogger",
"custom_components.powercalc.migrate.async_migrate_entity_id"
] | [((818, 845), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (835, 845), False, 'import logging\n'), ((1542, 1603), 'custom_components.powercalc.migrate.async_migrate_entity_id', 'async_migrate_entity_id', (['hass', '"""sensor"""', 'unique_id', 'entity_id'], {}), "(hass, 'sensor', unique_id, entity_id)\n", (1565, 1603), False, 'from custom_components.powercalc.migrate import async_migrate_entity_id\n')] |
import numpy as np
import matplotlib.pyplot as plt
import ipywidgets
from mesostat.utils.opencv_helper import cvWriter
from mesostat.utils.arrays import numpy_merge_dimensions
from sklearn.decomposition import PCA
def distance_matrix(data):
nDim, nTime = data.shape
dataExtr = np.repeat(data[..., None], nTime, axis=2)
delta = dataExtr - dataExtr.transpose((0,2,1))
return np.linalg.norm(delta, axis=0)
def decorrelate(data):
# Leading dimension must be channels
if data.ndim > 2:
dataEff = numpy_merge_dimensions(data, 1, data.ndim+1)
else:
dataEff = data
pca = PCA(n_components=48)
rez = pca.fit_transform(dataEff.T)
print(rez.shape)
rez /= np.std(rez, axis=0)
return rez.T.reshape(data.shape)
class RecurrencePlot:
def __init__(self, data, w0):
self.nPoint, self.nTime = data.shape
# Compute the recurrence plots with no threshold
self.dist = distance_matrix(data)
# Plot first
picdata = self.get_plotdata(w0)
self.fig, self.ax = plt.subplots(figsize=(5, 5))
self.pic = self.ax.imshow(picdata, cmap='binary', origin='lower')
self.ax.set_title('Recurrence Plot', fontsize=16)
def get_plotdata(self, w):
thr = np.percentile(self.dist.flatten(), w)
return self.dist <= thr
def update(self, w):
picnew = self.get_plotdata(w)
self.pic.set_data(picnew)
self.fig.canvas.draw_idle()
def interact(self):
ipywidgets.interact(self.update, w=(0, 100, 1))
def write_video(self, fname, frate=10.0, codec='XVID'):
frameDim = (self.nTime, self.nTime)
with cvWriter(fname, frameDim, frate=frate, codec=codec) as writer:
for w in range(101):
data = self.get_plotdata(w)
writer.write(data)
class RecurrencePlotMultitrial:
def __init__(self, data3D, w0):
self.nTrial, self.nPoint, self.nTime = data3D.shape
binarize = lambda dist, w: dist <= np.percentile(dist, w)
# Compute the recurrence plots with no threshold
self.rezMat = np.zeros((101, self.nTime, self.nTime), dtype=float)
for data in data3D:
dist = distance_matrix(data)
for w in range(101):
self.rezMat[w] += binarize(dist, w).astype(float)
self.rezMat /= self.nTrial
# Plot first
self.fig, self.ax = plt.subplots(figsize=(5, 5))
self.pic = self.ax.imshow(self.rezMat[w0], cmap='viridis', origin='lower', vmin=0, vmax=1)
self.ax.set_title('Recurrence Plot', fontsize=16)
def update(self, w):
self.pic.set_data(self.rezMat[w])
self.fig.canvas.draw_idle()
def interact(self):
ipywidgets.interact(self.update, w=(0, 100, 1))
def write_video(self, fname, frate=10.0, codec='XVID'):
frameDim = (self.nTime, self.nTime)
with cvWriter(fname, frameDim, frate=frate, codec=codec) as writer:
for w in range(101):
writer.write(self.rezMat[w])
| [
"numpy.repeat",
"sklearn.decomposition.PCA",
"numpy.linalg.norm",
"numpy.zeros",
"mesostat.utils.opencv_helper.cvWriter",
"ipywidgets.interact",
"numpy.std",
"numpy.percentile",
"mesostat.utils.arrays.numpy_merge_dimensions",
"matplotlib.pyplot.subplots"
] | [((288, 329), 'numpy.repeat', 'np.repeat', (['data[..., None]', 'nTime'], {'axis': '(2)'}), '(data[..., None], nTime, axis=2)\n', (297, 329), True, 'import numpy as np\n'), ((392, 421), 'numpy.linalg.norm', 'np.linalg.norm', (['delta'], {'axis': '(0)'}), '(delta, axis=0)\n', (406, 421), True, 'import numpy as np\n'), ((617, 637), 'sklearn.decomposition.PCA', 'PCA', ([], {'n_components': '(48)'}), '(n_components=48)\n', (620, 637), False, 'from sklearn.decomposition import PCA\n'), ((711, 730), 'numpy.std', 'np.std', (['rez'], {'axis': '(0)'}), '(rez, axis=0)\n', (717, 730), True, 'import numpy as np\n'), ((528, 574), 'mesostat.utils.arrays.numpy_merge_dimensions', 'numpy_merge_dimensions', (['data', '(1)', '(data.ndim + 1)'], {}), '(data, 1, data.ndim + 1)\n', (550, 574), False, 'from mesostat.utils.arrays import numpy_merge_dimensions\n'), ((1061, 1089), 'matplotlib.pyplot.subplots', 'plt.subplots', ([], {'figsize': '(5, 5)'}), '(figsize=(5, 5))\n', (1073, 1089), True, 'import matplotlib.pyplot as plt\n'), ((1505, 1552), 'ipywidgets.interact', 'ipywidgets.interact', (['self.update'], {'w': '(0, 100, 1)'}), '(self.update, w=(0, 100, 1))\n', (1524, 1552), False, 'import ipywidgets\n'), ((2123, 2175), 'numpy.zeros', 'np.zeros', (['(101, self.nTime, self.nTime)'], {'dtype': 'float'}), '((101, self.nTime, self.nTime), dtype=float)\n', (2131, 2175), True, 'import numpy as np\n'), ((2430, 2458), 'matplotlib.pyplot.subplots', 'plt.subplots', ([], {'figsize': '(5, 5)'}), '(figsize=(5, 5))\n', (2442, 2458), True, 'import matplotlib.pyplot as plt\n'), ((2753, 2800), 'ipywidgets.interact', 'ipywidgets.interact', (['self.update'], {'w': '(0, 100, 1)'}), '(self.update, w=(0, 100, 1))\n', (2772, 2800), False, 'import ipywidgets\n'), ((1672, 1723), 'mesostat.utils.opencv_helper.cvWriter', 'cvWriter', (['fname', 'frameDim'], {'frate': 'frate', 'codec': 'codec'}), '(fname, frameDim, frate=frate, codec=codec)\n', (1680, 1723), False, 'from mesostat.utils.opencv_helper import cvWriter\n'), ((2920, 2971), 'mesostat.utils.opencv_helper.cvWriter', 'cvWriter', (['fname', 'frameDim'], {'frate': 'frate', 'codec': 'codec'}), '(fname, frameDim, frate=frate, codec=codec)\n', (2928, 2971), False, 'from mesostat.utils.opencv_helper import cvWriter\n'), ((2020, 2042), 'numpy.percentile', 'np.percentile', (['dist', 'w'], {}), '(dist, w)\n', (2033, 2042), True, 'import numpy as np\n')] |
import sys
import os
import numpy as np
import scipy.io as sio
import random
from decimal import Decimal
import argparse
import csv
from keras.models import load_model
import f_model
from f_preprocess import fill_length
# Usage: python rematch_challenge.py test_file_path
def arg_parse():
"""
Parse arguements
"""
parser = argparse.ArgumentParser(description='Rematch test of ECG Contest')
parser.add_argument("--test_path", dest='test_path', help=
"the file path of Test Data",
default="your test_path", type=str)
#You need to write your test data path with the argparse parameter.
#For your convenience when testing with local data, you can write your local test set path to default
return parser.parse_args()
def main():
args = arg_parse()
test_path = args.test_path
print(test_path)
## Add your codes to classify normal and diseases.
model01_path = './model/model_01.h5'
#model02_path = '/media/uuser/data/final_run/model/model_02.h5'
#modelxg_path = './model/model.pkl'
#feature_path = './data/feature.csv'
keysname = ('I','II','III','aVR','aVL','aVF', \
'V1','V2','V3','V4','V5','V6','age','sex')
t_len = 25000
len_target=t_len
model_01 = f_model.build_model_01(num_classes=10,len_target=len_target)
model_01.load_weights(model01_path)
## Classify the samples of the test set and write the results into answers.txt,
## and each row representing a prediction of one sample.
## Here we use random numbers as prediction labels as an example and
## you should replace it with your own results.
Data_list = os.listdir(test_path)
classes = np.asarray([1,1,2,3,4,5,6,7,8,9])
with open('answers.csv', 'w') as csvfile:
writer = csv.writer(csvfile)
writer.writerow(['File_name', 'label1', 'label2', 'label3', 'label4', 'label5', 'label6', 'label7', 'label8', 'label9', 'label10'])
for file_name in Data_list:
if file_name.endswith('.mat'):
answer = []
record_name = file_name.strip('.mat')
answer.append(record_name)
# model 01
ecg = np.empty([t_len,12])
mypath=test_path+file_name
data = sio.loadmat(mypath)
# read 12 leads
for lead in range(12):
temp=data[keysname[lead]]
ecg[:,lead] = fill_length(temp,t_len)
data_x = ecg.reshape((1,t_len,12))
pred_1 = model_01.predict(data_x)
# model 02
# model xgboost
preds = pred_1
preds[preds>=0.5] = 1
preds[preds<0.5] = 0
pred_out = preds * classes
y_out =[]
for i in range(10):
if pred_out[0][i]==classes[i]:
y_out.append(i)
for x in range(10-len(y_out)):
y_out.append('')
writer.writerow(answer+y_out)
csvfile.close()
if __name__ == "__main__":
main()
| [
"f_preprocess.fill_length",
"os.listdir",
"argparse.ArgumentParser",
"csv.writer",
"numpy.asarray",
"scipy.io.loadmat",
"numpy.empty",
"f_model.build_model_01"
] | [((343, 409), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {'description': '"""Rematch test of ECG Contest"""'}), "(description='Rematch test of ECG Contest')\n", (366, 409), False, 'import argparse\n'), ((1293, 1354), 'f_model.build_model_01', 'f_model.build_model_01', ([], {'num_classes': '(10)', 'len_target': 'len_target'}), '(num_classes=10, len_target=len_target)\n', (1315, 1354), False, 'import f_model\n'), ((1685, 1706), 'os.listdir', 'os.listdir', (['test_path'], {}), '(test_path)\n', (1695, 1706), False, 'import os\n'), ((1721, 1763), 'numpy.asarray', 'np.asarray', (['[1, 1, 2, 3, 4, 5, 6, 7, 8, 9]'], {}), '([1, 1, 2, 3, 4, 5, 6, 7, 8, 9])\n', (1731, 1763), True, 'import numpy as np\n'), ((1818, 1837), 'csv.writer', 'csv.writer', (['csvfile'], {}), '(csvfile)\n', (1828, 1837), False, 'import csv\n'), ((2232, 2253), 'numpy.empty', 'np.empty', (['[t_len, 12]'], {}), '([t_len, 12])\n', (2240, 2253), True, 'import numpy as np\n'), ((2319, 2338), 'scipy.io.loadmat', 'sio.loadmat', (['mypath'], {}), '(mypath)\n', (2330, 2338), True, 'import scipy.io as sio\n'), ((2490, 2514), 'f_preprocess.fill_length', 'fill_length', (['temp', 't_len'], {}), '(temp, t_len)\n', (2501, 2514), False, 'from f_preprocess import fill_length\n')] |
#!/usr/bin/env python
# coding=utf-8
# This is free and unencumbered software released into the public domain.
#
# Anyone is free to copy, modify, publish, use, compile, sell, or
# distribute this software, either in source code form or as a compiled
# binary, for any purpose, commercial or non-commercial, and by any
# means.
#
# In jurisdictions that recognize copyright laws, the author or authors
# of this software dedicate any and all copyright interest in the
# software to the public domain. We make this dedication for the benefit
# of the public at large and to the detriment of our heirs and
# successors. We intend this dedication to be an overt act of
# relinquishment in perpetuity of all present and future rights to this
# software under copyright law.
#
# THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
# EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
# MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.
# IN NO EVENT SHALL THE AUTHORS BE LIABLE FOR ANY CLAIM, DAMAGES OR
# OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
# ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
# OTHER DEALINGS IN THE SOFTWARE.
#
# For more information, please refer to <http://unlicense.org/>
"""
This script demonstrates how to use the bmreports.UnitData class to
get information about Constraint Payments.
It does not attempt to cater for long or short days and simply assumes
that there will be 48 settlement periods.
"""
import argparse
from datetime import datetime, timedelta, date
from pywind.bmreports.unit import UnitData
def mkdate(datestr):
return datetime.strptime(datestr, '%Y-%m-%d').date()
def main():
parser = argparse.ArgumentParser(description='Get Constraint Payment information for yesterday')
parser.add_argument('--date', action='store', type=mkdate, help='Date to get results for')
parser.add_argument('--period', action='store', help='Period to get data for')
args = parser.parse_args()
data = {}
ud = UnitData({'date': args.date or date.today() - timedelta(days=2)})
pr = [args.period] or range(1,49)
for period in pr:
ud.period = period
if ud.get_data():
data[period] = ud.data
else:
print ("Unable to get data for %s, period %d" % (ud.date.strftime("%d %b %Y"), period))
for period, units in sorted(data.iteritems()):
print ("Period: ", period)
for unit in sorted(units, key=lambda x: x['ngc']):
print (" ", unit['ngc'], unit['lead'])
if unit['bid'].has_key('volume'):
print (" BID: ", unit['bid']['volume']+'MWh ', unit['bid']['cashflow'])
if unit['offer'].has_key('volume'):
print (" OFFER: ", unit['offer']['volume']+'MWh ', unit['offer']['cashflow'])
if __name__ == '__main__':
main()
| [
"datetime.datetime.strptime",
"datetime.date.today",
"argparse.ArgumentParser",
"datetime.timedelta"
] | [((1746, 1838), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {'description': '"""Get Constraint Payment information for yesterday"""'}), "(description=\n 'Get Constraint Payment information for yesterday')\n", (1769, 1838), False, 'import argparse\n'), ((1673, 1711), 'datetime.datetime.strptime', 'datetime.strptime', (['datestr', '"""%Y-%m-%d"""'], {}), "(datestr, '%Y-%m-%d')\n", (1690, 1711), False, 'from datetime import datetime, timedelta, date\n'), ((2098, 2110), 'datetime.date.today', 'date.today', ([], {}), '()\n', (2108, 2110), False, 'from datetime import datetime, timedelta, date\n'), ((2113, 2130), 'datetime.timedelta', 'timedelta', ([], {'days': '(2)'}), '(days=2)\n', (2122, 2130), False, 'from datetime import datetime, timedelta, date\n')] |
"""Run a minimal bot."""
import nonebot
from nonebot.adapters.cqhttp import Bot
nonebot.init()
app = nonebot.get_asgi()
driver = nonebot.get_driver()
driver.register_adapter("cqhttp", Bot)
nonebot.load_builtin_plugins()
nonebot.load_plugins("nonebot_plugin_guess")
if __name__ == "__main__":
# nonebot.run()
# nonebot.run(app="bot:app")
nonebot.run(app="minimal_bot:app")
| [
"nonebot.run",
"nonebot.load_builtin_plugins",
"nonebot.get_driver",
"nonebot.load_plugins",
"nonebot.get_asgi",
"nonebot.init"
] | [((81, 95), 'nonebot.init', 'nonebot.init', ([], {}), '()\n', (93, 95), False, 'import nonebot\n'), ((102, 120), 'nonebot.get_asgi', 'nonebot.get_asgi', ([], {}), '()\n', (118, 120), False, 'import nonebot\n'), ((131, 151), 'nonebot.get_driver', 'nonebot.get_driver', ([], {}), '()\n', (149, 151), False, 'import nonebot\n'), ((192, 222), 'nonebot.load_builtin_plugins', 'nonebot.load_builtin_plugins', ([], {}), '()\n', (220, 222), False, 'import nonebot\n'), ((223, 267), 'nonebot.load_plugins', 'nonebot.load_plugins', (['"""nonebot_plugin_guess"""'], {}), "('nonebot_plugin_guess')\n", (243, 267), False, 'import nonebot\n'), ((353, 387), 'nonebot.run', 'nonebot.run', ([], {'app': '"""minimal_bot:app"""'}), "(app='minimal_bot:app')\n", (364, 387), False, 'import nonebot\n')] |
from lib.appController import driver_queue
from lib.pyapp import Pyapp
import threading
# driver 多线程运行是进行线程之间的数据隔离
local = threading.local()
# 配置在实例化时,去mq中获取创建好的driver,如果调试page则需要传递driver
class BasePage(object):
def __init__(self, driver=None):
if driver is None:
local.driver = driver_queue.get()
local.pyapp = Pyapp(local.driver)
else:
local.driver = driver
local.pyapp = Pyapp(driver)
def quit(self):
local.app.quit()
def reset_package(self):
local.pyapp.reset()
class QQ_Login_Page(BasePage):
def login(self):
local.pyapp.click('android=>new UiSelector().text("登 录")')
def username(self):
local.pyapp.type('content=>请输入QQ号码或手机或邮箱', 3408467505)
def passwd(self):
local.pyapp.type('content=>密码 安全', '<PASSWORD>')
def login_check(self, name):
return local.pyapp.wait_and_save_exception('android=>new UiSelector().text("登 录")', name)
class Page(QQ_Login_Page):
pass
| [
"threading.local",
"lib.pyapp.Pyapp",
"lib.appController.driver_queue.get"
] | [((124, 141), 'threading.local', 'threading.local', ([], {}), '()\n', (139, 141), False, 'import threading\n'), ((306, 324), 'lib.appController.driver_queue.get', 'driver_queue.get', ([], {}), '()\n', (322, 324), False, 'from lib.appController import driver_queue\n'), ((351, 370), 'lib.pyapp.Pyapp', 'Pyapp', (['local.driver'], {}), '(local.driver)\n', (356, 370), False, 'from lib.pyapp import Pyapp\n'), ((445, 458), 'lib.pyapp.Pyapp', 'Pyapp', (['driver'], {}), '(driver)\n', (450, 458), False, 'from lib.pyapp import Pyapp\n')] |
from NoviCypher import FileCypher
from os import path
import sys
Version = '1.0.0b'
def encode(p, r, k, c):
print(f"encoding {p}\nrows={r}\nkey={k}\nchunk={c}")
FileCypher(p, rows=r, chunk=c, key=k)
input()
def decode(p, k):
print(f"Decoding {p}\nkey={k}")
if not FileCypher.decrypt_file(p, k):
input()
arg = sys.argv
len_argv = len(arg)
if len_argv > 1:
if arg[1] in ['-h', '--help']:
print(f"folder/File rows Cypher_size")
quit()
a = arg[1]
else:
a = input('File/Folder >')
if not path.exists(a):
print(f"Path doesn't exist: {a}")
quit()
key = input('Key >')
if path.isfile(a):
a, b = path.splitext(a)
if b != '.ncy':
print(f'File must be .ncy but it is {b}')
decode(a, key)
elif path.isdir(a):
if len_argv > 2:
row = arg[2]
else:
row = input('Number of rows >')
if len_argv > 3:
chunk = arg[3]
else:
chunk = input('Cypher size >')
encode(a, int(row), key, int(chunk))
else:
print("Unknown path")
| [
"os.path.exists",
"NoviCypher.FileCypher.decrypt_file",
"os.path.splitext",
"os.path.isfile",
"os.path.isdir",
"NoviCypher.FileCypher"
] | [((588, 602), 'os.path.isfile', 'path.isfile', (['a'], {}), '(a)\n', (599, 602), False, 'from os import path\n'), ((167, 204), 'NoviCypher.FileCypher', 'FileCypher', (['p'], {'rows': 'r', 'chunk': 'c', 'key': 'k'}), '(p, rows=r, chunk=c, key=k)\n', (177, 204), False, 'from NoviCypher import FileCypher\n'), ((505, 519), 'os.path.exists', 'path.exists', (['a'], {}), '(a)\n', (516, 519), False, 'from os import path\n'), ((612, 628), 'os.path.splitext', 'path.splitext', (['a'], {}), '(a)\n', (625, 628), False, 'from os import path\n'), ((711, 724), 'os.path.isdir', 'path.isdir', (['a'], {}), '(a)\n', (721, 724), False, 'from os import path\n'), ((275, 304), 'NoviCypher.FileCypher.decrypt_file', 'FileCypher.decrypt_file', (['p', 'k'], {}), '(p, k)\n', (298, 304), False, 'from NoviCypher import FileCypher\n')] |
from typing import List
import cv2
from fellbeast.bounding_box import BoundingBox
from fellbeast.configurations import CHECK_FOR_NEW_FACE_FREQUENCY
from fellbeast.utils import get_closest_coordinate
OPENCV_OBJECT_TRACKERS = {
"csrt": cv2.TrackerCSRT_create,
"kcf": cv2.TrackerKCF_create,
"boosting": cv2.TrackerBoosting_create,
"mil": cv2.TrackerMIL_create,
"tld": cv2.TrackerTLD_create,
"medianflow": cv2.TrackerMedianFlow_create,
"mosse": cv2.TrackerMOSSE_create
}
class BaseObjectTracker(object):
bounding_box = None
def init(self, _, bounding_box):
raise NotImplementedError
def update_tracker(self, _):
raise NotImplementedError
class ObjectTracker(BaseObjectTracker):
def __init__(self):
self.tracker = cv2.TrackerCSRT_create()
def init(self, frame, bounding_box):
self.tracker.init(frame, bounding_box.tracker_format)
def update_tracker(self, frame):
success, bounding_box = self.tracker.update(frame)
return success, BoundingBox.from_tracker(*bounding_box.astype(int))
class MultipleObjectTracker(BaseObjectTracker):
def __init__(self, logger):
self.logger = logger
self.tracker = None
self.lost_tracking = True
self.bounding_boxes = list()
self.objects_data = dict()
def init(self, frame, bounding_boxes: List[BoundingBox], tracker_type='medianflow'):
self.logger.info('Object Tracker initialized')
self.tracker = cv2.MultiTracker_create()
self.bounding_boxes = bounding_boxes
for bounding_box in bounding_boxes:
self.tracker.add(OPENCV_OBJECT_TRACKERS[tracker_type](), frame, bounding_box.tracker_format)
def update_tracker(self, frame):
success, tracker_bounding_boxes = self.tracker.update(frame)
bounding_boxes = [BoundingBox.from_tracker(*bounding_box.astype(int))
for bounding_box in tracker_bounding_boxes]
self.bounding_boxes = bounding_boxes
return success, bounding_boxes
def track_faces(self, frame, frame_number, camera):
scan_for_new_faces = frame_number % CHECK_FOR_NEW_FACE_FREQUENCY == 0
# Initial face detection
if self.lost_tracking:
self.logger.info('Object Tracker - Detecting Faces')
# Getting all faces and trying to recognise them
self.bounding_boxes = camera.face_detector.detect(frame, method='deepface')
# If faces where detected they are recognized
if len(self.bounding_boxes) > 0:
self.objects_data = {face_bounding_box.bounding_box_center: {
'name': camera.face_recognition.find_face_in_encodings(image=frame,
face_bounding_box=face_bounding_box),
'bounding_box': face_bounding_box}
for face_bounding_box in self.bounding_boxes}
self.init(frame, bounding_boxes=self.bounding_boxes)
self.lost_tracking = False
# Periodic scanning for new faces
elif scan_for_new_faces:
new_faces_bounding_box = camera.face_detector.detect(frame, method='deepface')
# If there are new faces setting the lost_tracking indicator to True
if len(new_faces_bounding_box) > len(self.bounding_boxes):
self.lost_tracking = True
# Updating tracker with new frame
else:
(success, faces_bounding_boxes) = self.update_tracker(frame)
old_coordinates = list(self.objects_data.keys())
updated_objects_data = {face_bounding_box.bounding_box_center: {
'name': self.objects_data[get_closest_coordinate(face_bounding_box.bounding_box_center,
old_coordinates)]['name'],
'bounding_box': face_bounding_box}
for face_bounding_box in faces_bounding_boxes}
self.objects_data = updated_objects_data
self.lost_tracking = not success
return self.objects_data
| [
"fellbeast.utils.get_closest_coordinate",
"cv2.MultiTracker_create",
"cv2.TrackerCSRT_create"
] | [((786, 810), 'cv2.TrackerCSRT_create', 'cv2.TrackerCSRT_create', ([], {}), '()\n', (808, 810), False, 'import cv2\n'), ((1501, 1526), 'cv2.MultiTracker_create', 'cv2.MultiTracker_create', ([], {}), '()\n', (1524, 1526), False, 'import cv2\n'), ((3766, 3844), 'fellbeast.utils.get_closest_coordinate', 'get_closest_coordinate', (['face_bounding_box.bounding_box_center', 'old_coordinates'], {}), '(face_bounding_box.bounding_box_center, old_coordinates)\n', (3788, 3844), False, 'from fellbeast.utils import get_closest_coordinate\n')] |
"""Views for Zinnia channels"""
from django.views.generic.list import ListView
from zinnia.models.entry import Entry
from zinnia.settings import PAGINATION
class EntryChannel(ListView):
"""View for displaying a custom selection of entries
based on a search pattern, useful for SEO/SMO pages"""
query = ''
paginate_by = PAGINATION
def get_queryset(self):
"""Override the get_queryset method to do the search"""
return Entry.published.search(self.query)
| [
"zinnia.models.entry.Entry.published.search"
] | [((457, 491), 'zinnia.models.entry.Entry.published.search', 'Entry.published.search', (['self.query'], {}), '(self.query)\n', (479, 491), False, 'from zinnia.models.entry import Entry\n')] |
from bge import logic, events, render
from mathutils import Vector
class mouseScroll:
def __init__ (self, cont):
#get Dependencies
self.cont = cont
self.camera = cont.owner
self.mouse = logic.mouse
x = render.getWindowWidth()//2
y = render.getWindowHeight()//2
self.screen_center = (x, y)
render.setMousePosition(x + 1, y + 1)
#show Mouse
render.showMouse(1)
def main (self):
#check for individual Mouse Position on Screen
if self.mouse.position[1] <= 0.0:
#print("Scroll Screen Foward")
self._scrollY(0.5)
if self.mouse.position[1] >= 1:
#print("Scroll Screen Down")
self._scrollY(-0.5)
if self.mouse.position[0] <= 0.0:
#print("Scroll Screen Left")
self._scrollX(-0.5)
if self.mouse.position[0] >= 1:
#print("Scroll Screen Right")
self._scrollX(0.5)
def _scrollY (self, dir):
#move Camera in Y axis
self.camera.position.y += dir
def _scrollX (self, dir):
#move Camera in X axis
self.camera.position.x += dir
class keyboardScroll:
def __init__(self, cont):
#get Dependencies
self.cont = cont
self.camera = cont.owner
self.keyboard = logic.keyboard
def main (self):
#check for individual Keyboard Events
if self.keyboard.events[events.UPARROWKEY] == logic.KX_INPUT_ACTIVE:
#print("Activate Forward!")
self._scrollY(0.1)
if self.keyboard.events[events.DOWNARROWKEY] == logic.KX_INPUT_ACTIVE:
#print("Activate Backward!")
self._scrollY(-0.1)
if self.keyboard.events[events.LEFTARROWKEY] == logic.KX_INPUT_ACTIVE:
#print("Activate Left!")
self._scrollX(-0.1)
if self.keyboard.events[events.RIGHTARROWKEY] == logic.KX_INPUT_ACTIVE:
#print("Activate Right!")
self._scrollX(0.1)
def _scrollY (self, dir):
#move Camera in Y axis
self.camera.position.y += dir
def _scrollX (self, dir):
#move Camera in X axis
self.camera.position.x += dir
| [
"bge.render.showMouse",
"bge.render.getWindowHeight",
"bge.render.setMousePosition",
"bge.render.getWindowWidth"
] | [((382, 419), 'bge.render.setMousePosition', 'render.setMousePosition', (['(x + 1)', '(y + 1)'], {}), '(x + 1, y + 1)\n', (405, 419), False, 'from bge import logic, events, render\n'), ((448, 467), 'bge.render.showMouse', 'render.showMouse', (['(1)'], {}), '(1)\n', (464, 467), False, 'from bge import logic, events, render\n'), ((271, 294), 'bge.render.getWindowWidth', 'render.getWindowWidth', ([], {}), '()\n', (292, 294), False, 'from bge import logic, events, render\n'), ((310, 334), 'bge.render.getWindowHeight', 'render.getWindowHeight', ([], {}), '()\n', (332, 334), False, 'from bge import logic, events, render\n')] |
# Copyright 2020 The Magenta Authors.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""Defines the SVGDecoder model."""
from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
import copy
from magenta.models.image_stylization import ops
# from magenta.models.svg_vae import image_vae_joint
from magenta.models.svg_vae import svg_decoder_loss_joint
from tensor2tensor.layers import common_hparams
from tensor2tensor.layers import common_layers
from tensor2tensor.utils import registry
from tensor2tensor.utils import t2t_model
from tensor2tensor.utils import trainer_lib
import tensorflow.compat.v1 as tf
from tensorflow.contrib import rnn as contrib_rnn
import numpy as np
import tensorflow_probability as tfp
tfd = tfp.distributions
@registry.register_model
class SVGDecoder(t2t_model.T2TModel):
"""Defines the SVGDecoder model."""
def body(self, features):
if self._hparams.initializer == 'orthogonal':
raise ValueError('LSTM models fail with orthogonal initializer.')
train = self._hparams.mode == tf.estimator.ModeKeys.TRAIN
return self.render2cmd_v3_internal(features, self._hparams, train)
def loss(self, logits, features):
# logits should be dict with 'outputs', which is image.
targets = tf.reshape(features['targets'], [-1, 64, 64, 1])
weights = common_layers.weights_all(targets)
loss_num = tf.pow(logits - targets, 2)
return tf.reduce_sum(loss_num * weights), tf.reduce_sum(weights)
def vae_internal(self, features, hparams, train):
# inputs and targets should both be images with dims [batch, 64, 64, 1]
inputs, targets = features['inputs'], features['targets']
inputs = tf.reshape(inputs, [-1, 64, 64, 1])
targets = tf.reshape(targets, [-1, 64, 64, 1])
clss = features['cls']
with tf.variable_scope('vae_internal', reuse=tf.AUTO_REUSE):
# encoder
enc_out = self.visual_encoder(inputs, clss, hparams, train)
enc_out = tf.reshape(enc_out, [-1, 2 * hparams.bottleneck_bits])
# bottleneck
sampled_bottleneck, b_loss = self.bottleneck(enc_out)
losses = {'bottleneck_kl': tf.reduce_mean(b_loss)}
if 'bottleneck' in features:
if common_layers.shape_list(features['bottleneck'])[0] == 0:
# return bottleneck for interpolation
# set losses['training'] = 0 so top() isn't called on it
# potential todo: use losses dict so we have kl_loss here for non stop
# gradient models
return sampled_bottleneck, {'training': 0.0}
else:
# we want to use the given bottleneck
sampled_bottleneck = features['bottleneck']
# finalize bottleneck
unbottleneck = sampled_bottleneck
# decoder.
dec_out = self.visual_decoder(unbottleneck, clss, hparams)
# calculate training loss here lol
rec_loss = -dec_out.log_prob(inputs)
elbo = tf.reduce_mean(-(b_loss + rec_loss))
losses['rec_loss'] = tf.reduce_mean(rec_loss)
losses['training'] = -elbo
if (not hasattr(self, 'summarized_imgs')
and self._hparams.mode != tf.estimator.ModeKeys.PREDICT):
self.summarized_imgs = True
with tf.name_scope(None), tf.name_scope('train' if train else 'test'):
tf.summary.image('rendered_out', dec_out.mean())
tf.summary.image('rendered_og', inputs)
# print(common_layers.shape_list(dec_out.mean()))
# print("sample bottle", common_layers.shape_list(sampled_bottleneck))
# print("b_loss", common_layers.shape_list(b_loss))
return sampled_bottleneck, dec_out.mean(), losses
def bottleneck(self, x):
z_size = self.hparams.bottleneck_bits
x_shape = common_layers.shape_list(x)
with tf.variable_scope('bottleneck', reuse=tf.AUTO_REUSE):
mu = x[..., :self.hparams.bottleneck_bits]
if self.hparams.mode != tf.estimator.ModeKeys.TRAIN:
return mu, 0.0 # No sampling or kl loss on eval.
log_sigma = x[..., self.hparams.bottleneck_bits:]
epsilon = tf.random_normal(x_shape[:-1] + [z_size])
z = mu + tf.exp(log_sigma / 2) * epsilon
kl = 0.5 * tf.reduce_mean(tf.exp(log_sigma) + tf.square(mu) - 1. - log_sigma, axis=-1)
# This is the 'free bits' trick mentioned in Kingma et al. (2016)
free_bits = self.hparams.free_bits
kl_loss = tf.reduce_mean(tf.maximum(kl - free_bits, 0.0))
return z, kl_loss * self.hparams.kl_beta
def visual_encoder(self, inputs, clss, hparams, train):
# goes from [batch, 64, 64, 1] to [batch, hidden_size]
with tf.variable_scope('visual_encoder', reuse=tf.AUTO_REUSE):
ret = inputs
clss = tf.reshape(clss, [-1])
# conv layer, followed by instance norm + FiLM
ret = tf.layers.Conv2D(hparams.base_depth, 5, 1,
padding='SAME', activation=None)(ret)
ret = ops.conditional_instance_norm(ret, clss, hparams.num_categories)
ret = tf.nn.relu(ret)
ret = tf.layers.Conv2D(hparams.base_depth, 5, 2,
padding='SAME', activation=None)(ret)
ret = ops.conditional_instance_norm(ret, clss, hparams.num_categories)
ret = tf.nn.relu(ret)
ret = tf.layers.Conv2D(2 * hparams.base_depth, 5, 1,
padding='SAME', activation=None)(ret)
ret = ops.conditional_instance_norm(ret, clss, hparams.num_categories)
ret = tf.nn.relu(ret)
ret = tf.layers.Conv2D(2 * hparams.base_depth, 5, 2,
padding='SAME', activation=None)(ret)
ret = ops.conditional_instance_norm(ret, clss, hparams.num_categories)
ret = tf.nn.relu(ret)
# new conv layer, to bring shape down
ret = tf.layers.Conv2D(2 * hparams.bottleneck_bits, 4, 2,
padding='SAME', activation=None)(ret)
ret = ops.conditional_instance_norm(ret, clss, hparams.num_categories)
ret = tf.nn.relu(ret)
# new conv layer, to bring shape down
ret = tf.layers.Conv2D(2 * hparams.bottleneck_bits, 4, 2,
padding='SAME', activation=None)(ret)
ret = ops.conditional_instance_norm(ret, clss, hparams.num_categories)
ret = tf.nn.relu(ret)
# ret has 1024
ret = tf.layers.flatten(ret)
ret = tf.layers.dense(ret, 2 * hparams.bottleneck_bits, activation=None)
return ret
def visual_decoder(self, bottleneck, clss, hparams):
# goes from [batch, bottleneck_bits] to [batch, 64, 64, 1]
with tf.variable_scope('visual_decoder', reuse=tf.AUTO_REUSE):
# unbottleneck
ret = tf.layers.dense(bottleneck, 1024, activation=None)
ret = tf.reshape(ret, [-1, 4, 4, 64])
clss = tf.reshape(clss, [-1])
# new deconv to bring shape up
ret = tf.layers.Conv2DTranspose(2 * hparams.base_depth, 4, 2,
padding='SAME', activation=None)(ret)
ret = ops.conditional_instance_norm(ret, clss, hparams.num_categories)
ret = tf.nn.relu(ret)
# new deconv to bring shape up
ret = tf.layers.Conv2DTranspose(2 * hparams.base_depth, 4, 2,
padding='SAME', activation=None)(ret)
ret = ops.conditional_instance_norm(ret, clss, hparams.num_categories)
ret = tf.nn.relu(ret)
ret = tf.layers.Conv2DTranspose(2 * hparams.base_depth, 5, padding='SAME',
activation=None)(ret)
ret = ops.conditional_instance_norm(ret, clss, hparams.num_categories)
ret = tf.nn.relu(ret)
ret = tf.layers.Conv2DTranspose(2 * hparams.base_depth, 5, 2,
padding='SAME', activation=None)(ret)
ret = ops.conditional_instance_norm(ret, clss, hparams.num_categories)
ret = tf.nn.relu(ret)
ret = tf.layers.Conv2DTranspose(hparams.base_depth, 5, padding='SAME',
activation=None)(ret)
ret = ops.conditional_instance_norm(ret, clss, hparams.num_categories)
ret = tf.nn.relu(ret)
ret = tf.layers.Conv2DTranspose(hparams.base_depth, 5, 2, padding='SAME',
activation=None)(ret)
ret = ops.conditional_instance_norm(ret, clss, hparams.num_categories)
ret = tf.nn.relu(ret)
ret = tf.layers.Conv2DTranspose(hparams.base_depth, 5, padding='SAME',
activation=None)(ret)
ret = ops.conditional_instance_norm(ret, clss, hparams.num_categories)
ret = tf.nn.relu(ret)
ret = tf.layers.Conv2D(1, 5, padding='SAME', activation=None)(ret)
ret = tfd.Independent(tfd.Bernoulli(logits=ret),
reinterpreted_batch_ndims=3,
name='image')
return ret
def pretrained_visual_encoder(self, features, hparams, train):
# we want the exact hparams used for training this vv
# vae_hparams = trainer_lib.create_hparams(
# hparams.vae_hparam_set, hparams.vae_hparams,
# data_dir=hparams.vae_data_dir, problem_name=hparams.vae_problem)
# go back to root variable scope
with tf.variable_scope(tf.VariableScope(tf.AUTO_REUSE, ''),
reuse=tf.AUTO_REUSE, auxiliary_name_scope=False):
# vae = image_vae_joint.ImageVAE(vae_hparams, mode=self._hparams.mode,
# problem_hparams=vae_hparams.problem_hparams)
# source image feat
vae_features_source = copy.copy(features)
vae_features_source['inputs'] = tf.reshape(vae_features_source['source_psr'][:, -1, :],
[-1, 64, 64, 1])
vae_features_source['targets'] = vae_features_source['inputs']
vae_features_source['cls'] = vae_features_source['targets_cls']
# we want vae to return bottleneck
# vae_features_source['bottleneck'] = tf.zeros((0, 128))
# we want vae return all
sampled_bottleneck_source, dec_out_source, losses_source = self.vae_internal(vae_features_source, hparams, train)
# vae.initialize_from_ckpt(hparams.vae_ckpt_dir)
if tf.executing_eagerly():
sampled_bottleneck_source, dec_out_source, losses_source = self.vae_internal(vae_features_source, hparams, train)
# the real input to vae will be features['rendered_targets']
vae_features_target = copy.copy(features)
#print('checking shape')
# print(vae_features_target['targets_psr'])
# input()
vae_features_target['inputs'] = tf.reshape(vae_features_target['targets_psr'][:, -1, :],
[-1, 64, 64, 1])
vae_features_target['targets'] = vae_features_target['inputs']
vae_features_target['cls'] = vae_features_target['targets_cls']
# we want vae to return bottleneck
# vae_features_target['bottleneck'] = tf.zeros((0, 128))
sampled_bottleneck_target, dec_out_target, losses_target = self.vae_internal(vae_features_target, hparams, train)
if tf.executing_eagerly():
sampled_bottleneck_target, dec_out_target, losses_target = self.vae_internal(vae_features_target, hparams, train)
vae_losses = {}
for k in losses_source.keys():
vae_losses[k] = losses_source[k] + losses_target[k]
return sampled_bottleneck_target - sampled_bottleneck_source, dec_out_source, dec_out_target, vae_losses
def render2cmd_v3_internal(self, features, hparams, train):
# inputs and targets are both sequences with
# shape = [batch, seq_len, 1, hparams.problem.feature_dim]
targets = features['targets']
source = features['source']
losses = {}
sampled_bottleneck, dec_out_source, dec_out_target, vae_losses = self.pretrained_visual_encoder(features, hparams, train)
losses.update(vae_losses)
# if hparams.sg_bottleneck:
# sampled_bottleneck = tf.stop_gradient(sampled_bottleneck)
with tf.variable_scope('render2cmd_v3_internal'):
# override bottleneck, or return it, if requested
if 'bottleneck' in features:
if common_layers.shape_list(features['bottleneck'])[0] == 0:
# return sampled_bottleneck,
# set losses['training'] = 0 so self.top() doesn't get called on it
return sampled_bottleneck, {'training': 0.0}
else:
# we want to use the given bottleneck
sampled_bottleneck = features['bottleneck']
# finalize bottleneck
unbottleneck_dim = hparams.hidden_size * 2 # twice because using LSTM
if hparams.twice_decoder:
unbottleneck_dim = unbottleneck_dim * 2
dec_initial_state = []
# LSTM encoder
_, encoder_output_states = self.lstm_encoder(
common_layers.flatten4d3d(source), hparams)
# print(features['targets'].shape)
# print('run stacking...')
# print(sampled_bottleneck.shape)
# print(source.shape)
# input()
for hi in range(hparams.num_hidden_layers):
unbottleneck = self.unbottleneck(sampled_bottleneck, unbottleneck_dim,
name_append='_{}'.format(hi))
c, h = encoder_output_states[hi]
# print(unbottleneck.shape)
# print(c.shape, h.shape)
# first_dim = common_layers.shape_list(unbottleneck)[0]
# print(first_dim)
#c = tf.tile(c,[first_dim,1])
#h = tf.tile(h,[first_dim,1])
# input()
dec_initial_state.append(
tf.nn.rnn_cell.LSTMStateTuple(
c=tf.concat([unbottleneck[:, :unbottleneck_dim // 2], c], 1),
h=tf.concat([unbottleneck[:, unbottleneck_dim // 2:], h], 1)))
dec_initial_state = tuple(dec_initial_state)
#print('checkshape dec_initial_state')
# print(dec_initial_state)
# input()
shifted_targets = common_layers.shift_right(targets)
# Add 1 to account for the padding added to the left from shift_right
targets_length = common_layers.length_from_embedding(
shifted_targets) + 1
# LSTM decoder
hparams_decoder = copy.copy(hparams)
if hparams.twice_decoder:
hparams_decoder.hidden_size = 2 * hparams.hidden_size
if hparams.mode == tf.estimator.ModeKeys.PREDICT:
decoder_outputs, _ = self.lstm_decoder_infer(
common_layers.flatten4d3d(shifted_targets),
targets_length, hparams_decoder, features['targets_cls'],
train, initial_state=dec_initial_state,
bottleneck=sampled_bottleneck)
else:
decoder_outputs, _ = self.lstm_decoder(
common_layers.flatten4d3d(shifted_targets),
targets_length, hparams_decoder, features['targets_cls'],
train, initial_state=dec_initial_state,
bottleneck=sampled_bottleneck)
ret = tf.expand_dims(decoder_outputs, axis=2)
return ret, losses
def lstm_decoder_infer(self, inputs, sequence_length, hparams, clss, train,
initial_state=None, bottleneck=None):
# IN PREDICT MODE, RUN tf.while RNN
max_decode_length = 51
batch_size = common_layers.shape_list(inputs)[0]
zero_pad, logits_so_far = self.create_initial_input_for_decode(batch_size)
layers = contrib_rnn.MultiRNNCell([
self.lstm_cell(hparams, train) for _ in range(hparams.num_hidden_layers)
])
if initial_state is None:
raise Exception('initial state should be init from bottleneck!')
# append one-hot class to bottleneck, which will be given per step
clss = tf.reshape(clss, [-1])
if not hparams.use_cls:
clss = tf.zeros_like(clss)
if hparams.condition_on_sln:
sln = tf.reshape(sequence_length, [-1])
bottleneck = tf.concat((bottleneck,
tf.one_hot(clss, hparams.num_categories),
tf.one_hot(sln, max_decode_length)), -1)
else:
bottleneck = tf.concat((bottleneck,
tf.one_hot(clss, hparams.num_categories)), -1)
def infer_step(logits_so_far, current_hidden):
"""Inference step of LSTM while loop."""
# unflatten hidden:
current_hidden = tuple(tf.nn.rnn_cell.LSTMStateTuple(c=s[0], h=s[1])
for s in current_hidden)
# put logits_so_far through top
tm = self._problem_hparams.modality['targets']
# need to reuse top params
reset_scope = tf.variable_scope(tf.VariableScope(tf.AUTO_REUSE, ''),
reuse=tf.AUTO_REUSE,
auxiliary_name_scope=False)
top_scope = tf.variable_scope('svg_decoder/{}_modality'.format(tm),
reuse=tf.AUTO_REUSE)
with reset_scope, top_scope:
samples_so_far = self.hparams.top['targets'](
logits_so_far, None, self.hparams, self.problem_hparams.vocab_size)
# append a zero pad to the samples. this effectively shifts the samples
# right, but, unlike shift_right, by not removing the last element, we
# allow an empty samples_so_far to not be empty after padding
samples_so_far = tf.concat([zero_pad, samples_so_far], axis=1)
shifted_targets = common_layers.flatten4d3d(samples_so_far)
# now take the very last one here, will be the actual input to the rnn
shifted_targets = shifted_targets[:, -1:, :]
# tile and append the bottleneck to inputs
sln_offset = 0
if hparams.condition_on_sln:
sln_offset = 51
pre_tile_y = tf.reshape(bottleneck,
[common_layers.shape_list(bottleneck)[0], 1,
hparams.bottleneck_bits + hparams.num_categories + sln_offset])
overlay_x = tf.tile(pre_tile_y,
[1, common_layers.shape_list(shifted_targets)[1], 1])
inputs = tf.concat([shifted_targets, overlay_x], -1)
seq_len_batch = tf.ones([common_layers.shape_list(inputs)[0]])
# RUN PRE-LSTM LAYER
with tf.variable_scope('pre_decoder', reuse=tf.AUTO_REUSE):
inputs = tf.layers.dense(inputs, hparams.hidden_size, name='bottom')
inputs = tf.nn.tanh(inputs)
# RUN LSTM
with tf.variable_scope('lstm_decoder', reuse=tf.AUTO_REUSE):
next_step, next_state = tf.nn.dynamic_rnn(
layers, inputs, seq_len_batch, initial_state=current_hidden,
dtype=tf.float32, time_major=False)
next_step = tf.expand_dims(next_step, [1])
logits_so_far = tf.concat([logits_so_far, next_step], 1)
#print('concat success')
# input()
# flatten state
next_state = tuple((s.c, s.h) for s in next_state)
return logits_so_far, next_state
def while_exit_cond(logits_so_far, unused_current_hidden):
length = common_layers.shape_list(logits_so_far)[1]
return length < max_decode_length
# passing state must be flattened:
initial_state = tuple([(s.c, s.h) for s in initial_state])
# actually run tf.while:
logits, final_state = tf.while_loop(
while_exit_cond, infer_step,
[logits_so_far, initial_state],
shape_invariants=[
tf.TensorShape([None, None, 1, hparams.hidden_size]),
tuple([(s[0].get_shape(), s[1].get_shape())
for s in initial_state]),
],
back_prop=False,
parallel_iterations=1
)
# logits should be returned in 3d mode:
logits = common_layers.flatten4d3d(logits)
return logits, final_state
def lstm_decoder(self, inputs, sequence_length, hparams, clss, train,
initial_state=None, bottleneck=None):
# NOT IN PREDICT MODE. JUST RUN TEACHER-FORCED RNN:
layers = contrib_rnn.MultiRNNCell([
self.lstm_cell(hparams, train) for _ in range(hparams.num_hidden_layers)
])
# append one-hot class to bottleneck, which will be given per step
clss = tf.reshape(clss, [-1])
if not hparams.use_cls:
clss = tf.zeros_like(clss)
if hparams.condition_on_sln:
sln = tf.reshape(sequence_length, [-1])
bottleneck = tf.concat((bottleneck,
tf.one_hot(clss, hparams.num_categories),
tf.one_hot(sln, 51)), -1)
else:
bottleneck = tf.concat((bottleneck,
tf.one_hot(clss, hparams.num_categories)), -1)
# tile and append the bottleneck to inputs
sln_offset = 0
if hparams.condition_on_sln:
sln_offset = 51
pre_tile_y = tf.reshape(
bottleneck,
[common_layers.shape_list(bottleneck)[0], 1,
hparams.bottleneck_bits + hparams.num_categories + sln_offset])
overlay_x = tf.tile(
pre_tile_y, [1, common_layers.shape_list(inputs)[1], 1])
inputs = tf.concat([inputs, overlay_x], -1)
with tf.variable_scope('pre_decoder', reuse=tf.AUTO_REUSE):
inputs = tf.layers.dense(
inputs, hparams.hidden_size, name='bottom')
inputs = tf.nn.tanh(inputs)
# print(inputs)
# print(initial_state)
# input()
with tf.variable_scope('lstm_decoder', reuse=tf.AUTO_REUSE):
return tf.nn.dynamic_rnn(
layers, inputs, sequence_length, initial_state=initial_state,
dtype=tf.float32, time_major=False)
def lstm_encoder(self, inputs, hparams):
batch_size = common_layers.shape_list(inputs)[0]
length = 51
# a 4-layer LSTM
cell = tf.nn.rnn_cell.LSTMCell(256, state_is_tuple=True)
if hparams.mode != tf.estimator.ModeKeys.PREDICT:
cell = tf.nn.rnn_cell.DropoutWrapper(
cell=cell, output_keep_prob=0.5)
cell1 = tf.nn.rnn_cell.LSTMCell(256, state_is_tuple=True)
if hparams.mode != tf.estimator.ModeKeys.PREDICT:
cell1 = tf.nn.rnn_cell.DropoutWrapper(
cell=cell1, output_keep_prob=0.5)
cell2 = tf.nn.rnn_cell.LSTMCell(256, state_is_tuple=True)
if hparams.mode != tf.estimator.ModeKeys.PREDICT:
cell2 = tf.nn.rnn_cell.DropoutWrapper(
cell=cell2, output_keep_prob=0.5)
cell3 = tf.nn.rnn_cell.LSTMCell(256, state_is_tuple=True)
if hparams.mode != tf.estimator.ModeKeys.PREDICT:
cell3 = tf.nn.rnn_cell.DropoutWrapper(
cell=cell3, output_keep_prob=0.5)
stack = tf.nn.rnn_cell.MultiRNNCell(
[cell, cell1, cell2, cell3], state_is_tuple=True)
initial_state = stack.zero_state(batch_size, dtype=tf.float32)
# a two layer LSTM
le_output, le_output_states = tf.nn.dynamic_rnn(
cell=stack,
inputs=inputs,
sequence_length=tf.fill([batch_size], length),
initial_state=initial_state,
dtype=tf.float32,
time_major=False
)
#print('check lstm encoder')
# print(le_output_states)
# input()
return le_output, le_output_states
def lstm_cell(self, hparams, train):
keep_prob = 1.0 - hparams.rec_dropout * tf.to_float(train)
recurrent_dropout_cell = contrib_rnn.LayerNormBasicLSTMCell(
hparams.hidden_size + 256,
layer_norm=hparams.layer_norm,
dropout_keep_prob=keep_prob)
if hparams.ff_dropout:
return contrib_rnn.DropoutWrapper(
recurrent_dropout_cell, input_keep_prob=keep_prob)
return recurrent_dropout_cell
def unbottleneck(self, x, res_size, reuse=tf.AUTO_REUSE, name_append=''):
with tf.variable_scope('unbottleneck{}'.format(name_append), reuse=reuse):
x = tf.layers.dense(x, res_size, name='dense', activation='tanh')
return x
def create_initial_input_for_decode(self, batch_size):
# Create an initial output tensor. This will be passed
# to the infer_step, which adds one timestep at every iteration.
dim = self._problem_hparams.vocab_size['targets']
hdim = self._hparams.hidden_size + 256
initial_output = tf.zeros((batch_size, 0, 1, hdim), dtype=tf.float32)
zero_pad = tf.zeros((batch_size, 1, 1, dim), dtype=tf.float32)
# Hack: foldl complains when the output shape is less specified than the
# input shape, so we confuse it about the input shape.
initial_output = tf.slice(initial_output, [0, 0, 0, 0],
common_layers.shape_list(initial_output))
zero_pad = tf.slice(zero_pad, [0, 0, 0, 0],
common_layers.shape_list(zero_pad))
return zero_pad, initial_output
def _greedy_infer(self, features, extra_decode_length, use_tpu=False):
# extra_decode_length is set to 0, unused.
infer_features = copy.copy(features)
if 'targets' not in infer_features:
infer_features['targets'] = infer_features['infer_targets']
logits, losses = self(infer_features) # pylint: disable=not-callable
return {
'outputs': logits,
'scores': None,
'logits': logits,
'losses': losses,
}
@registry.register_hparams
def svg_decoder():
"""Basic hparams for SVG decoder."""
hparams = common_hparams.basic_params1()
hparams.daisy_chain_variables = False
hparams.batch_size = 128
hparams.hidden_size = 1024
hparams.num_hidden_layers = 2
hparams.initializer = 'uniform_unit_scaling'
hparams.initializer_gain = 1.0
hparams.weight_decay = 0.0
hparams.num_hidden_layers = 4
hparams.force_full_predict = True
hparams.dropout = 0.5
hparams.learning_rate_warmup_steps = 100000
# LSTM-specific hparams
hparams.add_hparam('vocab_size', None)
# VAE params
hparams.add_hparam('base_depth', 32)
hparams.add_hparam('bottleneck_bits', 32)
# loss hparams
hparams.add_hparam('kl_beta', 300)
hparams.add_hparam('free_bits_div', 4)
hparams.add_hparam('free_bits', 0.15)
# loss params
hparams.add_hparam('soft_k', 10)
hparams.add_hparam('mdn_k', 1)
# params required by LayerNormLSTMCell, for us to just use recurrent dropout
hparams.add_hparam('layer_norm', False)
hparams.add_hparam('ff_dropout', True)
hparams.add_hparam('rec_dropout', 0.3)
# Decode architecture hparams
hparams.add_hparam('twice_decoder', False)
hparams.add_hparam('sg_bottleneck', False)
hparams.add_hparam('condition_on_sln', False)
hparams.add_hparam('use_cls', True)
# MDN loss hparams
hparams.add_hparam('num_mixture', 50)
hparams.add_hparam('mix_temperature', 0.0001)
hparams.add_hparam('gauss_temperature', 0.0001)
hparams.add_hparam('dont_reduce_loss', False)
# VAE meta hparams (to load image encoder)
# hparams.add_hparam('vae_ckpt_dir', '')
# hparams.add_hparam('vae_hparams', 'base_depth=32, hidden_size=32')
# hparams.add_hparam('vae_data_dir', '')
# hparams.add_hparam('vae_hparam_set', 'image_vae')
# hparams.add_hparam('vae_problem', 'glyph_azzn_problem')
# data format hparams
hparams.add_hparam('num_categories', 62)
# problem hparams (required, don't modify)
hparams.add_hparam('absolute', False)
hparams.add_hparam('just_render', False)
hparams.add_hparam('plus_render', False)
# modality hparams
hparams.bottom = {
'inputs': svg_decoder_loss_joint.real_svg_bottom,
'targets': svg_decoder_loss_joint.real_svg_bottom,
}
hparams.top = {'targets': svg_decoder_loss_joint.real_svg_top}
hparams.loss = {'targets': svg_decoder_loss_joint.real_svg_loss}
return hparams
| [
"tensorflow.compat.v1.one_hot",
"tensor2tensor.layers.common_layers.length_from_embedding",
"tensorflow.compat.v1.exp",
"tensor2tensor.layers.common_layers.shape_list",
"tensorflow.compat.v1.maximum",
"tensorflow.compat.v1.concat",
"tensorflow.compat.v1.zeros",
"tensorflow.compat.v1.nn.rnn_cell.MultiR... | [((27888, 27918), 'tensor2tensor.layers.common_hparams.basic_params1', 'common_hparams.basic_params1', ([], {}), '()\n', (27916, 27918), False, 'from tensor2tensor.layers import common_hparams\n'), ((1819, 1867), 'tensorflow.compat.v1.reshape', 'tf.reshape', (["features['targets']", '[-1, 64, 64, 1]'], {}), "(features['targets'], [-1, 64, 64, 1])\n", (1829, 1867), True, 'import tensorflow.compat.v1 as tf\n'), ((1886, 1920), 'tensor2tensor.layers.common_layers.weights_all', 'common_layers.weights_all', (['targets'], {}), '(targets)\n', (1911, 1920), False, 'from tensor2tensor.layers import common_layers\n'), ((1940, 1967), 'tensorflow.compat.v1.pow', 'tf.pow', (['(logits - targets)', '(2)'], {}), '(logits - targets, 2)\n', (1946, 1967), True, 'import tensorflow.compat.v1 as tf\n'), ((2259, 2294), 'tensorflow.compat.v1.reshape', 'tf.reshape', (['inputs', '[-1, 64, 64, 1]'], {}), '(inputs, [-1, 64, 64, 1])\n', (2269, 2294), True, 'import tensorflow.compat.v1 as tf\n'), ((2313, 2349), 'tensorflow.compat.v1.reshape', 'tf.reshape', (['targets', '[-1, 64, 64, 1]'], {}), '(targets, [-1, 64, 64, 1])\n', (2323, 2349), True, 'import tensorflow.compat.v1 as tf\n'), ((4537, 4564), 'tensor2tensor.layers.common_layers.shape_list', 'common_layers.shape_list', (['x'], {}), '(x)\n', (4561, 4564), False, 'from tensor2tensor.layers import common_layers\n'), ((17605, 17627), 'tensorflow.compat.v1.reshape', 'tf.reshape', (['clss', '[-1]'], {}), '(clss, [-1])\n', (17615, 17627), True, 'import tensorflow.compat.v1 as tf\n'), ((21964, 21997), 'tensor2tensor.layers.common_layers.flatten4d3d', 'common_layers.flatten4d3d', (['logits'], {}), '(logits)\n', (21989, 21997), False, 'from tensor2tensor.layers import common_layers\n'), ((22459, 22481), 'tensorflow.compat.v1.reshape', 'tf.reshape', (['clss', '[-1]'], {}), '(clss, [-1])\n', (22469, 22481), True, 'import tensorflow.compat.v1 as tf\n'), ((23421, 23455), 'tensorflow.compat.v1.concat', 'tf.concat', (['[inputs, overlay_x]', '(-1)'], {}), '([inputs, overlay_x], -1)\n', (23430, 23455), True, 'import tensorflow.compat.v1 as tf\n'), ((24136, 24185), 'tensorflow.compat.v1.nn.rnn_cell.LSTMCell', 'tf.nn.rnn_cell.LSTMCell', (['(256)'], {'state_is_tuple': '(True)'}), '(256, state_is_tuple=True)\n', (24159, 24185), True, 'import tensorflow.compat.v1 as tf\n'), ((24360, 24409), 'tensorflow.compat.v1.nn.rnn_cell.LSTMCell', 'tf.nn.rnn_cell.LSTMCell', (['(256)'], {'state_is_tuple': '(True)'}), '(256, state_is_tuple=True)\n', (24383, 24409), True, 'import tensorflow.compat.v1 as tf\n'), ((24586, 24635), 'tensorflow.compat.v1.nn.rnn_cell.LSTMCell', 'tf.nn.rnn_cell.LSTMCell', (['(256)'], {'state_is_tuple': '(True)'}), '(256, state_is_tuple=True)\n', (24609, 24635), True, 'import tensorflow.compat.v1 as tf\n'), ((24812, 24861), 'tensorflow.compat.v1.nn.rnn_cell.LSTMCell', 'tf.nn.rnn_cell.LSTMCell', (['(256)'], {'state_is_tuple': '(True)'}), '(256, state_is_tuple=True)\n', (24835, 24861), True, 'import tensorflow.compat.v1 as tf\n'), ((25038, 25115), 'tensorflow.compat.v1.nn.rnn_cell.MultiRNNCell', 'tf.nn.rnn_cell.MultiRNNCell', (['[cell, cell1, cell2, cell3]'], {'state_is_tuple': '(True)'}), '([cell, cell1, cell2, cell3], state_is_tuple=True)\n', (25065, 25115), True, 'import tensorflow.compat.v1 as tf\n'), ((25779, 25905), 'tensorflow.contrib.rnn.LayerNormBasicLSTMCell', 'contrib_rnn.LayerNormBasicLSTMCell', (['(hparams.hidden_size + 256)'], {'layer_norm': 'hparams.layer_norm', 'dropout_keep_prob': 'keep_prob'}), '(hparams.hidden_size + 256, layer_norm=\n hparams.layer_norm, dropout_keep_prob=keep_prob)\n', (25813, 25905), True, 'from tensorflow.contrib import rnn as contrib_rnn\n'), ((26709, 26761), 'tensorflow.compat.v1.zeros', 'tf.zeros', (['(batch_size, 0, 1, hdim)'], {'dtype': 'tf.float32'}), '((batch_size, 0, 1, hdim), dtype=tf.float32)\n', (26717, 26761), True, 'import tensorflow.compat.v1 as tf\n'), ((26781, 26832), 'tensorflow.compat.v1.zeros', 'tf.zeros', (['(batch_size, 1, 1, dim)'], {'dtype': 'tf.float32'}), '((batch_size, 1, 1, dim), dtype=tf.float32)\n', (26789, 26832), True, 'import tensorflow.compat.v1 as tf\n'), ((27425, 27444), 'copy.copy', 'copy.copy', (['features'], {}), '(features)\n', (27434, 27444), False, 'import copy\n'), ((1983, 2016), 'tensorflow.compat.v1.reduce_sum', 'tf.reduce_sum', (['(loss_num * weights)'], {}), '(loss_num * weights)\n', (1996, 2016), True, 'import tensorflow.compat.v1 as tf\n'), ((2018, 2040), 'tensorflow.compat.v1.reduce_sum', 'tf.reduce_sum', (['weights'], {}), '(weights)\n', (2031, 2040), True, 'import tensorflow.compat.v1 as tf\n'), ((2396, 2450), 'tensorflow.compat.v1.variable_scope', 'tf.variable_scope', (['"""vae_internal"""'], {'reuse': 'tf.AUTO_REUSE'}), "('vae_internal', reuse=tf.AUTO_REUSE)\n", (2413, 2450), True, 'import tensorflow.compat.v1 as tf\n'), ((2568, 2622), 'tensorflow.compat.v1.reshape', 'tf.reshape', (['enc_out', '[-1, 2 * hparams.bottleneck_bits]'], {}), '(enc_out, [-1, 2 * hparams.bottleneck_bits])\n', (2578, 2622), True, 'import tensorflow.compat.v1 as tf\n'), ((3662, 3698), 'tensorflow.compat.v1.reduce_mean', 'tf.reduce_mean', (['(-(b_loss + rec_loss))'], {}), '(-(b_loss + rec_loss))\n', (3676, 3698), True, 'import tensorflow.compat.v1 as tf\n'), ((3732, 3756), 'tensorflow.compat.v1.reduce_mean', 'tf.reduce_mean', (['rec_loss'], {}), '(rec_loss)\n', (3746, 3756), True, 'import tensorflow.compat.v1 as tf\n'), ((4578, 4630), 'tensorflow.compat.v1.variable_scope', 'tf.variable_scope', (['"""bottleneck"""'], {'reuse': 'tf.AUTO_REUSE'}), "('bottleneck', reuse=tf.AUTO_REUSE)\n", (4595, 4630), True, 'import tensorflow.compat.v1 as tf\n'), ((4902, 4943), 'tensorflow.compat.v1.random_normal', 'tf.random_normal', (['(x_shape[:-1] + [z_size])'], {}), '(x_shape[:-1] + [z_size])\n', (4918, 4943), True, 'import tensorflow.compat.v1 as tf\n'), ((5477, 5533), 'tensorflow.compat.v1.variable_scope', 'tf.variable_scope', (['"""visual_encoder"""'], {'reuse': 'tf.AUTO_REUSE'}), "('visual_encoder', reuse=tf.AUTO_REUSE)\n", (5494, 5533), True, 'import tensorflow.compat.v1 as tf\n'), ((5579, 5601), 'tensorflow.compat.v1.reshape', 'tf.reshape', (['clss', '[-1]'], {}), '(clss, [-1])\n', (5589, 5601), True, 'import tensorflow.compat.v1 as tf\n'), ((5814, 5878), 'magenta.models.image_stylization.ops.conditional_instance_norm', 'ops.conditional_instance_norm', (['ret', 'clss', 'hparams.num_categories'], {}), '(ret, clss, hparams.num_categories)\n', (5843, 5878), False, 'from magenta.models.image_stylization import ops\n'), ((5897, 5912), 'tensorflow.compat.v1.nn.relu', 'tf.nn.relu', (['ret'], {}), '(ret)\n', (5907, 5912), True, 'import tensorflow.compat.v1 as tf\n'), ((6066, 6130), 'magenta.models.image_stylization.ops.conditional_instance_norm', 'ops.conditional_instance_norm', (['ret', 'clss', 'hparams.num_categories'], {}), '(ret, clss, hparams.num_categories)\n', (6095, 6130), False, 'from magenta.models.image_stylization import ops\n'), ((6149, 6164), 'tensorflow.compat.v1.nn.relu', 'tf.nn.relu', (['ret'], {}), '(ret)\n', (6159, 6164), True, 'import tensorflow.compat.v1 as tf\n'), ((6322, 6386), 'magenta.models.image_stylization.ops.conditional_instance_norm', 'ops.conditional_instance_norm', (['ret', 'clss', 'hparams.num_categories'], {}), '(ret, clss, hparams.num_categories)\n', (6351, 6386), False, 'from magenta.models.image_stylization import ops\n'), ((6405, 6420), 'tensorflow.compat.v1.nn.relu', 'tf.nn.relu', (['ret'], {}), '(ret)\n', (6415, 6420), True, 'import tensorflow.compat.v1 as tf\n'), ((6578, 6642), 'magenta.models.image_stylization.ops.conditional_instance_norm', 'ops.conditional_instance_norm', (['ret', 'clss', 'hparams.num_categories'], {}), '(ret, clss, hparams.num_categories)\n', (6607, 6642), False, 'from magenta.models.image_stylization import ops\n'), ((6661, 6676), 'tensorflow.compat.v1.nn.relu', 'tf.nn.relu', (['ret'], {}), '(ret)\n', (6671, 6676), True, 'import tensorflow.compat.v1 as tf\n'), ((6889, 6953), 'magenta.models.image_stylization.ops.conditional_instance_norm', 'ops.conditional_instance_norm', (['ret', 'clss', 'hparams.num_categories'], {}), '(ret, clss, hparams.num_categories)\n', (6918, 6953), False, 'from magenta.models.image_stylization import ops\n'), ((6972, 6987), 'tensorflow.compat.v1.nn.relu', 'tf.nn.relu', (['ret'], {}), '(ret)\n', (6982, 6987), True, 'import tensorflow.compat.v1 as tf\n'), ((7200, 7264), 'magenta.models.image_stylization.ops.conditional_instance_norm', 'ops.conditional_instance_norm', (['ret', 'clss', 'hparams.num_categories'], {}), '(ret, clss, hparams.num_categories)\n', (7229, 7264), False, 'from magenta.models.image_stylization import ops\n'), ((7283, 7298), 'tensorflow.compat.v1.nn.relu', 'tf.nn.relu', (['ret'], {}), '(ret)\n', (7293, 7298), True, 'import tensorflow.compat.v1 as tf\n'), ((7345, 7367), 'tensorflow.compat.v1.layers.flatten', 'tf.layers.flatten', (['ret'], {}), '(ret)\n', (7362, 7367), True, 'import tensorflow.compat.v1 as tf\n'), ((7386, 7452), 'tensorflow.compat.v1.layers.dense', 'tf.layers.dense', (['ret', '(2 * hparams.bottleneck_bits)'], {'activation': 'None'}), '(ret, 2 * hparams.bottleneck_bits, activation=None)\n', (7401, 7452), True, 'import tensorflow.compat.v1 as tf\n'), ((7611, 7667), 'tensorflow.compat.v1.variable_scope', 'tf.variable_scope', (['"""visual_decoder"""'], {'reuse': 'tf.AUTO_REUSE'}), "('visual_decoder', reuse=tf.AUTO_REUSE)\n", (7628, 7667), True, 'import tensorflow.compat.v1 as tf\n'), ((7714, 7764), 'tensorflow.compat.v1.layers.dense', 'tf.layers.dense', (['bottleneck', '(1024)'], {'activation': 'None'}), '(bottleneck, 1024, activation=None)\n', (7729, 7764), True, 'import tensorflow.compat.v1 as tf\n'), ((7783, 7814), 'tensorflow.compat.v1.reshape', 'tf.reshape', (['ret', '[-1, 4, 4, 64]'], {}), '(ret, [-1, 4, 4, 64])\n', (7793, 7814), True, 'import tensorflow.compat.v1 as tf\n'), ((7834, 7856), 'tensorflow.compat.v1.reshape', 'tf.reshape', (['clss', '[-1]'], {}), '(clss, [-1])\n', (7844, 7856), True, 'import tensorflow.compat.v1 as tf\n'), ((8075, 8139), 'magenta.models.image_stylization.ops.conditional_instance_norm', 'ops.conditional_instance_norm', (['ret', 'clss', 'hparams.num_categories'], {}), '(ret, clss, hparams.num_categories)\n', (8104, 8139), False, 'from magenta.models.image_stylization import ops\n'), ((8158, 8173), 'tensorflow.compat.v1.nn.relu', 'tf.nn.relu', (['ret'], {}), '(ret)\n', (8168, 8173), True, 'import tensorflow.compat.v1 as tf\n'), ((8392, 8456), 'magenta.models.image_stylization.ops.conditional_instance_norm', 'ops.conditional_instance_norm', (['ret', 'clss', 'hparams.num_categories'], {}), '(ret, clss, hparams.num_categories)\n', (8421, 8456), False, 'from magenta.models.image_stylization import ops\n'), ((8475, 8490), 'tensorflow.compat.v1.nn.relu', 'tf.nn.relu', (['ret'], {}), '(ret)\n', (8485, 8490), True, 'import tensorflow.compat.v1 as tf\n'), ((8663, 8727), 'magenta.models.image_stylization.ops.conditional_instance_norm', 'ops.conditional_instance_norm', (['ret', 'clss', 'hparams.num_categories'], {}), '(ret, clss, hparams.num_categories)\n', (8692, 8727), False, 'from magenta.models.image_stylization import ops\n'), ((8746, 8761), 'tensorflow.compat.v1.nn.relu', 'tf.nn.relu', (['ret'], {}), '(ret)\n', (8756, 8761), True, 'import tensorflow.compat.v1 as tf\n'), ((8937, 9001), 'magenta.models.image_stylization.ops.conditional_instance_norm', 'ops.conditional_instance_norm', (['ret', 'clss', 'hparams.num_categories'], {}), '(ret, clss, hparams.num_categories)\n', (8966, 9001), False, 'from magenta.models.image_stylization import ops\n'), ((9020, 9035), 'tensorflow.compat.v1.nn.relu', 'tf.nn.relu', (['ret'], {}), '(ret)\n', (9030, 9035), True, 'import tensorflow.compat.v1 as tf\n'), ((9204, 9268), 'magenta.models.image_stylization.ops.conditional_instance_norm', 'ops.conditional_instance_norm', (['ret', 'clss', 'hparams.num_categories'], {}), '(ret, clss, hparams.num_categories)\n', (9233, 9268), False, 'from magenta.models.image_stylization import ops\n'), ((9287, 9302), 'tensorflow.compat.v1.nn.relu', 'tf.nn.relu', (['ret'], {}), '(ret)\n', (9297, 9302), True, 'import tensorflow.compat.v1 as tf\n'), ((9474, 9538), 'magenta.models.image_stylization.ops.conditional_instance_norm', 'ops.conditional_instance_norm', (['ret', 'clss', 'hparams.num_categories'], {}), '(ret, clss, hparams.num_categories)\n', (9503, 9538), False, 'from magenta.models.image_stylization import ops\n'), ((9557, 9572), 'tensorflow.compat.v1.nn.relu', 'tf.nn.relu', (['ret'], {}), '(ret)\n', (9567, 9572), True, 'import tensorflow.compat.v1 as tf\n'), ((9741, 9805), 'magenta.models.image_stylization.ops.conditional_instance_norm', 'ops.conditional_instance_norm', (['ret', 'clss', 'hparams.num_categories'], {}), '(ret, clss, hparams.num_categories)\n', (9770, 9805), False, 'from magenta.models.image_stylization import ops\n'), ((9824, 9839), 'tensorflow.compat.v1.nn.relu', 'tf.nn.relu', (['ret'], {}), '(ret)\n', (9834, 9839), True, 'import tensorflow.compat.v1 as tf\n'), ((10862, 10881), 'copy.copy', 'copy.copy', (['features'], {}), '(features)\n', (10871, 10881), False, 'import copy\n'), ((10926, 10998), 'tensorflow.compat.v1.reshape', 'tf.reshape', (["vae_features_source['source_psr'][:, -1, :]", '[-1, 64, 64, 1]'], {}), "(vae_features_source['source_psr'][:, -1, :], [-1, 64, 64, 1])\n", (10936, 10998), True, 'import tensorflow.compat.v1 as tf\n'), ((11556, 11578), 'tensorflow.compat.v1.executing_eagerly', 'tf.executing_eagerly', ([], {}), '()\n', (11576, 11578), True, 'import tensorflow.compat.v1 as tf\n'), ((11818, 11837), 'copy.copy', 'copy.copy', (['features'], {}), '(features)\n', (11827, 11837), False, 'import copy\n'), ((11997, 12070), 'tensorflow.compat.v1.reshape', 'tf.reshape', (["vae_features_target['targets_psr'][:, -1, :]", '[-1, 64, 64, 1]'], {}), "(vae_features_target['targets_psr'][:, -1, :], [-1, 64, 64, 1])\n", (12007, 12070), True, 'import tensorflow.compat.v1 as tf\n'), ((12529, 12551), 'tensorflow.compat.v1.executing_eagerly', 'tf.executing_eagerly', ([], {}), '()\n', (12549, 12551), True, 'import tensorflow.compat.v1 as tf\n'), ((13503, 13546), 'tensorflow.compat.v1.variable_scope', 'tf.variable_scope', (['"""render2cmd_v3_internal"""'], {}), "('render2cmd_v3_internal')\n", (13520, 13546), True, 'import tensorflow.compat.v1 as tf\n'), ((15704, 15738), 'tensor2tensor.layers.common_layers.shift_right', 'common_layers.shift_right', (['targets'], {}), '(targets)\n', (15729, 15738), False, 'from tensor2tensor.layers import common_layers\n'), ((15982, 16000), 'copy.copy', 'copy.copy', (['hparams'], {}), '(hparams)\n', (15991, 16000), False, 'import copy\n'), ((16833, 16872), 'tensorflow.compat.v1.expand_dims', 'tf.expand_dims', (['decoder_outputs'], {'axis': '(2)'}), '(decoder_outputs, axis=2)\n', (16847, 16872), True, 'import tensorflow.compat.v1 as tf\n'), ((17142, 17174), 'tensor2tensor.layers.common_layers.shape_list', 'common_layers.shape_list', (['inputs'], {}), '(inputs)\n', (17166, 17174), False, 'from tensor2tensor.layers import common_layers\n'), ((17679, 17698), 'tensorflow.compat.v1.zeros_like', 'tf.zeros_like', (['clss'], {}), '(clss)\n', (17692, 17698), True, 'import tensorflow.compat.v1 as tf\n'), ((17754, 17787), 'tensorflow.compat.v1.reshape', 'tf.reshape', (['sequence_length', '[-1]'], {}), '(sequence_length, [-1])\n', (17764, 17787), True, 'import tensorflow.compat.v1 as tf\n'), ((19383, 19428), 'tensorflow.compat.v1.concat', 'tf.concat', (['[zero_pad, samples_so_far]'], {'axis': '(1)'}), '([zero_pad, samples_so_far], axis=1)\n', (19392, 19428), True, 'import tensorflow.compat.v1 as tf\n'), ((19459, 19500), 'tensor2tensor.layers.common_layers.flatten4d3d', 'common_layers.flatten4d3d', (['samples_so_far'], {}), '(samples_so_far)\n', (19484, 19500), False, 'from tensor2tensor.layers import common_layers\n'), ((20177, 20220), 'tensorflow.compat.v1.concat', 'tf.concat', (['[shifted_targets, overlay_x]', '(-1)'], {}), '([shifted_targets, overlay_x], -1)\n', (20186, 20220), True, 'import tensorflow.compat.v1 as tf\n'), ((20850, 20880), 'tensorflow.compat.v1.expand_dims', 'tf.expand_dims', (['next_step', '[1]'], {}), '(next_step, [1])\n', (20864, 20880), True, 'import tensorflow.compat.v1 as tf\n'), ((20910, 20950), 'tensorflow.compat.v1.concat', 'tf.concat', (['[logits_so_far, next_step]', '(1)'], {}), '([logits_so_far, next_step], 1)\n', (20919, 20950), True, 'import tensorflow.compat.v1 as tf\n'), ((22533, 22552), 'tensorflow.compat.v1.zeros_like', 'tf.zeros_like', (['clss'], {}), '(clss)\n', (22546, 22552), True, 'import tensorflow.compat.v1 as tf\n'), ((22608, 22641), 'tensorflow.compat.v1.reshape', 'tf.reshape', (['sequence_length', '[-1]'], {}), '(sequence_length, [-1])\n', (22618, 22641), True, 'import tensorflow.compat.v1 as tf\n'), ((23470, 23523), 'tensorflow.compat.v1.variable_scope', 'tf.variable_scope', (['"""pre_decoder"""'], {'reuse': 'tf.AUTO_REUSE'}), "('pre_decoder', reuse=tf.AUTO_REUSE)\n", (23487, 23523), True, 'import tensorflow.compat.v1 as tf\n'), ((23546, 23605), 'tensorflow.compat.v1.layers.dense', 'tf.layers.dense', (['inputs', 'hparams.hidden_size'], {'name': '"""bottom"""'}), "(inputs, hparams.hidden_size, name='bottom')\n", (23561, 23605), True, 'import tensorflow.compat.v1 as tf\n'), ((23644, 23662), 'tensorflow.compat.v1.nn.tanh', 'tf.nn.tanh', (['inputs'], {}), '(inputs)\n', (23654, 23662), True, 'import tensorflow.compat.v1 as tf\n'), ((23749, 23803), 'tensorflow.compat.v1.variable_scope', 'tf.variable_scope', (['"""lstm_decoder"""'], {'reuse': 'tf.AUTO_REUSE'}), "('lstm_decoder', reuse=tf.AUTO_REUSE)\n", (23766, 23803), True, 'import tensorflow.compat.v1 as tf\n'), ((23824, 23944), 'tensorflow.compat.v1.nn.dynamic_rnn', 'tf.nn.dynamic_rnn', (['layers', 'inputs', 'sequence_length'], {'initial_state': 'initial_state', 'dtype': 'tf.float32', 'time_major': '(False)'}), '(layers, inputs, sequence_length, initial_state=\n initial_state, dtype=tf.float32, time_major=False)\n', (23841, 23944), True, 'import tensorflow.compat.v1 as tf\n'), ((24040, 24072), 'tensor2tensor.layers.common_layers.shape_list', 'common_layers.shape_list', (['inputs'], {}), '(inputs)\n', (24064, 24072), False, 'from tensor2tensor.layers import common_layers\n'), ((24263, 24325), 'tensorflow.compat.v1.nn.rnn_cell.DropoutWrapper', 'tf.nn.rnn_cell.DropoutWrapper', ([], {'cell': 'cell', 'output_keep_prob': '(0.5)'}), '(cell=cell, output_keep_prob=0.5)\n', (24292, 24325), True, 'import tensorflow.compat.v1 as tf\n'), ((24488, 24551), 'tensorflow.compat.v1.nn.rnn_cell.DropoutWrapper', 'tf.nn.rnn_cell.DropoutWrapper', ([], {'cell': 'cell1', 'output_keep_prob': '(0.5)'}), '(cell=cell1, output_keep_prob=0.5)\n', (24517, 24551), True, 'import tensorflow.compat.v1 as tf\n'), ((24714, 24777), 'tensorflow.compat.v1.nn.rnn_cell.DropoutWrapper', 'tf.nn.rnn_cell.DropoutWrapper', ([], {'cell': 'cell2', 'output_keep_prob': '(0.5)'}), '(cell=cell2, output_keep_prob=0.5)\n', (24743, 24777), True, 'import tensorflow.compat.v1 as tf\n'), ((24940, 25003), 'tensorflow.compat.v1.nn.rnn_cell.DropoutWrapper', 'tf.nn.rnn_cell.DropoutWrapper', ([], {'cell': 'cell3', 'output_keep_prob': '(0.5)'}), '(cell=cell3, output_keep_prob=0.5)\n', (24969, 25003), True, 'import tensorflow.compat.v1 as tf\n'), ((25989, 26066), 'tensorflow.contrib.rnn.DropoutWrapper', 'contrib_rnn.DropoutWrapper', (['recurrent_dropout_cell'], {'input_keep_prob': 'keep_prob'}), '(recurrent_dropout_cell, input_keep_prob=keep_prob)\n', (26015, 26066), True, 'from tensorflow.contrib import rnn as contrib_rnn\n'), ((26300, 26361), 'tensorflow.compat.v1.layers.dense', 'tf.layers.dense', (['x', 'res_size'], {'name': '"""dense"""', 'activation': '"""tanh"""'}), "(x, res_size, name='dense', activation='tanh')\n", (26315, 26361), True, 'import tensorflow.compat.v1 as tf\n'), ((27075, 27115), 'tensor2tensor.layers.common_layers.shape_list', 'common_layers.shape_list', (['initial_output'], {}), '(initial_output)\n', (27099, 27115), False, 'from tensor2tensor.layers import common_layers\n'), ((27197, 27231), 'tensor2tensor.layers.common_layers.shape_list', 'common_layers.shape_list', (['zero_pad'], {}), '(zero_pad)\n', (27221, 27231), False, 'from tensor2tensor.layers import common_layers\n'), ((2754, 2776), 'tensorflow.compat.v1.reduce_mean', 'tf.reduce_mean', (['b_loss'], {}), '(b_loss)\n', (2768, 2776), True, 'import tensorflow.compat.v1 as tf\n'), ((5258, 5289), 'tensorflow.compat.v1.maximum', 'tf.maximum', (['(kl - free_bits)', '(0.0)'], {}), '(kl - free_bits, 0.0)\n', (5268, 5289), True, 'import tensorflow.compat.v1 as tf\n'), ((5680, 5755), 'tensorflow.compat.v1.layers.Conv2D', 'tf.layers.Conv2D', (['hparams.base_depth', '(5)', '(1)'], {'padding': '"""SAME"""', 'activation': 'None'}), "(hparams.base_depth, 5, 1, padding='SAME', activation=None)\n", (5696, 5755), True, 'import tensorflow.compat.v1 as tf\n'), ((5932, 6007), 'tensorflow.compat.v1.layers.Conv2D', 'tf.layers.Conv2D', (['hparams.base_depth', '(5)', '(2)'], {'padding': '"""SAME"""', 'activation': 'None'}), "(hparams.base_depth, 5, 2, padding='SAME', activation=None)\n", (5948, 6007), True, 'import tensorflow.compat.v1 as tf\n'), ((6184, 6263), 'tensorflow.compat.v1.layers.Conv2D', 'tf.layers.Conv2D', (['(2 * hparams.base_depth)', '(5)', '(1)'], {'padding': '"""SAME"""', 'activation': 'None'}), "(2 * hparams.base_depth, 5, 1, padding='SAME', activation=None)\n", (6200, 6263), True, 'import tensorflow.compat.v1 as tf\n'), ((6440, 6519), 'tensorflow.compat.v1.layers.Conv2D', 'tf.layers.Conv2D', (['(2 * hparams.base_depth)', '(5)', '(2)'], {'padding': '"""SAME"""', 'activation': 'None'}), "(2 * hparams.base_depth, 5, 2, padding='SAME', activation=None)\n", (6456, 6519), True, 'import tensorflow.compat.v1 as tf\n'), ((6746, 6834), 'tensorflow.compat.v1.layers.Conv2D', 'tf.layers.Conv2D', (['(2 * hparams.bottleneck_bits)', '(4)', '(2)'], {'padding': '"""SAME"""', 'activation': 'None'}), "(2 * hparams.bottleneck_bits, 4, 2, padding='SAME',\n activation=None)\n", (6762, 6834), True, 'import tensorflow.compat.v1 as tf\n'), ((7057, 7145), 'tensorflow.compat.v1.layers.Conv2D', 'tf.layers.Conv2D', (['(2 * hparams.bottleneck_bits)', '(4)', '(2)'], {'padding': '"""SAME"""', 'activation': 'None'}), "(2 * hparams.bottleneck_bits, 4, 2, padding='SAME',\n activation=None)\n", (7073, 7145), True, 'import tensorflow.compat.v1 as tf\n'), ((7919, 8011), 'tensorflow.compat.v1.layers.Conv2DTranspose', 'tf.layers.Conv2DTranspose', (['(2 * hparams.base_depth)', '(4)', '(2)'], {'padding': '"""SAME"""', 'activation': 'None'}), "(2 * hparams.base_depth, 4, 2, padding='SAME',\n activation=None)\n", (7944, 8011), True, 'import tensorflow.compat.v1 as tf\n'), ((8236, 8328), 'tensorflow.compat.v1.layers.Conv2DTranspose', 'tf.layers.Conv2DTranspose', (['(2 * hparams.base_depth)', '(4)', '(2)'], {'padding': '"""SAME"""', 'activation': 'None'}), "(2 * hparams.base_depth, 4, 2, padding='SAME',\n activation=None)\n", (8261, 8328), True, 'import tensorflow.compat.v1 as tf\n'), ((8510, 8599), 'tensorflow.compat.v1.layers.Conv2DTranspose', 'tf.layers.Conv2DTranspose', (['(2 * hparams.base_depth)', '(5)'], {'padding': '"""SAME"""', 'activation': 'None'}), "(2 * hparams.base_depth, 5, padding='SAME',\n activation=None)\n", (8535, 8599), True, 'import tensorflow.compat.v1 as tf\n'), ((8781, 8873), 'tensorflow.compat.v1.layers.Conv2DTranspose', 'tf.layers.Conv2DTranspose', (['(2 * hparams.base_depth)', '(5)', '(2)'], {'padding': '"""SAME"""', 'activation': 'None'}), "(2 * hparams.base_depth, 5, 2, padding='SAME',\n activation=None)\n", (8806, 8873), True, 'import tensorflow.compat.v1 as tf\n'), ((9055, 9141), 'tensorflow.compat.v1.layers.Conv2DTranspose', 'tf.layers.Conv2DTranspose', (['hparams.base_depth', '(5)'], {'padding': '"""SAME"""', 'activation': 'None'}), "(hparams.base_depth, 5, padding='SAME', activation\n =None)\n", (9080, 9141), True, 'import tensorflow.compat.v1 as tf\n'), ((9322, 9410), 'tensorflow.compat.v1.layers.Conv2DTranspose', 'tf.layers.Conv2DTranspose', (['hparams.base_depth', '(5)', '(2)'], {'padding': '"""SAME"""', 'activation': 'None'}), "(hparams.base_depth, 5, 2, padding='SAME',\n activation=None)\n", (9347, 9410), True, 'import tensorflow.compat.v1 as tf\n'), ((9592, 9678), 'tensorflow.compat.v1.layers.Conv2DTranspose', 'tf.layers.Conv2DTranspose', (['hparams.base_depth', '(5)'], {'padding': '"""SAME"""', 'activation': 'None'}), "(hparams.base_depth, 5, padding='SAME', activation\n =None)\n", (9617, 9678), True, 'import tensorflow.compat.v1 as tf\n'), ((9859, 9914), 'tensorflow.compat.v1.layers.Conv2D', 'tf.layers.Conv2D', (['(1)', '(5)'], {'padding': '"""SAME"""', 'activation': 'None'}), "(1, 5, padding='SAME', activation=None)\n", (9875, 9914), True, 'import tensorflow.compat.v1 as tf\n'), ((10504, 10539), 'tensorflow.compat.v1.VariableScope', 'tf.VariableScope', (['tf.AUTO_REUSE', '""""""'], {}), "(tf.AUTO_REUSE, '')\n", (10520, 10539), True, 'import tensorflow.compat.v1 as tf\n'), ((14424, 14457), 'tensor2tensor.layers.common_layers.flatten4d3d', 'common_layers.flatten4d3d', (['source'], {}), '(source)\n', (14449, 14457), False, 'from tensor2tensor.layers import common_layers\n'), ((15850, 15902), 'tensor2tensor.layers.common_layers.length_from_embedding', 'common_layers.length_from_embedding', (['shifted_targets'], {}), '(shifted_targets)\n', (15885, 15902), False, 'from tensor2tensor.layers import common_layers\n'), ((18605, 18640), 'tensorflow.compat.v1.VariableScope', 'tf.VariableScope', (['tf.AUTO_REUSE', '""""""'], {}), "(tf.AUTO_REUSE, '')\n", (18621, 18640), True, 'import tensorflow.compat.v1 as tf\n'), ((20348, 20401), 'tensorflow.compat.v1.variable_scope', 'tf.variable_scope', (['"""pre_decoder"""'], {'reuse': 'tf.AUTO_REUSE'}), "('pre_decoder', reuse=tf.AUTO_REUSE)\n", (20365, 20401), True, 'import tensorflow.compat.v1 as tf\n'), ((20428, 20487), 'tensorflow.compat.v1.layers.dense', 'tf.layers.dense', (['inputs', 'hparams.hidden_size'], {'name': '"""bottom"""'}), "(inputs, hparams.hidden_size, name='bottom')\n", (20443, 20487), True, 'import tensorflow.compat.v1 as tf\n'), ((20513, 20531), 'tensorflow.compat.v1.nn.tanh', 'tf.nn.tanh', (['inputs'], {}), '(inputs)\n', (20523, 20531), True, 'import tensorflow.compat.v1 as tf\n'), ((20573, 20627), 'tensorflow.compat.v1.variable_scope', 'tf.variable_scope', (['"""lstm_decoder"""'], {'reuse': 'tf.AUTO_REUSE'}), "('lstm_decoder', reuse=tf.AUTO_REUSE)\n", (20590, 20627), True, 'import tensorflow.compat.v1 as tf\n'), ((20669, 20788), 'tensorflow.compat.v1.nn.dynamic_rnn', 'tf.nn.dynamic_rnn', (['layers', 'inputs', 'seq_len_batch'], {'initial_state': 'current_hidden', 'dtype': 'tf.float32', 'time_major': '(False)'}), '(layers, inputs, seq_len_batch, initial_state=\n current_hidden, dtype=tf.float32, time_major=False)\n', (20686, 20788), True, 'import tensorflow.compat.v1 as tf\n'), ((21236, 21275), 'tensor2tensor.layers.common_layers.shape_list', 'common_layers.shape_list', (['logits_so_far'], {}), '(logits_so_far)\n', (21260, 21275), False, 'from tensor2tensor.layers import common_layers\n'), ((25363, 25392), 'tensorflow.compat.v1.fill', 'tf.fill', (['[batch_size]', 'length'], {}), '([batch_size], length)\n', (25370, 25392), True, 'import tensorflow.compat.v1 as tf\n'), ((25726, 25744), 'tensorflow.compat.v1.to_float', 'tf.to_float', (['train'], {}), '(train)\n', (25737, 25744), True, 'import tensorflow.compat.v1 as tf\n'), ((3993, 4012), 'tensorflow.compat.v1.name_scope', 'tf.name_scope', (['None'], {}), '(None)\n', (4006, 4012), True, 'import tensorflow.compat.v1 as tf\n'), ((4014, 4057), 'tensorflow.compat.v1.name_scope', 'tf.name_scope', (["('train' if train else 'test')"], {}), "('train' if train else 'test')\n", (4027, 4057), True, 'import tensorflow.compat.v1 as tf\n'), ((4148, 4187), 'tensorflow.compat.v1.summary.image', 'tf.summary.image', (['"""rendered_og"""', 'inputs'], {}), "('rendered_og', inputs)\n", (4164, 4187), True, 'import tensorflow.compat.v1 as tf\n'), ((4965, 4986), 'tensorflow.compat.v1.exp', 'tf.exp', (['(log_sigma / 2)'], {}), '(log_sigma / 2)\n', (4971, 4986), True, 'import tensorflow.compat.v1 as tf\n'), ((16254, 16296), 'tensor2tensor.layers.common_layers.flatten4d3d', 'common_layers.flatten4d3d', (['shifted_targets'], {}), '(shifted_targets)\n', (16279, 16296), False, 'from tensor2tensor.layers import common_layers\n'), ((16581, 16623), 'tensor2tensor.layers.common_layers.flatten4d3d', 'common_layers.flatten4d3d', (['shifted_targets'], {}), '(shifted_targets)\n', (16606, 16623), False, 'from tensor2tensor.layers import common_layers\n'), ((17872, 17912), 'tensorflow.compat.v1.one_hot', 'tf.one_hot', (['clss', 'hparams.num_categories'], {}), '(clss, hparams.num_categories)\n', (17882, 17912), True, 'import tensorflow.compat.v1 as tf\n'), ((17950, 17984), 'tensorflow.compat.v1.one_hot', 'tf.one_hot', (['sln', 'max_decode_length'], {}), '(sln, max_decode_length)\n', (17960, 17984), True, 'import tensorflow.compat.v1 as tf\n'), ((18089, 18129), 'tensorflow.compat.v1.one_hot', 'tf.one_hot', (['clss', 'hparams.num_categories'], {}), '(clss, hparams.num_categories)\n', (18099, 18129), True, 'import tensorflow.compat.v1 as tf\n'), ((18312, 18357), 'tensorflow.compat.v1.nn.rnn_cell.LSTMStateTuple', 'tf.nn.rnn_cell.LSTMStateTuple', ([], {'c': 's[0]', 'h': 's[1]'}), '(c=s[0], h=s[1])\n', (18341, 18357), True, 'import tensorflow.compat.v1 as tf\n'), ((21647, 21699), 'tensorflow.compat.v1.TensorShape', 'tf.TensorShape', (['[None, None, 1, hparams.hidden_size]'], {}), '([None, None, 1, hparams.hidden_size])\n', (21661, 21699), True, 'import tensorflow.compat.v1 as tf\n'), ((22726, 22766), 'tensorflow.compat.v1.one_hot', 'tf.one_hot', (['clss', 'hparams.num_categories'], {}), '(clss, hparams.num_categories)\n', (22736, 22766), True, 'import tensorflow.compat.v1 as tf\n'), ((22804, 22823), 'tensorflow.compat.v1.one_hot', 'tf.one_hot', (['sln', '(51)'], {}), '(sln, 51)\n', (22814, 22823), True, 'import tensorflow.compat.v1 as tf\n'), ((22928, 22968), 'tensorflow.compat.v1.one_hot', 'tf.one_hot', (['clss', 'hparams.num_categories'], {}), '(clss, hparams.num_categories)\n', (22938, 22968), True, 'import tensorflow.compat.v1 as tf\n'), ((23185, 23221), 'tensor2tensor.layers.common_layers.shape_list', 'common_layers.shape_list', (['bottleneck'], {}), '(bottleneck)\n', (23209, 23221), False, 'from tensor2tensor.layers import common_layers\n'), ((23363, 23395), 'tensor2tensor.layers.common_layers.shape_list', 'common_layers.shape_list', (['inputs'], {}), '(inputs)\n', (23387, 23395), False, 'from tensor2tensor.layers import common_layers\n'), ((2839, 2887), 'tensor2tensor.layers.common_layers.shape_list', 'common_layers.shape_list', (["features['bottleneck']"], {}), "(features['bottleneck'])\n", (2863, 2887), False, 'from tensor2tensor.layers import common_layers\n'), ((13670, 13718), 'tensor2tensor.layers.common_layers.shape_list', 'common_layers.shape_list', (["features['bottleneck']"], {}), "(features['bottleneck'])\n", (13694, 13718), False, 'from tensor2tensor.layers import common_layers\n'), ((19882, 19918), 'tensor2tensor.layers.common_layers.shape_list', 'common_layers.shape_list', (['bottleneck'], {}), '(bottleneck)\n', (19906, 19918), False, 'from tensor2tensor.layers import common_layers\n'), ((20106, 20147), 'tensor2tensor.layers.common_layers.shape_list', 'common_layers.shape_list', (['shifted_targets'], {}), '(shifted_targets)\n', (20130, 20147), False, 'from tensor2tensor.layers import common_layers\n'), ((20259, 20291), 'tensor2tensor.layers.common_layers.shape_list', 'common_layers.shape_list', (['inputs'], {}), '(inputs)\n', (20283, 20291), False, 'from tensor2tensor.layers import common_layers\n'), ((15357, 15415), 'tensorflow.compat.v1.concat', 'tf.concat', (['[unbottleneck[:, :unbottleneck_dim // 2], c]', '(1)'], {}), '([unbottleneck[:, :unbottleneck_dim // 2], c], 1)\n', (15366, 15415), True, 'import tensorflow.compat.v1 as tf\n'), ((15443, 15501), 'tensorflow.compat.v1.concat', 'tf.concat', (['[unbottleneck[:, unbottleneck_dim // 2:], h]', '(1)'], {}), '([unbottleneck[:, unbottleneck_dim // 2:], h], 1)\n', (15452, 15501), True, 'import tensorflow.compat.v1 as tf\n'), ((5035, 5052), 'tensorflow.compat.v1.exp', 'tf.exp', (['log_sigma'], {}), '(log_sigma)\n', (5041, 5052), True, 'import tensorflow.compat.v1 as tf\n'), ((5055, 5068), 'tensorflow.compat.v1.square', 'tf.square', (['mu'], {}), '(mu)\n', (5064, 5068), True, 'import tensorflow.compat.v1 as tf\n')] |
import csv
import numpy as np
import time
from pathlib import Path
from Panalyzer.utils.wr_extractor import wr_extractor
from Panalyzer.TraceParser.logic_masking import *
def arm32buffered_csv2np(fcsv, buffersize, num_reg):
detailded_info = {'wr': None, 'regval': None, 'tick': None, 'masking': None, 'src1': None, 'src2': None,
'op': None}
tick_list = np.zeros([buffersize], dtype=np.int64)
wr_list = np.full([num_reg, 2, buffersize], False, dtype=bool)
reg_val_table = np.zeros([num_reg, buffersize], dtype=np.int64)
op_list = []
src1_list = []
src2_list = []
with open(fcsv, mode='r') as infocsv:
info_reader = csv.reader(infocsv)
buffer_idx = 0
chunk_counter = 0
for idx, row in enumerate(info_reader):
if idx % buffersize == 0:
buffer_idx = 0
chunk_counter = chunk_counter + 1
print(chunk_counter)
tick_list = np.zeros([buffersize], dtype=np.int64)
wr_list = np.full([num_reg, 2, buffersize], False, dtype=bool)
reg_val_table = np.zeros([num_reg, buffersize], dtype=np.int64)
op_list = []
src1_list = []
src2_list = []
else:
buffer_idx = buffer_idx + 1
tick_list[buffer_idx] = row[0] # Tick number list: an 1 x line_number np array
op_id = row[3]
op_list.append(op_id) # Opname is just a simple list of strings
# Variables required for utility.wr_extractor, feed into the function, then abstract the required
# data structure
op_dst1 = row[4]
op_dst2 = row[5]
op_src1 = row[6]
op_src2 = row[7]
src1_list.append(op_src1)
src2_list.append(op_src2)
data = row[-1]
for k in range(num_reg): # kth register
val_prev = reg_val_table[k, buffer_idx - 1]
reg_name = 'r' + str(k) # fp, lr, sp ,pc are renamed, simply
wr_list[k, 0, buffer_idx] = \
wr_extractor(reg_name, op_dst1, op_dst2, op_src1, op_src2, op_id, data, val_prev)[0]
wr_list[k, 1, buffer_idx] = \
wr_extractor(reg_name, op_dst1, op_dst2, op_src1, op_src2, op_id, data, val_prev)[1]
reg_val_table[k, buffer_idx] = \
wr_extractor(reg_name, op_dst1, op_dst2, op_src1, op_src2, op_id, data, val_prev)[2]
return tick_list
if __name__ == "__main__":
project_dir = Path(__file__).resolve().parent.parent
csv_dir = project_dir.joinpath('tempcsv')
fname = "fftbaseline.csv"
start_time = time.perf_counter() # Time counter starts
T = arm32buffered_csv2np(csv_dir / fname, 10000, 16)
elapsed_time_pandas = time.perf_counter() - start_time # Stop point of the timer
# tickexample = T['tick']
# wrexample = T['wr']
# regvalexample = T['regval']
# masking_table = T['masking']
# ops_list = T['op']
#
# # print('tick \n', tickexample, '\n wr: \n', wrexample, '\n regval:\n', regvalexample)
# print(ops_list) | [
"pathlib.Path",
"Panalyzer.utils.wr_extractor.wr_extractor",
"time.perf_counter",
"numpy.zeros",
"numpy.full",
"csv.reader"
] | [((402, 440), 'numpy.zeros', 'np.zeros', (['[buffersize]'], {'dtype': 'np.int64'}), '([buffersize], dtype=np.int64)\n', (410, 440), True, 'import numpy as np\n'), ((456, 508), 'numpy.full', 'np.full', (['[num_reg, 2, buffersize]', '(False)'], {'dtype': 'bool'}), '([num_reg, 2, buffersize], False, dtype=bool)\n', (463, 508), True, 'import numpy as np\n'), ((530, 577), 'numpy.zeros', 'np.zeros', (['[num_reg, buffersize]'], {'dtype': 'np.int64'}), '([num_reg, buffersize], dtype=np.int64)\n', (538, 577), True, 'import numpy as np\n'), ((2905, 2924), 'time.perf_counter', 'time.perf_counter', ([], {}), '()\n', (2922, 2924), False, 'import time\n'), ((706, 725), 'csv.reader', 'csv.reader', (['infocsv'], {}), '(infocsv)\n', (716, 725), False, 'import csv\n'), ((3033, 3052), 'time.perf_counter', 'time.perf_counter', ([], {}), '()\n', (3050, 3052), False, 'import time\n'), ((1017, 1055), 'numpy.zeros', 'np.zeros', (['[buffersize]'], {'dtype': 'np.int64'}), '([buffersize], dtype=np.int64)\n', (1025, 1055), True, 'import numpy as np\n'), ((1083, 1135), 'numpy.full', 'np.full', (['[num_reg, 2, buffersize]', '(False)'], {'dtype': 'bool'}), '([num_reg, 2, buffersize], False, dtype=bool)\n', (1090, 1135), True, 'import numpy as np\n'), ((1169, 1216), 'numpy.zeros', 'np.zeros', (['[num_reg, buffersize]'], {'dtype': 'np.int64'}), '([num_reg, buffersize], dtype=np.int64)\n', (1177, 1216), True, 'import numpy as np\n'), ((2768, 2782), 'pathlib.Path', 'Path', (['__file__'], {}), '(__file__)\n', (2772, 2782), False, 'from pathlib import Path\n'), ((2281, 2366), 'Panalyzer.utils.wr_extractor.wr_extractor', 'wr_extractor', (['reg_name', 'op_dst1', 'op_dst2', 'op_src1', 'op_src2', 'op_id', 'data', 'val_prev'], {}), '(reg_name, op_dst1, op_dst2, op_src1, op_src2, op_id, data,\n val_prev)\n', (2293, 2366), False, 'from Panalyzer.utils.wr_extractor import wr_extractor\n'), ((2442, 2527), 'Panalyzer.utils.wr_extractor.wr_extractor', 'wr_extractor', (['reg_name', 'op_dst1', 'op_dst2', 'op_src1', 'op_src2', 'op_id', 'data', 'val_prev'], {}), '(reg_name, op_dst1, op_dst2, op_src1, op_src2, op_id, data,\n val_prev)\n', (2454, 2527), False, 'from Panalyzer.utils.wr_extractor import wr_extractor\n'), ((2606, 2691), 'Panalyzer.utils.wr_extractor.wr_extractor', 'wr_extractor', (['reg_name', 'op_dst1', 'op_dst2', 'op_src1', 'op_src2', 'op_id', 'data', 'val_prev'], {}), '(reg_name, op_dst1, op_dst2, op_src1, op_src2, op_id, data,\n val_prev)\n', (2618, 2691), False, 'from Panalyzer.utils.wr_extractor import wr_extractor\n')] |
"""
define model for gp
"""
# from threading import Thread
# from queue import Queue
from multiprocessing import Pool
from random import random, randint
from math import floor
import operator
from autoprover.gp.gene import Gene
from autoprover.gp.rule import GeneRule
from autoprover.gp.action import GeneAction
from autoprover.gp.trigger import GeneTrigger
from autoprover.gp.restriction import Restriction
#TODO fox too many instant
class GPModel:
"""
gp model
"""
#TODO fix too many args
def __init__(self, args=None, populationSize=None, maxGeneration=None,
mutateRate=None, eliteRate=None, crossRate=None,
crossType=None, verifyNum=None, proof=None, tactics=None,
limit_hyp=None, limit_goal=None):
self.population_size = populationSize or args.populationSize
self.max_generation = maxGeneration or args.maxGeneration
self.mutate_rate = mutateRate or args.mutateRate
self.elite_rate = eliteRate or args.eliteRate
self.cross_rate = crossRate or args.crossRate
self.cross_type = crossType or args.crossType
self.verify_num = verifyNum or args.verifyNum
self.limit_hyp = limit_hyp or args.limit_hyp
self.limit_goal = limit_goal or args.limit_goal
self.debug = args.debug
self.proof = proof
self.tactics = tactics
self.population = None
self.current_generation = 1
self.proofs = []
self.rules = []
self.init_population(self.population_size)
self.pre_process()
def show_prop(self):
"""
display property for model
"""
print(self.population_size)
print(self.max_generation)
print(self.mutate_rate)
print(self.elite_rate)
print(self.cross_rate)
print(self.cross_type)
print(self.verify_num)
print(self.proof)
def init_population(self, size):
"""
create population by size
"""
print("Initializing population.")
self.population = []
for _ in range(size):
self.population.append(Gene(self.tactics))
def pre_process(self):
"""
run before start
"""
self.current_generation = 1
self.update_fitness_for_population()
self.fitness_sharing()
self.sort_population()
self.update_tactic_usage()
self.check_proof()
def is_proved(self):
"""
check population has a proof
"""
return len(self.proofs) > 0
def start(self, gen=None):
"""
run the model
"""
if gen is None:
# if gen is not set
local_gen_limit = self.max_generation + 1
else:
local_gen_limit = gen
if self.current_generation > self.max_generation:
return
for _ in range(local_gen_limit):
print("Generation No.{0}".format(self.current_generation))
if self.debug:
self.sort_population()
for index in range(0, 30):
self.population[index].print_lastest()
self.crossover()
self.update_fitness_for_population()
self.apply_rules()
self.fitness_sharing()
self.next_generation()
if self.current_generation > self.max_generation:
break
# self.printGeneByIndex(0, True)
def fitness_sharing(self):
"""Use fitness sharing to re-evaluate fitness"""
for gene in self.population:
gene.fitness = gene.raw_fitness
return
def dist(gene1, gene2):
"""Return distence between two gene"""
return abs(len(gene1.goal) - len(gene2.goal))
for gene in self.population:
raw_fitnesses = [e.raw_fitness for e in self.population if dist(e, gene) <= 5]
gene.fitness = sum(raw_fitnesses) / len(raw_fitnesses)
def next_generation(self):
"""
next generation
"""
print("Avg. fitness\tAvg. length")
print("{0:.8f}\t{1}".format(self.average_fitness(),
self.average_length_of_gene()))
self.current_generation += 1
self.sort_population()
self.update_tactic_usage()
self.check_proof()
def check_proof(self):
"""Check if there is a proof in population
"""
for gene in self.population:
if gene.is_proof:
print(gene.chromosome)
for state in gene.coq_states:
print(state)
self.proofs.append(Gene(chromosome=gene.valid_tactics))
def update_fitness_for_population(self):
"""
return individual if theorem is proved, o.w return None
"""
def wrapper(func, *args, **kwargs):
"""func wrapper"""
return func, args, kwargs
with Pool(processes=4) as pool:
for gene in self.population:
func, args, kargs = wrapper(gene.update_fitness_for_proof,
self.proof, self.limit_hyp,
self.limit_goal)
pool.apply_async(func(*args, **kargs))
def apply_rules(self):
"""Perform action by rules"""
if len(self.rules) == 0:
return
for gene in self.population:
for rule in self.rules:
if rule.type == "gene":
rule.check_and_apply(gene)
def crossover(self):
"""
the crossover operation for gp
"""
self.sort_population()
elite_amount = round(self.elite_rate * self.population_size)
# preserve from the top
new_population = [ele for ele in self.population if ele.ttl > 0]
for individual in new_population:
if individual.ttl > 0:
individual.ttl -= 1
new_population += self.population[:elite_amount]
while len(new_population) < self.population_size:
# newGene = self.crossBelowCrossRate()
new_gene, new_gene2 = self.cross_on_arb_seq()
if random() <= self.mutate_rate:
self.mutate_append(new_gene)
new_population.append(new_gene)
if len(new_population) == self.population_size:
break
if random() <= self.mutate_rate:
self.mutate_append(new_gene2)
new_population.append(new_gene2)
self.population = new_population
def sort_population(self):
"""
sort population by length and fitness
"""
self.population.sort(key=lambda x: x.fitness, reverse=True)
def update_tactic_usage(self):
"""update tactic statistic usage"""
usage = {e: 0 for e in self.tactics.all_tactics}
count = 0
for gene in self.population:
for tactic in gene.chromosome:
count += 1
try:
usage[tactic] += 1
except KeyError:
usage[tactic] = 1
for tactic in usage:
usage[tactic] = usage[tactic]/count
self.proof.tactics.usage = usage
def cross_below_cross_rate(self):
"""
select two parent by cross rate, crossover on random point
"""
p1_index = randint(0, floor(self.population_size * self.cross_rate)-1)
p2_index = randint(0, floor(self.population_size * self.cross_rate)-1)
gene_of_p1 = self.population[p1_index]
gene_of_p2 = self.population[p2_index]
cross_point = randint(0, int_min(len(gene_of_p1), len(gene_of_p2))-1)
new_chromosome = []
new_chromosome += gene_of_p1.chromosome[:cross_point]
new_chromosome += gene_of_p2.chromosome[cross_point:]
if (self.tactics.is_unrepeatable(new_chromosome[cross_point])
and cross_point < len(new_chromosome)-1):
if new_chromosome[cross_point] == new_chromosome[cross_point+1]:
del new_chromosome[cross_point]
return Gene(chromosome=new_chromosome)
def cross_on_arb_seq(self, slmax=6):
"""
select two parent by cross_rate, crossover by some seqence
"""
p1_index = randint(0, floor(self.population_size * self.cross_rate)-1)
p2_index = randint(0, floor(self.population_size * self.cross_rate)-1)
gene_of_p1 = self.population[p1_index]
gene_of_p2 = self.population[p2_index]
p1_begin = myrandint(0, len(gene_of_p1)-1)
p1_end = p1_begin + myrandint(1, int_min(slmax, len(gene_of_p1)-p1_begin))
p2_begin = myrandint(0, len(gene_of_p2)-1)
p2_end = p2_begin + myrandint(1, int_min(slmax, len(gene_of_p2)-p2_begin))
new_chromosome = []
new_chromosome += gene_of_p1.chromosome[:p1_begin]
new_chromosome += gene_of_p2.chromosome[p2_begin:p2_end]
new_chromosome += gene_of_p1.chromosome[p1_end:]
new_chromosome2 = []
new_chromosome2 += gene_of_p2.chromosome[:p2_begin]
new_chromosome2 += gene_of_p1.chromosome[p1_begin:p1_end]
new_chromosome2 += gene_of_p2.chromosome[p2_end:]
self.remove_repeatable(new_chromosome)
self.remove_repeatable(new_chromosome2)
return Gene(chromosome=new_chromosome), Gene(chromosome=new_chromosome2)
def remove_repeatable(self, chromosome):
"""
remove repeatable tactic
"""
tactic_set = set()
for index, tactic in enumerate(chromosome):
if self.tactics.is_unrepeatable(tactic):
if tactic in tactic_set:
del chromosome[index]
else:
tactic_set.add(tactic)
def mutate(self, gene):
"""
the mutate operation
"""
if len(gene) == 1:
gene.chromosome[0] = self.tactics.mutate_select()
else:
index = randint(len(gene)//2, len(gene)-1)
gene.chromosome[index] = self.tactics.mutate_select()
def mutate_append(self, gene):
"""append a tactic to chromosome"""
gene.chromosome.append(self.tactics.mutate_select())
def average_fitness(self):
"""Calculate the average fitness for population.
Returns:
double: avg. fitness
"""
return sum([e.fitness for e in self.population]) / len(self.population)
def average_length_of_gene(self):
"""Calculate the average fitness for population.
Returns:
double: avg. fitness
"""
return sum([len(e) for e in self.population]) / len(self.population)
def edit(self, index, data=None):
"""Human involved modification of some gene of the population
"""
if self.current_generation > self.max_generation:
return
print("Edit Gene {} now.".format(index))
gene = self.population[index]
gene.modification(data=data)
gene.update_fitness_for_proof(self.proof)
if gene.is_proof:
self.proofs.append(Gene(chromosome=gene.valid_tactics))
return
def show_proofs(self):
"""Show proofs found
"""
if self.proofs:
for gene in self.proofs:
print(gene.format_output(self.proof))
else:
print("There is not proof for now.")
def list(self, argv):
"""List property of some individual.
Args:
argv(list): sub command of list function.
"""
def get_interval(interval):
"""Get begin and end of interval
"""
interval_list = interval.split("-")
if len(interval_list) == 1:
return (int(interval_list[0]), int(interval_list[0])+1)
else:
return (int(interval_list[0]), int(interval_list[1])+1)
if not argv or not argv[0]:
return
(begin, end) = get_interval(argv[0])
if len(argv) == 1:
for index, gene in enumerate(self.population[begin:end]):
print("{0}: {1:.8f}".format(index, gene.fitness))
gene.print_progress()
elif argv[1] == "fitness":
for index, gene in enumerate(self.population[begin:end]):
print("{0}: {1:.8f}".format(index, gene.fitness))
elif argv[1] == "chromosome":
for index, gene in enumerate(self.population[begin:end]):
print("{0}: {1}".format(index, gene.chromosome))
elif argv[1] == "ttl":
for index, gene in enumerate(self.population[begin:end]):
print("{0}: {1}".format(index, gene.ttl))
def read_rule_from_file(self, file_name):
"""Read a rule from a JSON file"""
try:
self.rules.append(GeneRule(file_name=file_name, proof=self.proof))
except FileNotFoundError:
return
def delete_rule(self, index):
"""delete a rule from rule_list"""
del self.rules[index]
def remove_tactic(self):
"""remove a tactic in tactic_set and population"""
tactic_removed = input("Enter a tactic to be removed: ")
self.proof.tactics.remove(tactic_removed)
for gene in self.population:
gene.chromosome = [e for e in gene.chromosome if e != tactic_removed]
def set_attributes(self, argv):
"""Set attributes of population
"""
if argv[0] == "population" or argv[0] == "pop":
if argv[1] == "ttl":
self.population[int(argv[2])].ttl = int(argv[3])
def defrag(self, index_list):
"""Defrag some gene"""
for index in index_list:
self.population[index].defrag(self.proof)
def print_stats(self):
"""print tactic usage"""
sorted_stats = sorted(self.proof.tactics.usage.items(),
key=operator.itemgetter(1), reverse=True)
for tactic, usage in sorted_stats:
print("{0}: {1:.4f}%".format(tactic, usage*100))
def myrandint(begin, end):
"""
randint warrper for begin == end
"""
if begin == end:
return begin
else:
return randint(begin, end)
def int_max(int_a, int_b):
"""
max(a, b)
"""
if int_a > int_b:
return int_a
else:
return int_b
def int_min(int_a, int_b):
"""
min(a, b)
"""
if int_a < int_b:
return int_a
else:
return int_b
| [
"math.floor",
"autoprover.gp.gene.Gene",
"multiprocessing.Pool",
"autoprover.gp.rule.GeneRule",
"operator.itemgetter",
"random.random",
"random.randint"
] | [((8178, 8209), 'autoprover.gp.gene.Gene', 'Gene', ([], {'chromosome': 'new_chromosome'}), '(chromosome=new_chromosome)\n', (8182, 8209), False, 'from autoprover.gp.gene import Gene\n'), ((14306, 14325), 'random.randint', 'randint', (['begin', 'end'], {}), '(begin, end)\n', (14313, 14325), False, 'from random import random, randint\n'), ((4982, 4999), 'multiprocessing.Pool', 'Pool', ([], {'processes': '(4)'}), '(processes=4)\n', (4986, 4999), False, 'from multiprocessing import Pool\n'), ((9396, 9427), 'autoprover.gp.gene.Gene', 'Gene', ([], {'chromosome': 'new_chromosome'}), '(chromosome=new_chromosome)\n', (9400, 9427), False, 'from autoprover.gp.gene import Gene\n'), ((9429, 9461), 'autoprover.gp.gene.Gene', 'Gene', ([], {'chromosome': 'new_chromosome2'}), '(chromosome=new_chromosome2)\n', (9433, 9461), False, 'from autoprover.gp.gene import Gene\n'), ((2146, 2164), 'autoprover.gp.gene.Gene', 'Gene', (['self.tactics'], {}), '(self.tactics)\n', (2150, 2164), False, 'from autoprover.gp.gene import Gene\n'), ((6238, 6246), 'random.random', 'random', ([], {}), '()\n', (6244, 6246), False, 'from random import random, randint\n'), ((6455, 6463), 'random.random', 'random', ([], {}), '()\n', (6461, 6463), False, 'from random import random, randint\n'), ((7458, 7503), 'math.floor', 'floor', (['(self.population_size * self.cross_rate)'], {}), '(self.population_size * self.cross_rate)\n', (7463, 7503), False, 'from math import floor\n'), ((7537, 7582), 'math.floor', 'floor', (['(self.population_size * self.cross_rate)'], {}), '(self.population_size * self.cross_rate)\n', (7542, 7582), False, 'from math import floor\n'), ((8373, 8418), 'math.floor', 'floor', (['(self.population_size * self.cross_rate)'], {}), '(self.population_size * self.cross_rate)\n', (8378, 8418), False, 'from math import floor\n'), ((8452, 8497), 'math.floor', 'floor', (['(self.population_size * self.cross_rate)'], {}), '(self.population_size * self.cross_rate)\n', (8457, 8497), False, 'from math import floor\n'), ((11191, 11226), 'autoprover.gp.gene.Gene', 'Gene', ([], {'chromosome': 'gene.valid_tactics'}), '(chromosome=gene.valid_tactics)\n', (11195, 11226), False, 'from autoprover.gp.gene import Gene\n'), ((12928, 12975), 'autoprover.gp.rule.GeneRule', 'GeneRule', ([], {'file_name': 'file_name', 'proof': 'self.proof'}), '(file_name=file_name, proof=self.proof)\n', (12936, 12975), False, 'from autoprover.gp.rule import GeneRule\n'), ((14016, 14038), 'operator.itemgetter', 'operator.itemgetter', (['(1)'], {}), '(1)\n', (14035, 14038), False, 'import operator\n'), ((4684, 4719), 'autoprover.gp.gene.Gene', 'Gene', ([], {'chromosome': 'gene.valid_tactics'}), '(chromosome=gene.valid_tactics)\n', (4688, 4719), False, 'from autoprover.gp.gene import Gene\n')] |
import sys
sys.exit(1)
| [
"sys.exit"
] | [((12, 23), 'sys.exit', 'sys.exit', (['(1)'], {}), '(1)\n', (20, 23), False, 'import sys\n')] |
# -*- coding: utf-8 -*-
# Generated by Django 1.11.15 on 2018-11-26 10:42
from __future__ import unicode_literals
from django.db import migrations, models
import jsonfield.fields
class Migration(migrations.Migration):
initial = True
dependencies = [
]
operations = [
migrations.CreateModel(
name='Mapping',
fields=[
('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')),
('label', models.CharField(help_text='Label for your reference', max_length=255, verbose_name='label')),
('source', models.CharField(help_text='The source feed for your data', max_length=255, verbose_name='source')),
('parser', models.CharField(choices=[(b'feedmapper.parsers.AtomParser', b'Atom'), (b'feedmapper.parsers.XMLParser', b'XML')], help_text='Which parser to use when synchronizing', max_length=255, verbose_name='parser')),
('purge', models.BooleanField(default=False, help_text='Purge existing items on sync?', verbose_name='purge')),
('data_map', jsonfield.fields.JSONField(default=dict, verbose_name='data map')),
('notification_recipients', models.TextField(blank=True, help_text='Specify one email address per line to be notified of parsing errors.', verbose_name='notification recipients')),
('parse_attempted', models.DateTimeField(blank=True, null=True, verbose_name='parse attempted')),
('parse_succeeded', models.BooleanField(verbose_name='parse succeeded')),
('parse_log', models.TextField(blank=True, verbose_name='parse log')),
],
),
]
| [
"django.db.models.TextField",
"django.db.models.BooleanField",
"django.db.models.AutoField",
"django.db.models.DateTimeField",
"django.db.models.CharField"
] | [((393, 486), 'django.db.models.AutoField', 'models.AutoField', ([], {'auto_created': '(True)', 'primary_key': '(True)', 'serialize': '(False)', 'verbose_name': '"""ID"""'}), "(auto_created=True, primary_key=True, serialize=False,\n verbose_name='ID')\n", (409, 486), False, 'from django.db import migrations, models\n'), ((511, 607), 'django.db.models.CharField', 'models.CharField', ([], {'help_text': '"""Label for your reference"""', 'max_length': '(255)', 'verbose_name': '"""label"""'}), "(help_text='Label for your reference', max_length=255,\n verbose_name='label')\n", (527, 607), False, 'from django.db import migrations, models\n'), ((633, 735), 'django.db.models.CharField', 'models.CharField', ([], {'help_text': '"""The source feed for your data"""', 'max_length': '(255)', 'verbose_name': '"""source"""'}), "(help_text='The source feed for your data', max_length=255,\n verbose_name='source')\n", (649, 735), False, 'from django.db import migrations, models\n'), ((761, 981), 'django.db.models.CharField', 'models.CharField', ([], {'choices': "[(b'feedmapper.parsers.AtomParser', b'Atom'), (\n b'feedmapper.parsers.XMLParser', b'XML')]", 'help_text': '"""Which parser to use when synchronizing"""', 'max_length': '(255)', 'verbose_name': '"""parser"""'}), "(choices=[(b'feedmapper.parsers.AtomParser', b'Atom'), (\n b'feedmapper.parsers.XMLParser', b'XML')], help_text=\n 'Which parser to use when synchronizing', max_length=255, verbose_name=\n 'parser')\n", (777, 981), False, 'from django.db import migrations, models\n'), ((995, 1099), 'django.db.models.BooleanField', 'models.BooleanField', ([], {'default': '(False)', 'help_text': '"""Purge existing items on sync?"""', 'verbose_name': '"""purge"""'}), "(default=False, help_text=\n 'Purge existing items on sync?', verbose_name='purge')\n", (1014, 1099), False, 'from django.db import migrations, models\n'), ((1238, 1397), 'django.db.models.TextField', 'models.TextField', ([], {'blank': '(True)', 'help_text': '"""Specify one email address per line to be notified of parsing errors."""', 'verbose_name': '"""notification recipients"""'}), "(blank=True, help_text=\n 'Specify one email address per line to be notified of parsing errors.',\n verbose_name='notification recipients')\n", (1254, 1397), False, 'from django.db import migrations, models\n'), ((1427, 1502), 'django.db.models.DateTimeField', 'models.DateTimeField', ([], {'blank': '(True)', 'null': '(True)', 'verbose_name': '"""parse attempted"""'}), "(blank=True, null=True, verbose_name='parse attempted')\n", (1447, 1502), False, 'from django.db import migrations, models\n'), ((1541, 1592), 'django.db.models.BooleanField', 'models.BooleanField', ([], {'verbose_name': '"""parse succeeded"""'}), "(verbose_name='parse succeeded')\n", (1560, 1592), False, 'from django.db import migrations, models\n'), ((1625, 1679), 'django.db.models.TextField', 'models.TextField', ([], {'blank': '(True)', 'verbose_name': '"""parse log"""'}), "(blank=True, verbose_name='parse log')\n", (1641, 1679), False, 'from django.db import migrations, models\n')] |
"""
Markdownreveal local module tests.
"""
import json
import time
from hashlib import sha1
from pathlib import Path
from shutil import rmtree
from tarfile import TarInfo
from tempfile import mkdtemp
import pytest
from markdownreveal.local import clean_tar_members
from markdownreveal.local import initialize_localdir
from markdownreveal.local import latest_project_release
def test_latest_project_release():
"""
Test `latest_project_release()` function.
"""
# Latest reveal.js version
latest = latest_project_release(github='hakimel/reveal.js')
numbers = latest.split('.')
assert len(numbers) == 3
assert all(n.isdecimal() for n in numbers)
# Latest KaTeX version
latest = latest_project_release(github='khan/katex')
assert latest[0] == 'v'
latest = latest[1:]
numbers = latest.split('.')
assert len(numbers) == 3
assert all(n.isdecimal() for n in numbers)
def test_clean_tar_members():
"""
Test `clean_tar_members()` function.
"""
members = [
TarInfo('toplevel'),
TarInfo('toplevel/index.html'),
TarInfo('toplevel/foo/bar.xyz'),
]
output = [TarInfo('index.html'), TarInfo('foo/bar.xyz')]
result = clean_tar_members(members)
assert all(x.name == y.name for x, y in zip(output, result))
@pytest.mark.parametrize(
'reveal_version,katex_version,reveal_tag,katex_tag,style',
[
(
'latest',
'latest',
latest_project_release(github='hakimel/reveal.js'),
latest_project_release(github='khan/katex'),
'',
),
(
'3.4.0',
'v0.7.1',
'3.4.0',
'v0.7.1',
'https://github.com/markdownreveal/style-default/'
+ 'archive/master.tar.gz',
),
],
)
def test_initialize_localdir(
reveal_version, katex_version, reveal_tag, katex_tag, style
):
"""
Test `initialize_localdir()` function.
"""
localdir = Path(mkdtemp())
config = {
'local_path': localdir,
'reveal_version': reveal_version,
'katex_version': katex_version,
'style': style,
}
out = initialize_localdir(config)
package = json.load(open(str(out / 'revealjs' / 'package.json')))
assert package['version'] == reveal_tag
katex_readme = out / 'katex' / 'README.md'
assert katex_tag[1:] in katex_readme.read_text()
style_out = out / 'markdownrevealstyle'
if style:
style_dir = localdir / sha1(style.encode('utf')).hexdigest()
assert style_dir.exists()
assert style_out.exists()
else:
assert not style_out.exists()
# If version is already downloaded initialization should be pretty fast
t0 = time.time()
out = initialize_localdir(config)
assert time.time() - t0 < 0.01
rmtree(str(localdir))
| [
"markdownreveal.local.clean_tar_members",
"markdownreveal.local.initialize_localdir",
"tempfile.mkdtemp",
"markdownreveal.local.latest_project_release",
"time.time",
"tarfile.TarInfo"
] | [((518, 568), 'markdownreveal.local.latest_project_release', 'latest_project_release', ([], {'github': '"""hakimel/reveal.js"""'}), "(github='hakimel/reveal.js')\n", (540, 568), False, 'from markdownreveal.local import latest_project_release\n'), ((717, 760), 'markdownreveal.local.latest_project_release', 'latest_project_release', ([], {'github': '"""khan/katex"""'}), "(github='khan/katex')\n", (739, 760), False, 'from markdownreveal.local import latest_project_release\n'), ((1216, 1242), 'markdownreveal.local.clean_tar_members', 'clean_tar_members', (['members'], {}), '(members)\n', (1233, 1242), False, 'from markdownreveal.local import clean_tar_members\n'), ((2181, 2208), 'markdownreveal.local.initialize_localdir', 'initialize_localdir', (['config'], {}), '(config)\n', (2200, 2208), False, 'from markdownreveal.local import initialize_localdir\n'), ((2751, 2762), 'time.time', 'time.time', ([], {}), '()\n', (2760, 2762), False, 'import time\n'), ((2773, 2800), 'markdownreveal.local.initialize_localdir', 'initialize_localdir', (['config'], {}), '(config)\n', (2792, 2800), False, 'from markdownreveal.local import initialize_localdir\n'), ((1034, 1053), 'tarfile.TarInfo', 'TarInfo', (['"""toplevel"""'], {}), "('toplevel')\n", (1041, 1053), False, 'from tarfile import TarInfo\n'), ((1063, 1093), 'tarfile.TarInfo', 'TarInfo', (['"""toplevel/index.html"""'], {}), "('toplevel/index.html')\n", (1070, 1093), False, 'from tarfile import TarInfo\n'), ((1103, 1134), 'tarfile.TarInfo', 'TarInfo', (['"""toplevel/foo/bar.xyz"""'], {}), "('toplevel/foo/bar.xyz')\n", (1110, 1134), False, 'from tarfile import TarInfo\n'), ((1156, 1177), 'tarfile.TarInfo', 'TarInfo', (['"""index.html"""'], {}), "('index.html')\n", (1163, 1177), False, 'from tarfile import TarInfo\n'), ((1179, 1201), 'tarfile.TarInfo', 'TarInfo', (['"""foo/bar.xyz"""'], {}), "('foo/bar.xyz')\n", (1186, 1201), False, 'from tarfile import TarInfo\n'), ((2001, 2010), 'tempfile.mkdtemp', 'mkdtemp', ([], {}), '()\n', (2008, 2010), False, 'from tempfile import mkdtemp\n'), ((2812, 2823), 'time.time', 'time.time', ([], {}), '()\n', (2821, 2823), False, 'import time\n'), ((1471, 1521), 'markdownreveal.local.latest_project_release', 'latest_project_release', ([], {'github': '"""hakimel/reveal.js"""'}), "(github='hakimel/reveal.js')\n", (1493, 1521), False, 'from markdownreveal.local import latest_project_release\n'), ((1535, 1578), 'markdownreveal.local.latest_project_release', 'latest_project_release', ([], {'github': '"""khan/katex"""'}), "(github='khan/katex')\n", (1557, 1578), False, 'from markdownreveal.local import latest_project_release\n')] |
from models.MultiGMPmodel import MultiCopyGMPmodel
from models.MultiGGHPmodel import MultiCopyGGHPmodel
from util.calculatedCRBrateAndEstimationAccuracy import calculatedCRBrateAndEstimationAccuracy
from util.cutCircularChromosomes import cutCircularChromosomes
"""
Inferring ancestor species for Gramineae species.
Ancestor 4: Multi-copy GGHP model, result in outdutdata/Gramineae/Ancestor4
Ancestor 3: Multi-copy GMP model, result in outdutdata/Gramineae/Ancestor3
Ancestor 2: Multi-copy GMP model, result in outdutdata/Gramineae/Ancestor2
Ancestor 1: Multi-copy GMP model, result in outdutdata/Gramineae/Ancestor1
"""
path = 'D:/InferAncestorGenome/realData'
workdir = path + '/IAGS_version1.0/inputdata/Gramineae/'
"""
Inferring ancestor species for Gramineae species.
Ancestor 4 using Multi-copy GGHP model
"""
dup_child_file = workdir + 'Zmays.final.block'
outgroup_file = workdir + 'Sbicolor.final.block'
outAncestor4dir = path + '/IAGS_version1.0/outputdata/Gramineae/Ancestor4/'
dup_copy_number = 4
out_copy_number = 2
ancestor_target_copy_number = 2
ancestor_name = 'Ancestor4'
MultiCopyGGHPmodel(dup_child_file, outgroup_file, outAncestor4dir,
ancestor_name, dup_copy_number, out_copy_number, ancestor_target_copy_number)
# speciesAndCopyList = [
# [workdir + 'Zmays.final.block',dup_copy_number,'Z.mays'],
# [workdir + 'Sbicolor.final.block',out_copy_number,'S.bicolor']
# ]
# cutCircularChromosomes(outAncestor4dir + 'Ancestor4.block',
# ancestor_target_copy_number,
# ancestor_name,speciesAndCopyList,outAncestor4dir)
# Evaluation
matching_target_file = workdir + 'Sbicolor.final.block'
matching_target_copy_number = out_copy_number
matching_target_name = 'S.bicolor'
speciesAndCopyList = [
[workdir + 'Zmays.final.block',dup_copy_number,'Z.mays'],
[workdir + 'Sbicolor.final.block',out_copy_number,'S.bicolor']
]
model_type = 'MultiCopyGGHP'
calculatedCRBrateAndEstimationAccuracy(matching_target_file, matching_target_copy_number, matching_target_name,
speciesAndCopyList, outAncestor4dir, model_type)
"""
Inferring ancestor species for Gramineae species.
Ancestor 3 using Multi-copy GMP model
"""
species_file_list = [workdir + 'Sbicolor.final.block',
outAncestor4dir + 'Ancestor4.block',
workdir + 'Osativa.final.block']
guided_species_for_matching = workdir + 'Sbicolor.final.block'
ancestor_target_copy_number = 2
outAncestor3dir = path + '/IAGS_version1.0/outputdata/Gramineae/Ancestor3/'
ancestor_name = 'Ancestor3'
MultiCopyGMPmodel(species_file_list, outAncestor3dir, guided_species_for_matching, ancestor_name, ancestor_target_copy_number)
# Evaluation
matching_target_file = workdir + 'Sbicolor.final.block'
matching_target_copy_number = ancestor_target_copy_number
matching_target_name = 'S.bicolor'
speciesAndCopyList = [
[workdir + 'Sbicolor.final.block',ancestor_target_copy_number,'S.bicolor'],
[outAncestor4dir + 'Ancestor4.block',ancestor_target_copy_number,'Ancestor4'],
[workdir + 'Osativa.final.block',ancestor_target_copy_number,'O.sativa']
]
model_type = 'MultiCopyGMP'
calculatedCRBrateAndEstimationAccuracy(matching_target_file, matching_target_copy_number, matching_target_name,
speciesAndCopyList, outAncestor3dir, model_type)
"""
Inferring ancestor species for Gramineae species.
Ancestor 2 using Multi-copy GMP model
"""
species_file_list = [workdir + 'Bdistachyon.final.block',
workdir + 'Telongatum.final.block',
workdir + 'Osativa.final.block']
guided_species_for_matching = workdir + 'Bdistachyon.final.block'
ancestor_target_copy_number = 2
outAncestor2dir = path + '/IAGS_version1.0/outputdata/Gramineae/Ancestor2/'
ancestor_name = 'Ancestor2'
MultiCopyGMPmodel(species_file_list, outAncestor2dir, guided_species_for_matching, ancestor_name, ancestor_target_copy_number)
# Evaluation
matching_target_file = workdir + 'Bdistachyon.final.block'
matching_target_copy_number = ancestor_target_copy_number
matching_target_name = 'B.distachyon'
speciesAndCopyList = [
[workdir + 'Bdistachyon.final.block',ancestor_target_copy_number,'B.distachyon'],
[workdir + 'Telongatum.final.block',ancestor_target_copy_number,'T.elongatum'],
[workdir + 'Osativa.final.block',ancestor_target_copy_number,'O.sativa']
]
model_type = 'MultiCopyGMP'
calculatedCRBrateAndEstimationAccuracy(matching_target_file, matching_target_copy_number, matching_target_name,
speciesAndCopyList, outAncestor2dir, model_type)
"""
Inferring ancestor species for Gramineae species.
Ancestor 1 using Multi-copy GMP model
"""
species_file_list = [workdir + 'Osativa.final.block',
outAncestor2dir + 'Ancestor2.block',
outAncestor3dir + 'Ancestor3.block']
guided_species_for_matching = workdir + 'Osativa.final.block'
ancestor_target_copy_number = 2
outAncestor1dir = path + '/IAGS_version1.0/outputdata/Gramineae/Ancestor1/'
ancestor_name = 'Ancestor1'
MultiCopyGMPmodel(species_file_list, outAncestor1dir, guided_species_for_matching, ancestor_name, ancestor_target_copy_number)
# Evaluation
matching_target_file = workdir + 'Osativa.final.block'
matching_target_copy_number = ancestor_target_copy_number
matching_target_name = 'O.sativa'
speciesAndCopyList = [
[workdir + 'Osativa.final.block',ancestor_target_copy_number,'O.sativa'],
[outAncestor2dir + 'Ancestor2.block',ancestor_target_copy_number,'Ancestor2'],
[outAncestor3dir + 'Ancestor3.block',ancestor_target_copy_number,'Ancestor3']
]
model_type = 'MultiCopyGMP'
calculatedCRBrateAndEstimationAccuracy(matching_target_file, matching_target_copy_number, matching_target_name,
speciesAndCopyList, outAncestor1dir, model_type) | [
"util.calculatedCRBrateAndEstimationAccuracy.calculatedCRBrateAndEstimationAccuracy",
"models.MultiGMPmodel.MultiCopyGMPmodel",
"models.MultiGGHPmodel.MultiCopyGGHPmodel"
] | [((1094, 1246), 'models.MultiGGHPmodel.MultiCopyGGHPmodel', 'MultiCopyGGHPmodel', (['dup_child_file', 'outgroup_file', 'outAncestor4dir', 'ancestor_name', 'dup_copy_number', 'out_copy_number', 'ancestor_target_copy_number'], {}), '(dup_child_file, outgroup_file, outAncestor4dir,\n ancestor_name, dup_copy_number, out_copy_number,\n ancestor_target_copy_number)\n', (1112, 1246), False, 'from models.MultiGGHPmodel import MultiCopyGGHPmodel\n'), ((1949, 2117), 'util.calculatedCRBrateAndEstimationAccuracy.calculatedCRBrateAndEstimationAccuracy', 'calculatedCRBrateAndEstimationAccuracy', (['matching_target_file', 'matching_target_copy_number', 'matching_target_name', 'speciesAndCopyList', 'outAncestor4dir', 'model_type'], {}), '(matching_target_file,\n matching_target_copy_number, matching_target_name, speciesAndCopyList,\n outAncestor4dir, model_type)\n', (1987, 2117), False, 'from util.calculatedCRBrateAndEstimationAccuracy import calculatedCRBrateAndEstimationAccuracy\n'), ((2613, 2743), 'models.MultiGMPmodel.MultiCopyGMPmodel', 'MultiCopyGMPmodel', (['species_file_list', 'outAncestor3dir', 'guided_species_for_matching', 'ancestor_name', 'ancestor_target_copy_number'], {}), '(species_file_list, outAncestor3dir,\n guided_species_for_matching, ancestor_name, ancestor_target_copy_number)\n', (2630, 2743), False, 'from models.MultiGMPmodel import MultiCopyGMPmodel\n'), ((3197, 3365), 'util.calculatedCRBrateAndEstimationAccuracy.calculatedCRBrateAndEstimationAccuracy', 'calculatedCRBrateAndEstimationAccuracy', (['matching_target_file', 'matching_target_copy_number', 'matching_target_name', 'speciesAndCopyList', 'outAncestor3dir', 'model_type'], {}), '(matching_target_file,\n matching_target_copy_number, matching_target_name, speciesAndCopyList,\n outAncestor3dir, model_type)\n', (3235, 3365), False, 'from util.calculatedCRBrateAndEstimationAccuracy import calculatedCRBrateAndEstimationAccuracy\n'), ((3866, 3996), 'models.MultiGMPmodel.MultiCopyGMPmodel', 'MultiCopyGMPmodel', (['species_file_list', 'outAncestor2dir', 'guided_species_for_matching', 'ancestor_name', 'ancestor_target_copy_number'], {}), '(species_file_list, outAncestor2dir,\n guided_species_for_matching, ancestor_name, ancestor_target_copy_number)\n', (3883, 3996), False, 'from models.MultiGMPmodel import MultiCopyGMPmodel\n'), ((4463, 4631), 'util.calculatedCRBrateAndEstimationAccuracy.calculatedCRBrateAndEstimationAccuracy', 'calculatedCRBrateAndEstimationAccuracy', (['matching_target_file', 'matching_target_copy_number', 'matching_target_name', 'speciesAndCopyList', 'outAncestor2dir', 'model_type'], {}), '(matching_target_file,\n matching_target_copy_number, matching_target_name, speciesAndCopyList,\n outAncestor2dir, model_type)\n', (4501, 4631), False, 'from util.calculatedCRBrateAndEstimationAccuracy import calculatedCRBrateAndEstimationAccuracy\n'), ((5129, 5259), 'models.MultiGMPmodel.MultiCopyGMPmodel', 'MultiCopyGMPmodel', (['species_file_list', 'outAncestor1dir', 'guided_species_for_matching', 'ancestor_name', 'ancestor_target_copy_number'], {}), '(species_file_list, outAncestor1dir,\n guided_species_for_matching, ancestor_name, ancestor_target_copy_number)\n', (5146, 5259), False, 'from models.MultiGMPmodel import MultiCopyGMPmodel\n'), ((5714, 5882), 'util.calculatedCRBrateAndEstimationAccuracy.calculatedCRBrateAndEstimationAccuracy', 'calculatedCRBrateAndEstimationAccuracy', (['matching_target_file', 'matching_target_copy_number', 'matching_target_name', 'speciesAndCopyList', 'outAncestor1dir', 'model_type'], {}), '(matching_target_file,\n matching_target_copy_number, matching_target_name, speciesAndCopyList,\n outAncestor1dir, model_type)\n', (5752, 5882), False, 'from util.calculatedCRBrateAndEstimationAccuracy import calculatedCRBrateAndEstimationAccuracy\n')] |
from django.conf.urls import url
import views
urlpatterns = [
url(r'^user/', views.get_user, name='user'),
url(r'^all/', views.get_all, name='all'),
url(r'^fix/$', views.get_fix, name='fix'),
]
| [
"django.conf.urls.url"
] | [((67, 109), 'django.conf.urls.url', 'url', (['"""^user/"""', 'views.get_user'], {'name': '"""user"""'}), "('^user/', views.get_user, name='user')\n", (70, 109), False, 'from django.conf.urls import url\n'), ((116, 155), 'django.conf.urls.url', 'url', (['"""^all/"""', 'views.get_all'], {'name': '"""all"""'}), "('^all/', views.get_all, name='all')\n", (119, 155), False, 'from django.conf.urls import url\n'), ((162, 202), 'django.conf.urls.url', 'url', (['"""^fix/$"""', 'views.get_fix'], {'name': '"""fix"""'}), "('^fix/$', views.get_fix, name='fix')\n", (165, 202), False, 'from django.conf.urls import url\n')] |
import os
import argparse
import json
import tqdm
import numpy as np
from program_synthesis.datasets import dataset, executor
parser = argparse.ArgumentParser()
parser.add_argument("--data-pickle", default='data/karel/val.pkl')
parser.add_argument("--input-file", required=True, help="file containing list of beams, each of which is a list of programs")
parser.add_argument("--output-file", required=True)
args = parser.parse_args()
assert not os.path.exists(args.output_file)
with open(args.input_file) as f:
programs = json.load(f)
examples = dataset.KarelTorchDataset(
args.data_pickle,
lambda x: x)
def evaluate_code(eg, beam):
exe = executor.KarelExecutor()
tests = []
tests += list(eg.input_tests)
tests += list(eg.tests)
stats = executor.evaluate_code(beam[0], eg.schema.args, tests, exe.execute)
prediction = dict(
output=beam[0],
beams=beam,
beams_correct=[executor.evaluate_code(hypothesis, eg.schema.args, tests, exe.execute) for hypothesis in beam],
is_correct=stats['correct'] == stats['total'],
individual=stats['individual'],
passes_given_tests=all(stats['individual'][:len(eg.input_tests)])
)
return prediction
result = [evaluate_code(eg, beams) for eg, beams in zip(examples, tqdm.tqdm(programs))]
with open(args.output_file, "w") as f:
json.dump(result, f) | [
"os.path.exists",
"program_synthesis.datasets.executor.KarelExecutor",
"argparse.ArgumentParser",
"program_synthesis.datasets.dataset.KarelTorchDataset",
"tqdm.tqdm",
"program_synthesis.datasets.executor.evaluate_code",
"json.load",
"json.dump"
] | [((140, 165), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {}), '()\n', (163, 165), False, 'import argparse\n'), ((558, 614), 'program_synthesis.datasets.dataset.KarelTorchDataset', 'dataset.KarelTorchDataset', (['args.data_pickle', '(lambda x: x)'], {}), '(args.data_pickle, lambda x: x)\n', (583, 614), False, 'from program_synthesis.datasets import dataset, executor\n'), ((451, 483), 'os.path.exists', 'os.path.exists', (['args.output_file'], {}), '(args.output_file)\n', (465, 483), False, 'import os\n'), ((533, 545), 'json.load', 'json.load', (['f'], {}), '(f)\n', (542, 545), False, 'import json\n'), ((665, 689), 'program_synthesis.datasets.executor.KarelExecutor', 'executor.KarelExecutor', ([], {}), '()\n', (687, 689), False, 'from program_synthesis.datasets import dataset, executor\n'), ((780, 847), 'program_synthesis.datasets.executor.evaluate_code', 'executor.evaluate_code', (['beam[0]', 'eg.schema.args', 'tests', 'exe.execute'], {}), '(beam[0], eg.schema.args, tests, exe.execute)\n', (802, 847), False, 'from program_synthesis.datasets import dataset, executor\n'), ((1365, 1385), 'json.dump', 'json.dump', (['result', 'f'], {}), '(result, f)\n', (1374, 1385), False, 'import json\n'), ((1299, 1318), 'tqdm.tqdm', 'tqdm.tqdm', (['programs'], {}), '(programs)\n', (1308, 1318), False, 'import tqdm\n'), ((938, 1008), 'program_synthesis.datasets.executor.evaluate_code', 'executor.evaluate_code', (['hypothesis', 'eg.schema.args', 'tests', 'exe.execute'], {}), '(hypothesis, eg.schema.args, tests, exe.execute)\n', (960, 1008), False, 'from program_synthesis.datasets import dataset, executor\n')] |
# Copyright 2013 The Chromium Authors. All rights reserved.
# Use of this source code is governed by a BSD-style license that can be
# found in the LICENSE file.
from contrib.cluster_telemetry import ct_benchmarks_util
from contrib.cluster_telemetry import page_set
from contrib.cluster_telemetry import repaint_helpers
from benchmarks import rasterize_and_record_micro
# pylint: disable=protected-access
class RasterizeAndRecordMicroCT(
rasterize_and_record_micro._RasterizeAndRecordMicro):
"""Measures rasterize and record performance for Cluster Telemetry."""
@classmethod
def Name(cls):
return 'rasterize_and_record_micro_ct'
@classmethod
def AddBenchmarkCommandLineArgs(cls, parser):
(rasterize_and_record_micro._RasterizeAndRecordMicro.
AddBenchmarkCommandLineArgs(parser))
ct_benchmarks_util.AddBenchmarkCommandLineArgs(parser)
@classmethod
def ProcessCommandLineArgs(cls, parser, args):
ct_benchmarks_util.ValidateCommandLineArgs(parser, args)
def CreateStorySet(self, options):
return page_set.CTPageSet(
options.urls_list, options.user_agent, options.archive_data_file,
run_page_interaction_callback=repaint_helpers.WaitThenRepaint)
| [
"contrib.cluster_telemetry.ct_benchmarks_util.ValidateCommandLineArgs",
"contrib.cluster_telemetry.page_set.CTPageSet",
"benchmarks.rasterize_and_record_micro._RasterizeAndRecordMicro.AddBenchmarkCommandLineArgs",
"contrib.cluster_telemetry.ct_benchmarks_util.AddBenchmarkCommandLineArgs"
] | [((718, 810), 'benchmarks.rasterize_and_record_micro._RasterizeAndRecordMicro.AddBenchmarkCommandLineArgs', 'rasterize_and_record_micro._RasterizeAndRecordMicro.AddBenchmarkCommandLineArgs', (['parser'], {}), '(\n parser)\n', (797, 810), False, 'from benchmarks import rasterize_and_record_micro\n'), ((820, 874), 'contrib.cluster_telemetry.ct_benchmarks_util.AddBenchmarkCommandLineArgs', 'ct_benchmarks_util.AddBenchmarkCommandLineArgs', (['parser'], {}), '(parser)\n', (866, 874), False, 'from contrib.cluster_telemetry import ct_benchmarks_util\n'), ((944, 1000), 'contrib.cluster_telemetry.ct_benchmarks_util.ValidateCommandLineArgs', 'ct_benchmarks_util.ValidateCommandLineArgs', (['parser', 'args'], {}), '(parser, args)\n', (986, 1000), False, 'from contrib.cluster_telemetry import ct_benchmarks_util\n'), ((1050, 1207), 'contrib.cluster_telemetry.page_set.CTPageSet', 'page_set.CTPageSet', (['options.urls_list', 'options.user_agent', 'options.archive_data_file'], {'run_page_interaction_callback': 'repaint_helpers.WaitThenRepaint'}), '(options.urls_list, options.user_agent, options.\n archive_data_file, run_page_interaction_callback=repaint_helpers.\n WaitThenRepaint)\n', (1068, 1207), False, 'from contrib.cluster_telemetry import page_set\n')] |
import queue
ROOM_ENTERED = 'roomEntered'
class Event:
def __init__(self, eventType='', userParam=dict()):
self.type = eventType
self.userParam = userParam
class EventSystem:
def __init__(self):
self._eventQueue = queue.Queue()
self._eventHandlers = dict()
def registerEventHander(self, eventType, callback):
''' Register a handler to be called on the given event type.
eventType specifies the type of event the handler should process.
callback specifies the function that should be called on the event.
Its function header should look like "def myCallback(event):"
Returns the ID of the handler.
'''
if not eventType in self._eventHandlers:
self._eventHandlers[eventType] = []
handlerID = len(self._eventHandlers[eventType])
self._eventHandlers[eventType].append(callback)
return handlerID
def unregisterEventHandler(self, eventType, handlerID):
''' Unregister a handler, so it won't be called on the specified event.
eventType specifies the type of event the handler should process.
handlerID specifies the ID of the handler, which should be unregistered.
The ID was returned by the corresponding register-function.
Returns True on success, else False.
'''
if not eventType in self._eventHandlers:
return False
if handlerID >= len(self._eventHandlers[eventType]):
return False
self._eventHandlers[eventType].pop(handlerID)
return True
def createEvent(self, event):
self._eventQueue.put_nowait(event)
def processEvents(self):
while not self._eventQueue.empty():
event = self._eventQueue.get_nowait()
# check if eventhandler wants to process event
if not event.type in self._eventHandlers:
continue
for cb in self._eventHandlers[event.type]:
cb(event)
| [
"queue.Queue"
] | [((251, 264), 'queue.Queue', 'queue.Queue', ([], {}), '()\n', (262, 264), False, 'import queue\n')] |
import os
import FWCore.ParameterSet.Config as cms
process = cms.Process('TauDQMOffline')
process.load("FWCore.MessageLogger.MessageLogger_cfi")
process.MessageLogger.cerr.FwkReport.reportEvery = 1000
# import of standard configurations
from Configuration.StandardSequences.GeometryRecoDB_cff import *
process.load('Configuration.StandardSequences.GeometryRecoDB_cff')
process.load('Configuration.StandardSequences.MagneticField_cff')
process.load('Configuration.StandardSequences.Reconstruction_cff')
process.load('Configuration.StandardSequences.FrontierConditions_GlobalTag_cff')
process.load('Configuration.StandardSequences.Services_cff')
process.load('Configuration.EventContent.EventContent_cff')
process.GlobalTag.globaltag = '94X_dataRun2_ReReco_EOY17_v6'
#process.load("DQMServices.Components.DQMStoreStats_cfi")
#process.load('DQMOffline.Configuration.DQMOffline_cff')
readFiles = cms.untracked.vstring()
secFiles = cms.untracked.vstring()
process.source = cms.Source ("PoolSource",fileNames = readFiles, secondaryFileNames = secFiles)
readFiles.extend( [
'/store/data/Run2017D/Tau/MINIAOD/31Mar2018-v1/00000/02FE19AF-3837-E811-B3FF-44A842B4520B.root'
] );
process.maxEvents = cms.untracked.PSet(
input = cms.untracked.int32( -1 )
)
process.load('Validation.RecoTau.RecoTauValidation_cff')
#process.load('Validation.RecoTau.DQMSequences_cfi')
#process.load('Validation.RecoTau.RecoTauValidationMiniAOD_cfi')
process.dump = cms.EDAnalyzer("EventContentAnalyzer")
process.DQMoutput = cms.OutputModule("DQMRootOutputModule",
dataset = cms.untracked.PSet(
dataTier = cms.untracked.string('DQMIO'),
filterName = cms.untracked.string('')
),
fileName = cms.untracked.string('RECO_RAW2DIGI_L1Reco_RECO_EI_PAT_DQM_inDQM.root'),
outputCommands = process.DQMEventContent.outputCommands,
splitLevel = cms.untracked.int32(0)
)
#Setup FWK for multithreaded
process.options.numberOfThreads=cms.untracked.uint32(8)
process.options.numberOfStreams=cms.untracked.uint32(0)
#Have logErrorHarvester wait for the same EDProducers to finish as those providing data for the OutputModule
from FWCore.Modules.logErrorHarvester_cff import customiseLogErrorHarvesterUsingOutputCommands
process = customiseLogErrorHarvesterUsingOutputCommands(process)
| [
"FWCore.ParameterSet.Config.untracked.string",
"FWCore.Modules.logErrorHarvester_cff.customiseLogErrorHarvesterUsingOutputCommands",
"FWCore.ParameterSet.Config.Source",
"FWCore.ParameterSet.Config.untracked.int32",
"FWCore.ParameterSet.Config.untracked.uint32",
"FWCore.ParameterSet.Config.Process",
"FW... | [((63, 91), 'FWCore.ParameterSet.Config.Process', 'cms.Process', (['"""TauDQMOffline"""'], {}), "('TauDQMOffline')\n", (74, 91), True, 'import FWCore.ParameterSet.Config as cms\n'), ((902, 925), 'FWCore.ParameterSet.Config.untracked.vstring', 'cms.untracked.vstring', ([], {}), '()\n', (923, 925), True, 'import FWCore.ParameterSet.Config as cms\n'), ((937, 960), 'FWCore.ParameterSet.Config.untracked.vstring', 'cms.untracked.vstring', ([], {}), '()\n', (958, 960), True, 'import FWCore.ParameterSet.Config as cms\n'), ((979, 1053), 'FWCore.ParameterSet.Config.Source', 'cms.Source', (['"""PoolSource"""'], {'fileNames': 'readFiles', 'secondaryFileNames': 'secFiles'}), "('PoolSource', fileNames=readFiles, secondaryFileNames=secFiles)\n", (989, 1053), True, 'import FWCore.ParameterSet.Config as cms\n'), ((1461, 1499), 'FWCore.ParameterSet.Config.EDAnalyzer', 'cms.EDAnalyzer', (['"""EventContentAnalyzer"""'], {}), "('EventContentAnalyzer')\n", (1475, 1499), True, 'import FWCore.ParameterSet.Config as cms\n'), ((1953, 1976), 'FWCore.ParameterSet.Config.untracked.uint32', 'cms.untracked.uint32', (['(8)'], {}), '(8)\n', (1973, 1976), True, 'import FWCore.ParameterSet.Config as cms\n'), ((2009, 2032), 'FWCore.ParameterSet.Config.untracked.uint32', 'cms.untracked.uint32', (['(0)'], {}), '(0)\n', (2029, 2032), True, 'import FWCore.ParameterSet.Config as cms\n'), ((2248, 2302), 'FWCore.Modules.logErrorHarvester_cff.customiseLogErrorHarvesterUsingOutputCommands', 'customiseLogErrorHarvesterUsingOutputCommands', (['process'], {}), '(process)\n', (2293, 2302), False, 'from FWCore.Modules.logErrorHarvester_cff import customiseLogErrorHarvesterUsingOutputCommands\n'), ((1241, 1264), 'FWCore.ParameterSet.Config.untracked.int32', 'cms.untracked.int32', (['(-1)'], {}), '(-1)\n', (1260, 1264), True, 'import FWCore.ParameterSet.Config as cms\n'), ((1714, 1785), 'FWCore.ParameterSet.Config.untracked.string', 'cms.untracked.string', (['"""RECO_RAW2DIGI_L1Reco_RECO_EI_PAT_DQM_inDQM.root"""'], {}), "('RECO_RAW2DIGI_L1Reco_RECO_EI_PAT_DQM_inDQM.root')\n", (1734, 1785), True, 'import FWCore.ParameterSet.Config as cms\n'), ((1865, 1887), 'FWCore.ParameterSet.Config.untracked.int32', 'cms.untracked.int32', (['(0)'], {}), '(0)\n', (1884, 1887), True, 'import FWCore.ParameterSet.Config as cms\n'), ((1615, 1644), 'FWCore.ParameterSet.Config.untracked.string', 'cms.untracked.string', (['"""DQMIO"""'], {}), "('DQMIO')\n", (1635, 1644), True, 'import FWCore.ParameterSet.Config as cms\n'), ((1667, 1691), 'FWCore.ParameterSet.Config.untracked.string', 'cms.untracked.string', (['""""""'], {}), "('')\n", (1687, 1691), True, 'import FWCore.ParameterSet.Config as cms\n')] |
import os
import numpy as np
import tensorflow as tf
import tensorflow_addons as tfa
from loguru import logger
import config
from Train import train
from Model import EEGNet
class OptunaTrainer:
def __init__(self, checkpointPath, epochs, batchsize, logPath=None):
self.checkpointPath = checkpointPath
self.logpath = logPath
self.epochs = epochs
self.batchsize = batchsize
def __call__(self, trial, dataset, crossVal=False, **kwargs):
if isinstance(dataset, tuple):
dataset = {
"noname": dataset
}
info = "Trial #{} metric values:\n".format(trial.number)
metrics = []
for key, value in dataset.items():
if "augmenter" in kwargs:
kwargs["augmenter"].setState(False)
shape = list(value[0].shape[-2:])
shape[1] = int(config.window[1] * config.sampleRate) - int(config.window[0] * config.sampleRate)
model = self.buildModel(trial, shape)
auc, precision = train(
model=model,
dataset=value,
weightsPath=self.checkpointPath,
epochs=self.epochs,
batchsize=self.batchsize,
crossVal=crossVal,
**kwargs
)
info += "{}: auc {:.2f} pr {:.2f}\t".format(key, auc, precision)
metrics.append((auc, precision))
metrics = np.array(metrics)
mean = np.mean(metrics, axis=0).round(2)
median = np.median(metrics, axis=0).round(2)
for i, metric in enumerate(["auc", "precision"]):
info += "\nMetric - {}. Mean: {}\tMedian: {}".format(metric, mean[i], median[i])
logger.info(info)
logger.info(trial.params)
return mean[0]
@staticmethod
def chooseOptimizer(trial):
kwargs = {}
optimizer_options = ["RMSprop", "Adam"]
optimizer_selected = trial.suggest_categorical("optimizer", optimizer_options)
if optimizer_selected == "RMSprop":
kwargs["learning_rate"] = trial.suggest_loguniform("rmsprop_learning_rate", 1e-5, 1e-1)
kwargs["decay"] = trial.suggest_discrete_uniform("rmsprop_decay", 0.85, 0.99, 0.01)
kwargs["momentum"] = trial.suggest_loguniform("rmsprop_momentum", 1e-5, 1e-1)
elif optimizer_selected == "Adam":
kwargs["learning_rate"] = trial.suggest_loguniform("adam_learning_rate", 1e-5, 1e-1)
optimizer = getattr(tf.optimizers, optimizer_selected)(**kwargs)
return optimizer
@staticmethod
def chooseLoss(trial):
loss_functions = {
"binary_crossentropy": tf,
"sigmoid_focal_crossentropy": tfa
}
loss_selected = trial.suggest_categorical("loss", list(loss_functions.keys()))
loss = getattr(loss_functions[loss_selected].losses, loss_selected)
return loss
def buildModel(self, trial, shape):
samples = shape[-1]
assert samples // 2 > 16
temporalLength = int(trial.suggest_discrete_uniform("temporal_length", 16, samples // 2, 4))
dropoutRate = trial.suggest_discrete_uniform("dropout_rate", 0.1, 0.5, 0.05)
D = trial.suggest_int("depth_multiplier", 1, 4)
poolKernel = int(trial.suggest_discrete_uniform("pool_kernel", 4, 16, 2))
model = EEGNet(
categoriesN=2,
electrodes=shape[0],
samples=shape[1],
temporalLength=temporalLength,
dropoutRate=dropoutRate,
D=D,
poolPad="same",
poolKernel=poolKernel
)
optimizer = self.chooseOptimizer(trial)
loss = self.chooseLoss(trial)
model.compile(
loss=loss,
optimizer=optimizer,
metrics=["accuracy"]
)
return model
def studyInfo(study, bestN=7, file=None):
logger.info("Number of finished trials: {}", len(study.trials))
logger.info("Best {} trials:".format(bestN))
trials = sorted(study.trials, key=lambda elem: elem.value, reverse=True)[:bestN]
for i, trial in enumerate(trials):
logger.info("Trial {}", i)
logger.info("\tValue: {:.2f}", trial.value)
logger.info("\tParams: ")
for key, value in trial.params.items():
logger.info("\t\t{}: {}", key, value)
if file is not None:
os.makedirs(os.path.dirname(file), exist_ok=True)
studyDF = study.trials_dataframe()
studyDF.to_csv(file)
logger.info("Study file has been written to {}", file) | [
"numpy.mean",
"numpy.median",
"loguru.logger.info",
"Train.train",
"os.path.dirname",
"numpy.array",
"Model.EEGNet"
] | [((1200, 1217), 'numpy.array', 'np.array', (['metrics'], {}), '(metrics)\n', (1208, 1217), True, 'import numpy as np\n'), ((1449, 1466), 'loguru.logger.info', 'logger.info', (['info'], {}), '(info)\n', (1460, 1466), False, 'from loguru import logger\n'), ((1469, 1494), 'loguru.logger.info', 'logger.info', (['trial.params'], {}), '(trial.params)\n', (1480, 1494), False, 'from loguru import logger\n'), ((2908, 3077), 'Model.EEGNet', 'EEGNet', ([], {'categoriesN': '(2)', 'electrodes': 'shape[0]', 'samples': 'shape[1]', 'temporalLength': 'temporalLength', 'dropoutRate': 'dropoutRate', 'D': 'D', 'poolPad': '"""same"""', 'poolKernel': 'poolKernel'}), "(categoriesN=2, electrodes=shape[0], samples=shape[1], temporalLength\n =temporalLength, dropoutRate=dropoutRate, D=D, poolPad='same',\n poolKernel=poolKernel)\n", (2914, 3077), False, 'from Model import EEGNet\n'), ((3549, 3575), 'loguru.logger.info', 'logger.info', (['"""Trial {}"""', 'i'], {}), "('Trial {}', i)\n", (3560, 3575), False, 'from loguru import logger\n'), ((3578, 3621), 'loguru.logger.info', 'logger.info', (['"""\tValue: {:.2f}"""', 'trial.value'], {}), "('\\tValue: {:.2f}', trial.value)\n", (3589, 3621), False, 'from loguru import logger\n'), ((3625, 3650), 'loguru.logger.info', 'logger.info', (['"""\tParams: """'], {}), "('\\tParams: ')\n", (3636, 3650), False, 'from loguru import logger\n'), ((3872, 3926), 'loguru.logger.info', 'logger.info', (['"""Study file has been written to {}"""', 'file'], {}), "('Study file has been written to {}', file)\n", (3883, 3926), False, 'from loguru import logger\n'), ((907, 1053), 'Train.train', 'train', ([], {'model': 'model', 'dataset': 'value', 'weightsPath': 'self.checkpointPath', 'epochs': 'self.epochs', 'batchsize': 'self.batchsize', 'crossVal': 'crossVal'}), '(model=model, dataset=value, weightsPath=self.checkpointPath, epochs=\n self.epochs, batchsize=self.batchsize, crossVal=crossVal, **kwargs)\n', (912, 1053), False, 'from Train import train\n'), ((3696, 3733), 'loguru.logger.info', 'logger.info', (['"""\t\t{}: {}"""', 'key', 'value'], {}), "('\\t\\t{}: {}', key, value)\n", (3707, 3733), False, 'from loguru import logger\n'), ((3771, 3792), 'os.path.dirname', 'os.path.dirname', (['file'], {}), '(file)\n', (3786, 3792), False, 'import os\n'), ((1228, 1252), 'numpy.mean', 'np.mean', (['metrics'], {'axis': '(0)'}), '(metrics, axis=0)\n', (1235, 1252), True, 'import numpy as np\n'), ((1273, 1299), 'numpy.median', 'np.median', (['metrics'], {'axis': '(0)'}), '(metrics, axis=0)\n', (1282, 1299), True, 'import numpy as np\n')] |
from PyQt5 import QtGui, QtWidgets, QtCore
class RoundPixmapStyle(QtWidgets.QProxyStyle):
def __init__(self, radius=10, *args, **kwargs):
super(RoundPixmapStyle, self).__init__(*args, **kwargs)
self._radius = radius
def drawItemPixmap(self, painter, rectangle, alignment, pixmap):
painter.save()
pix = QtGui.QPixmap(pixmap.size())
pix.fill(QtCore.Qt.transparent)
p = QtGui.QPainter(pix)
p.setBrush(QtGui.QBrush(pixmap))
p.setPen(QtCore.Qt.NoPen)
p.drawRoundedRect(pixmap.rect(), self._radius, self._radius)
p.end()
super(RoundPixmapStyle, self).drawItemPixmap(painter, rectangle, alignment, pix)
painter.restore() | [
"PyQt5.QtGui.QBrush",
"PyQt5.QtGui.QPainter"
] | [((426, 445), 'PyQt5.QtGui.QPainter', 'QtGui.QPainter', (['pix'], {}), '(pix)\n', (440, 445), False, 'from PyQt5 import QtGui, QtWidgets, QtCore\n'), ((465, 485), 'PyQt5.QtGui.QBrush', 'QtGui.QBrush', (['pixmap'], {}), '(pixmap)\n', (477, 485), False, 'from PyQt5 import QtGui, QtWidgets, QtCore\n')] |
from qcodes import Instrument
try:
from spirack import D4_module
except ImportError:
raise ImportError(('The D4_module class could not be found. '
'Try installing it using pip install spirack'))
from functools import partial
class D4(Instrument):
"""
Qcodes driver for the D4 ADC SPI-rack module. Requires installation
of the 'spirack' package using pip.
Args:
name (str): name of the instrument.
spi_rack (SPI_rack): instance of the SPI_rack class as defined in
the spirack package. This class manages communication with the
individual modules.
module (int): module number as set on the hardware.
"""
def __init__(self, name, spi_rack, module, **kwargs):
super().__init__(name, **kwargs)
self.d4 = D4_module(spi_rack, module)
self.add_parameter('mode',
label='Mode',
get_cmd=self.get_mode)
self.add_parameter('filter_value',
label='Filter value',
get_cmd=self.get_filter_value)
self.add_parameter('buffers_enabled',
label='Buffers enabled',
get_cmd=self.get_buffers_enabled)
for i in range(2):
self.add_parameter('adc{}'.format(i + 1),
label='ADC {}'.format(i + 1),
get_cmd=partial(self.d4.singleConversion, i),
units='V')
def get_mode(self):
return self.d4.mode
def get_filter_value(self):
return self.d4.filter_val
def get_buffers_enabled(self):
return self.d4.buf_en
| [
"functools.partial",
"spirack.D4_module"
] | [((827, 854), 'spirack.D4_module', 'D4_module', (['spi_rack', 'module'], {}), '(spi_rack, module)\n', (836, 854), False, 'from spirack import D4_module\n'), ((1475, 1511), 'functools.partial', 'partial', (['self.d4.singleConversion', 'i'], {}), '(self.d4.singleConversion, i)\n', (1482, 1511), False, 'from functools import partial\n')] |
from typing import Optional
import yaml
from returns.result import Result, Success, Failure
from typistry.protos.invalid_object import InvalidObject, IgnorableObject
from typistry.protos.typed_dict import TypedDict
def safe_parse_yaml(file: str, include_file: bool = False, filter_type: Optional[str] = None) -> Result[TypedDict, InvalidObject]:
try:
with open(file, 'r') as stream:
try:
yaml_load = yaml.safe_load(stream)
if isinstance(yaml_load, dict):
to_type = yaml_load.get("type")
if isinstance(to_type, str):
if (filter_type == None) or (to_type == filter_type):
yaml_load.pop("type")
if include_file:
yaml_load["source"] = file
return Success(TypedDict(yaml_load, type=to_type))
else:
return Failure(IgnorableObject(f"Parsed object type: {to_type} does not match specified filter_type {filter_type}", file))
else:
return Failure(IgnorableObject("Invalid YAML {file}: {yaml_load}. Parsed object must contain 'type'", file))
else:
return Failure(InvalidObject(f"\nInvalid YAML {file}: {yaml_load}. Parsed object must be a dict", file))
except yaml.YAMLError as exc:
return Failure(InvalidObject(f"\nInvalid YAML {file}: {exc}\n", file))
except FileNotFoundError as e:
return Failure(InvalidObject(f"Specified YAML does not exist: {e}", file))
| [
"yaml.safe_load",
"typistry.protos.typed_dict.TypedDict",
"typistry.protos.invalid_object.IgnorableObject",
"typistry.protos.invalid_object.InvalidObject"
] | [((444, 466), 'yaml.safe_load', 'yaml.safe_load', (['stream'], {}), '(stream)\n', (458, 466), False, 'import yaml\n'), ((1603, 1661), 'typistry.protos.invalid_object.InvalidObject', 'InvalidObject', (['f"""Specified YAML does not exist: {e}"""', 'file'], {}), "(f'Specified YAML does not exist: {e}', file)\n", (1616, 1661), False, 'from typistry.protos.invalid_object import InvalidObject, IgnorableObject\n'), ((1325, 1426), 'typistry.protos.invalid_object.InvalidObject', 'InvalidObject', (['f"""\nInvalid YAML {file}: {yaml_load}. Parsed object must be a dict"""', 'file'], {}), '(\n f"""\nInvalid YAML {file}: {yaml_load}. Parsed object must be a dict""",\n file)\n', (1338, 1426), False, 'from typistry.protos.invalid_object import InvalidObject, IgnorableObject\n'), ((1489, 1545), 'typistry.protos.invalid_object.InvalidObject', 'InvalidObject', (['f"""\nInvalid YAML {file}: {exc}\n"""', 'file'], {}), '(f"""\nInvalid YAML {file}: {exc}\n""", file)\n', (1502, 1545), False, 'from typistry.protos.invalid_object import InvalidObject, IgnorableObject\n'), ((1173, 1275), 'typistry.protos.invalid_object.IgnorableObject', 'IgnorableObject', (['"""Invalid YAML {file}: {yaml_load}. Parsed object must contain \'type\'"""', 'file'], {}), '(\n "Invalid YAML {file}: {yaml_load}. Parsed object must contain \'type\'",\n file)\n', (1188, 1275), False, 'from typistry.protos.invalid_object import InvalidObject, IgnorableObject\n'), ((891, 925), 'typistry.protos.typed_dict.TypedDict', 'TypedDict', (['yaml_load'], {'type': 'to_type'}), '(yaml_load, type=to_type)\n', (900, 925), False, 'from typistry.protos.typed_dict import TypedDict\n'), ((1000, 1116), 'typistry.protos.invalid_object.IgnorableObject', 'IgnorableObject', (['f"""Parsed object type: {to_type} does not match specified filter_type {filter_type}"""', 'file'], {}), "(\n f'Parsed object type: {to_type} does not match specified filter_type {filter_type}'\n , file)\n", (1015, 1116), False, 'from typistry.protos.invalid_object import InvalidObject, IgnorableObject\n')] |
from datetime import datetime
import logging
from bs4 import BeautifulSoup
from db.models import Victim
from net.proxy import Proxy
from .sitecrawler import SiteCrawler
import time
class Nefilim(SiteCrawler):
actor = "Nefilim"
def _handle_page(self, soup):
victim_list = soup.find_all("header", class_="entry-header")
for victim in victim_list:
victim_title = victim.find("h2", class_="entry-title").text.strip()
victim_name = victim_title[0:victim_title.find(". Part")]
meta = victim.find("div", class_="entry-meta")
published = meta.find("time", class_="entry-date").attrs["datetime"]
published_dt = datetime.strptime(
published.strip()[:-6], "%Y-%m-%dT%H:%M:%S")
victim_leak_site = meta.find("span", class_="posted-on").find("a").attrs["href"]
q = self.session.query(Victim).filter_by(
url=victim_leak_site, site=self.site)
if q.count() == 0:
# new victim
v = Victim(name=victim_name, url=victim_leak_site, published=published_dt,
first_seen=datetime.utcnow(), last_seen=datetime.utcnow(), site=self.site)
self.session.add(v)
self.new_victims.append(v)
else:
# already seen, update last_seen
v = q.first()
v.last_seen = datetime.utcnow()
self.current_victims.append(v)
self.session.commit()
# server was timing out so slows it down a bit
time.sleep(1.0)
def scrape_victims(self):
with Proxy() as p:
r = p.get(f"{self.url}", headers=self.headers)
soup = BeautifulSoup(r.content.decode(), "html.parser")
page_count = 0
while True:
page_nav = soup.find("div", class_="nav-previous")
if page_nav is None:
break
url = page_nav.find("a").attrs["href"]
r = p.get(f"{url}", headers=self.headers)
soup = BeautifulSoup(r.content.decode(), "html.parser")
self._handle_page(soup) | [
"net.proxy.Proxy",
"time.sleep",
"datetime.datetime.utcnow"
] | [((1674, 1689), 'time.sleep', 'time.sleep', (['(1.0)'], {}), '(1.0)\n', (1684, 1689), False, 'import time\n'), ((1742, 1749), 'net.proxy.Proxy', 'Proxy', ([], {}), '()\n', (1747, 1749), False, 'from net.proxy import Proxy\n'), ((1511, 1528), 'datetime.datetime.utcnow', 'datetime.utcnow', ([], {}), '()\n', (1526, 1528), False, 'from datetime import datetime\n'), ((1241, 1258), 'datetime.datetime.utcnow', 'datetime.utcnow', ([], {}), '()\n', (1256, 1258), False, 'from datetime import datetime\n'), ((1270, 1287), 'datetime.datetime.utcnow', 'datetime.utcnow', ([], {}), '()\n', (1285, 1287), False, 'from datetime import datetime\n')] |
import numpy
import sympy
from matplotlib import pyplot
from sympy.utilities.lambdify import lambdify
# Set the font family and size to use for Matplotlib figures.
pyplot.rcParams['font.family'] = 'serif'
pyplot.rcParams['font.size'] = 16
sympy.init_printing()
x, nu, t = sympy.symbols('x nu t')
phi = (sympy.exp(-(x - 4 * t)**2 / (4 * nu * (t + 1))) +
sympy.exp(-(x - 4 * t - 2 * numpy.pi)**2 / (4 * nu * (t + 1))))
phiprime = phi.diff(x)
u = -2 * nu * (phiprime / phi) + 4
u_lamb = lambdify((t, x, nu), u)
# Set parameters.
nx = 101 # number of spatial grid points
L = 2.0 * numpy.pi # length of the domain
dx = L / (nx - 1) # spatial grid size
nu = 0.07 # viscosity
nt = 100 # number of time steps to compute
sigma = 0.1 # CFL limit
dt = sigma * dx**2 / nu # time-step size
# Discretize the domain.
x = numpy.linspace(0.0, L, num=nx)
# Set initial conditions.
t = 0.0
u0 = numpy.array([u_lamb(t, xi, nu) for xi in x])
# Integrate the Burgers' equation in time.
u = u0.copy()
for n in range(nt):
un = u.copy()
# Update all interior points.
u[1:-1] = (un[1:-1] -
un[1:-1] * dt / dx * (un[1:-1] - un[:-2]) +
nu * dt / dx**2 * (un[2:] - 2 * un[1:-1] + un[:-2]))
# Update boundary points.
u[0] = (un[0] -
un[0] * dt / dx * (un[0] - un[-1]) +
nu * dt / dx**2 * (un[1] - 2 * un[0] + un[-1]))
u[-1] = (un[-1] -
un[-1] * dt / dx * (un[-1] - un[-2]) +
nu * dt / dx**2 * (un[0] - 2 * un[-1] + un[-2]))
# Compute the analytical solution.
u_analytical = numpy.array([u_lamb(nt * dt, xi, nu) for xi in x])
# Plot the numerical solution along with the analytical solution.
pyplot.figure(figsize=(6.0, 4.0))
pyplot.xlabel('x')
pyplot.ylabel('u')
pyplot.grid()
pyplot.plot(x, u, label='Numerical',
color='C0', linestyle='-', linewidth=2)
pyplot.plot(x, u_analytical, label='Analytical',
color='C1', linestyle='--', linewidth=2)
pyplot.legend()
pyplot.xlim(0.0, L)
pyplot.ylim(0.0, 10.0);
pyplot.show()
pyplot.clf()
| [
"sympy.utilities.lambdify.lambdify",
"matplotlib.pyplot.grid",
"matplotlib.pyplot.ylabel",
"matplotlib.pyplot.xlabel",
"matplotlib.pyplot.plot",
"matplotlib.pyplot.clf",
"sympy.init_printing",
"sympy.symbols",
"numpy.linspace",
"matplotlib.pyplot.figure",
"sympy.exp",
"matplotlib.pyplot.ylim",... | [((241, 262), 'sympy.init_printing', 'sympy.init_printing', ([], {}), '()\n', (260, 262), False, 'import sympy\n'), ((275, 298), 'sympy.symbols', 'sympy.symbols', (['"""x nu t"""'], {}), "('x nu t')\n", (288, 298), False, 'import sympy\n'), ((496, 519), 'sympy.utilities.lambdify.lambdify', 'lambdify', (['(t, x, nu)', 'u'], {}), '((t, x, nu), u)\n', (504, 519), False, 'from sympy.utilities.lambdify import lambdify\n'), ((827, 857), 'numpy.linspace', 'numpy.linspace', (['(0.0)', 'L'], {'num': 'nx'}), '(0.0, L, num=nx)\n', (841, 857), False, 'import numpy\n'), ((1690, 1723), 'matplotlib.pyplot.figure', 'pyplot.figure', ([], {'figsize': '(6.0, 4.0)'}), '(figsize=(6.0, 4.0))\n', (1703, 1723), False, 'from matplotlib import pyplot\n'), ((1724, 1742), 'matplotlib.pyplot.xlabel', 'pyplot.xlabel', (['"""x"""'], {}), "('x')\n", (1737, 1742), False, 'from matplotlib import pyplot\n'), ((1743, 1761), 'matplotlib.pyplot.ylabel', 'pyplot.ylabel', (['"""u"""'], {}), "('u')\n", (1756, 1761), False, 'from matplotlib import pyplot\n'), ((1762, 1775), 'matplotlib.pyplot.grid', 'pyplot.grid', ([], {}), '()\n', (1773, 1775), False, 'from matplotlib import pyplot\n'), ((1776, 1852), 'matplotlib.pyplot.plot', 'pyplot.plot', (['x', 'u'], {'label': '"""Numerical"""', 'color': '"""C0"""', 'linestyle': '"""-"""', 'linewidth': '(2)'}), "(x, u, label='Numerical', color='C0', linestyle='-', linewidth=2)\n", (1787, 1852), False, 'from matplotlib import pyplot\n'), ((1865, 1958), 'matplotlib.pyplot.plot', 'pyplot.plot', (['x', 'u_analytical'], {'label': '"""Analytical"""', 'color': '"""C1"""', 'linestyle': '"""--"""', 'linewidth': '(2)'}), "(x, u_analytical, label='Analytical', color='C1', linestyle='--',\n linewidth=2)\n", (1876, 1958), False, 'from matplotlib import pyplot\n'), ((1967, 1982), 'matplotlib.pyplot.legend', 'pyplot.legend', ([], {}), '()\n', (1980, 1982), False, 'from matplotlib import pyplot\n'), ((1983, 2002), 'matplotlib.pyplot.xlim', 'pyplot.xlim', (['(0.0)', 'L'], {}), '(0.0, L)\n', (1994, 2002), False, 'from matplotlib import pyplot\n'), ((2003, 2025), 'matplotlib.pyplot.ylim', 'pyplot.ylim', (['(0.0)', '(10.0)'], {}), '(0.0, 10.0)\n', (2014, 2025), False, 'from matplotlib import pyplot\n'), ((2027, 2040), 'matplotlib.pyplot.show', 'pyplot.show', ([], {}), '()\n', (2038, 2040), False, 'from matplotlib import pyplot\n'), ((2041, 2053), 'matplotlib.pyplot.clf', 'pyplot.clf', ([], {}), '()\n', (2051, 2053), False, 'from matplotlib import pyplot\n'), ((306, 355), 'sympy.exp', 'sympy.exp', (['(-(x - 4 * t) ** 2 / (4 * nu * (t + 1)))'], {}), '(-(x - 4 * t) ** 2 / (4 * nu * (t + 1)))\n', (315, 355), False, 'import sympy\n'), ((363, 427), 'sympy.exp', 'sympy.exp', (['(-(x - 4 * t - 2 * numpy.pi) ** 2 / (4 * nu * (t + 1)))'], {}), '(-(x - 4 * t - 2 * numpy.pi) ** 2 / (4 * nu * (t + 1)))\n', (372, 427), False, 'import sympy\n')] |
from django import forms
from django.forms import ModelForm
from .models import Category,Image, Location
class ImageForm(forms.ModelForm):
class Meta:
model = Image
fields = '__all__'
CATEGORIES =(
("1", "Cars"),
("2", "Food"),
("3", "Travel"),
("4", "Animals"),
("5", "Nature"),
("6", "Sports"),
)
class ImagesForm(forms.Form):
image = forms.ImageField(required=True)
pic_name = forms.CharField(required=True)
description = forms.CharField(required=True)
location = forms.CharField(required=True)
pic_category = forms.ChoiceField(choices=CATEGORIES, required=True) | [
"django.forms.ImageField",
"django.forms.ChoiceField",
"django.forms.CharField"
] | [((398, 429), 'django.forms.ImageField', 'forms.ImageField', ([], {'required': '(True)'}), '(required=True)\n', (414, 429), False, 'from django import forms\n'), ((446, 476), 'django.forms.CharField', 'forms.CharField', ([], {'required': '(True)'}), '(required=True)\n', (461, 476), False, 'from django import forms\n'), ((495, 525), 'django.forms.CharField', 'forms.CharField', ([], {'required': '(True)'}), '(required=True)\n', (510, 525), False, 'from django import forms\n'), ((541, 571), 'django.forms.CharField', 'forms.CharField', ([], {'required': '(True)'}), '(required=True)\n', (556, 571), False, 'from django import forms\n'), ((591, 643), 'django.forms.ChoiceField', 'forms.ChoiceField', ([], {'choices': 'CATEGORIES', 'required': '(True)'}), '(choices=CATEGORIES, required=True)\n', (608, 643), False, 'from django import forms\n')] |
from .base import FunctionalTest
from selenium import webdriver
from selenium.webdriver.common.keys import Keys
class NewVisitorTest(FunctionalTest):
def test_can_start_a_list_and_retrieve_it_later(self):
self.browser.get(self.server_url)
self.assertIn("To-Do", self.browser.title)
header_text = self.browser.find_element_by_tag_name("h1").text
self.assertIn("To-Do", header_text)
inputbox = self.get_item_input_box()
self.assertEqual(
inputbox.get_attribute("placeholder"),
"Enter a to-do item"
)
inputbox.send_keys("Buy peacock feathers")
inputbox.send_keys(Keys.ENTER)
edith_list_url = self.browser.current_url
self.assertRegex(edith_list_url, "/lists/.+")
self.check_for_row_in_list_table('1: Buy peacock feathers')
inputbox = self.get_item_input_box()
inputbox.send_keys("Use peacock feathers to make a fly")
inputbox.send_keys(Keys.ENTER)
self.check_for_row_in_list_table("1: Buy peacock feathers")
self.check_for_row_in_list_table(
"2: Use peacock feathers to make a fly"
)
self.browser.quit()
self.browser = webdriver.Firefox()
self.browser.get(self.server_url)
page_text = self.browser.find_element_by_tag_name("body").text
self.assertNotIn("Buy peacock feathers", page_text)
self.assertNotIn("make a fly", page_text)
inputbox = self.get_item_input_box()
inputbox.send_keys("Buy milk")
inputbox.send_keys(Keys.ENTER)
francis_list_url = self.browser.current_url
self.assertRegex(francis_list_url, "/lists/.+")
self.assertNotEqual(francis_list_url, edith_list_url)
page_text = self.browser.find_element_by_tag_name("body").text
self.assertNotIn("Buy peacock feathers", page_text)
self.assertNotIn("make a fly", page_text)
| [
"selenium.webdriver.Firefox"
] | [((1226, 1245), 'selenium.webdriver.Firefox', 'webdriver.Firefox', ([], {}), '()\n', (1243, 1245), False, 'from selenium import webdriver\n')] |
# Copyright 2021 Google LLC
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""Tests for reservoir_nn.keras.rewiring."""
from absl.testing import absltest
from absl.testing import parameterized
import numpy as np
from reservoir_nn.keras import rewiring
import tensorflow as tf
class AdaptiveSparseReservoirTest(parameterized.TestCase):
def test_layer_with_num_connections_works(self):
x = tf.constant([1.])
layer = rewiring.AdaptiveSparseReservoir(
units=10,
reservoir_initializer=10,
)
layer(x)
def test_layer_with_large_num_connections_fails(self):
x = tf.constant([1.])
layer = rewiring.AdaptiveSparseReservoir(
units=10,
reservoir_initializer=100,
)
with self.assertRaisesRegex(ValueError,
"Cannot build layer.*(100).*(1).*(10)"):
layer(x)
def test_layer_with_reservoir_works(self):
initial_reservoir = np.arange(10).reshape(1, 10)
x = tf.constant([1.])
layer = rewiring.AdaptiveSparseReservoir(
units=10,
reservoir_initializer=initial_reservoir,
)
np.testing.assert_array_equal(layer(x), np.arange(10))
def test_layer_with_misshaped_reservoir_fails(self):
initial_reservoir = np.arange(10).reshape(2, 5)
x = tf.constant([1.])
layer = rewiring.AdaptiveSparseReservoir(
units=10,
reservoir_initializer=initial_reservoir,
)
with self.assertRaisesRegex(
ValueError,
r"Reservoir has a shape of \(2, 5\), but the layer expects \(1, 10\)"):
layer(x)
def test_get_coo_weight_matrix_works(self):
initial_reservoir = np.arange(10).reshape(2, 5)
x = tf.keras.Input(shape=(2,))
layer = rewiring.AdaptiveSparseReservoir(
units=5,
reservoir_initializer=initial_reservoir,
)
layer(x)
coo = layer.get_coo_weight_matrix().toarray()
np.testing.assert_array_equal(coo, initial_reservoir)
def test_get_coo_age_matrix_works(self):
initial_reservoir = np.arange(10).reshape(2, 5)
x = tf.keras.Input(shape=(2,))
layer = rewiring.AdaptiveSparseReservoir(
units=5,
reservoir_initializer=initial_reservoir,
)
layer(x)
coo = layer.get_coo_age_matrix().toarray()
np.testing.assert_array_equal(coo, np.zeros((2, 5)))
policy = rewiring.MutationPolicy(
candidate_fraction=0.0,
candidate_mutation_rate=1.0,
)
policy.mutation_step(layer)
coo = layer.get_coo_age_matrix().toarray()
np.testing.assert_array_equal(
coo,
[[0., 1, 1, 1, 1], [1., 1, 1, 1, 1]],
)
def test_apply_global_policy_works(self):
policy = rewiring.MutationPolicy(
candidate_fraction=0.5,
candidate_mutation_rate=0.5,
)
gpolicy = rewiring.GlobalPolicy(
scale_candidate_fraction=0.5, scale_candidate_mutation_rate=0.5)
policy = policy.apply_global_policy(gpolicy)
self.assertEqual(
policy,
rewiring.MutationPolicy(
candidate_fraction=0.25, candidate_mutation_rate=0.25))
def test_compute_mutation_probability_works(self):
initial_reservoir = np.arange(10).reshape(2, 5)
x = tf.keras.Input(shape=(2,))
layer = rewiring.AdaptiveSparseReservoir(
units=5,
reservoir_initializer=initial_reservoir,
)
layer(x)
policy = rewiring.MutationPolicy(
candidate_fraction=1.0,
candidate_mutation_rate=1.0,
)
p = policy.compute_mutation_probability(
sparse_values=layer.sparse_values.value(),
sparse_ages=layer.sparse_ages.value(),
)
coo = layer.get_coo_weight_matrix().copy()
coo.data[:] = p
np.testing.assert_allclose(
coo.toarray(), [
[0., 0.7, 0.3, 0, 0],
[0., 0, 0, 0, 0],
], atol=0.1)
def test_mutation_works(self):
initial_reservoir = np.arange(10).reshape(2, 5)
x = tf.keras.Input(shape=(2,))
layer = rewiring.AdaptiveSparseReservoir(
units=5,
reservoir_initializer=initial_reservoir,
)
layer(x)
coo = layer.get_coo_weight_matrix()
np.testing.assert_allclose(coo.toarray(), [
[0., 1, 2, 3, 4],
[5., 6, 7, 8, 9],
])
rng = np.random.RandomState(1234)
policy = rewiring.MutationPolicy(
candidate_fraction=0.2,
candidate_mutation_rate=1.0,
)
policy.mutation_step(layer, rng)
coo = layer.get_coo_weight_matrix()
# least active connections are replenished with zeros with 100%
# probability.
np.testing.assert_allclose(coo.toarray(), [
[0., 0, 2, 3, 4],
[5., 6, 7, 8, 9],
])
class SparseEvolutionEnd2EndTest(parameterized.TestCase):
def test_fit_small_model_works(self):
"""Testing fitting a 4x4 sparse reservoir with 4 truth connections."""
nunits = 4
inputs = tf.keras.Input(shape=(nunits,))
policy = rewiring.MutationPolicy(
candidate_fraction=0.5,
candidate_mutation_rate=0.1,
)
layer = rewiring.AdaptiveSparseReservoir(
units=nunits,
reservoir_initializer=2 * nunits,
# regularizer helps sparsify the redundant connections.
kernel_regularizer=tf.keras.regularizers.l2(1e-2),
)
outputs = layer(inputs)
rng = np.random.RandomState(1333)
model = tf.keras.Model(inputs, outputs)
# Low dimension model prefers SGD:
model.compile(
optimizer=tf.keras.optimizers.SGD(lr=0.1, nesterov=True),
loss="mse",
metrics=["mse"])
x = rng.uniform(size=(1000000, nunits)) - 0.5
y = x[:, ::-1].copy()
truth = np.eye(4)[::-1]
def mutation_schedule(epoch):
del epoch
return rewiring.GlobalPolicy()
model.fit(
x,
y,
batch_size=int(len(x) / 100),
epochs=10,
verbose=False,
callbacks=rewiring.MutationCallback(
policy={layer: policy},
mutation_schedule=mutation_schedule,
rng=rng,
verbose=1))
connection = layer.get_coo_weight_matrix().toarray()
# Use truth * 0.65 here because L2 biases the fit towards zero.
np.testing.assert_allclose(connection, truth * 0.65, atol=0.10)
def test_fit_large_model_works(self):
"""Testing fitting a 100x100 sparse reservoir with 100 truth connections."""
nunits = 100
inputs = tf.keras.Input(shape=(nunits,))
policy = rewiring.MutationPolicy(
candidate_fraction=0.2,
candidate_mutation_rate=0.8,
)
layer = rewiring.AdaptiveSparseReservoir(
units=nunits,
reservoir_initializer=2 * nunits,
)
outputs = layer(inputs)
model = tf.keras.Model(inputs, outputs)
model.compile(loss="mse", metrics=["mse"])
x = np.random.uniform(size=(100000, nunits)) - 0.5
truth = np.eye(nunits)[::-1].copy()
y = np.einsum("ij,jk->ik", x, truth)
class Reporter(tf.keras.callbacks.Callback):
def on_epoch_end(self, epoch, logs):
connection = layer.get_coo_weight_matrix().toarray()
ages = layer.get_coo_age_matrix().toarray()
cross = np.abs(connection) * truth
print(sorted(zip(*np.nonzero(cross))))
significant_elements = np.sum(cross > 0.02)
print(significant_elements)
print(ages[cross > 0.02])
model.fit(
x,
y,
batch_size=1600,
epochs=20,
verbose=True,
callbacks=[
rewiring.MutationCallback(policy=policy, verbose=1),
Reporter()
])
connection = layer.get_coo_weight_matrix().todense()
# We shall have some correlation with the truth after training for a while.
cross = np.abs(connection) * truth
np.testing.assert_allclose(np.sum(cross > 0.02), 80, atol=20)
if __name__ == "__main__":
absltest.main()
| [
"numpy.einsum",
"reservoir_nn.keras.rewiring.AdaptiveSparseReservoir",
"numpy.arange",
"numpy.random.RandomState",
"numpy.testing.assert_allclose",
"tensorflow.keras.optimizers.SGD",
"reservoir_nn.keras.rewiring.GlobalPolicy",
"reservoir_nn.keras.rewiring.MutationPolicy",
"numpy.testing.assert_array... | [((8276, 8291), 'absl.testing.absltest.main', 'absltest.main', ([], {}), '()\n', (8289, 8291), False, 'from absl.testing import absltest\n'), ((899, 917), 'tensorflow.constant', 'tf.constant', (['[1.0]'], {}), '([1.0])\n', (910, 917), True, 'import tensorflow as tf\n'), ((929, 997), 'reservoir_nn.keras.rewiring.AdaptiveSparseReservoir', 'rewiring.AdaptiveSparseReservoir', ([], {'units': '(10)', 'reservoir_initializer': '(10)'}), '(units=10, reservoir_initializer=10)\n', (961, 997), False, 'from reservoir_nn.keras import rewiring\n'), ((1100, 1118), 'tensorflow.constant', 'tf.constant', (['[1.0]'], {}), '([1.0])\n', (1111, 1118), True, 'import tensorflow as tf\n'), ((1130, 1199), 'reservoir_nn.keras.rewiring.AdaptiveSparseReservoir', 'rewiring.AdaptiveSparseReservoir', ([], {'units': '(10)', 'reservoir_initializer': '(100)'}), '(units=10, reservoir_initializer=100)\n', (1162, 1199), False, 'from reservoir_nn.keras import rewiring\n'), ((1464, 1482), 'tensorflow.constant', 'tf.constant', (['[1.0]'], {}), '([1.0])\n', (1475, 1482), True, 'import tensorflow as tf\n'), ((1494, 1582), 'reservoir_nn.keras.rewiring.AdaptiveSparseReservoir', 'rewiring.AdaptiveSparseReservoir', ([], {'units': '(10)', 'reservoir_initializer': 'initial_reservoir'}), '(units=10, reservoir_initializer=\n initial_reservoir)\n', (1526, 1582), False, 'from reservoir_nn.keras import rewiring\n'), ((1778, 1796), 'tensorflow.constant', 'tf.constant', (['[1.0]'], {}), '([1.0])\n', (1789, 1796), True, 'import tensorflow as tf\n'), ((1808, 1896), 'reservoir_nn.keras.rewiring.AdaptiveSparseReservoir', 'rewiring.AdaptiveSparseReservoir', ([], {'units': '(10)', 'reservoir_initializer': 'initial_reservoir'}), '(units=10, reservoir_initializer=\n initial_reservoir)\n', (1840, 1896), False, 'from reservoir_nn.keras import rewiring\n'), ((2172, 2198), 'tensorflow.keras.Input', 'tf.keras.Input', ([], {'shape': '(2,)'}), '(shape=(2,))\n', (2186, 2198), True, 'import tensorflow as tf\n'), ((2211, 2298), 'reservoir_nn.keras.rewiring.AdaptiveSparseReservoir', 'rewiring.AdaptiveSparseReservoir', ([], {'units': '(5)', 'reservoir_initializer': 'initial_reservoir'}), '(units=5, reservoir_initializer=\n initial_reservoir)\n', (2243, 2298), False, 'from reservoir_nn.keras import rewiring\n'), ((2385, 2438), 'numpy.testing.assert_array_equal', 'np.testing.assert_array_equal', (['coo', 'initial_reservoir'], {}), '(coo, initial_reservoir)\n', (2414, 2438), True, 'import numpy as np\n'), ((2544, 2570), 'tensorflow.keras.Input', 'tf.keras.Input', ([], {'shape': '(2,)'}), '(shape=(2,))\n', (2558, 2570), True, 'import tensorflow as tf\n'), ((2583, 2670), 'reservoir_nn.keras.rewiring.AdaptiveSparseReservoir', 'rewiring.AdaptiveSparseReservoir', ([], {'units': '(5)', 'reservoir_initializer': 'initial_reservoir'}), '(units=5, reservoir_initializer=\n initial_reservoir)\n', (2615, 2670), False, 'from reservoir_nn.keras import rewiring\n'), ((2821, 2897), 'reservoir_nn.keras.rewiring.MutationPolicy', 'rewiring.MutationPolicy', ([], {'candidate_fraction': '(0.0)', 'candidate_mutation_rate': '(1.0)'}), '(candidate_fraction=0.0, candidate_mutation_rate=1.0)\n', (2844, 2897), False, 'from reservoir_nn.keras import rewiring\n'), ((3004, 3078), 'numpy.testing.assert_array_equal', 'np.testing.assert_array_equal', (['coo', '[[0.0, 1, 1, 1, 1], [1.0, 1, 1, 1, 1]]'], {}), '(coo, [[0.0, 1, 1, 1, 1], [1.0, 1, 1, 1, 1]])\n', (3033, 3078), True, 'import numpy as np\n'), ((3158, 3234), 'reservoir_nn.keras.rewiring.MutationPolicy', 'rewiring.MutationPolicy', ([], {'candidate_fraction': '(0.5)', 'candidate_mutation_rate': '(0.5)'}), '(candidate_fraction=0.5, candidate_mutation_rate=0.5)\n', (3181, 3234), False, 'from reservoir_nn.keras import rewiring\n'), ((3272, 3362), 'reservoir_nn.keras.rewiring.GlobalPolicy', 'rewiring.GlobalPolicy', ([], {'scale_candidate_fraction': '(0.5)', 'scale_candidate_mutation_rate': '(0.5)'}), '(scale_candidate_fraction=0.5,\n scale_candidate_mutation_rate=0.5)\n', (3293, 3362), False, 'from reservoir_nn.keras import rewiring\n'), ((3673, 3699), 'tensorflow.keras.Input', 'tf.keras.Input', ([], {'shape': '(2,)'}), '(shape=(2,))\n', (3687, 3699), True, 'import tensorflow as tf\n'), ((3712, 3799), 'reservoir_nn.keras.rewiring.AdaptiveSparseReservoir', 'rewiring.AdaptiveSparseReservoir', ([], {'units': '(5)', 'reservoir_initializer': 'initial_reservoir'}), '(units=5, reservoir_initializer=\n initial_reservoir)\n', (3744, 3799), False, 'from reservoir_nn.keras import rewiring\n'), ((3846, 3922), 'reservoir_nn.keras.rewiring.MutationPolicy', 'rewiring.MutationPolicy', ([], {'candidate_fraction': '(1.0)', 'candidate_mutation_rate': '(1.0)'}), '(candidate_fraction=1.0, candidate_mutation_rate=1.0)\n', (3869, 3922), False, 'from reservoir_nn.keras import rewiring\n'), ((4402, 4428), 'tensorflow.keras.Input', 'tf.keras.Input', ([], {'shape': '(2,)'}), '(shape=(2,))\n', (4416, 4428), True, 'import tensorflow as tf\n'), ((4441, 4528), 'reservoir_nn.keras.rewiring.AdaptiveSparseReservoir', 'rewiring.AdaptiveSparseReservoir', ([], {'units': '(5)', 'reservoir_initializer': 'initial_reservoir'}), '(units=5, reservoir_initializer=\n initial_reservoir)\n', (4473, 4528), False, 'from reservoir_nn.keras import rewiring\n'), ((4718, 4745), 'numpy.random.RandomState', 'np.random.RandomState', (['(1234)'], {}), '(1234)\n', (4739, 4745), True, 'import numpy as np\n'), ((4760, 4836), 'reservoir_nn.keras.rewiring.MutationPolicy', 'rewiring.MutationPolicy', ([], {'candidate_fraction': '(0.2)', 'candidate_mutation_rate': '(1.0)'}), '(candidate_fraction=0.2, candidate_mutation_rate=1.0)\n', (4783, 4836), False, 'from reservoir_nn.keras import rewiring\n'), ((5338, 5369), 'tensorflow.keras.Input', 'tf.keras.Input', ([], {'shape': '(nunits,)'}), '(shape=(nunits,))\n', (5352, 5369), True, 'import tensorflow as tf\n'), ((5384, 5460), 'reservoir_nn.keras.rewiring.MutationPolicy', 'rewiring.MutationPolicy', ([], {'candidate_fraction': '(0.5)', 'candidate_mutation_rate': '(0.1)'}), '(candidate_fraction=0.5, candidate_mutation_rate=0.1)\n', (5407, 5460), False, 'from reservoir_nn.keras import rewiring\n'), ((5762, 5789), 'numpy.random.RandomState', 'np.random.RandomState', (['(1333)'], {}), '(1333)\n', (5783, 5789), True, 'import numpy as np\n'), ((5803, 5834), 'tensorflow.keras.Model', 'tf.keras.Model', (['inputs', 'outputs'], {}), '(inputs, outputs)\n', (5817, 5834), True, 'import tensorflow as tf\n'), ((6623, 6685), 'numpy.testing.assert_allclose', 'np.testing.assert_allclose', (['connection', '(truth * 0.65)'], {'atol': '(0.1)'}), '(connection, truth * 0.65, atol=0.1)\n', (6649, 6685), True, 'import numpy as np\n'), ((6839, 6870), 'tensorflow.keras.Input', 'tf.keras.Input', ([], {'shape': '(nunits,)'}), '(shape=(nunits,))\n', (6853, 6870), True, 'import tensorflow as tf\n'), ((6884, 6960), 'reservoir_nn.keras.rewiring.MutationPolicy', 'rewiring.MutationPolicy', ([], {'candidate_fraction': '(0.2)', 'candidate_mutation_rate': '(0.8)'}), '(candidate_fraction=0.2, candidate_mutation_rate=0.8)\n', (6907, 6960), False, 'from reservoir_nn.keras import rewiring\n'), ((6996, 7081), 'reservoir_nn.keras.rewiring.AdaptiveSparseReservoir', 'rewiring.AdaptiveSparseReservoir', ([], {'units': 'nunits', 'reservoir_initializer': '(2 * nunits)'}), '(units=nunits, reservoir_initializer=2 * nunits\n )\n', (7028, 7081), False, 'from reservoir_nn.keras import rewiring\n'), ((7141, 7172), 'tensorflow.keras.Model', 'tf.keras.Model', (['inputs', 'outputs'], {}), '(inputs, outputs)\n', (7155, 7172), True, 'import tensorflow as tf\n'), ((7326, 7358), 'numpy.einsum', 'np.einsum', (['"""ij,jk->ik"""', 'x', 'truth'], {}), "('ij,jk->ik', x, truth)\n", (7335, 7358), True, 'import numpy as np\n'), ((1646, 1659), 'numpy.arange', 'np.arange', (['(10)'], {}), '(10)\n', (1655, 1659), True, 'import numpy as np\n'), ((2789, 2805), 'numpy.zeros', 'np.zeros', (['(2, 5)'], {}), '((2, 5))\n', (2797, 2805), True, 'import numpy as np\n'), ((3465, 3543), 'reservoir_nn.keras.rewiring.MutationPolicy', 'rewiring.MutationPolicy', ([], {'candidate_fraction': '(0.25)', 'candidate_mutation_rate': '(0.25)'}), '(candidate_fraction=0.25, candidate_mutation_rate=0.25)\n', (3488, 3543), False, 'from reservoir_nn.keras import rewiring\n'), ((6095, 6104), 'numpy.eye', 'np.eye', (['(4)'], {}), '(4)\n', (6101, 6104), True, 'import numpy as np\n'), ((6175, 6198), 'reservoir_nn.keras.rewiring.GlobalPolicy', 'rewiring.GlobalPolicy', ([], {}), '()\n', (6196, 6198), False, 'from reservoir_nn.keras import rewiring\n'), ((7230, 7270), 'numpy.random.uniform', 'np.random.uniform', ([], {'size': '(100000, nunits)'}), '(size=(100000, nunits))\n', (7247, 7270), True, 'import numpy as np\n'), ((8152, 8170), 'numpy.abs', 'np.abs', (['connection'], {}), '(connection)\n', (8158, 8170), True, 'import numpy as np\n'), ((8210, 8230), 'numpy.sum', 'np.sum', (['(cross > 0.02)'], {}), '(cross > 0.02)\n', (8216, 8230), True, 'import numpy as np\n'), ((1426, 1439), 'numpy.arange', 'np.arange', (['(10)'], {}), '(10)\n', (1435, 1439), True, 'import numpy as np\n'), ((1741, 1754), 'numpy.arange', 'np.arange', (['(10)'], {}), '(10)\n', (1750, 1754), True, 'import numpy as np\n'), ((2135, 2148), 'numpy.arange', 'np.arange', (['(10)'], {}), '(10)\n', (2144, 2148), True, 'import numpy as np\n'), ((2507, 2520), 'numpy.arange', 'np.arange', (['(10)'], {}), '(10)\n', (2516, 2520), True, 'import numpy as np\n'), ((3636, 3649), 'numpy.arange', 'np.arange', (['(10)'], {}), '(10)\n', (3645, 3649), True, 'import numpy as np\n'), ((4365, 4378), 'numpy.arange', 'np.arange', (['(10)'], {}), '(10)\n', (4374, 4378), True, 'import numpy as np\n'), ((5685, 5715), 'tensorflow.keras.regularizers.l2', 'tf.keras.regularizers.l2', (['(0.01)'], {}), '(0.01)\n', (5709, 5715), True, 'import tensorflow as tf\n'), ((5912, 5958), 'tensorflow.keras.optimizers.SGD', 'tf.keras.optimizers.SGD', ([], {'lr': '(0.1)', 'nesterov': '(True)'}), '(lr=0.1, nesterov=True)\n', (5935, 5958), True, 'import tensorflow as tf\n'), ((6335, 6446), 'reservoir_nn.keras.rewiring.MutationCallback', 'rewiring.MutationCallback', ([], {'policy': '{layer: policy}', 'mutation_schedule': 'mutation_schedule', 'rng': 'rng', 'verbose': '(1)'}), '(policy={layer: policy}, mutation_schedule=\n mutation_schedule, rng=rng, verbose=1)\n', (6360, 6446), False, 'from reservoir_nn.keras import rewiring\n'), ((7687, 7707), 'numpy.sum', 'np.sum', (['(cross > 0.02)'], {}), '(cross > 0.02)\n', (7693, 7707), True, 'import numpy as np\n'), ((7289, 7303), 'numpy.eye', 'np.eye', (['nunits'], {}), '(nunits)\n', (7295, 7303), True, 'import numpy as np\n'), ((7582, 7600), 'numpy.abs', 'np.abs', (['connection'], {}), '(connection)\n', (7588, 7600), True, 'import numpy as np\n'), ((7914, 7965), 'reservoir_nn.keras.rewiring.MutationCallback', 'rewiring.MutationCallback', ([], {'policy': 'policy', 'verbose': '(1)'}), '(policy=policy, verbose=1)\n', (7939, 7965), False, 'from reservoir_nn.keras import rewiring\n'), ((7635, 7652), 'numpy.nonzero', 'np.nonzero', (['cross'], {}), '(cross)\n', (7645, 7652), True, 'import numpy as np\n')] |
from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
import cv2
import numpy as np
from progress.bar import Bar
import torch
import math
import copy
from .base_debugger import BaseDebugger
from models.utils import _tranpose_and_gather_feat, _gather_feat
from models.decode import _topk_original, _topk, _topk_channel, _nms
from datasets.dataset.utils import _bbox_overlaps
from utils.image import transform_preds
class TriCtdetDebugger(BaseDebugger):
def __init__(self, opt):
super(TriCtdetDebugger, self).__init__(opt)
def forward(self, images):
with torch.no_grad():
output = self.model(images)[-1]
tl = output['tl'].sigmoid_()
bl = output['bl'].sigmoid_()
br = output['br'].sigmoid_()
ct = output['ct'].sigmoid_()
tl_tag = output['tl_tag']
bl_tag = output['bl_tag']
br_tag = output['br_tag']
tl_reg = output['tl_reg']
bl_reg = output['bl_reg']
br_reg = output['br_reg']
ct_reg = output['ct_reg']
detections = {'tl_heatmap':tl, 'bl_heatmap':bl, 'br_heatmap':br, 'ct_heatmap':ct,
'tl_reg':tl_reg, 'bl_reg':bl_reg, 'br_reg':br_reg, 'ct_reg':ct_reg,
'tl_tag':tl_tag, 'bl_tag':bl_tag, 'br_tag':br_tag}
return detections
def debug(self, detections, targets, ae_threshold):
tl_heat = detections['tl_heatmap']
bl_heat = detections['bl_heatmap']
br_heat = detections['br_heatmap']
ct_heat = detections['ct_heatmap']
targets['tl_tag'] = targets['tl_tag'][targets['reg_mask']].unsqueeze(0)
targets['bl_tag'] = targets['bl_tag'][targets['reg_mask']].unsqueeze(0)
targets['br_tag'] = targets['br_tag'][targets['reg_mask']].unsqueeze(0)
targets['ct_tag'] = targets['ct_tag'][targets['reg_mask']].unsqueeze(0)
targets['tl_reg'] = targets['tl_reg'][targets['reg_mask']].unsqueeze(0)
targets['bl_reg'] = targets['bl_reg'][targets['reg_mask']].unsqueeze(0)
targets['br_reg'] = targets['br_reg'][targets['reg_mask']].unsqueeze(0)
targets['ct_reg'] = targets['ct_reg'][targets['reg_mask']].unsqueeze(0)
batch, cat, height, width = tl_heat.size()
# tl_scores, tl_inds, tl_clses, tl_ys, tl_xs = _topk(tl_heat, K=256)
# bl_scores, bl_inds, bl_clses, bl_ys, bl_xs = _topk(bl_heat, K=256)
# br_scores, br_inds, br_clses, br_ys, br_xs = _topk(br_heat, K=256)
# ct_scores, ct_inds, ct_clses, ct_ys, ct_xs = _topk(ct_heat, K=256)
tl_tag = detections['tl_tag']
bl_tag = detections['bl_tag']
br_tag = detections['br_tag']
tl_reg = detections['tl_reg']
bl_reg = detections['bl_reg']
br_reg = detections['br_reg']
ct_reg = detections['ct_reg']
# gather by gt
tl_tag = _tranpose_and_gather_feat(tl_tag, targets['tl_tag'].to(torch.device("cuda")))
bl_tag = _tranpose_and_gather_feat(bl_tag, targets['bl_tag'].to(torch.device("cuda")))
br_tag = _tranpose_and_gather_feat(br_tag, targets['br_tag'].to(torch.device("cuda")))
# gather by top k
# tl_tag = _tranpose_and_gather_feat(tl_tag, tl_inds)
# bl_tag = _tranpose_and_gather_feat(bl_tag, bl_inds)
# br_tag = _tranpose_and_gather_feat(br_tag, br_inds)
avg_tag = (tl_tag + bl_tag + br_tag) / 3
dists_tl = torch.abs(avg_tag - tl_tag).to(torch.device("cpu")).numpy()
dists_bl = torch.abs(bl_tag - avg_tag).to(torch.device("cpu")).numpy()
dists_br = torch.abs(avg_tag - br_tag).to(torch.device("cpu")).numpy()
dists_avg = (dists_tl.sum() + dists_bl.sum() + dists_br.sum()) / dists_tl.shape[1] / 3
min_tl = dists_tl.min()
max_tl = dists_tl.max()
min_bl = dists_bl.min()
max_bl = dists_bl.max()
min_br = dists_br.min()
max_br = dists_br.max()
# gather by gt
tl_reg = _tranpose_and_gather_feat(tl_reg, targets['tl_tag'].to(torch.device("cuda")))
bl_reg = _tranpose_and_gather_feat(bl_reg, targets['bl_tag'].to(torch.device("cuda")))
br_reg = _tranpose_and_gather_feat(br_reg, targets['br_tag'].to(torch.device("cuda")))
ct_reg = _tranpose_and_gather_feat(ct_reg, targets['ct_tag'].to(torch.device("cuda")))
# reg_diff_tl = tl_reg - targets['tl_reg'].to(torch.device("cuda"))
# reg_diff_tl = torch.sqrt(reg_diff_tl[..., 0]*reg_diff_tl[..., 0] + reg_diff_tl[..., 1]*reg_diff_tl[..., 1])
# reg_diff_bl = bl_reg - targets['bl_reg'].to(torch.device("cuda"))
# reg_diff_bl = torch.sqrt(reg_diff_bl[..., 0] * reg_diff_bl[..., 0] + reg_diff_bl[..., 1] * reg_diff_bl[..., 1])
# reg_diff_br = br_reg - targets['br_reg'].to(torch.device("cuda"))
# reg_diff_br = torch.sqrt(reg_diff_br[..., 0] * reg_diff_br[..., 0] + reg_diff_br[..., 1] * reg_diff_br[..., 1])
# reg_diff_ct = ct_reg - targets['ct_reg'].to(torch.device("cuda"))
# reg_diff_ct = torch.sqrt(reg_diff_ct[..., 0] * reg_diff_ct[..., 0] + reg_diff_ct[..., 1] * reg_diff_ct[..., 1])
tl_xs = ((targets['tl_tag'] % (width * height)) % width).int().float().to(torch.device("cuda"))
tl_ys = ((targets['tl_tag'] % (width * height)) / width).int().float().to(torch.device("cuda"))
bl_xs = ((targets['bl_tag'] % (width * height)) % width).int().float().to(torch.device("cuda"))
bl_ys = ((targets['bl_tag'] % (width * height)) / width).int().float().to(torch.device("cuda"))
br_xs = ((targets['br_tag'] % (width * height)) % width).int().float().to(torch.device("cuda"))
br_ys = ((targets['br_tag'] % (width * height)) / width).int().float().to(torch.device("cuda"))
ct_xs = ((targets['ct_tag'] % (width * height)) % width).int().float().to(torch.device("cuda"))
ct_ys = ((targets['ct_tag'] % (width * height)) / width).int().float().to(torch.device("cuda"))
tl_xs_pr = (tl_xs + tl_reg[..., 0]).squeeze(0).to(torch.device("cpu")).numpy()
tl_ys_pr = (tl_ys + tl_reg[..., 1]).squeeze(0).to(torch.device("cpu")).numpy()
bl_xs_pr = (bl_xs + bl_reg[..., 0]).squeeze(0).to(torch.device("cpu")).numpy()
bl_ys_pr = (bl_ys + bl_reg[..., 1]).squeeze(0).to(torch.device("cpu")).numpy()
br_xs_pr = (br_xs + br_reg[..., 0]).squeeze(0).to(torch.device("cpu")).numpy()
br_ys_pr = (br_ys + br_reg[..., 1]).squeeze(0).to(torch.device("cpu")).numpy()
ct_xs_pr = (ct_xs + ct_reg[..., 0]).squeeze(0).to(torch.device("cpu")).numpy()
ct_ys_pr = (ct_ys + ct_reg[..., 1]).squeeze(0).to(torch.device("cpu")).numpy()
tl_xs_gt = (tl_xs + targets['tl_reg'][..., 0].to(torch.device("cuda"))).squeeze(0).to(torch.device("cpu")).numpy()
tl_ys_gt = (tl_ys + targets['tl_reg'][..., 1].to(torch.device("cuda"))).squeeze(0).to(torch.device("cpu")).numpy()
bl_xs_gt = (bl_xs + targets['bl_reg'][..., 0].to(torch.device("cuda"))).squeeze(0).to(torch.device("cpu")).numpy()
bl_ys_gt = (bl_ys + targets['bl_reg'][..., 1].to(torch.device("cuda"))).squeeze(0).to(torch.device("cpu")).numpy()
br_xs_gt = (br_xs + targets['br_reg'][..., 0].to(torch.device("cuda"))).squeeze(0).to(torch.device("cpu")).numpy()
br_ys_gt = (br_ys + targets['br_reg'][..., 1].to(torch.device("cuda"))).squeeze(0).to(torch.device("cpu")).numpy()
ct_xs_gt = (ct_xs + targets['ct_reg'][..., 0].to(torch.device("cuda"))).squeeze(0).to(torch.device("cpu")).numpy()
ct_ys_gt = (ct_ys + targets['ct_reg'][..., 1].to(torch.device("cuda"))).squeeze(0).to(torch.device("cpu")).numpy()
bboxes_gt = targets['bbox'][targets['reg_mask']]
nm_instances =tl_xs_pr.shape[0]
for i in range(nm_instances):
bbox_gt = bboxes_gt[i, :]
# prediction
bbox_coord_pr = []
tl_x_pr = tl_xs_pr[i]
tl_y_pr = tl_ys_pr[i]
bl_x_pr = bl_xs_pr[i]
bl_y_pr = bl_ys_pr[i]
br_x_pr = br_xs_pr[i]
br_y_pr = br_ys_pr[i]
# center
x_c = (tl_x_pr + br_x_pr) / 2.
y_c = (tl_y_pr + br_y_pr) / 2.
if bl_x_pr == br_x_pr:
p_y = tl_y_pr
p_x = br_x_pr
if br_y_pr > bl_y_pr:
angle = np.pi / 2.
else:
angle = -np.pi / 2.
elif bl_y_pr == br_y_pr:
p_x = tl_x_pr
p_y = br_y_pr
angle = 0.
else:
# angle
angle = math.atan2(-(br_y_pr - bl_y_pr), br_x_pr - bl_x_pr)
# find intersected point
a = (br_x_pr - bl_x_pr) / (br_y_pr - bl_y_pr)
b = br_y_pr - a * br_x_pr
delta_x = br_x_pr - bl_x_pr
delta_y = br_y_pr - bl_y_pr
p_x = (delta_x * tl_x_pr + delta_y * tl_y_pr - delta_x * b) / (delta_x + delta_x * a)
p_y = a * p_x + b
# w, h
w = np.sqrt((br_x_pr - p_x) * (br_x_pr - p_x) + (br_y_pr - p_y) * (br_y_pr - p_y))
h = np.sqrt((tl_x_pr - p_x) * (tl_x_pr - p_x) + (tl_y_pr - p_y) * (tl_y_pr - p_y))
bbox_coord_pr.append([x_c - w / 2, y_c - h / 2, x_c + w / 2, y_c + h / 2, angle])
bbox_coord_pr = np.array(bbox_coord_pr)
# groundtruth
boxes_coord_gt = []
tl_x_gt = tl_xs_gt[i]
tl_y_gt = tl_ys_gt[i]
bl_x_gt = bl_xs_gt[i]
bl_y_gt = bl_ys_gt[i]
br_x_gt = br_xs_gt[i]
br_y_gt = br_ys_gt[i]
if bl_x_gt == br_x_gt:
p_y = tl_y_gt
p_x = bl_x_gt
if br_y_gt > bl_y_gt:
angle = np.pi / 4
else:
angle = -np.pi / 4
else:
# center
x_c = (tl_x_gt + br_x_gt) / 2.
y_c = (tl_y_gt + br_y_gt) / 2.
# angle
angle = math.atan(-(br_y_gt - bl_y_gt)/(br_x_gt - bl_x_gt))
# find intersected point
a = (br_y_gt - bl_y_gt) / (br_x_gt - bl_x_gt)
b = br_y_gt - a * br_x_gt
delta_x = br_x_gt - bl_x_gt
delta_y = br_y_gt - bl_y_gt
p_x = (delta_x * tl_x_gt + delta_y * tl_y_gt - delta_y * b) / (delta_x + delta_y * a)
p_y = a * p_x + b
# w, h
w = np.sqrt((br_x_gt - p_x) * (br_x_gt - p_x) + (br_y_gt - p_y) * (br_y_gt - p_y))
h = np.sqrt((tl_x_gt - p_x) * (tl_x_gt - p_x) + (tl_y_gt - p_y) * (tl_y_gt - p_y))
boxes_coord_gt.append([x_c - w / 2, y_c - h / 2, x_c + w / 2, y_c + h / 2, angle])
boxes_coord_gt = np.array(boxes_coord_gt)
# print(np.array_equal(bbox_gt, boxes_coord_gt))
overlaps = _bbox_overlaps(np.ascontiguousarray(bbox_coord_pr[:, :4], dtype=np.float32),
np.ascontiguousarray(boxes_coord_gt[:, :4], dtype=np.float32),
bbox_coord_pr[:, -1], boxes_coord_gt[:, -1],
128, 128)
flag_suc = False
flag_exit = 0
for i in range(overlaps.shape[0]):
for j in range(overlaps.shape[1]):
value_overlap = overlaps[i, j]
angle_diff = math.fabs(bbox_coord_pr[i, -1] - boxes_coord_gt[j, -1])
if value_overlap > 0.25 and angle_diff < np.pi / 6:
flag_suc = True
flag_exit = 1
break
if flag_exit:
break
if flag_exit:
break
return min_tl, max_tl, min_bl, max_bl, min_br, max_br, dists_avg, flag_suc
def process(self, images, kernel=1, ae_threshold=1, K=100, num_dets=100):
with torch.no_grad():
output = self.model(images)[-1]
tl_heat = output['tl'].sigmoid_()
bl_heat = output['bl'].sigmoid_()
br_heat = output['br'].sigmoid_()
ct_heat = output['ct'].sigmoid_()
tl_tag = output['tl_tag']
bl_tag = output['bl_tag']
br_tag = output['br_tag']
tl_reg = output['tl_reg']
bl_reg = output['bl_reg']
br_reg = output['br_reg']
ct_reg = output['ct_reg']
batch, cat, height, width = tl_heat.size()
tl_heat = _nms(tl_heat, kernel=3)
bl_heat = _nms(bl_heat, kernel=3)
br_heat = _nms(br_heat, kernel=3)
ct_heat = _nms(ct_heat, kernel=3)
tl_scores, tl_inds, tl_clses, tl_ys, tl_xs = _topk(tl_heat, K=K)
bl_scores, bl_inds, bl_clses, bl_ys, bl_xs = _topk(bl_heat, K=K)
br_scores, br_inds, br_clses, br_ys, br_xs = _topk(br_heat, K=K)
ct_scores, ct_inds, ct_clses, ct_ys, ct_xs = _topk(ct_heat, K=K)
tl_ys = tl_ys.view(batch, K, 1, 1).expand(batch, K, K, K)
tl_xs = tl_xs.view(batch, K, 1, 1).expand(batch, K, K, K)
bl_ys = bl_ys.view(batch, 1, K, 1).expand(batch, K, K, K)
bl_xs = bl_xs.view(batch, 1, K, 1).expand(batch, K, K, K)
br_ys = br_ys.view(batch, 1, 1, K).expand(batch, K, K, K)
br_xs = br_xs.view(batch, 1, 1, K).expand(batch, K, K, K)
ct_ys = ct_ys.view(batch, 1, K).expand(batch, K, K)
ct_xs = ct_xs.view(batch, 1, K).expand(batch, K, K)
if tl_reg is not None and bl_reg is not None and br_reg is not None:
tl_reg = _tranpose_and_gather_feat(tl_reg, tl_inds)
tl_reg = tl_reg.view(batch, K, 1, 1, 2)
bl_reg = _tranpose_and_gather_feat(bl_reg, bl_inds)
bl_reg = bl_reg.view(batch, 1, K, 1, 2)
br_reg = _tranpose_and_gather_feat(br_reg, br_inds)
br_reg = br_reg.view(batch, 1, 1, K, 2)
ct_reg = _tranpose_and_gather_feat(ct_reg, ct_inds)
ct_reg = ct_reg.view(batch, 1, K, 2)
tl_xs = tl_xs + tl_reg[..., 0]
tl_ys = tl_ys + tl_reg[..., 1]
bl_xs = bl_xs + bl_reg[..., 0]
bl_ys = bl_ys + bl_reg[..., 1]
br_xs = br_xs + br_reg[..., 0]
br_ys = br_ys + br_reg[..., 1]
ct_xs = ct_xs + ct_reg[..., 0]
ct_ys = ct_ys + ct_reg[..., 1]
# all possible boxes based on top k corners (ignoring class)
bboxes = torch.stack((tl_xs, tl_ys, bl_xs, bl_ys, br_xs, br_ys), dim=4)
tl_tag = _tranpose_and_gather_feat(tl_tag, tl_inds)
tl_tag = tl_tag.view(batch, K, 1, 1)
bl_tag = _tranpose_and_gather_feat(bl_tag, bl_inds)
bl_tag = bl_tag.view(batch, 1, K, 1)
br_tag = _tranpose_and_gather_feat(br_tag, br_inds)
br_tag = br_tag.view(batch, 1, 1, K)
avg_tag = (tl_tag + bl_tag + br_tag) / 3
dists = (torch.abs(tl_tag - avg_tag) + torch.abs(bl_tag - avg_tag) + torch.abs(br_tag - avg_tag)) / 3
tl_scores = tl_scores.view(batch, K, 1, 1).expand(batch, K, K, K)
bl_scores = bl_scores.view(batch, 1, K, 1).expand(batch, K, K, K)
br_scores = br_scores.view(batch, 1, 1, K).expand(batch, K, K, K)
# reject boxes based on corner scores
# sc_inds = (tl_scores < scores_thresh) | (bl_scores < scores_thresh) | (br_scores < scores_thresh)
scores = (tl_scores + bl_scores + br_scores) / 3
# reject boxes based on classes
tl_clses = tl_clses.view(batch, K, 1, 1).expand(batch, K, K, K)
bl_clses = bl_clses.view(batch, 1, K, 1).expand(batch, K, K, K)
br_clses = br_clses.view(batch, 1, 1, K).expand(batch, K, K, K)
cls_inds = (tl_clses != bl_clses) | (bl_clses != br_clses) | (tl_clses != br_clses)
# reject boxes based on distances
dist_inds = (dists > ae_threshold)
scores[cls_inds] = -1
scores[dist_inds] = -1
# scores[sc_inds] = -1
scores = scores.view(batch, -1)
scores, inds = torch.topk(scores, num_dets)
scores = scores.unsqueeze(2)
bboxes = bboxes.view(batch, -1, 6)
bboxes = _gather_feat(bboxes, inds)
clses = bl_clses.contiguous().view(batch, -1, 1)
clses = _gather_feat(clses, inds).float()
tl_scores = tl_scores.contiguous().view(batch, -1, 1)
tl_scores = _gather_feat(tl_scores, inds).float()
bl_scores = bl_scores.contiguous().view(batch, -1, 1)
bl_scores = _gather_feat(bl_scores, inds).float()
br_scores = br_scores.contiguous().view(batch, -1, 1)
br_scores = _gather_feat(br_scores, inds).float()
ct_xs = ct_xs[:, 0, :]
ct_ys = ct_ys[:, 0, :]
centers = torch.cat([ct_xs.unsqueeze(2), ct_ys.unsqueeze(2), ct_clses.float().unsqueeze(2), ct_scores.unsqueeze(2)], dim=2)
detections = torch.cat([bboxes, scores, tl_scores, bl_scores, br_scores, clses], dim=2)
# tl_heat = output['tl'].sigmoid_()
# bl_heat = output['bl'].sigmoid_()
# br_heat = output['br'].sigmoid_()
# ct_heat = output['ct'].sigmoid_()
#
# tl_tag = output['tl_tag']
# bl_tag = output['bl_tag']
# br_tag = output['br_tag']
#
# tl_reg = output['tl_reg']
# bl_reg = output['bl_reg']
# br_reg = output['br_reg']
# ct_reg = output['ct_reg']
#
# kernel = self.opt.nms_kernel
# ae_threshold = self.opt.ae_threshold
# K = self.opt.K
#
# batch, cat, height, width = tl_heat.size()
#
# # perform nms on heatmaps
# tl_heat = _nms(tl_heat, kernel=kernel)
# bl_heat = _nms(bl_heat, kernel=kernel)
# br_heat = _nms(br_heat, kernel=kernel)
# ct_heat = _nms(ct_heat, kernel=kernel)
#
# tl_scores, tl_inds, tl_clses, tl_ys, tl_xs = _topk(tl_heat, K=K)
# bl_scores, bl_inds, bl_clses, bl_ys, bl_xs = _topk(bl_heat, K=K)
# br_scores, br_inds, br_clses, br_ys, br_xs = _topk(br_heat, K=K)
# ct_scores, ct_inds, ct_clses, ct_ys, ct_xs = _topk(ct_heat, K=K)
#
# tl_ys = tl_ys.view(batch, K, 1, 1).expand(batch, K, K, K)
# tl_xs = tl_xs.view(batch, K, 1, 1).expand(batch, K, K, K)
# bl_ys = bl_ys.view(batch, 1, K, 1).expand(batch, K, K, K)
# bl_xs = bl_xs.view(batch, 1, K, 1).expand(batch, K, K, K)
# br_ys = br_ys.view(batch, 1, 1, K).expand(batch, K, K, K)
# br_xs = br_xs.view(batch, 1, 1, K).expand(batch, K, K, K)
# ct_ys = ct_ys.view(batch, 1, K).expand(batch, K, K)
# ct_xs = ct_xs.view(batch, 1, K).expand(batch, K, K)
#
# if tl_reg is not None and bl_reg is not None and br_reg is not None:
# tl_reg = _tranpose_and_gather_feat(tl_reg, tl_inds)
# tl_reg = tl_reg.view(batch, K, 1, 1, 2)
# bl_reg = _tranpose_and_gather_feat(bl_reg, bl_inds)
# bl_reg = bl_reg.view(batch, 1, K, 1, 2)
# br_reg = _tranpose_and_gather_feat(br_reg, br_inds)
# br_reg = br_reg.view(batch, 1, 1, K, 2)
# ct_reg = _tranpose_and_gather_feat(ct_reg, ct_inds)
# ct_reg = ct_reg.view(batch, 1, K, 2)
#
# tl_xs = tl_xs + tl_reg[..., 0]
# tl_ys = tl_ys + tl_reg[..., 1]
# bl_xs = bl_xs + bl_reg[..., 0]
# bl_ys = bl_ys + bl_reg[..., 1]
# br_xs = br_xs + br_reg[..., 0]
# br_ys = br_ys + br_reg[..., 1]
# ct_xs = ct_xs + ct_reg[..., 0]
# ct_ys = ct_ys + ct_reg[..., 1]
#
# # all possible boxes based on top k corners (ignoring class)
# bboxes = torch.stack((tl_xs, tl_ys, bl_xs, bl_ys, br_xs, br_ys), dim=4)
#
# tl_tag = _tranpose_and_gather_feat(tl_tag, tl_inds)
# tl_tag = tl_tag.view(batch, K, 1, 1).expand(batch, K, K, K)
# bl_tag = _tranpose_and_gather_feat(bl_tag, bl_inds)
# bl_tag = bl_tag.view(batch, 1, K, 1).expand(batch, K, K, K)
# br_tag = _tranpose_and_gather_feat(br_tag, br_inds)
# br_tag = br_tag.view(batch, 1, 1, K).expand(batch, K, K, K)
# avg_tag = (tl_tag + bl_tag + br_tag) / 3
# dists = (torch.abs(tl_tag - avg_tag) + torch.abs(bl_tag - avg_tag) + torch.abs(br_tag - avg_tag)) / 3
#
# tl_scores = tl_scores.view(batch, K, 1, 1).expand(batch, K, K, K)
# bl_scores = bl_scores.view(batch, 1, K, 1).expand(batch, K, K, K)
# br_scores = br_scores.view(batch, 1, 1, K).expand(batch, K, K, K)
# scores = (tl_scores + bl_scores + br_scores) / 3
#
# # reject boxes based on classes
# tl_clses = tl_clses.view(batch, K, 1, 1).expand(batch, K, K, K)
# bl_clses = bl_clses.view(batch, 1, K, 1).expand(batch, K, K, K)
# br_clses = br_clses.view(batch, 1, 1, K).expand(batch, K, K, K)
# cls_inds = (tl_clses != bl_clses) | (bl_clses != br_clses) | (tl_clses != br_clses)
#
# # reject boxes based on distances
# dist_inds = (dists > ae_threshold)
#
# # instead of filtering prediction according to the out-of-bound rotation, do data augmentation to mirror groundtruth
#
# scores[cls_inds] = -1
# scores[dist_inds] = -1
#
# scores = scores.view(batch, -1)
# scores, inds = torch.topk(scores, 100)
# scores = scores.unsqueeze(2)
#
# bboxes = bboxes.view(batch, -1, 6)
# bboxes = _gather_feat(bboxes, inds)
#
# tl_tag = tl_tag.contiguous().view(batch, -1, 1)
# tl_tag = _gather_feat(tl_tag, inds)
# bl_tag = bl_tag.contiguous().view(batch, -1, 1)
# bl_tag = _gather_feat(bl_tag, inds)
# br_tag = br_tag.contiguous().view(batch, -1, 1)
# br_tag = _gather_feat(br_tag, inds)
# avg_tag = avg_tag.contiguous().view(batch, -1, 1)
# avg_tag = _gather_feat(avg_tag, inds)
#
# clses = bl_clses.contiguous().view(batch, -1, 1)
# clses = _gather_feat(clses, inds).float()
#
# tl_scores = tl_scores.contiguous().view(batch, -1, 1)
# tl_scores = _gather_feat(tl_scores, inds).float()
# bl_scores = bl_scores.contiguous().view(batch, -1, 1)
# bl_scores = _gather_feat(bl_scores, inds).float()
# br_scores = br_scores.contiguous().view(batch, -1, 1)
# br_scores = _gather_feat(br_scores, inds).float()
#
# ct_xs = ct_xs[:, 0, :]
# ct_ys = ct_ys[:, 0, :]
#
# centers = torch.cat(
# [ct_xs.unsqueeze(2), ct_ys.unsqueeze(2), ct_clses.float().unsqueeze(2), ct_scores.unsqueeze(2)], dim=2)
# detections = torch.cat([bboxes, scores, tl_scores, bl_scores, br_scores, clses, tl_tag, bl_tag, br_tag, avg_tag], dim=2)
return detections, centers
def post_process(self, detections, centers, num_classes, bbox_size_threshold, ori_threshold):
detections = detections.detach().cpu().numpy()
centers = centers.detach().cpu().numpy()
detections = detections.reshape(1, -1, detections.shape[2])
centers = centers.reshape(1, -1, centers.shape[2])
ret = []
for i in range(detections.shape[0]):
top_preds = {}
detections[i, :, 0:2] *= 4.
detections[i, :, 2:4] *= 4.
detections[i, :, 4:6] *= 4.
centers[i, :, 0:2] *= 4.
# Dump bbox whose central region has no center point
detections = np.concatenate(detections, axis=1)
centers = np.concatenate(centers, axis=1)
# filter by orientation distance between quantized and continuous predicted angle
classes = detections[..., -1]
quant_ori = (5.0 * classes - 85.0) / 180 * np.pi
bl_x = detections[..., 2]
bl_y = detections[..., 3]
br_x = detections[..., 4]
br_y = detections[..., 5]
cont_ori = np.arctan(-(br_y - bl_y) / (br_x - bl_x))
dist_ori = np.fabs(quant_ori - cont_ori)
ori_ind = dist_ori < ori_threshold
valid_detections = detections[ori_ind]
valid_ind = valid_detections[:, 6] > -1
valid_detections = valid_detections[valid_ind]
# valid_ind = detections[:, 6] > -1
# valid_detections = detections[valid_ind]
box_width = np.sqrt(np.power(valid_detections[:, 2] - valid_detections[:, 4], 2) + \
np.power(valid_detections[:, 3] - valid_detections[:, 5], 2))
box_height = np.sqrt(np.power(valid_detections[:, 2] - valid_detections[:, 0], 2) + \
np.power(valid_detections[:, 3] - valid_detections[:, 1], 2))
s_ind = (box_width * box_height <= bbox_size_threshold)
l_ind = (box_width * box_height > bbox_size_threshold)
s_detections = valid_detections[s_ind]
l_detections = valid_detections[l_ind]
# pro-process for small bounding box
s_tl_x = (2 * s_detections[:, 0] + s_detections[:, 4]) / 3
s_br_x = (s_detections[:, 0] + 2 * s_detections[:, 4]) / 3
s_tl_y = (2 * s_detections[:, 1] + s_detections[:, 5]) / 3
s_br_y = (s_detections[:, 1] + 2 * s_detections[:, 5]) / 3
s_temp_score = copy.copy(s_detections[:, 6])
s_detections[:, 6] = -1
center_x = centers[:, 0][:, np.newaxis]
center_y = centers[:, 1][:, np.newaxis]
s_tl_x = s_tl_x[np.newaxis, :]
s_br_x = s_br_x[np.newaxis, :]
s_tl_y = s_tl_y[np.newaxis, :]
s_br_y = s_br_y[np.newaxis, :]
ind_x1 = (center_x > s_tl_x) & (center_x < s_br_x)
ind_x2 = (center_x < s_tl_x) & (center_x > s_br_x)
ind_y1 = (center_y > s_tl_y) & (center_y < s_br_y)
ind_y2 = (center_y < s_tl_y) & (center_y > s_br_y)
ind_cls = (centers[:, 2][:, np.newaxis] - s_detections[:, -1][np.newaxis, :]) == 0
ind_s_new_score = np.max((((ind_x1 + 0) & (ind_y1 + 0) & (ind_cls + 0)) |
((ind_x1 + 0) & (ind_y2 + 0) & (ind_cls + 0)) |
((ind_x2 + 0) & (ind_y2 + 0) & (ind_cls + 0))), axis=0) == 1
index_s_new_score = np.argmax((((ind_x1 + 0) & (ind_y1 + 0) & (ind_cls + 0)) |
((ind_x1 + 0) & (ind_y2 + 0) & (ind_cls + 0)) |
((ind_x2 + 0) & (ind_y2 + 0) & (ind_cls + 0)))[:, ind_s_new_score], axis=0)
s_corner_score = s_temp_score[ind_s_new_score]
s_center_score = centers[index_s_new_score, 3]
s_detections[:, 6][ind_s_new_score] = (s_corner_score * 3 + s_center_score) / 4
# pro-process for large bounding box
l_tl_x = (2 * l_detections[:, 0] + l_detections[:, 4]) / 3
l_br_x = (l_detections[:, 0] + 2 * l_detections[:, 4]) / 3
l_tl_y = (2 * l_detections[:, 1] + l_detections[:, 5]) / 3
l_br_y = (l_detections[:, 1] + 2 * l_detections[:, 5]) / 3
l_temp_score = copy.copy(l_detections[:, 6])
l_detections[:, 6] = -1
center_x = centers[:, 0][:, np.newaxis]
center_y = centers[:, 1][:, np.newaxis]
l_tl_x = l_tl_x[np.newaxis, :]
l_br_x = l_br_x[np.newaxis, :]
l_tl_y = l_tl_y[np.newaxis, :]
l_br_y = l_br_y[np.newaxis, :]
ind_x1 = (center_x > l_tl_x) & (center_x < l_br_x)
ind_x2 = (center_x < l_tl_x) & (center_x > l_br_x)
ind_y1 = (center_y > l_tl_y) & (center_y < l_br_y)
ind_y2 = (center_y < l_tl_y) & (center_y > l_br_y)
ind_cls = (centers[:, 2][:, np.newaxis] - l_detections[:, -1][np.newaxis, :]) == 0
ind_l_new_score = np.max((((ind_x1 + 0) & (ind_y1 + 0) & (ind_cls + 0)) |
((ind_x1 + 0) & (ind_y2 + 0) & (ind_cls + 0)) |
((ind_x2 + 0) & (ind_y2 + 0) & (ind_cls + 0))), axis=0) == 1
index_l_new_score = np.argmax((((ind_x1 + 0) & (ind_y1 + 0) & (ind_cls + 0)) |
((ind_x1 + 0) & (ind_y2 + 0) & (ind_cls + 0)) |
((ind_x2 + 0) & (ind_y2 + 0) & (ind_cls + 0)))[:, ind_l_new_score], axis=0)
l_corner_score = l_temp_score[ind_l_new_score]
l_center_score = centers[index_l_new_score, 3]
l_detections[:, 6][ind_l_new_score] = (l_corner_score * 3 + l_center_score) / 4
detections = np.concatenate([l_detections, s_detections], axis=0)
detections = detections[np.argsort(-detections[:, 6])]
classes = detections[..., -1]
# reject detections with negative scores
keep_inds = (detections[:, 6] > -1)
detections = detections[keep_inds]
classes = classes[keep_inds]
detections = np.expand_dims(detections, axis=0)
for j in range(num_classes):
inds = (classes == j)
top_preds[j + 1] = detections[i, inds, :].astype(np.float32).tolist()
ret.append(top_preds)
for j in range(1, num_classes + 1):
ret[0][j] = np.array(ret[0][j], dtype=np.float32).reshape(-1, 11)
return ret[0]
def merge_outputs(self, detections):
results = {}
for j in range(1, self.num_classes + 1):
results[j] = np.concatenate(
[detection[j] for detection in detections], axis=0).astype(np.float32)
if len(self.scales) > 1 or self.opt.nms:
soft_nms(results[j], Nt=0.5, method=2)
scores = np.hstack(
[results[j][:, 6] for j in range(1, self.num_classes + 1)])
if len(scores) > self.max_per_image:
kth = len(scores) - self.max_per_image
thresh = np.partition(scores, kth)[kth]
for j in range(1, self.num_classes + 1):
keep_inds = (results[j][:, 6] >= thresh)
results[j] = results[j][keep_inds]
return results
| [
"numpy.sqrt",
"models.decode._topk",
"models.utils._tranpose_and_gather_feat",
"numpy.ascontiguousarray",
"numpy.argsort",
"numpy.array",
"copy.copy",
"math.atan",
"numpy.partition",
"numpy.max",
"math.fabs",
"numpy.concatenate",
"numpy.arctan",
"torch.abs",
"torch.topk",
"numpy.argmax... | [((626, 641), 'torch.no_grad', 'torch.no_grad', ([], {}), '()\n', (639, 641), False, 'import torch\n'), ((5076, 5096), 'torch.device', 'torch.device', (['"""cuda"""'], {}), "('cuda')\n", (5088, 5096), False, 'import torch\n'), ((5178, 5198), 'torch.device', 'torch.device', (['"""cuda"""'], {}), "('cuda')\n", (5190, 5198), False, 'import torch\n'), ((5280, 5300), 'torch.device', 'torch.device', (['"""cuda"""'], {}), "('cuda')\n", (5292, 5300), False, 'import torch\n'), ((5382, 5402), 'torch.device', 'torch.device', (['"""cuda"""'], {}), "('cuda')\n", (5394, 5402), False, 'import torch\n'), ((5484, 5504), 'torch.device', 'torch.device', (['"""cuda"""'], {}), "('cuda')\n", (5496, 5504), False, 'import torch\n'), ((5586, 5606), 'torch.device', 'torch.device', (['"""cuda"""'], {}), "('cuda')\n", (5598, 5606), False, 'import torch\n'), ((5688, 5708), 'torch.device', 'torch.device', (['"""cuda"""'], {}), "('cuda')\n", (5700, 5708), False, 'import torch\n'), ((5790, 5810), 'torch.device', 'torch.device', (['"""cuda"""'], {}), "('cuda')\n", (5802, 5810), False, 'import torch\n'), ((8815, 8893), 'numpy.sqrt', 'np.sqrt', (['((br_x_pr - p_x) * (br_x_pr - p_x) + (br_y_pr - p_y) * (br_y_pr - p_y))'], {}), '((br_x_pr - p_x) * (br_x_pr - p_x) + (br_y_pr - p_y) * (br_y_pr - p_y))\n', (8822, 8893), True, 'import numpy as np\n'), ((8908, 8986), 'numpy.sqrt', 'np.sqrt', (['((tl_x_pr - p_x) * (tl_x_pr - p_x) + (tl_y_pr - p_y) * (tl_y_pr - p_y))'], {}), '((tl_x_pr - p_x) * (tl_x_pr - p_x) + (tl_y_pr - p_y) * (tl_y_pr - p_y))\n', (8915, 8986), True, 'import numpy as np\n'), ((9106, 9129), 'numpy.array', 'np.array', (['bbox_coord_pr'], {}), '(bbox_coord_pr)\n', (9114, 9129), True, 'import numpy as np\n'), ((10210, 10288), 'numpy.sqrt', 'np.sqrt', (['((br_x_gt - p_x) * (br_x_gt - p_x) + (br_y_gt - p_y) * (br_y_gt - p_y))'], {}), '((br_x_gt - p_x) * (br_x_gt - p_x) + (br_y_gt - p_y) * (br_y_gt - p_y))\n', (10217, 10288), True, 'import numpy as np\n'), ((10303, 10381), 'numpy.sqrt', 'np.sqrt', (['((tl_x_gt - p_x) * (tl_x_gt - p_x) + (tl_y_gt - p_y) * (tl_y_gt - p_y))'], {}), '((tl_x_gt - p_x) * (tl_x_gt - p_x) + (tl_y_gt - p_y) * (tl_y_gt - p_y))\n', (10310, 10381), True, 'import numpy as np\n'), ((10502, 10526), 'numpy.array', 'np.array', (['boxes_coord_gt'], {}), '(boxes_coord_gt)\n', (10510, 10526), True, 'import numpy as np\n'), ((11632, 11647), 'torch.no_grad', 'torch.no_grad', ([], {}), '()\n', (11645, 11647), False, 'import torch\n'), ((12197, 12220), 'models.decode._nms', '_nms', (['tl_heat'], {'kernel': '(3)'}), '(tl_heat, kernel=3)\n', (12201, 12220), False, 'from models.decode import _topk_original, _topk, _topk_channel, _nms\n'), ((12241, 12264), 'models.decode._nms', '_nms', (['bl_heat'], {'kernel': '(3)'}), '(bl_heat, kernel=3)\n', (12245, 12264), False, 'from models.decode import _topk_original, _topk, _topk_channel, _nms\n'), ((12285, 12308), 'models.decode._nms', '_nms', (['br_heat'], {'kernel': '(3)'}), '(br_heat, kernel=3)\n', (12289, 12308), False, 'from models.decode import _topk_original, _topk, _topk_channel, _nms\n'), ((12329, 12352), 'models.decode._nms', '_nms', (['ct_heat'], {'kernel': '(3)'}), '(ct_heat, kernel=3)\n', (12333, 12352), False, 'from models.decode import _topk_original, _topk, _topk_channel, _nms\n'), ((12409, 12428), 'models.decode._topk', '_topk', (['tl_heat'], {'K': 'K'}), '(tl_heat, K=K)\n', (12414, 12428), False, 'from models.decode import _topk_original, _topk, _topk_channel, _nms\n'), ((12484, 12503), 'models.decode._topk', '_topk', (['bl_heat'], {'K': 'K'}), '(bl_heat, K=K)\n', (12489, 12503), False, 'from models.decode import _topk_original, _topk, _topk_channel, _nms\n'), ((12559, 12578), 'models.decode._topk', '_topk', (['br_heat'], {'K': 'K'}), '(br_heat, K=K)\n', (12564, 12578), False, 'from models.decode import _topk_original, _topk, _topk_channel, _nms\n'), ((12634, 12653), 'models.decode._topk', '_topk', (['ct_heat'], {'K': 'K'}), '(ct_heat, K=K)\n', (12639, 12653), False, 'from models.decode import _topk_original, _topk, _topk_channel, _nms\n'), ((14196, 14258), 'torch.stack', 'torch.stack', (['(tl_xs, tl_ys, bl_xs, bl_ys, br_xs, br_ys)'], {'dim': '(4)'}), '((tl_xs, tl_ys, bl_xs, bl_ys, br_xs, br_ys), dim=4)\n', (14207, 14258), False, 'import torch\n'), ((14279, 14321), 'models.utils._tranpose_and_gather_feat', '_tranpose_and_gather_feat', (['tl_tag', 'tl_inds'], {}), '(tl_tag, tl_inds)\n', (14304, 14321), False, 'from models.utils import _tranpose_and_gather_feat, _gather_feat\n'), ((14388, 14430), 'models.utils._tranpose_and_gather_feat', '_tranpose_and_gather_feat', (['bl_tag', 'bl_inds'], {}), '(bl_tag, bl_inds)\n', (14413, 14430), False, 'from models.utils import _tranpose_and_gather_feat, _gather_feat\n'), ((14497, 14539), 'models.utils._tranpose_and_gather_feat', '_tranpose_and_gather_feat', (['br_tag', 'br_inds'], {}), '(br_tag, br_inds)\n', (14522, 14539), False, 'from models.utils import _tranpose_and_gather_feat, _gather_feat\n'), ((15812, 15840), 'torch.topk', 'torch.topk', (['scores', 'num_dets'], {}), '(scores, num_dets)\n', (15822, 15840), False, 'import torch\n'), ((15945, 15971), 'models.utils._gather_feat', '_gather_feat', (['bboxes', 'inds'], {}), '(bboxes, inds)\n', (15957, 15971), False, 'from models.utils import _tranpose_and_gather_feat, _gather_feat\n'), ((16681, 16755), 'torch.cat', 'torch.cat', (['[bboxes, scores, tl_scores, bl_scores, br_scores, clses]'], {'dim': '(2)'}), '([bboxes, scores, tl_scores, bl_scores, br_scores, clses], dim=2)\n', (16690, 16755), False, 'import torch\n'), ((23561, 23595), 'numpy.concatenate', 'np.concatenate', (['detections'], {'axis': '(1)'}), '(detections, axis=1)\n', (23575, 23595), True, 'import numpy as np\n'), ((23614, 23645), 'numpy.concatenate', 'np.concatenate', (['centers'], {'axis': '(1)'}), '(centers, axis=1)\n', (23628, 23645), True, 'import numpy as np\n'), ((23987, 24028), 'numpy.arctan', 'np.arctan', (['(-(br_y - bl_y) / (br_x - bl_x))'], {}), '(-(br_y - bl_y) / (br_x - bl_x))\n', (23996, 24028), True, 'import numpy as np\n'), ((24048, 24077), 'numpy.fabs', 'np.fabs', (['(quant_ori - cont_ori)'], {}), '(quant_ori - cont_ori)\n', (24055, 24077), True, 'import numpy as np\n'), ((25298, 25327), 'copy.copy', 'copy.copy', (['s_detections[:, 6]'], {}), '(s_detections[:, 6])\n', (25307, 25327), False, 'import copy\n'), ((26228, 26395), 'numpy.argmax', 'np.argmax', (['(ind_x1 + 0 & ind_y1 + 0 & ind_cls + 0 | ind_x1 + 0 & ind_y2 + 0 & ind_cls +\n 0 | ind_x2 + 0 & ind_y2 + 0 & ind_cls + 0)[:, ind_s_new_score]'], {'axis': '(0)'}), '((ind_x1 + 0 & ind_y1 + 0 & ind_cls + 0 | ind_x1 + 0 & ind_y2 + 0 &\n ind_cls + 0 | ind_x2 + 0 & ind_y2 + 0 & ind_cls + 0)[:, ind_s_new_score\n ], axis=0)\n', (26237, 26395), True, 'import numpy as np\n'), ((27025, 27054), 'copy.copy', 'copy.copy', (['l_detections[:, 6]'], {}), '(l_detections[:, 6])\n', (27034, 27054), False, 'import copy\n'), ((27955, 28122), 'numpy.argmax', 'np.argmax', (['(ind_x1 + 0 & ind_y1 + 0 & ind_cls + 0 | ind_x1 + 0 & ind_y2 + 0 & ind_cls +\n 0 | ind_x2 + 0 & ind_y2 + 0 & ind_cls + 0)[:, ind_l_new_score]'], {'axis': '(0)'}), '((ind_x1 + 0 & ind_y1 + 0 & ind_cls + 0 | ind_x1 + 0 & ind_y2 + 0 &\n ind_cls + 0 | ind_x2 + 0 & ind_y2 + 0 & ind_cls + 0)[:, ind_l_new_score\n ], axis=0)\n', (27964, 28122), True, 'import numpy as np\n'), ((28436, 28488), 'numpy.concatenate', 'np.concatenate', (['[l_detections, s_detections]'], {'axis': '(0)'}), '([l_detections, s_detections], axis=0)\n', (28450, 28488), True, 'import numpy as np\n'), ((28786, 28820), 'numpy.expand_dims', 'np.expand_dims', (['detections'], {'axis': '(0)'}), '(detections, axis=0)\n', (28800, 28820), True, 'import numpy as np\n'), ((2861, 2881), 'torch.device', 'torch.device', (['"""cuda"""'], {}), "('cuda')\n", (2873, 2881), False, 'import torch\n'), ((2954, 2974), 'torch.device', 'torch.device', (['"""cuda"""'], {}), "('cuda')\n", (2966, 2974), False, 'import torch\n'), ((3047, 3067), 'torch.device', 'torch.device', (['"""cuda"""'], {}), "('cuda')\n", (3059, 3067), False, 'import torch\n'), ((3919, 3939), 'torch.device', 'torch.device', (['"""cuda"""'], {}), "('cuda')\n", (3931, 3939), False, 'import torch\n'), ((4012, 4032), 'torch.device', 'torch.device', (['"""cuda"""'], {}), "('cuda')\n", (4024, 4032), False, 'import torch\n'), ((4105, 4125), 'torch.device', 'torch.device', (['"""cuda"""'], {}), "('cuda')\n", (4117, 4125), False, 'import torch\n'), ((4198, 4218), 'torch.device', 'torch.device', (['"""cuda"""'], {}), "('cuda')\n", (4210, 4218), False, 'import torch\n'), ((9768, 9821), 'math.atan', 'math.atan', (['(-(br_y_gt - bl_y_gt) / (br_x_gt - bl_x_gt))'], {}), '(-(br_y_gt - bl_y_gt) / (br_x_gt - bl_x_gt))\n', (9777, 9821), False, 'import math\n'), ((10623, 10683), 'numpy.ascontiguousarray', 'np.ascontiguousarray', (['bbox_coord_pr[:, :4]'], {'dtype': 'np.float32'}), '(bbox_coord_pr[:, :4], dtype=np.float32)\n', (10643, 10683), True, 'import numpy as np\n'), ((10721, 10782), 'numpy.ascontiguousarray', 'np.ascontiguousarray', (['boxes_coord_gt[:, :4]'], {'dtype': 'np.float32'}), '(boxes_coord_gt[:, :4], dtype=np.float32)\n', (10741, 10782), True, 'import numpy as np\n'), ((13290, 13332), 'models.utils._tranpose_and_gather_feat', '_tranpose_and_gather_feat', (['tl_reg', 'tl_inds'], {}), '(tl_reg, tl_inds)\n', (13315, 13332), False, 'from models.utils import _tranpose_and_gather_feat, _gather_feat\n'), ((13410, 13452), 'models.utils._tranpose_and_gather_feat', '_tranpose_and_gather_feat', (['bl_reg', 'bl_inds'], {}), '(bl_reg, bl_inds)\n', (13435, 13452), False, 'from models.utils import _tranpose_and_gather_feat, _gather_feat\n'), ((13530, 13572), 'models.utils._tranpose_and_gather_feat', '_tranpose_and_gather_feat', (['br_reg', 'br_inds'], {}), '(br_reg, br_inds)\n', (13555, 13572), False, 'from models.utils import _tranpose_and_gather_feat, _gather_feat\n'), ((13650, 13692), 'models.utils._tranpose_and_gather_feat', '_tranpose_and_gather_feat', (['ct_reg', 'ct_inds'], {}), '(ct_reg, ct_inds)\n', (13675, 13692), False, 'from models.utils import _tranpose_and_gather_feat, _gather_feat\n'), ((25967, 26105), 'numpy.max', 'np.max', (['(ind_x1 + 0 & ind_y1 + 0 & ind_cls + 0 | ind_x1 + 0 & ind_y2 + 0 & ind_cls +\n 0 | ind_x2 + 0 & ind_y2 + 0 & ind_cls + 0)'], {'axis': '(0)'}), '(ind_x1 + 0 & ind_y1 + 0 & ind_cls + 0 | ind_x1 + 0 & ind_y2 + 0 & \n ind_cls + 0 | ind_x2 + 0 & ind_y2 + 0 & ind_cls + 0, axis=0)\n', (25973, 26105), True, 'import numpy as np\n'), ((27694, 27832), 'numpy.max', 'np.max', (['(ind_x1 + 0 & ind_y1 + 0 & ind_cls + 0 | ind_x1 + 0 & ind_y2 + 0 & ind_cls +\n 0 | ind_x2 + 0 & ind_y2 + 0 & ind_cls + 0)'], {'axis': '(0)'}), '(ind_x1 + 0 & ind_y1 + 0 & ind_cls + 0 | ind_x1 + 0 & ind_y2 + 0 & \n ind_cls + 0 | ind_x2 + 0 & ind_y2 + 0 & ind_cls + 0, axis=0)\n', (27700, 27832), True, 'import numpy as np\n'), ((28521, 28550), 'numpy.argsort', 'np.argsort', (['(-detections[:, 6])'], {}), '(-detections[:, 6])\n', (28531, 28550), True, 'import numpy as np\n'), ((29678, 29703), 'numpy.partition', 'np.partition', (['scores', 'kth'], {}), '(scores, kth)\n', (29690, 29703), True, 'import numpy as np\n'), ((3371, 3390), 'torch.device', 'torch.device', (['"""cpu"""'], {}), "('cpu')\n", (3383, 3390), False, 'import torch\n'), ((3448, 3467), 'torch.device', 'torch.device', (['"""cpu"""'], {}), "('cpu')\n", (3460, 3467), False, 'import torch\n'), ((3525, 3544), 'torch.device', 'torch.device', (['"""cpu"""'], {}), "('cpu')\n", (3537, 3544), False, 'import torch\n'), ((5869, 5888), 'torch.device', 'torch.device', (['"""cpu"""'], {}), "('cpu')\n", (5881, 5888), False, 'import torch\n'), ((5954, 5973), 'torch.device', 'torch.device', (['"""cpu"""'], {}), "('cpu')\n", (5966, 5973), False, 'import torch\n'), ((6039, 6058), 'torch.device', 'torch.device', (['"""cpu"""'], {}), "('cpu')\n", (6051, 6058), False, 'import torch\n'), ((6124, 6143), 'torch.device', 'torch.device', (['"""cpu"""'], {}), "('cpu')\n", (6136, 6143), False, 'import torch\n'), ((6209, 6228), 'torch.device', 'torch.device', (['"""cpu"""'], {}), "('cpu')\n", (6221, 6228), False, 'import torch\n'), ((6294, 6313), 'torch.device', 'torch.device', (['"""cpu"""'], {}), "('cpu')\n", (6306, 6313), False, 'import torch\n'), ((6379, 6398), 'torch.device', 'torch.device', (['"""cpu"""'], {}), "('cpu')\n", (6391, 6398), False, 'import torch\n'), ((6464, 6483), 'torch.device', 'torch.device', (['"""cpu"""'], {}), "('cpu')\n", (6476, 6483), False, 'import torch\n'), ((6586, 6605), 'torch.device', 'torch.device', (['"""cpu"""'], {}), "('cpu')\n", (6598, 6605), False, 'import torch\n'), ((6707, 6726), 'torch.device', 'torch.device', (['"""cpu"""'], {}), "('cpu')\n", (6719, 6726), False, 'import torch\n'), ((6828, 6847), 'torch.device', 'torch.device', (['"""cpu"""'], {}), "('cpu')\n", (6840, 6847), False, 'import torch\n'), ((6949, 6968), 'torch.device', 'torch.device', (['"""cpu"""'], {}), "('cpu')\n", (6961, 6968), False, 'import torch\n'), ((7070, 7089), 'torch.device', 'torch.device', (['"""cpu"""'], {}), "('cpu')\n", (7082, 7089), False, 'import torch\n'), ((7191, 7210), 'torch.device', 'torch.device', (['"""cpu"""'], {}), "('cpu')\n", (7203, 7210), False, 'import torch\n'), ((7312, 7331), 'torch.device', 'torch.device', (['"""cpu"""'], {}), "('cpu')\n", (7324, 7331), False, 'import torch\n'), ((7433, 7452), 'torch.device', 'torch.device', (['"""cpu"""'], {}), "('cpu')\n", (7445, 7452), False, 'import torch\n'), ((8373, 8424), 'math.atan2', 'math.atan2', (['(-(br_y_pr - bl_y_pr))', '(br_x_pr - bl_x_pr)'], {}), '(-(br_y_pr - bl_y_pr), br_x_pr - bl_x_pr)\n', (8383, 8424), False, 'import math\n'), ((11137, 11192), 'math.fabs', 'math.fabs', (['(bbox_coord_pr[i, -1] - boxes_coord_gt[j, -1])'], {}), '(bbox_coord_pr[i, -1] - boxes_coord_gt[j, -1])\n', (11146, 11192), False, 'import math\n'), ((14717, 14744), 'torch.abs', 'torch.abs', (['(br_tag - avg_tag)'], {}), '(br_tag - avg_tag)\n', (14726, 14744), False, 'import torch\n'), ((16050, 16075), 'models.utils._gather_feat', '_gather_feat', (['clses', 'inds'], {}), '(clses, inds)\n', (16062, 16075), False, 'from models.utils import _tranpose_and_gather_feat, _gather_feat\n'), ((16170, 16199), 'models.utils._gather_feat', '_gather_feat', (['tl_scores', 'inds'], {}), '(tl_scores, inds)\n', (16182, 16199), False, 'from models.utils import _tranpose_and_gather_feat, _gather_feat\n'), ((16294, 16323), 'models.utils._gather_feat', '_gather_feat', (['bl_scores', 'inds'], {}), '(bl_scores, inds)\n', (16306, 16323), False, 'from models.utils import _tranpose_and_gather_feat, _gather_feat\n'), ((16418, 16447), 'models.utils._gather_feat', '_gather_feat', (['br_scores', 'inds'], {}), '(br_scores, inds)\n', (16430, 16447), False, 'from models.utils import _tranpose_and_gather_feat, _gather_feat\n'), ((24397, 24457), 'numpy.power', 'np.power', (['(valid_detections[:, 2] - valid_detections[:, 4])', '(2)'], {}), '(valid_detections[:, 2] - valid_detections[:, 4], 2)\n', (24405, 24457), True, 'import numpy as np\n'), ((24490, 24550), 'numpy.power', 'np.power', (['(valid_detections[:, 3] - valid_detections[:, 5])', '(2)'], {}), '(valid_detections[:, 3] - valid_detections[:, 5], 2)\n', (24498, 24550), True, 'import numpy as np\n'), ((24581, 24641), 'numpy.power', 'np.power', (['(valid_detections[:, 2] - valid_detections[:, 0])', '(2)'], {}), '(valid_detections[:, 2] - valid_detections[:, 0], 2)\n', (24589, 24641), True, 'import numpy as np\n'), ((24675, 24735), 'numpy.power', 'np.power', (['(valid_detections[:, 3] - valid_detections[:, 1])', '(2)'], {}), '(valid_detections[:, 3] - valid_detections[:, 1], 2)\n', (24683, 24735), True, 'import numpy as np\n'), ((29064, 29101), 'numpy.array', 'np.array', (['ret[0][j]'], {'dtype': 'np.float32'}), '(ret[0][j], dtype=np.float32)\n', (29072, 29101), True, 'import numpy as np\n'), ((29266, 29332), 'numpy.concatenate', 'np.concatenate', (['[detection[j] for detection in detections]'], {'axis': '(0)'}), '([detection[j] for detection in detections], axis=0)\n', (29280, 29332), True, 'import numpy as np\n'), ((3340, 3367), 'torch.abs', 'torch.abs', (['(avg_tag - tl_tag)'], {}), '(avg_tag - tl_tag)\n', (3349, 3367), False, 'import torch\n'), ((3417, 3444), 'torch.abs', 'torch.abs', (['(bl_tag - avg_tag)'], {}), '(bl_tag - avg_tag)\n', (3426, 3444), False, 'import torch\n'), ((3494, 3521), 'torch.abs', 'torch.abs', (['(avg_tag - br_tag)'], {}), '(avg_tag - br_tag)\n', (3503, 3521), False, 'import torch\n'), ((14657, 14684), 'torch.abs', 'torch.abs', (['(tl_tag - avg_tag)'], {}), '(tl_tag - avg_tag)\n', (14666, 14684), False, 'import torch\n'), ((14687, 14714), 'torch.abs', 'torch.abs', (['(bl_tag - avg_tag)'], {}), '(bl_tag - avg_tag)\n', (14696, 14714), False, 'import torch\n'), ((6549, 6569), 'torch.device', 'torch.device', (['"""cuda"""'], {}), "('cuda')\n", (6561, 6569), False, 'import torch\n'), ((6670, 6690), 'torch.device', 'torch.device', (['"""cuda"""'], {}), "('cuda')\n", (6682, 6690), False, 'import torch\n'), ((6791, 6811), 'torch.device', 'torch.device', (['"""cuda"""'], {}), "('cuda')\n", (6803, 6811), False, 'import torch\n'), ((6912, 6932), 'torch.device', 'torch.device', (['"""cuda"""'], {}), "('cuda')\n", (6924, 6932), False, 'import torch\n'), ((7033, 7053), 'torch.device', 'torch.device', (['"""cuda"""'], {}), "('cuda')\n", (7045, 7053), False, 'import torch\n'), ((7154, 7174), 'torch.device', 'torch.device', (['"""cuda"""'], {}), "('cuda')\n", (7166, 7174), False, 'import torch\n'), ((7275, 7295), 'torch.device', 'torch.device', (['"""cuda"""'], {}), "('cuda')\n", (7287, 7295), False, 'import torch\n'), ((7396, 7416), 'torch.device', 'torch.device', (['"""cuda"""'], {}), "('cuda')\n", (7408, 7416), False, 'import torch\n')] |
import logbook
import pandas as pd
import zipline as zl
from datetime import datetime, timedelta
import pathlib
import azul
import numpy as np
from typing import List, Tuple
log = logbook.Logger('BasePriceManager')
class BasePriceManager(object):
def __init__(self, calendar_name='NYSE'):
self._calendar = zl.get_calendar(name=calendar_name)
self._cols = ['open', 'high', 'low', 'close', 'volume', 'dividend', 'split']
# The number of days the price manager will keep trying to pull data for a symbol that is not returning data.
self.MISSING_DATE_THRESHOLD = 5
def get_price_data(self, symbols: List[str], output_dir: str, start_date: datetime, end_date: datetime) -> None:
minute_dir_path = pathlib.Path(output_dir, 'minute')
daily_dir_path = pathlib.Path(output_dir, 'daily')
for ticker in symbols:
self._download_and_process_data(
ticker, start_date, end_date, minute_dir_path, daily_dir_path)
def _download_and_process_data(
self,
ticker: str,
start_date: datetime,
end_date: datetime,
minute_dir_path: pathlib.Path,
daily_dir_path: pathlib.Path
) -> None:
df = self._minute_dataframe_for_dates(ticker, start_date, end_date)
if df.empty:
return
df = self._check_sessions(df, ticker, frequency='minute')
minute_dir_path.mkdir(parents=True, exist_ok=True)
filename = pathlib.Path(minute_dir_path, ticker + '.csv')
df.to_csv(filename)
daily_df = self._resample_minute_data_to_daily_data(df)
daily_df = self._check_sessions(daily_df, ticker, frequency='daily')
daily_dir_path.mkdir(parents=True, exist_ok=True)
filename = pathlib.Path(daily_dir_path, ticker + '.csv')
daily_df.to_csv(filename)
log.notice('Retrieved: {}'.format(ticker))
def _resample_minute_data_to_daily_data(self, df):
ohlc_dict = {
'open': 'first',
'high': 'max',
'low': 'min',
'close': 'last',
'volume': 'sum',
'dividend': 'last',
'split': 'last'
}
if df.empty:
daily_df = df
else:
daily_df = df.resample('D', closed='left', label='left').agg(ohlc_dict).dropna(how='any')
# Resample mixes the columns so lets re-arrange them
daily_df = daily_df[self._cols]
return daily_df
def _validated_start_and_end_dates(
self,
start_date: datetime,
end_date: datetime
) -> Tuple[datetime, datetime]:
"""
Creates valid start and end dates. Defaults to a start date of 30 calendar days ago and end date of today.
Args:
start_date (datetime): The start date.
end_date (datetime): The end date.
Returns:
start_date (datetime): The validated start date.
end_date (datetime): The validated end date.
"""
today = datetime.today()
if start_date is None:
start_date = today - timedelta(days=30)
if end_date is None:
end_date = today
if start_date > end_date:
temp_date = start_date
start_date = end_date
end_date = temp_date
return start_date, end_date
def _minute_dataframe_for_dates(
self,
ticker: str,
start_date: datetime,
end_date: datetime
) -> pd.DataFrame:
"""
Returns a DataFrame containing the all the minute bars for stock between the start and end dates.
Args:
ticker (str): Ticker symbol for the stock.
start_date (datetime): Date to start pulling data.
end_date (datetime): Date to stop pulling data.
Returns:
combined_df (DataFrame): Contains the all the minute bars for a stock between the start and end dates.
"""
start_date, end_date = self._validated_start_and_end_dates(start_date, end_date)
combined_df = pd.DataFrame(columns=self._cols)
combined_df.index.name = 'date'
#
# # Get the date the symbol was listed on the exchange.
# list_date = self._list_date(ticker)
#
# if list_date is not None:
# # If the we are asking for data from before the stock was listed, then set the start date to the day
# # the stock was listed.
# if list_date > start_date:
# log.info('The symbol {} was not listed until: {}. Adjusting start time.', ticker, list_date)
# start_date = list_date
# Build a list of the trading days from the dates passed in.
session_dates = self._calendar.sessions_in_range(start_date, end_date)
if session_dates.empty:
log.info('The symbol {} did not trade between {} and {} ', ticker, start_date, end_date)
return combined_df
# Iterate over the trading dates backwards. This means we don't need to know exactly
# when the stock started trading. Note: this won't pull data for stocks that have been delisted.
# TODO: Add code to capture data for delisted stocks.
num_missing_dates = 0
for timestamp in reversed(session_dates):
df = self._minute_dataframe_for_date(ticker, timestamp)
if df.empty:
# Start counting the number of consecutive trading dates we are missing data.
num_missing_dates += 1
log.info('No minute data for {} on {}'.format(ticker, timestamp.date()))
else:
# reset missing date counter
num_missing_dates = 0
log.info('Retrieved minute data for {} on {}'.format(ticker, timestamp.date()))
combined_df = pd.concat([combined_df, df])
if num_missing_dates >= self.MISSING_DATE_THRESHOLD:
log.info('No minute data for {} for {} days. Quitting.'.format(ticker, self.MISSING_DATE_THRESHOLD))
break
# Sort the dataframe oldest first, newest last.
combined_df.sort_index(inplace=True)
return combined_df
# def _list_date(self, ticker: str) -> datetime:
# return None
def _minute_dataframe_for_date(self, ticker: str, start_timestamp: pd.Timestamp) -> pd.DataFrame:
raise NotImplementedError
def _fixna(self, df, symbol):
cols = ['close', 'high', 'low', 'open']
df[cols] = df[cols].replace({0: np.nan})
df[cols] = df[cols].replace({-1.0: np.nan})
if df.isnull().sum().sum() > 0:
# fixna_list.append(symbol)
df['open'] = df['open'].bfill().ffill()
df['close'] = df['close'].bfill().ffill()
df.loc[df['low'].isnull(), 'low'] = df['open']
df.loc[df['high'].isnull(), 'high'] = df['open']
df.loc[df['close'].isnull(), 'close'] = df['open']
return df
def _check_sessions(self, df, ticker, frequency='daily'):
# Remove any data that are outside of the trading sessions for the calendar.
if df.empty:
return df
asset_first_day = df.index[0]
asset_last_day = df.index[-1]
sessions = self._calendar.sessions_in_range(asset_first_day, asset_last_day)
asset_sessions = sessions[sessions.slice_indexer(asset_first_day, asset_last_day)]
if frequency == 'minute':
minutes_passed = len(df)
asset_first_day = self._calendar.minute_to_session_label(asset_first_day, direction='next')
asset_last_day = self._calendar.minute_to_session_label(asset_last_day, direction='previous')
minutes_in_session = self._calendar.minutes_for_sessions_in_range(asset_first_day, asset_last_day)
df = df[df.index.isin(minutes_in_session)]
if (minutes_passed) > len(minutes_in_session):
# print('Removed ' + str((minutes_passed) - len(minutes_in_session)) + ' minutes')
pass
elif minutes_passed < len(minutes_in_session):
num_missing_sessions = len(minutes_in_session) - minutes_passed
log.info('Missing sessions for {}'.format(ticker))
elif frequency == 'daily' and len(df) != len(asset_sessions):
missing_sessions = asset_sessions.difference(
pd.to_datetime(np.array(df.index), unit='s', utc=True, )).tolist()
extra_sessions = pd.to_datetime(np.array(df.index), unit='s', utc=True, ).difference(
asset_sessions).tolist()
if len(missing_sessions) > 0:
# missing_sessions_list.append(symbol)
# print('Adding ' + str(len(missing_sessions)) + ' sessions for ' + str(ticker))
pass
if len(extra_sessions) > 0:
# extra_sessions_list.append(symbol)
# print('Removing ' + str(len(extra_sessions)) + ' sessions for ' + str(symbol))
pass
for missing_session in missing_sessions:
prev_date = self._calendar.previous_session_label(missing_session)
row_to_copy = df[(df.index == prev_date)]
row_to_copy_val = row_to_copy.values
# from IPython import embed; embed()
df.loc[missing_session] = row_to_copy_val[0]
df.loc[missing_session].volume = 0
# row = row_to_copy
# table.append(row)
for extra_session in extra_sessions:
# delete stuff
df.drop(extra_session)
if frequency == 'minute':
log.info('Downloaded and processed {} minute bars for {}', len(df), ticker)
else:
log.info('Downsampled {} daily bars for {}', len(df), ticker)
return df
| [
"logbook.Logger",
"pandas.DataFrame",
"pathlib.Path",
"numpy.array",
"zipline.get_calendar",
"datetime.datetime.today",
"datetime.timedelta",
"pandas.concat"
] | [((181, 215), 'logbook.Logger', 'logbook.Logger', (['"""BasePriceManager"""'], {}), "('BasePriceManager')\n", (195, 215), False, 'import logbook\n'), ((322, 357), 'zipline.get_calendar', 'zl.get_calendar', ([], {'name': 'calendar_name'}), '(name=calendar_name)\n', (337, 357), True, 'import zipline as zl\n'), ((746, 780), 'pathlib.Path', 'pathlib.Path', (['output_dir', '"""minute"""'], {}), "(output_dir, 'minute')\n", (758, 780), False, 'import pathlib\n'), ((806, 839), 'pathlib.Path', 'pathlib.Path', (['output_dir', '"""daily"""'], {}), "(output_dir, 'daily')\n", (818, 839), False, 'import pathlib\n'), ((1503, 1549), 'pathlib.Path', 'pathlib.Path', (['minute_dir_path', "(ticker + '.csv')"], {}), "(minute_dir_path, ticker + '.csv')\n", (1515, 1549), False, 'import pathlib\n'), ((1797, 1842), 'pathlib.Path', 'pathlib.Path', (['daily_dir_path', "(ticker + '.csv')"], {}), "(daily_dir_path, ticker + '.csv')\n", (1809, 1842), False, 'import pathlib\n'), ((3071, 3087), 'datetime.datetime.today', 'datetime.today', ([], {}), '()\n', (3085, 3087), False, 'from datetime import datetime, timedelta\n'), ((4143, 4175), 'pandas.DataFrame', 'pd.DataFrame', ([], {'columns': 'self._cols'}), '(columns=self._cols)\n', (4155, 4175), True, 'import pandas as pd\n'), ((3152, 3170), 'datetime.timedelta', 'timedelta', ([], {'days': '(30)'}), '(days=30)\n', (3161, 3170), False, 'from datetime import datetime, timedelta\n'), ((5926, 5954), 'pandas.concat', 'pd.concat', (['[combined_df, df]'], {}), '([combined_df, df])\n', (5935, 5954), True, 'import pandas as pd\n'), ((8511, 8529), 'numpy.array', 'np.array', (['df.index'], {}), '(df.index)\n', (8519, 8529), True, 'import numpy as np\n'), ((8607, 8625), 'numpy.array', 'np.array', (['df.index'], {}), '(df.index)\n', (8615, 8625), True, 'import numpy as np\n')] |
from unittest import TestCase
from unittest import skip
from cs3api4lab.api.cs3_share_api import Cs3ShareApi
from cs3api4lab.api.cs3_file_api import Cs3FileApi
from cs3api4lab.config.config_manager import Cs3ConfigManager
from traitlets.config import LoggingConfigurable
class TestCs3ShareApi(TestCase, LoggingConfigurable):
api = None
config = None
share_id = None
receiver_id = 'f7fbf8c8-139b-4376-b307-cf0a8c2d0d9c'
receiver_idp = '<EMAIL>'
receiver2_id = '932b4540-8d16-481e-8ef4-588e4b6b151c'
receiver2_idp = 'example.<EMAIL>'
receiver_role = 'viewer'
receiver_grantee_type = 'user'
file_path = '/test.txt'
file_name = 'test.txt'
storage_id = '123e4567-e89b-12d3-a456-426655440000'
def setUp(self):
self.config = Cs3ConfigManager.get_config()
self.storage = Cs3FileApi(self.log)
self.api = Cs3ShareApi(self.log)
def test_create_and_list_directory_model(self):
created_share = self._create_share()
self.share_id = created_share['opaque_id']
share_list = self.api.list()
try:
if not list(filter(lambda s: s['name'] == self.file_name, share_list['content'])):
raise Exception("Share not created")
finally:
self._clear_shares()
def test_create_duplicate_and_list_directory_model(self):
created_share = self._create_share()
self.share_id = created_share['opaque_id']
self._create_test_share(self.receiver2_id, self.receiver2_idp)
share_list = self.api.list()
self.assertEqual(len(share_list['content']), 1)
try:
if not list(filter(lambda s: s['name'] == self.file_name, share_list['content'])):
raise Exception("Share not created")
finally:
self._clear_shares()
def test_list_grantees_for_file(self):
self._create_share()
response = self.api.list_grantees_for_file(self.file_path)
try:
if not response:
raise Exception("Failed to retrieve grantees of the file")
if response['shares'][0]['grantee']['opaque_id'] != self.receiver_id:
raise Exception("Incorrect grantee")
if response['shares'][0]['grantee']['permissions'] != self.receiver_role:
raise Exception("Incorrect permissions")
finally:
self._clear_shares()
def test_remove(self):
created_share = self._create_share()
self.share_id = created_share['opaque_id']
share_list = self.api.list()
try:
if not list(filter(lambda s: s['name'] == self.file_name, share_list['content'])):
raise Exception("Share not created")
finally:
self.api.remove(self.share_id)
share_list = self.api.list()
if list(filter(lambda s: s['name'] == self.file_name, share_list['content'])):
raise Exception("Share not removed")
def test_update(self):
created_share = self._create_share()
self.share_id = created_share['opaque_id']
self.api.update(self.share_id, 'editor')
share_list = self.api.list_grantees_for_file(self.file_path)
try:
if not list(filter(
lambda s: s['grantee']['opaque_id'] == self.receiver_id
and s['grantee']['permissions'] == 'editor',
share_list['shares'])):
raise Exception("Share not updated")
finally:
self._clear_shares()
@skip
def test_list_received(self):
self.api.list_received()
@skip
def test_update_received(self):
self.api.update_received("1", "accepted")
def _create_share(self):
self._create_test_file()
return self._create_test_share()
def _clear_shares(self):
shares = self.api.list_grantees_for_file(self.file_path)
for share in shares['shares']:
self._remove_test_share(share['opaque_id'])
try:
self._remove_test_file()
except IOError as e:
print("Error remove file:", e)
def _create_test_share(self, receiver_id='f7fbf8c8-139b-4376-b307-cf0a8c2d0d9c', receiver_idp='<EMAIL>'):
file_path = self.config['home_dir'] + self.file_path
return self.api.create(self.config['endpoint'],
file_path,
receiver_id,
receiver_idp,
self.receiver_role,
self.receiver_grantee_type)
def _remove_test_share(self, share_id):
self.api.remove(share_id)
def _create_test_file(self):
self.storage.write_file(self.file_path,
"Lorem ipsum dolor sit amet...",
self.config['endpoint'])
def _remove_test_file(self):
self.storage.remove(self.file_path,
self.config['endpoint'])
| [
"cs3api4lab.config.config_manager.Cs3ConfigManager.get_config",
"cs3api4lab.api.cs3_share_api.Cs3ShareApi",
"cs3api4lab.api.cs3_file_api.Cs3FileApi"
] | [((784, 813), 'cs3api4lab.config.config_manager.Cs3ConfigManager.get_config', 'Cs3ConfigManager.get_config', ([], {}), '()\n', (811, 813), False, 'from cs3api4lab.config.config_manager import Cs3ConfigManager\n'), ((837, 857), 'cs3api4lab.api.cs3_file_api.Cs3FileApi', 'Cs3FileApi', (['self.log'], {}), '(self.log)\n', (847, 857), False, 'from cs3api4lab.api.cs3_file_api import Cs3FileApi\n'), ((877, 898), 'cs3api4lab.api.cs3_share_api.Cs3ShareApi', 'Cs3ShareApi', (['self.log'], {}), '(self.log)\n', (888, 898), False, 'from cs3api4lab.api.cs3_share_api import Cs3ShareApi\n')] |
# -*- coding: utf-8 -*-
"""
Created on Mon Jul 1 23:11:03 2019
@author: Relieak
"""
'''
DCARD 熱門前30文章搜尋
純用 BeautifulSoup 寫法
'''
from bs4 import BeautifulSoup
import requests
import re
url = "https://www.dcard.tw/f/"
html = requests.get(url)
html.encoding = "utf-8"
#print(html.text)
sp = BeautifulSoup(html.text,"html.parser")
#print(sp.prettify()) #排版後更容易分析
#data = sp.select(".PostEntry_root_V6g0rd")
data = sp.select(".PostList_entry_1rq5Lf") #若要搜尋標簽中的內容 必須先搜尋上一個標籤 否則會找不到 ex : href 先<div 不能先<a
#print(data[0])
#for link in data[1].find_all("a",{"class" : "PostEntry_root_V6g0rd"}) :
#
# #print(link)
#
# http = link.get("href")
#
# print("https://www.dcard.tw%s" %(http))
while True :
try :
number = input("~~~歡迎來到Dcard 前30熱門文章搜尋~~~\n""\n你想看第幾則文章(離開 請按Enter)>> ")
if number == "" :
break
elif 0 < int(number) <= 30 :
try :
for i in range(int(number)-1,int(number)) :
data1 = data[i].find_all("h3",{"class" : "Title__Text-v196i6-0 gmfDU"})
data2 = data[i].find_all("div",{"class" : "PostEntry_excerpt_2eHlNn"})
data3 = data[i].find_all("div",{"class" : "ActionBar__LikeCount-pwz3il-1 cGEHtj"})
data4 = data[i].find_all("span",{"class" : "ActionBar__CommentWrapper-pwz3il-5 hkpJwJ"})
data5 = data[i].find_all("span",{"class" : "Header__PublishedDateTime-xvcbwe-3 MDszy"})
print("")
print("第%d則" %(int(number)))
print("標題 :",data1[0].text)
print("簡介 :",data2[0].text)
print("表達心情數 :",data3[0].text)
# .strip() 方法用於移除字符串頭尾指定的字符(默認為空格或換行符)或字符序列 -->文字<--
print("回應數 :",data4[0].text.strip("回應"))
print("發表時間 :",data5[0].text)
# 搜尋標簽中的內容的方法 #href
for link in data[i].find_all("a",{"class" : "PostEntry_root_V6g0rd"}) :
http = link.get("href")
#print(http)
#正規表示法
A = re.compile("[0-9]+") #表示任意數字串
B = re.compile("[A-Za-z0-9\./_]+") #表達任意數字,任意英文字母和底線字元的組合,也可寫成 \w
# search(string)的用法是傳回第一組符合正規表示法的字串
Search1 = A.search(http)
Search2 = B.search(http) #去掉網址後面的中文字
#print(Search1)
#print(Search2)
print("文章ID :",Search1.group()) #傳回儲存在match物件中的值 group()
print("網頁 :https://www.dcard.tw%s" %(Search2.group()))
print("")
except :
""" 若文章簡介開頭是 "前情提要" 會搜尋不到內文 "PostEntry_excerpt_2eHlNn" 產生錯誤 須去搜尋 "PostEntry_reply_1oU-6z" """
data2 = data[i].find_all("div",{"class" : "PostEntry_reply_1oU-6z"})
data3 = data[i].find_all("div",{"class" : "ActionBar__LikeCount-pwz3il-1 cGEHtj"})
data4 = data[i].find_all("span",{"class" : "ActionBar__CommentWrapper-pwz3il-5 hkpJwJ"})
data5 = data[i].find_all("span",{"class" : "Header__PublishedDateTime-xvcbwe-3 MDszy"})
print("簡介 :",data2[0].text)
print("表達心情數 :",data3[0].text)
print("回應數 :",data4[0].text.strip("回應"))
print("發表時間 :",data5[0].text)
for link in data[i].find_all("a",{"class" : "PostEntry_root_V6g0rd"}) :
http = link.get("href")
A = re.compile("[0-9]+")
B = re.compile("[A-Za-z0-9\./_]+")
Search1 = A.search(http)
Search2 = B.search(http)
print("文章ID :",Search1.group())
print("網頁 :https://www.dcard.tw%s" %(Search2.group()))
print("")
else :
print("")
print("只能搜尋前30篇文章喔!!")
except :
print("")
print("輸入錯誤請重新輸入!!!")
| [
"bs4.BeautifulSoup",
"requests.get",
"re.compile"
] | [((230, 247), 'requests.get', 'requests.get', (['url'], {}), '(url)\n', (242, 247), False, 'import requests\n'), ((297, 336), 'bs4.BeautifulSoup', 'BeautifulSoup', (['html.text', '"""html.parser"""'], {}), "(html.text, 'html.parser')\n", (310, 336), False, 'from bs4 import BeautifulSoup\n'), ((2656, 2676), 're.compile', 're.compile', (['"""[0-9]+"""'], {}), "('[0-9]+')\n", (2666, 2676), False, 'import re\n'), ((2746, 2777), 're.compile', 're.compile', (['"""[A-Za-z0-9\\\\./_]+"""'], {}), "('[A-Za-z0-9\\\\./_]+')\n", (2756, 2777), False, 'import re\n'), ((4575, 4595), 're.compile', 're.compile', (['"""[0-9]+"""'], {}), "('[0-9]+')\n", (4585, 4595), False, 'import re\n'), ((4620, 4651), 're.compile', 're.compile', (['"""[A-Za-z0-9\\\\./_]+"""'], {}), "('[A-Za-z0-9\\\\./_]+')\n", (4630, 4651), False, 'import re\n')] |
import random
import sys
print("")
print("4d6, drop lowest:")
print("")
def roll4d6():
global threshold
target_met = False
total = []
for attr in range(0,6):
result = []
raw = []
for roll in range(0,4):
dice = random.randint(1,6)
result.append(dice)
raw.append(dice)
result.remove(min(result))
print(raw, result, sum(result))
total.append(sum(result))
if sum(result) >= int(threshold):
target_met = True
print("------------------")
print("Total: " + str(sum(total)))
return sum(total), target_met
try:
global target_met, threshold
target = 0
count = 1
if "-" in sys.argv[1]:
# count
tag, threshold = sys.argv[1].split("-")
for i in range(int(threshold)):
print("")
print("*******************")
target, target_met = roll4d6()
print("4d6 Total: " + str(target) + " Total " + str(threshold) + " Sets: " + str(threshold) + " Rolls: " + str(count))
count = count + 1
print("")
elif "+" in sys.argv[1]:
# at least one stat matches this number
tag, threshold = sys.argv[1].split("+")
target_met = False
while target_met == False:
print("")
print("*******************")
target, target_met = roll4d6()
print("4d6 Total: " + str(target) + " Threshold (At Least): " + str(threshold) + " Rolls: " + str(count))
count = count + 1
print("")
else:
threshold = sys.argv[1]
while int(target) <= int(threshold):
print("")
print("*******************")
target, result = roll4d6()
print("4d6 Total: " + str(target) + " Threshold: " + str(threshold) + " Rolls: " + str(count))
count = count + 1
print("")
except:
threshold = 0
roll4d6()
print("4d6.py +<num> for at least one stat in a set at num or above")
print("4d6.py -<num> will roll up that many stat sets")
print("4d6.py <num> will roll until the total of a set meets or exceeds num") | [
"random.randint"
] | [((265, 285), 'random.randint', 'random.randint', (['(1)', '(6)'], {}), '(1, 6)\n', (279, 285), False, 'import random\n')] |
from __future__ import print_function
import os
import os.path
from xml.etree import ElementTree
def read_machine_file():
"""
Reads the machine definition xml ROS launch file MCT_CONFIG/machine/mct.machine
Returns a list containg a dictionary of attributes for each machine in the
machine file.
"""
mct_config_dir = os.environ['MCT_CONFIG']
mct_machine_file = os.path.join(mct_config_dir, 'machine', 'mct.machine')
tree = ElementTree.parse(mct_machine_file)
machine_elem_list = tree.findall('machine')
machine_list = []
for machine_elem in machine_elem_list:
machine_list.append(machine_elem.attrib)
return machine_list
# -----------------------------------------------------------------------------
if __name__ == '__main__':
machine_list = read_machine_file()
print(machine_list)
| [
"os.path.join",
"xml.etree.ElementTree.parse"
] | [((393, 447), 'os.path.join', 'os.path.join', (['mct_config_dir', '"""machine"""', '"""mct.machine"""'], {}), "(mct_config_dir, 'machine', 'mct.machine')\n", (405, 447), False, 'import os\n'), ((459, 494), 'xml.etree.ElementTree.parse', 'ElementTree.parse', (['mct_machine_file'], {}), '(mct_machine_file)\n', (476, 494), False, 'from xml.etree import ElementTree\n')] |
#!/usr/local/python/bin/python
# script to check the previously unsolved files
#
# to do:
# Sanity check all the image_ids in the table actually have a png
# quick check shows 1349 in DB and 1353 pngs, 4 out, not bad
#
from create_movie import create_movie
import os,sys,getpass,time
import glob as g
from astropy.io import fits
from collections import defaultdict
import pyds9
import argparse as ap
me=getpass.getuser()
if me=='ops':
w_dir="/ngts/staging/archive/minisurvey/junk"
astrom_loc="/usr/local/astrometry.net/bin/"
elif me=='James':
w_dir='/Users/James/Desktop/junk'
astrom_loc="/usr/local/bin/"
else:
print("WHOAMI?")
sys.exit(1)
# check for w_dir
if os.path.exists(w_dir)==False:
print("I'm dying... (no w_dir)")
sys.exit(1)
# get command line args
def argParse():
parser=ap.ArgumentParser(description="A script to redo the failed minisurvey publishing step")
parser.add_argument('--astrometry', help = "try redoing the astrometry?", action='store_true')
parser.add_argument('--manual', help = "manually analyse the images with DS9/DSS", action='store_true')
parser.add_argument('--yes2all', help = "select this to skip prompting - used if pass already has been made through the imagaes and all are good", action='store_true')
args=parser.parse_args()
return args
# read in astrometry.net log files
def getAstromFromFile(astromfile):
f = open(astromfile, "r").readlines()
ra=None
dec=None
for i in f:
if i.startswith("Field center: (RA H:M:S"):
tmp = i.split('=')
ra,dec = tmp[1].split(',')
ra = ra.strip()
ra = ra.replace('(','')
dec = dec.strip()
dec = dec.replace(').','')
break
return ra, dec
# do the astrometry
def astrometry(image,scale_l,scale_h,ra=None,dec=None,radius=5.0,cpulimit=90):
astromfile="astrometry_%s.log" % (image)
command = "%s/solve-field %s --scale-low %s --scale-high %s --cpulimit %s --no-plots --overwrite" % (astrom_loc,image, scale_l, scale_h, cpulimit)
command = "%s > %s" % (command,astromfile)
os.system(command)
ra,dec=getAstromFromFile(astromfile)
return ra,dec
args=argParse()
os.chdir(w_dir)
t=sorted(g.glob('*.fits'))
if args.astrometry:
RA,DEC=[],[]
for i in t:
ra,dec=astrometry(i,2.83,2.93,cpulimit=2)
if ra:
RA.append(ra)
DEC.append(dec)
else:
RA.append("0")
DEC.append("0")
if args.manual:
fields=defaultdict(list)
done=defaultdict(list)
# loop over the and check for multiples
# of the same field, if so work on the last one only
for i in t:
h=fits.open(i)[0].header['FIELD']
fields[h].append(i)
d=pyds9.DS9()
time.sleep(5)
d.set('scale zscale')
d.set('preserve scale')
d.set('preserve pan')
print("Remeber to DELETE duplicate images")
rm_string=""
for i in fields:
image=fields[i][-1]
d.set('frame clear all')
h=fits.open(image)[0]
ra=h.header['CMD_RA']
dec=h.header['CMD_DEC']
# print this so we can see which have duplicates to delete
print(fields[i])
if len(fields[i])>1:
for k in range(0,len(fields[i])-1):
rm_string=rm_string+"%s " % (fields[i][k])
# display the image in DS9 and load the correct region of sky beside it
d.set('tile yes')
d.set('frame 1')
d.set('file %s' % (image))
d.set('zoom 2')
d.set('wcs align yes')
d.set('cmap invert yes')
d.set('frame 2')
d.set('dsseso coord %.6f %.6f degrees size 30 30 arcmin' % (ra,dec))
d.set('zoom to fit')
d.set('wcs align yes')
d.set('cmap invert yes')
d.set('frame center all')
if args.yes2all:
done[i].append(image)
else:
yn=input("Do the fields match? (y/n): ")
if yn.lower().startswith('y'):
done[i].append(image)
else:
continue
print(rm_string)
# need to make an astrometry* log file for the manually solved images?
# and also a png too, then update the database as with the others, manually?
table_update_string=""
if len(done)> 0:
print("Check the UPDATE strings as use them to UPDATE the minisurvey table")
for i in done:
create_movie(done[i],images_directory="%s/" % (w_dir),no_time_series=True,include_increment=False,clobber_images_directory=False,resize_factor=4,multiprocess=False)
table_update_string=table_update_string+"UPDATE mini_survey SET checked_out=0,astrometry=1,done=1,png=1,fails=0 where image_id=\"%s\";\n" % (done[i][0][5:-5]) # image name minus IMAGE and .fits
print(table_update_string)
| [
"os.path.exists",
"pyds9.DS9",
"argparse.ArgumentParser",
"time.sleep",
"astropy.io.fits.open",
"os.chdir",
"create_movie.create_movie",
"collections.defaultdict",
"sys.exit",
"getpass.getuser",
"os.system",
"glob.glob"
] | [((408, 425), 'getpass.getuser', 'getpass.getuser', ([], {}), '()\n', (423, 425), False, 'import os, sys, getpass, time\n'), ((2255, 2270), 'os.chdir', 'os.chdir', (['w_dir'], {}), '(w_dir)\n', (2263, 2270), False, 'import os, sys, getpass, time\n'), ((692, 713), 'os.path.exists', 'os.path.exists', (['w_dir'], {}), '(w_dir)\n', (706, 713), False, 'import os, sys, getpass, time\n'), ((763, 774), 'sys.exit', 'sys.exit', (['(1)'], {}), '(1)\n', (771, 774), False, 'import os, sys, getpass, time\n'), ((827, 919), 'argparse.ArgumentParser', 'ap.ArgumentParser', ([], {'description': '"""A script to redo the failed minisurvey publishing step"""'}), "(description=\n 'A script to redo the failed minisurvey publishing step')\n", (844, 919), True, 'import argparse as ap\n'), ((2132, 2150), 'os.system', 'os.system', (['command'], {}), '(command)\n', (2141, 2150), False, 'import os, sys, getpass, time\n'), ((2280, 2296), 'glob.glob', 'g.glob', (['"""*.fits"""'], {}), "('*.fits')\n", (2286, 2296), True, 'import glob as g\n'), ((2569, 2586), 'collections.defaultdict', 'defaultdict', (['list'], {}), '(list)\n', (2580, 2586), False, 'from collections import defaultdict\n'), ((2596, 2613), 'collections.defaultdict', 'defaultdict', (['list'], {}), '(list)\n', (2607, 2613), False, 'from collections import defaultdict\n'), ((2808, 2819), 'pyds9.DS9', 'pyds9.DS9', ([], {}), '()\n', (2817, 2819), False, 'import pyds9\n'), ((2824, 2837), 'time.sleep', 'time.sleep', (['(5)'], {}), '(5)\n', (2834, 2837), False, 'import os, sys, getpass, time\n'), ((658, 669), 'sys.exit', 'sys.exit', (['(1)'], {}), '(1)\n', (666, 669), False, 'import os, sys, getpass, time\n'), ((3076, 3092), 'astropy.io.fits.open', 'fits.open', (['image'], {}), '(image)\n', (3085, 3092), False, 'from astropy.io import fits\n'), ((4485, 4662), 'create_movie.create_movie', 'create_movie', (['done[i]'], {'images_directory': "('%s/' % w_dir)", 'no_time_series': '(True)', 'include_increment': '(False)', 'clobber_images_directory': '(False)', 'resize_factor': '(4)', 'multiprocess': '(False)'}), "(done[i], images_directory='%s/' % w_dir, no_time_series=True,\n include_increment=False, clobber_images_directory=False, resize_factor=\n 4, multiprocess=False)\n", (4497, 4662), False, 'from create_movie import create_movie\n'), ((2741, 2753), 'astropy.io.fits.open', 'fits.open', (['i'], {}), '(i)\n', (2750, 2753), False, 'from astropy.io import fits\n')] |
from csv import DictReader
from os.path import dirname, join, exists
from os import makedirs, stat
from concurrent.futures import ThreadPoolExecutor, as_completed
import urllib.request
from time import time
def get_file_by_case(case):
D = r'C:\Users\andy\supreme-court-scrap\documents-from-student-decoding'
parts = case.split('/')
year = parts[-1]
case = parts[0].zfill(6)
a, b, c = case[:2], case[2:4], case[4:6]
dir = join(D, year, a, b)
outfile = join(dir, f'{c}.html')
return outfile
def write_out_file(case, text):
outfile = get_file_by_case(case)
makedirs(dirname(outfile), 777, True)
open(outfile, 'w', encoding='Windows-1255').write(text)
print(f'Wrote to {outfile}')
def parse_case(case):
url = case['url']
case_id = case['casename'].split()[1]
return case_id, url
def exists_case(case_id):
return exists(get_file_by_case(case_id))
def download(case):
case_id, url = parse_case(case)
print(f'Downloading {case_id}')
s = time()
response = urllib.request.urlopen(url)
data = response.read() # a `bytes` object
text = data.decode('Windows-1255') # a `str`; this
elapsed = time() -s
print(f'Downloaded {case_id} in {elapsed}')
write_out_file(case_id, text)
return
f = r'C:\Users\andy\supreme-court-scrap\answers-13.csv'
f = open(f, 'r', encoding='utf-8')
f = DictReader(f)
futures = []
with ThreadPoolExecutor(max_workers=20) as tpe:
for case in f:
case_id, url = parse_case(case)
if exists_case(case_id):
print(f'Skipping {case_id}')
continue
future = tpe.submit(download, case)
futures.append(future)
for future in as_completed(futures):
future.result()
| [
"csv.DictReader",
"concurrent.futures.ThreadPoolExecutor",
"os.path.join",
"concurrent.futures.as_completed",
"os.path.dirname",
"time.time"
] | [((1378, 1391), 'csv.DictReader', 'DictReader', (['f'], {}), '(f)\n', (1388, 1391), False, 'from csv import DictReader\n'), ((447, 466), 'os.path.join', 'join', (['D', 'year', 'a', 'b'], {}), '(D, year, a, b)\n', (451, 466), False, 'from os.path import dirname, join, exists\n'), ((481, 503), 'os.path.join', 'join', (['dir', 'f"""{c}.html"""'], {}), "(dir, f'{c}.html')\n", (485, 503), False, 'from os.path import dirname, join, exists\n'), ((1012, 1018), 'time.time', 'time', ([], {}), '()\n', (1016, 1018), False, 'from time import time\n'), ((1411, 1445), 'concurrent.futures.ThreadPoolExecutor', 'ThreadPoolExecutor', ([], {'max_workers': '(20)'}), '(max_workers=20)\n', (1429, 1445), False, 'from concurrent.futures import ThreadPoolExecutor, as_completed\n'), ((1703, 1724), 'concurrent.futures.as_completed', 'as_completed', (['futures'], {}), '(futures)\n', (1715, 1724), False, 'from concurrent.futures import ThreadPoolExecutor, as_completed\n'), ((606, 622), 'os.path.dirname', 'dirname', (['outfile'], {}), '(outfile)\n', (613, 622), False, 'from os.path import dirname, join, exists\n'), ((1179, 1185), 'time.time', 'time', ([], {}), '()\n', (1183, 1185), False, 'from time import time\n')] |
from pathlib import Path
from shutil import copyfile
import logging
def move_files(dir_from: Path, dir_to: Path):
logging.info(f"Moving files from '{dir_from}' to '{dir_to}'")
p = dir_from.glob("**/*")
input_paths = [x for x in p if x.is_file()]
for input_path in input_paths:
filename = input_path.name
output_path = dir_to / filename
copyfile(input_path, output_path)
logging.info(f"Moved file: {filename}")
| [
"shutil.copyfile",
"logging.info"
] | [((120, 181), 'logging.info', 'logging.info', (['f"""Moving files from \'{dir_from}\' to \'{dir_to}\'"""'], {}), '(f"Moving files from \'{dir_from}\' to \'{dir_to}\'")\n', (132, 181), False, 'import logging\n'), ((378, 411), 'shutil.copyfile', 'copyfile', (['input_path', 'output_path'], {}), '(input_path, output_path)\n', (386, 411), False, 'from shutil import copyfile\n'), ((420, 459), 'logging.info', 'logging.info', (['f"""Moved file: {filename}"""'], {}), "(f'Moved file: {filename}')\n", (432, 459), False, 'import logging\n')] |
#!/usr/bin/python
#-*-coding=utf-8-*-
import json
import jieba
class emo:
dict = {}
emoVal = {}
emodir = "emo_dict"
def __init__(self,dictname):
self.dict = self.loadDict(dictname)
def loadDict(self,dictname):
path = "./"+self.emodir+"/"+dictname+".json"
print ("loading "+ path +"......")
f = open(path, mode='r')
try:
js = json.load(f)
print (["new dictionary loaded!",js])
f.close()
return js
except Exception :
print (Exception)
def update(self,dictname):
new_dict = self.loadDict(dictname)
for w in new_dict:
try:
dict[w].append(new_dict[w])
except KeyError:
dict[w] = (dict[w]+new_dict[w])/2.0
def getEmo(self, txts, name):
Eval = 0
wordsNum = 0
for w in txts:
if w in self.dict:
Eval += self.dict[w]
wordsNum+=1
if name in self.emoVal:
self.emoVal[name]['totalV'] += Eval
self.emoVal[name]['wnum'] += wordsNum
else:
self.emoVal[name] = dict(wnum = wordsNum, totalV = Eval)
return self.emoVal[name]
def clear(self):
self.emoVal = {}
#test
#emo = emo("emodata")
#print emo.dict
#txt = "2011年底泽熙在重庆啤酒的抢反弹广为人知,徐翔事后说:“重庆啤酒不是股票,是彩票。第一次刮出来‘谢谢你’,第二次刮出来还是‘谢谢你’,这时候大家都把它当废纸扔了,但彩票还没刮"
#test_txts = jieba.cut(txt)
#print emo.getEmo(test_txts,"重庆啤酒")
| [
"json.load"
] | [((412, 424), 'json.load', 'json.load', (['f'], {}), '(f)\n', (421, 424), False, 'import json\n')] |
import random
import numpy as np
from gym_multigrid.multigrid import World
from gym_multigrid.multigrid import DIR_TO_VEC
from gym_multigrid.multigrid import Actions
class Agent:
def __init__(self, agent_id, agent_type=0):
self.id = agent_id
self.total_reward = 0
self.action_probabilities = [0.1, 0.2, 0.2, 0.4, 0.1]
self.agent_type = agent_type
self.observation = None
def next_action(self, observation, reward, round_id):
pass
def start_simulation(self, observation, rounds):
pass
def end_simulation(self, observation, reward, round_id):
pass
def random_action(self):
action = random.choices(np.arange(5), weights=self.action_probabilities, k=1)[0]
return action
def get_my_position(self):
width = len(self.observation)
height = len(self.observation[0])
for x in range(width):
for y in range(height):
if self.observation[x][y][0] == World.OBJECT_TO_IDX["agent"] and self.observation[x][y][2] == self.id:
return x, y
return -1, -1
def get_all_ball_positions(self):
width = len(self.observation)
height = len(self.observation[0])
positions_x = []
positions_y = []
for x in range(width):
for y in range(height):
if self.observation[x][y][0] == World.OBJECT_TO_IDX["ball"]:
positions_x.append(x)
positions_y.append(y)
return positions_x, positions_y
"""
actions:
still = 0
left = 1
right = 2
forward = 3
pickup = 4
"""
class RandomAgent(Agent):
def __init__(self, agent_id):
super().__init__(agent_id, agent_type=1)
def start_simulation(self, observation, rounds):
""" Nothing to be done """
def next_action(self, observation, reward, round_id):
#print("random index: ", self.id, " type: ", self.agent_type)
return self.random_action()
def end_simulation(self, observation, reward, round_id):
""" Nothing to be done """
class GreedyAgent(Agent):
def __init__(self, agent_id):
super().__init__(agent_id, agent_type=2)
self.width = 0
self.height = 0
def get_ball_positions(self):
positions = []
for x in range(self.width):
for y in range(self.height):
if self.observation[x][y][0] == World.OBJECT_TO_IDX["ball"]:
positions.append([x, y])
return positions
def greedy_action(self):
pos_x, pos_y = self.get_my_position()
direction = self.observation[pos_x][pos_y][1]
ball_positions = self.get_ball_positions()
target_ball_positions = get_closest_balls(pos_x, pos_y, direction, ball_positions)
target_ball_position = random.choice(target_ball_positions)
return move_towards_ball(pos_x, pos_y, direction, target_ball_position[0], target_ball_position[1])
def start_simulation(self, observation, rounds):
self.width = len(observation)
self.height = len(observation[0])
def next_action(self, observation, reward, round_id):
self.observation = observation
x, y = self.get_my_position()
#print("greedy index: ", self.id, " type: ", x, " ", y)
return self.greedy_action()
def end_simulation(self, observation, reward, round_id):
""" Nothing to be done """
def sign(x):
if x > 0:
return 1
elif x < 0:
return -1
else:
return 0
def distance_from_ball(pos_x, pos_y, direction, ball_x, ball_y):
dx = ball_x - pos_x
dy = ball_y - pos_y
turns_x = abs(sign(dx) - DIR_TO_VEC[direction][0])
turns_y = abs(sign(dy) - DIR_TO_VEC[direction][1])
return abs(dx) + abs(dy) + max(turns_x, turns_y)
def get_closest_balls(pos_x, pos_y, direction, ball_positions):
if len(ball_positions) == 0:
return [[0, 0]]
best_positions = []
best_distance = -1
for index, [x, y] in enumerate(ball_positions):
current_distance = distance_from_ball(pos_x, pos_y, direction, x, y)
if best_distance == -1 or current_distance < best_distance:
best_distance = current_distance
best_positions = [[x, y]]
elif current_distance == best_distance:
best_positions.append([x, y])
return best_positions
def get_next_state(pos_x, pos_y, direction, action):
if action == Actions.still:
return pos_x, pos_y, direction
if action == Actions.left:
new_direction = (direction + 3) % 4
return pos_x, pos_y, new_direction
if action == Actions.right:
new_direction = (direction + 1) % 4
return pos_x, pos_y, new_direction
if action == Actions.forward:
return pos_x + DIR_TO_VEC[direction][0], pos_y + DIR_TO_VEC[direction][1], direction
def move_towards_ball(pos_x, pos_y, direction, ball_x, ball_y):
distance = distance_from_ball(pos_x, pos_y, direction, ball_x, ball_y)
if distance == 1:
return Actions.pickup
best_action = Actions.still
best_next_distance = distance
for action in [Actions.left, Actions.right, Actions.forward]:
next_x, next_y, next_direction = get_next_state(pos_x, pos_y, direction, action)
current_next_distance = distance_from_ball(next_x, next_y, next_direction, ball_x, ball_y)
if current_next_distance < best_next_distance:
best_next_distance = current_next_distance
best_action = action
return best_action
| [
"random.choice",
"numpy.arange"
] | [((2837, 2873), 'random.choice', 'random.choice', (['target_ball_positions'], {}), '(target_ball_positions)\n', (2850, 2873), False, 'import random\n'), ((693, 705), 'numpy.arange', 'np.arange', (['(5)'], {}), '(5)\n', (702, 705), True, 'import numpy as np\n')] |
# -*- coding: UTF-8 -*-
import glob
import numpy as np
import pandas as pd
from PIL import Image
import random
# h,w = 60,50
h, w = (60, 50)
size = h * w
# Receding_Hairline Wearing_Necktie Rosy_Cheeks Eyeglasses Goatee Chubby
# Sideburns Blurry Wearing_Hat Double_Chin Pale_Skin Gray_Hair Mustache Bald
label_cls = 'Eyeglasses'
pngs = sorted(glob.glob('./data/img_align_celeba/*.jpg'))
data = pd.read_table('./data/list_attr_celeba.txt',
delim_whitespace=True, error_bad_lines=False)
eyeglasses = np.array(data[label_cls])
eyeglasses_cls = (eyeglasses + 1)/2
label_glasses = np.zeros((202599, 2))
correct_list = []
correct_list_test = []
false_list = []
false_list_test = []
for i in range(len(label_glasses)):
if eyeglasses_cls[i] == 1:
label_glasses[i][1] = 1
if i < 160000:
correct_list.append(i)
else:
correct_list_test.append(i)
else:
label_glasses[i][0] = 1
if i < 160000:
false_list.append(i)
else:
false_list_test.append(i)
print(len(correct_list_test), len(false_list_test))
training_set_label = label_glasses[0:160000, :]
test_set_label = label_glasses[160000:, :]
training_set_cls = eyeglasses_cls[0:160000]
test_set_cls = eyeglasses_cls[160000:]
def create_trainbatch(num=10, channel=0):
train_num = random.sample(false_list, num)
if channel == 0:
train_set = np.zeros((num, h, w))
else:
train_set = np.zeros((num, h, w, 3))
train_set_label_ = []
train_set_cls_ = []
for i in range(num):
img = Image.open(pngs[train_num[i]])
img_grey = img.resize((w, h))
if channel == 0:
img_grey = np.array(img_grey.convert('L'))
train_set[i, :, :] = img_grey
else:
img_grey = np.array(img_grey)
train_set[i, :, :, :] = img_grey
train_set_label_.append(training_set_label[train_num[i]])
train_set_cls_.append(training_set_cls[train_num[i]])
# if channel == 0:
# train_set = train_set.reshape(size,num).T
train_set_label_new = np.array(train_set_label_)
train_set_cls_new = np.array(train_set_cls_)
return train_set/255, train_set_label_new, train_set_cls_new
def create_trainbatch_all_correct(num=10, channel=0):
train_num = random.sample(correct_list, num)
if channel == 0:
train_set = np.zeros((num, h, w))
else:
train_set = np.zeros((num, h, w, 3))
train_set_label_ = []
train_set_cls_ = []
n = 0
for i in range(num):
img = Image.open(pngs[train_num[i]])
img_grey = img.resize((w, h))
if channel == 0:
img_grey = np.array(img_grey.convert('L'))
train_set[i, :, :] = img_grey
else:
img_grey = np.array(img_grey)
train_set[i, :, :, :] = img_grey
train_set_label_.append(training_set_label[train_num[i]])
train_set_cls_.append(training_set_cls[train_num[i]])
# if channel == 0:
# train_set = train_set.reshape(size,num).T
train_set_label_new = np.array(train_set_label_)
train_set_cls_new = np.array(train_set_cls_)
return train_set/255, train_set_label_new, train_set_cls_new
def create_trainbatch_(num=10, channel=0):
train_num1 = random.sample(correct_list, int(num/2))
train_num2 = random.sample(false_list, int(num/2))
train_num = train_num1+train_num2
if channel == 0:
train_set = np.zeros((num, h, w))
else:
train_set = np.zeros((num, h, w, 3))
train_set_label_ = []
train_set_cls_ = []
n = 0
for i in range(num):
img = Image.open(pngs[train_num[i]])
img_grey = img.resize((w, h))
if channel == 0:
img_grey = np.array(img_grey.convert('L'))
train_set[i, :, :] = img_grey
else:
img_grey = np.array(img_grey)
train_set[i, :, :, :] = img_grey
train_set_label_.append(training_set_label[train_num[i]])
train_set_cls_.append(training_set_cls[train_num[i]])
# if channel == 0:
# train_set = train_set.reshape(size,num).T
train_set_label_new = np.array(train_set_label_)
train_set_cls_new = np.array(train_set_cls_)
return train_set/255, train_set_label_new, train_set_cls_new
def create_trainbatch_grad(num=200, channel=0):
train_num1 = random.sample(correct_list, int(10))
train_num2 = random.sample(false_list, int(190))
train_num = train_num1+train_num2
if channel == 0:
train_set = np.zeros((num, h, w))
else:
train_set = np.zeros((num, h, w, 3))
train_set_label_ = []
train_set_cls_ = []
n = 0
for i in range(num):
img = Image.open(pngs[train_num[i]])
img_grey = img.resize((w, h))
if channel == 0:
img_grey = np.array(img_grey.convert('L'))
train_set[i, :, :] = img_grey
else:
img_grey = np.array(img_grey)
train_set[i, :, :, :] = img_grey
train_set_label_.append(training_set_label[train_num[i]])
train_set_cls_.append(training_set_cls[train_num[i]])
# if channel == 0:
# train_set = train_set.reshape(size,num).T
train_set_label_new = np.array(train_set_label_)
train_set_cls_new = np.array(train_set_cls_)
return train_set/255, train_set_label_new, train_set_cls_new
def create_testset(num=100, channel=0):
test_num1 = random.sample(correct_list_test, num)
test_num2 = random.sample(false_list_test, num)
test_num = test_num1 + test_num2
if channel == 0:
test_set = np.zeros((num*2, h, w))
else:
test_set = np.zeros((num*2, h, w, 3))
test_set_label_ = []
test_set_cls_ = []
for i in range(num*2):
img = Image.open(pngs[test_num[i]])
img_grey = img.resize((w, h))
if channel == 0:
img_grey = np.array(img_grey.convert('L'))
test_set[i, :, :] = img_grey
else:
img_grey = np.array(img_grey)
test_set[i, :, :, :] = img_grey
test_set_label_.append(label_glasses[test_num[i]])
test_set_cls_.append(eyeglasses_cls[test_num[i]])
# if channel == 0:
# test_set = test_set.reshape(size,num*2).T
test_set_label_new = np.array(test_set_label_)
test_set_cls_new = np.array(test_set_cls_)
return test_set/255, test_set_label_new, test_set_cls_new, test_set_cls_new.mean()*100
def create_testset_all(channel=0):
test_num1 = random.sample(correct_list_test, len(correct_list_test))
test_num2 = random.sample(false_list_test, len(false_list_test))
test_num = test_num1 + test_num2
# test_num =
num = len(test_num)
if channel == 0:
test_set = np.zeros((num, h, w))
else:
test_set = np.zeros((num, h, w, 3))
test_set_label_ = []
test_set_cls_ = []
for i in range(num):
img = Image.open(pngs[test_num[i]])
img_grey = img.resize((w, h))
if channel == 0:
img_grey = np.array(img_grey.convert('L'))
test_set[i, :, :] = img_grey
else:
img_grey = np.array(img_grey)
test_set[i, :, :, :] = img_grey
test_set_label_.append(label_glasses[test_num[i]])
test_set_cls_.append(eyeglasses_cls[test_num[i]])
# if channel == 0:
# test_set = test_set.reshape(size,num).T
test_set_label_new = np.array(test_set_label_)
test_set_cls_new = np.array(test_set_cls_)
return test_set/255, test_set_label_new, test_set_cls_new, test_set_cls_new.mean()*100
def create_testset_unbalanced(channel=0):
test_num1 = random.sample(correct_list_test, 10)
test_num2 = random.sample(false_list_test, 190)
test_num = test_num1 + test_num2
# test_num =
num = len(test_num)
if channel == 0:
test_set = np.zeros((num, h, w))
else:
test_set = np.zeros((num, h, w, 3))
test_set_label_ = []
test_set_cls_ = []
for i in range(num):
img = Image.open(pngs[test_num[i]])
img_grey = img.resize((w, h))
if channel == 0:
img_grey = np.array(img_grey.convert('L'))
test_set[i, :, :] = img_grey
else:
img_grey = np.array(img_grey)
test_set[i, :, :, :] = img_grey
test_set_label_.append(label_glasses[test_num[i]])
test_set_cls_.append(eyeglasses_cls[test_num[i]])
# if channel == 0:
# test_set = test_set.reshape(size,num).T
test_set_label_new = np.array(test_set_label_)
test_set_cls_new = np.array(test_set_cls_)
return test_set/255, test_set_label_new, test_set_cls_new, test_set_cls_new.mean()*100
| [
"random.sample",
"PIL.Image.open",
"numpy.array",
"numpy.zeros",
"pandas.read_table",
"glob.glob"
] | [((411, 505), 'pandas.read_table', 'pd.read_table', (['"""./data/list_attr_celeba.txt"""'], {'delim_whitespace': '(True)', 'error_bad_lines': '(False)'}), "('./data/list_attr_celeba.txt', delim_whitespace=True,\n error_bad_lines=False)\n", (424, 505), True, 'import pandas as pd\n'), ((537, 562), 'numpy.array', 'np.array', (['data[label_cls]'], {}), '(data[label_cls])\n', (545, 562), True, 'import numpy as np\n'), ((616, 637), 'numpy.zeros', 'np.zeros', (['(202599, 2)'], {}), '((202599, 2))\n', (624, 637), True, 'import numpy as np\n'), ((360, 402), 'glob.glob', 'glob.glob', (['"""./data/img_align_celeba/*.jpg"""'], {}), "('./data/img_align_celeba/*.jpg')\n", (369, 402), False, 'import glob\n'), ((1365, 1395), 'random.sample', 'random.sample', (['false_list', 'num'], {}), '(false_list, num)\n', (1378, 1395), False, 'import random\n'), ((2127, 2153), 'numpy.array', 'np.array', (['train_set_label_'], {}), '(train_set_label_)\n', (2135, 2153), True, 'import numpy as np\n'), ((2178, 2202), 'numpy.array', 'np.array', (['train_set_cls_'], {}), '(train_set_cls_)\n', (2186, 2202), True, 'import numpy as np\n'), ((2342, 2374), 'random.sample', 'random.sample', (['correct_list', 'num'], {}), '(correct_list, num)\n', (2355, 2374), False, 'import random\n'), ((3114, 3140), 'numpy.array', 'np.array', (['train_set_label_'], {}), '(train_set_label_)\n', (3122, 3140), True, 'import numpy as np\n'), ((3165, 3189), 'numpy.array', 'np.array', (['train_set_cls_'], {}), '(train_set_cls_)\n', (3173, 3189), True, 'import numpy as np\n'), ((4192, 4218), 'numpy.array', 'np.array', (['train_set_label_'], {}), '(train_set_label_)\n', (4200, 4218), True, 'import numpy as np\n'), ((4243, 4267), 'numpy.array', 'np.array', (['train_set_cls_'], {}), '(train_set_cls_)\n', (4251, 4267), True, 'import numpy as np\n'), ((5272, 5298), 'numpy.array', 'np.array', (['train_set_label_'], {}), '(train_set_label_)\n', (5280, 5298), True, 'import numpy as np\n'), ((5323, 5347), 'numpy.array', 'np.array', (['train_set_cls_'], {}), '(train_set_cls_)\n', (5331, 5347), True, 'import numpy as np\n'), ((5473, 5510), 'random.sample', 'random.sample', (['correct_list_test', 'num'], {}), '(correct_list_test, num)\n', (5486, 5510), False, 'import random\n'), ((5527, 5562), 'random.sample', 'random.sample', (['false_list_test', 'num'], {}), '(false_list_test, num)\n', (5540, 5562), False, 'import random\n'), ((6318, 6343), 'numpy.array', 'np.array', (['test_set_label_'], {}), '(test_set_label_)\n', (6326, 6343), True, 'import numpy as np\n'), ((6367, 6390), 'numpy.array', 'np.array', (['test_set_cls_'], {}), '(test_set_cls_)\n', (6375, 6390), True, 'import numpy as np\n'), ((7451, 7476), 'numpy.array', 'np.array', (['test_set_label_'], {}), '(test_set_label_)\n', (7459, 7476), True, 'import numpy as np\n'), ((7500, 7523), 'numpy.array', 'np.array', (['test_set_cls_'], {}), '(test_set_cls_)\n', (7508, 7523), True, 'import numpy as np\n'), ((7677, 7713), 'random.sample', 'random.sample', (['correct_list_test', '(10)'], {}), '(correct_list_test, 10)\n', (7690, 7713), False, 'import random\n'), ((7730, 7765), 'random.sample', 'random.sample', (['false_list_test', '(190)'], {}), '(false_list_test, 190)\n', (7743, 7765), False, 'import random\n'), ((8554, 8579), 'numpy.array', 'np.array', (['test_set_label_'], {}), '(test_set_label_)\n', (8562, 8579), True, 'import numpy as np\n'), ((8603, 8626), 'numpy.array', 'np.array', (['test_set_cls_'], {}), '(test_set_cls_)\n', (8611, 8626), True, 'import numpy as np\n'), ((1437, 1458), 'numpy.zeros', 'np.zeros', (['(num, h, w)'], {}), '((num, h, w))\n', (1445, 1458), True, 'import numpy as np\n'), ((1489, 1513), 'numpy.zeros', 'np.zeros', (['(num, h, w, 3)'], {}), '((num, h, w, 3))\n', (1497, 1513), True, 'import numpy as np\n'), ((1605, 1635), 'PIL.Image.open', 'Image.open', (['pngs[train_num[i]]'], {}), '(pngs[train_num[i]])\n', (1615, 1635), False, 'from PIL import Image\n'), ((2416, 2437), 'numpy.zeros', 'np.zeros', (['(num, h, w)'], {}), '((num, h, w))\n', (2424, 2437), True, 'import numpy as np\n'), ((2468, 2492), 'numpy.zeros', 'np.zeros', (['(num, h, w, 3)'], {}), '((num, h, w, 3))\n', (2476, 2492), True, 'import numpy as np\n'), ((2592, 2622), 'PIL.Image.open', 'Image.open', (['pngs[train_num[i]]'], {}), '(pngs[train_num[i]])\n', (2602, 2622), False, 'from PIL import Image\n'), ((3493, 3514), 'numpy.zeros', 'np.zeros', (['(num, h, w)'], {}), '((num, h, w))\n', (3501, 3514), True, 'import numpy as np\n'), ((3545, 3569), 'numpy.zeros', 'np.zeros', (['(num, h, w, 3)'], {}), '((num, h, w, 3))\n', (3553, 3569), True, 'import numpy as np\n'), ((3669, 3699), 'PIL.Image.open', 'Image.open', (['pngs[train_num[i]]'], {}), '(pngs[train_num[i]])\n', (3679, 3699), False, 'from PIL import Image\n'), ((4572, 4593), 'numpy.zeros', 'np.zeros', (['(num, h, w)'], {}), '((num, h, w))\n', (4580, 4593), True, 'import numpy as np\n'), ((4624, 4648), 'numpy.zeros', 'np.zeros', (['(num, h, w, 3)'], {}), '((num, h, w, 3))\n', (4632, 4648), True, 'import numpy as np\n'), ((4749, 4779), 'PIL.Image.open', 'Image.open', (['pngs[train_num[i]]'], {}), '(pngs[train_num[i]])\n', (4759, 4779), False, 'from PIL import Image\n'), ((5640, 5665), 'numpy.zeros', 'np.zeros', (['(num * 2, h, w)'], {}), '((num * 2, h, w))\n', (5648, 5665), True, 'import numpy as np\n'), ((5693, 5721), 'numpy.zeros', 'np.zeros', (['(num * 2, h, w, 3)'], {}), '((num * 2, h, w, 3))\n', (5701, 5721), True, 'import numpy as np\n'), ((5811, 5840), 'PIL.Image.open', 'Image.open', (['pngs[test_num[i]]'], {}), '(pngs[test_num[i]])\n', (5821, 5840), False, 'from PIL import Image\n'), ((6781, 6802), 'numpy.zeros', 'np.zeros', (['(num, h, w)'], {}), '((num, h, w))\n', (6789, 6802), True, 'import numpy as np\n'), ((6832, 6856), 'numpy.zeros', 'np.zeros', (['(num, h, w, 3)'], {}), '((num, h, w, 3))\n', (6840, 6856), True, 'import numpy as np\n'), ((6946, 6975), 'PIL.Image.open', 'Image.open', (['pngs[test_num[i]]'], {}), '(pngs[test_num[i]])\n', (6956, 6975), False, 'from PIL import Image\n'), ((7884, 7905), 'numpy.zeros', 'np.zeros', (['(num, h, w)'], {}), '((num, h, w))\n', (7892, 7905), True, 'import numpy as np\n'), ((7935, 7959), 'numpy.zeros', 'np.zeros', (['(num, h, w, 3)'], {}), '((num, h, w, 3))\n', (7943, 7959), True, 'import numpy as np\n'), ((8049, 8078), 'PIL.Image.open', 'Image.open', (['pngs[test_num[i]]'], {}), '(pngs[test_num[i]])\n', (8059, 8078), False, 'from PIL import Image\n'), ((1833, 1851), 'numpy.array', 'np.array', (['img_grey'], {}), '(img_grey)\n', (1841, 1851), True, 'import numpy as np\n'), ((2820, 2838), 'numpy.array', 'np.array', (['img_grey'], {}), '(img_grey)\n', (2828, 2838), True, 'import numpy as np\n'), ((3898, 3916), 'numpy.array', 'np.array', (['img_grey'], {}), '(img_grey)\n', (3906, 3916), True, 'import numpy as np\n'), ((4978, 4996), 'numpy.array', 'np.array', (['img_grey'], {}), '(img_grey)\n', (4986, 4996), True, 'import numpy as np\n'), ((6037, 6055), 'numpy.array', 'np.array', (['img_grey'], {}), '(img_grey)\n', (6045, 6055), True, 'import numpy as np\n'), ((7172, 7190), 'numpy.array', 'np.array', (['img_grey'], {}), '(img_grey)\n', (7180, 7190), True, 'import numpy as np\n'), ((8275, 8293), 'numpy.array', 'np.array', (['img_grey'], {}), '(img_grey)\n', (8283, 8293), True, 'import numpy as np\n')] |
# Generated by Django 3.2.4 on 2021-08-03 15:02
from django.contrib.sites.management import create_default_site
from django.db import migrations
def create_privacy_policy_flatpage(apps, schema_editor):
Site = apps.get_model("sites", "Site")
site = Site.objects.first()
if not site:
create_default_site(None)
site = Site.objects.first()
FlatPage = apps.get_model("flatpages", "FlatPage")
page, created = \
FlatPage.objects.get_or_create(url='/privacy-policy/',
defaults={'title': 'Privacy Policy'})
if created:
page.sites.add(site)
def delete_privacy_policy_flatpage(apps, schema_editor):
FlatPage = apps.get_model("flatpages", "FlatPage")
FlatPage.objects.filter(url='/privacy-policy/').delete()
class Migration(migrations.Migration):
dependencies = [
('playlist_creation', '0001_initial'),
("sites", "0002_alter_domain_unique")
]
operations = [
migrations.RunPython(create_privacy_policy_flatpage,
delete_privacy_policy_flatpage),
]
| [
"django.contrib.sites.management.create_default_site",
"django.db.migrations.RunPython"
] | [((304, 329), 'django.contrib.sites.management.create_default_site', 'create_default_site', (['None'], {}), '(None)\n', (323, 329), False, 'from django.contrib.sites.management import create_default_site\n'), ((994, 1082), 'django.db.migrations.RunPython', 'migrations.RunPython', (['create_privacy_policy_flatpage', 'delete_privacy_policy_flatpage'], {}), '(create_privacy_policy_flatpage,\n delete_privacy_policy_flatpage)\n', (1014, 1082), False, 'from django.db import migrations\n')] |
# Generated by Django 3.2.8 on 2021-10-15 16:58
from django.db import migrations
class Migration(migrations.Migration):
dependencies = [
('app', '0038_auto_20211012_1634'),
]
operations = [
migrations.RenameField(
model_name='device',
old_name='erase_wifi_credentials_at',
new_name='last_event_sent_at',
),
migrations.RemoveField(
model_name='device',
name='zero_config_at',
),
migrations.RenameField(
model_name='devicelogentry',
old_name='erase_wifi_credentials_at',
new_name='last_event_sent_at',
),
migrations.RemoveField(
model_name='devicelogentry',
name='zero_config_at',
),
] | [
"django.db.migrations.RemoveField",
"django.db.migrations.RenameField"
] | [((223, 340), 'django.db.migrations.RenameField', 'migrations.RenameField', ([], {'model_name': '"""device"""', 'old_name': '"""erase_wifi_credentials_at"""', 'new_name': '"""last_event_sent_at"""'}), "(model_name='device', old_name=\n 'erase_wifi_credentials_at', new_name='last_event_sent_at')\n", (245, 340), False, 'from django.db import migrations\n'), ((392, 458), 'django.db.migrations.RemoveField', 'migrations.RemoveField', ([], {'model_name': '"""device"""', 'name': '"""zero_config_at"""'}), "(model_name='device', name='zero_config_at')\n", (414, 458), False, 'from django.db import migrations\n'), ((503, 628), 'django.db.migrations.RenameField', 'migrations.RenameField', ([], {'model_name': '"""devicelogentry"""', 'old_name': '"""erase_wifi_credentials_at"""', 'new_name': '"""last_event_sent_at"""'}), "(model_name='devicelogentry', old_name=\n 'erase_wifi_credentials_at', new_name='last_event_sent_at')\n", (525, 628), False, 'from django.db import migrations\n'), ((680, 754), 'django.db.migrations.RemoveField', 'migrations.RemoveField', ([], {'model_name': '"""devicelogentry"""', 'name': '"""zero_config_at"""'}), "(model_name='devicelogentry', name='zero_config_at')\n", (702, 754), False, 'from django.db import migrations\n')] |
from rest_framework.views import APIView
from rest_framework.response import Response
from rest_framework import status
from rest_framework import viewsets
from rest_framework.authentication import TokenAuthentication
from rest_framework import filters
from profiles_api import serializers
from profiles_api import models
from profiles_api import permissions
class HelloApiView(APIView):
"""Test API View"""
serializer_class = serializers.HelloSerializer
def get(self,request, format=None):
"""returns a list of APIView features"""
an_apiview = [
'uses HTTTP mothod as function(get, post, patch, put, delete) ',
'Is simmilar to Traditional Django View',
'Gives you the most control over your application logic',
'Is mapped manually to URLs',
]
return Response({'message':'Hello!','an_apiview':an_apiview})
def post(self, request):
"""Create a hello message with our name"""
serializer = self.serializer_class(data=request.data)
if serializer.is_valid():
name = serializer.validated_data.get('name')
message = f'Hello {name}'
return Response({'message':message})
else:
return Response(
serializer.errors,
status = status.HTTP_400_BAD_REQUEST
)
def put(self, request, pk=None):
"""Handle updating an object"""
return Response({'method': 'PUT'})
def patch(self, request, pk=None):
"""Handle partial update of object"""
return Response({'method': 'PATCH'})
def delete(self, request, pk=None):
"""Delete an object"""
return Response({'method': 'DELETE'})
class HelloViewSet(viewsets.ViewSet):
"""Test API ViewSet"""
serializer_class = serializers.HelloSerializer
def list(self, request):
"""Return a hello message."""
a_viewset = [
'Uses actions (list, create, retrieve, update, partial_update)',
'Automatically maps to URLS using Routers',
'Provides more functionality with less code',
]
return Response({'message': 'Hello!', 'a_viewset': a_viewset})
def create(self, request):
"""create a new hello mesage"""
serializer= self.serializer_class(data=request.data)
if serializer.is_valid():
name = serializer.validated_data.get('name')
message = f'hello{name}!'
return Response({'message':message})
else:
return Response(
serializer.errors,
status = status.HTTP_400_BAD_REQUEST
)
def retrieve(self, request,pk=None):
"""habdle getting an object by its id"""
return Response({'http_method': 'GET'})
def update(self, request, pk=None):
"""handle updateing an object"""
return Response({'http_method':'PUT'})
def partial_update(self, request, pk=None):
"""handle updating part of an object"""
return Response({'http_method':'PATCH'})
def destroy(self, request, pk=None):
"""handle removing an object"""
return Response({'http_method':'DELETE'})
class UserProfileViewSet(viewsets.ModelViewSet):
"""handle creating and updating profiles"""
serializer_class = serializers.UserProfileSerializer
queryset = models.UserProfile.objects.all()
authentication_classes = (TokenAuthentication,)
permission_class = (permissions.UpdateOwnProfile,)
filter_backends = (filters.SearchFilter,)
search_fields = ('name', 'email',)
| [
"rest_framework.response.Response",
"profiles_api.models.UserProfile.objects.all"
] | [((3382, 3414), 'profiles_api.models.UserProfile.objects.all', 'models.UserProfile.objects.all', ([], {}), '()\n', (3412, 3414), False, 'from profiles_api import models\n'), ((834, 891), 'rest_framework.response.Response', 'Response', (["{'message': 'Hello!', 'an_apiview': an_apiview}"], {}), "({'message': 'Hello!', 'an_apiview': an_apiview})\n", (842, 891), False, 'from rest_framework.response import Response\n'), ((1449, 1476), 'rest_framework.response.Response', 'Response', (["{'method': 'PUT'}"], {}), "({'method': 'PUT'})\n", (1457, 1476), False, 'from rest_framework.response import Response\n'), ((1578, 1607), 'rest_framework.response.Response', 'Response', (["{'method': 'PATCH'}"], {}), "({'method': 'PATCH'})\n", (1586, 1607), False, 'from rest_framework.response import Response\n'), ((1695, 1725), 'rest_framework.response.Response', 'Response', (["{'method': 'DELETE'}"], {}), "({'method': 'DELETE'})\n", (1703, 1725), False, 'from rest_framework.response import Response\n'), ((2158, 2213), 'rest_framework.response.Response', 'Response', (["{'message': 'Hello!', 'a_viewset': a_viewset}"], {}), "({'message': 'Hello!', 'a_viewset': a_viewset})\n", (2166, 2213), False, 'from rest_framework.response import Response\n'), ((2771, 2803), 'rest_framework.response.Response', 'Response', (["{'http_method': 'GET'}"], {}), "({'http_method': 'GET'})\n", (2779, 2803), False, 'from rest_framework.response import Response\n'), ((2901, 2933), 'rest_framework.response.Response', 'Response', (["{'http_method': 'PUT'}"], {}), "({'http_method': 'PUT'})\n", (2909, 2933), False, 'from rest_framework.response import Response\n'), ((3045, 3079), 'rest_framework.response.Response', 'Response', (["{'http_method': 'PATCH'}"], {}), "({'http_method': 'PATCH'})\n", (3053, 3079), False, 'from rest_framework.response import Response\n'), ((3176, 3211), 'rest_framework.response.Response', 'Response', (["{'http_method': 'DELETE'}"], {}), "({'http_method': 'DELETE'})\n", (3184, 3211), False, 'from rest_framework.response import Response\n'), ((1181, 1211), 'rest_framework.response.Response', 'Response', (["{'message': message}"], {}), "({'message': message})\n", (1189, 1211), False, 'from rest_framework.response import Response\n'), ((1244, 1307), 'rest_framework.response.Response', 'Response', (['serializer.errors'], {'status': 'status.HTTP_400_BAD_REQUEST'}), '(serializer.errors, status=status.HTTP_400_BAD_REQUEST)\n', (1252, 1307), False, 'from rest_framework.response import Response\n'), ((2496, 2526), 'rest_framework.response.Response', 'Response', (["{'message': message}"], {}), "({'message': message})\n", (2504, 2526), False, 'from rest_framework.response import Response\n'), ((2560, 2623), 'rest_framework.response.Response', 'Response', (['serializer.errors'], {'status': 'status.HTTP_400_BAD_REQUEST'}), '(serializer.errors, status=status.HTTP_400_BAD_REQUEST)\n', (2568, 2623), False, 'from rest_framework.response import Response\n')] |
# -*- coding: utf-8 -*-
import random, json, string, warnings
from typing import Union, List, Set, Dict
from simalign import SentenceAligner
from nltk.translate.phrase_based import phrase_extraction
class CodeMixer(object):
def __init__(self, matrix_lg: str, embedded_lgs: Union[List[str],Set[str]], device='cuda', precomputed_phrases=False):
warnings.filterwarnings("ignore", category=FutureWarning)
if not precomputed_phrases:
self.aligner = SentenceAligner(model="xlmr", token_type="bpe", matching_methods="m", device=device)
self.rtl_lgs = {'ar', 'he'}
self.matrix_lg = matrix_lg
self.embedded_lgs = set(embedded_lgs)
def get_phrases(self, matrix_sentence: str, translations: Dict[str,str],
sample_lgs: Union[List[str],Set[str]]=None):
if not sample_lgs:
sample_lgs = self.embedded_lgs
filtered_translations = {k: v for k,v in translations.items() if k in sample_lgs}
matrix_tokens = matrix_sentence.split()
phrases = []
for lg, embedded_sentence in filtered_translations.items():
tokenized_embedded = embedded_sentence.split()
alignments = self.aligner.get_word_aligns(matrix_tokens, tokenized_embedded)
candidate_phrases = phrase_extraction(matrix_sentence, embedded_sentence, alignments['mwmf'], 4)
candidate_phrases = [(p[0], p[1], p[2], p[3], lg) for p in candidate_phrases if p[2][0] not in string.punctuation]
if lg == 'zh':
candidate_phrases = [(p[0], p[1], p[2], p[3].replace(' ', ''), p[4]) for p in candidate_phrases]
if lg == 'th':
candidate_phrases = [(p[0], p[1], p[2], p[3].replace(' ', ' ').replace(' ,', ','), p[4]) for p in candidate_phrases]
phrases += candidate_phrases
sorted_phrases = sorted(phrases, key=lambda x: (x[0][0], -x[0][1]))
grouped_phrases = {i:[] for i in range(len(matrix_tokens))}
for phrase in sorted_phrases:
grouped_phrases[phrase[0][0]].append(phrase)
return grouped_phrases
def swap_phrase(self, tokens, replace_start_idx, replace_end_idx, to_replace):
return tokens[0:replace_start_idx] + [to_replace] + tokens[replace_end_idx:]
def get_weights(self, lg_counts):
filtered_lg_counts = {k: v for k,v in lg_counts.items() if k in self.embedded_lgs or k == self.matrix_lg}
total_count = sum(filtered_lg_counts.values())
return {k: v/total_count for k,v in filtered_lg_counts.items()}
def generate(self, sentence, reference_translations, probability=0.15, lg_counts: Dict[str,int]=None):
phrases = self.get_phrases(sentence, reference_translations)
return generate_precomputed_alignments(sentence, phrases, probability, lg_counts)
def generate_precomputed_alignments(self, sentence, phrase_alignments, probability=0.15):#, lg_counts: Dict[str,int]=None):
tokens = sentence.split()
token_length = len(tokens)
pos = 0
prev_lg = self.matrix_lg
prev_replacement_pos = pos
while pos < token_length:
candidates = phrase_alignments.get(pos)
pos += 1
if random.random() >= probability or not candidates:
prev_lg = self.matrix_lg
continue
eligible_candidates = []
for candidate in candidates:
phrase_to_replace = candidate[2]
replacement = candidate[3]
replacement_lg = candidate[4]
replace_start_idx = candidate[0][0] - token_length
replace_end_idx = candidate[0][1] - token_length
if phrase_to_replace.split() != tokens[replace_start_idx:replace_end_idx]:
continue
if replacement_lg not in self.rtl_lgs and replacement_lg == prev_lg and candidate[1][1] <= prev_replacement_pos:
continue
eligible_candidates.append(candidate)
if eligible_candidates:
chosen_candidate = random.choice(eligible_candidates)
replacement_lg = chosen_candidate[4]
replacement = chosen_candidate[3]
replace_start_idx = chosen_candidate[0][0] - token_length
replace_end_idx = chosen_candidate[0][1] - token_length
tokens = self.swap_phrase(tokens, replace_start_idx, replace_end_idx, replacement)
prev_lg = replacement_lg
prev_replacement_pos = pos
pos = max(replace_end_idx, pos)
else:
prev_lg = self.matrix_lg
continue
return ' '.join(tokens)
| [
"random.choice",
"nltk.translate.phrase_based.phrase_extraction",
"random.random",
"warnings.filterwarnings",
"simalign.SentenceAligner"
] | [((357, 414), 'warnings.filterwarnings', 'warnings.filterwarnings', (['"""ignore"""'], {'category': 'FutureWarning'}), "('ignore', category=FutureWarning)\n", (380, 414), False, 'import random, json, string, warnings\n'), ((478, 566), 'simalign.SentenceAligner', 'SentenceAligner', ([], {'model': '"""xlmr"""', 'token_type': '"""bpe"""', 'matching_methods': '"""m"""', 'device': 'device'}), "(model='xlmr', token_type='bpe', matching_methods='m',\n device=device)\n", (493, 566), False, 'from simalign import SentenceAligner\n'), ((1301, 1377), 'nltk.translate.phrase_based.phrase_extraction', 'phrase_extraction', (['matrix_sentence', 'embedded_sentence', "alignments['mwmf']", '(4)'], {}), "(matrix_sentence, embedded_sentence, alignments['mwmf'], 4)\n", (1318, 1377), False, 'from nltk.translate.phrase_based import phrase_extraction\n'), ((4118, 4152), 'random.choice', 'random.choice', (['eligible_candidates'], {}), '(eligible_candidates)\n', (4131, 4152), False, 'import random, json, string, warnings\n'), ((3246, 3261), 'random.random', 'random.random', ([], {}), '()\n', (3259, 3261), False, 'import random, json, string, warnings\n')] |
#! python3
# To write files and make a directory
import os
# To filter string
import re
# For sleeping
import time
# Googles Text to speech lib
from gtts import gTTS
# File info reader, in this case, it is for determening the length of a mp3
from mutagen.mp3 import MP3
def format_text(text: str):
"""
Removes unwanted characters
Replaces spaces with underscores
"""
whitelist = re.compile(r'[^a-zA-Z ]+')
return re.sub(whitelist, '', text).replace(' ', '_')
def create_dir(dir_name: str):
"""Creates directory when not made yet"""
if not os.path.isdir(dir_name):
os.makedirs(dir_name)
def create_spell_out(text: str):
"""Maps the characters to the NATO words"""
words_for_letters = ['Alfa', 'Bravo', 'Charlie', 'Delta', 'Echo', 'Foxtrot',
'Golf', 'Hotel', 'India', 'Juliett', 'Kilo', 'Lima',
'Mike', 'November', 'Oscar', 'Papa', 'Quebec', 'Romeo',
'Sierra', 'Tango', 'Uniform', 'Victor', 'Whiskey',
'X-ray', 'Yankee', 'Zulu', 'underscore']
# ord('{') - 97 == words_for_letters.index('underscore')
words = [word.replace('_', '{') for word in text.split(' ')]
return [words_for_letters[ord(
letter.lower()) - 97] for word in words for letter in word]
def save_audio(text: str, filename: str, dir: str):
"""
Converts text to audio and saves
Notes
-----
If the .mp3 file extension is missing in the filename, it will be added
If a file with the same name exists, it will not save, only notify the user
Returns
_______
Path : str
"""
# Make the path to the folder
path = '{0}/{1}'.format(dir, filename)
if not filename.endswith('.mp3'):
path += '.mp3'
# Generates and saves audio file
tts = gTTS(text=text, lang='en')
# Only saves when file does not exist
if os.path.isfile(path):
print("File named {0} already exist, will not safe".format(path))
else:
tts.save(path)
return path
def play_audio(path: str):
os.startfile(os.getcwd() + path[1:])
duration = MP3(path).info.length
time.sleep(duration)
# ---------- MAIN-PROGRAM ----------
if __name__ == '__main__':
print('Welcome!')
output_dir = './audio'
create_dir(output_dir)
while True:
text = format_text(input('What is the sentence?'))
output_words = 'Message incoming' + ', '.join(create_spell_out(text))
path = save_audio(output_words, text, output_dir)
play_audio(path)
if(input('Continue? y/n') not in ('y', 'yes')):
break
| [
"os.makedirs",
"re.compile",
"time.sleep",
"os.getcwd",
"os.path.isfile",
"mutagen.mp3.MP3",
"os.path.isdir",
"gtts.gTTS",
"re.sub"
] | [((406, 431), 're.compile', 're.compile', (['"""[^a-zA-Z ]+"""'], {}), "('[^a-zA-Z ]+')\n", (416, 431), False, 'import re\n'), ((1843, 1869), 'gtts.gTTS', 'gTTS', ([], {'text': 'text', 'lang': '"""en"""'}), "(text=text, lang='en')\n", (1847, 1869), False, 'from gtts import gTTS\n'), ((1920, 1940), 'os.path.isfile', 'os.path.isfile', (['path'], {}), '(path)\n', (1934, 1940), False, 'import os\n'), ((2176, 2196), 'time.sleep', 'time.sleep', (['duration'], {}), '(duration)\n', (2186, 2196), False, 'import time\n'), ((580, 603), 'os.path.isdir', 'os.path.isdir', (['dir_name'], {}), '(dir_name)\n', (593, 603), False, 'import os\n'), ((613, 634), 'os.makedirs', 'os.makedirs', (['dir_name'], {}), '(dir_name)\n', (624, 634), False, 'import os\n'), ((444, 471), 're.sub', 're.sub', (['whitelist', '""""""', 'text'], {}), "(whitelist, '', text)\n", (450, 471), False, 'import re\n'), ((2111, 2122), 'os.getcwd', 'os.getcwd', ([], {}), '()\n', (2120, 2122), False, 'import os\n'), ((2150, 2159), 'mutagen.mp3.MP3', 'MP3', (['path'], {}), '(path)\n', (2153, 2159), False, 'from mutagen.mp3 import MP3\n')] |
#!/usr/bin/env python
import json
jsonFile = open('temp.json', 'r')
jsonData = json.load(jsonFile)
places = jsonData["places"]
for item in places:
originalArray = places[item]
newArray = ["default"]
for originalPiece in originalArray:
newArray.append(originalPiece)
places[item] = newArray
jsonData["places"] = places
newJsonFile = open('newStuff.json', 'w')
json.dump(jsonData, newJsonFile)
jsonFile.close()
newJsonFile.close()
| [
"json.load",
"json.dump"
] | [((81, 100), 'json.load', 'json.load', (['jsonFile'], {}), '(jsonFile)\n', (90, 100), False, 'import json\n'), ((372, 404), 'json.dump', 'json.dump', (['jsonData', 'newJsonFile'], {}), '(jsonData, newJsonFile)\n', (381, 404), False, 'import json\n')] |
from setuptools import setup, Extension
setup(ext_modules=[
Extension('_module', sources=["module_wrap.c"])
]) | [
"setuptools.Extension"
] | [((65, 112), 'setuptools.Extension', 'Extension', (['"""_module"""'], {'sources': "['module_wrap.c']"}), "('_module', sources=['module_wrap.c'])\n", (74, 112), False, 'from setuptools import setup, Extension\n')] |
from gaesessions import SessionMiddleware
# suggestion: generate your own random key using os.urandom(64)
# WARNING: Make sure you run os.urandom(64) OFFLINE and copy/paste the output to
# this file. If you use os.urandom() to *dynamically* generate your key at
# runtime then any existing sessions will become junk every time you start,
# deploy, or update your app!
import os
# https://cloud.google.com/appengine/docs/standard/python/tools/using-libraries-python-27
# appengine_config.py
from google.appengine.ext import vendor
# Add any libraries install in the "lib" folder.
# vendor.add('lib')
vendor.add(os.path.join(os.path.dirname(os.path.realpath(__file__)), 'lib'))
# def add_vendor_packages(vendor_folder):
# """
# Adds our vendor packages folder to sys.path so that third-party
# packages can be imported.
# """
# import site
# import os.path
# import sys
#
# # Use site.addsitedir() because it appropriately reads .pth
# # files for namespaced packages. Unfortunately, there's not an
# # option to choose where addsitedir() puts its paths in sys.path
# # so we have to do a little bit of magic to make it play along.
#
# # We're going to grab the current sys.path and split it up into
# # the first entry and then the rest. Essentially turning
# # ['.', '/site-packages/x', 'site-packages/y']
# # into
# # ['.'] and ['/site-packages/x', 'site-packages/y']
# # The reason for this is we want '.' to remain at the top of the
# # list but we want our vendor files to override everything else.
# sys.path, remainder = sys.path[:1], sys.path[1:]
#
# # Now we call addsitedir which will append our vendor directories
# # to sys.path (which was truncated by the last step.)
# site.addsitedir(os.path.join(os.path.dirname(__file__), vendor_folder))
#
# # Finally, we'll add the paths we removed back.
# sys.path.extend(remainder)
#
# # Change 'lib' to whichever directory you use for your vendored packages.
# add_vendor_packages('lib')
COOKIE_KEY = "" # TODO: Generate new COOKIE_KEY using os.urandom(64) See above
def webapp_add_wsgi_middleware(app):
from google.appengine.ext.appstats import recording
app = SessionMiddleware(app, cookie_key=COOKIE_KEY)
app = recording.appstats_wsgi_middleware(app)
return app
| [
"os.path.realpath",
"google.appengine.ext.appstats.recording.appstats_wsgi_middleware",
"gaesessions.SessionMiddleware"
] | [((2228, 2273), 'gaesessions.SessionMiddleware', 'SessionMiddleware', (['app'], {'cookie_key': 'COOKIE_KEY'}), '(app, cookie_key=COOKIE_KEY)\n', (2245, 2273), False, 'from gaesessions import SessionMiddleware\n'), ((2282, 2321), 'google.appengine.ext.appstats.recording.appstats_wsgi_middleware', 'recording.appstats_wsgi_middleware', (['app'], {}), '(app)\n', (2316, 2321), False, 'from google.appengine.ext.appstats import recording\n'), ((642, 668), 'os.path.realpath', 'os.path.realpath', (['__file__'], {}), '(__file__)\n', (658, 668), False, 'import os\n')] |
import flask
blueprint = flask.Blueprint('rooms', __name__, template_folder='templates')
from . import routes
| [
"flask.Blueprint"
] | [((26, 89), 'flask.Blueprint', 'flask.Blueprint', (['"""rooms"""', '__name__'], {'template_folder': '"""templates"""'}), "('rooms', __name__, template_folder='templates')\n", (41, 89), False, 'import flask\n')] |
# Debugging
#T# Table of contents
#C# Python debugger (pdb)
#T# Beginning of content
#C# Python debugger (pdb)
# |-------------------------------------------------------------
#T# pdb is the builtin Python debugger, it has breakpoints, stepping through the code, printing the values of variables, post-mortem debugging, debugging of modules, functions, scripts, among other features
#T# pdb can be executed with an script argument to debug said script, the following syntax is done in the operating system shell
# SYNTAX python3 -m pdb script1.py
#T# python3 is the Python executable, -m pdb script1.py runs pdb to debug script1.py (see the file titled Interpreter), this automatically enters post-mortem if script1 crashes
#T# the pdb module is imported to use the pdb debugger as part of a script
import pdb
#T# the run function of the pdb module allows debugging the execution of a Python string
# SYNTAX pdb.run('string1')
#T# the pdb debugger is started right before the execution of string1, and is used to debug whatever string1 executes
pdb.run('import S01_Basic_syntax') # this debugs the S01_Basic_syntax.py file, because the import statement executes the imported module
#T# the following code is used to show the syntax of the pdb debugger in its interactive mode
output_var1 = "help variable to show the different output of the pdb debugger"
var1 = [5, 2, 3]
var2 = 7
def func1(num1, num2):
num3 = num1 + num2
print("func1_string1")
return num3
def func2():
func1(var1[0], var1[2])
loc1 = 72
for i1 in [1, 2, 3]:
print("i1 is", i1)
func2()
#T# create a breakpoint with the breakpoint function, this starts the (Pdb) interpreter to do interactive debugging
breakpoint()
# |--------------------------------------------------\
#T# the following syntaxes are written in the pdb debugger language, so they can't be written outside of a comment because they are not valid Python syntax and this .py file would show errors in an IDE (IDE stands for Integrated Development Environment)
#T# the 'output_var1' variable used in the following is used as a helper to signal and display the output of the pdb debugger commands
#T# the pdb debugger prompt is (Pdb), so anything shown after a (Pdb) means that it was typed in said prompt, e.g. '(Pdb) prompt_typings1' prompt_typings1 was typed directly in the pdb debugger prompt
#T# when source code is printed, the current line is shown with '->' after the line number
# SYNTAX next
#T# the next command executes code up to the next line of code (not entering functions)
output_var1 # (Pdb) next # this shows output of the script or program under debugging
# SYNTAX step
#T# the step command steps into functions or the next line
output_var1 # (Pdb) step # this shows output of the script or program under debugging, possibly inside a function
# SYNTAX continue
#T# the continue command continues execution until a breakpoint is found
output_var1 # (Pdb) continue # this shows output of the script or program under debugging, up to the next breakpoint or the end of the file
# SYNTAX until int1
#T# the until command continues execution until a line of number int1 or greater is reached, without int1 it continues until the next bigger line number
output_var1 # (Pdb) until # this shows output of the script or program under debugging, up to the next bigger line number
# SYNTAX return
#T# the return command continues execution until arriving at the return keyword of the current function, so this is used inside functions
output_var1 # (Pdb) return # inside a function, this shows output of the script or program under debugging, up to the return keyword of the current function
# SYNTAX run
# SYNTAX restart
#T# the run command and its alias the restart command restart the script or program under debugging, preserving the options and created breakpoints
output_var1 # (Pdb) restart # the script or program restarts
# SYNTAX p var1
#T# the p command (for print) prints the value of var1, if the name var1 is defined, this syntax is an alias for print(var1)
output_var1 # (Pdb) p var1 # [5, 2, 3]
output_var1 # (Pdb) print(var1) # [5, 2, 3]
# SYNTAX p func1
#T# same as before, but when used with a function, its address is printed
output_var1 # (Pdb) p func1 # <function func1 at 0x7f35210401f0> # or similar
# SYNTAX p func1(arg1, arg2)
#T# same as before, but this prints any output from func1 and its return value using arg1, arg2 as arguments, and any other arguments present
output_var1 # (Pdb) p func1(2, 3)
#T# the former prints
# func1_string1
# 5
# SYNTAX args
#T# the args command displays the arguments passed to a function
output_var1 # (Pdb) args # inside func1(5, 3)
#T# the former prints
# num1 = 5
# num2 = 3
# SYNTAX display var1
#T# the display command prints a variable each time it changes
# SYNTAX undisplay var1
#T# stop displaying a variable var1 with the undisplay command
# SYNTAX l int1
#T# the l command lists 11 source code lines, this is done around line int1, 5 lines above and 5 lines below it, this syntax is an alias for list int1
output_var1 # (Pdb) l 7 # (Pdb) list 7
#T# the former prints
# 2 # Debugging
# 3
# 4 #T# Table of contents
# 5
# 6 #C# Python debugger (pdb)
# 7
# 8 #T# Beginning of content
# 9
# 10 #C# Python debugger (pdb)
# 11
# 12 # |-------------------------------------------------------------
# SYNTAX l .
#T# same as before, but list source code lines around the current line
output_var1 # (Pdb) l . # this prints similar as before
# SYNTAX ll
#T# the ll commands does a long list of the source code local to the current line
output_var1 # (Pdb) ll # the output is too large to put here, more than 150 lines
# SYNTAX break
#T# the break command alone displays all breakpoints
output_var1 # (Pdb) break # with two breakpoints already created
#T# the former prints
# Num Type Disp Enb Where
# 1 breakpoint keep yes at /path/to/S13_Debugging.py:20
# 2 breakpoint keep yes at /path/to/S13_Debugging.py:35
# stop only if var2 < 10
# SYNTAX break file1:int1
#T# the break command is used to create breakpoints in file1 (the current python script name without the .py extension), in line int1
output_var1 # (Pdb) break S13_Debugging:20 # Breakpoint 1 at /path/to/S13_Debugging.py:20
# SYNTAX break file1.func1, condition1
#T# same as before, but the breakpoint is created in the first line of func1 (its def line), and the breakpoint only activates if condition1 evaluates to True using Python boolean syntax
output_var1 # (Pdb) break S13_Debugging.func1, var2 < 10 # Breakpoint 2 at /path/to/S13_Debugging.py:35
# SYNTAX disable int1
#T# the disable command disables the breakpoint numbered with the number int1
output_var1 # (Pdb) disable 1 # Disabled breakpoint 1 at /path/to/S13_Debugging.py:20
# SYNTAX enable int1
#T# the enable command enables the breakpoint numbered with the number int1
output_var1 # (Pdb) enable 1 # Enabled breakpoint 1 at /path/to/S13_Debugging.py:20
# SYNTAX clear int1
#T# the clear command completely deletes a breakpoint
output_var1 # (Pdb) clear 1 # Deleted breakpoint 1 at /path/to/S13_Debugging.py:20
# SYNTAX where
#T# the where command prints the stack_frame trace
output_var1 # (Pdb) where
#T# the former prints
# /path/to/S13_Debugging.py(46)<module>()
#-> func2()
# /path/to/S13_Debugging.py(42)func2()
#-> func1(var1[0], var1[2])
#> /path/to/S13_Debugging.py(40)func1()->8
#-> return num3
# SYNTAX up int1
#T# the up command goes up to an older frame in the stack trace, the amount of frames that go up is int1
output_var1 # (Pdb) up 1
#T# the former prints
#> /path/to/S13_Debugging.py(42)func2()
#-> func1(var1[0], var1[2])
# SYNTAX down int1
#T# the down command goes down to a newer frame in the stack trace, the amount of frames that go down is int1
output_var1 # (Pdb) down 1 # *** Newest frame # this is the output at the lowest frame
# SYNTAX help
#T# print the debugger pdb help with the help command
# SYNTAX quit
#T# quit the debugger with the quit command
# |--------------------------------------------------/
# |------------------------------------------------------------- | [
"pdb.run"
] | [((1058, 1092), 'pdb.run', 'pdb.run', (['"""import S01_Basic_syntax"""'], {}), "('import S01_Basic_syntax')\n", (1065, 1092), False, 'import pdb\n')] |
"""
Created on Sat Mar 23 00:23:27 2019
@author: nahid
"""
#https://docs.scipy.org/doc/numpy/reference/generated/numpy.absolute.html
import numpy as np
import matplotlib.pyplot as plt
x = np.array([-1.2, 1.2])
x = np.absolute(x)
print(x)
print(np.absolute(1 + 2j))
#Plot the function over [-10, 10]:
x = np.linspace(-10, 10, 101); #start, end, totalElements you want to create
plt.plot(np.absolute(x))
plt.show()
plt.plot(x)
plt.show()
xx = x + 1j * x[:, np.newaxis]
plt.imshow(np.abs(xx), extent=[-10, 10, -10, 10], cmap='gray')
plt.show() | [
"numpy.abs",
"numpy.absolute",
"matplotlib.pyplot.plot",
"numpy.array",
"numpy.linspace",
"matplotlib.pyplot.show"
] | [((189, 210), 'numpy.array', 'np.array', (['[-1.2, 1.2]'], {}), '([-1.2, 1.2])\n', (197, 210), True, 'import numpy as np\n'), ((215, 229), 'numpy.absolute', 'np.absolute', (['x'], {}), '(x)\n', (226, 229), True, 'import numpy as np\n'), ((306, 331), 'numpy.linspace', 'np.linspace', (['(-10)', '(10)', '(101)'], {}), '(-10, 10, 101)\n', (317, 331), True, 'import numpy as np\n'), ((404, 414), 'matplotlib.pyplot.show', 'plt.show', ([], {}), '()\n', (412, 414), True, 'import matplotlib.pyplot as plt\n'), ((415, 426), 'matplotlib.pyplot.plot', 'plt.plot', (['x'], {}), '(x)\n', (423, 426), True, 'import matplotlib.pyplot as plt\n'), ((427, 437), 'matplotlib.pyplot.show', 'plt.show', ([], {}), '()\n', (435, 437), True, 'import matplotlib.pyplot as plt\n'), ((533, 543), 'matplotlib.pyplot.show', 'plt.show', ([], {}), '()\n', (541, 543), True, 'import matplotlib.pyplot as plt\n'), ((245, 266), 'numpy.absolute', 'np.absolute', (['(1 + 2.0j)'], {}), '(1 + 2.0j)\n', (256, 266), True, 'import numpy as np\n'), ((388, 402), 'numpy.absolute', 'np.absolute', (['x'], {}), '(x)\n', (399, 402), True, 'import numpy as np\n'), ((481, 491), 'numpy.abs', 'np.abs', (['xx'], {}), '(xx)\n', (487, 491), True, 'import numpy as np\n')] |
from __future__ import annotations
import asyncio
import typing
import types
import pandas as pd
import tooltime
from ctc import evm
from ctc import spec
async def async_get_lending_flows(
wallet: spec.Address,
pool_token: spec.ERC20Reference,
protocol: typing.Literal['aave', 'compound', 'rari'],
wallet_deposits: spec.DataFrame | None = None,
deposits: spec.DataFrame | None = None,
wallet_withdrawals: spec.DataFrame | None = None,
withdrawals: spec.DataFrame | None = None,
include_latest: bool = True,
provider: spec.ProviderSpec = None,
replace_symbols: bool = True,
normalize: bool = True,
include_rewards: bool = True,
) -> spec.DataFrame:
if protocol == 'aave':
from ctc.protocols import aave_v2_utils
protocol_module: types.ModuleType = aave_v2_utils
elif protocol == 'compound':
from ctc.protocols import compound_utils
protocol_module = compound_utils
elif protocol == 'rari':
from ctc.protocols import rari_utils
protocol_module = rari_utils
else:
raise Exception('unknown protocol: ' + str(protocol))
df = await _async_create_raw_wallet_flows_df(
wallet=wallet,
wallet_deposits=wallet_deposits,
deposits=deposits,
wallet_withdrawals=wallet_withdrawals,
withdrawals=withdrawals,
include_latest=include_latest,
provider=provider,
)
underlying = await protocol_module.async_get_underlying_asset(
pool_token=pool_token,
provider=provider,
)
# add time data
blocks = df.index.values
blocks_before = blocks - 1
# queue tasks
timestamps_coroutine = evm.async_get_block_timestamps(
blocks=blocks,
provider=provider,
)
timestamps_task = asyncio.create_task(timestamps_coroutine)
pool_token_balances_before_coroutine = (
evm.async_get_erc20_balance_of_by_block(
token=pool_token,
address=wallet,
blocks=blocks_before,
provider=provider,
)
)
pool_token_balances_before_task = asyncio.create_task(
pool_token_balances_before_coroutine
)
pool_token_balances_after_coroutine = (
evm.async_get_erc20_balance_of_by_block(
token=pool_token,
address=wallet,
blocks=blocks,
provider=provider,
)
)
pool_token_balances_after_task = asyncio.create_task(
pool_token_balances_after_coroutine
)
asset_prices_coroutine = protocol_module.async_get_asset_price_by_block(
asset=underlying,
blocks=blocks,
provider=provider,
)
asset_prices_task = asyncio.create_task(asset_prices_coroutine)
# queue optional tasks
if include_rewards:
reward_coroutine = protocol_module.async_compute_wallet_rewards(
wallet=wallet,
blocks=blocks,
provider=provider,
replace_symbol=replace_symbols,
)
reward_task = asyncio.create_task(reward_coroutine)
if normalize:
decimals_coroutine = evm.async_get_erc20_decimals(
underlying,
provider=provider,
)
decimals_task = asyncio.create_task(decimals_coroutine)
if replace_symbols:
underlying_symbol_coroutine = evm.async_get_erc20_symbol(
underlying,
provider=provider,
)
underlying_symbol_task = asyncio.create_task(
underlying_symbol_coroutine
)
pool_token_coroutine = evm.async_get_erc20_symbol(
pool_token,
provider=provider,
)
pool_token_symbol_task = asyncio.create_task(pool_token_coroutine)
# normalize deposits and withdrawals
if normalize:
decimals = await decimals_task
df['asset_deposit'] /= 10 ** decimals
df['asset_withdrawal'] /= 10 ** decimals
# compute time columns
timestamps = await timestamps_task
df.insert(loc=0, column='timestamp', value=timestamps) # type: ignore
df.insert(
loc=1,
column='time',
value=df['timestamp'].map(tooltime.timestamp_to_iso),
)
# add pool token balances
df['pool_token_balance_before'] = await pool_token_balances_before_task
df['pool_token_balance_after'] = await pool_token_balances_after_task
# add underlying balances
df['asset_balance_before'] = df['pool_token_balance_before']
df['asset_balance_after'] = df['pool_token_balance_after']
# add asset price
df['asset_price'] = await asset_prices_task
df['asset_balance_usd'] = df['asset_balance_after'] * df['asset_price']
# add rewards
rewards = await reward_task
for key, value in rewards.items():
df[key] = value
# replace symbols
if replace_symbols:
rename_columns = {}
underlying_symbol = await underlying_symbol_task
pool_token_symbol = await pool_token_symbol_task
for column in df.columns:
if 'asset' in column:
rename_columns[column] = column.replace(
'asset', underlying_symbol
)
if 'pool_token' in column:
rename_columns[column] = column.replace(
'pool_token', pool_token_symbol
)
df = df.rename(columns=rename_columns)
return df
async def _async_create_raw_wallet_flows_df(
wallet: spec.Address,
wallet_deposits: spec.DataFrame | None = None,
deposits: spec.DataFrame | None = None,
wallet_withdrawals: spec.DataFrame | None = None,
withdrawals: spec.DataFrame | None = None,
include_latest: bool = True,
provider: spec.ProviderSpec = None,
) -> spec.DataFrame:
from ctc.protocols import aave_v2_utils
no_deposits = wallet_deposits is None and deposits is None
no_withdrawals = wallet_withdrawals is None and withdrawals is None
if no_deposits and not no_withdrawals:
deposits = await aave_v2_utils.async_get_deposits()
elif not no_deposits and no_withdrawals:
withdrawals = await aave_v2_utils.async_get_withdrawals()
elif no_deposits and no_withdrawals:
deposits, withdrawals = await asyncio.gather(
aave_v2_utils.async_get_deposits(provider=provider),
aave_v2_utils.async_get_withdrawals(provider=provider),
)
wallet = wallet.lower()
if wallet_deposits is None:
if deposits is None:
raise Exception('could not determine deposits')
wallet_deposits = deposits[deposits['arg__user'] == wallet]
if isinstance(wallet_deposits.index, pd.MultiIndex):
wallet_deposits = wallet_deposits.groupby(level='block_number').sum()
if isinstance(wallet_deposits, pd.DataFrame):
wallet_deposits_series = wallet_deposits['arg__amount']
if wallet_withdrawals is None:
if withdrawals is None:
raise Exception('could not determine withdrawals')
wallet_withdrawals = withdrawals[withdrawals['arg__user'] == wallet]
if isinstance(wallet_withdrawals.index, pd.MultiIndex):
wallet_withdrawals = wallet_withdrawals.groupby(
level='block_number'
).sum()
if isinstance(wallet_withdrawals, pd.DataFrame):
wallet_withdrawals_series = wallet_withdrawals['arg__amount']
raw_data = {
'asset_deposit': wallet_deposits_series,
'asset_withdrawal': wallet_withdrawals_series,
}
raw_df = pd.DataFrame(raw_data)
raw_df = raw_df.fillna(0)
if include_latest:
block = await evm.async_get_latest_block_number(provider=provider)
raw_df.loc[block] = [0, 0]
return raw_df
| [
"ctc.evm.async_get_erc20_decimals",
"ctc.protocols.aave_v2_utils.async_get_deposits",
"ctc.protocols.aave_v2_utils.async_get_withdrawals",
"ctc.evm.async_get_block_timestamps",
"ctc.evm.async_get_erc20_balance_of_by_block",
"ctc.evm.async_get_latest_block_number",
"pandas.DataFrame",
"ctc.evm.async_ge... | [((1697, 1761), 'ctc.evm.async_get_block_timestamps', 'evm.async_get_block_timestamps', ([], {'blocks': 'blocks', 'provider': 'provider'}), '(blocks=blocks, provider=provider)\n', (1727, 1761), False, 'from ctc import evm\n'), ((1807, 1848), 'asyncio.create_task', 'asyncio.create_task', (['timestamps_coroutine'], {}), '(timestamps_coroutine)\n', (1826, 1848), False, 'import asyncio\n'), ((1902, 2020), 'ctc.evm.async_get_erc20_balance_of_by_block', 'evm.async_get_erc20_balance_of_by_block', ([], {'token': 'pool_token', 'address': 'wallet', 'blocks': 'blocks_before', 'provider': 'provider'}), '(token=pool_token, address=wallet,\n blocks=blocks_before, provider=provider)\n', (1941, 2020), False, 'from ctc import evm\n'), ((2120, 2177), 'asyncio.create_task', 'asyncio.create_task', (['pool_token_balances_before_coroutine'], {}), '(pool_token_balances_before_coroutine)\n', (2139, 2177), False, 'import asyncio\n'), ((2244, 2355), 'ctc.evm.async_get_erc20_balance_of_by_block', 'evm.async_get_erc20_balance_of_by_block', ([], {'token': 'pool_token', 'address': 'wallet', 'blocks': 'blocks', 'provider': 'provider'}), '(token=pool_token, address=wallet,\n blocks=blocks, provider=provider)\n', (2283, 2355), False, 'from ctc import evm\n'), ((2454, 2510), 'asyncio.create_task', 'asyncio.create_task', (['pool_token_balances_after_coroutine'], {}), '(pool_token_balances_after_coroutine)\n', (2473, 2510), False, 'import asyncio\n'), ((2708, 2751), 'asyncio.create_task', 'asyncio.create_task', (['asset_prices_coroutine'], {}), '(asset_prices_coroutine)\n', (2727, 2751), False, 'import asyncio\n'), ((7505, 7527), 'pandas.DataFrame', 'pd.DataFrame', (['raw_data'], {}), '(raw_data)\n', (7517, 7527), True, 'import pandas as pd\n'), ((3038, 3075), 'asyncio.create_task', 'asyncio.create_task', (['reward_coroutine'], {}), '(reward_coroutine)\n', (3057, 3075), False, 'import asyncio\n'), ((3123, 3182), 'ctc.evm.async_get_erc20_decimals', 'evm.async_get_erc20_decimals', (['underlying'], {'provider': 'provider'}), '(underlying, provider=provider)\n', (3151, 3182), False, 'from ctc import evm\n'), ((3242, 3281), 'asyncio.create_task', 'asyncio.create_task', (['decimals_coroutine'], {}), '(decimals_coroutine)\n', (3261, 3281), False, 'import asyncio\n'), ((3344, 3401), 'ctc.evm.async_get_erc20_symbol', 'evm.async_get_erc20_symbol', (['underlying'], {'provider': 'provider'}), '(underlying, provider=provider)\n', (3370, 3401), False, 'from ctc import evm\n'), ((3470, 3518), 'asyncio.create_task', 'asyncio.create_task', (['underlying_symbol_coroutine'], {}), '(underlying_symbol_coroutine)\n', (3489, 3518), False, 'import asyncio\n'), ((3572, 3629), 'ctc.evm.async_get_erc20_symbol', 'evm.async_get_erc20_symbol', (['pool_token'], {'provider': 'provider'}), '(pool_token, provider=provider)\n', (3598, 3629), False, 'from ctc import evm\n'), ((3698, 3739), 'asyncio.create_task', 'asyncio.create_task', (['pool_token_coroutine'], {}), '(pool_token_coroutine)\n', (3717, 3739), False, 'import asyncio\n'), ((6017, 6051), 'ctc.protocols.aave_v2_utils.async_get_deposits', 'aave_v2_utils.async_get_deposits', ([], {}), '()\n', (6049, 6051), False, 'from ctc.protocols import aave_v2_utils\n'), ((7604, 7656), 'ctc.evm.async_get_latest_block_number', 'evm.async_get_latest_block_number', ([], {'provider': 'provider'}), '(provider=provider)\n', (7637, 7656), False, 'from ctc import evm\n'), ((6125, 6162), 'ctc.protocols.aave_v2_utils.async_get_withdrawals', 'aave_v2_utils.async_get_withdrawals', ([], {}), '()\n', (6160, 6162), False, 'from ctc.protocols import aave_v2_utils\n'), ((6270, 6321), 'ctc.protocols.aave_v2_utils.async_get_deposits', 'aave_v2_utils.async_get_deposits', ([], {'provider': 'provider'}), '(provider=provider)\n', (6302, 6321), False, 'from ctc.protocols import aave_v2_utils\n'), ((6335, 6389), 'ctc.protocols.aave_v2_utils.async_get_withdrawals', 'aave_v2_utils.async_get_withdrawals', ([], {'provider': 'provider'}), '(provider=provider)\n', (6370, 6389), False, 'from ctc.protocols import aave_v2_utils\n')] |
from PIL import Image, ImageDraw
im = Image.open("a.jpg")
im = im.resize((150,150))
result = im.convert('P', palette=Image.ADAPTIVE, colors=8)
result.putalpha(0)
colors = result.getcolors(150*150)
newimg = Image.new('RGB', (64*8, 64))
draw = ImageDraw.Draw(newimg)
posx = 0
for row, col in colors:
draw.rectangle([posx, 0, posx+64, 64], fill=col)
posx = posx + 64
del draw
newimg.save("swatch", "JPEG")
| [
"PIL.Image.new",
"PIL.ImageDraw.Draw",
"PIL.Image.open"
] | [((39, 58), 'PIL.Image.open', 'Image.open', (['"""a.jpg"""'], {}), "('a.jpg')\n", (49, 58), False, 'from PIL import Image, ImageDraw\n'), ((208, 238), 'PIL.Image.new', 'Image.new', (['"""RGB"""', '(64 * 8, 64)'], {}), "('RGB', (64 * 8, 64))\n", (217, 238), False, 'from PIL import Image, ImageDraw\n'), ((244, 266), 'PIL.ImageDraw.Draw', 'ImageDraw.Draw', (['newimg'], {}), '(newimg)\n', (258, 266), False, 'from PIL import Image, ImageDraw\n')] |
# Copyright 2020 Google LLC
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""Handles account deletion logic.
Deletes all objects that are related to a user. This includes:
NDB objects:
- User by user_id
- PageRating by user_id.
- Connection by publisher_id
- Connection by subscriber_id
- PastRecommendation by user_id
- RecommendationSession by user_id
- Category by parent
- export.ExportRatingResult by key
Non NDB objects:
- Memcache: "ri:<user_id>"
- Clear text search indexes:
- rating_history:<user_id>
- saved_for_later:<user_id>
"""
from google.appengine.api import memcache
from google.appengine.ext import deferred
from google.appengine.ext import ndb
from recommender import models
def _DeleteUser(user_id):
models.UserKey(user_id).delete()
def _DeleteAll(keys, continuation_fn, user_id):
if keys:
ndb.delete_multi(keys)
deferred.defer(continuation_fn, user_id)
def _DeletePageRating(user_id):
_DeleteAll(
models.PageRating.query(ancestor=models.UserKey(user_id)).fetch(
keys_only=True, limit=500), _DeletePageRating, user_id)
def _DeleteConnectionPublisher(user_id):
_DeleteAll(
models.Connection.query(models.Connection.publisher_id == user_id).fetch(
keys_only=True, limit=500), _DeleteConnectionPublisher, user_id)
def _DeleteConnectionSubscriber(user_id):
_DeleteAll(
models.Connection.query(models.Connection.subscriber_id == user_id).fetch(
keys_only=True, limit=500), _DeleteConnectionSubscriber, user_id)
def _DeletePastRecommendation(user_id):
_DeleteAll(
models.PastRecommendation.query(
models.PastRecommendation.user_id == user_id).fetch(
keys_only=True, limit=500), _DeletePastRecommendation, user_id)
def _DeleteRecommendationSession(user_id):
_DeleteAll(
models.RecommendationSession.query(
models.RecommendationSession.user_id == user_id).fetch(
keys_only=True, limit=500), _DeleteRecommendationSession, user_id)
def _DeleteCategory(user_id):
_DeleteAll(
models.Category.query(ancestor=models.UserKey(user_id)).fetch(
keys_only=True, limit=500), _DeleteCategory, user_id)
def _DeleteCachedRatings(user_id):
memcache.delete(models.GetUserRatedItemsCacheKey(user_id))
HANDLERS = [
_DeleteUser,
_DeletePageRating,
_DeleteConnectionPublisher,
_DeleteConnectionSubscriber,
_DeletePastRecommendation,
_DeleteRecommendationSession,
_DeleteCategory,
_DeleteCachedRatings,
]
def DeleteAccount(user_id):
for handler in HANDLERS:
deferred.defer(handler, user_id)
| [
"recommender.models.GetUserRatedItemsCacheKey",
"google.appengine.ext.deferred.defer",
"recommender.models.UserKey",
"recommender.models.PastRecommendation.query",
"recommender.models.RecommendationSession.query",
"google.appengine.ext.ndb.delete_multi",
"recommender.models.Connection.query"
] | [((1330, 1352), 'google.appengine.ext.ndb.delete_multi', 'ndb.delete_multi', (['keys'], {}), '(keys)\n', (1346, 1352), False, 'from google.appengine.ext import ndb\n'), ((1357, 1397), 'google.appengine.ext.deferred.defer', 'deferred.defer', (['continuation_fn', 'user_id'], {}), '(continuation_fn, user_id)\n', (1371, 1397), False, 'from google.appengine.ext import deferred\n'), ((2728, 2769), 'recommender.models.GetUserRatedItemsCacheKey', 'models.GetUserRatedItemsCacheKey', (['user_id'], {}), '(user_id)\n', (2760, 2769), False, 'from recommender import models\n'), ((3066, 3098), 'google.appengine.ext.deferred.defer', 'deferred.defer', (['handler', 'user_id'], {}), '(handler, user_id)\n', (3080, 3098), False, 'from google.appengine.ext import deferred\n'), ((1232, 1255), 'recommender.models.UserKey', 'models.UserKey', (['user_id'], {}), '(user_id)\n', (1246, 1255), False, 'from recommender import models\n'), ((1646, 1712), 'recommender.models.Connection.query', 'models.Connection.query', (['(models.Connection.publisher_id == user_id)'], {}), '(models.Connection.publisher_id == user_id)\n', (1669, 1712), False, 'from recommender import models\n'), ((1859, 1926), 'recommender.models.Connection.query', 'models.Connection.query', (['(models.Connection.subscriber_id == user_id)'], {}), '(models.Connection.subscriber_id == user_id)\n', (1882, 1926), False, 'from recommender import models\n'), ((2072, 2149), 'recommender.models.PastRecommendation.query', 'models.PastRecommendation.query', (['(models.PastRecommendation.user_id == user_id)'], {}), '(models.PastRecommendation.user_id == user_id)\n', (2103, 2149), False, 'from recommender import models\n'), ((2311, 2398), 'recommender.models.RecommendationSession.query', 'models.RecommendationSession.query', (['(models.RecommendationSession.user_id == user_id)'], {}), '(models.RecommendationSession.user_id ==\n user_id)\n', (2345, 2398), False, 'from recommender import models\n'), ((1485, 1508), 'recommender.models.UserKey', 'models.UserKey', (['user_id'], {}), '(user_id)\n', (1499, 1508), False, 'from recommender import models\n'), ((2577, 2600), 'recommender.models.UserKey', 'models.UserKey', (['user_id'], {}), '(user_id)\n', (2591, 2600), False, 'from recommender import models\n')] |
from os.path import abspath, dirname, join
WORLDGEN_ROOT_PATH = abspath(join(dirname(__file__), '..', '..'))
def worldgen_path(*args):
"""
Returns an absolute path from a path relative to the mujoco_worldgen repository
root directory.
"""
return join(WORLDGEN_ROOT_PATH, *args)
| [
"os.path.dirname",
"os.path.join"
] | [((269, 300), 'os.path.join', 'join', (['WORLDGEN_ROOT_PATH', '*args'], {}), '(WORLDGEN_ROOT_PATH, *args)\n', (273, 300), False, 'from os.path import abspath, dirname, join\n'), ((78, 95), 'os.path.dirname', 'dirname', (['__file__'], {}), '(__file__)\n', (85, 95), False, 'from os.path import abspath, dirname, join\n')] |
# encoding: utf-8
import os
import re
import sys
import gzip
import time
import json
import socket
import random
import weakref
import datetime
import functools
import threading
import collections
import urllib.error
import urllib.parse
import urllib.request
import collections.abc
import json_dict
from . import utils
class ProxyURLRefreshError(Exception):
pass
class AliveProxiesNotFound(Exception):
pass
class NoFreeProxies(Exception):
pass
def _get_missing(target, source):
"""Возвращает присутствующие в `target`, но отсутствующие в `source` элементы
"""
old_target = set(target)
new_target = old_target.intersection(source)
return old_target.difference(new_target)
def _build_opener(proxy=None):
if proxy is not None:
parsed = urllib.parse.urlparse(proxy)
handler = urllib.request.ProxyHandler({parsed.scheme: proxy})
return urllib.request.build_opener(handler)
else:
return urllib.request.build_opener()
class Proxies:
default_opener = _build_opener()
def __init__(
self,
proxies=None,
proxies_url=None,
proxies_url_gateway=None,
proxies_file=None,
options=None,
):
"""
@param proxies: список адресов прокси-серверов
@param proxies_url: ссылка на список прокси-серверов
@param proxies_file: путь до файла со списком прокси-серверов
@param options: доп. параметры
"""
if options is None:
options = {}
shuffle = options.get('shuffle', False)
if proxies is not None:
proxies = list(proxies)
if shuffle:
random.shuffle(proxies)
auto_refresh_period = options.get('auto_refresh_period')
if auto_refresh_period:
auto_refresh_period = datetime.timedelta(**auto_refresh_period)
blacklist = utils.get_json_dict(json_dict.JsonLastUpdatedOrderedDict, filename=options.get('blacklist'))
cooling_down = utils.get_json_dict(json_dict.JsonOrderedDict, filename=options.get('cooldown'))
stats = utils.get_json_dict(json_dict.JsonDict, filename=options.get('stats'))
if proxies_url_gateway:
url_opener = _build_opener(proxies_url_gateway)
else:
url_opener = None
self._url_opener = url_opener
self._proxies = proxies
self.proxies_url = proxies_url
self.proxies_file = proxies_file
self._shuffle = shuffle
self.slice = options.get('slice')
self.force_type = options.get('type')
self.auto_refresh_period = auto_refresh_period
self._blacklist = blacklist
self._cooling_down = cooling_down
self._stats = stats
self._cleanup_lock = threading.RLock()
self._last_auto_refresh = None
self._auto_refresh_lock = threading.Lock()
self._load_lock = threading.Lock()
self._modified_at = time.perf_counter()
self.__pool = None
self._smart_holdout_start = options.get('smart_holdout_start')
self._options = options
if self._proxies is not None:
proxies = set(self._proxies)
self._cleanup_internals(proxies)
@property
def proxies(self):
if self._proxies is None:
with self._load_lock:
# Вышли из состояния гонки, теперь можно удостовериться в реальной необходимости
if self._proxies is None:
self._proxies = self._load()
self._cleanup_internals(self._proxies)
self._modified_at = time.perf_counter()
return self._proxies
def _load(self):
if self.proxies_url:
proxies = self.read_url(self.proxies_url, opener=self._url_opener)
elif self.proxies_file:
proxies = self.read_file(self.proxies_file)
else:
raise NotImplementedError(
"Can't load proxies: "
"please specify one of the sources ('proxies_url' or 'proxies_file')"
)
if self.slice:
proxies = proxies[slice(*self.slice)]
if self.force_type:
new_type = self.force_type + '://' # `socks` format
proxies = [
re.sub(r'^(?:(.*?)://)?', new_type, proxy)
for proxy in proxies
]
if self._shuffle:
random.shuffle(proxies)
return proxies
def _cleanup_internals(self, proxies):
with self._cleanup_lock:
self._cleanup_blacklist(proxies)
self._cleanup_cooling_down(proxies)
self._cleanup_stats(proxies)
def _cleanup_cooling_down(self, proxies):
for proxy in _get_missing(self._cooling_down, proxies):
self._cooling_down.pop(proxy)
def _cleanup_blacklist(self, proxies):
for proxy in _get_missing(self._blacklist, proxies):
self._blacklist.pop(proxy)
def _cleanup_stats(self, proxies):
for proxy in _get_missing(self._stats, proxies):
self._stats.pop(proxy)
def _get_options(self, *options, missing_ok=True):
if missing_ok:
return {k: self._options.get(k) for k in options}
else:
return {k: self._options[k] for k in options}
@classmethod
def read_string(cls, string, sep=','):
return list(x for x in map(str.strip, string.split(sep)) if x)
@classmethod
def read_url(cls, url, sep='\n', retry=10, sleep_range=(2, 10), timeout=2, opener=None):
if opener is None:
opener = cls.default_opener
while True:
try:
resp = opener.open(url, timeout=timeout)
break
except (urllib.error.HTTPError, socket.timeout):
if not retry:
raise
retry -= 1
time.sleep(random.randint(*sleep_range))
content = resp.read()
if resp.headers.get('Content-Encoding', 'identity') == 'gzip':
content = gzip.decompress(content)
charset = resp.headers.get_content_charset('utf-8')
content = content.decode(charset)
return cls.read_string(content, sep=sep)
@classmethod
def read_file(cls, file_name, sep='\n'):
with open(file_name) as f:
return cls.read_string(f.read(), sep=sep)
def refresh(self):
if not self.proxies_url and not self.proxies_file:
return
try:
self._proxies = self._load()
self._cleanup_internals(self._proxies)
except urllib.error.HTTPError:
import problems
problems.handle(ProxyURLRefreshError, extra={'url': self.proxies_url})
else:
self._modified_at = time.perf_counter()
def _auto_refresh(self):
if self.proxies_file:
with self._auto_refresh_lock:
modification_time = datetime.datetime.fromtimestamp(os.stat(self.proxies_file).st_mtime)
if modification_time == self._last_auto_refresh:
return
self.refresh()
self._last_auto_refresh = modification_time
elif self.proxies_url:
if self.auto_refresh_period is None:
return
with self._auto_refresh_lock:
now = datetime.datetime.now()
if self._last_auto_refresh is not None:
if now - self._last_auto_refresh < self.auto_refresh_period:
return
self.refresh()
self._last_auto_refresh = now
def get_random_address(self):
self._auto_refresh()
return random.choice(self.proxies)
def get_pool(self):
if self.__pool is None:
with self._cleanup_lock: # оптимизация: используем уже существующий лок
# Вышли из состояния гонки, теперь можно удостовериться в реальной необходимости
if self.__pool is None:
options = self._get_options('default_holdout', 'default_bad_holdout', 'force_defaults')
if self._smart_holdout_start is not None:
options['smart_holdout'] = True
options['smart_holdout_start'] = self._smart_holdout_start
options.update(self._get_options('smart_holdout_min', 'smart_holdout_max'))
self.__pool = _Pool(
self, self._cooling_down, self._blacklist, self._stats, self._cleanup_lock,
**options
)
return self.__pool
@classmethod
def from_cfg_string(cls, cfg_string):
"""Возвращает список прокси с тем исключением что список опций берется автоматически.
Формат: json
Доступные опции:
type ('socks5', 'http'; для полного списка типов см. модуль socks):
все прокси будут автоматически промаркированы этип типом
slice (tuple c аргументами для builtins.slice):
будет взят только указанный фрагмент списка прокси-серверов
auto_refresh_period (dict): {'days': ..., 'hours': ..., 'minutes': ...}
как часто необходимо обновлять список прокси-серверов (только для `url` и `file`)
url_gateway:
адрес proxy, через которые будет загружаться список прокси по url
(url, file, list) - может быть именем файла, ссылкой или списком в формате json
Параметры slice и force_type являются необязательными
Примеры:
option = {"list": ["127.0.0.1:3128"]}
option = {"list": ["127.0.0.1:3128", "127.0.0.1:9999"]}
option = {"file": "./my_new_proxies.txt", "type": "socks5"}
option = {"url": "http://example.com/get/proxy_list/", "slice": [35, null], "type": "http"}
option = {"url": "http://example.com/get/proxy_list/", "auto_refresh_period": {"days": 1}}
option = {"url": "http://example.com/get/proxy_list/", "url_gateway": "http://proxy.example.com:9999"}
"""
cfg = json.loads(cfg_string)
proxies = cfg.pop('list', None)
proxies_url = cfg.pop('url', None)
proxies_url_gateway = cfg.pop('url_gateway', None)
proxies_file = cfg.pop('file', None)
return cls(
proxies=proxies,
proxies_url=proxies_url,
proxies_url_gateway=proxies_url_gateway,
proxies_file=proxies_file,
options=cfg
)
class _Pool:
def __init__(
self, proxies: "`Proxies` instance", cooling_down, blacklist, stats, _cleanup_lock=None,
smart_holdout=False, smart_holdout_start=None, smart_holdout_min=None, smart_holdout_max=None,
default_holdout=None, default_bad_holdout=None, force_defaults=False,
):
if smart_holdout:
if smart_holdout_start in (None, 0):
raise RuntimeError("Вы должны указать начальное время охлаждения")
if smart_holdout_max is None:
smart_holdout_max = float('inf')
self._used = set()
self._cond = threading.Condition(lock=_cleanup_lock)
self._free = collections.deque(
p for p in proxies.proxies
if (
p not in blacklist and
p not in cooling_down
)
)
self._proxies = proxies
self._cooling_down = cooling_down
self._blacklist = blacklist
self._stats = stats
self._smart_holdout = smart_holdout
self._smart_holdout_start = smart_holdout_start
self._smart_holdout_min = smart_holdout_min or 0
self._smart_holdout_max = smart_holdout_max
self._default_holdout = default_holdout
self._default_bad_holdout = default_bad_holdout
self._force_defaults = force_defaults
self._proxies_modified_at = proxies._modified_at
@property
def _size(self):
return len(self._free) + len(self._used) + len(self._cooling_down) + len(self._blacklist)
def _cool_released(self):
now = time.time()
cooled = []
for proxy, holdout in self._cooling_down.items():
if now >= holdout:
cooled.append(proxy)
for proxy in cooled:
self._cooling_down.pop(proxy, None)
if proxy not in self._blacklist:
self._free.append(proxy)
def _is_proxies_changed(self):
self._proxies._auto_refresh()
return self._proxies._modified_at != self._proxies_modified_at
def _remove_outdated(self):
# список прокси изменился, оставляем только актуальные
full_list = set(self._proxies.proxies)
for proxy in _get_missing(self._blacklist, full_list):
self._blacklist.pop(proxy, None)
for proxy in _get_missing(self._cooling_down, full_list):
self._cooling_down.pop(proxy, None)
for proxy in _get_missing(self._used, full_list):
self._used.remove(proxy)
for proxy in _get_missing(self._stats, full_list):
self._stats.pop(proxy, None)
free = set(
p for p in full_list
if (
p not in self._used and
p not in self._blacklist and
p not in self._cooling_down
)
)
old_free = set(self._free)
new_free = old_free.intersection(free)
if old_free.difference(new_free):
self._free.clear()
self._free.extend(new_free)
self._proxies_modified_at = self._proxies._modified_at
def _update_stats(self, proxy, bad=False, holdout=None):
proxy_stat = self._stats.get(proxy) or {}
ok, fail = proxy_stat.get('uptime', (0, 0))
if not bad:
ok += 1
else:
fail += 1
proxy_stat['uptime'] = ok, fail
proxy_stat['last_holdout'] = holdout
if (
not bad or
(
holdout is not None and
holdout >= (proxy_stat.get('last_good_holdout') or 0)
)
):
proxy_stat['last_good_holdout'] = holdout
# универсальный способ сказать что статистика обновилась
# тк без вызова метода .save будет работать и с обычным словарем (не только с JsonDict)
self._stats[proxy] = proxy_stat
def _get_next_holdout(self, proxy, bad=False):
"""Рассчитывает время охлаждения.
@param proxy: прокси, для которого необходимо вычислить
@param bad: True - вычисляем охлаждение для неудачи, иначе False
@return: рекомендуемое время охлаждения в секундах или None, если недостаточно данных
"""
# Алгоритм основан на бинарном поиске,
# в отличии от которого нам не известна верхняя граница
proxy_stat = self._stats.get(proxy)
if proxy_stat is None:
return None
last_holdout = proxy_stat['last_holdout']
last_good_holdout = proxy_stat.get('last_good_holdout', 0)
lo = last_holdout # предыдущее время охлаждения (нижняя граница)
if bad:
# Мы получили "бан" ...
if lo < last_good_holdout:
# ... возвращаемся к предыдущему хорошему значению ...
holdout = last_good_holdout
else:
# ... или сдвигаем границу дальше
holdout = lo * 2
else:
# возвращаемся к предыдущей границе (lo / 2)
# но с небольшим отступом - на середину отрезка [(lo / 2), lo]
holdout = lo * 0.75
return holdout
def acquire(self, timeout=None):
start = time.perf_counter()
with self._cond:
while True:
if self._is_proxies_changed():
self._remove_outdated()
self._cool_released()
if self._free:
proxy = self._free.popleft()
self._used.add(proxy)
return proxy
if self._blacklist:
# Возвращаем самый стабильный из блеклиста. Возможно бан снят.
def _uptime(p):
uptime = float('inf')
p_stat = self._stats.get(p)
if p_stat is not None:
ok, failed = p_stat.get('uptime', (0, 0))
if failed != 0:
uptime = ok // failed
else:
uptime = ok
return uptime
proxy = next((
p for p in sorted(self._blacklist, key=_uptime, reverse=True)
if p not in self._cooling_down
), None)
if proxy is not None:
self._blacklist.pop(proxy)
self._used.add(proxy)
return proxy
else:
# Все прокси из блеклиста находятся на охлаждении
pass
if self._cooling_down:
self._cond.wait(1)
else:
self._cond.wait(timeout)
if timeout is not None:
if time.perf_counter() - start > timeout:
raise NoFreeProxies
def release(self, proxy, bad=False, holdout=None, bad_reason=None):
"""Возвращает прокси в пул
@param proxy: прокси
@param holdout (сек): None - вернуть сразу, иначе прокси не будет использован до истечения указанного интервала
"""
with self._cond:
is_outdated = proxy not in self._used
if is_outdated:
# Скорее всего прокси уже не актуален
# И был удален из списка
return
self._used.remove(proxy)
if holdout is None or self._force_defaults:
holdout = self._default_holdout if not bad else self._default_bad_holdout
if self._smart_holdout:
_holdout = (
self._get_next_holdout(proxy, bad=bad) or
holdout or
self._smart_holdout_start
)
# Не позволяем границе опуститься слишком низко
if _holdout < self._smart_holdout_min:
holdout = self._smart_holdout_min
elif _holdout > self._smart_holdout_max:
holdout = self._smart_holdout_max
else:
holdout = max(self._smart_holdout_min, _holdout)
if holdout is not None:
self._cooling_down[proxy] = time.time() + holdout
if bad:
self._blacklist[proxy] = bad_reason
elif holdout is None:
# прокси не требует остывания
self._free.append(proxy)
self._cond.notify()
self._update_stats(proxy, bad=bad, holdout=holdout)
class IChain:
def switch(self, bad=False, holdout=None, bad_reason=None, lazy=False):
raise NotImplementedError
def get_adapter(self):
raise NotImplementedError
def get_handler(self):
raise NotImplementedError
def get_path(self):
raise NotImplementedError
def wrap_session(self, session):
raise NotImplementedError
def wrap_module(self, module, all_threads=False):
"""
Attempts to replace a module's socket library with a SOCKS socket.
This will only work on modules that import socket directly into the
namespace; most of the Python Standard Library falls into this category.
"""
import socks
import socks.monkey_socket
routes = socks.RoutingTable.from_addresses(self.get_path())
if not all_threads:
socks.monkey_socket.socks_wrap_module_thread(routes, module)
else:
socks.monkey_socket.socks_wrap_module_global(routes, module)
class Chain(IChain):
"""
Не является потокобезопасным.
"""
def __init__(self, proxies, proxy_gw=None, use_pool=False, pool_acquire_timeout=None):
"""
@param proxies: список адресов прокси-серверов
@param proxy_gw: прокси-сервер, который должен стоять во главе цепочки
(все запросы к другим прокси-серверам будут проходить через него)
@param use_pool: использовать список прокси в качестве пула
@param pool_acquire_timeout (сек.): если за указанный период не удастся получить свободный прокси
будет брошено исключение `NoFreeProxies`, None - ждать до появления свободного адреса
"""
if not isinstance(proxies, Proxies) and isinstance(proxies, collections.Sequence):
proxies = Proxies(proxies)
if use_pool:
pool = proxies.get_pool()
else:
pool = None
self.proxies = proxies
self.proxy_gw = proxy_gw
self._proxies_pool = pool
self._current_pool_proxy = None
self._pool_acquire_timeout = pool_acquire_timeout
self.__path = []
# fix http://bugs.python.org/issue23841
if sys.version_info >= (3, 4, 0):
self.finalizer = weakref.finalize(self, self.finalize)
else:
self.finalizer = None
def __del__(self):
if self.finalizer is None:
self.finalize()
def finalize(self):
if self._proxies_pool is not None:
self._release_pool_proxy()
def _build_path(self, proxy):
path = []
if self.proxy_gw:
path.append(self.proxy_gw)
path.append(proxy)
return path
def _release_pool_proxy(self, bad=False, holdout=None, bad_reason=None):
if self._current_pool_proxy:
proxy = self._current_pool_proxy
self._current_pool_proxy = None
self._proxies_pool.release(proxy, bad=bad, holdout=holdout, bad_reason=bad_reason)
def _acquire_pool_proxy(self):
proxy = self._proxies_pool.acquire(timeout=self._pool_acquire_timeout)
self._current_pool_proxy = proxy
return proxy
def _get_proxy(self):
if self._proxies_pool is not None:
return self._acquire_pool_proxy()
else:
return self.proxies.get_random_address()
@property
def _path(self):
if not self.__path:
self.__path = self._build_path(self._get_proxy())
return self.__path
def get_path(self):
return self._path
def switch(self, bad=False, holdout=None, bad_reason=None, lazy=False):
self.__path.clear()
if self._proxies_pool is not None:
self._release_pool_proxy(bad, holdout, bad_reason)
if not lazy:
self.__path = self._build_path(self._get_proxy())
def get_adapter(self):
import socks.adapters
return socks.adapters.ChainedProxyHTTPAdapter(chain=self._path)
def get_handler(self):
import socks.handlers
return socks.handlers.ChainProxyHandler(chain=self._path)
def wrap_session(self, session):
adapter = self.get_adapter()
session.mount('http://', adapter)
session.mount('https://', adapter)
return session
@classmethod
def from_config(cls, cfg):
proxy_cfg_string = cfg.get('Прокси')
if proxy_cfg_string is None:
return None
proxy_gw = cfg.get('Шлюз')
proxies = Proxies.from_cfg_string(proxy_cfg_string)
return cls(proxies, proxy_gw=proxy_gw)
class MultiChain(IChain):
def __init__(self, *proxies_all, use_pool=True, pool_acquire_timeout=None):
if use_pool:
pool_kw = {'use_pool': True, 'pool_acquire_timeout': 1}
else:
pool_kw = {}
self._pool_acquire_timeout = pool_acquire_timeout
self._chains = collections.deque(
Chain(p, gw, **pool_kw)
for p, gw in self._unwrap_proxies_all(proxies_all)
)
@staticmethod
def _unwrap_proxies_all(proxies_all):
for p in proxies_all:
if isinstance(p, tuple):
# (Proxies, Gateway)
p, gw = p
else:
# Proxies
p, gw = p, None
yield p, gw
def _self_auto_rotate(func):
@functools.wraps(func)
def wrapped(self, *args, **kw):
start = time.perf_counter()
while True:
try:
return func(self, *args, **kw)
except NoFreeProxies:
self._rotate() # FIXME: cycle rotate is normal?
if (
self._pool_acquire_timeout is not None and
time.perf_counter() - start > self._pool_acquire_timeout
):
raise
return wrapped
@property
def _current(self):
return self._chains[-1]
def get_path(self):
return self._current.get_path()
def _rotate(self):
self._chains.rotate(1)
def switch(self, bad=False, holdout=None, bad_reason=None, lazy=False):
self._current.switch(bad=bad, holdout=holdout, bad_reason=bad_reason, lazy=True)
self._rotate()
if not lazy:
self._enforce_current_path_build()
@_self_auto_rotate
def _enforce_current_path_build(self):
_ = self._current._path # FIXME: ugly enforce path building after switching
@_self_auto_rotate
def get_adapter(self):
return self._current.get_adapter()
@_self_auto_rotate
def get_handler(self):
return self._current.get_handler()
@_self_auto_rotate
def wrap_session(self, session):
return self._current.wrap_session(session)
@_self_auto_rotate
def wrap_module(self, module):
return self._current.wrap_module(module)
| [
"gzip.decompress",
"datetime.timedelta",
"weakref.finalize",
"collections.deque",
"threading.Lock",
"threading.RLock",
"time.perf_counter",
"functools.wraps",
"threading.Condition",
"random.randint",
"json.loads",
"random.choice",
"random.shuffle",
"socks.handlers.ChainProxyHandler",
"so... | [((2788, 2805), 'threading.RLock', 'threading.RLock', ([], {}), '()\n', (2803, 2805), False, 'import threading\n'), ((2880, 2896), 'threading.Lock', 'threading.Lock', ([], {}), '()\n', (2894, 2896), False, 'import threading\n'), ((2924, 2940), 'threading.Lock', 'threading.Lock', ([], {}), '()\n', (2938, 2940), False, 'import threading\n'), ((2969, 2988), 'time.perf_counter', 'time.perf_counter', ([], {}), '()\n', (2986, 2988), False, 'import time\n'), ((7759, 7786), 'random.choice', 'random.choice', (['self.proxies'], {}), '(self.proxies)\n', (7772, 7786), False, 'import random\n'), ((10186, 10208), 'json.loads', 'json.loads', (['cfg_string'], {}), '(cfg_string)\n', (10196, 10208), False, 'import json\n'), ((11239, 11278), 'threading.Condition', 'threading.Condition', ([], {'lock': '_cleanup_lock'}), '(lock=_cleanup_lock)\n', (11258, 11278), False, 'import threading\n'), ((11301, 11398), 'collections.deque', 'collections.deque', (['(p for p in proxies.proxies if p not in blacklist and p not in cooling_down)'], {}), '(p for p in proxies.proxies if p not in blacklist and p not in\n cooling_down)\n', (11318, 11398), False, 'import collections\n'), ((12214, 12225), 'time.time', 'time.time', ([], {}), '()\n', (12223, 12225), False, 'import time\n'), ((15809, 15828), 'time.perf_counter', 'time.perf_counter', ([], {}), '()\n', (15826, 15828), False, 'import time\n'), ((23158, 23214), 'socks.adapters.ChainedProxyHTTPAdapter', 'socks.adapters.ChainedProxyHTTPAdapter', ([], {'chain': 'self._path'}), '(chain=self._path)\n', (23196, 23214), False, 'import socks\n'), ((23288, 23338), 'socks.handlers.ChainProxyHandler', 'socks.handlers.ChainProxyHandler', ([], {'chain': 'self._path'}), '(chain=self._path)\n', (23320, 23338), False, 'import socks\n'), ((24602, 24623), 'functools.wraps', 'functools.wraps', (['func'], {}), '(func)\n', (24617, 24623), False, 'import functools\n'), ((1840, 1881), 'datetime.timedelta', 'datetime.timedelta', ([], {}), '(**auto_refresh_period)\n', (1858, 1881), False, 'import datetime\n'), ((4440, 4463), 'random.shuffle', 'random.shuffle', (['proxies'], {}), '(proxies)\n', (4454, 4463), False, 'import random\n'), ((6095, 6119), 'gzip.decompress', 'gzip.decompress', (['content'], {}), '(content)\n', (6110, 6119), False, 'import gzip\n'), ((6829, 6848), 'time.perf_counter', 'time.perf_counter', ([], {}), '()\n', (6846, 6848), False, 'import time\n'), ((20090, 20150), 'socks.monkey_socket.socks_wrap_module_thread', 'socks.monkey_socket.socks_wrap_module_thread', (['routes', 'module'], {}), '(routes, module)\n', (20134, 20150), False, 'import socks\n'), ((20177, 20237), 'socks.monkey_socket.socks_wrap_module_global', 'socks.monkey_socket.socks_wrap_module_global', (['routes', 'module'], {}), '(routes, module)\n', (20221, 20237), False, 'import socks\n'), ((21477, 21514), 'weakref.finalize', 'weakref.finalize', (['self', 'self.finalize'], {}), '(self, self.finalize)\n', (21493, 21514), False, 'import weakref\n'), ((24684, 24703), 'time.perf_counter', 'time.perf_counter', ([], {}), '()\n', (24701, 24703), False, 'import time\n'), ((1684, 1707), 'random.shuffle', 'random.shuffle', (['proxies'], {}), '(proxies)\n', (1698, 1707), False, 'import random\n'), ((4307, 4348), 're.sub', 're.sub', (['"""^(?:(.*?)://)?"""', 'new_type', 'proxy'], {}), "('^(?:(.*?)://)?', new_type, proxy)\n", (4313, 4348), False, 'import re\n'), ((6712, 6782), 'problems.handle', 'problems.handle', (['ProxyURLRefreshError'], {'extra': "{'url': self.proxies_url}"}), "(ProxyURLRefreshError, extra={'url': self.proxies_url})\n", (6727, 6782), False, 'import problems\n'), ((3639, 3658), 'time.perf_counter', 'time.perf_counter', ([], {}), '()\n', (3656, 3658), False, 'import time\n'), ((7409, 7432), 'datetime.datetime.now', 'datetime.datetime.now', ([], {}), '()\n', (7430, 7432), False, 'import datetime\n'), ((18915, 18926), 'time.time', 'time.time', ([], {}), '()\n', (18924, 18926), False, 'import time\n'), ((5940, 5968), 'random.randint', 'random.randint', (['*sleep_range'], {}), '(*sleep_range)\n', (5954, 5968), False, 'import random\n'), ((7019, 7045), 'os.stat', 'os.stat', (['self.proxies_file'], {}), '(self.proxies_file)\n', (7026, 7045), False, 'import os\n'), ((17475, 17494), 'time.perf_counter', 'time.perf_counter', ([], {}), '()\n', (17492, 17494), False, 'import time\n'), ((25023, 25042), 'time.perf_counter', 'time.perf_counter', ([], {}), '()\n', (25040, 25042), False, 'import time\n')] |
from project.aquarium.freshwater_aquarium import FreshwaterAquarium # noqa
from project.aquarium.saltwater_aquarium import SaltwaterAquarium # noqa
from project.decoration.decoration_repository import DecorationRepository
from project.decoration.ornament import Ornament # noqa
from project.decoration.plant import Plant # noqa
from project.fish.freshwater_fish import FreshwaterFish # noqa
from project.fish.saltwater_fish import SaltwaterFish # noqa
class Controller:
def __init__(self):
self.decorations_repository = DecorationRepository()
self.aquariums = []
def add_aquarium(self, aquarium_type, aquarium_name):
try:
aquarium = globals()[aquarium_type]
aquarium = aquarium(aquarium_name)
except KeyError:
return "Invalid aquarium type."
self.aquariums.append(aquarium)
return f"Successfully added {aquarium_type}."
def add_decoration(self, decoration_type):
try:
decoration = globals()[decoration_type]
decoration = decoration()
except KeyError:
return "Invalid decoration type."
self.decorations_repository.add(decoration)
return f"Successfully added {decoration_type}."
def insert_decoration(self, aquarium_name, decoration_type):
decoration = self.decorations_repository.find_by_type(decoration_type)
if decoration == "None":
return f"There isn't a decoration of type {decoration_type}."
aquarium = self.__find_aquarium(aquarium_name)
if aquarium:
aquarium.add_decoration(decoration)
self.decorations_repository.remove(decoration)
return f"Successfully added {decoration_type} to {aquarium_name}."
def add_fish(self, aquarium_name, fish_type, fish_name, fish_species, price):
try:
fish = globals()[fish_type]
fish = fish(fish_name, fish_species, price)
except KeyError:
return f"There isn't a fish of type {fish_type}."
aquarium = self.__find_aquarium(aquarium_name)
result = aquarium.add_fish(fish)
return result
def feed_fish(self, aquarium_name):
aquarium = self.__find_aquarium(aquarium_name)
aquarium.feed()
return f"Fish fed: {len(aquarium.fish)}"
def calculate_value(self, aquarium_name):
aquarium = self.__find_aquarium(aquarium_name)
value = aquarium.total_fish_price + aquarium.total_decoration_price
return f"The value of Aquarium {aquarium_name} is {value:.2f}."
def report(self):
return "\n".join(str(a) for a in self.aquariums)
def __find_aquarium(self, aquarium_name):
for aquarium in self.aquariums:
if aquarium.name == aquarium_name:
return aquarium
| [
"project.decoration.decoration_repository.DecorationRepository"
] | [((540, 562), 'project.decoration.decoration_repository.DecorationRepository', 'DecorationRepository', ([], {}), '()\n', (560, 562), False, 'from project.decoration.decoration_repository import DecorationRepository\n')] |
import json
from Note import Note
n = Note("Friction", "introduction to friction", "UC Berkeley", 0, True, False, False, "https://google.com", 'lec.pdf')
print(n.toJSON())
| [
"Note.Note"
] | [((39, 158), 'Note.Note', 'Note', (['"""Friction"""', '"""introduction to friction"""', '"""UC Berkeley"""', '(0)', '(True)', '(False)', '(False)', '"""https://google.com"""', '"""lec.pdf"""'], {}), "('Friction', 'introduction to friction', 'UC Berkeley', 0, True, False,\n False, 'https://google.com', 'lec.pdf')\n", (43, 158), False, 'from Note import Note\n')] |
# Reference/tutorial to take a look at
# https://help.autodesk.com/view/MOBPRO/2019/ENU/?guid=__py_ref__tasks_2_assign_rigid_body_8py_example_html
# https://help.autodesk.com/view/MOBPRO/2019/ENU/?guid=__files_GUID_A1189AA0_3816_4350_B8F3_5383DEC25A33_htm
# https://mocappys.com/complete-guide-to-poses-in-motionbuilder/#07_Using_the_Pose_Controls
# https://github.com/eksod/Retargeter
from pyfbsdk import *
import os
import re
import random
def addJoint(jointMap, newJoint, parent, localTrans, adjChildList=[], adjTrans=False, jointOrder=list()):
jointMap[newJoint] = (parent, localTrans)
for k in jointMap:
v = jointMap[k]
if v[0] == parent and k!=newJoint:
if k in adjChildList:
if adjTrans:
adjTrans = [v[1][0]- localTrans[0], v[1][1]- localTrans[1], v[1][2]- localTrans[2]]
jointMap[k] = (newJoint, adjTrans)
else:
jointMap[k] = (newJoint, v[1])
if parent in jointOrder:
n = jointOrder.index(parent)
jointOrder.insert(n+1, newJoint)
else:
jointOrder.insert(0, newJoint)
# Create a skeleton in a T-pose facing along the positive Z axis
def createSkeleton(pNamespace):
# reference jointMap is copied from mixamo character tpose.
jointMap = {
#jointName, (parentName, translation )
#'Reference': (None, ( 0, 0, 0)),
'Hips': ('None', ( 0, 0, 0)),
'LeftUpLeg': ('Hips', (8.20778, -6.75171, -1.59956)),
'LeftLeg': ('LeftUpLeg', (0, -44.3705, 0.284643)),
'LeftFoot': ('LeftLeg', (0, -44.4279, -2.98219)),
'LeftToeBase': ('LeftFoot', (0, -8.72867, 10.7106)),
'RightUpLeg': ('Hips', (-8.2078, -6.75166, -1.59956)),
'RightLeg': ('RightUpLeg', (0, -44.3705, 0.286156)),
'RightFoot': ('RightLeg', (0, -44.4277, -2.98379)),
'RightToeBase': ('RightFoot', (0, -8.72867, 10.7106)),
'Spine': ('Hips', (0. , 29.36511 , -2.242699)),
'LeftArm': ('Spine', (15.16277, 10.421402, -4.86012)),
'LeftForeArm': ('LeftArm', (27.8415, -8.94286e-05, 3.74589e-05)),
'LeftHand': ('LeftForeArm', (28.3288, -1.74407e-05, 3.78045e-05)),
'RightArm': ('Spine', (-15.16277, 10.421402, -4.86012)),
'RightForeArm': ('RightArm', (-27.8415, -3.30792e-05, 1.16763e-05)),
'RightHand': ('RightForeArm', (-28.3288, 0, 5.5816e-05)),
'Neck': ('Spine', (0, 16.6717, -2.51617)),
'Head': ('Neck', (0, 9.61788, 1.68501)),
}
skeleton = {}
headOrder, legOrder, armOrder= ['Hips', 'Spine', 'Neck', 'Head'], ['LeftUpLeg', 'LeftLeg', 'LeftFoot', 'LeftToeBase'], ['LeftArm', 'LeftForeArm', 'LeftHand']
if random.random() < 0.3:
addJoint(jointMap, "LeftHipJoint", "Hips", (0,0,0), ["LeftUpLeg"], False, legOrder)
addJoint(jointMap, "RightHipJoint", "Hips", (0,0,0), ["RightUpLeg"], False)
if random.random() < 0.3:
addJoint(jointMap, "LowerBack", "Hips", (0,0,0), ["Spine"], False, headOrder)
spineCntRange = range(1, 7)
weights=[1.5+1/x for x in spineCntRange]
sum_weights= sum(weights)
weights =[x/sum_weights for x in weights]
spineRnd = random.random()
sum_w = 0
spineCnt = -1
for i, w in enumerate(weights):
sum_w += w
if spineRnd < sum_w:
spineCnt = i+1
break
if spineCnt == -1: spineCnt = 6
lastSpine = 'Spine'
if spineCnt > 1:
origTrans = jointMap['Neck'][1]
unitTrans = [origTrans[0]/spineCnt, origTrans[1]/spineCnt, origTrans[2]/spineCnt]
jointMap['Spine'] = (jointMap['Spine'][0], unitTrans)
for i in range(1, spineCnt):
addJoint(jointMap, 'Spine'+str(i), lastSpine, unitTrans, ['LeftArm', 'RightArm', 'Neck'], False, headOrder)
lastSpine= 'Spine'+str(i)
if random.random() < 0.5:
addJoint(jointMap, 'LeftShoulder', lastSpine, (4.57045, 10.946, -2.62799), ['LeftArm'], True, armOrder)
addJoint(jointMap, 'RightShoulder', lastSpine, (-4.56997, 10.9462, -2.62802), ['RightArm'], True)
elif random.random() < 0.3 :
addJoint(jointMap, 'LeftShoulder', lastSpine, (0, 0, 0), ['LeftArm'], True, armOrder)
addJoint(jointMap, 'RightShoulder', lastSpine, (0, 0, 0), ['RightArm'], True)
if random.random() < 0.5:
neckHeadCnt = 2
origTrans = jointMap['Head'][1]
unitTrans = [origTrans[0]/neckHeadCnt, origTrans[1]/neckHeadCnt, origTrans[2]/neckHeadCnt]
addJoint(jointMap, 'Neck1', 'Neck', unitTrans, ['Head'], True, headOrder)
#End Sites
#Head
if random.random() < 0.5:
headTrans = (0, 19, 1)
addJoint(jointMap, 'Head_End', 'Head', headTrans, [], False, headOrder)
# Leg
if random.random() < 1:
toeTrans = (0, 0, 6)
addJoint(jointMap, 'LeftToeBase_End', 'LeftToeBase', toeTrans, [], False, legOrder)
addJoint(jointMap, 'RightToeBase_End', 'RightToeBase', toeTrans, [], False)
# Arm
if random.random() < 1:
addJoint(jointMap, 'LeftHand_End', 'LeftHand', (6, 0, 0), [], False, armOrder)
addJoint(jointMap, 'RightHand_End', 'RightHand', (-6, 0, 0), [], False)
scale = 0.6+ 0.8*random.random()
link_scale = {}
# Populate the skeleton with joints.
other_root = findRoot()
for jointName, (parentName, translation) in jointMap.iteritems():
jointName_ = jointName if not skelExists(other_root, jointName) else jointName+"_"
if jointName == 'Reference' or jointName == 'Hips':
# If it is the reference node, create an FBModelRoot.
joint = FBModelRoot(jointName_)
else:
# Otherwise, create an FBModelSkeleton.
joint = FBModelSkeleton(jointName_)
joint.LongName = pNamespace + ':' + joint.Name # Apply the specified namespace to each joint.
joint.Color = FBColor(0.3, 0.8, 1) # Cyan
joint.Size = 150 # Arbitrary size: big enough to see in viewport
joint.Show = True # Make the joint visible in the scene.
# Add the joint to our skeleton.
skeleton[jointName] = joint
if "Right" not in jointName:
ls = [0.6+ 0.8*random.random(), 0.6+ 0.8*random.random(), 0.6+ 0.8*random.random()]
jointMap[jointName] = (parentName, (translation[0]*ls[0], translation[1]*ls[1], translation[2]*ls[2]))
#print(jointName, ls)
if "Left" in jointName:
cor_right = "Right"+jointName[4:]
r_parentName, r_translation = jointMap[cor_right]
jointMap[cor_right] = (r_parentName, (r_translation[0]*ls[0], r_translation[1]*ls[1], r_translation[2]*ls[2]))
# move Hips so that foot touches the ground (2cm above considering foot depth(?))
joint_lower= 'LeftToeBase'
lower_sum= 0
footOffset = 2
while joint_lower != 'Hips':
translation = jointMap[joint_lower][1]
lower_sum= lower_sum +translation[1]
joint_lower= jointMap[joint_lower][0]
hips_trans = jointMap['Hips'][1]
jointMap['Hips'] = ('None', (hips_trans[0], -lower_sum+footOffset, hips_trans[2]))
# Once all the joints have been created, apply the parent/child
# relationships to each of the skeleton's joints.
jointOrder= headOrder+legOrder+armOrder
def connectPlaceJoint(jointName):
(parentName, translation)= jointMap[jointName]
# Only assign a parent if it exists.
if parentName != None and parentName in jointMap.keys():
skeleton[jointName].Parent = skeleton[parentName]
# The translation should be set after the parent has been assigned.
skeleton[jointName].Translation = FBVector3d(translation)*scale
for jointName in reversed(jointOrder):
connectPlaceJoint(jointName)
if "Left" in jointName:
cor_right = "Right"+jointName[4:]
connectPlaceJoint(cor_right)
return skeleton
# Characterize the skeleton and create a control rig.
def characterizeSkeleton(pCharacterName, pSkeleton, ctrlRig=False):
# Create a new character.
character = FBCharacter(pCharacterName)
FBApplication().CurrentCharacter = character
# Add each joint in our skeleton to the character.
for jointName, joint in pSkeleton.iteritems():
slot = character.PropertyList.Find(jointName + 'Link')
# skip dummy nodes (e.g.: LeftHipJoint)
if slot is not None:
slot.append(joint)
# Flag that the character has been characterized.
character.SetCharacterizeOn(True)
if ctrlRig:
# Create a control rig using Forward and Inverse Kinematics,
# as specified by the "True" parameter.
character.CreateControlRig(True)
# Set the control rig to active.
character.ActiveInput = True
return character
# This is the Motionbuilder mapping to use the same function. Edit this list or create your own.
mobuMap = {'Reference' : 'reference',
'Hips':'Hips',
'LeftUpLeg' : 'LeftUpLeg',
'LeftLeg' : 'LeftLeg',
'LeftFoot' : 'LeftFoot',
'LeftToeBase': 'LeftToeBase',
'RightUpLeg' : 'RightUpLeg',
'RightLeg' : 'RightLeg',
'RightFoot' : 'RightFoot',
'RightToeBase': 'RightToeBase',
'Spine' : 'Spine',
'LeftArm' : 'LeftArm',
'LeftForeArm' : 'LeftForeArm',
'LeftHand' : 'LeftHand',
'RightArm' : 'RightArm',
'RightForeArm' : 'RightForeArm',
'RightHand' : 'RightHand',
'Head' : 'Head',
'LeftShoulder' : 'LeftShoulder',
'RightShoulder' : 'RightShoulder',
'Neck' : 'Neck',
'Spine1' : 'Spine1',
'Spine2' : 'Spine2',
'Spine3' : 'Spine3',
'Spine4' : 'Spine4',
'Spine5' : 'Spine5',
'Spine6' : 'Spine6',
'Spine7' : 'Spine7',
'Spine8' : 'Spine8',
'Spine9' : 'Spine9',
'Neck1' : 'Neck1',
'Neck2' : 'Neck2',
'Neck3' : 'Neck3',
'Neck4' : 'Neck4',
'Neck5' : 'Neck5',
'Neck6' : 'Neck6',
'Neck7' : 'Neck7',
'Neck8' : 'Neck8',
'Neck9' : 'Neck9',
'LeftHandThumb1' : 'LeftHandThumb1',
'LeftHandThumb2' : 'LeftHandThumb2',
'LeftHandThumb3' : 'LeftHandThumb3',
'LeftHandIndex1' : 'LeftHandIndex1',
'LeftHandIndex2' : 'LeftHandIndex2',
'LeftHandIndex3' : 'LeftHandIndex3',
'LeftHandMiddle1' : 'LeftHandMiddle1',
'LeftHandMiddle2' : 'LeftHandMiddle2',
'LeftHandMiddle3' : 'LeftHandMiddle3',
'LeftHandRing1' : 'LeftHandRing1',
'LeftHandRing2' : 'LeftHandRing2',
'LeftHandRing3' : 'LeftHandRing3',
'LeftHandPinky1' : 'LeftHandPinky1',
'LeftHandPinky2' : 'LeftHandPinky2',
'LeftHandPinky3' : 'LeftHandPinky3',
'RightHandThumb1' : 'RightHandThumb1',
'RightHandThumb2' : 'RightHandThumb2',
'RightHandThumb3' : 'RightHandThumb3',
'RightHandIndex1' : 'RightHandIndex1',
'RightHandIndex2' : 'RightHandIndex2',
'RightHandIndex3' : 'RightHandIndex3',
'RightHandMiddle1' : 'RightHandMiddle1',
'RightHandMiddle2' : 'RightHandMiddle2',
'RightHandMiddle3' : 'RightHandMiddle3',
'RightHandRing1' : 'RightHandRing1',
'RightHandRing2' : 'RightHandRing2',
'RightHandRing3' : 'RightHandRing3',
'RightHandPinky1' : 'RightHandPinky1',
'RightHandPinky2' : 'RightHandPinky2',
'RightHandPinky3' : 'RightHandPinky3',
'LeftFootThumb1' : 'LeftFootThumb1',
'LeftFootThumb2' : 'LeftFootThumb2',
'LeftFootThumb3' : 'LeftFootThumb3',
'LeftFootIndex1' : 'LeftFootIndex1',
'LeftFootIndex2' : 'LeftFootIndex2',
'LeftFootIndex3' : 'LeftFootIndex3',
'LeftFootMiddle1' : 'LeftFootMiddle1',
'LeftFootMiddle2' : 'LeftFootMiddle2',
'LeftFootMiddle3' : 'LeftFootMiddle3',
'LeftFootRing1' : 'LeftFootRing1',
'LeftFootRing2' : 'LeftFootRing2',
'LeftFootRing3' : 'LeftFootRing3',
'LeftFootPinky1' : 'LeftFootPinky1',
'LeftFootPinky2' : 'LeftFootPinky2',
'LeftFootPinky3' : 'LeftFootPinky3',
'RightFootThumb1' : 'RightFootThumb1',
'RightFootThumb2' : 'RightFootThumb2',
'RightFootThumb3' : 'RightFootThumb3',
'RightFootIndex1' : 'RightFootIndex1',
'RightFootIndex2' : 'RightFootIndex2',
'RightFootIndex3' : 'RightFootIndex3',
'RightFootMiddle1' : 'RightFootMiddle1',
'RightFootMiddle2' : 'RightFootMiddle2',
'RightFootMiddle3' : 'RightFootMiddle3',
'RightFootRing1' : 'RightFootRing1',
'RightFootRing2' : 'RightFootRing2',
'RightFootRing3' : 'RightFootRing3',
'RightFootPinky1' : 'RightFootPinky1',
'RightFootPinky2' : 'RightFootPinky2',
'RightFootPinky3' : 'RightFootPinky3',
'LeftUpLegRoll' : 'LeftUpLegRoll',
'LeftLegRoll' : 'LeftLegRoll',
'RightUpLegRoll' : 'RightUpLegRoll',
'RightLegRoll' : 'RightLegRoll',
'LeftArmRoll' : 'LeftArmRoll',
'LeftForeArmRoll' : 'LeftForeArmRoll',
'RightArmRoll' : 'RightArmRoll',
'RightForeArmRoll' : 'RightForeArmRoll' }
def addJointToCharacter ( characterObject, slot, jointName ):
myJoint = FBFindModelByLabelName(jointName)
if myJoint:
proplist = characterObject.PropertyList.Find(slot + "Link")
proplist.append (myJoint)
def CharacterizeBiped(rootname, useBipedPrefixNamingScheme, nameprefix, boneMap, models):
system = FBSystem()
app = FBApplication()
longname = models.LongName
namespaceindex = longname.rfind(":")
if namespaceindex != -1:
namespace = longname[0:namespaceindex+1]
name = longname[namespaceindex + 1:]
else:
namespace = ""
name = longname
myBiped = FBCharacter("mycharacter")
app.CurrentCharacter = myBiped
# If in Biped mode, extract the character prefix name
if useBipedPrefixNamingScheme:
splitname = name.split()
nameprefix = splitname[0] + " "
# Override the rootname so it is the character orefix name
rootname = splitname[0]
myBiped.LongName = namespace + rootname
else:
myBiped.LongName = namespace + nameprefix + rootname
# Create a FBProgress object and set default values for the caption and text.
fbp = FBProgress()
fbp.Caption = ""
fbp.Text = " ----------------------------------- Creating Biped character"
progress = 0.0
progresssteps = len(boneMap)
# assign Biped to Character Mapping.
for pslot, pjointName in boneMap.iteritems():
if not pjointName:
addJointToCharacter(myBiped, pslot, namespace + rootname)
else:
addJointToCharacter(myBiped, pslot, namespace + nameprefix + pjointName)
progress += 1
val = progress / len(boneMap) * 100
fbp.Percent = int(val)
switchOn = myBiped.SetCharacterizeOn( True )
# print "Character mapping created for " + (myBiped.LongName)
# We must call FBDelete when the FBProgress object is no longer needed.
fbp.FBDelete()
return myBiped
def createControlRig(character, activeInput=True):
# Create a control rig using Forward and Inverse Kinematics,
# as specified by the "True" parameter.
character.CreateControlRig(True)
# Set the control rig to active.
if activeInput:
character.ActiveInput = True
def plotAnim(char, animChar):
"""
Receives two characters, sets the input of the first character to the second
and plot. Return ploted character.
"""
#if char.GetCharacterize:
# switchOn = char.SetCharacterizeOn(True)
plotoBla = FBPlotOptions()
plotoBla.ConstantKeyReducerKeepOneKey = True
plotoBla.PlotAllTakes = True
plotoBla.PlotOnFrame = True
plotoBla.PlotPeriod = FBTime( 0, 0, 0, 1 )
plotoBla.PlotTranslationOnRootOnly = True
plotoBla.PreciseTimeDiscontinuities = True
#plotoBla.RotationFilterToApply = FBRotationFilter.kFBRotationFilterGimbleKiller
plotoBla.UseConstantKeyReducer = False
plotoBla.ConstantKeyReducerKeepOneKey = True
char.InputCharacter = animChar
char.InputType = FBCharacterInputType.kFBCharacterInputCharacter
char.ActiveInput = True
if (not char.PlotAnimation(FBCharacterPlotWhere.kFBCharacterPlotOnSkeleton, plotoBla)):
FBMessageBox( "Something went wrong", "Plot animation returned false, cannot continue", "OK", None, None )
return False
return char
def SwitchTake( pTakeName ):
iDestName= pTakeName
for iTake in FBSystem().Scene.Takes:
if iTake.Name == iDestName:
FBSystem().CurrentTake = iTake
def findRoot():
for child in FBSystem().Scene.RootModel.Children:
if "Hips" in child.Name:
return child
return None
def skelExists(root, name):
if root == None: return False
if root.Name == name: return True
for child in root.Children:
if skelExists(child, name):
return True
return False
###############################################################
# Main. #
###############################################################
rootDir = "C:/Users/sunmin/Documents/GitHub/MBScript/"
bvhFileDir = os.path.join(rootDir, "bvh/")
dirList = [os.path.join(bvhFileDir, subDir) for subDir in os.listdir(bvhFileDir) if os.path.isdir(os.path.join(bvhFileDir, subDir))]
fileList = []
## filter short clips
for subDir in dirList:
allList = [os.path.join(subDir, f) for f in os.listdir(subDir) if f.endswith('.bvh')]
fileList += allList
#dirFileList.append(allList)
system = FBSystem()
app = FBApplication()
newFileDir = os.path.join(rootDir, "result/")
if not os.path.exists(newFileDir):
os.makedirs(newFileDir)
logPath = os.path.join(newFileDir, "result_log.txt")
logFile = open(logPath, 'w')
for idx, filePath in enumerate(fileList):
print(idx, filePath)
fileName = filePath.split(os.sep)[-1]
fileType = filePath.split(os.sep)[-2]
logStr = str(idx)+','+fileType+','+fileName+"\n"
logFile.write(logStr)
userRoot = "Hips"
boneMap = mobuMap
bipedPrefixNamingScheme = False
prefix = ""
app.FileNew()
success = app.FileImport(filePath, True)
poseOptions = FBCharacterPoseOptions()
poseOptions.mCharacterPoseKeyingMode = FBCharacterPoseKeyingMode.kFBCharacterPoseKeyingModeFullBody
animRoot = findRoot()
animChar = CharacterizeBiped(userRoot, bipedPrefixNamingScheme, prefix, boneMap, animRoot)
animChar.SelectModels(True, True, True, False)
lPlayer = FBPlayerControl()
lPlayer.Goto(FBTime(0, 0, 0, 0))
createControlRig(animChar, False)
# first key all frame for animation bvh to prevent unwanted interpolation between frames
lEndTime = system.CurrentTake.LocalTimeSpan.GetStop()
lEndFrame = system.CurrentTake.LocalTimeSpan.GetStop().GetFrame()
lStartFrameTime = system.CurrentTake.LocalTimeSpan.GetStart()
lStartFrame = system.CurrentTake.LocalTimeSpan.GetStart().GetFrame()
lRange = min(int(lEndFrame)+1, 50)
lPlayer = FBPlayerControl()
for i in range(lRange):
lPlayer.Goto(FBTime(0, 0, 0, i))
system.Scene.Evaluate()
lPlayer.Key()
FBSystem().Scene.Evaluate()
# Create a new skeleton.
characterName = 'SkelTestCharacter'
skeleton = createSkeleton(characterName)
character = characterizeSkeleton(characterName, skeleton, False)
plotAnim(character, animChar)
animChar.SelectModels(False, True, True, True)
character.SelectModels(True, True, True, True)
# setup save options (for some reason, they were not working outside this loop...)
sOptions = FBFbxOptions(False) # false = save options
sOptions.SaveCharacter = True
sOptions.SaveControlSet = False
sOptions.SaveCharacterExtension = False
sOptions.ShowFileDialog = False
sOptions.ShowOptionsDialog = False
fileName = str(idx)+'.bvh'
newPath = os.path.join(newFileDir, fileName)
FBApplication().FileExport(newPath)
logFile.close()
| [
"os.path.exists",
"os.listdir",
"os.makedirs",
"os.path.join",
"random.random"
] | [((18487, 18516), 'os.path.join', 'os.path.join', (['rootDir', '"""bvh/"""'], {}), "(rootDir, 'bvh/')\n", (18499, 18516), False, 'import os\n'), ((18919, 18951), 'os.path.join', 'os.path.join', (['rootDir', '"""result/"""'], {}), "(rootDir, 'result/')\n", (18931, 18951), False, 'import os\n'), ((19026, 19068), 'os.path.join', 'os.path.join', (['newFileDir', '"""result_log.txt"""'], {}), "(newFileDir, 'result_log.txt')\n", (19038, 19068), False, 'import os\n'), ((3424, 3439), 'random.random', 'random.random', ([], {}), '()\n', (3437, 3439), False, 'import random\n'), ((18528, 18560), 'os.path.join', 'os.path.join', (['bvhFileDir', 'subDir'], {}), '(bvhFileDir, subDir)\n', (18540, 18560), False, 'import os\n'), ((18959, 18985), 'os.path.exists', 'os.path.exists', (['newFileDir'], {}), '(newFileDir)\n', (18973, 18985), False, 'import os\n'), ((18991, 19014), 'os.makedirs', 'os.makedirs', (['newFileDir'], {}), '(newFileDir)\n', (19002, 19014), False, 'import os\n'), ((21226, 21260), 'os.path.join', 'os.path.join', (['newFileDir', 'fileName'], {}), '(newFileDir, fileName)\n', (21238, 21260), False, 'import os\n'), ((2935, 2950), 'random.random', 'random.random', ([], {}), '()\n', (2948, 2950), False, 'import random\n'), ((3141, 3156), 'random.random', 'random.random', ([], {}), '()\n', (3154, 3156), False, 'import random\n'), ((4083, 4098), 'random.random', 'random.random', ([], {}), '()\n', (4096, 4098), False, 'import random\n'), ((4564, 4579), 'random.random', 'random.random', ([], {}), '()\n', (4577, 4579), False, 'import random\n'), ((4888, 4903), 'random.random', 'random.random', ([], {}), '()\n', (4901, 4903), False, 'import random\n'), ((5040, 5055), 'random.random', 'random.random', ([], {}), '()\n', (5053, 5055), False, 'import random\n'), ((5284, 5299), 'random.random', 'random.random', ([], {}), '()\n', (5297, 5299), False, 'import random\n'), ((18575, 18597), 'os.listdir', 'os.listdir', (['bvhFileDir'], {}), '(bvhFileDir)\n', (18585, 18597), False, 'import os\n'), ((18725, 18748), 'os.path.join', 'os.path.join', (['subDir', 'f'], {}), '(subDir, f)\n', (18737, 18748), False, 'import os\n'), ((4343, 4358), 'random.random', 'random.random', ([], {}), '()\n', (4356, 4358), False, 'import random\n'), ((5497, 5512), 'random.random', 'random.random', ([], {}), '()\n', (5510, 5512), False, 'import random\n'), ((18615, 18647), 'os.path.join', 'os.path.join', (['bvhFileDir', 'subDir'], {}), '(bvhFileDir, subDir)\n', (18627, 18647), False, 'import os\n'), ((18758, 18776), 'os.listdir', 'os.listdir', (['subDir'], {}), '(subDir)\n', (18768, 18776), False, 'import os\n'), ((6599, 6614), 'random.random', 'random.random', ([], {}), '()\n', (6612, 6614), False, 'import random\n'), ((6625, 6640), 'random.random', 'random.random', ([], {}), '()\n', (6638, 6640), False, 'import random\n'), ((6651, 6666), 'random.random', 'random.random', ([], {}), '()\n', (6664, 6666), False, 'import random\n')] |
import logging
import pytest
from ophyd.sim import make_fake_device
from pcdsdevices.gon import (BaseGon, Goniometer, GonWithDetArm, Kappa, SamPhi,
XYZStage)
logger = logging.getLogger(__name__)
def test_gon_factory():
logger.debug('test_gon_factory')
assert isinstance(Goniometer(name='gon', prefix_hor='a', prefix_ver='b',
prefix_rot='c', prefix_tip='d',
prefix_tilt='e'), BaseGon)
assert isinstance(Goniometer(name='gon', prefix_hor='a', prefix_ver='b',
prefix_rot='c', prefix_tip='d',
prefix_tilt='e', prefix_detver='i',
prefix_dettilt='j', prefix_2theta='k'),
GonWithDetArm)
@pytest.mark.timeout(5)
def test_gon_init():
logger.debug('test_gon_init')
FakeGon = make_fake_device(BaseGon)
FakeGon(name='test', prefix_hor='hor', prefix_ver='ver',
prefix_rot='rot', prefix_tip='tip', prefix_tilt='tilt')
FakeGon = make_fake_device(GonWithDetArm)
FakeGon(name='test', prefix_hor='hor', prefix_ver='ver',
prefix_rot='rot', prefix_tip='tip', prefix_tilt='tilt',
prefix_detver='detver', prefix_dettilt='dettilt',
prefix_2theta='2theta')
FakeGon = make_fake_device(XYZStage)
FakeGon(name='test', prefix_x='x', prefix_y='y', prefix_z='z')
FakeGon = make_fake_device(SamPhi)
FakeGon(name='test', prefix_samz='samz', prefix_samphi='samphi')
FakeGon = make_fake_device(Kappa)
FakeGon(name='test', prefix_x='x', prefix_y='y', prefix_z='z',
prefix_eta='eta', prefix_kappa='kappa', prefix_phi='phi')
@pytest.mark.timeout(5)
def test_gon_disconnected():
logger.debug('test_gon_disconnected')
BaseGon(name='test1', prefix_hor='hor', prefix_ver='ver',
prefix_rot='rot', prefix_tip='tip', prefix_tilt='tilt')
GonWithDetArm(name='test2', prefix_hor='hor', prefix_ver='ver',
prefix_rot='rot', prefix_tip='tip', prefix_tilt='tilt',
prefix_detver='detver', prefix_dettilt='dettilt',
prefix_2theta='2theta')
XYZStage(name='test3', prefix_x='x', prefix_y='y', prefix_z='z')
SamPhi(name='test4', prefix_samz='samz', prefix_samphi='samphi')
Kappa(name='test5', prefix_x='x', prefix_y='y', prefix_z='z',
prefix_eta='eta', prefix_kappa='kappa', prefix_phi='phi')
| [
"logging.getLogger",
"pcdsdevices.gon.BaseGon",
"pcdsdevices.gon.SamPhi",
"pcdsdevices.gon.Kappa",
"pcdsdevices.gon.XYZStage",
"pcdsdevices.gon.Goniometer",
"pytest.mark.timeout",
"pcdsdevices.gon.GonWithDetArm",
"ophyd.sim.make_fake_device"
] | [((199, 226), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (216, 226), False, 'import logging\n'), ((816, 838), 'pytest.mark.timeout', 'pytest.mark.timeout', (['(5)'], {}), '(5)\n', (835, 838), False, 'import pytest\n'), ((1730, 1752), 'pytest.mark.timeout', 'pytest.mark.timeout', (['(5)'], {}), '(5)\n', (1749, 1752), False, 'import pytest\n'), ((908, 933), 'ophyd.sim.make_fake_device', 'make_fake_device', (['BaseGon'], {}), '(BaseGon)\n', (924, 933), False, 'from ophyd.sim import make_fake_device\n'), ((1077, 1108), 'ophyd.sim.make_fake_device', 'make_fake_device', (['GonWithDetArm'], {}), '(GonWithDetArm)\n', (1093, 1108), False, 'from ophyd.sim import make_fake_device\n'), ((1350, 1376), 'ophyd.sim.make_fake_device', 'make_fake_device', (['XYZStage'], {}), '(XYZStage)\n', (1366, 1376), False, 'from ophyd.sim import make_fake_device\n'), ((1458, 1482), 'ophyd.sim.make_fake_device', 'make_fake_device', (['SamPhi'], {}), '(SamPhi)\n', (1474, 1482), False, 'from ophyd.sim import make_fake_device\n'), ((1566, 1589), 'ophyd.sim.make_fake_device', 'make_fake_device', (['Kappa'], {}), '(Kappa)\n', (1582, 1589), False, 'from ophyd.sim import make_fake_device\n'), ((1828, 1945), 'pcdsdevices.gon.BaseGon', 'BaseGon', ([], {'name': '"""test1"""', 'prefix_hor': '"""hor"""', 'prefix_ver': '"""ver"""', 'prefix_rot': '"""rot"""', 'prefix_tip': '"""tip"""', 'prefix_tilt': '"""tilt"""'}), "(name='test1', prefix_hor='hor', prefix_ver='ver', prefix_rot='rot',\n prefix_tip='tip', prefix_tilt='tilt')\n", (1835, 1945), False, 'from pcdsdevices.gon import BaseGon, Goniometer, GonWithDetArm, Kappa, SamPhi, XYZStage\n'), ((1958, 2160), 'pcdsdevices.gon.GonWithDetArm', 'GonWithDetArm', ([], {'name': '"""test2"""', 'prefix_hor': '"""hor"""', 'prefix_ver': '"""ver"""', 'prefix_rot': '"""rot"""', 'prefix_tip': '"""tip"""', 'prefix_tilt': '"""tilt"""', 'prefix_detver': '"""detver"""', 'prefix_dettilt': '"""dettilt"""', 'prefix_2theta': '"""2theta"""'}), "(name='test2', prefix_hor='hor', prefix_ver='ver', prefix_rot=\n 'rot', prefix_tip='tip', prefix_tilt='tilt', prefix_detver='detver',\n prefix_dettilt='dettilt', prefix_2theta='2theta')\n", (1971, 2160), False, 'from pcdsdevices.gon import BaseGon, Goniometer, GonWithDetArm, Kappa, SamPhi, XYZStage\n'), ((2210, 2274), 'pcdsdevices.gon.XYZStage', 'XYZStage', ([], {'name': '"""test3"""', 'prefix_x': '"""x"""', 'prefix_y': '"""y"""', 'prefix_z': '"""z"""'}), "(name='test3', prefix_x='x', prefix_y='y', prefix_z='z')\n", (2218, 2274), False, 'from pcdsdevices.gon import BaseGon, Goniometer, GonWithDetArm, Kappa, SamPhi, XYZStage\n'), ((2279, 2343), 'pcdsdevices.gon.SamPhi', 'SamPhi', ([], {'name': '"""test4"""', 'prefix_samz': '"""samz"""', 'prefix_samphi': '"""samphi"""'}), "(name='test4', prefix_samz='samz', prefix_samphi='samphi')\n", (2285, 2343), False, 'from pcdsdevices.gon import BaseGon, Goniometer, GonWithDetArm, Kappa, SamPhi, XYZStage\n'), ((2348, 2472), 'pcdsdevices.gon.Kappa', 'Kappa', ([], {'name': '"""test5"""', 'prefix_x': '"""x"""', 'prefix_y': '"""y"""', 'prefix_z': '"""z"""', 'prefix_eta': '"""eta"""', 'prefix_kappa': '"""kappa"""', 'prefix_phi': '"""phi"""'}), "(name='test5', prefix_x='x', prefix_y='y', prefix_z='z', prefix_eta=\n 'eta', prefix_kappa='kappa', prefix_phi='phi')\n", (2353, 2472), False, 'from pcdsdevices.gon import BaseGon, Goniometer, GonWithDetArm, Kappa, SamPhi, XYZStage\n'), ((312, 419), 'pcdsdevices.gon.Goniometer', 'Goniometer', ([], {'name': '"""gon"""', 'prefix_hor': '"""a"""', 'prefix_ver': '"""b"""', 'prefix_rot': '"""c"""', 'prefix_tip': '"""d"""', 'prefix_tilt': '"""e"""'}), "(name='gon', prefix_hor='a', prefix_ver='b', prefix_rot='c',\n prefix_tip='d', prefix_tilt='e')\n", (322, 419), False, 'from pcdsdevices.gon import BaseGon, Goniometer, GonWithDetArm, Kappa, SamPhi, XYZStage\n'), ((514, 683), 'pcdsdevices.gon.Goniometer', 'Goniometer', ([], {'name': '"""gon"""', 'prefix_hor': '"""a"""', 'prefix_ver': '"""b"""', 'prefix_rot': '"""c"""', 'prefix_tip': '"""d"""', 'prefix_tilt': '"""e"""', 'prefix_detver': '"""i"""', 'prefix_dettilt': '"""j"""', 'prefix_2theta': '"""k"""'}), "(name='gon', prefix_hor='a', prefix_ver='b', prefix_rot='c',\n prefix_tip='d', prefix_tilt='e', prefix_detver='i', prefix_dettilt='j',\n prefix_2theta='k')\n", (524, 683), False, 'from pcdsdevices.gon import BaseGon, Goniometer, GonWithDetArm, Kappa, SamPhi, XYZStage\n')] |
#!/usr/bin/env python3
import sys
import os
import argparse
import time
import serial
import csv
import math
import pickle
from collections import defaultdict
import numpy as np
from sklearn.decomposition import PCA, FastICA
from sklearn.svm import SVC
# Graph
WINDOW_WIDTH = 800
WINDOW_HEIGHT = 800
PLOT_SCROLL = 3 # higher is faster
CHANNELS = 8
FONT_SIZE = 25
# Data
FREQUENCY = 200 # Hz
CSV_HEADER_EMG = ["timestamp", "emg1", "emg2", "emg3", "emg4", "emg5", "emg6", "emg7", "emg8"]
CSV_HEADER_CA = ["timestamp", "ca1", "ca2", "ca3", "ca4", "ca5", "ca6", "ca7", "ca8"]
# Processing
RMS_WINDOW_SIZE = 50
SVM_WINDOW_SIZE = 5 # higher is smoother but more delay
SVM_IDLE_WEIGHT_FACTOR = 100.0 # higher makes "idle" move more important
VERBOSE = False
# Plotting (Pygame) window interface
class Plotter():
def __init__(self, live=False):
if "pygame" not in sys.modules:
print("Error! pygame not loaded! Plotter not available for library use.")
return None
self.screen = pygame.display.set_mode((WINDOW_WIDTH, WINDOW_HEIGHT))
pygame.display.set_caption("Electromyography Processor")
self.font = pygame.font.Font(None, FONT_SIZE)
self.live = live
self.last_values = None
self.last_rms_values = None
self.last_ca_values = None
self.plots = 0
def plot(self, values, rms_values=[], ca_values=[], ca="", gesture="", frequency=None, recording=False):
if self.last_values is None:
self.last_values = values
self.last_rms_values = rms_values
self.last_ca_values = ca_values
self.plots = len(values) + len(ca_values)
return
self.screen.scroll(-PLOT_SCROLL)
self.screen.fill(pygame.Color("black"), (WINDOW_WIDTH - PLOT_SCROLL, 0, WINDOW_WIDTH, WINDOW_HEIGHT))
self.screen.fill(pygame.Color("black"), (0, 0, 60, WINDOW_HEIGHT))
self.clear_info()
# Subplot base
for i in range(self.plots):
base_height = self.subplot_height(i)
pygame.draw.line(self.screen, pygame.Color("darkgrey"),
(WINDOW_WIDTH - PLOT_SCROLL, base_height),
(WINDOW_WIDTH, base_height))
if i < 8 and self.plots >= 8: # Raw / RMS
plot_text = self.font.render(f"RAW {i}", True, pygame.Color("darkgrey"))
rms_offset = 10 if rms_values else 0
if rms_values:
plot_rms = self.font.render(f"RMS {i}", True, pygame.Color("blue"))
self.screen.blit(plot_rms, (0, base_height - rms_offset - FONT_SIZE // 2))
self.screen.blit(plot_text, (0, base_height + rms_offset - FONT_SIZE // 2))
else: # PCA/ICA
plot_text = self.font.render(f" {ca.upper()} {i - len(values)}", True, pygame.Color("green"))
self.screen.blit(plot_text, (0, base_height - FONT_SIZE // 2))
# Raw signal
for i, (u, v) in enumerate(zip(self.last_values, values)):
pygame.draw.line(self.screen, pygame.Color("darkslategrey"),
(WINDOW_WIDTH - PLOT_SCROLL, self.subplot_height(i, u)),
(WINDOW_WIDTH, self.subplot_height(i, v)))
# Processed signals
if rms_values:
for i, (u, v) in enumerate(zip(self.last_rms_values, rms_values)):
pygame.draw.line(self.screen, pygame.Color("blue"),
(WINDOW_WIDTH - PLOT_SCROLL, self.subplot_height(i, u)),
(WINDOW_WIDTH, self.subplot_height(i, v)))
if ca_values:
for i, (u, v) in enumerate(zip(self.last_ca_values, ca_values)):
pygame.draw.line(self.screen, pygame.Color("green"),
(WINDOW_WIDTH - PLOT_SCROLL, self.subplot_height(i + len(rms_values), u)),
(WINDOW_WIDTH, self.subplot_height(i + len(rms_values), v)))
# Information
if frequency:
self.render_frequency(frequency)
self.render_mode()
self.render_controls(recording)
if gesture:
self.render_classification(gesture)
pygame.display.flip()
self.last_values = values
self.last_rms_values = rms_values
self.last_ca_values = ca_values
def subplot_height(self, i, value=0):
scaled_value = value * 1.5
return int(WINDOW_HEIGHT / (self.plots + 1) * (i + 1 - scaled_value))
def clear_info(self):
self.screen.fill(pygame.Color("black"), (0, 0, WINDOW_WIDTH, FONT_SIZE))
self.screen.fill(pygame.Color("black"), (0, WINDOW_HEIGHT - FONT_SIZE, WINDOW_WIDTH, WINDOW_HEIGHT))
def render_mode(self):
mode_text = "LIVE" if self.live else "PLAYBACK"
mode = self.font.render("LIVE" if self.live else "PLAYBACK",
True, pygame.Color("green"))
self.screen.blit(mode, (WINDOW_WIDTH // 2 - len(mode_text) * FONT_SIZE // 2, 0))
def render_frequency(self, frequency):
framerate = self.font.render(f"{frequency} Hz", True,
pygame.Color("green") if frequency > 180 else pygame.Color("red"))
self.screen.fill(pygame.Color("black"), (0, 0, 75, FONT_SIZE)) # Clear old framerate
self.screen.blit(framerate, (0, 0))
def render_controls(self, recording):
pause = self.font.render("P (pause)", True, pygame.Color("white"))
self.screen.blit(pause, (WINDOW_WIDTH - 250, 0))
if self.live: # Can only record live
record = self.font.render("R (stop rec)" if recording else "R (record)", True,
pygame.Color("red") if recording else pygame.Color("white"))
self.screen.blit(record, (WINDOW_WIDTH - 150, 0))
def render_classification(self, gesture):
plot_gesture = self.font.render(f"Classification: {gesture}", True, pygame.Color("green"))
self.screen.blit(plot_gesture, (WINDOW_WIDTH // 2 - 225, WINDOW_HEIGHT - FONT_SIZE))
def pause(self):
self.clear_info()
pause = self.font.render("P (resume)", True, pygame.Color("red"))
self.screen.blit(pause, (WINDOW_WIDTH - 250, 0))
self.render_mode()
pygame.display.flip()
def end(self):
self.clear_info()
pause = self.font.render("END", True, pygame.Color("red"))
self.screen.blit(pause, (WINDOW_WIDTH - 250, 0))
self.render_mode()
pygame.display.flip()
# Interface for data streaming from either live Myo device or recorded playback
class Stream():
def __init__(self, do_rms=False, pca_train_set=[], ica_train_set=[], ca_components=3, svm_train_set=[]):
self.plotter = None # Late setup (display modes)
self.reset()
# Processing
self.do_rms = do_rms
self.ca_components = ca_components
self.pca = self.init_pca(pca_train_set) if pca_train_set else None
self.ica = self.init_ica(ica_train_set) if ica_train_set else None
self.svm = self.init_svm(svm_train_set) if svm_train_set else None
self.gesture = ""
def create_plot(self, live=False):
self.plotter = Plotter(live=live)
def plot(self, data, ca=False, recording=False):
self.calc_frequency()
# Processing
rms_data, ca_data = [], []
if ca:
ca_data, data = data, []
else:
if self.do_rms or self.pca is not None or self.ica is not None:
rms_data = self.calc_rms(data)
ca_data = []
if self.pca is not None:
ca_data = self.calc_pca(rms_data)
elif self.ica is not None:
ca_data = self.calc_ica(rms_data)
if self.svm is not None:
self.gesture = self.class_svm(ca_data)
if not self.paused and self.plotter is not None:
self.plotter.plot([x / 500. for x in data],
rms_values=[x / 500. for x in rms_data],
ca_values=[x / 500. for x in ca_data],
ca=self.current_model()[1],
gesture=self.gesture,
frequency=self.frequency,
recording=recording)
return rms_data, ca_data, self.gesture
def calc_frequency(self):
self.times.append(time.time())
if len(self.times) >= 100:
self.frequency = int((len(self.times) - 1) / (self.times[-1] - self.times[0]))
self.times.clear()
def pause(self, state=False, toggle=False):
if toggle:
self.paused = not self.paused
else:
self.paused = state
if self.paused and not self.ended:
self.plotter.pause()
def end(self):
self.ended = True
if self.plotter is not None:
self.plotter.end()
def reset(self):
self.paused = False
self.ended = False
# Frequency measuring
self.times = []
self.frequency = 0
# Processing
self.rms_window = []
self.svm_window = []
# Processing
def calc_rms(self, data):
# Gather samples, up to RMS_WINDOW_SIZE
self.rms_window.append(data)
if len(self.rms_window) >= RMS_WINDOW_SIZE:
self.rms_window.pop(0)
# Calculate RMS for each channel
rms_data = [0] * CHANNELS
for channel in range(CHANNELS):
samples = [item[channel] for item in self.rms_window]
total = sum([sample ** 2 for sample in samples])
rms_data[channel] = math.sqrt(1.0 / RMS_WINDOW_SIZE * total)
if VERBOSE:
print(f"rms: {rms_data}")
return rms_data
def read_ca_train_set(self, train_set, stype="?"):
emg_data = []
for file in train_set:
print(f"Training {stype.upper()} with '{file}'...")
emg_file = open(file, "r", newline="")
emg_reader = csv.reader(emg_file, csv.unix_dialect, quoting=csv.QUOTE_MINIMAL)
# Read file
header = next(emg_reader)
if header == CSV_HEADER_EMG:
try:
while True:
data = next(emg_reader)
_, emg = data[0], list(map(int, data[1:]))
emg_data.append(self.calc_rms(emg))
except StopIteration:
pass
else:
print("-> Error! Incorrect header! Expected 'RAW'.")
self.rms_window.clear()
emg_file.close()
emg_data = np.array(emg_data)
return emg_data
def read_model(self, model, stype="?"):
print(f"Reading {stype.upper()} model '{model}'...")
with open(model, "rb") as f:
return pickle.load(f)
def init_pca(self, train_set):
if isinstance(train_set, list):
emg_data = self.read_ca_train_set(train_set, "pca")
# Initialize and train
pca = PCA(n_components=self.ca_components)
pca.fit(emg_data)
else:
pca = self.read_model(train_set, "pca")
return pca
def calc_pca(self, rms_data):
emg_data = np.array(rms_data).reshape(1, -1) # Reshape to 1 sample, N features
pca_data = self.pca.transform(emg_data)[0] # Take 1 sample from array of samples (contains only one)
if VERBOSE:
print(f"pca: {pca_data}")
return pca_data
def init_ica(self, train_set):
if isinstance(train_set, list):
emg_data = self.read_ca_train_set(train_set, "ica")
# Initialize and train
ica = FastICA(n_components=self.ca_components, random_state=0)
ica.fit(emg_data)
else:
ica = self.read_model(train_set, "ica")
return ica
def calc_ica(self, rms_data):
emg_data = np.array(rms_data).reshape(1, -1) # Reshape to 1 sample, N features
ica_data = self.ica.transform(emg_data)[0] # Take 1 sample from array of samples (contains only one)
ica_data *= 5000 # Scale up
if VERBOSE:
print(f"ica: {ica_data}")
return ica_data
def read_class_train_set(self, train_set, stype="?"):
emg_data = []
classes = []
for file in train_set:
classification = os.path.basename(file).split("_")[0]
print(f"Training {stype.upper()} '{classification}' with '{file}'...")
emg_file = open(file, "r", newline="")
emg_reader = csv.reader(emg_file, csv.unix_dialect, quoting=csv.QUOTE_MINIMAL)
# Read file
header = next(emg_reader)
if header == CSV_HEADER_CA[:self.ca_components + 1]:
try:
while True:
data = next(emg_reader)
_, emg = data[0], list(map(float, data[1:]))
emg_data.append(emg)
classes.append(classification)
except StopIteration:
pass
else:
print("-> Error! Incorrect header! Expected 'PCA/ICA'.")
emg_file.close()
if "idle" not in classes:
print("Warning! No 'idle' move trained!")
emg_data, classes = np.array(emg_data), np.array(classes)
return emg_data, classes
def init_svm(self, train_set):
if isinstance(train_set, list):
emg_data, classes = self.read_class_train_set(train_set, "svm")
svm = SVC(random_state=0, kernel="rbf", class_weight={"idle": SVM_IDLE_WEIGHT_FACTOR})
svm.fit(emg_data, classes)
else:
svm = self.read_model(train_set, "svm")
return svm
def class_svm(self, ca_data):
# Gather samples, up to SVM_WINDOW_SIZE to smooth classification
self.svm_window.append(ca_data)
if len(self.svm_window) > SVM_WINDOW_SIZE:
self.svm_window.pop(0)
window = np.array(self.svm_window)
svm_classes = self.svm.predict(window) # predict each sample in window
# Take classification with most occurences in the window
d = defaultdict(int)
for svm_class in svm_classes:
d[svm_class] += 1
svm_class = max(d.items(), key=lambda x: x[1])[0]
if VERBOSE:
print(f"svm: {svm_class}")
return svm_class
return ""
def current_model(self):
if self.svm is not None:
return self.svm, "svm"
elif self.pca is not None:
return self.pca, "pca"
elif self.ica is not None:
return self.ica, "ica"
return None, ""
# Live Myo device interface
class Myo():
def __init__(self, stream, tty, native, mac):
# Instantiate
self.myo = MyoRaw(tty, native, mac)
self.stream = stream
self.recording = False
self.recording_type = self.init_recording()
# Recording
self.emg_file = None
self.emg_writer = None
# Setup
self.setup()
def close(self):
self.myo.disconnect()
self.record(False)
def setup(self):
# Add handles to process EMG and battery level data
self.myo.add_handler(DataCategory.EMG, self.handle_emg)
self.myo.add_handler(DataCategory.BATTERY, self.handle_battery)
# Subscribe to all data services in full RAW mode (200 Hz)
self.myo.subscribe(EMGMode.RAW)
# Disable sleep to a void disconnects while retrieving data
self.myo.set_sleep_mode(1)
# Vibrate to signalise a successful setup
# myo.vibrate(1)
def run(self):
self.myo.run(1)
def disconnect(self):
self.myo.disconnect()
def sleep(self):
self.myo.deep_sleep()
def handle_emg(self, timestamp, emg, moving, characteristic_num):
emg = list(emg)
_, ca_data, _ = self.stream.plot(emg, recording=self.recording)
record_data = ca_data if len(ca_data) > 0 else emg
if self.recording:
csv_data = [timestamp]
csv_data.extend(record_data)
try:
self.emg_writer.writerow(csv_data)
except AttributeError:
print("Error! Unable to write to CSV!")
if VERBOSE:
print(f"[myo] {self.recording_type}: {timestamp}, {record_data}")
def handle_battery(self, timestamp, battery_level):
if battery_level < 5:
self.myo.set_leds([255, 0, 0], [255, 0, 0]) # red logo, red bar
else:
self.myo.set_leds([128, 128, 255], [128, 128, 255]) # purple logo, purple bar
if VERBOSE:
print(f"[myo] battery level: {timestamp}, {battery_level}")
def init_recording(self):
if self.stream.pca is not None:
return "pca"
elif self.stream.ica is not None:
return "ica"
return "raw"
def record(self, state=False, toggle=False):
if toggle:
recording = not self.recording
else:
recording = state
if recording:
filename = f"recordings/{self.recording_type}/{time.strftime('%Y%m%d-%H%M%S')}.csv"
os.makedirs(os.path.dirname(filename), exist_ok=True)
self.emg_file = open(filename, "w", newline="")
self.emg_writer = csv.writer(self.emg_file, csv.unix_dialect, quoting=csv.QUOTE_MINIMAL)
if self.recording_type == "raw":
self.emg_writer.writerow(CSV_HEADER_EMG)
else:
self.emg_writer.writerow(CSV_HEADER_CA[:self.stream.ca_components + 1])
elif self.emg_file is not None:
self.emg_file.close()
self.emg_file = None
self.emg_writer = None
self.recording = recording
# Recorded Myo data playback interface
class Playback():
def __init__(self, stream, filename):
self.stream = stream
self.valid = False
self.type = ""
try:
self.emg_file = open(filename, "r", newline="")
self.emg_reader = csv.reader(self.emg_file, csv.unix_dialect, quoting=csv.QUOTE_MINIMAL)
self.read_header()
except FileNotFoundError:
self.emg_file = None
def close(self):
if self.emg_file:
self.emg_file.close()
def read_header(self):
try:
header = next(self.emg_reader)
if header == CSV_HEADER_EMG:
self.valid = True
self.type = "raw"
if header[:2] == CSV_HEADER_CA[:2]:
self.valid = True
self.type = "ca"
except StopIteration:
pass
def is_valid(self):
return self.valid
# Plays a frame from the recording and indicating end of recording on subsequent calls
def play_frame(self):
if not self.stream.paused:
try:
data = next(self.emg_reader)
if self.type == "raw":
timestamp, emg = data[0], list(map(int, data[1:]))
rms_data, ca_data, gesture = self.stream.plot(emg)
else:
timestamp, emg = data[0], list(map(float, data[1:]))
rms_data, ca_data, gesture = self.stream.plot(emg, ca=True)
if VERBOSE:
print(f"[playback] emg: {timestamp}, {emg}")
return timestamp, rms_data, ca_data, gesture
except StopIteration:
self.stream.end()
return 0, [], [], ""
def main():
# Parse arguments
parser = argparse.ArgumentParser(description="Electromyography Processor")
group1 = parser.add_mutually_exclusive_group()
group1.add_argument("-r", "--recording", default=None, metavar="REC", help="playback recorded Myo data stream")
group1.add_argument("-s", "--sleep", default=False, action="store_true", help="put Myo into deep sleep (turn off)")
parser.add_argument("--rms", default=False, action="store_true", help="process stream using RMS smoothing")
group2 = parser.add_mutually_exclusive_group()
group2.add_argument("--pca", nargs="+", metavar="REC", help="process stream using RAW training set or PCA model")
group2.add_argument("--ica", nargs="+", metavar="REC", help="process stream using RAW training set or ICA model")
parser.add_argument("-c", "--components", default=3, type=int, help="PCA/ICA components to use")
group3 = parser.add_mutually_exclusive_group()
group3.add_argument("--svm", nargs="+", metavar="REC", help="classify using PCA/ICA training set or SVM model")
group4 = parser.add_mutually_exclusive_group()
group4.add_argument("--tty", default=None, help="Myo dongle device (autodetected if omitted)")
group4.add_argument("--native", default=False, action="store_true", help="use a native Bluetooth stack")
parser.add_argument("--mac", default=None, help="Myo MAC address (arbitrarily detected if omitted)")
parser.add_argument("-v", "--verbose", default=False, action="store_true", help="verbose output")
args = parser.parse_args()
if args.svm and not args.pca and not args.ica:
parser.error("the following arguments are required for 'svm': 'pca' or 'ica'")
# Model was given instead of trainining set
if args.pca is not None and len(args.pca) == 1 and not args.pca[0].endswith(".csv"):
args.pca = args.pca[0]
if args.ica is not None and len(args.ica) == 1 and not args.ica[0].endswith(".csv"):
args.ica = args.ica[0]
if args.svm is not None and len(args.svm) == 1 and not args.svm[0].endswith(".csv"):
args.svm = args.svm[0]
if args.verbose:
global VERBOSE
VERBOSE = args.verbose
live_myo = args.recording is None
# Setup common stream interface for Myo or Playback
stream = Stream(do_rms=args.rms, pca_train_set=args.pca, ica_train_set=args.ica, svm_train_set=args.svm,
ca_components=args.components)
# Setup Myo or Playback
if live_myo:
try:
print("Connecting to Myo...")
myo = Myo(stream, args.tty, args.native, args.mac)
print("Connected to Myo!")
except (ValueError, KeyboardInterrupt) as e:
print(f"Error! Unable to connect!\n{e}")
return 1
else:
playback = Playback(stream, args.recording)
if not playback.is_valid():
print("Error! Invalid CSV file!")
return 2
# Run main logic
if args.sleep:
if live_myo:
myo.sleep()
else:
pygame.init()
stream.create_plot(live=live_myo)
# Run until terminated by user or recording ended
try:
starttime = time.time()
while True:
if live_myo:
try:
myo.run()
except serial.serialutil.SerialException:
print("Error! Myo exception! Attempting reboot...")
myo.disconnect()
myo = Myo(stream, args.tty, args.native, args.mac)
else:
playback.play_frame()
# Delay by (1 second / FREQUENCY Hz) including execution time
delay = 1 / FREQUENCY
diff = min(time.time() - starttime, 1 / FREQUENCY)
time.sleep(delay - diff)
starttime = time.time()
# Handle Pygame events
for ev in pygame.event.get():
if ev.type == pygame.QUIT:
raise KeyboardInterrupt()
elif ev.type == pygame.KEYDOWN:
if ev.key == pygame.K_q:
raise KeyboardInterrupt()
elif ev.key == pygame.K_p:
stream.pause(toggle=True)
elif ev.key == pygame.K_r:
if live_myo:
myo.record(toggle=True)
except KeyboardInterrupt:
pass
if live_myo:
myo.close()
else:
playback.close()
return 0
# Conditional imports
if __name__ == "__main__" or os.environ.get("EMGPROC_LOAD_GAME", False):
import pygame
if __name__ == "__main__" or os.environ.get("EMGPROC_LOAD_MYO", False):
from myo_raw import MyoRaw, DataCategory, EMGMode
if __name__ == "__main__":
sys.exit(main())
| [
"pygame.init",
"math.sqrt",
"time.sleep",
"numpy.array",
"sklearn.decomposition.FastICA",
"pygame.font.Font",
"argparse.ArgumentParser",
"sklearn.decomposition.PCA",
"pygame.display.set_mode",
"pygame.display.flip",
"csv.reader",
"csv.writer",
"pickle.load",
"os.path.dirname",
"time.time... | [((19921, 19986), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {'description': '"""Electromyography Processor"""'}), "(description='Electromyography Processor')\n", (19944, 19986), False, 'import argparse\n'), ((24584, 24626), 'os.environ.get', 'os.environ.get', (['"""EMGPROC_LOAD_GAME"""', '(False)'], {}), "('EMGPROC_LOAD_GAME', False)\n", (24598, 24626), False, 'import os\n'), ((24675, 24716), 'os.environ.get', 'os.environ.get', (['"""EMGPROC_LOAD_MYO"""', '(False)'], {}), "('EMGPROC_LOAD_MYO', False)\n", (24689, 24716), False, 'import os\n'), ((1025, 1079), 'pygame.display.set_mode', 'pygame.display.set_mode', (['(WINDOW_WIDTH, WINDOW_HEIGHT)'], {}), '((WINDOW_WIDTH, WINDOW_HEIGHT))\n', (1048, 1079), False, 'import pygame\n'), ((1088, 1144), 'pygame.display.set_caption', 'pygame.display.set_caption', (['"""Electromyography Processor"""'], {}), "('Electromyography Processor')\n", (1114, 1144), False, 'import pygame\n'), ((1165, 1198), 'pygame.font.Font', 'pygame.font.Font', (['None', 'FONT_SIZE'], {}), '(None, FONT_SIZE)\n', (1181, 1198), False, 'import pygame\n'), ((4272, 4293), 'pygame.display.flip', 'pygame.display.flip', ([], {}), '()\n', (4291, 4293), False, 'import pygame\n'), ((6363, 6384), 'pygame.display.flip', 'pygame.display.flip', ([], {}), '()\n', (6382, 6384), False, 'import pygame\n'), ((6591, 6612), 'pygame.display.flip', 'pygame.display.flip', ([], {}), '()\n', (6610, 6612), False, 'import pygame\n'), ((10782, 10800), 'numpy.array', 'np.array', (['emg_data'], {}), '(emg_data)\n', (10790, 10800), True, 'import numpy as np\n'), ((15081, 15105), 'myo_raw.MyoRaw', 'MyoRaw', (['tty', 'native', 'mac'], {}), '(tty, native, mac)\n', (15087, 15105), False, 'from myo_raw import MyoRaw, DataCategory, EMGMode\n'), ((22923, 22936), 'pygame.init', 'pygame.init', ([], {}), '()\n', (22934, 22936), False, 'import pygame\n'), ((1767, 1788), 'pygame.Color', 'pygame.Color', (['"""black"""'], {}), "('black')\n", (1779, 1788), False, 'import pygame\n'), ((1877, 1898), 'pygame.Color', 'pygame.Color', (['"""black"""'], {}), "('black')\n", (1889, 1898), False, 'import pygame\n'), ((4619, 4640), 'pygame.Color', 'pygame.Color', (['"""black"""'], {}), "('black')\n", (4631, 4640), False, 'import pygame\n'), ((4700, 4721), 'pygame.Color', 'pygame.Color', (['"""black"""'], {}), "('black')\n", (4712, 4721), False, 'import pygame\n'), ((4975, 4996), 'pygame.Color', 'pygame.Color', (['"""green"""'], {}), "('green')\n", (4987, 4996), False, 'import pygame\n'), ((5322, 5343), 'pygame.Color', 'pygame.Color', (['"""black"""'], {}), "('black')\n", (5334, 5343), False, 'import pygame\n'), ((5530, 5551), 'pygame.Color', 'pygame.Color', (['"""white"""'], {}), "('white')\n", (5542, 5551), False, 'import pygame\n'), ((6032, 6053), 'pygame.Color', 'pygame.Color', (['"""green"""'], {}), "('green')\n", (6044, 6053), False, 'import pygame\n'), ((6249, 6268), 'pygame.Color', 'pygame.Color', (['"""red"""'], {}), "('red')\n", (6261, 6268), False, 'import pygame\n'), ((6477, 6496), 'pygame.Color', 'pygame.Color', (['"""red"""'], {}), "('red')\n", (6489, 6496), False, 'import pygame\n'), ((8520, 8531), 'time.time', 'time.time', ([], {}), '()\n', (8529, 8531), False, 'import time\n'), ((9771, 9811), 'math.sqrt', 'math.sqrt', (['(1.0 / RMS_WINDOW_SIZE * total)'], {}), '(1.0 / RMS_WINDOW_SIZE * total)\n', (9780, 9811), False, 'import math\n'), ((10147, 10212), 'csv.reader', 'csv.reader', (['emg_file', 'csv.unix_dialect'], {'quoting': 'csv.QUOTE_MINIMAL'}), '(emg_file, csv.unix_dialect, quoting=csv.QUOTE_MINIMAL)\n', (10157, 10212), False, 'import csv\n'), ((10987, 11001), 'pickle.load', 'pickle.load', (['f'], {}), '(f)\n', (10998, 11001), False, 'import pickle\n'), ((11196, 11232), 'sklearn.decomposition.PCA', 'PCA', ([], {'n_components': 'self.ca_components'}), '(n_components=self.ca_components)\n', (11199, 11232), False, 'from sklearn.decomposition import PCA, FastICA\n'), ((11860, 11916), 'sklearn.decomposition.FastICA', 'FastICA', ([], {'n_components': 'self.ca_components', 'random_state': '(0)'}), '(n_components=self.ca_components, random_state=0)\n', (11867, 11916), False, 'from sklearn.decomposition import PCA, FastICA\n'), ((12747, 12812), 'csv.reader', 'csv.reader', (['emg_file', 'csv.unix_dialect'], {'quoting': 'csv.QUOTE_MINIMAL'}), '(emg_file, csv.unix_dialect, quoting=csv.QUOTE_MINIMAL)\n', (12757, 12812), False, 'import csv\n'), ((13514, 13532), 'numpy.array', 'np.array', (['emg_data'], {}), '(emg_data)\n', (13522, 13532), True, 'import numpy as np\n'), ((13534, 13551), 'numpy.array', 'np.array', (['classes'], {}), '(classes)\n', (13542, 13551), True, 'import numpy as np\n'), ((13756, 13841), 'sklearn.svm.SVC', 'SVC', ([], {'random_state': '(0)', 'kernel': '"""rbf"""', 'class_weight': "{'idle': SVM_IDLE_WEIGHT_FACTOR}"}), "(random_state=0, kernel='rbf', class_weight={'idle': SVM_IDLE_WEIGHT_FACTOR}\n )\n", (13759, 13841), False, 'from sklearn.svm import SVC\n'), ((14218, 14243), 'numpy.array', 'np.array', (['self.svm_window'], {}), '(self.svm_window)\n', (14226, 14243), True, 'import numpy as np\n'), ((14414, 14430), 'collections.defaultdict', 'defaultdict', (['int'], {}), '(int)\n', (14425, 14430), False, 'from collections import defaultdict\n'), ((17642, 17712), 'csv.writer', 'csv.writer', (['self.emg_file', 'csv.unix_dialect'], {'quoting': 'csv.QUOTE_MINIMAL'}), '(self.emg_file, csv.unix_dialect, quoting=csv.QUOTE_MINIMAL)\n', (17652, 17712), False, 'import csv\n'), ((18383, 18453), 'csv.reader', 'csv.reader', (['self.emg_file', 'csv.unix_dialect'], {'quoting': 'csv.QUOTE_MINIMAL'}), '(self.emg_file, csv.unix_dialect, quoting=csv.QUOTE_MINIMAL)\n', (18393, 18453), False, 'import csv\n'), ((23075, 23086), 'time.time', 'time.time', ([], {}), '()\n', (23084, 23086), False, 'import time\n'), ((2104, 2128), 'pygame.Color', 'pygame.Color', (['"""darkgrey"""'], {}), "('darkgrey')\n", (2116, 2128), False, 'import pygame\n'), ((3113, 3142), 'pygame.Color', 'pygame.Color', (['"""darkslategrey"""'], {}), "('darkslategrey')\n", (3125, 3142), False, 'import pygame\n'), ((5230, 5251), 'pygame.Color', 'pygame.Color', (['"""green"""'], {}), "('green')\n", (5242, 5251), False, 'import pygame\n'), ((5276, 5295), 'pygame.Color', 'pygame.Color', (['"""red"""'], {}), "('red')\n", (5288, 5295), False, 'import pygame\n'), ((11403, 11421), 'numpy.array', 'np.array', (['rms_data'], {}), '(rms_data)\n', (11411, 11421), True, 'import numpy as np\n'), ((12087, 12105), 'numpy.array', 'np.array', (['rms_data'], {}), '(rms_data)\n', (12095, 12105), True, 'import numpy as np\n'), ((17510, 17535), 'os.path.dirname', 'os.path.dirname', (['filename'], {}), '(filename)\n', (17525, 17535), False, 'import os\n'), ((23868, 23886), 'pygame.event.get', 'pygame.event.get', ([], {}), '()\n', (23884, 23886), False, 'import pygame\n'), ((2379, 2403), 'pygame.Color', 'pygame.Color', (['"""darkgrey"""'], {}), "('darkgrey')\n", (2391, 2403), False, 'import pygame\n'), ((2880, 2901), 'pygame.Color', 'pygame.Color', (['"""green"""'], {}), "('green')\n", (2892, 2901), False, 'import pygame\n'), ((3479, 3499), 'pygame.Color', 'pygame.Color', (['"""blue"""'], {}), "('blue')\n", (3491, 3499), False, 'import pygame\n'), ((3813, 3834), 'pygame.Color', 'pygame.Color', (['"""green"""'], {}), "('green')\n", (3825, 3834), False, 'import pygame\n'), ((5786, 5805), 'pygame.Color', 'pygame.Color', (['"""red"""'], {}), "('red')\n", (5798, 5805), False, 'import pygame\n'), ((5824, 5845), 'pygame.Color', 'pygame.Color', (['"""white"""'], {}), "('white')\n", (5836, 5845), False, 'import pygame\n'), ((17449, 17479), 'time.strftime', 'time.strftime', (['"""%Y%m%d-%H%M%S"""'], {}), "('%Y%m%d-%H%M%S')\n", (17462, 17479), False, 'import time\n'), ((23733, 23757), 'time.sleep', 'time.sleep', (['(delay - diff)'], {}), '(delay - diff)\n', (23743, 23757), False, 'import time\n'), ((23790, 23801), 'time.time', 'time.time', ([], {}), '()\n', (23799, 23801), False, 'import time\n'), ((2555, 2575), 'pygame.Color', 'pygame.Color', (['"""blue"""'], {}), "('blue')\n", (2567, 2575), False, 'import pygame\n'), ((12550, 12572), 'os.path.basename', 'os.path.basename', (['file'], {}), '(file)\n', (12566, 12572), False, 'import os\n'), ((23673, 23684), 'time.time', 'time.time', ([], {}), '()\n', (23682, 23684), False, 'import time\n')] |
import numpy as np
from preprocess import Vectorizer
from flask import render_template, make_response
from google.oauth2.id_token import verify_oauth2_token
from google.auth.transport.requests import Request
from google.cloud import firestore
from os.path import join, abspath, dirname
from random import randint
from pickle import load
from scipy.sparse import load_npz
database = firestore.Client()
hedgy_path = dirname(abspath(__file__))
with open(join(hedgy_path, 'chapters.p'), 'rb') as chapters_f, open(join(hedgy_path, 'vectorizer.p'), 'rb') as vectorizer_f:
chapters = load(chapters_f)
vectorizer = load(vectorizer_f)
tfidf_matrix = load_npz(join(hedgy_path, 'tfidf.npz'))
similarity_matrix = np.load(join(hedgy_path, 'similarity.npy'))
def hedgy(request):
ranking, sliced, max_request, seed, token = [], False, 50, None, None
credential, clicks = None, None
if request.method == 'POST':
credential = request.form['credential'] if 'credential' in request.form else None
elif request.cookies.get('__session'):
credential, clicks = request.cookies.get('__session').split('#')
if credential:
try:
token = verify_oauth2_token(credential, Request(), '1080182836213-psdjtgo2u10a1fb6e4sbdfpdlmco5i63.apps.googleusercontent.com')
except:
pass
if token:
user_doc = database.collection('users').document(token['sub'])
if not user_doc.get().exists:
user_doc.set({'email': token['email'], 'given_name': token['given_name'], 'family_name': token['family_name'], 'picture': token['picture'], 'clicks': []})
if clicks:
user_doc.update({'clicks': firestore.ArrayUnion(clicks[:-1].split(','))})
if 'max' in request.args:
max_request = int(request.args.get('max'))
if 'query' in request.args or 'similar' in request.args:
if 'query' in request.args:
query_vector = vectorizer.transform([request.args.get('query')])
similarity_vector = (tfidf_matrix @ query_vector.T).toarray().squeeze()
else:
similarity_vector = similarity_matrix[int(request.args.get('similar'))]
if np.any(similarity_vector):
max_chapters = np.count_nonzero(similarity_vector)
if max_request < max_chapters:
max_chapters = max_request
sliced = True
ranking = np.argsort(similarity_vector)[::-1][:max_chapters].tolist()
elif 'seed' in request.args:
seed = int(request.args.get('seed'))
else:
seed = randint(1, 1000000)
if seed:
np.random.seed(seed)
ranking = np.random.permutation(len(chapters))[:max_request].tolist()
sliced = True
response = make_response(render_template('hedgy.html', chapters=chapters, ranking=ranking, sliced=sliced, max_request=max_request, seed=seed, token=token, args=request.args))
if token:
response.set_cookie('__session', credential + '#', secure=True)
response.headers['Cache-Control'] = 'private'
elif '__session' in request.cookies:
response.set_cookie('__session', '', expires=0)
response.headers['Cache-Control'] = 'private'
return response
| [
"flask.render_template",
"google.cloud.firestore.Client",
"google.auth.transport.requests.Request",
"pickle.load",
"os.path.join",
"numpy.any",
"numpy.count_nonzero",
"numpy.argsort",
"numpy.random.seed",
"os.path.abspath",
"random.randint"
] | [((383, 401), 'google.cloud.firestore.Client', 'firestore.Client', ([], {}), '()\n', (399, 401), False, 'from google.cloud import firestore\n'), ((424, 441), 'os.path.abspath', 'abspath', (['__file__'], {}), '(__file__)\n', (431, 441), False, 'from os.path import join, abspath, dirname\n'), ((583, 599), 'pickle.load', 'load', (['chapters_f'], {}), '(chapters_f)\n', (587, 599), False, 'from pickle import load\n'), ((617, 635), 'pickle.load', 'load', (['vectorizer_f'], {}), '(vectorizer_f)\n', (621, 635), False, 'from pickle import load\n'), ((660, 689), 'os.path.join', 'join', (['hedgy_path', '"""tfidf.npz"""'], {}), "(hedgy_path, 'tfidf.npz')\n", (664, 689), False, 'from os.path import join, abspath, dirname\n'), ((719, 753), 'os.path.join', 'join', (['hedgy_path', '"""similarity.npy"""'], {}), "(hedgy_path, 'similarity.npy')\n", (723, 753), False, 'from os.path import join, abspath, dirname\n'), ((453, 483), 'os.path.join', 'join', (['hedgy_path', '"""chapters.p"""'], {}), "(hedgy_path, 'chapters.p')\n", (457, 483), False, 'from os.path import join, abspath, dirname\n'), ((511, 543), 'os.path.join', 'join', (['hedgy_path', '"""vectorizer.p"""'], {}), "(hedgy_path, 'vectorizer.p')\n", (515, 543), False, 'from os.path import join, abspath, dirname\n'), ((2621, 2640), 'random.randint', 'randint', (['(1)', '(1000000)'], {}), '(1, 1000000)\n', (2628, 2640), False, 'from random import randint\n'), ((2662, 2682), 'numpy.random.seed', 'np.random.seed', (['seed'], {}), '(seed)\n', (2676, 2682), True, 'import numpy as np\n'), ((2812, 2965), 'flask.render_template', 'render_template', (['"""hedgy.html"""'], {'chapters': 'chapters', 'ranking': 'ranking', 'sliced': 'sliced', 'max_request': 'max_request', 'seed': 'seed', 'token': 'token', 'args': 'request.args'}), "('hedgy.html', chapters=chapters, ranking=ranking, sliced=\n sliced, max_request=max_request, seed=seed, token=token, args=request.args)\n", (2827, 2965), False, 'from flask import render_template, make_response\n'), ((2202, 2227), 'numpy.any', 'np.any', (['similarity_vector'], {}), '(similarity_vector)\n', (2208, 2227), True, 'import numpy as np\n'), ((1210, 1219), 'google.auth.transport.requests.Request', 'Request', ([], {}), '()\n', (1217, 1219), False, 'from google.auth.transport.requests import Request\n'), ((2260, 2295), 'numpy.count_nonzero', 'np.count_nonzero', (['similarity_vector'], {}), '(similarity_vector)\n', (2276, 2295), True, 'import numpy as np\n'), ((2450, 2479), 'numpy.argsort', 'np.argsort', (['similarity_vector'], {}), '(similarity_vector)\n', (2460, 2479), True, 'import numpy as np\n')] |
#-*- coding:utf-8 -*-
import sys, traceback
from core_backend import context
from core_backend.libs.exception import Error
import logging
#logger = Log.getDebugLogger()
#logger.setLevel(logging.INFO)
logger = logging.getLogger(__name__)
class plugin(object):
def __init__(self, handler, session):
self.handler = handler
self.session = session
self.context = self.handler.context
self.request = self.context.request
self.service = self.handler._service
def process(self):
pass
def post_process(self):
pass
class PluginHandler(object):
def __init__(self, handler, session):
self.handler = handler
self.session = session
self.plg_modules = []
self.plg_inst_list = []
def import_module(self, module, fromlist):
# return __import__(self.get_module(m), fromlist=["plugins"])
return __import__(module, fromlist=fromlist)
def load_plugins(self, plg_module):
plgconfig = self.import_module(plg_module, [plg_module])
module_files = plgconfig.plugins_modules
for f in module_files:
m = self.import_module(plg_module + '.' + f, [plg_module])
self.plg_modules.append(m)
ins = m.Plugin(self.handler, self.session)
self.plg_inst_list.append(ins)
def run_plugins(self):
for ins in self.plg_inst_list:
ins.process()
def run_post_plugins(self):
for ins in self.plg_inst_list:
ins.post_process()
| [
"logging.getLogger"
] | [((210, 237), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (227, 237), False, 'import logging\n')] |
'''
Filename: predict.py
Python Version: 3.6.5
Project: Neutrophil Identifier
Author: <NAME>
Created date: Sep 5, 2018 4:13 PM
-----
Last Modified: Oct 9, 2018 3:48 PM
Modified By: <NAME>
-----
License: MIT
http://www.opensource.org/licenses/MIT
'''
import os
import sys
import logging
from math import ceil
from keras.models import load_model
import numpy as np
import tables as tb
from paths import Paths
def read_hdf5(hdf5_file, dataset="pred", batch_size=32):
"""
"""
data_img = dataset + "_img"
m_data = hdf5_file.root.__getitem__(data_img).shape[0]
batch_list = list(range(int(ceil(m_data / batch_size))))
while True:
for num in batch_list:
n_start = num * batch_size
n_end = min((num + 1) * batch_size, m_data)
inputs = hdf5_file.root.__getitem__(data_img)[n_start:n_end, ...]
yield inputs
def predict(model_path, hdf5_file_path=None):
"""
"""
try:
BATCH_SIZE = 32
model = load_model(model_path)
default_path = Paths.tiles_80
if hdf5_file_path:
hdf5_file = tb.open_file(hdf5_file_path, mode='r')
else:
hdf5_file = tb.open_file(default_path, mode='r')
m_samples = hdf5_file.root.__getitem__("pred_img").shape[0]
logging.debug(f'# of samples: {m_samples}')
steps = int(ceil(m_samples / BATCH_SIZE))
generator = read_hdf5(hdf5_file, dataset="pred", batch_size=BATCH_SIZE)
preds = model.predict_generator(generator, steps=steps, verbose=1)
logging.info(preds[0:100])
base_name = os.path.basename(model_path).split('.')[0]
save_path = os.path.join(
Paths.data_test, base_name + "_tiles_80_preds.csv")
np.savetxt(save_path, preds, delimiter=',')
except Exception as e:
hdf5_file.close()
logging.debug(e.with_traceback())
finally:
hdf5_file.close()
if __name__ == "__main__":
logging.basicConfig(level=logging.DEBUG)
predict(sys.argv[1])
| [
"logging.basicConfig",
"keras.models.load_model",
"logging.debug",
"math.ceil",
"os.path.join",
"tables.open_file",
"os.path.basename",
"numpy.savetxt",
"logging.info"
] | [((1965, 2005), 'logging.basicConfig', 'logging.basicConfig', ([], {'level': 'logging.DEBUG'}), '(level=logging.DEBUG)\n', (1984, 2005), False, 'import logging\n'), ((996, 1018), 'keras.models.load_model', 'load_model', (['model_path'], {}), '(model_path)\n', (1006, 1018), False, 'from keras.models import load_model\n'), ((1300, 1343), 'logging.debug', 'logging.debug', (['f"""# of samples: {m_samples}"""'], {}), "(f'# of samples: {m_samples}')\n", (1313, 1343), False, 'import logging\n'), ((1557, 1583), 'logging.info', 'logging.info', (['preds[0:100]'], {}), '(preds[0:100])\n', (1569, 1583), False, 'import logging\n'), ((1668, 1732), 'os.path.join', 'os.path.join', (['Paths.data_test', "(base_name + '_tiles_80_preds.csv')"], {}), "(Paths.data_test, base_name + '_tiles_80_preds.csv')\n", (1680, 1732), False, 'import os\n'), ((1754, 1797), 'numpy.savetxt', 'np.savetxt', (['save_path', 'preds'], {'delimiter': '""","""'}), "(save_path, preds, delimiter=',')\n", (1764, 1797), True, 'import numpy as np\n'), ((1108, 1146), 'tables.open_file', 'tb.open_file', (['hdf5_file_path'], {'mode': '"""r"""'}), "(hdf5_file_path, mode='r')\n", (1120, 1146), True, 'import tables as tb\n'), ((1186, 1222), 'tables.open_file', 'tb.open_file', (['default_path'], {'mode': '"""r"""'}), "(default_path, mode='r')\n", (1198, 1222), True, 'import tables as tb\n'), ((1364, 1392), 'math.ceil', 'ceil', (['(m_samples / BATCH_SIZE)'], {}), '(m_samples / BATCH_SIZE)\n', (1368, 1392), False, 'from math import ceil\n'), ((608, 633), 'math.ceil', 'ceil', (['(m_data / batch_size)'], {}), '(m_data / batch_size)\n', (612, 633), False, 'from math import ceil\n'), ((1605, 1633), 'os.path.basename', 'os.path.basename', (['model_path'], {}), '(model_path)\n', (1621, 1633), False, 'import os\n')] |
import argparse
import json
from tqdm import tqdm
from common.dataset.reader import JSONLineReader
from common.util.log_helper import LogHelper
def _sent_to_str(sent):
return sent[-2] + "$$$" + str(sent[-1])
def _replace_sent_with_str(sent, string):
segments = string.split(r"$$$")
if len(segments) != 2:
raise Exception("Illegal string: " + string)
sent[-2] = segments[0]
sent[-1] = int(segments[1])
return sent
def _build_new_sent_with_str(string, num_of_segments=2):
if num_of_segments == 2:
sent = ["", -1]
elif num_of_segments == 4:
sent = [-1, -1, "", -1]
else:
raise Exception("Illegal num_of_segments: " + str(num_of_segments))
return _replace_sent_with_str(sent, string)
def _sents_from_evidences(evidences):
sents = set()
for evidence in evidences:
for s in evidence:
sent = _sent_to_str(s)
sents.add(sent)
return sents
def _fill_pred_sents_with_gold(pred_sents, gold_sents, max_sent):
selected_sents = pred_sents[:max_sent]
neg_indices = []
for i, selected in enumerate(selected_sents):
key_selected = _sent_to_str(selected)
if key_selected in gold_sents:
gold_sents.remove(key_selected)
else:
neg_indices.append(i)
if len(gold_sents) == 0:
return selected_sents
if len(selected_sents) <= max_sent:
for _ in range(max_sent - len(selected_sents)):
selected_sents.append(_build_new_sent_with_str(gold_sents.pop()))
if len(gold_sents) == 0:
return selected_sents
if len(neg_indices) > 0:
neg_indices = reversed(neg_indices)
for i in neg_indices:
sent = selected_sents[i]
selected_sents[i] = _replace_sent_with_str(sent, gold_sents.pop())
if len(gold_sents) == 0:
return selected_sents
if len(gold_sents) > 0:
logger.warn(str(len(gold_sents)) +
" gold sentences cannot be filled into prediction")
return selected_sents
if __name__ == '__main__':
LogHelper.setup()
logger = LogHelper.get_logger('fill_gold_sentences')
parser = argparse.ArgumentParser()
parser.add_argument(
'--input', help='/path/to/input/file', required=True)
parser.add_argument(
'--output', help='/path/to/output/file', required=True)
parser.add_argument(
'--max-sent', type=int, help='Maximal number of sentences per claim', default=10)
args = parser.parse_args()
jlr = JSONLineReader()
data = jlr.read(args.input)
with open(args.output, "w+") as output_file:
for data in tqdm(data):
if data['verifiable'] != 'NOT VERIFIABLE':
pred_sents = data['predicted_sentences']
gold_evidences = data['evidence']
gold_sents = _sents_from_evidences(gold_evidences)
filled_pred_sents = _fill_pred_sents_with_gold(
pred_sents, gold_sents, args.max_sent)
data['predicted_sentences'] = filled_pred_sents
output_file.write(json.dumps(data) + "\n")
| [
"argparse.ArgumentParser",
"common.util.log_helper.LogHelper.setup",
"common.dataset.reader.JSONLineReader",
"common.util.log_helper.LogHelper.get_logger",
"tqdm.tqdm",
"json.dumps"
] | [((2117, 2134), 'common.util.log_helper.LogHelper.setup', 'LogHelper.setup', ([], {}), '()\n', (2132, 2134), False, 'from common.util.log_helper import LogHelper\n'), ((2148, 2191), 'common.util.log_helper.LogHelper.get_logger', 'LogHelper.get_logger', (['"""fill_gold_sentences"""'], {}), "('fill_gold_sentences')\n", (2168, 2191), False, 'from common.util.log_helper import LogHelper\n'), ((2205, 2230), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {}), '()\n', (2228, 2230), False, 'import argparse\n'), ((2563, 2579), 'common.dataset.reader.JSONLineReader', 'JSONLineReader', ([], {}), '()\n', (2577, 2579), False, 'from common.dataset.reader import JSONLineReader\n'), ((2681, 2691), 'tqdm.tqdm', 'tqdm', (['data'], {}), '(data)\n', (2685, 2691), False, 'from tqdm import tqdm\n'), ((3139, 3155), 'json.dumps', 'json.dumps', (['data'], {}), '(data)\n', (3149, 3155), False, 'import json\n')] |
# Code generated by `typeddictgen`. DO NOT EDIT.
"""V1ISCSIVolumeSourceDict generated type."""
from typing import TypedDict, List
from kubernetes_typed.client import V1LocalObjectReferenceDict
V1ISCSIVolumeSourceDict = TypedDict(
"V1ISCSIVolumeSourceDict",
{
"chapAuthDiscovery": bool,
"chapAuthSession": bool,
"fsType": str,
"initiatorName": str,
"iqn": str,
"iscsiInterface": str,
"lun": int,
"portals": List[str],
"readOnly": bool,
"secretRef": V1LocalObjectReferenceDict,
"targetPortal": str,
},
total=False,
)
| [
"typing.TypedDict"
] | [((221, 526), 'typing.TypedDict', 'TypedDict', (['"""V1ISCSIVolumeSourceDict"""', "{'chapAuthDiscovery': bool, 'chapAuthSession': bool, 'fsType': str,\n 'initiatorName': str, 'iqn': str, 'iscsiInterface': str, 'lun': int,\n 'portals': List[str], 'readOnly': bool, 'secretRef':\n V1LocalObjectReferenceDict, 'targetPortal': str}"], {'total': '(False)'}), "('V1ISCSIVolumeSourceDict', {'chapAuthDiscovery': bool,\n 'chapAuthSession': bool, 'fsType': str, 'initiatorName': str, 'iqn':\n str, 'iscsiInterface': str, 'lun': int, 'portals': List[str],\n 'readOnly': bool, 'secretRef': V1LocalObjectReferenceDict,\n 'targetPortal': str}, total=False)\n", (230, 526), False, 'from typing import TypedDict, List\n')] |
from pytpp.attributes._helper import IterableMeta, Attribute
from pytpp.attributes.workflow import WorkflowAttributes
class AdaptableWorkflowAttributes(WorkflowAttributes, metaclass=IterableMeta):
__config_class__ = "Adaptable Workflow"
adaptable_workflow_text_field_1 = Attribute('Adaptable Workflow Text Field 1', min_version='18.3')
adaptable_workflow_text_field_10 = Attribute('Adaptable Workflow Text Field 10', min_version='18.3')
adaptable_workflow_text_field_11 = Attribute('Adaptable Workflow Text Field 11', min_version='18.3')
adaptable_workflow_text_field_12 = Attribute('Adaptable Workflow Text Field 12', min_version='18.3')
adaptable_workflow_text_field_2 = Attribute('Adaptable Workflow Text Field 2', min_version='18.3')
adaptable_workflow_text_field_3 = Attribute('Adaptable Workflow Text Field 3', min_version='18.3')
adaptable_workflow_text_field_4 = Attribute('Adaptable Workflow Text Field 4', min_version='18.3')
adaptable_workflow_text_field_5 = Attribute('Adaptable Workflow Text Field 5', min_version='18.3')
adaptable_workflow_text_field_6 = Attribute('Adaptable Workflow Text Field 6', min_version='18.3')
adaptable_workflow_text_field_7 = Attribute('Adaptable Workflow Text Field 7', min_version='18.3')
adaptable_workflow_text_field_8 = Attribute('Adaptable Workflow Text Field 8', min_version='18.3')
adaptable_workflow_text_field_9 = Attribute('Adaptable Workflow Text Field 9', min_version='18.3')
credential = Attribute('Credential', min_version='18.3')
log_debug = Attribute('Log Debug', min_version='19.3')
oauth_token_application_id = Attribute('OAuth Token Application Id', min_version='21.3')
oauth_token_credential = Attribute('OAuth Token Credential', min_version='21.3')
oauth_token_scope = Attribute('OAuth Token Scope', min_version='21.3')
powershell_script = Attribute('PowerShell Script', min_version='18.3')
powershell_script_hash_vault_id = Attribute('PowerShell Script Hash Vault Id', min_version='19.1')
script_execution_timeout = Attribute('Script Execution Timeout', min_version='20.2')
secondary_credential = Attribute('Secondary Credential', min_version='18.3')
service_address = Attribute('Service Address', min_version='18.3')
| [
"pytpp.attributes._helper.Attribute"
] | [((275, 339), 'pytpp.attributes._helper.Attribute', 'Attribute', (['"""Adaptable Workflow Text Field 1"""'], {'min_version': '"""18.3"""'}), "('Adaptable Workflow Text Field 1', min_version='18.3')\n", (284, 339), False, 'from pytpp.attributes._helper import IterableMeta, Attribute\n'), ((376, 441), 'pytpp.attributes._helper.Attribute', 'Attribute', (['"""Adaptable Workflow Text Field 10"""'], {'min_version': '"""18.3"""'}), "('Adaptable Workflow Text Field 10', min_version='18.3')\n", (385, 441), False, 'from pytpp.attributes._helper import IterableMeta, Attribute\n'), ((478, 543), 'pytpp.attributes._helper.Attribute', 'Attribute', (['"""Adaptable Workflow Text Field 11"""'], {'min_version': '"""18.3"""'}), "('Adaptable Workflow Text Field 11', min_version='18.3')\n", (487, 543), False, 'from pytpp.attributes._helper import IterableMeta, Attribute\n'), ((580, 645), 'pytpp.attributes._helper.Attribute', 'Attribute', (['"""Adaptable Workflow Text Field 12"""'], {'min_version': '"""18.3"""'}), "('Adaptable Workflow Text Field 12', min_version='18.3')\n", (589, 645), False, 'from pytpp.attributes._helper import IterableMeta, Attribute\n'), ((681, 745), 'pytpp.attributes._helper.Attribute', 'Attribute', (['"""Adaptable Workflow Text Field 2"""'], {'min_version': '"""18.3"""'}), "('Adaptable Workflow Text Field 2', min_version='18.3')\n", (690, 745), False, 'from pytpp.attributes._helper import IterableMeta, Attribute\n'), ((781, 845), 'pytpp.attributes._helper.Attribute', 'Attribute', (['"""Adaptable Workflow Text Field 3"""'], {'min_version': '"""18.3"""'}), "('Adaptable Workflow Text Field 3', min_version='18.3')\n", (790, 845), False, 'from pytpp.attributes._helper import IterableMeta, Attribute\n'), ((881, 945), 'pytpp.attributes._helper.Attribute', 'Attribute', (['"""Adaptable Workflow Text Field 4"""'], {'min_version': '"""18.3"""'}), "('Adaptable Workflow Text Field 4', min_version='18.3')\n", (890, 945), False, 'from pytpp.attributes._helper import IterableMeta, Attribute\n'), ((981, 1045), 'pytpp.attributes._helper.Attribute', 'Attribute', (['"""Adaptable Workflow Text Field 5"""'], {'min_version': '"""18.3"""'}), "('Adaptable Workflow Text Field 5', min_version='18.3')\n", (990, 1045), False, 'from pytpp.attributes._helper import IterableMeta, Attribute\n'), ((1081, 1145), 'pytpp.attributes._helper.Attribute', 'Attribute', (['"""Adaptable Workflow Text Field 6"""'], {'min_version': '"""18.3"""'}), "('Adaptable Workflow Text Field 6', min_version='18.3')\n", (1090, 1145), False, 'from pytpp.attributes._helper import IterableMeta, Attribute\n'), ((1181, 1245), 'pytpp.attributes._helper.Attribute', 'Attribute', (['"""Adaptable Workflow Text Field 7"""'], {'min_version': '"""18.3"""'}), "('Adaptable Workflow Text Field 7', min_version='18.3')\n", (1190, 1245), False, 'from pytpp.attributes._helper import IterableMeta, Attribute\n'), ((1281, 1345), 'pytpp.attributes._helper.Attribute', 'Attribute', (['"""Adaptable Workflow Text Field 8"""'], {'min_version': '"""18.3"""'}), "('Adaptable Workflow Text Field 8', min_version='18.3')\n", (1290, 1345), False, 'from pytpp.attributes._helper import IterableMeta, Attribute\n'), ((1381, 1445), 'pytpp.attributes._helper.Attribute', 'Attribute', (['"""Adaptable Workflow Text Field 9"""'], {'min_version': '"""18.3"""'}), "('Adaptable Workflow Text Field 9', min_version='18.3')\n", (1390, 1445), False, 'from pytpp.attributes._helper import IterableMeta, Attribute\n'), ((1460, 1503), 'pytpp.attributes._helper.Attribute', 'Attribute', (['"""Credential"""'], {'min_version': '"""18.3"""'}), "('Credential', min_version='18.3')\n", (1469, 1503), False, 'from pytpp.attributes._helper import IterableMeta, Attribute\n'), ((1517, 1559), 'pytpp.attributes._helper.Attribute', 'Attribute', (['"""Log Debug"""'], {'min_version': '"""19.3"""'}), "('Log Debug', min_version='19.3')\n", (1526, 1559), False, 'from pytpp.attributes._helper import IterableMeta, Attribute\n'), ((1590, 1649), 'pytpp.attributes._helper.Attribute', 'Attribute', (['"""OAuth Token Application Id"""'], {'min_version': '"""21.3"""'}), "('OAuth Token Application Id', min_version='21.3')\n", (1599, 1649), False, 'from pytpp.attributes._helper import IterableMeta, Attribute\n'), ((1676, 1731), 'pytpp.attributes._helper.Attribute', 'Attribute', (['"""OAuth Token Credential"""'], {'min_version': '"""21.3"""'}), "('OAuth Token Credential', min_version='21.3')\n", (1685, 1731), False, 'from pytpp.attributes._helper import IterableMeta, Attribute\n'), ((1753, 1803), 'pytpp.attributes._helper.Attribute', 'Attribute', (['"""OAuth Token Scope"""'], {'min_version': '"""21.3"""'}), "('OAuth Token Scope', min_version='21.3')\n", (1762, 1803), False, 'from pytpp.attributes._helper import IterableMeta, Attribute\n'), ((1825, 1875), 'pytpp.attributes._helper.Attribute', 'Attribute', (['"""PowerShell Script"""'], {'min_version': '"""18.3"""'}), "('PowerShell Script', min_version='18.3')\n", (1834, 1875), False, 'from pytpp.attributes._helper import IterableMeta, Attribute\n'), ((1911, 1975), 'pytpp.attributes._helper.Attribute', 'Attribute', (['"""PowerShell Script Hash Vault Id"""'], {'min_version': '"""19.1"""'}), "('PowerShell Script Hash Vault Id', min_version='19.1')\n", (1920, 1975), False, 'from pytpp.attributes._helper import IterableMeta, Attribute\n'), ((2004, 2061), 'pytpp.attributes._helper.Attribute', 'Attribute', (['"""Script Execution Timeout"""'], {'min_version': '"""20.2"""'}), "('Script Execution Timeout', min_version='20.2')\n", (2013, 2061), False, 'from pytpp.attributes._helper import IterableMeta, Attribute\n'), ((2086, 2139), 'pytpp.attributes._helper.Attribute', 'Attribute', (['"""Secondary Credential"""'], {'min_version': '"""18.3"""'}), "('Secondary Credential', min_version='18.3')\n", (2095, 2139), False, 'from pytpp.attributes._helper import IterableMeta, Attribute\n'), ((2159, 2207), 'pytpp.attributes._helper.Attribute', 'Attribute', (['"""Service Address"""'], {'min_version': '"""18.3"""'}), "('Service Address', min_version='18.3')\n", (2168, 2207), False, 'from pytpp.attributes._helper import IterableMeta, Attribute\n')] |
#!/usr/env/python
"""
Generalized Cyclic Multi-Pass Sweep-line Algorithm
The cyclic multi-pass sweep-line algorithm, simply repeatedly
sweeps across the same Timeline, until a specified number of passes
has been completed or some signal is given to stop sweeping.
Classes:
- CycleSweep
"""
from typing import TypeVar, Union
from sources.abstract import Subscriber, Timeline
from .basesweep import Sweepline
T = TypeVar('T')
class CycleSweep(Sweepline[T]):
"""
The generalized cyclic multi-pass sweep-line algorithm.
Generics:
T: Objects type within the Timeline.
Extends:
Sweepline[T]
"""
### Methods: Evaluation
def evaluate(self, iterations: int = -1, *args, evparams_kw = {}, **kwargs):
"""
Execute the cyclic multi-pass sweep-line algorithm over the
attached Timeline. Broadcast Events to the Observers.
Args:
iterations:
The number of passes to run the algorithm over
the Timeline. If negative value, the algorithm will
continue until one of the subscriber calls the added
stopiteration() method in the Events as a parameter.
evparams_kw:
Arguments for event.setparams().
args, kwargs:
Arguments for timeline.events().
"""
N = 0
status = {'stop': False}
def stopiter():
status['stop'] = True
while 0 > iterations or iterations > N:
# cycle through each event in the timeline, one-pass
for event in self.timeline.events(*args, **kwargs):
event.setparams(iteration=N, stopiteration=stopiter, **evparams_kw)
self.on_next(event)
# stopiteration called
if status['stop']:
iterations = 0
# increment sweep counter
N += 1
self.on_completed()
| [
"typing.TypeVar"
] | [((419, 431), 'typing.TypeVar', 'TypeVar', (['"""T"""'], {}), "('T')\n", (426, 431), False, 'from typing import TypeVar, Union\n')] |
import numpy
import pandas as pd
from boo.columns import SHORT_COLUMNS
from boo.errors import UnclassifiableCodeError
QUOTE_CHAR = '"'
EMPTY = int(0)
NUMERIC_COLUMNS = SHORT_COLUMNS.numeric
def adjust_rub(df, cols=NUMERIC_COLUMNS):
rows = (df.unit == "385")
df.loc[rows, cols] = df.loc[rows, cols].multiply(1000)
df.loc[rows, "unit"] = "384"
rows = (df.unit == "383")
df.loc[rows, cols] = df.loc[rows, cols].divide(1000).round(0).astype(int)
df.loc[rows, "unit"] = "384"
return df
def dequote(name: str):
"""Split company *name* to organisation and title."""
# Warning: will not work well on company names with more than 4 quotechars
parts = name.split(QUOTE_CHAR)
org = parts[0].strip()
cnt = name.count(QUOTE_CHAR)
if cnt == 2:
title = parts[1].strip()
elif cnt > 2:
title = QUOTE_CHAR.join(parts[1:])
else:
title = name
return org, title.strip()
def replace_names(title: str):
return title .replace(
"ПУБЛИЧНОЕ АКЦИОНЕРНОЕ ОБЩЕСТВО",
"ПАО") .replace(
"ОТКРЫТОЕ АКЦИОНЕРНОЕ ОБЩЕСТВО",
"ОАО") .replace(
"АКЦИОНЕРНОЕ ОБЩЕСТВО ЭНЕРГЕТИКИ И ЭЛЕКТРИФИКАЦИИ",
"AO энерго") .replace(
"НЕФТЕПЕРЕРАБАТЫВАЮЩИЙ ЗАВОД",
"НПЗ") .replace(
"ГЕНЕРИРУЮЩАЯ КОМПАНИЯ ОПТОВОГО РЫНКА ЭЛЕКТРОЭНЕРГИИ",
"ОГК") .replace(
"ГОРНО-ОБОГАТИТЕЛЬНЫЙ КОМБИНАТ",
"ГОК")
def add_title(df):
s_ = df.name.apply(dequote)
df['org'] = s_.apply(lambda x: x[0])
df['title'] = s_.apply(lambda x: replace_names(x[1]))
return df
def rename_rows(df):
RENAME_DICT = {
'2460066195': "РусГидро",
'4716016979': "ФСК ЕЭС",
'7702038150': "Московский метрополитен",
'7721632827': "Концерн Росэнергоатом",
'7706664260': "Атомэнергопром",
'7703683145': "Холдинг ВТБ Капитал АЙ БИ",
'9102048801': "Черноморнефтегаз",
'7736036626': "РИТЭК"
}
keys = RENAME_DICT.keys()
ix = df.index.isin(keys)
if not ix.any():
return df
sub = df.loc[ix, 'title']
for k, v in RENAME_DICT.items():
if k in sub.index:
sub.loc[k] = v
df.loc[ix, 'title'] = sub
return df
def split_okved(code_string: str):
"""Get 3 levels of OKVED codes from *code_string*."""
if code_string.count(".") > 2:
raise UnclassifiableCodeError(code_string)
try:
codes = [int(x) for x in code_string.split(".")]
except ValueError:
raise UnclassifiableCodeError(code_string)
return codes + [0] * (3 - len(codes))
def add_okved_subcode(df):
df['ok1'], df['ok2'], df['ok3'] = zip(*df.okved.apply(split_okved))
return df
def fst(x):
try:
return int(x[0:2])
except TypeError:
return 0
def add_region(df):
df['region'] = df.inn.apply(fst)
return df
def more_columns(df):
return add_okved_subcode(add_region(add_title(df)))
def canonic_df(df):
"""Преобразовать данные внтури датафрейма:
- Привести все строки к одинаковым единицам измерения (тыс. руб.)
- Убрать неиспользуемые колонки (date_revised, report_type)
- Новые колонки:
* короткое название компании
* три уровня кода ОКВЭД
* регион (по ИНН)
"""
df_ = add_okved_subcode(add_region(add_title(df)))
df_ = adjust_rub(df_)
df_ = df_.set_index('inn')
df_ = rename_rows(df_)
return df_[canonic_columns()]
def canonic_columns(numeric=SHORT_COLUMNS.numeric):
return (['title', 'org', 'okpo', 'okopf', 'okfs', 'okved'] +
['unit'] +
['ok1', 'ok2', 'ok3', 'region'] +
numeric)
def is_numeric_column(name, numeric=SHORT_COLUMNS.numeric):
return name in numeric
def columns_typed_as_integer(numeric=SHORT_COLUMNS.numeric):
return numeric + ['ok1', 'ok2', 'ok3', 'region']
def canonic_dtypes():
def switch(col):
int_columns = columns_typed_as_integer()
return numpy.int64 if (col in int_columns) else str
result = {col: switch(col) for col in canonic_columns()}
result['inn'] = str
return result
| [
"boo.errors.UnclassifiableCodeError"
] | [((2447, 2483), 'boo.errors.UnclassifiableCodeError', 'UnclassifiableCodeError', (['code_string'], {}), '(code_string)\n', (2470, 2483), False, 'from boo.errors import UnclassifiableCodeError\n'), ((2587, 2623), 'boo.errors.UnclassifiableCodeError', 'UnclassifiableCodeError', (['code_string'], {}), '(code_string)\n', (2610, 2623), False, 'from boo.errors import UnclassifiableCodeError\n')] |
# 3rd party
import pytest
from coincidence.regressions import AdvancedFileRegressionFixture
from consolekit.testing import CliRunner, Result
from domdf_python_tools.paths import in_directory
from github3.exceptions import UnprocessableEntity
# this package
from repo_helper_github.cli import new
@pytest.mark.usefixtures("module_cassette")
def test_create_repo(github_manager, temp_github_repo):
with in_directory(temp_github_repo):
github_manager.new()
github_manager.github.repository("domdfcoding", "repo_helper_demo")
with pytest.raises(UnprocessableEntity, match="422 Repository creation failed."):
github_manager.new()
@pytest.mark.usefixtures("betamax_github_session", "module_cassette")
def test_via_cli(
temp_github_repo,
advanced_file_regression: AdvancedFileRegressionFixture,
github_manager,
):
with in_directory(temp_github_repo):
runner = CliRunner()
result: Result = runner.invoke(new)
assert result.exit_code == 0
result.check_stdout(advanced_file_regression, extension=".md")
# Check the repository now exists
github_manager.github.repository("domdfcoding", "repo_helper_demo")
with pytest.raises(UnprocessableEntity, match="422 Repository creation failed."):
github_manager.new()
| [
"domdf_python_tools.paths.in_directory",
"pytest.raises",
"pytest.mark.usefixtures",
"consolekit.testing.CliRunner"
] | [((300, 342), 'pytest.mark.usefixtures', 'pytest.mark.usefixtures', (['"""module_cassette"""'], {}), "('module_cassette')\n", (323, 342), False, 'import pytest\n'), ((643, 711), 'pytest.mark.usefixtures', 'pytest.mark.usefixtures', (['"""betamax_github_session"""', '"""module_cassette"""'], {}), "('betamax_github_session', 'module_cassette')\n", (666, 711), False, 'import pytest\n'), ((405, 435), 'domdf_python_tools.paths.in_directory', 'in_directory', (['temp_github_repo'], {}), '(temp_github_repo)\n', (417, 435), False, 'from domdf_python_tools.paths import in_directory\n'), ((838, 868), 'domdf_python_tools.paths.in_directory', 'in_directory', (['temp_github_repo'], {}), '(temp_github_repo)\n', (850, 868), False, 'from domdf_python_tools.paths import in_directory\n'), ((881, 892), 'consolekit.testing.CliRunner', 'CliRunner', ([], {}), '()\n', (890, 892), False, 'from consolekit.testing import CliRunner, Result\n'), ((1138, 1213), 'pytest.raises', 'pytest.raises', (['UnprocessableEntity'], {'match': '"""422 Repository creation failed."""'}), "(UnprocessableEntity, match='422 Repository creation failed.')\n", (1151, 1213), False, 'import pytest\n'), ((539, 614), 'pytest.raises', 'pytest.raises', (['UnprocessableEntity'], {'match': '"""422 Repository creation failed."""'}), "(UnprocessableEntity, match='422 Repository creation failed.')\n", (552, 614), False, 'import pytest\n')] |
import numpy as np
import os
import tensorflow as tf
EPS = 1e-8
def placeholder(dim=None):
return tf.placeholder(dtype=tf.float32, shape=(None,dim) if dim else (None,))
def placeholders(*args):
return [placeholder(dim) for dim in args]
def mlp(x, hidden_sizes=(32,), activation=tf.tanh, output_activation=None):
init_fn = tf.keras.initializers.Orthogonal(1.0)
for h in hidden_sizes[:-1]:
x = tf.layers.dense(x, units=h, activation=activation, kernel_initializer=init_fn)
return tf.layers.dense(x, units=hidden_sizes[-1], activation=output_activation, kernel_initializer=init_fn)
def get_vars(scope):
return [x for x in tf.global_variables() if scope in x.name]
def count_vars(scope):
v = get_vars(scope)
return sum([np.prod(var.shape.as_list()) for var in v])
def gaussian_likelihood(x, mu, log_std):
pre_sum = -0.5 * (((x-mu)/(tf.exp(log_std)+EPS))**2 + 2*log_std + np.log(2*np.pi))
return tf.reduce_sum(pre_sum, axis=1)
def clip_but_pass_gradient(x, l=-1., u=1.):
clip_up = tf.cast(x > u, tf.float32)
clip_low = tf.cast(x < l, tf.float32)
return x + tf.stop_gradient((u - x)*clip_up + (l - x)*clip_low)
"""
Policies
"""
def gumbel_policy(x, act_dim, hidden_sizes, activation):
# policy network outputs
net = mlp(x, list(hidden_sizes), activation, activation)
logits = tf.layers.dense(net, act_dim, activation='linear')
# action and log action probabilites (log_softmax covers numerical problems)
action_probs = tf.nn.softmax([logits], axis=-1)
log_action_probs = tf.nn.log_softmax([logits], axis=-1)
# policy with no noise
mu = tf.argmax(logits, axis=-1)
# add gumbel noise to action distributions
temperature = tf.convert_to_tensor(1.0) # 0 --> argmax, inf --> uniform
uniform_noise = tf.random_uniform(shape=tf.shape(logits),
minval=np.finfo(np.float32).tiny, # (0,1) range
maxval=1.)
gumbel_noise = -tf.log(-tf.log(uniform_noise))
noisy_logits = logits + gumbel_noise
pi_dist = tf.nn.softmax(noisy_logits / temperature[..., tf.newaxis])
# dont use tf.dist.relaxedCategorical for log_prob, seems to give wrong results
logp_pi = -tf.reduce_sum(-pi_dist * tf.nn.log_softmax(logits, axis=-1), axis=1)
return mu, pi_dist, logp_pi
"""
Actor-Critics
"""
def a_out_mlp_actor_critic(x, a, hidden_sizes=[400,300], activation=tf.nn.relu, policy=gumbel_policy):
act_dim = a.shape.as_list()[-1]
with tf.variable_scope('pi'):
mu, pi_dist, logp_pi = policy(x, act_dim, hidden_sizes, activation)
# vfs
with tf.variable_scope('q1'):
q1 = mlp(x, list(hidden_sizes)+[act_dim], activation, None)
q1_a = tf.reduce_sum(tf.multiply(q1, a), axis=1)
with tf.variable_scope('q2'):
q2 = mlp(x, list(hidden_sizes)+[act_dim], activation, None)
q2_a = tf.reduce_sum(tf.multiply(q2, a), axis=1)
return mu, pi_dist, logp_pi, q1_a, q2_a
def a_in_mlp_actor_critic(x, a, hidden_sizes=[400,300], activation=tf.nn.relu, policy=gumbel_policy):
act_dim = a.shape.as_list()[-1]
with tf.variable_scope('pi'):
mu, pi_dist, logp_pi = policy(x, act_dim, hidden_sizes, activation)
# vfs
with tf.variable_scope('q1'):
q1_a = tf.squeeze(mlp(tf.concat([x,a], axis=-1), list(hidden_sizes)+[1], activation, None), axis=1)
with tf.variable_scope('q2'):
q2_a = tf.squeeze(mlp(tf.concat([x,a], axis=-1), list(hidden_sizes)+[1], activation, None), axis=1)
return mu, pi_dist, logp_pi, q1_a, q2_a
| [
"tensorflow.shape",
"tensorflow.reduce_sum",
"numpy.log",
"tensorflow.multiply",
"tensorflow.nn.softmax",
"tensorflow.keras.initializers.Orthogonal",
"tensorflow.cast",
"tensorflow.log",
"tensorflow.placeholder",
"tensorflow.concat",
"tensorflow.convert_to_tensor",
"tensorflow.variable_scope",... | [((104, 175), 'tensorflow.placeholder', 'tf.placeholder', ([], {'dtype': 'tf.float32', 'shape': '((None, dim) if dim else (None,))'}), '(dtype=tf.float32, shape=(None, dim) if dim else (None,))\n', (118, 175), True, 'import tensorflow as tf\n'), ((338, 375), 'tensorflow.keras.initializers.Orthogonal', 'tf.keras.initializers.Orthogonal', (['(1.0)'], {}), '(1.0)\n', (370, 375), True, 'import tensorflow as tf\n'), ((510, 614), 'tensorflow.layers.dense', 'tf.layers.dense', (['x'], {'units': 'hidden_sizes[-1]', 'activation': 'output_activation', 'kernel_initializer': 'init_fn'}), '(x, units=hidden_sizes[-1], activation=output_activation,\n kernel_initializer=init_fn)\n', (525, 614), True, 'import tensorflow as tf\n'), ((946, 976), 'tensorflow.reduce_sum', 'tf.reduce_sum', (['pre_sum'], {'axis': '(1)'}), '(pre_sum, axis=1)\n', (959, 976), True, 'import tensorflow as tf\n'), ((1036, 1062), 'tensorflow.cast', 'tf.cast', (['(x > u)', 'tf.float32'], {}), '(x > u, tf.float32)\n', (1043, 1062), True, 'import tensorflow as tf\n'), ((1078, 1104), 'tensorflow.cast', 'tf.cast', (['(x < l)', 'tf.float32'], {}), '(x < l, tf.float32)\n', (1085, 1104), True, 'import tensorflow as tf\n'), ((1353, 1403), 'tensorflow.layers.dense', 'tf.layers.dense', (['net', 'act_dim'], {'activation': '"""linear"""'}), "(net, act_dim, activation='linear')\n", (1368, 1403), True, 'import tensorflow as tf\n'), ((1505, 1537), 'tensorflow.nn.softmax', 'tf.nn.softmax', (['[logits]'], {'axis': '(-1)'}), '([logits], axis=-1)\n', (1518, 1537), True, 'import tensorflow as tf\n'), ((1561, 1597), 'tensorflow.nn.log_softmax', 'tf.nn.log_softmax', (['[logits]'], {'axis': '(-1)'}), '([logits], axis=-1)\n', (1578, 1597), True, 'import tensorflow as tf\n'), ((1635, 1661), 'tensorflow.argmax', 'tf.argmax', (['logits'], {'axis': '(-1)'}), '(logits, axis=-1)\n', (1644, 1661), True, 'import tensorflow as tf\n'), ((1728, 1753), 'tensorflow.convert_to_tensor', 'tf.convert_to_tensor', (['(1.0)'], {}), '(1.0)\n', (1748, 1753), True, 'import tensorflow as tf\n'), ((2091, 2149), 'tensorflow.nn.softmax', 'tf.nn.softmax', (['(noisy_logits / temperature[..., tf.newaxis])'], {}), '(noisy_logits / temperature[..., tf.newaxis])\n', (2104, 2149), True, 'import tensorflow as tf\n'), ((420, 498), 'tensorflow.layers.dense', 'tf.layers.dense', (['x'], {'units': 'h', 'activation': 'activation', 'kernel_initializer': 'init_fn'}), '(x, units=h, activation=activation, kernel_initializer=init_fn)\n', (435, 498), True, 'import tensorflow as tf\n'), ((1120, 1176), 'tensorflow.stop_gradient', 'tf.stop_gradient', (['((u - x) * clip_up + (l - x) * clip_low)'], {}), '((u - x) * clip_up + (l - x) * clip_low)\n', (1136, 1176), True, 'import tensorflow as tf\n'), ((2525, 2548), 'tensorflow.variable_scope', 'tf.variable_scope', (['"""pi"""'], {}), "('pi')\n", (2542, 2548), True, 'import tensorflow as tf\n'), ((2646, 2669), 'tensorflow.variable_scope', 'tf.variable_scope', (['"""q1"""'], {}), "('q1')\n", (2663, 2669), True, 'import tensorflow as tf\n'), ((2809, 2832), 'tensorflow.variable_scope', 'tf.variable_scope', (['"""q2"""'], {}), "('q2')\n", (2826, 2832), True, 'import tensorflow as tf\n'), ((3159, 3182), 'tensorflow.variable_scope', 'tf.variable_scope', (['"""pi"""'], {}), "('pi')\n", (3176, 3182), True, 'import tensorflow as tf\n'), ((3280, 3303), 'tensorflow.variable_scope', 'tf.variable_scope', (['"""q1"""'], {}), "('q1')\n", (3297, 3303), True, 'import tensorflow as tf\n'), ((3424, 3447), 'tensorflow.variable_scope', 'tf.variable_scope', (['"""q2"""'], {}), "('q2')\n", (3441, 3447), True, 'import tensorflow as tf\n'), ((656, 677), 'tensorflow.global_variables', 'tf.global_variables', ([], {}), '()\n', (675, 677), True, 'import tensorflow as tf\n'), ((918, 935), 'numpy.log', 'np.log', (['(2 * np.pi)'], {}), '(2 * np.pi)\n', (924, 935), True, 'import numpy as np\n'), ((1830, 1846), 'tensorflow.shape', 'tf.shape', (['logits'], {}), '(logits)\n', (1838, 1846), True, 'import tensorflow as tf\n'), ((2772, 2790), 'tensorflow.multiply', 'tf.multiply', (['q1', 'a'], {}), '(q1, a)\n', (2783, 2790), True, 'import tensorflow as tf\n'), ((2935, 2953), 'tensorflow.multiply', 'tf.multiply', (['q2', 'a'], {}), '(q2, a)\n', (2946, 2953), True, 'import tensorflow as tf\n'), ((1893, 1913), 'numpy.finfo', 'np.finfo', (['np.float32'], {}), '(np.float32)\n', (1901, 1913), True, 'import numpy as np\n'), ((2012, 2033), 'tensorflow.log', 'tf.log', (['uniform_noise'], {}), '(uniform_noise)\n', (2018, 2033), True, 'import tensorflow as tf\n'), ((2275, 2309), 'tensorflow.nn.log_softmax', 'tf.nn.log_softmax', (['logits'], {'axis': '(-1)'}), '(logits, axis=-1)\n', (2292, 2309), True, 'import tensorflow as tf\n'), ((3336, 3362), 'tensorflow.concat', 'tf.concat', (['[x, a]'], {'axis': '(-1)'}), '([x, a], axis=-1)\n', (3345, 3362), True, 'import tensorflow as tf\n'), ((3480, 3506), 'tensorflow.concat', 'tf.concat', (['[x, a]'], {'axis': '(-1)'}), '([x, a], axis=-1)\n', (3489, 3506), True, 'import tensorflow as tf\n'), ((879, 894), 'tensorflow.exp', 'tf.exp', (['log_std'], {}), '(log_std)\n', (885, 894), True, 'import tensorflow as tf\n')] |
from objects.lighting.lighting_interactions import SwitchLightImmediateInteraction
from objects.object_state_utils import ObjectStateHelper
import sims4
logger = sims4.log.Logger('LightingAndObjectState', default_owner='mkartika')
class SwitchLightAndStateImmediateInteraction(SwitchLightImmediateInteraction):
INSTANCE_TUNABLES = {'state_settings': ObjectStateHelper.TunableFactory(description='\n Find objects in the same room or lot based on the tags and \n set state to the desired state.\n ')}
def _run_interaction_gen(self, timeline):
yield from super()._run_interaction_gen(timeline)
self.state_settings.execute_helper(self)
| [
"objects.object_state_utils.ObjectStateHelper.TunableFactory",
"sims4.log.Logger"
] | [((162, 230), 'sims4.log.Logger', 'sims4.log.Logger', (['"""LightingAndObjectState"""'], {'default_owner': '"""mkartika"""'}), "('LightingAndObjectState', default_owner='mkartika')\n", (178, 230), False, 'import sims4\n'), ((355, 546), 'objects.object_state_utils.ObjectStateHelper.TunableFactory', 'ObjectStateHelper.TunableFactory', ([], {'description': '"""\n Find objects in the same room or lot based on the tags and \n set state to the desired state.\n """'}), '(description=\n """\n Find objects in the same room or lot based on the tags and \n set state to the desired state.\n """\n )\n', (387, 546), False, 'from objects.object_state_utils import ObjectStateHelper\n')] |
import mock
def test_setup(GPIO, spidev):
from unicornhatmini import UnicornHATMini
unicornhatmini = UnicornHATMini()
spidev.SpiDev.assert_has_calls((
mock.call(0, 0),
mock.call(0, 1)
), any_order=True)
GPIO.setwarnings.assert_called_once_with(False)
GPIO.setmode.assert_called_once_with(GPIO.BCM)
del unicornhatmini
def test_shutdown(GPIO, spidev, atexit):
from unicornhatmini import UnicornHATMini
unicornhatmini = UnicornHATMini()
atexit.register.assert_called_once_with(unicornhatmini._exit)
unicornhatmini._exit()
| [
"mock.call",
"unicornhatmini.UnicornHATMini"
] | [((111, 127), 'unicornhatmini.UnicornHATMini', 'UnicornHATMini', ([], {}), '()\n', (125, 127), False, 'from unicornhatmini import UnicornHATMini\n'), ((476, 492), 'unicornhatmini.UnicornHATMini', 'UnicornHATMini', ([], {}), '()\n', (490, 492), False, 'from unicornhatmini import UnicornHATMini\n'), ((174, 189), 'mock.call', 'mock.call', (['(0)', '(0)'], {}), '(0, 0)\n', (183, 189), False, 'import mock\n'), ((199, 214), 'mock.call', 'mock.call', (['(0)', '(1)'], {}), '(0, 1)\n', (208, 214), False, 'import mock\n')] |
import time
from ..base import order as od
from .api import BybitApi
class BybitOrderManager(od.OrderManagerBase):
def __init__(self, api, ws=None, retention=60):
super().__init__(api, ws, retention)
self.ws.subscribe('execution', self.__on_events, True)
self.ws.subscribe('position', self.__on_events, True)
self.ws.subscribe('order', self.__on_events, True)
def _generate_order_object(self, e):
info = e.info
if e.type != od.EVENT_OPEN:
self.log.warning(f'event for unknown order: {e}')
return None
api = BybitApi.ccxt_instance()
symbol = api.markets_by_id[info['symbol']]['symbol']
return od.Order(
symbol, info['order_type'].lower(), info['side'].lower(),
info['qty'], float(info['price']))
def __on_events(self, msg):
topic = msg['topic']
for e in msg['data']:
oe = od.OrderEvent()
oe.info = e
oe.ts = time.time()
if topic == 'order':
oe.id = e['order_id']
st = e['order_status']
if st == 'New':
oe.type = od.EVENT_OPEN
elif st == 'Filled':
oe.type = od.EVENT_CLOSE
elif st in ['Cancelled', 'Rejected']:
oe.type = od.EVENT_CANCEL
else: # ignore(PartiallyFilled, Created, PendingCancel)
continue
elif topic == 'execution':
oe.type = od.EVENT_EXECUTION
oe.id = e['order_id']
oe.price = float(e['price'])
size = e['exec_qty']
oe.size = -size if e['side'] == 'Sell' else size
oe.fee = float(e['exec_fee']) * size
elif topic == 'position':
break
else:
assert False
self._handle_order_event(oe)
class BybitPositionGroup(od.PositionGroupBase):
INVERSE = True
class BybitOrderGroup(od.OrderGroupBase):
PositionGroup = BybitPositionGroup
class BybitOrderGroupManager(od.OrderGroupManagerBase):
OrderGroup = BybitOrderGroup
# Future
class BybitUsdtOrderManager(BybitOrderManager):
pass
class BybitUsdtPositionGroup(BybitPositionGroup):
INVERSE = False
class BybitUsdtOrderGroup(BybitOrderGroup):
PositionGroup = BybitUsdtPositionGroup
class BybitUsdtOrderGroupManager(BybitOrderGroupManager):
OrderGroup = BybitUsdtOrderGroup
| [
"time.time"
] | [((998, 1009), 'time.time', 'time.time', ([], {}), '()\n', (1007, 1009), False, 'import time\n')] |
import os
import wget
import paddle
from .tokenizer import Tokenizer
from .model import CLIP
from paddle.vision.transforms import Compose, Resize, CenterCrop, ToTensor, Normalize
tokenizer = Tokenizer()
def get_transforms(image_resolution):
transforms = Compose([
Resize(image_resolution, interpolation='bicubic'),
CenterCrop(image_resolution),
lambda image: image.convert("RGB"),
ToTensor(),
Normalize((0.48145466, 0.4578275, 0.40821073),
(0.26862954, 0.26130258, 0.27577711)),
])
return transforms
def clip_rn50():
model = CLIP(
embed_dim=1024,
image_resolution=224,
vision_layers=(3, 4, 6, 3),
vision_width=64,
vision_patch_size=None,
context_length=77,
vocab_size=49408,
transformer_width=512,
transformer_heads=8,
transformer_layers=12
)
return model, get_transforms(224)
def clip_rn101():
model = CLIP(
embed_dim=512,
image_resolution=224,
vision_layers=(3, 4, 23, 3),
vision_width=64,
vision_patch_size=None,
context_length=77,
vocab_size=49408,
transformer_width=512,
transformer_heads=8,
transformer_layers=12
)
return model, get_transforms(224)
def clip_rn50x4():
model = CLIP(
embed_dim=640,
image_resolution=288,
vision_layers=(4, 6, 10, 6),
vision_width=80,
vision_patch_size=None,
context_length=77,
vocab_size=49408,
transformer_width=640,
transformer_heads=10,
transformer_layers=12
)
return model, get_transforms(288)
def clip_vit_b_32():
model = CLIP(
embed_dim=512,
image_resolution=224,
vision_layers=12,
vision_width=768,
vision_patch_size=32,
context_length=77,
vocab_size=49408,
transformer_width=512,
transformer_heads=8,
transformer_layers=12
)
return model, get_transforms(224)
def tokenize(texts, context_length=77):
"""
Returns the tokenized representation of given input string(s)
Parameters
----------
texts : Union[str, List[str]]
An input string or a list of input strings to tokenize
context_length : int
The context length to use; all CLIP models use 77 as the context length
Returns
-------
A two-dimensional tensor containing the resulting tokens, shape = [number of input strings, context_length]
"""
if isinstance(texts, str):
texts = [texts]
sot_token = tokenizer.encoder["<|startoftext|>"]
eot_token = tokenizer.encoder["<|endoftext|>"]
all_tokens = [[sot_token] +
tokenizer.encode(text) + [eot_token] for text in texts]
result = paddle.zeros((len(all_tokens), context_length), dtype='int64')
for i, tokens in enumerate(all_tokens):
if len(tokens) > context_length:
raise RuntimeError(
f"Input {texts[i]} is too long for context length {context_length}")
result[i, :len(tokens)] = paddle.to_tensor(tokens)
return result
model_dict = {
'RN50': [clip_rn50, r'https://bj.bcebos.com/v1/ai-studio-online/6ffc89246e974a809e6e4b40fdb58063a112a0153e674dae8ed5b6dfe5d46d86?responseContentDisposition=attachment%3B%20filename%3DRN50.pdparams', 'RN50.pdparams'],
'RN50x4': [clip_rn50x4, r'https://bj.bcebos.com/v1/ai-studio-online/9f874e0174da48ffbd7c17e77b1fb278632620a9995e476ba873e334caec9037?responseContentDisposition=attachment%3B%20filename%3DRN50x4.pdparams', 'RN50x4.pdparams'],
'RN101': [clip_rn101, r'https://bj.bcebos.com/v1/ai-studio-online/484592d98c584785bc8f6f9f7badbf4a9fb7a96f6102470697ed974e8eeee2a9?responseContentDisposition=attachment%3B%20filename%3DRN101.pdparams', 'RN101.pdparams'],
'ViT_B_32': [clip_vit_b_32, r'https://bj.bcebos.com/v1/ai-studio-online/eb5e4dbf1ec142caa003a27cefd510ef46a8a6c3932a4d60bfecb3f3ab746c02?responseContentDisposition=attachment%3B%20filename%3DViT-B-32.pdparams', 'ViT-B-32.pdparams']
}
def load_model(model_name, pretrained=False):
model_fn, url, file_name = model_dict[model_name]
model, transforms = model_fn()
if pretrained:
model_path = os.path.join('pretrained_models', file_name)
if not os.path.isfile(model_path):
if not os.path.exists('pretrained_models'):
os.mkdir('pretrained_models')
wget.download(url, out=model_path)
params = paddle.load(model_path)
model.set_dict(params)
model.eval()
return model, transforms
| [
"paddle.vision.transforms.ToTensor",
"wget.download",
"os.path.exists",
"paddle.vision.transforms.Normalize",
"paddle.vision.transforms.CenterCrop",
"os.path.join",
"os.path.isfile",
"paddle.to_tensor",
"os.mkdir",
"paddle.load",
"paddle.vision.transforms.Resize"
] | [((3125, 3149), 'paddle.to_tensor', 'paddle.to_tensor', (['tokens'], {}), '(tokens)\n', (3141, 3149), False, 'import paddle\n'), ((4277, 4321), 'os.path.join', 'os.path.join', (['"""pretrained_models"""', 'file_name'], {}), "('pretrained_models', file_name)\n", (4289, 4321), False, 'import os\n'), ((4531, 4554), 'paddle.load', 'paddle.load', (['model_path'], {}), '(model_path)\n', (4542, 4554), False, 'import paddle\n'), ((280, 329), 'paddle.vision.transforms.Resize', 'Resize', (['image_resolution'], {'interpolation': '"""bicubic"""'}), "(image_resolution, interpolation='bicubic')\n", (286, 329), False, 'from paddle.vision.transforms import Compose, Resize, CenterCrop, ToTensor, Normalize\n'), ((339, 367), 'paddle.vision.transforms.CenterCrop', 'CenterCrop', (['image_resolution'], {}), '(image_resolution)\n', (349, 367), False, 'from paddle.vision.transforms import Compose, Resize, CenterCrop, ToTensor, Normalize\n'), ((421, 431), 'paddle.vision.transforms.ToTensor', 'ToTensor', ([], {}), '()\n', (429, 431), False, 'from paddle.vision.transforms import Compose, Resize, CenterCrop, ToTensor, Normalize\n'), ((441, 530), 'paddle.vision.transforms.Normalize', 'Normalize', (['(0.48145466, 0.4578275, 0.40821073)', '(0.26862954, 0.26130258, 0.27577711)'], {}), '((0.48145466, 0.4578275, 0.40821073), (0.26862954, 0.26130258, \n 0.27577711))\n', (450, 530), False, 'from paddle.vision.transforms import Compose, Resize, CenterCrop, ToTensor, Normalize\n'), ((4337, 4363), 'os.path.isfile', 'os.path.isfile', (['model_path'], {}), '(model_path)\n', (4351, 4363), False, 'import os\n'), ((4479, 4513), 'wget.download', 'wget.download', (['url'], {'out': 'model_path'}), '(url, out=model_path)\n', (4492, 4513), False, 'import wget\n'), ((4384, 4419), 'os.path.exists', 'os.path.exists', (['"""pretrained_models"""'], {}), "('pretrained_models')\n", (4398, 4419), False, 'import os\n'), ((4437, 4466), 'os.mkdir', 'os.mkdir', (['"""pretrained_models"""'], {}), "('pretrained_models')\n", (4445, 4466), False, 'import os\n')] |
# coding=utf-8
# *** WARNING: this file was generated by the Pulumi Terraform Bridge (tfgen) Tool. ***
# *** Do not edit by hand unless you're certain you know what you are doing! ***
import warnings
import pulumi
import pulumi.runtime
from typing import Any, Mapping, Optional, Sequence, Union, overload
from . import _utilities
__all__ = [
'AlbBackendGroupGrpcBackendArgs',
'AlbBackendGroupGrpcBackendHealthcheckArgs',
'AlbBackendGroupGrpcBackendHealthcheckGrpcHealthcheckArgs',
'AlbBackendGroupGrpcBackendHealthcheckHttpHealthcheckArgs',
'AlbBackendGroupGrpcBackendHealthcheckStreamHealthcheckArgs',
'AlbBackendGroupGrpcBackendLoadBalancingConfigArgs',
'AlbBackendGroupGrpcBackendTlsArgs',
'AlbBackendGroupGrpcBackendTlsValidationContextArgs',
'AlbBackendGroupHttpBackendArgs',
'AlbBackendGroupHttpBackendHealthcheckArgs',
'AlbBackendGroupHttpBackendHealthcheckGrpcHealthcheckArgs',
'AlbBackendGroupHttpBackendHealthcheckHttpHealthcheckArgs',
'AlbBackendGroupHttpBackendHealthcheckStreamHealthcheckArgs',
'AlbBackendGroupHttpBackendLoadBalancingConfigArgs',
'AlbBackendGroupHttpBackendTlsArgs',
'AlbBackendGroupHttpBackendTlsValidationContextArgs',
'AlbLoadBalancerAllocationPolicyArgs',
'AlbLoadBalancerAllocationPolicyLocationArgs',
'AlbLoadBalancerListenerArgs',
'AlbLoadBalancerListenerEndpointArgs',
'AlbLoadBalancerListenerEndpointAddressArgs',
'AlbLoadBalancerListenerEndpointAddressExternalIpv4AddressArgs',
'AlbLoadBalancerListenerEndpointAddressExternalIpv6AddressArgs',
'AlbLoadBalancerListenerEndpointAddressInternalIpv4AddressArgs',
'AlbLoadBalancerListenerHttpArgs',
'AlbLoadBalancerListenerHttpHandlerArgs',
'AlbLoadBalancerListenerHttpHandlerHttp2OptionsArgs',
'AlbLoadBalancerListenerHttpRedirectsArgs',
'AlbLoadBalancerListenerTlsArgs',
'AlbLoadBalancerListenerTlsDefaultHandlerArgs',
'AlbLoadBalancerListenerTlsDefaultHandlerHttpHandlerArgs',
'AlbLoadBalancerListenerTlsDefaultHandlerHttpHandlerHttp2OptionsArgs',
'AlbLoadBalancerListenerTlsSniHandlerArgs',
'AlbLoadBalancerListenerTlsSniHandlerHandlerArgs',
'AlbLoadBalancerListenerTlsSniHandlerHandlerHttpHandlerArgs',
'AlbLoadBalancerListenerTlsSniHandlerHandlerHttpHandlerHttp2OptionsArgs',
'AlbTargetGroupTargetArgs',
'AlbVirtualHostModifyRequestHeaderArgs',
'AlbVirtualHostModifyResponseHeaderArgs',
'AlbVirtualHostRouteArgs',
'AlbVirtualHostRouteGrpcRouteArgs',
'AlbVirtualHostRouteGrpcRouteGrpcMatchArgs',
'AlbVirtualHostRouteGrpcRouteGrpcMatchFqmnArgs',
'AlbVirtualHostRouteGrpcRouteGrpcRouteActionArgs',
'AlbVirtualHostRouteGrpcRouteGrpcStatusResponseActionArgs',
'AlbVirtualHostRouteHttpRouteArgs',
'AlbVirtualHostRouteHttpRouteDirectResponseActionArgs',
'AlbVirtualHostRouteHttpRouteHttpMatchArgs',
'AlbVirtualHostRouteHttpRouteHttpMatchPathArgs',
'AlbVirtualHostRouteHttpRouteHttpRouteActionArgs',
'AlbVirtualHostRouteHttpRouteRedirectActionArgs',
'ComputeDiskDiskPlacementPolicyArgs',
'ComputeInstanceBootDiskArgs',
'ComputeInstanceBootDiskInitializeParamsArgs',
'ComputeInstanceGroupAllocationPolicyArgs',
'ComputeInstanceGroupApplicationLoadBalancerArgs',
'ComputeInstanceGroupDeployPolicyArgs',
'ComputeInstanceGroupHealthCheckArgs',
'ComputeInstanceGroupHealthCheckHttpOptionArgs',
'ComputeInstanceGroupHealthCheckTcpOptionArgs',
'ComputeInstanceGroupInstanceArgs',
'ComputeInstanceGroupInstanceNetworkInterfaceArgs',
'ComputeInstanceGroupInstanceTemplateArgs',
'ComputeInstanceGroupInstanceTemplateBootDiskArgs',
'ComputeInstanceGroupInstanceTemplateBootDiskInitializeParamsArgs',
'ComputeInstanceGroupInstanceTemplateNetworkInterfaceArgs',
'ComputeInstanceGroupInstanceTemplateNetworkInterfaceDnsRecordArgs',
'ComputeInstanceGroupInstanceTemplateNetworkInterfaceIpv6DnsRecordArgs',
'ComputeInstanceGroupInstanceTemplateNetworkInterfaceNatDnsRecordArgs',
'ComputeInstanceGroupInstanceTemplateNetworkSettingArgs',
'ComputeInstanceGroupInstanceTemplatePlacementPolicyArgs',
'ComputeInstanceGroupInstanceTemplateResourcesArgs',
'ComputeInstanceGroupInstanceTemplateSchedulingPolicyArgs',
'ComputeInstanceGroupInstanceTemplateSecondaryDiskArgs',
'ComputeInstanceGroupInstanceTemplateSecondaryDiskInitializeParamsArgs',
'ComputeInstanceGroupLoadBalancerArgs',
'ComputeInstanceGroupScalePolicyArgs',
'ComputeInstanceGroupScalePolicyAutoScaleArgs',
'ComputeInstanceGroupScalePolicyAutoScaleCustomRuleArgs',
'ComputeInstanceGroupScalePolicyFixedScaleArgs',
'ComputeInstanceGroupScalePolicyTestAutoScaleArgs',
'ComputeInstanceGroupScalePolicyTestAutoScaleCustomRuleArgs',
'ComputeInstanceNetworkInterfaceArgs',
'ComputeInstanceNetworkInterfaceDnsRecordArgs',
'ComputeInstanceNetworkInterfaceIpv6DnsRecordArgs',
'ComputeInstanceNetworkInterfaceNatDnsRecordArgs',
'ComputeInstancePlacementPolicyArgs',
'ComputeInstanceResourcesArgs',
'ComputeInstanceSchedulingPolicyArgs',
'ComputeInstanceSecondaryDiskArgs',
'DataprocClusterClusterConfigArgs',
'DataprocClusterClusterConfigHadoopArgs',
'DataprocClusterClusterConfigSubclusterSpecArgs',
'DataprocClusterClusterConfigSubclusterSpecAutoscalingConfigArgs',
'DataprocClusterClusterConfigSubclusterSpecResourcesArgs',
'FunctionContentArgs',
'FunctionPackageArgs',
'FunctionScalingPolicyPolicyArgs',
'FunctionTriggerDlqArgs',
'FunctionTriggerFunctionArgs',
'FunctionTriggerIotArgs',
'FunctionTriggerLogGroupArgs',
'FunctionTriggerLoggingArgs',
'FunctionTriggerMessageQueueArgs',
'FunctionTriggerObjectStorageArgs',
'FunctionTriggerTimerArgs',
'KubernetesClusterKmsProviderArgs',
'KubernetesClusterMasterArgs',
'KubernetesClusterMasterMaintenancePolicyArgs',
'KubernetesClusterMasterMaintenancePolicyMaintenanceWindowArgs',
'KubernetesClusterMasterRegionalArgs',
'KubernetesClusterMasterRegionalLocationArgs',
'KubernetesClusterMasterVersionInfoArgs',
'KubernetesClusterMasterZonalArgs',
'KubernetesClusterNetworkImplementationArgs',
'KubernetesClusterNetworkImplementationCiliumArgs',
'KubernetesNodeGroupAllocationPolicyArgs',
'KubernetesNodeGroupAllocationPolicyLocationArgs',
'KubernetesNodeGroupDeployPolicyArgs',
'KubernetesNodeGroupInstanceTemplateArgs',
'KubernetesNodeGroupInstanceTemplateBootDiskArgs',
'KubernetesNodeGroupInstanceTemplateNetworkInterfaceArgs',
'KubernetesNodeGroupInstanceTemplatePlacementPolicyArgs',
'KubernetesNodeGroupInstanceTemplateResourcesArgs',
'KubernetesNodeGroupInstanceTemplateSchedulingPolicyArgs',
'KubernetesNodeGroupMaintenancePolicyArgs',
'KubernetesNodeGroupMaintenancePolicyMaintenanceWindowArgs',
'KubernetesNodeGroupScalePolicyArgs',
'KubernetesNodeGroupScalePolicyAutoScaleArgs',
'KubernetesNodeGroupScalePolicyFixedScaleArgs',
'KubernetesNodeGroupVersionInfoArgs',
'LbNetworkLoadBalancerAttachedTargetGroupArgs',
'LbNetworkLoadBalancerAttachedTargetGroupHealthcheckArgs',
'LbNetworkLoadBalancerAttachedTargetGroupHealthcheckHttpOptionsArgs',
'LbNetworkLoadBalancerAttachedTargetGroupHealthcheckTcpOptionsArgs',
'LbNetworkLoadBalancerListenerArgs',
'LbNetworkLoadBalancerListenerExternalAddressSpecArgs',
'LbNetworkLoadBalancerListenerInternalAddressSpecArgs',
'LbTargetGroupTargetArgs',
'MdbClickhouseClusterAccessArgs',
'MdbClickhouseClusterBackupWindowStartArgs',
'MdbClickhouseClusterClickhouseArgs',
'MdbClickhouseClusterClickhouseConfigArgs',
'MdbClickhouseClusterClickhouseConfigCompressionArgs',
'MdbClickhouseClusterClickhouseConfigGraphiteRollupArgs',
'MdbClickhouseClusterClickhouseConfigGraphiteRollupPatternArgs',
'MdbClickhouseClusterClickhouseConfigGraphiteRollupPatternRetentionArgs',
'MdbClickhouseClusterClickhouseConfigKafkaArgs',
'MdbClickhouseClusterClickhouseConfigKafkaTopicArgs',
'MdbClickhouseClusterClickhouseConfigKafkaTopicSettingsArgs',
'MdbClickhouseClusterClickhouseConfigMergeTreeArgs',
'MdbClickhouseClusterClickhouseConfigRabbitmqArgs',
'MdbClickhouseClusterClickhouseResourcesArgs',
'MdbClickhouseClusterCloudStorageArgs',
'MdbClickhouseClusterDatabaseArgs',
'MdbClickhouseClusterFormatSchemaArgs',
'MdbClickhouseClusterHostArgs',
'MdbClickhouseClusterMaintenanceWindowArgs',
'MdbClickhouseClusterMlModelArgs',
'MdbClickhouseClusterShardGroupArgs',
'MdbClickhouseClusterUserArgs',
'MdbClickhouseClusterUserPermissionArgs',
'MdbClickhouseClusterUserQuotaArgs',
'MdbClickhouseClusterUserSettingsArgs',
'MdbClickhouseClusterZookeeperArgs',
'MdbClickhouseClusterZookeeperResourcesArgs',
'MdbElasticSearchClusterConfigArgs',
'MdbElasticSearchClusterConfigDataNodeArgs',
'MdbElasticSearchClusterConfigDataNodeResourcesArgs',
'MdbElasticSearchClusterConfigMasterNodeArgs',
'MdbElasticSearchClusterConfigMasterNodeResourcesArgs',
'MdbElasticSearchClusterHostArgs',
'MdbGreenplumClusterAccessArgs',
'MdbGreenplumClusterBackupWindowStartArgs',
'MdbGreenplumClusterMasterHostArgs',
'MdbGreenplumClusterMasterSubclusterArgs',
'MdbGreenplumClusterMasterSubclusterResourcesArgs',
'MdbGreenplumClusterSegmentHostArgs',
'MdbGreenplumClusterSegmentSubclusterArgs',
'MdbGreenplumClusterSegmentSubclusterResourcesArgs',
'MdbKafkaClusterConfigArgs',
'MdbKafkaClusterConfigKafkaArgs',
'MdbKafkaClusterConfigKafkaKafkaConfigArgs',
'MdbKafkaClusterConfigKafkaResourcesArgs',
'MdbKafkaClusterConfigZookeeperArgs',
'MdbKafkaClusterConfigZookeeperResourcesArgs',
'MdbKafkaClusterHostArgs',
'MdbKafkaClusterTopicArgs',
'MdbKafkaClusterTopicTopicConfigArgs',
'MdbKafkaClusterUserArgs',
'MdbKafkaClusterUserPermissionArgs',
'MdbKafkaTopicTopicConfigArgs',
'MdbMongodbClusterClusterConfigArgs',
'MdbMongodbClusterClusterConfigAccessArgs',
'MdbMongodbClusterClusterConfigBackupWindowStartArgs',
'MdbMongodbClusterDatabaseArgs',
'MdbMongodbClusterHostArgs',
'MdbMongodbClusterMaintenanceWindowArgs',
'MdbMongodbClusterResourcesArgs',
'MdbMongodbClusterUserArgs',
'MdbMongodbClusterUserPermissionArgs',
'MdbMysqlClusterAccessArgs',
'MdbMysqlClusterBackupWindowStartArgs',
'MdbMysqlClusterDatabaseArgs',
'MdbMysqlClusterHostArgs',
'MdbMysqlClusterMaintenanceWindowArgs',
'MdbMysqlClusterResourcesArgs',
'MdbMysqlClusterRestoreArgs',
'MdbMysqlClusterUserArgs',
'MdbMysqlClusterUserConnectionLimitsArgs',
'MdbMysqlClusterUserPermissionArgs',
'MdbRedisClusterConfigArgs',
'MdbRedisClusterHostArgs',
'MdbRedisClusterMaintenanceWindowArgs',
'MdbRedisClusterResourcesArgs',
'MdbSqlServerClusterBackupWindowStartArgs',
'MdbSqlServerClusterDatabaseArgs',
'MdbSqlServerClusterHostArgs',
'MdbSqlServerClusterResourcesArgs',
'MdbSqlServerClusterUserArgs',
'MdbSqlServerClusterUserPermissionArgs',
'StorageBucketCorsRuleArgs',
'StorageBucketGrantArgs',
'StorageBucketLifecycleRuleArgs',
'StorageBucketLifecycleRuleExpirationArgs',
'StorageBucketLifecycleRuleNoncurrentVersionExpirationArgs',
'StorageBucketLifecycleRuleNoncurrentVersionTransitionArgs',
'StorageBucketLifecycleRuleTransitionArgs',
'StorageBucketLoggingArgs',
'StorageBucketServerSideEncryptionConfigurationArgs',
'StorageBucketServerSideEncryptionConfigurationRuleArgs',
'StorageBucketServerSideEncryptionConfigurationRuleApplyServerSideEncryptionByDefaultArgs',
'StorageBucketVersioningArgs',
'StorageBucketWebsiteArgs',
'VpcAddressExternalIpv4AddressArgs',
'VpcDefaultSecurityGroupEgressArgs',
'VpcDefaultSecurityGroupIngressArgs',
'VpcRouteTableStaticRouteArgs',
'VpcSecurityGroupEgressArgs',
'VpcSecurityGroupIngressArgs',
'VpcSubnetDhcpOptionsArgs',
'YdbDatabaseDedicatedLocationArgs',
'YdbDatabaseDedicatedLocationRegionArgs',
'YdbDatabaseDedicatedScalePolicyArgs',
'YdbDatabaseDedicatedScalePolicyFixedScaleArgs',
'YdbDatabaseDedicatedStorageConfigArgs',
'GetAlbBackendGroupGrpcBackendArgs',
'GetAlbBackendGroupGrpcBackendHealthcheckArgs',
'GetAlbBackendGroupGrpcBackendHealthcheckGrpcHealthcheckArgs',
'GetAlbBackendGroupGrpcBackendHealthcheckHttpHealthcheckArgs',
'GetAlbBackendGroupGrpcBackendHealthcheckStreamHealthcheckArgs',
'GetAlbBackendGroupGrpcBackendLoadBalancingConfigArgs',
'GetAlbBackendGroupGrpcBackendTlsArgs',
'GetAlbBackendGroupGrpcBackendTlsValidationContextArgs',
'GetAlbBackendGroupHttpBackendArgs',
'GetAlbBackendGroupHttpBackendHealthcheckArgs',
'GetAlbBackendGroupHttpBackendHealthcheckGrpcHealthcheckArgs',
'GetAlbBackendGroupHttpBackendHealthcheckHttpHealthcheckArgs',
'GetAlbBackendGroupHttpBackendHealthcheckStreamHealthcheckArgs',
'GetAlbBackendGroupHttpBackendLoadBalancingConfigArgs',
'GetAlbBackendGroupHttpBackendTlsArgs',
'GetAlbBackendGroupHttpBackendTlsValidationContextArgs',
'GetComputeDiskDiskPlacementPolicyArgs',
'GetComputeInstancePlacementPolicyArgs',
'GetComputeInstanceSchedulingPolicyArgs',
'GetFunctionScalingPolicyPolicyArgs',
'GetIamPolicyBindingArgs',
'GetMdbClickhouseClusterCloudStorageArgs',
'GetMdbKafkaClusterConfigArgs',
'GetMdbKafkaClusterConfigKafkaArgs',
'GetMdbKafkaClusterConfigKafkaKafkaConfigArgs',
'GetMdbKafkaClusterConfigKafkaResourcesArgs',
'GetMdbKafkaClusterConfigZookeeperArgs',
'GetMdbKafkaClusterConfigZookeeperResourcesArgs',
'GetMdbKafkaClusterTopicArgs',
'GetMdbKafkaClusterTopicTopicConfigArgs',
'GetMdbKafkaClusterUserArgs',
'GetMdbKafkaClusterUserPermissionArgs',
'GetMdbMysqlClusterAccessArgs',
]
@pulumi.input_type
class AlbBackendGroupGrpcBackendArgs:
def __init__(__self__, *,
name: pulumi.Input[str],
target_group_ids: pulumi.Input[Sequence[pulumi.Input[str]]],
healthcheck: Optional[pulumi.Input['AlbBackendGroupGrpcBackendHealthcheckArgs']] = None,
load_balancing_config: Optional[pulumi.Input['AlbBackendGroupGrpcBackendLoadBalancingConfigArgs']] = None,
port: Optional[pulumi.Input[int]] = None,
tls: Optional[pulumi.Input['AlbBackendGroupGrpcBackendTlsArgs']] = None,
weight: Optional[pulumi.Input[int]] = None):
"""
:param pulumi.Input[str] name: Name of the backend.
:param pulumi.Input[Sequence[pulumi.Input[str]]] target_group_ids: References target groups for the backend.
:param pulumi.Input['AlbBackendGroupGrpcBackendHealthcheckArgs'] healthcheck: Healthcheck specification that will be used by this backend. Structure is documented below.
:param pulumi.Input['AlbBackendGroupGrpcBackendLoadBalancingConfigArgs'] load_balancing_config: Load Balancing Config specification that will be used by this backend. Structure is documented below.
:param pulumi.Input[int] port: Port for incoming traffic.
:param pulumi.Input['AlbBackendGroupGrpcBackendTlsArgs'] tls: Tls specification that will be used by this backend. Structure is documented below.
:param pulumi.Input[int] weight: Weight of the backend. Traffic will be split between backends of the same BackendGroup according to their weights.
"""
pulumi.set(__self__, "name", name)
pulumi.set(__self__, "target_group_ids", target_group_ids)
if healthcheck is not None:
pulumi.set(__self__, "healthcheck", healthcheck)
if load_balancing_config is not None:
pulumi.set(__self__, "load_balancing_config", load_balancing_config)
if port is not None:
pulumi.set(__self__, "port", port)
if tls is not None:
pulumi.set(__self__, "tls", tls)
if weight is not None:
pulumi.set(__self__, "weight", weight)
@property
@pulumi.getter
def name(self) -> pulumi.Input[str]:
"""
Name of the backend.
"""
return pulumi.get(self, "name")
@name.setter
def name(self, value: pulumi.Input[str]):
pulumi.set(self, "name", value)
@property
@pulumi.getter(name="targetGroupIds")
def target_group_ids(self) -> pulumi.Input[Sequence[pulumi.Input[str]]]:
"""
References target groups for the backend.
"""
return pulumi.get(self, "target_group_ids")
@target_group_ids.setter
def target_group_ids(self, value: pulumi.Input[Sequence[pulumi.Input[str]]]):
pulumi.set(self, "target_group_ids", value)
@property
@pulumi.getter
def healthcheck(self) -> Optional[pulumi.Input['AlbBackendGroupGrpcBackendHealthcheckArgs']]:
"""
Healthcheck specification that will be used by this backend. Structure is documented below.
"""
return pulumi.get(self, "healthcheck")
@healthcheck.setter
def healthcheck(self, value: Optional[pulumi.Input['AlbBackendGroupGrpcBackendHealthcheckArgs']]):
pulumi.set(self, "healthcheck", value)
@property
@pulumi.getter(name="loadBalancingConfig")
def load_balancing_config(self) -> Optional[pulumi.Input['AlbBackendGroupGrpcBackendLoadBalancingConfigArgs']]:
"""
Load Balancing Config specification that will be used by this backend. Structure is documented below.
"""
return pulumi.get(self, "load_balancing_config")
@load_balancing_config.setter
def load_balancing_config(self, value: Optional[pulumi.Input['AlbBackendGroupGrpcBackendLoadBalancingConfigArgs']]):
pulumi.set(self, "load_balancing_config", value)
@property
@pulumi.getter
def port(self) -> Optional[pulumi.Input[int]]:
"""
Port for incoming traffic.
"""
return pulumi.get(self, "port")
@port.setter
def port(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "port", value)
@property
@pulumi.getter
def tls(self) -> Optional[pulumi.Input['AlbBackendGroupGrpcBackendTlsArgs']]:
"""
Tls specification that will be used by this backend. Structure is documented below.
"""
return pulumi.get(self, "tls")
@tls.setter
def tls(self, value: Optional[pulumi.Input['AlbBackendGroupGrpcBackendTlsArgs']]):
pulumi.set(self, "tls", value)
@property
@pulumi.getter
def weight(self) -> Optional[pulumi.Input[int]]:
"""
Weight of the backend. Traffic will be split between backends of the same BackendGroup according to their weights.
"""
return pulumi.get(self, "weight")
@weight.setter
def weight(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "weight", value)
@pulumi.input_type
class AlbBackendGroupGrpcBackendHealthcheckArgs:
def __init__(__self__, *,
interval: pulumi.Input[str],
timeout: pulumi.Input[str],
grpc_healthcheck: Optional[pulumi.Input['AlbBackendGroupGrpcBackendHealthcheckGrpcHealthcheckArgs']] = None,
healthcheck_port: Optional[pulumi.Input[int]] = None,
healthy_threshold: Optional[pulumi.Input[int]] = None,
http_healthcheck: Optional[pulumi.Input['AlbBackendGroupGrpcBackendHealthcheckHttpHealthcheckArgs']] = None,
interval_jitter_percent: Optional[pulumi.Input[float]] = None,
stream_healthcheck: Optional[pulumi.Input['AlbBackendGroupGrpcBackendHealthcheckStreamHealthcheckArgs']] = None,
unhealthy_threshold: Optional[pulumi.Input[int]] = None):
"""
:param pulumi.Input[str] interval: Interval between health checks.
:param pulumi.Input[str] timeout: Time to wait for a health check response.
:param pulumi.Input['AlbBackendGroupGrpcBackendHealthcheckGrpcHealthcheckArgs'] grpc_healthcheck: Grpc Healthcheck specification that will be used by this healthcheck. Structure is documented below.
:param pulumi.Input[int] healthcheck_port: Optional alternative port for health checking.
:param pulumi.Input[int] healthy_threshold: Number of consecutive successful health checks required to promote endpoint into the healthy state. 0 means 1. Note that during startup, only a single successful health check is required to mark a host healthy.
:param pulumi.Input['AlbBackendGroupGrpcBackendHealthcheckHttpHealthcheckArgs'] http_healthcheck: Http Healthcheck specification that will be used by this healthcheck. Structure is documented below.
:param pulumi.Input[float] interval_jitter_percent: An optional jitter amount as a percentage of interval. If specified, during every interval value of (interval_ms * interval_jitter_percent / 100) will be added to the wait time.
:param pulumi.Input['AlbBackendGroupGrpcBackendHealthcheckStreamHealthcheckArgs'] stream_healthcheck: Stream Healthcheck specification that will be used by this healthcheck. Structure is documented below.
:param pulumi.Input[int] unhealthy_threshold: Number of consecutive failed health checks required to demote endpoint into the unhealthy state. 0 means 1. Note that for HTTP health checks, a single 503 immediately makes endpoint unhealthy.
"""
pulumi.set(__self__, "interval", interval)
pulumi.set(__self__, "timeout", timeout)
if grpc_healthcheck is not None:
pulumi.set(__self__, "grpc_healthcheck", grpc_healthcheck)
if healthcheck_port is not None:
pulumi.set(__self__, "healthcheck_port", healthcheck_port)
if healthy_threshold is not None:
pulumi.set(__self__, "healthy_threshold", healthy_threshold)
if http_healthcheck is not None:
pulumi.set(__self__, "http_healthcheck", http_healthcheck)
if interval_jitter_percent is not None:
pulumi.set(__self__, "interval_jitter_percent", interval_jitter_percent)
if stream_healthcheck is not None:
pulumi.set(__self__, "stream_healthcheck", stream_healthcheck)
if unhealthy_threshold is not None:
pulumi.set(__self__, "unhealthy_threshold", unhealthy_threshold)
@property
@pulumi.getter
def interval(self) -> pulumi.Input[str]:
"""
Interval between health checks.
"""
return pulumi.get(self, "interval")
@interval.setter
def interval(self, value: pulumi.Input[str]):
pulumi.set(self, "interval", value)
@property
@pulumi.getter
def timeout(self) -> pulumi.Input[str]:
"""
Time to wait for a health check response.
"""
return pulumi.get(self, "timeout")
@timeout.setter
def timeout(self, value: pulumi.Input[str]):
pulumi.set(self, "timeout", value)
@property
@pulumi.getter(name="grpcHealthcheck")
def grpc_healthcheck(self) -> Optional[pulumi.Input['AlbBackendGroupGrpcBackendHealthcheckGrpcHealthcheckArgs']]:
"""
Grpc Healthcheck specification that will be used by this healthcheck. Structure is documented below.
"""
return pulumi.get(self, "grpc_healthcheck")
@grpc_healthcheck.setter
def grpc_healthcheck(self, value: Optional[pulumi.Input['AlbBackendGroupGrpcBackendHealthcheckGrpcHealthcheckArgs']]):
pulumi.set(self, "grpc_healthcheck", value)
@property
@pulumi.getter(name="healthcheckPort")
def healthcheck_port(self) -> Optional[pulumi.Input[int]]:
"""
Optional alternative port for health checking.
"""
return pulumi.get(self, "healthcheck_port")
@healthcheck_port.setter
def healthcheck_port(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "healthcheck_port", value)
@property
@pulumi.getter(name="healthyThreshold")
def healthy_threshold(self) -> Optional[pulumi.Input[int]]:
"""
Number of consecutive successful health checks required to promote endpoint into the healthy state. 0 means 1. Note that during startup, only a single successful health check is required to mark a host healthy.
"""
return pulumi.get(self, "healthy_threshold")
@healthy_threshold.setter
def healthy_threshold(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "healthy_threshold", value)
@property
@pulumi.getter(name="httpHealthcheck")
def http_healthcheck(self) -> Optional[pulumi.Input['AlbBackendGroupGrpcBackendHealthcheckHttpHealthcheckArgs']]:
"""
Http Healthcheck specification that will be used by this healthcheck. Structure is documented below.
"""
return pulumi.get(self, "http_healthcheck")
@http_healthcheck.setter
def http_healthcheck(self, value: Optional[pulumi.Input['AlbBackendGroupGrpcBackendHealthcheckHttpHealthcheckArgs']]):
pulumi.set(self, "http_healthcheck", value)
@property
@pulumi.getter(name="intervalJitterPercent")
def interval_jitter_percent(self) -> Optional[pulumi.Input[float]]:
"""
An optional jitter amount as a percentage of interval. If specified, during every interval value of (interval_ms * interval_jitter_percent / 100) will be added to the wait time.
"""
return pulumi.get(self, "interval_jitter_percent")
@interval_jitter_percent.setter
def interval_jitter_percent(self, value: Optional[pulumi.Input[float]]):
pulumi.set(self, "interval_jitter_percent", value)
@property
@pulumi.getter(name="streamHealthcheck")
def stream_healthcheck(self) -> Optional[pulumi.Input['AlbBackendGroupGrpcBackendHealthcheckStreamHealthcheckArgs']]:
"""
Stream Healthcheck specification that will be used by this healthcheck. Structure is documented below.
"""
return pulumi.get(self, "stream_healthcheck")
@stream_healthcheck.setter
def stream_healthcheck(self, value: Optional[pulumi.Input['AlbBackendGroupGrpcBackendHealthcheckStreamHealthcheckArgs']]):
pulumi.set(self, "stream_healthcheck", value)
@property
@pulumi.getter(name="unhealthyThreshold")
def unhealthy_threshold(self) -> Optional[pulumi.Input[int]]:
"""
Number of consecutive failed health checks required to demote endpoint into the unhealthy state. 0 means 1. Note that for HTTP health checks, a single 503 immediately makes endpoint unhealthy.
"""
return pulumi.get(self, "unhealthy_threshold")
@unhealthy_threshold.setter
def unhealthy_threshold(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "unhealthy_threshold", value)
@pulumi.input_type
class AlbBackendGroupGrpcBackendHealthcheckGrpcHealthcheckArgs:
def __init__(__self__, *,
service_name: Optional[pulumi.Input[str]] = None):
"""
:param pulumi.Input[str] service_name: Service name for grpc.health.v1.HealthCheckRequest message.
"""
if service_name is not None:
pulumi.set(__self__, "service_name", service_name)
@property
@pulumi.getter(name="serviceName")
def service_name(self) -> Optional[pulumi.Input[str]]:
"""
Service name for grpc.health.v1.HealthCheckRequest message.
"""
return pulumi.get(self, "service_name")
@service_name.setter
def service_name(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "service_name", value)
@pulumi.input_type
class AlbBackendGroupGrpcBackendHealthcheckHttpHealthcheckArgs:
def __init__(__self__, *,
path: pulumi.Input[str],
host: Optional[pulumi.Input[str]] = None,
http2: Optional[pulumi.Input[bool]] = None):
"""
:param pulumi.Input[str] path: HTTP path.
:param pulumi.Input[str] host: "Host" HTTP header value.
:param pulumi.Input[bool] http2: If set, health checks will use HTTP2.
"""
pulumi.set(__self__, "path", path)
if host is not None:
pulumi.set(__self__, "host", host)
if http2 is not None:
pulumi.set(__self__, "http2", http2)
@property
@pulumi.getter
def path(self) -> pulumi.Input[str]:
"""
HTTP path.
"""
return pulumi.get(self, "path")
@path.setter
def path(self, value: pulumi.Input[str]):
pulumi.set(self, "path", value)
@property
@pulumi.getter
def host(self) -> Optional[pulumi.Input[str]]:
"""
"Host" HTTP header value.
"""
return pulumi.get(self, "host")
@host.setter
def host(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "host", value)
@property
@pulumi.getter
def http2(self) -> Optional[pulumi.Input[bool]]:
"""
If set, health checks will use HTTP2.
"""
return pulumi.get(self, "http2")
@http2.setter
def http2(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "http2", value)
@pulumi.input_type
class AlbBackendGroupGrpcBackendHealthcheckStreamHealthcheckArgs:
def __init__(__self__, *,
receive: Optional[pulumi.Input[str]] = None,
send: Optional[pulumi.Input[str]] = None):
"""
:param pulumi.Input[str] receive: Text to search in reply.
:param pulumi.Input[str] send: Message to send. If empty, it's a connect-only health check.
"""
if receive is not None:
pulumi.set(__self__, "receive", receive)
if send is not None:
pulumi.set(__self__, "send", send)
@property
@pulumi.getter
def receive(self) -> Optional[pulumi.Input[str]]:
"""
Text to search in reply.
"""
return pulumi.get(self, "receive")
@receive.setter
def receive(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "receive", value)
@property
@pulumi.getter
def send(self) -> Optional[pulumi.Input[str]]:
"""
Message to send. If empty, it's a connect-only health check.
"""
return pulumi.get(self, "send")
@send.setter
def send(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "send", value)
@pulumi.input_type
class AlbBackendGroupGrpcBackendLoadBalancingConfigArgs:
def __init__(__self__, *,
locality_aware_routing_percent: Optional[pulumi.Input[int]] = None,
panic_threshold: Optional[pulumi.Input[int]] = None,
strict_locality: Optional[pulumi.Input[bool]] = None):
"""
:param pulumi.Input[int] locality_aware_routing_percent: Percent of traffic to be sent to the same availability zone. The rest will be equally divided between other zones.
:param pulumi.Input[int] panic_threshold: If percentage of healthy hosts in the backend is lower than panic_threshold, traffic will be routed to all backends no matter what the health status is. This helps to avoid healthy backends overloading when everything is bad. Zero means no panic threshold.
:param pulumi.Input[bool] strict_locality: If set, will route requests only to the same availability zone. Balancer won't know about endpoints in other zones.
"""
if locality_aware_routing_percent is not None:
pulumi.set(__self__, "locality_aware_routing_percent", locality_aware_routing_percent)
if panic_threshold is not None:
pulumi.set(__self__, "panic_threshold", panic_threshold)
if strict_locality is not None:
pulumi.set(__self__, "strict_locality", strict_locality)
@property
@pulumi.getter(name="localityAwareRoutingPercent")
def locality_aware_routing_percent(self) -> Optional[pulumi.Input[int]]:
"""
Percent of traffic to be sent to the same availability zone. The rest will be equally divided between other zones.
"""
return pulumi.get(self, "locality_aware_routing_percent")
@locality_aware_routing_percent.setter
def locality_aware_routing_percent(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "locality_aware_routing_percent", value)
@property
@pulumi.getter(name="panicThreshold")
def panic_threshold(self) -> Optional[pulumi.Input[int]]:
"""
If percentage of healthy hosts in the backend is lower than panic_threshold, traffic will be routed to all backends no matter what the health status is. This helps to avoid healthy backends overloading when everything is bad. Zero means no panic threshold.
"""
return pulumi.get(self, "panic_threshold")
@panic_threshold.setter
def panic_threshold(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "panic_threshold", value)
@property
@pulumi.getter(name="strictLocality")
def strict_locality(self) -> Optional[pulumi.Input[bool]]:
"""
If set, will route requests only to the same availability zone. Balancer won't know about endpoints in other zones.
"""
return pulumi.get(self, "strict_locality")
@strict_locality.setter
def strict_locality(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "strict_locality", value)
@pulumi.input_type
class AlbBackendGroupGrpcBackendTlsArgs:
def __init__(__self__, *,
sni: Optional[pulumi.Input[str]] = None,
validation_context: Optional[pulumi.Input['AlbBackendGroupGrpcBackendTlsValidationContextArgs']] = None):
"""
:param pulumi.Input[str] sni: [SNI](https://en.wikipedia.org/wiki/Server_Name_Indication) string for TLS connections.
* `validation_context.0.trusted_ca_id` - (Optional) Trusted CA certificate ID in the Certificate Manager.
* `validation_context.0.trusted_ca_bytes` - (Optional) PEM-encoded trusted CA certificate chain.
"""
if sni is not None:
pulumi.set(__self__, "sni", sni)
if validation_context is not None:
pulumi.set(__self__, "validation_context", validation_context)
@property
@pulumi.getter
def sni(self) -> Optional[pulumi.Input[str]]:
"""
[SNI](https://en.wikipedia.org/wiki/Server_Name_Indication) string for TLS connections.
* `validation_context.0.trusted_ca_id` - (Optional) Trusted CA certificate ID in the Certificate Manager.
* `validation_context.0.trusted_ca_bytes` - (Optional) PEM-encoded trusted CA certificate chain.
"""
return pulumi.get(self, "sni")
@sni.setter
def sni(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "sni", value)
@property
@pulumi.getter(name="validationContext")
def validation_context(self) -> Optional[pulumi.Input['AlbBackendGroupGrpcBackendTlsValidationContextArgs']]:
return pulumi.get(self, "validation_context")
@validation_context.setter
def validation_context(self, value: Optional[pulumi.Input['AlbBackendGroupGrpcBackendTlsValidationContextArgs']]):
pulumi.set(self, "validation_context", value)
@pulumi.input_type
class AlbBackendGroupGrpcBackendTlsValidationContextArgs:
def __init__(__self__, *,
trusted_ca_bytes: Optional[pulumi.Input[str]] = None,
trusted_ca_id: Optional[pulumi.Input[str]] = None):
if trusted_ca_bytes is not None:
pulumi.set(__self__, "trusted_ca_bytes", trusted_ca_bytes)
if trusted_ca_id is not None:
pulumi.set(__self__, "trusted_ca_id", trusted_ca_id)
@property
@pulumi.getter(name="trustedCaBytes")
def trusted_ca_bytes(self) -> Optional[pulumi.Input[str]]:
return pulumi.get(self, "trusted_ca_bytes")
@trusted_ca_bytes.setter
def trusted_ca_bytes(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "trusted_ca_bytes", value)
@property
@pulumi.getter(name="trustedCaId")
def trusted_ca_id(self) -> Optional[pulumi.Input[str]]:
return pulumi.get(self, "trusted_ca_id")
@trusted_ca_id.setter
def trusted_ca_id(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "trusted_ca_id", value)
@pulumi.input_type
class AlbBackendGroupHttpBackendArgs:
def __init__(__self__, *,
name: pulumi.Input[str],
target_group_ids: pulumi.Input[Sequence[pulumi.Input[str]]],
healthcheck: Optional[pulumi.Input['AlbBackendGroupHttpBackendHealthcheckArgs']] = None,
http2: Optional[pulumi.Input[bool]] = None,
load_balancing_config: Optional[pulumi.Input['AlbBackendGroupHttpBackendLoadBalancingConfigArgs']] = None,
port: Optional[pulumi.Input[int]] = None,
tls: Optional[pulumi.Input['AlbBackendGroupHttpBackendTlsArgs']] = None,
weight: Optional[pulumi.Input[int]] = None):
"""
:param pulumi.Input[str] name: Name of the backend.
:param pulumi.Input[Sequence[pulumi.Input[str]]] target_group_ids: References target groups for the backend.
:param pulumi.Input['AlbBackendGroupHttpBackendHealthcheckArgs'] healthcheck: Healthcheck specification that will be used by this backend. Structure is documented below.
:param pulumi.Input[bool] http2: If set, health checks will use HTTP2.
:param pulumi.Input['AlbBackendGroupHttpBackendLoadBalancingConfigArgs'] load_balancing_config: Load Balancing Config specification that will be used by this backend. Structure is documented below.
:param pulumi.Input[int] port: Port for incoming traffic.
:param pulumi.Input['AlbBackendGroupHttpBackendTlsArgs'] tls: Tls specification that will be used by this backend. Structure is documented below.
:param pulumi.Input[int] weight: Weight of the backend. Traffic will be split between backends of the same BackendGroup according to their weights.
"""
pulumi.set(__self__, "name", name)
pulumi.set(__self__, "target_group_ids", target_group_ids)
if healthcheck is not None:
pulumi.set(__self__, "healthcheck", healthcheck)
if http2 is not None:
pulumi.set(__self__, "http2", http2)
if load_balancing_config is not None:
pulumi.set(__self__, "load_balancing_config", load_balancing_config)
if port is not None:
pulumi.set(__self__, "port", port)
if tls is not None:
pulumi.set(__self__, "tls", tls)
if weight is not None:
pulumi.set(__self__, "weight", weight)
@property
@pulumi.getter
def name(self) -> pulumi.Input[str]:
"""
Name of the backend.
"""
return pulumi.get(self, "name")
@name.setter
def name(self, value: pulumi.Input[str]):
pulumi.set(self, "name", value)
@property
@pulumi.getter(name="targetGroupIds")
def target_group_ids(self) -> pulumi.Input[Sequence[pulumi.Input[str]]]:
"""
References target groups for the backend.
"""
return pulumi.get(self, "target_group_ids")
@target_group_ids.setter
def target_group_ids(self, value: pulumi.Input[Sequence[pulumi.Input[str]]]):
pulumi.set(self, "target_group_ids", value)
@property
@pulumi.getter
def healthcheck(self) -> Optional[pulumi.Input['AlbBackendGroupHttpBackendHealthcheckArgs']]:
"""
Healthcheck specification that will be used by this backend. Structure is documented below.
"""
return pulumi.get(self, "healthcheck")
@healthcheck.setter
def healthcheck(self, value: Optional[pulumi.Input['AlbBackendGroupHttpBackendHealthcheckArgs']]):
pulumi.set(self, "healthcheck", value)
@property
@pulumi.getter
def http2(self) -> Optional[pulumi.Input[bool]]:
"""
If set, health checks will use HTTP2.
"""
return pulumi.get(self, "http2")
@http2.setter
def http2(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "http2", value)
@property
@pulumi.getter(name="loadBalancingConfig")
def load_balancing_config(self) -> Optional[pulumi.Input['AlbBackendGroupHttpBackendLoadBalancingConfigArgs']]:
"""
Load Balancing Config specification that will be used by this backend. Structure is documented below.
"""
return pulumi.get(self, "load_balancing_config")
@load_balancing_config.setter
def load_balancing_config(self, value: Optional[pulumi.Input['AlbBackendGroupHttpBackendLoadBalancingConfigArgs']]):
pulumi.set(self, "load_balancing_config", value)
@property
@pulumi.getter
def port(self) -> Optional[pulumi.Input[int]]:
"""
Port for incoming traffic.
"""
return pulumi.get(self, "port")
@port.setter
def port(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "port", value)
@property
@pulumi.getter
def tls(self) -> Optional[pulumi.Input['AlbBackendGroupHttpBackendTlsArgs']]:
"""
Tls specification that will be used by this backend. Structure is documented below.
"""
return pulumi.get(self, "tls")
@tls.setter
def tls(self, value: Optional[pulumi.Input['AlbBackendGroupHttpBackendTlsArgs']]):
pulumi.set(self, "tls", value)
@property
@pulumi.getter
def weight(self) -> Optional[pulumi.Input[int]]:
"""
Weight of the backend. Traffic will be split between backends of the same BackendGroup according to their weights.
"""
return pulumi.get(self, "weight")
@weight.setter
def weight(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "weight", value)
@pulumi.input_type
class AlbBackendGroupHttpBackendHealthcheckArgs:
def __init__(__self__, *,
interval: pulumi.Input[str],
timeout: pulumi.Input[str],
grpc_healthcheck: Optional[pulumi.Input['AlbBackendGroupHttpBackendHealthcheckGrpcHealthcheckArgs']] = None,
healthcheck_port: Optional[pulumi.Input[int]] = None,
healthy_threshold: Optional[pulumi.Input[int]] = None,
http_healthcheck: Optional[pulumi.Input['AlbBackendGroupHttpBackendHealthcheckHttpHealthcheckArgs']] = None,
interval_jitter_percent: Optional[pulumi.Input[float]] = None,
stream_healthcheck: Optional[pulumi.Input['AlbBackendGroupHttpBackendHealthcheckStreamHealthcheckArgs']] = None,
unhealthy_threshold: Optional[pulumi.Input[int]] = None):
"""
:param pulumi.Input[str] interval: Interval between health checks.
:param pulumi.Input[str] timeout: Time to wait for a health check response.
:param pulumi.Input['AlbBackendGroupHttpBackendHealthcheckGrpcHealthcheckArgs'] grpc_healthcheck: Grpc Healthcheck specification that will be used by this healthcheck. Structure is documented below.
:param pulumi.Input[int] healthcheck_port: Optional alternative port for health checking.
:param pulumi.Input[int] healthy_threshold: Number of consecutive successful health checks required to promote endpoint into the healthy state. 0 means 1. Note that during startup, only a single successful health check is required to mark a host healthy.
:param pulumi.Input['AlbBackendGroupHttpBackendHealthcheckHttpHealthcheckArgs'] http_healthcheck: Http Healthcheck specification that will be used by this healthcheck. Structure is documented below.
:param pulumi.Input[float] interval_jitter_percent: An optional jitter amount as a percentage of interval. If specified, during every interval value of (interval_ms * interval_jitter_percent / 100) will be added to the wait time.
:param pulumi.Input['AlbBackendGroupHttpBackendHealthcheckStreamHealthcheckArgs'] stream_healthcheck: Stream Healthcheck specification that will be used by this healthcheck. Structure is documented below.
:param pulumi.Input[int] unhealthy_threshold: Number of consecutive failed health checks required to demote endpoint into the unhealthy state. 0 means 1. Note that for HTTP health checks, a single 503 immediately makes endpoint unhealthy.
"""
pulumi.set(__self__, "interval", interval)
pulumi.set(__self__, "timeout", timeout)
if grpc_healthcheck is not None:
pulumi.set(__self__, "grpc_healthcheck", grpc_healthcheck)
if healthcheck_port is not None:
pulumi.set(__self__, "healthcheck_port", healthcheck_port)
if healthy_threshold is not None:
pulumi.set(__self__, "healthy_threshold", healthy_threshold)
if http_healthcheck is not None:
pulumi.set(__self__, "http_healthcheck", http_healthcheck)
if interval_jitter_percent is not None:
pulumi.set(__self__, "interval_jitter_percent", interval_jitter_percent)
if stream_healthcheck is not None:
pulumi.set(__self__, "stream_healthcheck", stream_healthcheck)
if unhealthy_threshold is not None:
pulumi.set(__self__, "unhealthy_threshold", unhealthy_threshold)
@property
@pulumi.getter
def interval(self) -> pulumi.Input[str]:
"""
Interval between health checks.
"""
return pulumi.get(self, "interval")
@interval.setter
def interval(self, value: pulumi.Input[str]):
pulumi.set(self, "interval", value)
@property
@pulumi.getter
def timeout(self) -> pulumi.Input[str]:
"""
Time to wait for a health check response.
"""
return pulumi.get(self, "timeout")
@timeout.setter
def timeout(self, value: pulumi.Input[str]):
pulumi.set(self, "timeout", value)
@property
@pulumi.getter(name="grpcHealthcheck")
def grpc_healthcheck(self) -> Optional[pulumi.Input['AlbBackendGroupHttpBackendHealthcheckGrpcHealthcheckArgs']]:
"""
Grpc Healthcheck specification that will be used by this healthcheck. Structure is documented below.
"""
return pulumi.get(self, "grpc_healthcheck")
@grpc_healthcheck.setter
def grpc_healthcheck(self, value: Optional[pulumi.Input['AlbBackendGroupHttpBackendHealthcheckGrpcHealthcheckArgs']]):
pulumi.set(self, "grpc_healthcheck", value)
@property
@pulumi.getter(name="healthcheckPort")
def healthcheck_port(self) -> Optional[pulumi.Input[int]]:
"""
Optional alternative port for health checking.
"""
return pulumi.get(self, "healthcheck_port")
@healthcheck_port.setter
def healthcheck_port(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "healthcheck_port", value)
@property
@pulumi.getter(name="healthyThreshold")
def healthy_threshold(self) -> Optional[pulumi.Input[int]]:
"""
Number of consecutive successful health checks required to promote endpoint into the healthy state. 0 means 1. Note that during startup, only a single successful health check is required to mark a host healthy.
"""
return pulumi.get(self, "healthy_threshold")
@healthy_threshold.setter
def healthy_threshold(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "healthy_threshold", value)
@property
@pulumi.getter(name="httpHealthcheck")
def http_healthcheck(self) -> Optional[pulumi.Input['AlbBackendGroupHttpBackendHealthcheckHttpHealthcheckArgs']]:
"""
Http Healthcheck specification that will be used by this healthcheck. Structure is documented below.
"""
return pulumi.get(self, "http_healthcheck")
@http_healthcheck.setter
def http_healthcheck(self, value: Optional[pulumi.Input['AlbBackendGroupHttpBackendHealthcheckHttpHealthcheckArgs']]):
pulumi.set(self, "http_healthcheck", value)
@property
@pulumi.getter(name="intervalJitterPercent")
def interval_jitter_percent(self) -> Optional[pulumi.Input[float]]:
"""
An optional jitter amount as a percentage of interval. If specified, during every interval value of (interval_ms * interval_jitter_percent / 100) will be added to the wait time.
"""
return pulumi.get(self, "interval_jitter_percent")
@interval_jitter_percent.setter
def interval_jitter_percent(self, value: Optional[pulumi.Input[float]]):
pulumi.set(self, "interval_jitter_percent", value)
@property
@pulumi.getter(name="streamHealthcheck")
def stream_healthcheck(self) -> Optional[pulumi.Input['AlbBackendGroupHttpBackendHealthcheckStreamHealthcheckArgs']]:
"""
Stream Healthcheck specification that will be used by this healthcheck. Structure is documented below.
"""
return pulumi.get(self, "stream_healthcheck")
@stream_healthcheck.setter
def stream_healthcheck(self, value: Optional[pulumi.Input['AlbBackendGroupHttpBackendHealthcheckStreamHealthcheckArgs']]):
pulumi.set(self, "stream_healthcheck", value)
@property
@pulumi.getter(name="unhealthyThreshold")
def unhealthy_threshold(self) -> Optional[pulumi.Input[int]]:
"""
Number of consecutive failed health checks required to demote endpoint into the unhealthy state. 0 means 1. Note that for HTTP health checks, a single 503 immediately makes endpoint unhealthy.
"""
return pulumi.get(self, "unhealthy_threshold")
@unhealthy_threshold.setter
def unhealthy_threshold(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "unhealthy_threshold", value)
@pulumi.input_type
class AlbBackendGroupHttpBackendHealthcheckGrpcHealthcheckArgs:
def __init__(__self__, *,
service_name: Optional[pulumi.Input[str]] = None):
"""
:param pulumi.Input[str] service_name: Service name for grpc.health.v1.HealthCheckRequest message.
"""
if service_name is not None:
pulumi.set(__self__, "service_name", service_name)
@property
@pulumi.getter(name="serviceName")
def service_name(self) -> Optional[pulumi.Input[str]]:
"""
Service name for grpc.health.v1.HealthCheckRequest message.
"""
return pulumi.get(self, "service_name")
@service_name.setter
def service_name(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "service_name", value)
@pulumi.input_type
class AlbBackendGroupHttpBackendHealthcheckHttpHealthcheckArgs:
def __init__(__self__, *,
path: pulumi.Input[str],
host: Optional[pulumi.Input[str]] = None,
http2: Optional[pulumi.Input[bool]] = None):
"""
:param pulumi.Input[str] path: HTTP path.
:param pulumi.Input[str] host: "Host" HTTP header value.
:param pulumi.Input[bool] http2: If set, health checks will use HTTP2.
"""
pulumi.set(__self__, "path", path)
if host is not None:
pulumi.set(__self__, "host", host)
if http2 is not None:
pulumi.set(__self__, "http2", http2)
@property
@pulumi.getter
def path(self) -> pulumi.Input[str]:
"""
HTTP path.
"""
return pulumi.get(self, "path")
@path.setter
def path(self, value: pulumi.Input[str]):
pulumi.set(self, "path", value)
@property
@pulumi.getter
def host(self) -> Optional[pulumi.Input[str]]:
"""
"Host" HTTP header value.
"""
return pulumi.get(self, "host")
@host.setter
def host(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "host", value)
@property
@pulumi.getter
def http2(self) -> Optional[pulumi.Input[bool]]:
"""
If set, health checks will use HTTP2.
"""
return pulumi.get(self, "http2")
@http2.setter
def http2(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "http2", value)
@pulumi.input_type
class AlbBackendGroupHttpBackendHealthcheckStreamHealthcheckArgs:
def __init__(__self__, *,
receive: Optional[pulumi.Input[str]] = None,
send: Optional[pulumi.Input[str]] = None):
"""
:param pulumi.Input[str] receive: Text to search in reply.
:param pulumi.Input[str] send: Message to send. If empty, it's a connect-only health check.
"""
if receive is not None:
pulumi.set(__self__, "receive", receive)
if send is not None:
pulumi.set(__self__, "send", send)
@property
@pulumi.getter
def receive(self) -> Optional[pulumi.Input[str]]:
"""
Text to search in reply.
"""
return pulumi.get(self, "receive")
@receive.setter
def receive(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "receive", value)
@property
@pulumi.getter
def send(self) -> Optional[pulumi.Input[str]]:
"""
Message to send. If empty, it's a connect-only health check.
"""
return pulumi.get(self, "send")
@send.setter
def send(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "send", value)
@pulumi.input_type
class AlbBackendGroupHttpBackendLoadBalancingConfigArgs:
def __init__(__self__, *,
locality_aware_routing_percent: Optional[pulumi.Input[int]] = None,
panic_threshold: Optional[pulumi.Input[int]] = None,
strict_locality: Optional[pulumi.Input[bool]] = None):
"""
:param pulumi.Input[int] locality_aware_routing_percent: Percent of traffic to be sent to the same availability zone. The rest will be equally divided between other zones.
:param pulumi.Input[int] panic_threshold: If percentage of healthy hosts in the backend is lower than panic_threshold, traffic will be routed to all backends no matter what the health status is. This helps to avoid healthy backends overloading when everything is bad. Zero means no panic threshold.
:param pulumi.Input[bool] strict_locality: If set, will route requests only to the same availability zone. Balancer won't know about endpoints in other zones.
"""
if locality_aware_routing_percent is not None:
pulumi.set(__self__, "locality_aware_routing_percent", locality_aware_routing_percent)
if panic_threshold is not None:
pulumi.set(__self__, "panic_threshold", panic_threshold)
if strict_locality is not None:
pulumi.set(__self__, "strict_locality", strict_locality)
@property
@pulumi.getter(name="localityAwareRoutingPercent")
def locality_aware_routing_percent(self) -> Optional[pulumi.Input[int]]:
"""
Percent of traffic to be sent to the same availability zone. The rest will be equally divided between other zones.
"""
return pulumi.get(self, "locality_aware_routing_percent")
@locality_aware_routing_percent.setter
def locality_aware_routing_percent(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "locality_aware_routing_percent", value)
@property
@pulumi.getter(name="panicThreshold")
def panic_threshold(self) -> Optional[pulumi.Input[int]]:
"""
If percentage of healthy hosts in the backend is lower than panic_threshold, traffic will be routed to all backends no matter what the health status is. This helps to avoid healthy backends overloading when everything is bad. Zero means no panic threshold.
"""
return pulumi.get(self, "panic_threshold")
@panic_threshold.setter
def panic_threshold(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "panic_threshold", value)
@property
@pulumi.getter(name="strictLocality")
def strict_locality(self) -> Optional[pulumi.Input[bool]]:
"""
If set, will route requests only to the same availability zone. Balancer won't know about endpoints in other zones.
"""
return pulumi.get(self, "strict_locality")
@strict_locality.setter
def strict_locality(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "strict_locality", value)
@pulumi.input_type
class AlbBackendGroupHttpBackendTlsArgs:
def __init__(__self__, *,
sni: Optional[pulumi.Input[str]] = None,
validation_context: Optional[pulumi.Input['AlbBackendGroupHttpBackendTlsValidationContextArgs']] = None):
"""
:param pulumi.Input[str] sni: [SNI](https://en.wikipedia.org/wiki/Server_Name_Indication) string for TLS connections.
* `validation_context.0.trusted_ca_id` - (Optional) Trusted CA certificate ID in the Certificate Manager.
* `validation_context.0.trusted_ca_bytes` - (Optional) PEM-encoded trusted CA certificate chain.
"""
if sni is not None:
pulumi.set(__self__, "sni", sni)
if validation_context is not None:
pulumi.set(__self__, "validation_context", validation_context)
@property
@pulumi.getter
def sni(self) -> Optional[pulumi.Input[str]]:
"""
[SNI](https://en.wikipedia.org/wiki/Server_Name_Indication) string for TLS connections.
* `validation_context.0.trusted_ca_id` - (Optional) Trusted CA certificate ID in the Certificate Manager.
* `validation_context.0.trusted_ca_bytes` - (Optional) PEM-encoded trusted CA certificate chain.
"""
return pulumi.get(self, "sni")
@sni.setter
def sni(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "sni", value)
@property
@pulumi.getter(name="validationContext")
def validation_context(self) -> Optional[pulumi.Input['AlbBackendGroupHttpBackendTlsValidationContextArgs']]:
return pulumi.get(self, "validation_context")
@validation_context.setter
def validation_context(self, value: Optional[pulumi.Input['AlbBackendGroupHttpBackendTlsValidationContextArgs']]):
pulumi.set(self, "validation_context", value)
@pulumi.input_type
class AlbBackendGroupHttpBackendTlsValidationContextArgs:
def __init__(__self__, *,
trusted_ca_bytes: Optional[pulumi.Input[str]] = None,
trusted_ca_id: Optional[pulumi.Input[str]] = None):
if trusted_ca_bytes is not None:
pulumi.set(__self__, "trusted_ca_bytes", trusted_ca_bytes)
if trusted_ca_id is not None:
pulumi.set(__self__, "trusted_ca_id", trusted_ca_id)
@property
@pulumi.getter(name="trustedCaBytes")
def trusted_ca_bytes(self) -> Optional[pulumi.Input[str]]:
return pulumi.get(self, "trusted_ca_bytes")
@trusted_ca_bytes.setter
def trusted_ca_bytes(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "trusted_ca_bytes", value)
@property
@pulumi.getter(name="trustedCaId")
def trusted_ca_id(self) -> Optional[pulumi.Input[str]]:
return pulumi.get(self, "trusted_ca_id")
@trusted_ca_id.setter
def trusted_ca_id(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "trusted_ca_id", value)
@pulumi.input_type
class AlbLoadBalancerAllocationPolicyArgs:
def __init__(__self__, *,
locations: pulumi.Input[Sequence[pulumi.Input['AlbLoadBalancerAllocationPolicyLocationArgs']]]):
"""
:param pulumi.Input[Sequence[pulumi.Input['AlbLoadBalancerAllocationPolicyLocationArgs']]] locations: Unique set of locations. The structure is documented below.
"""
pulumi.set(__self__, "locations", locations)
@property
@pulumi.getter
def locations(self) -> pulumi.Input[Sequence[pulumi.Input['AlbLoadBalancerAllocationPolicyLocationArgs']]]:
"""
Unique set of locations. The structure is documented below.
"""
return pulumi.get(self, "locations")
@locations.setter
def locations(self, value: pulumi.Input[Sequence[pulumi.Input['AlbLoadBalancerAllocationPolicyLocationArgs']]]):
pulumi.set(self, "locations", value)
@pulumi.input_type
class AlbLoadBalancerAllocationPolicyLocationArgs:
def __init__(__self__, *,
subnet_id: pulumi.Input[str],
zone_id: pulumi.Input[str],
disable_traffic: Optional[pulumi.Input[bool]] = None):
"""
:param pulumi.Input[str] subnet_id: Provided by the client or computed automatically.
:param pulumi.Input[str] zone_id: ID of the zone that location is located at.
:param pulumi.Input[bool] disable_traffic: If set, will disable all L7 instances in the zone for request handling.
"""
pulumi.set(__self__, "subnet_id", subnet_id)
pulumi.set(__self__, "zone_id", zone_id)
if disable_traffic is not None:
pulumi.set(__self__, "disable_traffic", disable_traffic)
@property
@pulumi.getter(name="subnetId")
def subnet_id(self) -> pulumi.Input[str]:
"""
Provided by the client or computed automatically.
"""
return pulumi.get(self, "subnet_id")
@subnet_id.setter
def subnet_id(self, value: pulumi.Input[str]):
pulumi.set(self, "subnet_id", value)
@property
@pulumi.getter(name="zoneId")
def zone_id(self) -> pulumi.Input[str]:
"""
ID of the zone that location is located at.
"""
return pulumi.get(self, "zone_id")
@zone_id.setter
def zone_id(self, value: pulumi.Input[str]):
pulumi.set(self, "zone_id", value)
@property
@pulumi.getter(name="disableTraffic")
def disable_traffic(self) -> Optional[pulumi.Input[bool]]:
"""
If set, will disable all L7 instances in the zone for request handling.
"""
return pulumi.get(self, "disable_traffic")
@disable_traffic.setter
def disable_traffic(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "disable_traffic", value)
@pulumi.input_type
class AlbLoadBalancerListenerArgs:
def __init__(__self__, *,
name: pulumi.Input[str],
endpoints: Optional[pulumi.Input[Sequence[pulumi.Input['AlbLoadBalancerListenerEndpointArgs']]]] = None,
http: Optional[pulumi.Input['AlbLoadBalancerListenerHttpArgs']] = None,
tls: Optional[pulumi.Input['AlbLoadBalancerListenerTlsArgs']] = None):
"""
:param pulumi.Input[str] name: name of SNI match.
:param pulumi.Input[Sequence[pulumi.Input['AlbLoadBalancerListenerEndpointArgs']]] endpoints: Network endpoints (addresses and ports) of the listener. The structure is documented below.
:param pulumi.Input['AlbLoadBalancerListenerHttpArgs'] http: HTTP listener resource. The structure is documented below.
:param pulumi.Input['AlbLoadBalancerListenerTlsArgs'] tls: TLS listener resource. The structure is documented below.
"""
pulumi.set(__self__, "name", name)
if endpoints is not None:
pulumi.set(__self__, "endpoints", endpoints)
if http is not None:
pulumi.set(__self__, "http", http)
if tls is not None:
pulumi.set(__self__, "tls", tls)
@property
@pulumi.getter
def name(self) -> pulumi.Input[str]:
"""
name of SNI match.
"""
return pulumi.get(self, "name")
@name.setter
def name(self, value: pulumi.Input[str]):
pulumi.set(self, "name", value)
@property
@pulumi.getter
def endpoints(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['AlbLoadBalancerListenerEndpointArgs']]]]:
"""
Network endpoints (addresses and ports) of the listener. The structure is documented below.
"""
return pulumi.get(self, "endpoints")
@endpoints.setter
def endpoints(self, value: Optional[pulumi.Input[Sequence[pulumi.Input['AlbLoadBalancerListenerEndpointArgs']]]]):
pulumi.set(self, "endpoints", value)
@property
@pulumi.getter
def http(self) -> Optional[pulumi.Input['AlbLoadBalancerListenerHttpArgs']]:
"""
HTTP listener resource. The structure is documented below.
"""
return pulumi.get(self, "http")
@http.setter
def http(self, value: Optional[pulumi.Input['AlbLoadBalancerListenerHttpArgs']]):
pulumi.set(self, "http", value)
@property
@pulumi.getter
def tls(self) -> Optional[pulumi.Input['AlbLoadBalancerListenerTlsArgs']]:
"""
TLS listener resource. The structure is documented below.
"""
return pulumi.get(self, "tls")
@tls.setter
def tls(self, value: Optional[pulumi.Input['AlbLoadBalancerListenerTlsArgs']]):
pulumi.set(self, "tls", value)
@pulumi.input_type
class AlbLoadBalancerListenerEndpointArgs:
def __init__(__self__, *,
addresses: pulumi.Input[Sequence[pulumi.Input['AlbLoadBalancerListenerEndpointAddressArgs']]],
ports: pulumi.Input[Sequence[pulumi.Input[int]]]):
"""
:param pulumi.Input[Sequence[pulumi.Input['AlbLoadBalancerListenerEndpointAddressArgs']]] addresses: Provided by the client or computed automatically.
:param pulumi.Input[Sequence[pulumi.Input[int]]] ports: One or more ports to listen on.
"""
pulumi.set(__self__, "addresses", addresses)
pulumi.set(__self__, "ports", ports)
@property
@pulumi.getter
def addresses(self) -> pulumi.Input[Sequence[pulumi.Input['AlbLoadBalancerListenerEndpointAddressArgs']]]:
"""
Provided by the client or computed automatically.
"""
return pulumi.get(self, "addresses")
@addresses.setter
def addresses(self, value: pulumi.Input[Sequence[pulumi.Input['AlbLoadBalancerListenerEndpointAddressArgs']]]):
pulumi.set(self, "addresses", value)
@property
@pulumi.getter
def ports(self) -> pulumi.Input[Sequence[pulumi.Input[int]]]:
"""
One or more ports to listen on.
"""
return pulumi.get(self, "ports")
@ports.setter
def ports(self, value: pulumi.Input[Sequence[pulumi.Input[int]]]):
pulumi.set(self, "ports", value)
@pulumi.input_type
class AlbLoadBalancerListenerEndpointAddressArgs:
def __init__(__self__, *,
external_ipv4_address: Optional[pulumi.Input['AlbLoadBalancerListenerEndpointAddressExternalIpv4AddressArgs']] = None,
external_ipv6_address: Optional[pulumi.Input['AlbLoadBalancerListenerEndpointAddressExternalIpv6AddressArgs']] = None,
internal_ipv4_address: Optional[pulumi.Input['AlbLoadBalancerListenerEndpointAddressInternalIpv4AddressArgs']] = None):
"""
:param pulumi.Input['AlbLoadBalancerListenerEndpointAddressExternalIpv4AddressArgs'] external_ipv4_address: External IPv4 address. The structure is documented below.
:param pulumi.Input['AlbLoadBalancerListenerEndpointAddressExternalIpv6AddressArgs'] external_ipv6_address: External IPv6 address. The structure is documented below.
:param pulumi.Input['AlbLoadBalancerListenerEndpointAddressInternalIpv4AddressArgs'] internal_ipv4_address: Internal IPv4 address. The structure is documented below.
"""
if external_ipv4_address is not None:
pulumi.set(__self__, "external_ipv4_address", external_ipv4_address)
if external_ipv6_address is not None:
pulumi.set(__self__, "external_ipv6_address", external_ipv6_address)
if internal_ipv4_address is not None:
pulumi.set(__self__, "internal_ipv4_address", internal_ipv4_address)
@property
@pulumi.getter(name="externalIpv4Address")
def external_ipv4_address(self) -> Optional[pulumi.Input['AlbLoadBalancerListenerEndpointAddressExternalIpv4AddressArgs']]:
"""
External IPv4 address. The structure is documented below.
"""
return pulumi.get(self, "external_ipv4_address")
@external_ipv4_address.setter
def external_ipv4_address(self, value: Optional[pulumi.Input['AlbLoadBalancerListenerEndpointAddressExternalIpv4AddressArgs']]):
pulumi.set(self, "external_ipv4_address", value)
@property
@pulumi.getter(name="externalIpv6Address")
def external_ipv6_address(self) -> Optional[pulumi.Input['AlbLoadBalancerListenerEndpointAddressExternalIpv6AddressArgs']]:
"""
External IPv6 address. The structure is documented below.
"""
return pulumi.get(self, "external_ipv6_address")
@external_ipv6_address.setter
def external_ipv6_address(self, value: Optional[pulumi.Input['AlbLoadBalancerListenerEndpointAddressExternalIpv6AddressArgs']]):
pulumi.set(self, "external_ipv6_address", value)
@property
@pulumi.getter(name="internalIpv4Address")
def internal_ipv4_address(self) -> Optional[pulumi.Input['AlbLoadBalancerListenerEndpointAddressInternalIpv4AddressArgs']]:
"""
Internal IPv4 address. The structure is documented below.
"""
return pulumi.get(self, "internal_ipv4_address")
@internal_ipv4_address.setter
def internal_ipv4_address(self, value: Optional[pulumi.Input['AlbLoadBalancerListenerEndpointAddressInternalIpv4AddressArgs']]):
pulumi.set(self, "internal_ipv4_address", value)
@pulumi.input_type
class AlbLoadBalancerListenerEndpointAddressExternalIpv4AddressArgs:
def __init__(__self__, *,
address: Optional[pulumi.Input[str]] = None):
"""
:param pulumi.Input[str] address: Provided by the client or computed automatically.
"""
if address is not None:
pulumi.set(__self__, "address", address)
@property
@pulumi.getter
def address(self) -> Optional[pulumi.Input[str]]:
"""
Provided by the client or computed automatically.
"""
return pulumi.get(self, "address")
@address.setter
def address(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "address", value)
@pulumi.input_type
class AlbLoadBalancerListenerEndpointAddressExternalIpv6AddressArgs:
def __init__(__self__, *,
address: Optional[pulumi.Input[str]] = None):
"""
:param pulumi.Input[str] address: Provided by the client or computed automatically.
"""
if address is not None:
pulumi.set(__self__, "address", address)
@property
@pulumi.getter
def address(self) -> Optional[pulumi.Input[str]]:
"""
Provided by the client or computed automatically.
"""
return pulumi.get(self, "address")
@address.setter
def address(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "address", value)
@pulumi.input_type
class AlbLoadBalancerListenerEndpointAddressInternalIpv4AddressArgs:
def __init__(__self__, *,
address: Optional[pulumi.Input[str]] = None,
subnet_id: Optional[pulumi.Input[str]] = None):
"""
:param pulumi.Input[str] address: Provided by the client or computed automatically.
:param pulumi.Input[str] subnet_id: Provided by the client or computed automatically.
"""
if address is not None:
pulumi.set(__self__, "address", address)
if subnet_id is not None:
pulumi.set(__self__, "subnet_id", subnet_id)
@property
@pulumi.getter
def address(self) -> Optional[pulumi.Input[str]]:
"""
Provided by the client or computed automatically.
"""
return pulumi.get(self, "address")
@address.setter
def address(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "address", value)
@property
@pulumi.getter(name="subnetId")
def subnet_id(self) -> Optional[pulumi.Input[str]]:
"""
Provided by the client or computed automatically.
"""
return pulumi.get(self, "subnet_id")
@subnet_id.setter
def subnet_id(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "subnet_id", value)
@pulumi.input_type
class AlbLoadBalancerListenerHttpArgs:
def __init__(__self__, *,
handler: Optional[pulumi.Input['AlbLoadBalancerListenerHttpHandlerArgs']] = None,
redirects: Optional[pulumi.Input['AlbLoadBalancerListenerHttpRedirectsArgs']] = None):
"""
:param pulumi.Input['AlbLoadBalancerListenerHttpHandlerArgs'] handler: HTTP handler that sets plaintext HTTP router. The structure is documented below.
:param pulumi.Input['AlbLoadBalancerListenerHttpRedirectsArgs'] redirects: Shortcut for adding http > https redirects. The structure is documented below.
"""
if handler is not None:
pulumi.set(__self__, "handler", handler)
if redirects is not None:
pulumi.set(__self__, "redirects", redirects)
@property
@pulumi.getter
def handler(self) -> Optional[pulumi.Input['AlbLoadBalancerListenerHttpHandlerArgs']]:
"""
HTTP handler that sets plaintext HTTP router. The structure is documented below.
"""
return pulumi.get(self, "handler")
@handler.setter
def handler(self, value: Optional[pulumi.Input['AlbLoadBalancerListenerHttpHandlerArgs']]):
pulumi.set(self, "handler", value)
@property
@pulumi.getter
def redirects(self) -> Optional[pulumi.Input['AlbLoadBalancerListenerHttpRedirectsArgs']]:
"""
Shortcut for adding http > https redirects. The structure is documented below.
"""
return pulumi.get(self, "redirects")
@redirects.setter
def redirects(self, value: Optional[pulumi.Input['AlbLoadBalancerListenerHttpRedirectsArgs']]):
pulumi.set(self, "redirects", value)
@pulumi.input_type
class AlbLoadBalancerListenerHttpHandlerArgs:
def __init__(__self__, *,
allow_http10: Optional[pulumi.Input[bool]] = None,
http2_options: Optional[pulumi.Input['AlbLoadBalancerListenerHttpHandlerHttp2OptionsArgs']] = None,
http_router_id: Optional[pulumi.Input[str]] = None):
"""
:param pulumi.Input[bool] allow_http10: If set, will enable only HTTP1 protocol with HTTP1.0 support.
:param pulumi.Input['AlbLoadBalancerListenerHttpHandlerHttp2OptionsArgs'] http2_options: If set, will enable HTTP2 protocol for the handler. The structure is documented below.
:param pulumi.Input[str] http_router_id: HTTP router id.
"""
if allow_http10 is not None:
pulumi.set(__self__, "allow_http10", allow_http10)
if http2_options is not None:
pulumi.set(__self__, "http2_options", http2_options)
if http_router_id is not None:
pulumi.set(__self__, "http_router_id", http_router_id)
@property
@pulumi.getter(name="allowHttp10")
def allow_http10(self) -> Optional[pulumi.Input[bool]]:
"""
If set, will enable only HTTP1 protocol with HTTP1.0 support.
"""
return pulumi.get(self, "allow_http10")
@allow_http10.setter
def allow_http10(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "allow_http10", value)
@property
@pulumi.getter(name="http2Options")
def http2_options(self) -> Optional[pulumi.Input['AlbLoadBalancerListenerHttpHandlerHttp2OptionsArgs']]:
"""
If set, will enable HTTP2 protocol for the handler. The structure is documented below.
"""
return pulumi.get(self, "http2_options")
@http2_options.setter
def http2_options(self, value: Optional[pulumi.Input['AlbLoadBalancerListenerHttpHandlerHttp2OptionsArgs']]):
pulumi.set(self, "http2_options", value)
@property
@pulumi.getter(name="httpRouterId")
def http_router_id(self) -> Optional[pulumi.Input[str]]:
"""
HTTP router id.
"""
return pulumi.get(self, "http_router_id")
@http_router_id.setter
def http_router_id(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "http_router_id", value)
@pulumi.input_type
class AlbLoadBalancerListenerHttpHandlerHttp2OptionsArgs:
def __init__(__self__, *,
max_concurrent_streams: Optional[pulumi.Input[int]] = None):
"""
:param pulumi.Input[int] max_concurrent_streams: Maximum number of concurrent streams.
"""
if max_concurrent_streams is not None:
pulumi.set(__self__, "max_concurrent_streams", max_concurrent_streams)
@property
@pulumi.getter(name="maxConcurrentStreams")
def max_concurrent_streams(self) -> Optional[pulumi.Input[int]]:
"""
Maximum number of concurrent streams.
"""
return pulumi.get(self, "max_concurrent_streams")
@max_concurrent_streams.setter
def max_concurrent_streams(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "max_concurrent_streams", value)
@pulumi.input_type
class AlbLoadBalancerListenerHttpRedirectsArgs:
def __init__(__self__, *,
http_to_https: Optional[pulumi.Input[bool]] = None):
if http_to_https is not None:
pulumi.set(__self__, "http_to_https", http_to_https)
@property
@pulumi.getter(name="httpToHttps")
def http_to_https(self) -> Optional[pulumi.Input[bool]]:
return pulumi.get(self, "http_to_https")
@http_to_https.setter
def http_to_https(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "http_to_https", value)
@pulumi.input_type
class AlbLoadBalancerListenerTlsArgs:
def __init__(__self__, *,
default_handler: pulumi.Input['AlbLoadBalancerListenerTlsDefaultHandlerArgs'],
sni_handlers: Optional[pulumi.Input[Sequence[pulumi.Input['AlbLoadBalancerListenerTlsSniHandlerArgs']]]] = None):
"""
:param pulumi.Input['AlbLoadBalancerListenerTlsDefaultHandlerArgs'] default_handler: TLS handler resource. The structure is documented below.
:param pulumi.Input[Sequence[pulumi.Input['AlbLoadBalancerListenerTlsSniHandlerArgs']]] sni_handlers: SNI match resource. The structure is documented below.
"""
pulumi.set(__self__, "default_handler", default_handler)
if sni_handlers is not None:
pulumi.set(__self__, "sni_handlers", sni_handlers)
@property
@pulumi.getter(name="defaultHandler")
def default_handler(self) -> pulumi.Input['AlbLoadBalancerListenerTlsDefaultHandlerArgs']:
"""
TLS handler resource. The structure is documented below.
"""
return pulumi.get(self, "default_handler")
@default_handler.setter
def default_handler(self, value: pulumi.Input['AlbLoadBalancerListenerTlsDefaultHandlerArgs']):
pulumi.set(self, "default_handler", value)
@property
@pulumi.getter(name="sniHandlers")
def sni_handlers(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['AlbLoadBalancerListenerTlsSniHandlerArgs']]]]:
"""
SNI match resource. The structure is documented below.
"""
return pulumi.get(self, "sni_handlers")
@sni_handlers.setter
def sni_handlers(self, value: Optional[pulumi.Input[Sequence[pulumi.Input['AlbLoadBalancerListenerTlsSniHandlerArgs']]]]):
pulumi.set(self, "sni_handlers", value)
@pulumi.input_type
class AlbLoadBalancerListenerTlsDefaultHandlerArgs:
def __init__(__self__, *,
certificate_ids: pulumi.Input[Sequence[pulumi.Input[str]]],
http_handler: Optional[pulumi.Input['AlbLoadBalancerListenerTlsDefaultHandlerHttpHandlerArgs']] = None):
"""
:param pulumi.Input[Sequence[pulumi.Input[str]]] certificate_ids: Certificate IDs in the Certificate Manager. Multiple TLS certificates can be associated
with the same context to allow both RSA and ECDSA certificates. Only the first certificate of each type will be used.
:param pulumi.Input['AlbLoadBalancerListenerTlsDefaultHandlerHttpHandlerArgs'] http_handler: HTTP handler resource. The structure is documented below.
"""
pulumi.set(__self__, "certificate_ids", certificate_ids)
if http_handler is not None:
pulumi.set(__self__, "http_handler", http_handler)
@property
@pulumi.getter(name="certificateIds")
def certificate_ids(self) -> pulumi.Input[Sequence[pulumi.Input[str]]]:
"""
Certificate IDs in the Certificate Manager. Multiple TLS certificates can be associated
with the same context to allow both RSA and ECDSA certificates. Only the first certificate of each type will be used.
"""
return pulumi.get(self, "certificate_ids")
@certificate_ids.setter
def certificate_ids(self, value: pulumi.Input[Sequence[pulumi.Input[str]]]):
pulumi.set(self, "certificate_ids", value)
@property
@pulumi.getter(name="httpHandler")
def http_handler(self) -> Optional[pulumi.Input['AlbLoadBalancerListenerTlsDefaultHandlerHttpHandlerArgs']]:
"""
HTTP handler resource. The structure is documented below.
"""
return pulumi.get(self, "http_handler")
@http_handler.setter
def http_handler(self, value: Optional[pulumi.Input['AlbLoadBalancerListenerTlsDefaultHandlerHttpHandlerArgs']]):
pulumi.set(self, "http_handler", value)
@pulumi.input_type
class AlbLoadBalancerListenerTlsDefaultHandlerHttpHandlerArgs:
def __init__(__self__, *,
allow_http10: Optional[pulumi.Input[bool]] = None,
http2_options: Optional[pulumi.Input['AlbLoadBalancerListenerTlsDefaultHandlerHttpHandlerHttp2OptionsArgs']] = None,
http_router_id: Optional[pulumi.Input[str]] = None):
"""
:param pulumi.Input[bool] allow_http10: If set, will enable only HTTP1 protocol with HTTP1.0 support.
:param pulumi.Input['AlbLoadBalancerListenerTlsDefaultHandlerHttpHandlerHttp2OptionsArgs'] http2_options: If set, will enable HTTP2 protocol for the handler. The structure is documented below.
:param pulumi.Input[str] http_router_id: HTTP router id.
"""
if allow_http10 is not None:
pulumi.set(__self__, "allow_http10", allow_http10)
if http2_options is not None:
pulumi.set(__self__, "http2_options", http2_options)
if http_router_id is not None:
pulumi.set(__self__, "http_router_id", http_router_id)
@property
@pulumi.getter(name="allowHttp10")
def allow_http10(self) -> Optional[pulumi.Input[bool]]:
"""
If set, will enable only HTTP1 protocol with HTTP1.0 support.
"""
return pulumi.get(self, "allow_http10")
@allow_http10.setter
def allow_http10(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "allow_http10", value)
@property
@pulumi.getter(name="http2Options")
def http2_options(self) -> Optional[pulumi.Input['AlbLoadBalancerListenerTlsDefaultHandlerHttpHandlerHttp2OptionsArgs']]:
"""
If set, will enable HTTP2 protocol for the handler. The structure is documented below.
"""
return pulumi.get(self, "http2_options")
@http2_options.setter
def http2_options(self, value: Optional[pulumi.Input['AlbLoadBalancerListenerTlsDefaultHandlerHttpHandlerHttp2OptionsArgs']]):
pulumi.set(self, "http2_options", value)
@property
@pulumi.getter(name="httpRouterId")
def http_router_id(self) -> Optional[pulumi.Input[str]]:
"""
HTTP router id.
"""
return pulumi.get(self, "http_router_id")
@http_router_id.setter
def http_router_id(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "http_router_id", value)
@pulumi.input_type
class AlbLoadBalancerListenerTlsDefaultHandlerHttpHandlerHttp2OptionsArgs:
def __init__(__self__, *,
max_concurrent_streams: Optional[pulumi.Input[int]] = None):
"""
:param pulumi.Input[int] max_concurrent_streams: Maximum number of concurrent streams.
"""
if max_concurrent_streams is not None:
pulumi.set(__self__, "max_concurrent_streams", max_concurrent_streams)
@property
@pulumi.getter(name="maxConcurrentStreams")
def max_concurrent_streams(self) -> Optional[pulumi.Input[int]]:
"""
Maximum number of concurrent streams.
"""
return pulumi.get(self, "max_concurrent_streams")
@max_concurrent_streams.setter
def max_concurrent_streams(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "max_concurrent_streams", value)
@pulumi.input_type
class AlbLoadBalancerListenerTlsSniHandlerArgs:
def __init__(__self__, *,
handler: pulumi.Input['AlbLoadBalancerListenerTlsSniHandlerHandlerArgs'],
name: pulumi.Input[str],
server_names: pulumi.Input[Sequence[pulumi.Input[str]]]):
"""
:param pulumi.Input['AlbLoadBalancerListenerTlsSniHandlerHandlerArgs'] handler: HTTP handler that sets plaintext HTTP router. The structure is documented below.
:param pulumi.Input[str] name: name of SNI match.
:param pulumi.Input[Sequence[pulumi.Input[str]]] server_names: A set of server names.
"""
pulumi.set(__self__, "handler", handler)
pulumi.set(__self__, "name", name)
pulumi.set(__self__, "server_names", server_names)
@property
@pulumi.getter
def handler(self) -> pulumi.Input['AlbLoadBalancerListenerTlsSniHandlerHandlerArgs']:
"""
HTTP handler that sets plaintext HTTP router. The structure is documented below.
"""
return pulumi.get(self, "handler")
@handler.setter
def handler(self, value: pulumi.Input['AlbLoadBalancerListenerTlsSniHandlerHandlerArgs']):
pulumi.set(self, "handler", value)
@property
@pulumi.getter
def name(self) -> pulumi.Input[str]:
"""
name of SNI match.
"""
return pulumi.get(self, "name")
@name.setter
def name(self, value: pulumi.Input[str]):
pulumi.set(self, "name", value)
@property
@pulumi.getter(name="serverNames")
def server_names(self) -> pulumi.Input[Sequence[pulumi.Input[str]]]:
"""
A set of server names.
"""
return pulumi.get(self, "server_names")
@server_names.setter
def server_names(self, value: pulumi.Input[Sequence[pulumi.Input[str]]]):
pulumi.set(self, "server_names", value)
@pulumi.input_type
class AlbLoadBalancerListenerTlsSniHandlerHandlerArgs:
def __init__(__self__, *,
certificate_ids: pulumi.Input[Sequence[pulumi.Input[str]]],
http_handler: Optional[pulumi.Input['AlbLoadBalancerListenerTlsSniHandlerHandlerHttpHandlerArgs']] = None):
"""
:param pulumi.Input[Sequence[pulumi.Input[str]]] certificate_ids: Certificate IDs in the Certificate Manager. Multiple TLS certificates can be associated
with the same context to allow both RSA and ECDSA certificates. Only the first certificate of each type will be used.
:param pulumi.Input['AlbLoadBalancerListenerTlsSniHandlerHandlerHttpHandlerArgs'] http_handler: HTTP handler resource. The structure is documented below.
"""
pulumi.set(__self__, "certificate_ids", certificate_ids)
if http_handler is not None:
pulumi.set(__self__, "http_handler", http_handler)
@property
@pulumi.getter(name="certificateIds")
def certificate_ids(self) -> pulumi.Input[Sequence[pulumi.Input[str]]]:
"""
Certificate IDs in the Certificate Manager. Multiple TLS certificates can be associated
with the same context to allow both RSA and ECDSA certificates. Only the first certificate of each type will be used.
"""
return pulumi.get(self, "certificate_ids")
@certificate_ids.setter
def certificate_ids(self, value: pulumi.Input[Sequence[pulumi.Input[str]]]):
pulumi.set(self, "certificate_ids", value)
@property
@pulumi.getter(name="httpHandler")
def http_handler(self) -> Optional[pulumi.Input['AlbLoadBalancerListenerTlsSniHandlerHandlerHttpHandlerArgs']]:
"""
HTTP handler resource. The structure is documented below.
"""
return pulumi.get(self, "http_handler")
@http_handler.setter
def http_handler(self, value: Optional[pulumi.Input['AlbLoadBalancerListenerTlsSniHandlerHandlerHttpHandlerArgs']]):
pulumi.set(self, "http_handler", value)
@pulumi.input_type
class AlbLoadBalancerListenerTlsSniHandlerHandlerHttpHandlerArgs:
def __init__(__self__, *,
allow_http10: Optional[pulumi.Input[bool]] = None,
http2_options: Optional[pulumi.Input['AlbLoadBalancerListenerTlsSniHandlerHandlerHttpHandlerHttp2OptionsArgs']] = None,
http_router_id: Optional[pulumi.Input[str]] = None):
"""
:param pulumi.Input[bool] allow_http10: If set, will enable only HTTP1 protocol with HTTP1.0 support.
:param pulumi.Input['AlbLoadBalancerListenerTlsSniHandlerHandlerHttpHandlerHttp2OptionsArgs'] http2_options: If set, will enable HTTP2 protocol for the handler. The structure is documented below.
:param pulumi.Input[str] http_router_id: HTTP router id.
"""
if allow_http10 is not None:
pulumi.set(__self__, "allow_http10", allow_http10)
if http2_options is not None:
pulumi.set(__self__, "http2_options", http2_options)
if http_router_id is not None:
pulumi.set(__self__, "http_router_id", http_router_id)
@property
@pulumi.getter(name="allowHttp10")
def allow_http10(self) -> Optional[pulumi.Input[bool]]:
"""
If set, will enable only HTTP1 protocol with HTTP1.0 support.
"""
return pulumi.get(self, "allow_http10")
@allow_http10.setter
def allow_http10(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "allow_http10", value)
@property
@pulumi.getter(name="http2Options")
def http2_options(self) -> Optional[pulumi.Input['AlbLoadBalancerListenerTlsSniHandlerHandlerHttpHandlerHttp2OptionsArgs']]:
"""
If set, will enable HTTP2 protocol for the handler. The structure is documented below.
"""
return pulumi.get(self, "http2_options")
@http2_options.setter
def http2_options(self, value: Optional[pulumi.Input['AlbLoadBalancerListenerTlsSniHandlerHandlerHttpHandlerHttp2OptionsArgs']]):
pulumi.set(self, "http2_options", value)
@property
@pulumi.getter(name="httpRouterId")
def http_router_id(self) -> Optional[pulumi.Input[str]]:
"""
HTTP router id.
"""
return pulumi.get(self, "http_router_id")
@http_router_id.setter
def http_router_id(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "http_router_id", value)
@pulumi.input_type
class AlbLoadBalancerListenerTlsSniHandlerHandlerHttpHandlerHttp2OptionsArgs:
def __init__(__self__, *,
max_concurrent_streams: Optional[pulumi.Input[int]] = None):
"""
:param pulumi.Input[int] max_concurrent_streams: Maximum number of concurrent streams.
"""
if max_concurrent_streams is not None:
pulumi.set(__self__, "max_concurrent_streams", max_concurrent_streams)
@property
@pulumi.getter(name="maxConcurrentStreams")
def max_concurrent_streams(self) -> Optional[pulumi.Input[int]]:
"""
Maximum number of concurrent streams.
"""
return pulumi.get(self, "max_concurrent_streams")
@max_concurrent_streams.setter
def max_concurrent_streams(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "max_concurrent_streams", value)
@pulumi.input_type
class AlbTargetGroupTargetArgs:
def __init__(__self__, *,
ip_address: pulumi.Input[str],
subnet_id: pulumi.Input[str]):
"""
:param pulumi.Input[str] ip_address: IP address of the target.
:param pulumi.Input[str] subnet_id: ID of the subnet that targets are connected to.
All targets in the target group must be connected to the same subnet within a single availability zone.
"""
pulumi.set(__self__, "ip_address", ip_address)
pulumi.set(__self__, "subnet_id", subnet_id)
@property
@pulumi.getter(name="ipAddress")
def ip_address(self) -> pulumi.Input[str]:
"""
IP address of the target.
"""
return pulumi.get(self, "ip_address")
@ip_address.setter
def ip_address(self, value: pulumi.Input[str]):
pulumi.set(self, "ip_address", value)
@property
@pulumi.getter(name="subnetId")
def subnet_id(self) -> pulumi.Input[str]:
"""
ID of the subnet that targets are connected to.
All targets in the target group must be connected to the same subnet within a single availability zone.
"""
return pulumi.get(self, "subnet_id")
@subnet_id.setter
def subnet_id(self, value: pulumi.Input[str]):
pulumi.set(self, "subnet_id", value)
@pulumi.input_type
class AlbVirtualHostModifyRequestHeaderArgs:
def __init__(__self__, *,
append: Optional[pulumi.Input[str]] = None,
name: Optional[pulumi.Input[str]] = None,
remove: Optional[pulumi.Input[bool]] = None,
replace: Optional[pulumi.Input[str]] = None):
"""
:param pulumi.Input[str] append: Append string to the header value.
:param pulumi.Input[str] name: name of the route.
:param pulumi.Input[bool] remove: If set, remove the header.
:param pulumi.Input[str] replace: New value for a header. Header values support the following
[formatters](https://www.envoyproxy.io/docs/envoy/latest/configuration/http/http_conn_man/headers#custom-request-response-headers).
"""
if append is not None:
pulumi.set(__self__, "append", append)
if name is not None:
pulumi.set(__self__, "name", name)
if remove is not None:
pulumi.set(__self__, "remove", remove)
if replace is not None:
pulumi.set(__self__, "replace", replace)
@property
@pulumi.getter
def append(self) -> Optional[pulumi.Input[str]]:
"""
Append string to the header value.
"""
return pulumi.get(self, "append")
@append.setter
def append(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "append", value)
@property
@pulumi.getter
def name(self) -> Optional[pulumi.Input[str]]:
"""
name of the route.
"""
return pulumi.get(self, "name")
@name.setter
def name(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "name", value)
@property
@pulumi.getter
def remove(self) -> Optional[pulumi.Input[bool]]:
"""
If set, remove the header.
"""
return pulumi.get(self, "remove")
@remove.setter
def remove(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "remove", value)
@property
@pulumi.getter
def replace(self) -> Optional[pulumi.Input[str]]:
"""
New value for a header. Header values support the following
[formatters](https://www.envoyproxy.io/docs/envoy/latest/configuration/http/http_conn_man/headers#custom-request-response-headers).
"""
return pulumi.get(self, "replace")
@replace.setter
def replace(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "replace", value)
@pulumi.input_type
class AlbVirtualHostModifyResponseHeaderArgs:
def __init__(__self__, *,
append: Optional[pulumi.Input[str]] = None,
name: Optional[pulumi.Input[str]] = None,
remove: Optional[pulumi.Input[bool]] = None,
replace: Optional[pulumi.Input[str]] = None):
"""
:param pulumi.Input[str] append: Append string to the header value.
:param pulumi.Input[str] name: name of the route.
:param pulumi.Input[bool] remove: If set, remove the header.
:param pulumi.Input[str] replace: New value for a header. Header values support the following
[formatters](https://www.envoyproxy.io/docs/envoy/latest/configuration/http/http_conn_man/headers#custom-request-response-headers).
"""
if append is not None:
pulumi.set(__self__, "append", append)
if name is not None:
pulumi.set(__self__, "name", name)
if remove is not None:
pulumi.set(__self__, "remove", remove)
if replace is not None:
pulumi.set(__self__, "replace", replace)
@property
@pulumi.getter
def append(self) -> Optional[pulumi.Input[str]]:
"""
Append string to the header value.
"""
return pulumi.get(self, "append")
@append.setter
def append(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "append", value)
@property
@pulumi.getter
def name(self) -> Optional[pulumi.Input[str]]:
"""
name of the route.
"""
return pulumi.get(self, "name")
@name.setter
def name(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "name", value)
@property
@pulumi.getter
def remove(self) -> Optional[pulumi.Input[bool]]:
"""
If set, remove the header.
"""
return pulumi.get(self, "remove")
@remove.setter
def remove(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "remove", value)
@property
@pulumi.getter
def replace(self) -> Optional[pulumi.Input[str]]:
"""
New value for a header. Header values support the following
[formatters](https://www.envoyproxy.io/docs/envoy/latest/configuration/http/http_conn_man/headers#custom-request-response-headers).
"""
return pulumi.get(self, "replace")
@replace.setter
def replace(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "replace", value)
@pulumi.input_type
class AlbVirtualHostRouteArgs:
def __init__(__self__, *,
grpc_route: Optional[pulumi.Input['AlbVirtualHostRouteGrpcRouteArgs']] = None,
http_route: Optional[pulumi.Input['AlbVirtualHostRouteHttpRouteArgs']] = None,
name: Optional[pulumi.Input[str]] = None):
"""
:param pulumi.Input['AlbVirtualHostRouteGrpcRouteArgs'] grpc_route: GRPC route resource. The structure is documented below.
:param pulumi.Input['AlbVirtualHostRouteHttpRouteArgs'] http_route: HTTP route resource. The structure is documented below.
:param pulumi.Input[str] name: name of the route.
"""
if grpc_route is not None:
pulumi.set(__self__, "grpc_route", grpc_route)
if http_route is not None:
pulumi.set(__self__, "http_route", http_route)
if name is not None:
pulumi.set(__self__, "name", name)
@property
@pulumi.getter(name="grpcRoute")
def grpc_route(self) -> Optional[pulumi.Input['AlbVirtualHostRouteGrpcRouteArgs']]:
"""
GRPC route resource. The structure is documented below.
"""
return pulumi.get(self, "grpc_route")
@grpc_route.setter
def grpc_route(self, value: Optional[pulumi.Input['AlbVirtualHostRouteGrpcRouteArgs']]):
pulumi.set(self, "grpc_route", value)
@property
@pulumi.getter(name="httpRoute")
def http_route(self) -> Optional[pulumi.Input['AlbVirtualHostRouteHttpRouteArgs']]:
"""
HTTP route resource. The structure is documented below.
"""
return pulumi.get(self, "http_route")
@http_route.setter
def http_route(self, value: Optional[pulumi.Input['AlbVirtualHostRouteHttpRouteArgs']]):
pulumi.set(self, "http_route", value)
@property
@pulumi.getter
def name(self) -> Optional[pulumi.Input[str]]:
"""
name of the route.
"""
return pulumi.get(self, "name")
@name.setter
def name(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "name", value)
@pulumi.input_type
class AlbVirtualHostRouteGrpcRouteArgs:
def __init__(__self__, *,
grpc_matches: Optional[pulumi.Input[Sequence[pulumi.Input['AlbVirtualHostRouteGrpcRouteGrpcMatchArgs']]]] = None,
grpc_route_action: Optional[pulumi.Input['AlbVirtualHostRouteGrpcRouteGrpcRouteActionArgs']] = None,
grpc_status_response_action: Optional[pulumi.Input['AlbVirtualHostRouteGrpcRouteGrpcStatusResponseActionArgs']] = None):
"""
:param pulumi.Input[Sequence[pulumi.Input['AlbVirtualHostRouteGrpcRouteGrpcMatchArgs']]] grpc_matches: Checks "/" prefix by default. The structure is documented below.
:param pulumi.Input['AlbVirtualHostRouteGrpcRouteGrpcRouteActionArgs'] grpc_route_action: GRPC route action resource. The structure is documented below.
:param pulumi.Input['AlbVirtualHostRouteGrpcRouteGrpcStatusResponseActionArgs'] grpc_status_response_action: GRPC status response action resource. The structure is documented below.
"""
if grpc_matches is not None:
pulumi.set(__self__, "grpc_matches", grpc_matches)
if grpc_route_action is not None:
pulumi.set(__self__, "grpc_route_action", grpc_route_action)
if grpc_status_response_action is not None:
pulumi.set(__self__, "grpc_status_response_action", grpc_status_response_action)
@property
@pulumi.getter(name="grpcMatches")
def grpc_matches(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['AlbVirtualHostRouteGrpcRouteGrpcMatchArgs']]]]:
"""
Checks "/" prefix by default. The structure is documented below.
"""
return pulumi.get(self, "grpc_matches")
@grpc_matches.setter
def grpc_matches(self, value: Optional[pulumi.Input[Sequence[pulumi.Input['AlbVirtualHostRouteGrpcRouteGrpcMatchArgs']]]]):
pulumi.set(self, "grpc_matches", value)
@property
@pulumi.getter(name="grpcRouteAction")
def grpc_route_action(self) -> Optional[pulumi.Input['AlbVirtualHostRouteGrpcRouteGrpcRouteActionArgs']]:
"""
GRPC route action resource. The structure is documented below.
"""
return pulumi.get(self, "grpc_route_action")
@grpc_route_action.setter
def grpc_route_action(self, value: Optional[pulumi.Input['AlbVirtualHostRouteGrpcRouteGrpcRouteActionArgs']]):
pulumi.set(self, "grpc_route_action", value)
@property
@pulumi.getter(name="grpcStatusResponseAction")
def grpc_status_response_action(self) -> Optional[pulumi.Input['AlbVirtualHostRouteGrpcRouteGrpcStatusResponseActionArgs']]:
"""
GRPC status response action resource. The structure is documented below.
"""
return pulumi.get(self, "grpc_status_response_action")
@grpc_status_response_action.setter
def grpc_status_response_action(self, value: Optional[pulumi.Input['AlbVirtualHostRouteGrpcRouteGrpcStatusResponseActionArgs']]):
pulumi.set(self, "grpc_status_response_action", value)
@pulumi.input_type
class AlbVirtualHostRouteGrpcRouteGrpcMatchArgs:
def __init__(__self__, *,
fqmn: Optional[pulumi.Input['AlbVirtualHostRouteGrpcRouteGrpcMatchFqmnArgs']] = None):
"""
:param pulumi.Input['AlbVirtualHostRouteGrpcRouteGrpcMatchFqmnArgs'] fqmn: If not set, all services/methods are assumed. The structure is documented below.
"""
if fqmn is not None:
pulumi.set(__self__, "fqmn", fqmn)
@property
@pulumi.getter
def fqmn(self) -> Optional[pulumi.Input['AlbVirtualHostRouteGrpcRouteGrpcMatchFqmnArgs']]:
"""
If not set, all services/methods are assumed. The structure is documented below.
"""
return pulumi.get(self, "fqmn")
@fqmn.setter
def fqmn(self, value: Optional[pulumi.Input['AlbVirtualHostRouteGrpcRouteGrpcMatchFqmnArgs']]):
pulumi.set(self, "fqmn", value)
@pulumi.input_type
class AlbVirtualHostRouteGrpcRouteGrpcMatchFqmnArgs:
def __init__(__self__, *,
exact: Optional[pulumi.Input[str]] = None,
prefix: Optional[pulumi.Input[str]] = None):
if exact is not None:
pulumi.set(__self__, "exact", exact)
if prefix is not None:
pulumi.set(__self__, "prefix", prefix)
@property
@pulumi.getter
def exact(self) -> Optional[pulumi.Input[str]]:
return pulumi.get(self, "exact")
@exact.setter
def exact(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "exact", value)
@property
@pulumi.getter
def prefix(self) -> Optional[pulumi.Input[str]]:
return pulumi.get(self, "prefix")
@prefix.setter
def prefix(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "prefix", value)
@pulumi.input_type
class AlbVirtualHostRouteGrpcRouteGrpcRouteActionArgs:
def __init__(__self__, *,
backend_group_id: pulumi.Input[str],
auto_host_rewrite: Optional[pulumi.Input[bool]] = None,
host_rewrite: Optional[pulumi.Input[str]] = None,
idle_timeout: Optional[pulumi.Input[str]] = None,
max_timeout: Optional[pulumi.Input[str]] = None):
"""
:param pulumi.Input[str] backend_group_id: Backend group to route requests.
:param pulumi.Input[bool] auto_host_rewrite: If set, will automatically rewrite host.
:param pulumi.Input[str] host_rewrite: Host rewrite specifier.
:param pulumi.Input[str] idle_timeout: Specifies the idle timeout (time without any data transfer for the active request) for the
route. It is useful for streaming scenarios - one should set idle_timeout to something meaningful and max_timeout
to the maximum time the stream is allowed to be alive. If not specified, there is no
per-route idle timeout.
:param pulumi.Input[str] max_timeout: Lower timeout may be specified by the client (using grpc-timeout header). If not set, default is
60 seconds.
"""
pulumi.set(__self__, "backend_group_id", backend_group_id)
if auto_host_rewrite is not None:
pulumi.set(__self__, "auto_host_rewrite", auto_host_rewrite)
if host_rewrite is not None:
pulumi.set(__self__, "host_rewrite", host_rewrite)
if idle_timeout is not None:
pulumi.set(__self__, "idle_timeout", idle_timeout)
if max_timeout is not None:
pulumi.set(__self__, "max_timeout", max_timeout)
@property
@pulumi.getter(name="backendGroupId")
def backend_group_id(self) -> pulumi.Input[str]:
"""
Backend group to route requests.
"""
return pulumi.get(self, "backend_group_id")
@backend_group_id.setter
def backend_group_id(self, value: pulumi.Input[str]):
pulumi.set(self, "backend_group_id", value)
@property
@pulumi.getter(name="autoHostRewrite")
def auto_host_rewrite(self) -> Optional[pulumi.Input[bool]]:
"""
If set, will automatically rewrite host.
"""
return pulumi.get(self, "auto_host_rewrite")
@auto_host_rewrite.setter
def auto_host_rewrite(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "auto_host_rewrite", value)
@property
@pulumi.getter(name="hostRewrite")
def host_rewrite(self) -> Optional[pulumi.Input[str]]:
"""
Host rewrite specifier.
"""
return pulumi.get(self, "host_rewrite")
@host_rewrite.setter
def host_rewrite(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "host_rewrite", value)
@property
@pulumi.getter(name="idleTimeout")
def idle_timeout(self) -> Optional[pulumi.Input[str]]:
"""
Specifies the idle timeout (time without any data transfer for the active request) for the
route. It is useful for streaming scenarios - one should set idle_timeout to something meaningful and max_timeout
to the maximum time the stream is allowed to be alive. If not specified, there is no
per-route idle timeout.
"""
return pulumi.get(self, "idle_timeout")
@idle_timeout.setter
def idle_timeout(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "idle_timeout", value)
@property
@pulumi.getter(name="maxTimeout")
def max_timeout(self) -> Optional[pulumi.Input[str]]:
"""
Lower timeout may be specified by the client (using grpc-timeout header). If not set, default is
60 seconds.
"""
return pulumi.get(self, "max_timeout")
@max_timeout.setter
def max_timeout(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "max_timeout", value)
@pulumi.input_type
class AlbVirtualHostRouteGrpcRouteGrpcStatusResponseActionArgs:
def __init__(__self__, *,
status: Optional[pulumi.Input[str]] = None):
"""
:param pulumi.Input[str] status: The status of the response. Supported values are: ok, invalid_argumet, not_found,
permission_denied, unauthenticated, unimplemented, internal, unavailable.
"""
if status is not None:
pulumi.set(__self__, "status", status)
@property
@pulumi.getter
def status(self) -> Optional[pulumi.Input[str]]:
"""
The status of the response. Supported values are: ok, invalid_argumet, not_found,
permission_denied, unauthenticated, unimplemented, internal, unavailable.
"""
return pulumi.get(self, "status")
@status.setter
def status(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "status", value)
@pulumi.input_type
class AlbVirtualHostRouteHttpRouteArgs:
def __init__(__self__, *,
direct_response_action: Optional[pulumi.Input['AlbVirtualHostRouteHttpRouteDirectResponseActionArgs']] = None,
http_matches: Optional[pulumi.Input[Sequence[pulumi.Input['AlbVirtualHostRouteHttpRouteHttpMatchArgs']]]] = None,
http_route_action: Optional[pulumi.Input['AlbVirtualHostRouteHttpRouteHttpRouteActionArgs']] = None,
redirect_action: Optional[pulumi.Input['AlbVirtualHostRouteHttpRouteRedirectActionArgs']] = None):
"""
:param pulumi.Input['AlbVirtualHostRouteHttpRouteDirectResponseActionArgs'] direct_response_action: Direct response action resource. The structure is documented below.
:param pulumi.Input[Sequence[pulumi.Input['AlbVirtualHostRouteHttpRouteHttpMatchArgs']]] http_matches: Checks "/" prefix by default. The structure is documented below.
:param pulumi.Input['AlbVirtualHostRouteHttpRouteHttpRouteActionArgs'] http_route_action: HTTP route action resource. The structure is documented below.
:param pulumi.Input['AlbVirtualHostRouteHttpRouteRedirectActionArgs'] redirect_action: Redirect action resource. The structure is documented below.
"""
if direct_response_action is not None:
pulumi.set(__self__, "direct_response_action", direct_response_action)
if http_matches is not None:
pulumi.set(__self__, "http_matches", http_matches)
if http_route_action is not None:
pulumi.set(__self__, "http_route_action", http_route_action)
if redirect_action is not None:
pulumi.set(__self__, "redirect_action", redirect_action)
@property
@pulumi.getter(name="directResponseAction")
def direct_response_action(self) -> Optional[pulumi.Input['AlbVirtualHostRouteHttpRouteDirectResponseActionArgs']]:
"""
Direct response action resource. The structure is documented below.
"""
return pulumi.get(self, "direct_response_action")
@direct_response_action.setter
def direct_response_action(self, value: Optional[pulumi.Input['AlbVirtualHostRouteHttpRouteDirectResponseActionArgs']]):
pulumi.set(self, "direct_response_action", value)
@property
@pulumi.getter(name="httpMatches")
def http_matches(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['AlbVirtualHostRouteHttpRouteHttpMatchArgs']]]]:
"""
Checks "/" prefix by default. The structure is documented below.
"""
return pulumi.get(self, "http_matches")
@http_matches.setter
def http_matches(self, value: Optional[pulumi.Input[Sequence[pulumi.Input['AlbVirtualHostRouteHttpRouteHttpMatchArgs']]]]):
pulumi.set(self, "http_matches", value)
@property
@pulumi.getter(name="httpRouteAction")
def http_route_action(self) -> Optional[pulumi.Input['AlbVirtualHostRouteHttpRouteHttpRouteActionArgs']]:
"""
HTTP route action resource. The structure is documented below.
"""
return pulumi.get(self, "http_route_action")
@http_route_action.setter
def http_route_action(self, value: Optional[pulumi.Input['AlbVirtualHostRouteHttpRouteHttpRouteActionArgs']]):
pulumi.set(self, "http_route_action", value)
@property
@pulumi.getter(name="redirectAction")
def redirect_action(self) -> Optional[pulumi.Input['AlbVirtualHostRouteHttpRouteRedirectActionArgs']]:
"""
Redirect action resource. The structure is documented below.
"""
return pulumi.get(self, "redirect_action")
@redirect_action.setter
def redirect_action(self, value: Optional[pulumi.Input['AlbVirtualHostRouteHttpRouteRedirectActionArgs']]):
pulumi.set(self, "redirect_action", value)
@pulumi.input_type
class AlbVirtualHostRouteHttpRouteDirectResponseActionArgs:
def __init__(__self__, *,
body: Optional[pulumi.Input[str]] = None,
status: Optional[pulumi.Input[int]] = None):
"""
:param pulumi.Input[str] body: Response body text.
:param pulumi.Input[int] status: The status of the response. Supported values are: ok, invalid_argumet, not_found,
permission_denied, unauthenticated, unimplemented, internal, unavailable.
"""
if body is not None:
pulumi.set(__self__, "body", body)
if status is not None:
pulumi.set(__self__, "status", status)
@property
@pulumi.getter
def body(self) -> Optional[pulumi.Input[str]]:
"""
Response body text.
"""
return pulumi.get(self, "body")
@body.setter
def body(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "body", value)
@property
@pulumi.getter
def status(self) -> Optional[pulumi.Input[int]]:
"""
The status of the response. Supported values are: ok, invalid_argumet, not_found,
permission_denied, unauthenticated, unimplemented, internal, unavailable.
"""
return pulumi.get(self, "status")
@status.setter
def status(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "status", value)
@pulumi.input_type
class AlbVirtualHostRouteHttpRouteHttpMatchArgs:
def __init__(__self__, *,
http_methods: Optional[pulumi.Input[Sequence[Any]]] = None,
path: Optional[pulumi.Input['AlbVirtualHostRouteHttpRouteHttpMatchPathArgs']] = None):
"""
:param pulumi.Input[Sequence[Any]] http_methods: List of methods(strings).
:param pulumi.Input['AlbVirtualHostRouteHttpRouteHttpMatchPathArgs'] path: If not set, '/' is assumed. The structure is documented below.
"""
if http_methods is not None:
pulumi.set(__self__, "http_methods", http_methods)
if path is not None:
pulumi.set(__self__, "path", path)
@property
@pulumi.getter(name="httpMethods")
def http_methods(self) -> Optional[pulumi.Input[Sequence[Any]]]:
"""
List of methods(strings).
"""
return pulumi.get(self, "http_methods")
@http_methods.setter
def http_methods(self, value: Optional[pulumi.Input[Sequence[Any]]]):
pulumi.set(self, "http_methods", value)
@property
@pulumi.getter
def path(self) -> Optional[pulumi.Input['AlbVirtualHostRouteHttpRouteHttpMatchPathArgs']]:
"""
If not set, '/' is assumed. The structure is documented below.
"""
return pulumi.get(self, "path")
@path.setter
def path(self, value: Optional[pulumi.Input['AlbVirtualHostRouteHttpRouteHttpMatchPathArgs']]):
pulumi.set(self, "path", value)
@pulumi.input_type
class AlbVirtualHostRouteHttpRouteHttpMatchPathArgs:
def __init__(__self__, *,
exact: Optional[pulumi.Input[str]] = None,
prefix: Optional[pulumi.Input[str]] = None):
if exact is not None:
pulumi.set(__self__, "exact", exact)
if prefix is not None:
pulumi.set(__self__, "prefix", prefix)
@property
@pulumi.getter
def exact(self) -> Optional[pulumi.Input[str]]:
return pulumi.get(self, "exact")
@exact.setter
def exact(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "exact", value)
@property
@pulumi.getter
def prefix(self) -> Optional[pulumi.Input[str]]:
return pulumi.get(self, "prefix")
@prefix.setter
def prefix(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "prefix", value)
@pulumi.input_type
class AlbVirtualHostRouteHttpRouteHttpRouteActionArgs:
def __init__(__self__, *,
backend_group_id: pulumi.Input[str],
auto_host_rewrite: Optional[pulumi.Input[bool]] = None,
host_rewrite: Optional[pulumi.Input[str]] = None,
idle_timeout: Optional[pulumi.Input[str]] = None,
prefix_rewrite: Optional[pulumi.Input[str]] = None,
timeout: Optional[pulumi.Input[str]] = None,
upgrade_types: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None):
"""
:param pulumi.Input[str] backend_group_id: Backend group to route requests.
:param pulumi.Input[bool] auto_host_rewrite: If set, will automatically rewrite host.
:param pulumi.Input[str] host_rewrite: Host rewrite specifier.
:param pulumi.Input[str] idle_timeout: Specifies the idle timeout (time without any data transfer for the active request) for the
route. It is useful for streaming scenarios - one should set idle_timeout to something meaningful and max_timeout
to the maximum time the stream is allowed to be alive. If not specified, there is no
per-route idle timeout.
:param pulumi.Input[str] prefix_rewrite: If not empty, matched path prefix will be replaced by this value.
:param pulumi.Input[str] timeout: Specifies the request timeout (overall time request processing is allowed to take) for the
route. If not set, default is 60 seconds.
:param pulumi.Input[Sequence[pulumi.Input[str]]] upgrade_types: List of upgrade types. Only specified upgrade types will be allowed. For example,
"websocket".
"""
pulumi.set(__self__, "backend_group_id", backend_group_id)
if auto_host_rewrite is not None:
pulumi.set(__self__, "auto_host_rewrite", auto_host_rewrite)
if host_rewrite is not None:
pulumi.set(__self__, "host_rewrite", host_rewrite)
if idle_timeout is not None:
pulumi.set(__self__, "idle_timeout", idle_timeout)
if prefix_rewrite is not None:
pulumi.set(__self__, "prefix_rewrite", prefix_rewrite)
if timeout is not None:
pulumi.set(__self__, "timeout", timeout)
if upgrade_types is not None:
pulumi.set(__self__, "upgrade_types", upgrade_types)
@property
@pulumi.getter(name="backendGroupId")
def backend_group_id(self) -> pulumi.Input[str]:
"""
Backend group to route requests.
"""
return pulumi.get(self, "backend_group_id")
@backend_group_id.setter
def backend_group_id(self, value: pulumi.Input[str]):
pulumi.set(self, "backend_group_id", value)
@property
@pulumi.getter(name="autoHostRewrite")
def auto_host_rewrite(self) -> Optional[pulumi.Input[bool]]:
"""
If set, will automatically rewrite host.
"""
return pulumi.get(self, "auto_host_rewrite")
@auto_host_rewrite.setter
def auto_host_rewrite(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "auto_host_rewrite", value)
@property
@pulumi.getter(name="hostRewrite")
def host_rewrite(self) -> Optional[pulumi.Input[str]]:
"""
Host rewrite specifier.
"""
return pulumi.get(self, "host_rewrite")
@host_rewrite.setter
def host_rewrite(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "host_rewrite", value)
@property
@pulumi.getter(name="idleTimeout")
def idle_timeout(self) -> Optional[pulumi.Input[str]]:
"""
Specifies the idle timeout (time without any data transfer for the active request) for the
route. It is useful for streaming scenarios - one should set idle_timeout to something meaningful and max_timeout
to the maximum time the stream is allowed to be alive. If not specified, there is no
per-route idle timeout.
"""
return pulumi.get(self, "idle_timeout")
@idle_timeout.setter
def idle_timeout(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "idle_timeout", value)
@property
@pulumi.getter(name="prefixRewrite")
def prefix_rewrite(self) -> Optional[pulumi.Input[str]]:
"""
If not empty, matched path prefix will be replaced by this value.
"""
return pulumi.get(self, "prefix_rewrite")
@prefix_rewrite.setter
def prefix_rewrite(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "prefix_rewrite", value)
@property
@pulumi.getter
def timeout(self) -> Optional[pulumi.Input[str]]:
"""
Specifies the request timeout (overall time request processing is allowed to take) for the
route. If not set, default is 60 seconds.
"""
return pulumi.get(self, "timeout")
@timeout.setter
def timeout(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "timeout", value)
@property
@pulumi.getter(name="upgradeTypes")
def upgrade_types(self) -> Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]:
"""
List of upgrade types. Only specified upgrade types will be allowed. For example,
"websocket".
"""
return pulumi.get(self, "upgrade_types")
@upgrade_types.setter
def upgrade_types(self, value: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]):
pulumi.set(self, "upgrade_types", value)
@pulumi.input_type
class AlbVirtualHostRouteHttpRouteRedirectActionArgs:
def __init__(__self__, *,
remove_query: Optional[pulumi.Input[bool]] = None,
replace_host: Optional[pulumi.Input[str]] = None,
replace_path: Optional[pulumi.Input[str]] = None,
replace_port: Optional[pulumi.Input[int]] = None,
replace_prefix: Optional[pulumi.Input[str]] = None,
replace_scheme: Optional[pulumi.Input[str]] = None,
response_code: Optional[pulumi.Input[str]] = None):
"""
:param pulumi.Input[str] replace_host: Replaces hostname.
:param pulumi.Input[str] replace_path: Replace path.
:param pulumi.Input[int] replace_port: Replaces port.
:param pulumi.Input[str] replace_prefix: Replace only matched prefix. Example:<br/> match:{ prefix_match: "/some" } <br/>
redirect: { replace_prefix: "/other" } <br/> will redirect "/something" to "/otherthing".
:param pulumi.Input[str] replace_scheme: Replaces scheme. If the original scheme is `http` or `https`, will also remove the
80 or 443 port, if present.
:param pulumi.Input[str] response_code: The HTTP status code to use in the redirect response. Supported values are:
moved_permanently, found, see_other, temporary_redirect, permanent_redirect.
"""
if remove_query is not None:
pulumi.set(__self__, "remove_query", remove_query)
if replace_host is not None:
pulumi.set(__self__, "replace_host", replace_host)
if replace_path is not None:
pulumi.set(__self__, "replace_path", replace_path)
if replace_port is not None:
pulumi.set(__self__, "replace_port", replace_port)
if replace_prefix is not None:
pulumi.set(__self__, "replace_prefix", replace_prefix)
if replace_scheme is not None:
pulumi.set(__self__, "replace_scheme", replace_scheme)
if response_code is not None:
pulumi.set(__self__, "response_code", response_code)
@property
@pulumi.getter(name="removeQuery")
def remove_query(self) -> Optional[pulumi.Input[bool]]:
return pulumi.get(self, "remove_query")
@remove_query.setter
def remove_query(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "remove_query", value)
@property
@pulumi.getter(name="replaceHost")
def replace_host(self) -> Optional[pulumi.Input[str]]:
"""
Replaces hostname.
"""
return pulumi.get(self, "replace_host")
@replace_host.setter
def replace_host(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "replace_host", value)
@property
@pulumi.getter(name="replacePath")
def replace_path(self) -> Optional[pulumi.Input[str]]:
"""
Replace path.
"""
return pulumi.get(self, "replace_path")
@replace_path.setter
def replace_path(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "replace_path", value)
@property
@pulumi.getter(name="replacePort")
def replace_port(self) -> Optional[pulumi.Input[int]]:
"""
Replaces port.
"""
return pulumi.get(self, "replace_port")
@replace_port.setter
def replace_port(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "replace_port", value)
@property
@pulumi.getter(name="replacePrefix")
def replace_prefix(self) -> Optional[pulumi.Input[str]]:
"""
Replace only matched prefix. Example:<br/> match:{ prefix_match: "/some" } <br/>
redirect: { replace_prefix: "/other" } <br/> will redirect "/something" to "/otherthing".
"""
return pulumi.get(self, "replace_prefix")
@replace_prefix.setter
def replace_prefix(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "replace_prefix", value)
@property
@pulumi.getter(name="replaceScheme")
def replace_scheme(self) -> Optional[pulumi.Input[str]]:
"""
Replaces scheme. If the original scheme is `http` or `https`, will also remove the
80 or 443 port, if present.
"""
return pulumi.get(self, "replace_scheme")
@replace_scheme.setter
def replace_scheme(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "replace_scheme", value)
@property
@pulumi.getter(name="responseCode")
def response_code(self) -> Optional[pulumi.Input[str]]:
"""
The HTTP status code to use in the redirect response. Supported values are:
moved_permanently, found, see_other, temporary_redirect, permanent_redirect.
"""
return pulumi.get(self, "response_code")
@response_code.setter
def response_code(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "response_code", value)
@pulumi.input_type
class ComputeDiskDiskPlacementPolicyArgs:
def __init__(__self__, *,
disk_placement_group_id: pulumi.Input[str]):
"""
:param pulumi.Input[str] disk_placement_group_id: Specifies Disk Placement Group id.
"""
pulumi.set(__self__, "disk_placement_group_id", disk_placement_group_id)
@property
@pulumi.getter(name="diskPlacementGroupId")
def disk_placement_group_id(self) -> pulumi.Input[str]:
"""
Specifies Disk Placement Group id.
"""
return pulumi.get(self, "disk_placement_group_id")
@disk_placement_group_id.setter
def disk_placement_group_id(self, value: pulumi.Input[str]):
pulumi.set(self, "disk_placement_group_id", value)
@pulumi.input_type
class ComputeInstanceBootDiskArgs:
def __init__(__self__, *,
auto_delete: Optional[pulumi.Input[bool]] = None,
device_name: Optional[pulumi.Input[str]] = None,
disk_id: Optional[pulumi.Input[str]] = None,
initialize_params: Optional[pulumi.Input['ComputeInstanceBootDiskInitializeParamsArgs']] = None,
mode: Optional[pulumi.Input[str]] = None):
"""
:param pulumi.Input[bool] auto_delete: Whether the disk is auto-deleted when the instance
is deleted. The default value is false.
:param pulumi.Input[str] device_name: Name that can be used to access an attached disk
under `/dev/disk/by-id/`.
:param pulumi.Input[str] disk_id: ID of the disk that is attached to the instance.
:param pulumi.Input['ComputeInstanceBootDiskInitializeParamsArgs'] initialize_params: Parameters for a new disk that will be created
alongside the new instance. Either `initialize_params` or `disk_id` must be set. The structure is documented below.
:param pulumi.Input[str] mode: Type of access to the disk resource. By default, a disk is attached in `READ_WRITE` mode.
"""
if auto_delete is not None:
pulumi.set(__self__, "auto_delete", auto_delete)
if device_name is not None:
pulumi.set(__self__, "device_name", device_name)
if disk_id is not None:
pulumi.set(__self__, "disk_id", disk_id)
if initialize_params is not None:
pulumi.set(__self__, "initialize_params", initialize_params)
if mode is not None:
pulumi.set(__self__, "mode", mode)
@property
@pulumi.getter(name="autoDelete")
def auto_delete(self) -> Optional[pulumi.Input[bool]]:
"""
Whether the disk is auto-deleted when the instance
is deleted. The default value is false.
"""
return pulumi.get(self, "auto_delete")
@auto_delete.setter
def auto_delete(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "auto_delete", value)
@property
@pulumi.getter(name="deviceName")
def device_name(self) -> Optional[pulumi.Input[str]]:
"""
Name that can be used to access an attached disk
under `/dev/disk/by-id/`.
"""
return pulumi.get(self, "device_name")
@device_name.setter
def device_name(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "device_name", value)
@property
@pulumi.getter(name="diskId")
def disk_id(self) -> Optional[pulumi.Input[str]]:
"""
ID of the disk that is attached to the instance.
"""
return pulumi.get(self, "disk_id")
@disk_id.setter
def disk_id(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "disk_id", value)
@property
@pulumi.getter(name="initializeParams")
def initialize_params(self) -> Optional[pulumi.Input['ComputeInstanceBootDiskInitializeParamsArgs']]:
"""
Parameters for a new disk that will be created
alongside the new instance. Either `initialize_params` or `disk_id` must be set. The structure is documented below.
"""
return pulumi.get(self, "initialize_params")
@initialize_params.setter
def initialize_params(self, value: Optional[pulumi.Input['ComputeInstanceBootDiskInitializeParamsArgs']]):
pulumi.set(self, "initialize_params", value)
@property
@pulumi.getter
def mode(self) -> Optional[pulumi.Input[str]]:
"""
Type of access to the disk resource. By default, a disk is attached in `READ_WRITE` mode.
"""
return pulumi.get(self, "mode")
@mode.setter
def mode(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "mode", value)
@pulumi.input_type
class ComputeInstanceBootDiskInitializeParamsArgs:
def __init__(__self__, *,
description: Optional[pulumi.Input[str]] = None,
image_id: Optional[pulumi.Input[str]] = None,
name: Optional[pulumi.Input[str]] = None,
size: Optional[pulumi.Input[int]] = None,
snapshot_id: Optional[pulumi.Input[str]] = None,
type: Optional[pulumi.Input[str]] = None):
"""
:param pulumi.Input[str] description: Description of the boot disk.
:param pulumi.Input[str] image_id: A disk image to initialize this disk from.
:param pulumi.Input[str] name: Name of the boot disk.
:param pulumi.Input[int] size: Size of the disk in GB.
:param pulumi.Input[str] snapshot_id: A snapshot to initialize this disk from.
:param pulumi.Input[str] type: Disk type.
"""
if description is not None:
pulumi.set(__self__, "description", description)
if image_id is not None:
pulumi.set(__self__, "image_id", image_id)
if name is not None:
pulumi.set(__self__, "name", name)
if size is not None:
pulumi.set(__self__, "size", size)
if snapshot_id is not None:
pulumi.set(__self__, "snapshot_id", snapshot_id)
if type is not None:
pulumi.set(__self__, "type", type)
@property
@pulumi.getter
def description(self) -> Optional[pulumi.Input[str]]:
"""
Description of the boot disk.
"""
return pulumi.get(self, "description")
@description.setter
def description(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "description", value)
@property
@pulumi.getter(name="imageId")
def image_id(self) -> Optional[pulumi.Input[str]]:
"""
A disk image to initialize this disk from.
"""
return pulumi.get(self, "image_id")
@image_id.setter
def image_id(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "image_id", value)
@property
@pulumi.getter
def name(self) -> Optional[pulumi.Input[str]]:
"""
Name of the boot disk.
"""
return pulumi.get(self, "name")
@name.setter
def name(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "name", value)
@property
@pulumi.getter
def size(self) -> Optional[pulumi.Input[int]]:
"""
Size of the disk in GB.
"""
return pulumi.get(self, "size")
@size.setter
def size(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "size", value)
@property
@pulumi.getter(name="snapshotId")
def snapshot_id(self) -> Optional[pulumi.Input[str]]:
"""
A snapshot to initialize this disk from.
"""
return pulumi.get(self, "snapshot_id")
@snapshot_id.setter
def snapshot_id(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "snapshot_id", value)
@property
@pulumi.getter
def type(self) -> Optional[pulumi.Input[str]]:
"""
Disk type.
"""
return pulumi.get(self, "type")
@type.setter
def type(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "type", value)
@pulumi.input_type
class ComputeInstanceGroupAllocationPolicyArgs:
def __init__(__self__, *,
zones: pulumi.Input[Sequence[pulumi.Input[str]]]):
"""
:param pulumi.Input[Sequence[pulumi.Input[str]]] zones: A list of availability zones.
"""
pulumi.set(__self__, "zones", zones)
@property
@pulumi.getter
def zones(self) -> pulumi.Input[Sequence[pulumi.Input[str]]]:
"""
A list of availability zones.
"""
return pulumi.get(self, "zones")
@zones.setter
def zones(self, value: pulumi.Input[Sequence[pulumi.Input[str]]]):
pulumi.set(self, "zones", value)
@pulumi.input_type
class ComputeInstanceGroupApplicationLoadBalancerArgs:
def __init__(__self__, *,
max_opening_traffic_duration: Optional[pulumi.Input[int]] = None,
status_message: Optional[pulumi.Input[str]] = None,
target_group_description: Optional[pulumi.Input[str]] = None,
target_group_id: Optional[pulumi.Input[str]] = None,
target_group_labels: Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]] = None,
target_group_name: Optional[pulumi.Input[str]] = None):
"""
:param pulumi.Input[int] max_opening_traffic_duration: Timeout for waiting for the VM to be checked by the load balancer. If the timeout is exceeded, the VM will be turned off based on the deployment policy. Specified in seconds.
:param pulumi.Input[str] status_message: The status message of the instance.
:param pulumi.Input[str] target_group_description: A description of the target group.
:param pulumi.Input[Mapping[str, pulumi.Input[str]]] target_group_labels: A set of key/value label pairs.
:param pulumi.Input[str] target_group_name: The name of the target group.
"""
if max_opening_traffic_duration is not None:
pulumi.set(__self__, "max_opening_traffic_duration", max_opening_traffic_duration)
if status_message is not None:
pulumi.set(__self__, "status_message", status_message)
if target_group_description is not None:
pulumi.set(__self__, "target_group_description", target_group_description)
if target_group_id is not None:
pulumi.set(__self__, "target_group_id", target_group_id)
if target_group_labels is not None:
pulumi.set(__self__, "target_group_labels", target_group_labels)
if target_group_name is not None:
pulumi.set(__self__, "target_group_name", target_group_name)
@property
@pulumi.getter(name="maxOpeningTrafficDuration")
def max_opening_traffic_duration(self) -> Optional[pulumi.Input[int]]:
"""
Timeout for waiting for the VM to be checked by the load balancer. If the timeout is exceeded, the VM will be turned off based on the deployment policy. Specified in seconds.
"""
return pulumi.get(self, "max_opening_traffic_duration")
@max_opening_traffic_duration.setter
def max_opening_traffic_duration(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "max_opening_traffic_duration", value)
@property
@pulumi.getter(name="statusMessage")
def status_message(self) -> Optional[pulumi.Input[str]]:
"""
The status message of the instance.
"""
return pulumi.get(self, "status_message")
@status_message.setter
def status_message(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "status_message", value)
@property
@pulumi.getter(name="targetGroupDescription")
def target_group_description(self) -> Optional[pulumi.Input[str]]:
"""
A description of the target group.
"""
return pulumi.get(self, "target_group_description")
@target_group_description.setter
def target_group_description(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "target_group_description", value)
@property
@pulumi.getter(name="targetGroupId")
def target_group_id(self) -> Optional[pulumi.Input[str]]:
return pulumi.get(self, "target_group_id")
@target_group_id.setter
def target_group_id(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "target_group_id", value)
@property
@pulumi.getter(name="targetGroupLabels")
def target_group_labels(self) -> Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]]:
"""
A set of key/value label pairs.
"""
return pulumi.get(self, "target_group_labels")
@target_group_labels.setter
def target_group_labels(self, value: Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]]):
pulumi.set(self, "target_group_labels", value)
@property
@pulumi.getter(name="targetGroupName")
def target_group_name(self) -> Optional[pulumi.Input[str]]:
"""
The name of the target group.
"""
return pulumi.get(self, "target_group_name")
@target_group_name.setter
def target_group_name(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "target_group_name", value)
@pulumi.input_type
class ComputeInstanceGroupDeployPolicyArgs:
def __init__(__self__, *,
max_expansion: pulumi.Input[int],
max_unavailable: pulumi.Input[int],
max_creating: Optional[pulumi.Input[int]] = None,
max_deleting: Optional[pulumi.Input[int]] = None,
startup_duration: Optional[pulumi.Input[int]] = None,
strategy: Optional[pulumi.Input[str]] = None):
"""
:param pulumi.Input[int] max_expansion: The maximum number of instances that can be temporarily allocated above the group's target size
during the update process.
:param pulumi.Input[int] max_unavailable: The maximum number of running instances that can be taken offline (stopped or deleted) at the same time
during the update process.
:param pulumi.Input[int] max_creating: The maximum number of instances that can be created at the same time.
:param pulumi.Input[int] max_deleting: The maximum number of instances that can be deleted at the same time.
:param pulumi.Input[int] startup_duration: The amount of time in seconds to allow for an instance to start.
Instance will be considered up and running (and start receiving traffic) only after the startup_duration
has elapsed and all health checks are passed.
:param pulumi.Input[str] strategy: Affects the lifecycle of the instance during deployment. If set to `proactive` (default), Instance Groups
can forcefully stop a running instance. If `opportunistic`, Instance Groups does not stop a running instance. Instead,
it will wait until the instance stops itself or becomes unhealthy.
"""
pulumi.set(__self__, "max_expansion", max_expansion)
pulumi.set(__self__, "max_unavailable", max_unavailable)
if max_creating is not None:
pulumi.set(__self__, "max_creating", max_creating)
if max_deleting is not None:
pulumi.set(__self__, "max_deleting", max_deleting)
if startup_duration is not None:
pulumi.set(__self__, "startup_duration", startup_duration)
if strategy is not None:
pulumi.set(__self__, "strategy", strategy)
@property
@pulumi.getter(name="maxExpansion")
def max_expansion(self) -> pulumi.Input[int]:
"""
The maximum number of instances that can be temporarily allocated above the group's target size
during the update process.
"""
return pulumi.get(self, "max_expansion")
@max_expansion.setter
def max_expansion(self, value: pulumi.Input[int]):
pulumi.set(self, "max_expansion", value)
@property
@pulumi.getter(name="maxUnavailable")
def max_unavailable(self) -> pulumi.Input[int]:
"""
The maximum number of running instances that can be taken offline (stopped or deleted) at the same time
during the update process.
"""
return pulumi.get(self, "max_unavailable")
@max_unavailable.setter
def max_unavailable(self, value: pulumi.Input[int]):
pulumi.set(self, "max_unavailable", value)
@property
@pulumi.getter(name="maxCreating")
def max_creating(self) -> Optional[pulumi.Input[int]]:
"""
The maximum number of instances that can be created at the same time.
"""
return pulumi.get(self, "max_creating")
@max_creating.setter
def max_creating(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "max_creating", value)
@property
@pulumi.getter(name="maxDeleting")
def max_deleting(self) -> Optional[pulumi.Input[int]]:
"""
The maximum number of instances that can be deleted at the same time.
"""
return pulumi.get(self, "max_deleting")
@max_deleting.setter
def max_deleting(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "max_deleting", value)
@property
@pulumi.getter(name="startupDuration")
def startup_duration(self) -> Optional[pulumi.Input[int]]:
"""
The amount of time in seconds to allow for an instance to start.
Instance will be considered up and running (and start receiving traffic) only after the startup_duration
has elapsed and all health checks are passed.
"""
return pulumi.get(self, "startup_duration")
@startup_duration.setter
def startup_duration(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "startup_duration", value)
@property
@pulumi.getter
def strategy(self) -> Optional[pulumi.Input[str]]:
"""
Affects the lifecycle of the instance during deployment. If set to `proactive` (default), Instance Groups
can forcefully stop a running instance. If `opportunistic`, Instance Groups does not stop a running instance. Instead,
it will wait until the instance stops itself or becomes unhealthy.
"""
return pulumi.get(self, "strategy")
@strategy.setter
def strategy(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "strategy", value)
@pulumi.input_type
class ComputeInstanceGroupHealthCheckArgs:
def __init__(__self__, *,
healthy_threshold: Optional[pulumi.Input[int]] = None,
http_options: Optional[pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupHealthCheckHttpOptionArgs']]]] = None,
interval: Optional[pulumi.Input[int]] = None,
tcp_options: Optional[pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupHealthCheckTcpOptionArgs']]]] = None,
timeout: Optional[pulumi.Input[int]] = None,
unhealthy_threshold: Optional[pulumi.Input[int]] = None):
"""
:param pulumi.Input[int] healthy_threshold: The number of successful health checks before the managed instance is declared healthy.
:param pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupHealthCheckHttpOptionArgs']]] http_options: HTTP check options. The structure is documented below.
:param pulumi.Input[int] interval: The interval to wait between health checks in seconds.
:param pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupHealthCheckTcpOptionArgs']]] tcp_options: TCP check options. The structure is documented below.
:param pulumi.Input[int] timeout: The length of time to wait for a response before the health check times out in seconds.
:param pulumi.Input[int] unhealthy_threshold: The number of failed health checks before the managed instance is declared unhealthy.
"""
if healthy_threshold is not None:
pulumi.set(__self__, "healthy_threshold", healthy_threshold)
if http_options is not None:
pulumi.set(__self__, "http_options", http_options)
if interval is not None:
pulumi.set(__self__, "interval", interval)
if tcp_options is not None:
pulumi.set(__self__, "tcp_options", tcp_options)
if timeout is not None:
pulumi.set(__self__, "timeout", timeout)
if unhealthy_threshold is not None:
pulumi.set(__self__, "unhealthy_threshold", unhealthy_threshold)
@property
@pulumi.getter(name="healthyThreshold")
def healthy_threshold(self) -> Optional[pulumi.Input[int]]:
"""
The number of successful health checks before the managed instance is declared healthy.
"""
return pulumi.get(self, "healthy_threshold")
@healthy_threshold.setter
def healthy_threshold(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "healthy_threshold", value)
@property
@pulumi.getter(name="httpOptions")
def http_options(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupHealthCheckHttpOptionArgs']]]]:
"""
HTTP check options. The structure is documented below.
"""
return pulumi.get(self, "http_options")
@http_options.setter
def http_options(self, value: Optional[pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupHealthCheckHttpOptionArgs']]]]):
pulumi.set(self, "http_options", value)
@property
@pulumi.getter
def interval(self) -> Optional[pulumi.Input[int]]:
"""
The interval to wait between health checks in seconds.
"""
return pulumi.get(self, "interval")
@interval.setter
def interval(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "interval", value)
@property
@pulumi.getter(name="tcpOptions")
def tcp_options(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupHealthCheckTcpOptionArgs']]]]:
"""
TCP check options. The structure is documented below.
"""
return pulumi.get(self, "tcp_options")
@tcp_options.setter
def tcp_options(self, value: Optional[pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupHealthCheckTcpOptionArgs']]]]):
pulumi.set(self, "tcp_options", value)
@property
@pulumi.getter
def timeout(self) -> Optional[pulumi.Input[int]]:
"""
The length of time to wait for a response before the health check times out in seconds.
"""
return pulumi.get(self, "timeout")
@timeout.setter
def timeout(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "timeout", value)
@property
@pulumi.getter(name="unhealthyThreshold")
def unhealthy_threshold(self) -> Optional[pulumi.Input[int]]:
"""
The number of failed health checks before the managed instance is declared unhealthy.
"""
return pulumi.get(self, "unhealthy_threshold")
@unhealthy_threshold.setter
def unhealthy_threshold(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "unhealthy_threshold", value)
@pulumi.input_type
class ComputeInstanceGroupHealthCheckHttpOptionArgs:
def __init__(__self__, *,
path: pulumi.Input[str],
port: pulumi.Input[int]):
"""
:param pulumi.Input[str] path: The URL path used for health check requests.
:param pulumi.Input[int] port: The port used for TCP health checks.
"""
pulumi.set(__self__, "path", path)
pulumi.set(__self__, "port", port)
@property
@pulumi.getter
def path(self) -> pulumi.Input[str]:
"""
The URL path used for health check requests.
"""
return pulumi.get(self, "path")
@path.setter
def path(self, value: pulumi.Input[str]):
pulumi.set(self, "path", value)
@property
@pulumi.getter
def port(self) -> pulumi.Input[int]:
"""
The port used for TCP health checks.
"""
return pulumi.get(self, "port")
@port.setter
def port(self, value: pulumi.Input[int]):
pulumi.set(self, "port", value)
@pulumi.input_type
class ComputeInstanceGroupHealthCheckTcpOptionArgs:
def __init__(__self__, *,
port: pulumi.Input[int]):
"""
:param pulumi.Input[int] port: The port used for TCP health checks.
"""
pulumi.set(__self__, "port", port)
@property
@pulumi.getter
def port(self) -> pulumi.Input[int]:
"""
The port used for TCP health checks.
"""
return pulumi.get(self, "port")
@port.setter
def port(self, value: pulumi.Input[int]):
pulumi.set(self, "port", value)
@pulumi.input_type
class ComputeInstanceGroupInstanceArgs:
def __init__(__self__, *,
fqdn: Optional[pulumi.Input[str]] = None,
instance_id: Optional[pulumi.Input[str]] = None,
name: Optional[pulumi.Input[str]] = None,
network_interfaces: Optional[pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupInstanceNetworkInterfaceArgs']]]] = None,
status: Optional[pulumi.Input[str]] = None,
status_changed_at: Optional[pulumi.Input[str]] = None,
status_message: Optional[pulumi.Input[str]] = None,
zone_id: Optional[pulumi.Input[str]] = None):
"""
:param pulumi.Input[str] fqdn: DNS record fqdn (must have dot at the end).
:param pulumi.Input[str] instance_id: The ID of the instance.
:param pulumi.Input[str] name: Name template of the instance.
In order to be unique it must contain at least one of instance unique placeholders:
{instance.short_id}
{instance.index}
combination of {instance.zone_id} and {instance.index_in_zone}
Example: my-instance-{instance.index}
If not set, default is used: {instance_group.id}-{instance.short_id}
It may also contain another placeholders, see metadata doc for full list.
:param pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupInstanceNetworkInterfaceArgs']]] network_interfaces: Network specifications for the instance. This can be used multiple times for adding multiple interfaces. The structure is documented below.
:param pulumi.Input[str] status: The status of the instance.
:param pulumi.Input[str] status_message: The status message of the instance.
:param pulumi.Input[str] zone_id: The ID of the availability zone where the instance resides.
"""
if fqdn is not None:
pulumi.set(__self__, "fqdn", fqdn)
if instance_id is not None:
pulumi.set(__self__, "instance_id", instance_id)
if name is not None:
pulumi.set(__self__, "name", name)
if network_interfaces is not None:
pulumi.set(__self__, "network_interfaces", network_interfaces)
if status is not None:
pulumi.set(__self__, "status", status)
if status_changed_at is not None:
pulumi.set(__self__, "status_changed_at", status_changed_at)
if status_message is not None:
pulumi.set(__self__, "status_message", status_message)
if zone_id is not None:
pulumi.set(__self__, "zone_id", zone_id)
@property
@pulumi.getter
def fqdn(self) -> Optional[pulumi.Input[str]]:
"""
DNS record fqdn (must have dot at the end).
"""
return pulumi.get(self, "fqdn")
@fqdn.setter
def fqdn(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "fqdn", value)
@property
@pulumi.getter(name="instanceId")
def instance_id(self) -> Optional[pulumi.Input[str]]:
"""
The ID of the instance.
"""
return pulumi.get(self, "instance_id")
@instance_id.setter
def instance_id(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "instance_id", value)
@property
@pulumi.getter
def name(self) -> Optional[pulumi.Input[str]]:
"""
Name template of the instance.
In order to be unique it must contain at least one of instance unique placeholders:
{instance.short_id}
{instance.index}
combination of {instance.zone_id} and {instance.index_in_zone}
Example: my-instance-{instance.index}
If not set, default is used: {instance_group.id}-{instance.short_id}
It may also contain another placeholders, see metadata doc for full list.
"""
return pulumi.get(self, "name")
@name.setter
def name(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "name", value)
@property
@pulumi.getter(name="networkInterfaces")
def network_interfaces(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupInstanceNetworkInterfaceArgs']]]]:
"""
Network specifications for the instance. This can be used multiple times for adding multiple interfaces. The structure is documented below.
"""
return pulumi.get(self, "network_interfaces")
@network_interfaces.setter
def network_interfaces(self, value: Optional[pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupInstanceNetworkInterfaceArgs']]]]):
pulumi.set(self, "network_interfaces", value)
@property
@pulumi.getter
def status(self) -> Optional[pulumi.Input[str]]:
"""
The status of the instance.
"""
return pulumi.get(self, "status")
@status.setter
def status(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "status", value)
@property
@pulumi.getter(name="statusChangedAt")
def status_changed_at(self) -> Optional[pulumi.Input[str]]:
return pulumi.get(self, "status_changed_at")
@status_changed_at.setter
def status_changed_at(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "status_changed_at", value)
@property
@pulumi.getter(name="statusMessage")
def status_message(self) -> Optional[pulumi.Input[str]]:
"""
The status message of the instance.
"""
return pulumi.get(self, "status_message")
@status_message.setter
def status_message(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "status_message", value)
@property
@pulumi.getter(name="zoneId")
def zone_id(self) -> Optional[pulumi.Input[str]]:
"""
The ID of the availability zone where the instance resides.
"""
return pulumi.get(self, "zone_id")
@zone_id.setter
def zone_id(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "zone_id", value)
@pulumi.input_type
class ComputeInstanceGroupInstanceNetworkInterfaceArgs:
def __init__(__self__, *,
index: Optional[pulumi.Input[int]] = None,
ip_address: Optional[pulumi.Input[str]] = None,
ipv4: Optional[pulumi.Input[bool]] = None,
ipv6: Optional[pulumi.Input[bool]] = None,
ipv6_address: Optional[pulumi.Input[str]] = None,
mac_address: Optional[pulumi.Input[str]] = None,
nat: Optional[pulumi.Input[bool]] = None,
nat_ip_address: Optional[pulumi.Input[str]] = None,
nat_ip_version: Optional[pulumi.Input[str]] = None,
subnet_id: Optional[pulumi.Input[str]] = None):
"""
:param pulumi.Input[int] index: The index of the network interface as generated by the server.
:param pulumi.Input[str] ip_address: Manual set static IP address.
:param pulumi.Input[bool] ipv4: True if IPv4 address allocated for the network interface.
:param pulumi.Input[str] ipv6_address: Manual set static IPv6 address.
:param pulumi.Input[str] mac_address: The MAC address assigned to the network interface.
:param pulumi.Input[bool] nat: Flag for using NAT.
:param pulumi.Input[str] nat_ip_address: A public address that can be used to access the internet over NAT. Use `variables` to set.
:param pulumi.Input[str] nat_ip_version: The IP version for the public address.
:param pulumi.Input[str] subnet_id: The ID of the subnet to attach this interface to. The subnet must reside in the same zone where this instance was created.
"""
if index is not None:
pulumi.set(__self__, "index", index)
if ip_address is not None:
pulumi.set(__self__, "ip_address", ip_address)
if ipv4 is not None:
pulumi.set(__self__, "ipv4", ipv4)
if ipv6 is not None:
pulumi.set(__self__, "ipv6", ipv6)
if ipv6_address is not None:
pulumi.set(__self__, "ipv6_address", ipv6_address)
if mac_address is not None:
pulumi.set(__self__, "mac_address", mac_address)
if nat is not None:
pulumi.set(__self__, "nat", nat)
if nat_ip_address is not None:
pulumi.set(__self__, "nat_ip_address", nat_ip_address)
if nat_ip_version is not None:
pulumi.set(__self__, "nat_ip_version", nat_ip_version)
if subnet_id is not None:
pulumi.set(__self__, "subnet_id", subnet_id)
@property
@pulumi.getter
def index(self) -> Optional[pulumi.Input[int]]:
"""
The index of the network interface as generated by the server.
"""
return pulumi.get(self, "index")
@index.setter
def index(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "index", value)
@property
@pulumi.getter(name="ipAddress")
def ip_address(self) -> Optional[pulumi.Input[str]]:
"""
Manual set static IP address.
"""
return pulumi.get(self, "ip_address")
@ip_address.setter
def ip_address(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "ip_address", value)
@property
@pulumi.getter
def ipv4(self) -> Optional[pulumi.Input[bool]]:
"""
True if IPv4 address allocated for the network interface.
"""
return pulumi.get(self, "ipv4")
@ipv4.setter
def ipv4(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "ipv4", value)
@property
@pulumi.getter
def ipv6(self) -> Optional[pulumi.Input[bool]]:
return pulumi.get(self, "ipv6")
@ipv6.setter
def ipv6(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "ipv6", value)
@property
@pulumi.getter(name="ipv6Address")
def ipv6_address(self) -> Optional[pulumi.Input[str]]:
"""
Manual set static IPv6 address.
"""
return pulumi.get(self, "ipv6_address")
@ipv6_address.setter
def ipv6_address(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "ipv6_address", value)
@property
@pulumi.getter(name="macAddress")
def mac_address(self) -> Optional[pulumi.Input[str]]:
"""
The MAC address assigned to the network interface.
"""
return pulumi.get(self, "mac_address")
@mac_address.setter
def mac_address(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "mac_address", value)
@property
@pulumi.getter
def nat(self) -> Optional[pulumi.Input[bool]]:
"""
Flag for using NAT.
"""
return pulumi.get(self, "nat")
@nat.setter
def nat(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "nat", value)
@property
@pulumi.getter(name="natIpAddress")
def nat_ip_address(self) -> Optional[pulumi.Input[str]]:
"""
A public address that can be used to access the internet over NAT. Use `variables` to set.
"""
return pulumi.get(self, "nat_ip_address")
@nat_ip_address.setter
def nat_ip_address(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "nat_ip_address", value)
@property
@pulumi.getter(name="natIpVersion")
def nat_ip_version(self) -> Optional[pulumi.Input[str]]:
"""
The IP version for the public address.
"""
return pulumi.get(self, "nat_ip_version")
@nat_ip_version.setter
def nat_ip_version(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "nat_ip_version", value)
@property
@pulumi.getter(name="subnetId")
def subnet_id(self) -> Optional[pulumi.Input[str]]:
"""
The ID of the subnet to attach this interface to. The subnet must reside in the same zone where this instance was created.
"""
return pulumi.get(self, "subnet_id")
@subnet_id.setter
def subnet_id(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "subnet_id", value)
@pulumi.input_type
class ComputeInstanceGroupInstanceTemplateArgs:
def __init__(__self__, *,
boot_disk: pulumi.Input['ComputeInstanceGroupInstanceTemplateBootDiskArgs'],
network_interfaces: pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupInstanceTemplateNetworkInterfaceArgs']]],
resources: pulumi.Input['ComputeInstanceGroupInstanceTemplateResourcesArgs'],
description: Optional[pulumi.Input[str]] = None,
hostname: Optional[pulumi.Input[str]] = None,
labels: Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]] = None,
metadata: Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]] = None,
name: Optional[pulumi.Input[str]] = None,
network_settings: Optional[pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupInstanceTemplateNetworkSettingArgs']]]] = None,
placement_policy: Optional[pulumi.Input['ComputeInstanceGroupInstanceTemplatePlacementPolicyArgs']] = None,
platform_id: Optional[pulumi.Input[str]] = None,
scheduling_policy: Optional[pulumi.Input['ComputeInstanceGroupInstanceTemplateSchedulingPolicyArgs']] = None,
secondary_disks: Optional[pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupInstanceTemplateSecondaryDiskArgs']]]] = None,
service_account_id: Optional[pulumi.Input[str]] = None):
"""
:param pulumi.Input['ComputeInstanceGroupInstanceTemplateBootDiskArgs'] boot_disk: Boot disk specifications for the instance. The structure is documented below.
:param pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupInstanceTemplateNetworkInterfaceArgs']]] network_interfaces: Network specifications for the instance. This can be used multiple times for adding multiple interfaces. The structure is documented below.
:param pulumi.Input['ComputeInstanceGroupInstanceTemplateResourcesArgs'] resources: Compute resource specifications for the instance. The structure is documented below.
:param pulumi.Input[str] description: A description of the boot disk.
:param pulumi.Input[str] hostname: Hostname template for the instance.
This field is used to generate the FQDN value of instance.
The hostname must be unique within the network and region.
If not specified, the hostname will be equal to id of the instance
and FQDN will be `<id>.auto.internal`. Otherwise FQDN will be `<hostname>.<region_id>.internal`.
In order to be unique it must contain at least on of instance unique placeholders:
{instance.short_id}
{instance.index}
combination of {instance.zone_id} and {instance.index_in_zone}
Example: my-instance-{instance.index}
If not set, `name` value will be used
It may also contain another placeholders, see metadata doc for full list.
:param pulumi.Input[Mapping[str, pulumi.Input[str]]] labels: A map of labels of metric.
:param pulumi.Input[Mapping[str, pulumi.Input[str]]] metadata: A set of metadata key/value pairs to make available from within the instance.
:param pulumi.Input[str] name: Name template of the instance.
In order to be unique it must contain at least one of instance unique placeholders:
{instance.short_id}
{instance.index}
combination of {instance.zone_id} and {instance.index_in_zone}
Example: my-instance-{instance.index}
If not set, default is used: {instance_group.id}-{instance.short_id}
It may also contain another placeholders, see metadata doc for full list.
:param pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupInstanceTemplateNetworkSettingArgs']]] network_settings: Network acceleration type for instance. The structure is documented below.
:param pulumi.Input['ComputeInstanceGroupInstanceTemplatePlacementPolicyArgs'] placement_policy: The placement policy configuration. The structure is documented below.
:param pulumi.Input[str] platform_id: The ID of the hardware platform configuration for the instance. The default is 'standard-v1'.
:param pulumi.Input['ComputeInstanceGroupInstanceTemplateSchedulingPolicyArgs'] scheduling_policy: The scheduling policy configuration. The structure is documented below.
:param pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupInstanceTemplateSecondaryDiskArgs']]] secondary_disks: A list of disks to attach to the instance. The structure is documented below.
:param pulumi.Input[str] service_account_id: The ID of the service account authorized for this instance.
"""
pulumi.set(__self__, "boot_disk", boot_disk)
pulumi.set(__self__, "network_interfaces", network_interfaces)
pulumi.set(__self__, "resources", resources)
if description is not None:
pulumi.set(__self__, "description", description)
if hostname is not None:
pulumi.set(__self__, "hostname", hostname)
if labels is not None:
pulumi.set(__self__, "labels", labels)
if metadata is not None:
pulumi.set(__self__, "metadata", metadata)
if name is not None:
pulumi.set(__self__, "name", name)
if network_settings is not None:
pulumi.set(__self__, "network_settings", network_settings)
if placement_policy is not None:
pulumi.set(__self__, "placement_policy", placement_policy)
if platform_id is not None:
pulumi.set(__self__, "platform_id", platform_id)
if scheduling_policy is not None:
pulumi.set(__self__, "scheduling_policy", scheduling_policy)
if secondary_disks is not None:
pulumi.set(__self__, "secondary_disks", secondary_disks)
if service_account_id is not None:
pulumi.set(__self__, "service_account_id", service_account_id)
@property
@pulumi.getter(name="bootDisk")
def boot_disk(self) -> pulumi.Input['ComputeInstanceGroupInstanceTemplateBootDiskArgs']:
"""
Boot disk specifications for the instance. The structure is documented below.
"""
return pulumi.get(self, "boot_disk")
@boot_disk.setter
def boot_disk(self, value: pulumi.Input['ComputeInstanceGroupInstanceTemplateBootDiskArgs']):
pulumi.set(self, "boot_disk", value)
@property
@pulumi.getter(name="networkInterfaces")
def network_interfaces(self) -> pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupInstanceTemplateNetworkInterfaceArgs']]]:
"""
Network specifications for the instance. This can be used multiple times for adding multiple interfaces. The structure is documented below.
"""
return pulumi.get(self, "network_interfaces")
@network_interfaces.setter
def network_interfaces(self, value: pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupInstanceTemplateNetworkInterfaceArgs']]]):
pulumi.set(self, "network_interfaces", value)
@property
@pulumi.getter
def resources(self) -> pulumi.Input['ComputeInstanceGroupInstanceTemplateResourcesArgs']:
"""
Compute resource specifications for the instance. The structure is documented below.
"""
return pulumi.get(self, "resources")
@resources.setter
def resources(self, value: pulumi.Input['ComputeInstanceGroupInstanceTemplateResourcesArgs']):
pulumi.set(self, "resources", value)
@property
@pulumi.getter
def description(self) -> Optional[pulumi.Input[str]]:
"""
A description of the boot disk.
"""
return pulumi.get(self, "description")
@description.setter
def description(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "description", value)
@property
@pulumi.getter
def hostname(self) -> Optional[pulumi.Input[str]]:
"""
Hostname template for the instance.
This field is used to generate the FQDN value of instance.
The hostname must be unique within the network and region.
If not specified, the hostname will be equal to id of the instance
and FQDN will be `<id>.auto.internal`. Otherwise FQDN will be `<hostname>.<region_id>.internal`.
In order to be unique it must contain at least on of instance unique placeholders:
{instance.short_id}
{instance.index}
combination of {instance.zone_id} and {instance.index_in_zone}
Example: my-instance-{instance.index}
If not set, `name` value will be used
It may also contain another placeholders, see metadata doc for full list.
"""
return pulumi.get(self, "hostname")
@hostname.setter
def hostname(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "hostname", value)
@property
@pulumi.getter
def labels(self) -> Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]]:
"""
A map of labels of metric.
"""
return pulumi.get(self, "labels")
@labels.setter
def labels(self, value: Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]]):
pulumi.set(self, "labels", value)
@property
@pulumi.getter
def metadata(self) -> Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]]:
"""
A set of metadata key/value pairs to make available from within the instance.
"""
return pulumi.get(self, "metadata")
@metadata.setter
def metadata(self, value: Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]]):
pulumi.set(self, "metadata", value)
@property
@pulumi.getter
def name(self) -> Optional[pulumi.Input[str]]:
"""
Name template of the instance.
In order to be unique it must contain at least one of instance unique placeholders:
{instance.short_id}
{instance.index}
combination of {instance.zone_id} and {instance.index_in_zone}
Example: my-instance-{instance.index}
If not set, default is used: {instance_group.id}-{instance.short_id}
It may also contain another placeholders, see metadata doc for full list.
"""
return pulumi.get(self, "name")
@name.setter
def name(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "name", value)
@property
@pulumi.getter(name="networkSettings")
def network_settings(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupInstanceTemplateNetworkSettingArgs']]]]:
"""
Network acceleration type for instance. The structure is documented below.
"""
return pulumi.get(self, "network_settings")
@network_settings.setter
def network_settings(self, value: Optional[pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupInstanceTemplateNetworkSettingArgs']]]]):
pulumi.set(self, "network_settings", value)
@property
@pulumi.getter(name="placementPolicy")
def placement_policy(self) -> Optional[pulumi.Input['ComputeInstanceGroupInstanceTemplatePlacementPolicyArgs']]:
"""
The placement policy configuration. The structure is documented below.
"""
return pulumi.get(self, "placement_policy")
@placement_policy.setter
def placement_policy(self, value: Optional[pulumi.Input['ComputeInstanceGroupInstanceTemplatePlacementPolicyArgs']]):
pulumi.set(self, "placement_policy", value)
@property
@pulumi.getter(name="platformId")
def platform_id(self) -> Optional[pulumi.Input[str]]:
"""
The ID of the hardware platform configuration for the instance. The default is 'standard-v1'.
"""
return pulumi.get(self, "platform_id")
@platform_id.setter
def platform_id(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "platform_id", value)
@property
@pulumi.getter(name="schedulingPolicy")
def scheduling_policy(self) -> Optional[pulumi.Input['ComputeInstanceGroupInstanceTemplateSchedulingPolicyArgs']]:
"""
The scheduling policy configuration. The structure is documented below.
"""
return pulumi.get(self, "scheduling_policy")
@scheduling_policy.setter
def scheduling_policy(self, value: Optional[pulumi.Input['ComputeInstanceGroupInstanceTemplateSchedulingPolicyArgs']]):
pulumi.set(self, "scheduling_policy", value)
@property
@pulumi.getter(name="secondaryDisks")
def secondary_disks(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupInstanceTemplateSecondaryDiskArgs']]]]:
"""
A list of disks to attach to the instance. The structure is documented below.
"""
return pulumi.get(self, "secondary_disks")
@secondary_disks.setter
def secondary_disks(self, value: Optional[pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupInstanceTemplateSecondaryDiskArgs']]]]):
pulumi.set(self, "secondary_disks", value)
@property
@pulumi.getter(name="serviceAccountId")
def service_account_id(self) -> Optional[pulumi.Input[str]]:
"""
The ID of the service account authorized for this instance.
"""
return pulumi.get(self, "service_account_id")
@service_account_id.setter
def service_account_id(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "service_account_id", value)
@pulumi.input_type
class ComputeInstanceGroupInstanceTemplateBootDiskArgs:
def __init__(__self__, *,
device_name: Optional[pulumi.Input[str]] = None,
disk_id: Optional[pulumi.Input[str]] = None,
initialize_params: Optional[pulumi.Input['ComputeInstanceGroupInstanceTemplateBootDiskInitializeParamsArgs']] = None,
mode: Optional[pulumi.Input[str]] = None):
"""
:param pulumi.Input[str] device_name: This value can be used to reference the device under `/dev/disk/by-id/`.
:param pulumi.Input[str] disk_id: ID of the existing disk. To set use variables.
:param pulumi.Input['ComputeInstanceGroupInstanceTemplateBootDiskInitializeParamsArgs'] initialize_params: Parameters for creating a disk alongside the instance. The structure is documented below.
:param pulumi.Input[str] mode: The access mode to the disk resource. By default a disk is attached in `READ_WRITE` mode.
"""
if device_name is not None:
pulumi.set(__self__, "device_name", device_name)
if disk_id is not None:
pulumi.set(__self__, "disk_id", disk_id)
if initialize_params is not None:
pulumi.set(__self__, "initialize_params", initialize_params)
if mode is not None:
pulumi.set(__self__, "mode", mode)
@property
@pulumi.getter(name="deviceName")
def device_name(self) -> Optional[pulumi.Input[str]]:
"""
This value can be used to reference the device under `/dev/disk/by-id/`.
"""
return pulumi.get(self, "device_name")
@device_name.setter
def device_name(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "device_name", value)
@property
@pulumi.getter(name="diskId")
def disk_id(self) -> Optional[pulumi.Input[str]]:
"""
ID of the existing disk. To set use variables.
"""
return pulumi.get(self, "disk_id")
@disk_id.setter
def disk_id(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "disk_id", value)
@property
@pulumi.getter(name="initializeParams")
def initialize_params(self) -> Optional[pulumi.Input['ComputeInstanceGroupInstanceTemplateBootDiskInitializeParamsArgs']]:
"""
Parameters for creating a disk alongside the instance. The structure is documented below.
"""
return pulumi.get(self, "initialize_params")
@initialize_params.setter
def initialize_params(self, value: Optional[pulumi.Input['ComputeInstanceGroupInstanceTemplateBootDiskInitializeParamsArgs']]):
pulumi.set(self, "initialize_params", value)
@property
@pulumi.getter
def mode(self) -> Optional[pulumi.Input[str]]:
"""
The access mode to the disk resource. By default a disk is attached in `READ_WRITE` mode.
"""
return pulumi.get(self, "mode")
@mode.setter
def mode(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "mode", value)
@pulumi.input_type
class ComputeInstanceGroupInstanceTemplateBootDiskInitializeParamsArgs:
def __init__(__self__, *,
description: Optional[pulumi.Input[str]] = None,
image_id: Optional[pulumi.Input[str]] = None,
size: Optional[pulumi.Input[int]] = None,
snapshot_id: Optional[pulumi.Input[str]] = None,
type: Optional[pulumi.Input[str]] = None):
"""
:param pulumi.Input[str] description: A description of the boot disk.
:param pulumi.Input[str] image_id: The disk image to initialize this disk from.
:param pulumi.Input[int] size: The number of instances in the instance group.
:param pulumi.Input[str] snapshot_id: The snapshot to initialize this disk from.
:param pulumi.Input[str] type: Network acceleration type. By default a network is in `STANDARD` mode.
"""
if description is not None:
pulumi.set(__self__, "description", description)
if image_id is not None:
pulumi.set(__self__, "image_id", image_id)
if size is not None:
pulumi.set(__self__, "size", size)
if snapshot_id is not None:
pulumi.set(__self__, "snapshot_id", snapshot_id)
if type is not None:
pulumi.set(__self__, "type", type)
@property
@pulumi.getter
def description(self) -> Optional[pulumi.Input[str]]:
"""
A description of the boot disk.
"""
return pulumi.get(self, "description")
@description.setter
def description(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "description", value)
@property
@pulumi.getter(name="imageId")
def image_id(self) -> Optional[pulumi.Input[str]]:
"""
The disk image to initialize this disk from.
"""
return pulumi.get(self, "image_id")
@image_id.setter
def image_id(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "image_id", value)
@property
@pulumi.getter
def size(self) -> Optional[pulumi.Input[int]]:
"""
The number of instances in the instance group.
"""
return pulumi.get(self, "size")
@size.setter
def size(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "size", value)
@property
@pulumi.getter(name="snapshotId")
def snapshot_id(self) -> Optional[pulumi.Input[str]]:
"""
The snapshot to initialize this disk from.
"""
return pulumi.get(self, "snapshot_id")
@snapshot_id.setter
def snapshot_id(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "snapshot_id", value)
@property
@pulumi.getter
def type(self) -> Optional[pulumi.Input[str]]:
"""
Network acceleration type. By default a network is in `STANDARD` mode.
"""
return pulumi.get(self, "type")
@type.setter
def type(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "type", value)
@pulumi.input_type
class ComputeInstanceGroupInstanceTemplateNetworkInterfaceArgs:
def __init__(__self__, *,
dns_records: Optional[pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupInstanceTemplateNetworkInterfaceDnsRecordArgs']]]] = None,
ip_address: Optional[pulumi.Input[str]] = None,
ipv4: Optional[pulumi.Input[bool]] = None,
ipv6: Optional[pulumi.Input[bool]] = None,
ipv6_address: Optional[pulumi.Input[str]] = None,
ipv6_dns_records: Optional[pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupInstanceTemplateNetworkInterfaceIpv6DnsRecordArgs']]]] = None,
nat: Optional[pulumi.Input[bool]] = None,
nat_dns_records: Optional[pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupInstanceTemplateNetworkInterfaceNatDnsRecordArgs']]]] = None,
nat_ip_address: Optional[pulumi.Input[str]] = None,
network_id: Optional[pulumi.Input[str]] = None,
security_group_ids: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None,
subnet_ids: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None):
"""
:param pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupInstanceTemplateNetworkInterfaceDnsRecordArgs']]] dns_records: List of dns records. The structure is documented below.
:param pulumi.Input[str] ip_address: Manual set static IP address.
:param pulumi.Input[bool] ipv4: True if IPv4 address allocated for the network interface.
:param pulumi.Input[str] ipv6_address: Manual set static IPv6 address.
:param pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupInstanceTemplateNetworkInterfaceIpv6DnsRecordArgs']]] ipv6_dns_records: List of ipv6 dns records. The structure is documented below.
:param pulumi.Input[bool] nat: Flag for using NAT.
:param pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupInstanceTemplateNetworkInterfaceNatDnsRecordArgs']]] nat_dns_records: List of nat dns records. The structure is documented below.
:param pulumi.Input[str] nat_ip_address: A public address that can be used to access the internet over NAT. Use `variables` to set.
:param pulumi.Input[str] network_id: The ID of the network.
:param pulumi.Input[Sequence[pulumi.Input[str]]] security_group_ids: Security group ids for network interface.
:param pulumi.Input[Sequence[pulumi.Input[str]]] subnet_ids: The ID of the subnets to attach this interface to.
"""
if dns_records is not None:
pulumi.set(__self__, "dns_records", dns_records)
if ip_address is not None:
pulumi.set(__self__, "ip_address", ip_address)
if ipv4 is not None:
pulumi.set(__self__, "ipv4", ipv4)
if ipv6 is not None:
pulumi.set(__self__, "ipv6", ipv6)
if ipv6_address is not None:
pulumi.set(__self__, "ipv6_address", ipv6_address)
if ipv6_dns_records is not None:
pulumi.set(__self__, "ipv6_dns_records", ipv6_dns_records)
if nat is not None:
pulumi.set(__self__, "nat", nat)
if nat_dns_records is not None:
pulumi.set(__self__, "nat_dns_records", nat_dns_records)
if nat_ip_address is not None:
pulumi.set(__self__, "nat_ip_address", nat_ip_address)
if network_id is not None:
pulumi.set(__self__, "network_id", network_id)
if security_group_ids is not None:
pulumi.set(__self__, "security_group_ids", security_group_ids)
if subnet_ids is not None:
pulumi.set(__self__, "subnet_ids", subnet_ids)
@property
@pulumi.getter(name="dnsRecords")
def dns_records(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupInstanceTemplateNetworkInterfaceDnsRecordArgs']]]]:
"""
List of dns records. The structure is documented below.
"""
return pulumi.get(self, "dns_records")
@dns_records.setter
def dns_records(self, value: Optional[pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupInstanceTemplateNetworkInterfaceDnsRecordArgs']]]]):
pulumi.set(self, "dns_records", value)
@property
@pulumi.getter(name="ipAddress")
def ip_address(self) -> Optional[pulumi.Input[str]]:
"""
Manual set static IP address.
"""
return pulumi.get(self, "ip_address")
@ip_address.setter
def ip_address(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "ip_address", value)
@property
@pulumi.getter
def ipv4(self) -> Optional[pulumi.Input[bool]]:
"""
True if IPv4 address allocated for the network interface.
"""
return pulumi.get(self, "ipv4")
@ipv4.setter
def ipv4(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "ipv4", value)
@property
@pulumi.getter
def ipv6(self) -> Optional[pulumi.Input[bool]]:
return pulumi.get(self, "ipv6")
@ipv6.setter
def ipv6(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "ipv6", value)
@property
@pulumi.getter(name="ipv6Address")
def ipv6_address(self) -> Optional[pulumi.Input[str]]:
"""
Manual set static IPv6 address.
"""
return pulumi.get(self, "ipv6_address")
@ipv6_address.setter
def ipv6_address(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "ipv6_address", value)
@property
@pulumi.getter(name="ipv6DnsRecords")
def ipv6_dns_records(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupInstanceTemplateNetworkInterfaceIpv6DnsRecordArgs']]]]:
"""
List of ipv6 dns records. The structure is documented below.
"""
return pulumi.get(self, "ipv6_dns_records")
@ipv6_dns_records.setter
def ipv6_dns_records(self, value: Optional[pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupInstanceTemplateNetworkInterfaceIpv6DnsRecordArgs']]]]):
pulumi.set(self, "ipv6_dns_records", value)
@property
@pulumi.getter
def nat(self) -> Optional[pulumi.Input[bool]]:
"""
Flag for using NAT.
"""
return pulumi.get(self, "nat")
@nat.setter
def nat(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "nat", value)
@property
@pulumi.getter(name="natDnsRecords")
def nat_dns_records(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupInstanceTemplateNetworkInterfaceNatDnsRecordArgs']]]]:
"""
List of nat dns records. The structure is documented below.
"""
return pulumi.get(self, "nat_dns_records")
@nat_dns_records.setter
def nat_dns_records(self, value: Optional[pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupInstanceTemplateNetworkInterfaceNatDnsRecordArgs']]]]):
pulumi.set(self, "nat_dns_records", value)
@property
@pulumi.getter(name="natIpAddress")
def nat_ip_address(self) -> Optional[pulumi.Input[str]]:
"""
A public address that can be used to access the internet over NAT. Use `variables` to set.
"""
return pulumi.get(self, "nat_ip_address")
@nat_ip_address.setter
def nat_ip_address(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "nat_ip_address", value)
@property
@pulumi.getter(name="networkId")
def network_id(self) -> Optional[pulumi.Input[str]]:
"""
The ID of the network.
"""
return pulumi.get(self, "network_id")
@network_id.setter
def network_id(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "network_id", value)
@property
@pulumi.getter(name="securityGroupIds")
def security_group_ids(self) -> Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]:
"""
Security group ids for network interface.
"""
return pulumi.get(self, "security_group_ids")
@security_group_ids.setter
def security_group_ids(self, value: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]):
pulumi.set(self, "security_group_ids", value)
@property
@pulumi.getter(name="subnetIds")
def subnet_ids(self) -> Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]:
"""
The ID of the subnets to attach this interface to.
"""
return pulumi.get(self, "subnet_ids")
@subnet_ids.setter
def subnet_ids(self, value: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]):
pulumi.set(self, "subnet_ids", value)
@pulumi.input_type
class ComputeInstanceGroupInstanceTemplateNetworkInterfaceDnsRecordArgs:
def __init__(__self__, *,
fqdn: pulumi.Input[str],
dns_zone_id: Optional[pulumi.Input[str]] = None,
ptr: Optional[pulumi.Input[bool]] = None,
ttl: Optional[pulumi.Input[int]] = None):
"""
:param pulumi.Input[str] fqdn: DNS record fqdn (must have dot at the end).
:param pulumi.Input[str] dns_zone_id: DNS zone id (if not set, private zone used).
:param pulumi.Input[bool] ptr: When set to true, also create PTR DNS record.
:param pulumi.Input[int] ttl: DNS record TTL.
"""
pulumi.set(__self__, "fqdn", fqdn)
if dns_zone_id is not None:
pulumi.set(__self__, "dns_zone_id", dns_zone_id)
if ptr is not None:
pulumi.set(__self__, "ptr", ptr)
if ttl is not None:
pulumi.set(__self__, "ttl", ttl)
@property
@pulumi.getter
def fqdn(self) -> pulumi.Input[str]:
"""
DNS record fqdn (must have dot at the end).
"""
return pulumi.get(self, "fqdn")
@fqdn.setter
def fqdn(self, value: pulumi.Input[str]):
pulumi.set(self, "fqdn", value)
@property
@pulumi.getter(name="dnsZoneId")
def dns_zone_id(self) -> Optional[pulumi.Input[str]]:
"""
DNS zone id (if not set, private zone used).
"""
return pulumi.get(self, "dns_zone_id")
@dns_zone_id.setter
def dns_zone_id(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "dns_zone_id", value)
@property
@pulumi.getter
def ptr(self) -> Optional[pulumi.Input[bool]]:
"""
When set to true, also create PTR DNS record.
"""
return pulumi.get(self, "ptr")
@ptr.setter
def ptr(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "ptr", value)
@property
@pulumi.getter
def ttl(self) -> Optional[pulumi.Input[int]]:
"""
DNS record TTL.
"""
return pulumi.get(self, "ttl")
@ttl.setter
def ttl(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "ttl", value)
@pulumi.input_type
class ComputeInstanceGroupInstanceTemplateNetworkInterfaceIpv6DnsRecordArgs:
def __init__(__self__, *,
fqdn: pulumi.Input[str],
dns_zone_id: Optional[pulumi.Input[str]] = None,
ptr: Optional[pulumi.Input[bool]] = None,
ttl: Optional[pulumi.Input[int]] = None):
"""
:param pulumi.Input[str] fqdn: DNS record fqdn (must have dot at the end).
:param pulumi.Input[str] dns_zone_id: DNS zone id (if not set, private zone used).
:param pulumi.Input[bool] ptr: When set to true, also create PTR DNS record.
:param pulumi.Input[int] ttl: DNS record TTL.
"""
pulumi.set(__self__, "fqdn", fqdn)
if dns_zone_id is not None:
pulumi.set(__self__, "dns_zone_id", dns_zone_id)
if ptr is not None:
pulumi.set(__self__, "ptr", ptr)
if ttl is not None:
pulumi.set(__self__, "ttl", ttl)
@property
@pulumi.getter
def fqdn(self) -> pulumi.Input[str]:
"""
DNS record fqdn (must have dot at the end).
"""
return pulumi.get(self, "fqdn")
@fqdn.setter
def fqdn(self, value: pulumi.Input[str]):
pulumi.set(self, "fqdn", value)
@property
@pulumi.getter(name="dnsZoneId")
def dns_zone_id(self) -> Optional[pulumi.Input[str]]:
"""
DNS zone id (if not set, private zone used).
"""
return pulumi.get(self, "dns_zone_id")
@dns_zone_id.setter
def dns_zone_id(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "dns_zone_id", value)
@property
@pulumi.getter
def ptr(self) -> Optional[pulumi.Input[bool]]:
"""
When set to true, also create PTR DNS record.
"""
return pulumi.get(self, "ptr")
@ptr.setter
def ptr(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "ptr", value)
@property
@pulumi.getter
def ttl(self) -> Optional[pulumi.Input[int]]:
"""
DNS record TTL.
"""
return pulumi.get(self, "ttl")
@ttl.setter
def ttl(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "ttl", value)
@pulumi.input_type
class ComputeInstanceGroupInstanceTemplateNetworkInterfaceNatDnsRecordArgs:
def __init__(__self__, *,
fqdn: pulumi.Input[str],
dns_zone_id: Optional[pulumi.Input[str]] = None,
ptr: Optional[pulumi.Input[bool]] = None,
ttl: Optional[pulumi.Input[int]] = None):
"""
:param pulumi.Input[str] fqdn: DNS record fqdn (must have dot at the end).
:param pulumi.Input[str] dns_zone_id: DNS zone id (if not set, private zone used).
:param pulumi.Input[bool] ptr: When set to true, also create PTR DNS record.
:param pulumi.Input[int] ttl: DNS record TTL.
"""
pulumi.set(__self__, "fqdn", fqdn)
if dns_zone_id is not None:
pulumi.set(__self__, "dns_zone_id", dns_zone_id)
if ptr is not None:
pulumi.set(__self__, "ptr", ptr)
if ttl is not None:
pulumi.set(__self__, "ttl", ttl)
@property
@pulumi.getter
def fqdn(self) -> pulumi.Input[str]:
"""
DNS record fqdn (must have dot at the end).
"""
return pulumi.get(self, "fqdn")
@fqdn.setter
def fqdn(self, value: pulumi.Input[str]):
pulumi.set(self, "fqdn", value)
@property
@pulumi.getter(name="dnsZoneId")
def dns_zone_id(self) -> Optional[pulumi.Input[str]]:
"""
DNS zone id (if not set, private zone used).
"""
return pulumi.get(self, "dns_zone_id")
@dns_zone_id.setter
def dns_zone_id(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "dns_zone_id", value)
@property
@pulumi.getter
def ptr(self) -> Optional[pulumi.Input[bool]]:
"""
When set to true, also create PTR DNS record.
"""
return pulumi.get(self, "ptr")
@ptr.setter
def ptr(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "ptr", value)
@property
@pulumi.getter
def ttl(self) -> Optional[pulumi.Input[int]]:
"""
DNS record TTL.
"""
return pulumi.get(self, "ttl")
@ttl.setter
def ttl(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "ttl", value)
@pulumi.input_type
class ComputeInstanceGroupInstanceTemplateNetworkSettingArgs:
def __init__(__self__, *,
type: Optional[pulumi.Input[str]] = None):
"""
:param pulumi.Input[str] type: Network acceleration type. By default a network is in `STANDARD` mode.
"""
if type is not None:
pulumi.set(__self__, "type", type)
@property
@pulumi.getter
def type(self) -> Optional[pulumi.Input[str]]:
"""
Network acceleration type. By default a network is in `STANDARD` mode.
"""
return pulumi.get(self, "type")
@type.setter
def type(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "type", value)
@pulumi.input_type
class ComputeInstanceGroupInstanceTemplatePlacementPolicyArgs:
def __init__(__self__, *,
placement_group_id: pulumi.Input[str]):
"""
:param pulumi.Input[str] placement_group_id: Specifies the id of the Placement Group to assign to the instances.
"""
pulumi.set(__self__, "placement_group_id", placement_group_id)
@property
@pulumi.getter(name="placementGroupId")
def placement_group_id(self) -> pulumi.Input[str]:
"""
Specifies the id of the Placement Group to assign to the instances.
"""
return pulumi.get(self, "placement_group_id")
@placement_group_id.setter
def placement_group_id(self, value: pulumi.Input[str]):
pulumi.set(self, "placement_group_id", value)
@pulumi.input_type
class ComputeInstanceGroupInstanceTemplateResourcesArgs:
def __init__(__self__, *,
cores: pulumi.Input[int],
memory: pulumi.Input[float],
core_fraction: Optional[pulumi.Input[int]] = None,
gpus: Optional[pulumi.Input[int]] = None):
"""
:param pulumi.Input[int] cores: The number of CPU cores for the instance.
:param pulumi.Input[float] memory: The memory size in GB.
:param pulumi.Input[int] core_fraction: If provided, specifies baseline core performance as a percent.
"""
pulumi.set(__self__, "cores", cores)
pulumi.set(__self__, "memory", memory)
if core_fraction is not None:
pulumi.set(__self__, "core_fraction", core_fraction)
if gpus is not None:
pulumi.set(__self__, "gpus", gpus)
@property
@pulumi.getter
def cores(self) -> pulumi.Input[int]:
"""
The number of CPU cores for the instance.
"""
return pulumi.get(self, "cores")
@cores.setter
def cores(self, value: pulumi.Input[int]):
pulumi.set(self, "cores", value)
@property
@pulumi.getter
def memory(self) -> pulumi.Input[float]:
"""
The memory size in GB.
"""
return pulumi.get(self, "memory")
@memory.setter
def memory(self, value: pulumi.Input[float]):
pulumi.set(self, "memory", value)
@property
@pulumi.getter(name="coreFraction")
def core_fraction(self) -> Optional[pulumi.Input[int]]:
"""
If provided, specifies baseline core performance as a percent.
"""
return pulumi.get(self, "core_fraction")
@core_fraction.setter
def core_fraction(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "core_fraction", value)
@property
@pulumi.getter
def gpus(self) -> Optional[pulumi.Input[int]]:
return pulumi.get(self, "gpus")
@gpus.setter
def gpus(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "gpus", value)
@pulumi.input_type
class ComputeInstanceGroupInstanceTemplateSchedulingPolicyArgs:
def __init__(__self__, *,
preemptible: Optional[pulumi.Input[bool]] = None):
"""
:param pulumi.Input[bool] preemptible: Specifies if the instance is preemptible. Defaults to false.
"""
if preemptible is not None:
pulumi.set(__self__, "preemptible", preemptible)
@property
@pulumi.getter
def preemptible(self) -> Optional[pulumi.Input[bool]]:
"""
Specifies if the instance is preemptible. Defaults to false.
"""
return pulumi.get(self, "preemptible")
@preemptible.setter
def preemptible(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "preemptible", value)
@pulumi.input_type
class ComputeInstanceGroupInstanceTemplateSecondaryDiskArgs:
def __init__(__self__, *,
device_name: Optional[pulumi.Input[str]] = None,
disk_id: Optional[pulumi.Input[str]] = None,
initialize_params: Optional[pulumi.Input['ComputeInstanceGroupInstanceTemplateSecondaryDiskInitializeParamsArgs']] = None,
mode: Optional[pulumi.Input[str]] = None):
"""
:param pulumi.Input[str] device_name: This value can be used to reference the device under `/dev/disk/by-id/`.
:param pulumi.Input[str] disk_id: ID of the existing disk. To set use variables.
:param pulumi.Input['ComputeInstanceGroupInstanceTemplateSecondaryDiskInitializeParamsArgs'] initialize_params: Parameters for creating a disk alongside the instance. The structure is documented below.
:param pulumi.Input[str] mode: The access mode to the disk resource. By default a disk is attached in `READ_WRITE` mode.
"""
if device_name is not None:
pulumi.set(__self__, "device_name", device_name)
if disk_id is not None:
pulumi.set(__self__, "disk_id", disk_id)
if initialize_params is not None:
pulumi.set(__self__, "initialize_params", initialize_params)
if mode is not None:
pulumi.set(__self__, "mode", mode)
@property
@pulumi.getter(name="deviceName")
def device_name(self) -> Optional[pulumi.Input[str]]:
"""
This value can be used to reference the device under `/dev/disk/by-id/`.
"""
return pulumi.get(self, "device_name")
@device_name.setter
def device_name(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "device_name", value)
@property
@pulumi.getter(name="diskId")
def disk_id(self) -> Optional[pulumi.Input[str]]:
"""
ID of the existing disk. To set use variables.
"""
return pulumi.get(self, "disk_id")
@disk_id.setter
def disk_id(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "disk_id", value)
@property
@pulumi.getter(name="initializeParams")
def initialize_params(self) -> Optional[pulumi.Input['ComputeInstanceGroupInstanceTemplateSecondaryDiskInitializeParamsArgs']]:
"""
Parameters for creating a disk alongside the instance. The structure is documented below.
"""
return pulumi.get(self, "initialize_params")
@initialize_params.setter
def initialize_params(self, value: Optional[pulumi.Input['ComputeInstanceGroupInstanceTemplateSecondaryDiskInitializeParamsArgs']]):
pulumi.set(self, "initialize_params", value)
@property
@pulumi.getter
def mode(self) -> Optional[pulumi.Input[str]]:
"""
The access mode to the disk resource. By default a disk is attached in `READ_WRITE` mode.
"""
return pulumi.get(self, "mode")
@mode.setter
def mode(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "mode", value)
@pulumi.input_type
class ComputeInstanceGroupInstanceTemplateSecondaryDiskInitializeParamsArgs:
def __init__(__self__, *,
description: Optional[pulumi.Input[str]] = None,
image_id: Optional[pulumi.Input[str]] = None,
size: Optional[pulumi.Input[int]] = None,
snapshot_id: Optional[pulumi.Input[str]] = None,
type: Optional[pulumi.Input[str]] = None):
"""
:param pulumi.Input[str] description: A description of the boot disk.
:param pulumi.Input[str] image_id: The disk image to initialize this disk from.
:param pulumi.Input[int] size: The number of instances in the instance group.
:param pulumi.Input[str] snapshot_id: The snapshot to initialize this disk from.
:param pulumi.Input[str] type: Network acceleration type. By default a network is in `STANDARD` mode.
"""
if description is not None:
pulumi.set(__self__, "description", description)
if image_id is not None:
pulumi.set(__self__, "image_id", image_id)
if size is not None:
pulumi.set(__self__, "size", size)
if snapshot_id is not None:
pulumi.set(__self__, "snapshot_id", snapshot_id)
if type is not None:
pulumi.set(__self__, "type", type)
@property
@pulumi.getter
def description(self) -> Optional[pulumi.Input[str]]:
"""
A description of the boot disk.
"""
return pulumi.get(self, "description")
@description.setter
def description(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "description", value)
@property
@pulumi.getter(name="imageId")
def image_id(self) -> Optional[pulumi.Input[str]]:
"""
The disk image to initialize this disk from.
"""
return pulumi.get(self, "image_id")
@image_id.setter
def image_id(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "image_id", value)
@property
@pulumi.getter
def size(self) -> Optional[pulumi.Input[int]]:
"""
The number of instances in the instance group.
"""
return pulumi.get(self, "size")
@size.setter
def size(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "size", value)
@property
@pulumi.getter(name="snapshotId")
def snapshot_id(self) -> Optional[pulumi.Input[str]]:
"""
The snapshot to initialize this disk from.
"""
return pulumi.get(self, "snapshot_id")
@snapshot_id.setter
def snapshot_id(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "snapshot_id", value)
@property
@pulumi.getter
def type(self) -> Optional[pulumi.Input[str]]:
"""
Network acceleration type. By default a network is in `STANDARD` mode.
"""
return pulumi.get(self, "type")
@type.setter
def type(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "type", value)
@pulumi.input_type
class ComputeInstanceGroupLoadBalancerArgs:
def __init__(__self__, *,
max_opening_traffic_duration: Optional[pulumi.Input[int]] = None,
status_message: Optional[pulumi.Input[str]] = None,
target_group_description: Optional[pulumi.Input[str]] = None,
target_group_id: Optional[pulumi.Input[str]] = None,
target_group_labels: Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]] = None,
target_group_name: Optional[pulumi.Input[str]] = None):
"""
:param pulumi.Input[int] max_opening_traffic_duration: Timeout for waiting for the VM to be checked by the load balancer. If the timeout is exceeded, the VM will be turned off based on the deployment policy. Specified in seconds.
:param pulumi.Input[str] status_message: The status message of the instance.
:param pulumi.Input[str] target_group_description: A description of the target group.
:param pulumi.Input[Mapping[str, pulumi.Input[str]]] target_group_labels: A set of key/value label pairs.
:param pulumi.Input[str] target_group_name: The name of the target group.
"""
if max_opening_traffic_duration is not None:
pulumi.set(__self__, "max_opening_traffic_duration", max_opening_traffic_duration)
if status_message is not None:
pulumi.set(__self__, "status_message", status_message)
if target_group_description is not None:
pulumi.set(__self__, "target_group_description", target_group_description)
if target_group_id is not None:
pulumi.set(__self__, "target_group_id", target_group_id)
if target_group_labels is not None:
pulumi.set(__self__, "target_group_labels", target_group_labels)
if target_group_name is not None:
pulumi.set(__self__, "target_group_name", target_group_name)
@property
@pulumi.getter(name="maxOpeningTrafficDuration")
def max_opening_traffic_duration(self) -> Optional[pulumi.Input[int]]:
"""
Timeout for waiting for the VM to be checked by the load balancer. If the timeout is exceeded, the VM will be turned off based on the deployment policy. Specified in seconds.
"""
return pulumi.get(self, "max_opening_traffic_duration")
@max_opening_traffic_duration.setter
def max_opening_traffic_duration(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "max_opening_traffic_duration", value)
@property
@pulumi.getter(name="statusMessage")
def status_message(self) -> Optional[pulumi.Input[str]]:
"""
The status message of the instance.
"""
return pulumi.get(self, "status_message")
@status_message.setter
def status_message(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "status_message", value)
@property
@pulumi.getter(name="targetGroupDescription")
def target_group_description(self) -> Optional[pulumi.Input[str]]:
"""
A description of the target group.
"""
return pulumi.get(self, "target_group_description")
@target_group_description.setter
def target_group_description(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "target_group_description", value)
@property
@pulumi.getter(name="targetGroupId")
def target_group_id(self) -> Optional[pulumi.Input[str]]:
return pulumi.get(self, "target_group_id")
@target_group_id.setter
def target_group_id(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "target_group_id", value)
@property
@pulumi.getter(name="targetGroupLabels")
def target_group_labels(self) -> Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]]:
"""
A set of key/value label pairs.
"""
return pulumi.get(self, "target_group_labels")
@target_group_labels.setter
def target_group_labels(self, value: Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]]):
pulumi.set(self, "target_group_labels", value)
@property
@pulumi.getter(name="targetGroupName")
def target_group_name(self) -> Optional[pulumi.Input[str]]:
"""
The name of the target group.
"""
return pulumi.get(self, "target_group_name")
@target_group_name.setter
def target_group_name(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "target_group_name", value)
@pulumi.input_type
class ComputeInstanceGroupScalePolicyArgs:
def __init__(__self__, *,
auto_scale: Optional[pulumi.Input['ComputeInstanceGroupScalePolicyAutoScaleArgs']] = None,
fixed_scale: Optional[pulumi.Input['ComputeInstanceGroupScalePolicyFixedScaleArgs']] = None,
test_auto_scale: Optional[pulumi.Input['ComputeInstanceGroupScalePolicyTestAutoScaleArgs']] = None):
"""
:param pulumi.Input['ComputeInstanceGroupScalePolicyAutoScaleArgs'] auto_scale: The auto scaling policy of the instance group. The structure is documented below.
:param pulumi.Input['ComputeInstanceGroupScalePolicyFixedScaleArgs'] fixed_scale: The fixed scaling policy of the instance group. The structure is documented below.
:param pulumi.Input['ComputeInstanceGroupScalePolicyTestAutoScaleArgs'] test_auto_scale: The test auto scaling policy of the instance group. Use it to test how the auto scale works. The structure is documented below.
"""
if auto_scale is not None:
pulumi.set(__self__, "auto_scale", auto_scale)
if fixed_scale is not None:
pulumi.set(__self__, "fixed_scale", fixed_scale)
if test_auto_scale is not None:
pulumi.set(__self__, "test_auto_scale", test_auto_scale)
@property
@pulumi.getter(name="autoScale")
def auto_scale(self) -> Optional[pulumi.Input['ComputeInstanceGroupScalePolicyAutoScaleArgs']]:
"""
The auto scaling policy of the instance group. The structure is documented below.
"""
return pulumi.get(self, "auto_scale")
@auto_scale.setter
def auto_scale(self, value: Optional[pulumi.Input['ComputeInstanceGroupScalePolicyAutoScaleArgs']]):
pulumi.set(self, "auto_scale", value)
@property
@pulumi.getter(name="fixedScale")
def fixed_scale(self) -> Optional[pulumi.Input['ComputeInstanceGroupScalePolicyFixedScaleArgs']]:
"""
The fixed scaling policy of the instance group. The structure is documented below.
"""
return pulumi.get(self, "fixed_scale")
@fixed_scale.setter
def fixed_scale(self, value: Optional[pulumi.Input['ComputeInstanceGroupScalePolicyFixedScaleArgs']]):
pulumi.set(self, "fixed_scale", value)
@property
@pulumi.getter(name="testAutoScale")
def test_auto_scale(self) -> Optional[pulumi.Input['ComputeInstanceGroupScalePolicyTestAutoScaleArgs']]:
"""
The test auto scaling policy of the instance group. Use it to test how the auto scale works. The structure is documented below.
"""
return pulumi.get(self, "test_auto_scale")
@test_auto_scale.setter
def test_auto_scale(self, value: Optional[pulumi.Input['ComputeInstanceGroupScalePolicyTestAutoScaleArgs']]):
pulumi.set(self, "test_auto_scale", value)
@pulumi.input_type
class ComputeInstanceGroupScalePolicyAutoScaleArgs:
def __init__(__self__, *,
initial_size: pulumi.Input[int],
measurement_duration: pulumi.Input[int],
cpu_utilization_target: Optional[pulumi.Input[float]] = None,
custom_rules: Optional[pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupScalePolicyAutoScaleCustomRuleArgs']]]] = None,
max_size: Optional[pulumi.Input[int]] = None,
min_zone_size: Optional[pulumi.Input[int]] = None,
stabilization_duration: Optional[pulumi.Input[int]] = None,
warmup_duration: Optional[pulumi.Input[int]] = None):
"""
:param pulumi.Input[int] initial_size: The initial number of instances in the instance group.
:param pulumi.Input[int] measurement_duration: The amount of time, in seconds, that metrics are averaged for.
If the average value at the end of the interval is higher than the `cpu_utilization_target`,
the instance group will increase the number of virtual machines in the group.
:param pulumi.Input[float] cpu_utilization_target: Target CPU load level.
:param pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupScalePolicyAutoScaleCustomRuleArgs']]] custom_rules: A list of custom rules. The structure is documented below.
:param pulumi.Input[int] max_size: The maximum number of virtual machines in the group.
:param pulumi.Input[int] min_zone_size: The minimum number of virtual machines in a single availability zone.
:param pulumi.Input[int] stabilization_duration: The minimum time interval, in seconds, to monitor the load before
an instance group can reduce the number of virtual machines in the group. During this time, the group
will not decrease even if the average load falls below the value of `cpu_utilization_target`.
:param pulumi.Input[int] warmup_duration: The warm-up time of the virtual machine, in seconds. During this time,
traffic is fed to the virtual machine, but load metrics are not taken into account.
"""
pulumi.set(__self__, "initial_size", initial_size)
pulumi.set(__self__, "measurement_duration", measurement_duration)
if cpu_utilization_target is not None:
pulumi.set(__self__, "cpu_utilization_target", cpu_utilization_target)
if custom_rules is not None:
pulumi.set(__self__, "custom_rules", custom_rules)
if max_size is not None:
pulumi.set(__self__, "max_size", max_size)
if min_zone_size is not None:
pulumi.set(__self__, "min_zone_size", min_zone_size)
if stabilization_duration is not None:
pulumi.set(__self__, "stabilization_duration", stabilization_duration)
if warmup_duration is not None:
pulumi.set(__self__, "warmup_duration", warmup_duration)
@property
@pulumi.getter(name="initialSize")
def initial_size(self) -> pulumi.Input[int]:
"""
The initial number of instances in the instance group.
"""
return pulumi.get(self, "initial_size")
@initial_size.setter
def initial_size(self, value: pulumi.Input[int]):
pulumi.set(self, "initial_size", value)
@property
@pulumi.getter(name="measurementDuration")
def measurement_duration(self) -> pulumi.Input[int]:
"""
The amount of time, in seconds, that metrics are averaged for.
If the average value at the end of the interval is higher than the `cpu_utilization_target`,
the instance group will increase the number of virtual machines in the group.
"""
return pulumi.get(self, "measurement_duration")
@measurement_duration.setter
def measurement_duration(self, value: pulumi.Input[int]):
pulumi.set(self, "measurement_duration", value)
@property
@pulumi.getter(name="cpuUtilizationTarget")
def cpu_utilization_target(self) -> Optional[pulumi.Input[float]]:
"""
Target CPU load level.
"""
return pulumi.get(self, "cpu_utilization_target")
@cpu_utilization_target.setter
def cpu_utilization_target(self, value: Optional[pulumi.Input[float]]):
pulumi.set(self, "cpu_utilization_target", value)
@property
@pulumi.getter(name="customRules")
def custom_rules(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupScalePolicyAutoScaleCustomRuleArgs']]]]:
"""
A list of custom rules. The structure is documented below.
"""
return pulumi.get(self, "custom_rules")
@custom_rules.setter
def custom_rules(self, value: Optional[pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupScalePolicyAutoScaleCustomRuleArgs']]]]):
pulumi.set(self, "custom_rules", value)
@property
@pulumi.getter(name="maxSize")
def max_size(self) -> Optional[pulumi.Input[int]]:
"""
The maximum number of virtual machines in the group.
"""
return pulumi.get(self, "max_size")
@max_size.setter
def max_size(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "max_size", value)
@property
@pulumi.getter(name="minZoneSize")
def min_zone_size(self) -> Optional[pulumi.Input[int]]:
"""
The minimum number of virtual machines in a single availability zone.
"""
return pulumi.get(self, "min_zone_size")
@min_zone_size.setter
def min_zone_size(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "min_zone_size", value)
@property
@pulumi.getter(name="stabilizationDuration")
def stabilization_duration(self) -> Optional[pulumi.Input[int]]:
"""
The minimum time interval, in seconds, to monitor the load before
an instance group can reduce the number of virtual machines in the group. During this time, the group
will not decrease even if the average load falls below the value of `cpu_utilization_target`.
"""
return pulumi.get(self, "stabilization_duration")
@stabilization_duration.setter
def stabilization_duration(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "stabilization_duration", value)
@property
@pulumi.getter(name="warmupDuration")
def warmup_duration(self) -> Optional[pulumi.Input[int]]:
"""
The warm-up time of the virtual machine, in seconds. During this time,
traffic is fed to the virtual machine, but load metrics are not taken into account.
"""
return pulumi.get(self, "warmup_duration")
@warmup_duration.setter
def warmup_duration(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "warmup_duration", value)
@pulumi.input_type
class ComputeInstanceGroupScalePolicyAutoScaleCustomRuleArgs:
def __init__(__self__, *,
metric_name: pulumi.Input[str],
metric_type: pulumi.Input[str],
rule_type: pulumi.Input[str],
target: pulumi.Input[float],
folder_id: Optional[pulumi.Input[str]] = None,
labels: Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]] = None,
service: Optional[pulumi.Input[str]] = None):
"""
:param pulumi.Input[str] metric_name: The name of metric.
:param pulumi.Input[str] metric_type: Metric type, `GAUGE` or `COUNTER`.
:param pulumi.Input[str] rule_type: Rule type: `UTILIZATION` - This type means that the metric applies to one instance.
First, Instance Groups calculates the average metric value for each instance,
then averages the values for instances in one availability zone.
This type of metric must have the `instance_id` label. `WORKLOAD` - This type means that the metric applies to instances in one availability zone.
This type of metric must have the `zone_id` label.
:param pulumi.Input[float] target: Target metric value level.
:param pulumi.Input[str] folder_id: Folder ID of custom metric in Yandex Monitoring that should be used for scaling.
:param pulumi.Input[Mapping[str, pulumi.Input[str]]] labels: A map of labels of metric.
:param pulumi.Input[str] service: Service of custom metric in Yandex Monitoring that should be used for scaling.
"""
pulumi.set(__self__, "metric_name", metric_name)
pulumi.set(__self__, "metric_type", metric_type)
pulumi.set(__self__, "rule_type", rule_type)
pulumi.set(__self__, "target", target)
if folder_id is not None:
pulumi.set(__self__, "folder_id", folder_id)
if labels is not None:
pulumi.set(__self__, "labels", labels)
if service is not None:
pulumi.set(__self__, "service", service)
@property
@pulumi.getter(name="metricName")
def metric_name(self) -> pulumi.Input[str]:
"""
The name of metric.
"""
return pulumi.get(self, "metric_name")
@metric_name.setter
def metric_name(self, value: pulumi.Input[str]):
pulumi.set(self, "metric_name", value)
@property
@pulumi.getter(name="metricType")
def metric_type(self) -> pulumi.Input[str]:
"""
Metric type, `GAUGE` or `COUNTER`.
"""
return pulumi.get(self, "metric_type")
@metric_type.setter
def metric_type(self, value: pulumi.Input[str]):
pulumi.set(self, "metric_type", value)
@property
@pulumi.getter(name="ruleType")
def rule_type(self) -> pulumi.Input[str]:
"""
Rule type: `UTILIZATION` - This type means that the metric applies to one instance.
First, Instance Groups calculates the average metric value for each instance,
then averages the values for instances in one availability zone.
This type of metric must have the `instance_id` label. `WORKLOAD` - This type means that the metric applies to instances in one availability zone.
This type of metric must have the `zone_id` label.
"""
return pulumi.get(self, "rule_type")
@rule_type.setter
def rule_type(self, value: pulumi.Input[str]):
pulumi.set(self, "rule_type", value)
@property
@pulumi.getter
def target(self) -> pulumi.Input[float]:
"""
Target metric value level.
"""
return pulumi.get(self, "target")
@target.setter
def target(self, value: pulumi.Input[float]):
pulumi.set(self, "target", value)
@property
@pulumi.getter(name="folderId")
def folder_id(self) -> Optional[pulumi.Input[str]]:
"""
Folder ID of custom metric in Yandex Monitoring that should be used for scaling.
"""
return pulumi.get(self, "folder_id")
@folder_id.setter
def folder_id(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "folder_id", value)
@property
@pulumi.getter
def labels(self) -> Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]]:
"""
A map of labels of metric.
"""
return pulumi.get(self, "labels")
@labels.setter
def labels(self, value: Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]]):
pulumi.set(self, "labels", value)
@property
@pulumi.getter
def service(self) -> Optional[pulumi.Input[str]]:
"""
Service of custom metric in Yandex Monitoring that should be used for scaling.
"""
return pulumi.get(self, "service")
@service.setter
def service(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "service", value)
@pulumi.input_type
class ComputeInstanceGroupScalePolicyFixedScaleArgs:
def __init__(__self__, *,
size: pulumi.Input[int]):
"""
:param pulumi.Input[int] size: The number of instances in the instance group.
"""
pulumi.set(__self__, "size", size)
@property
@pulumi.getter
def size(self) -> pulumi.Input[int]:
"""
The number of instances in the instance group.
"""
return pulumi.get(self, "size")
@size.setter
def size(self, value: pulumi.Input[int]):
pulumi.set(self, "size", value)
@pulumi.input_type
class ComputeInstanceGroupScalePolicyTestAutoScaleArgs:
def __init__(__self__, *,
initial_size: pulumi.Input[int],
measurement_duration: pulumi.Input[int],
cpu_utilization_target: Optional[pulumi.Input[float]] = None,
custom_rules: Optional[pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupScalePolicyTestAutoScaleCustomRuleArgs']]]] = None,
max_size: Optional[pulumi.Input[int]] = None,
min_zone_size: Optional[pulumi.Input[int]] = None,
stabilization_duration: Optional[pulumi.Input[int]] = None,
warmup_duration: Optional[pulumi.Input[int]] = None):
"""
:param pulumi.Input[int] initial_size: The initial number of instances in the instance group.
:param pulumi.Input[int] measurement_duration: The amount of time, in seconds, that metrics are averaged for.
If the average value at the end of the interval is higher than the `cpu_utilization_target`,
the instance group will increase the number of virtual machines in the group.
:param pulumi.Input[float] cpu_utilization_target: Target CPU load level.
:param pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupScalePolicyTestAutoScaleCustomRuleArgs']]] custom_rules: A list of custom rules. The structure is documented below.
:param pulumi.Input[int] max_size: The maximum number of virtual machines in the group.
:param pulumi.Input[int] min_zone_size: The minimum number of virtual machines in a single availability zone.
:param pulumi.Input[int] stabilization_duration: The minimum time interval, in seconds, to monitor the load before
an instance group can reduce the number of virtual machines in the group. During this time, the group
will not decrease even if the average load falls below the value of `cpu_utilization_target`.
:param pulumi.Input[int] warmup_duration: The warm-up time of the virtual machine, in seconds. During this time,
traffic is fed to the virtual machine, but load metrics are not taken into account.
"""
pulumi.set(__self__, "initial_size", initial_size)
pulumi.set(__self__, "measurement_duration", measurement_duration)
if cpu_utilization_target is not None:
pulumi.set(__self__, "cpu_utilization_target", cpu_utilization_target)
if custom_rules is not None:
pulumi.set(__self__, "custom_rules", custom_rules)
if max_size is not None:
pulumi.set(__self__, "max_size", max_size)
if min_zone_size is not None:
pulumi.set(__self__, "min_zone_size", min_zone_size)
if stabilization_duration is not None:
pulumi.set(__self__, "stabilization_duration", stabilization_duration)
if warmup_duration is not None:
pulumi.set(__self__, "warmup_duration", warmup_duration)
@property
@pulumi.getter(name="initialSize")
def initial_size(self) -> pulumi.Input[int]:
"""
The initial number of instances in the instance group.
"""
return pulumi.get(self, "initial_size")
@initial_size.setter
def initial_size(self, value: pulumi.Input[int]):
pulumi.set(self, "initial_size", value)
@property
@pulumi.getter(name="measurementDuration")
def measurement_duration(self) -> pulumi.Input[int]:
"""
The amount of time, in seconds, that metrics are averaged for.
If the average value at the end of the interval is higher than the `cpu_utilization_target`,
the instance group will increase the number of virtual machines in the group.
"""
return pulumi.get(self, "measurement_duration")
@measurement_duration.setter
def measurement_duration(self, value: pulumi.Input[int]):
pulumi.set(self, "measurement_duration", value)
@property
@pulumi.getter(name="cpuUtilizationTarget")
def cpu_utilization_target(self) -> Optional[pulumi.Input[float]]:
"""
Target CPU load level.
"""
return pulumi.get(self, "cpu_utilization_target")
@cpu_utilization_target.setter
def cpu_utilization_target(self, value: Optional[pulumi.Input[float]]):
pulumi.set(self, "cpu_utilization_target", value)
@property
@pulumi.getter(name="customRules")
def custom_rules(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupScalePolicyTestAutoScaleCustomRuleArgs']]]]:
"""
A list of custom rules. The structure is documented below.
"""
return pulumi.get(self, "custom_rules")
@custom_rules.setter
def custom_rules(self, value: Optional[pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupScalePolicyTestAutoScaleCustomRuleArgs']]]]):
pulumi.set(self, "custom_rules", value)
@property
@pulumi.getter(name="maxSize")
def max_size(self) -> Optional[pulumi.Input[int]]:
"""
The maximum number of virtual machines in the group.
"""
return pulumi.get(self, "max_size")
@max_size.setter
def max_size(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "max_size", value)
@property
@pulumi.getter(name="minZoneSize")
def min_zone_size(self) -> Optional[pulumi.Input[int]]:
"""
The minimum number of virtual machines in a single availability zone.
"""
return pulumi.get(self, "min_zone_size")
@min_zone_size.setter
def min_zone_size(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "min_zone_size", value)
@property
@pulumi.getter(name="stabilizationDuration")
def stabilization_duration(self) -> Optional[pulumi.Input[int]]:
"""
The minimum time interval, in seconds, to monitor the load before
an instance group can reduce the number of virtual machines in the group. During this time, the group
will not decrease even if the average load falls below the value of `cpu_utilization_target`.
"""
return pulumi.get(self, "stabilization_duration")
@stabilization_duration.setter
def stabilization_duration(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "stabilization_duration", value)
@property
@pulumi.getter(name="warmupDuration")
def warmup_duration(self) -> Optional[pulumi.Input[int]]:
"""
The warm-up time of the virtual machine, in seconds. During this time,
traffic is fed to the virtual machine, but load metrics are not taken into account.
"""
return pulumi.get(self, "warmup_duration")
@warmup_duration.setter
def warmup_duration(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "warmup_duration", value)
@pulumi.input_type
class ComputeInstanceGroupScalePolicyTestAutoScaleCustomRuleArgs:
def __init__(__self__, *,
metric_name: pulumi.Input[str],
metric_type: pulumi.Input[str],
rule_type: pulumi.Input[str],
target: pulumi.Input[float],
folder_id: Optional[pulumi.Input[str]] = None,
labels: Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]] = None,
service: Optional[pulumi.Input[str]] = None):
"""
:param pulumi.Input[str] metric_name: The name of metric.
:param pulumi.Input[str] metric_type: Metric type, `GAUGE` or `COUNTER`.
:param pulumi.Input[str] rule_type: Rule type: `UTILIZATION` - This type means that the metric applies to one instance.
First, Instance Groups calculates the average metric value for each instance,
then averages the values for instances in one availability zone.
This type of metric must have the `instance_id` label. `WORKLOAD` - This type means that the metric applies to instances in one availability zone.
This type of metric must have the `zone_id` label.
:param pulumi.Input[float] target: Target metric value level.
:param pulumi.Input[str] folder_id: Folder ID of custom metric in Yandex Monitoring that should be used for scaling.
:param pulumi.Input[Mapping[str, pulumi.Input[str]]] labels: A map of labels of metric.
:param pulumi.Input[str] service: Service of custom metric in Yandex Monitoring that should be used for scaling.
"""
pulumi.set(__self__, "metric_name", metric_name)
pulumi.set(__self__, "metric_type", metric_type)
pulumi.set(__self__, "rule_type", rule_type)
pulumi.set(__self__, "target", target)
if folder_id is not None:
pulumi.set(__self__, "folder_id", folder_id)
if labels is not None:
pulumi.set(__self__, "labels", labels)
if service is not None:
pulumi.set(__self__, "service", service)
@property
@pulumi.getter(name="metricName")
def metric_name(self) -> pulumi.Input[str]:
"""
The name of metric.
"""
return pulumi.get(self, "metric_name")
@metric_name.setter
def metric_name(self, value: pulumi.Input[str]):
pulumi.set(self, "metric_name", value)
@property
@pulumi.getter(name="metricType")
def metric_type(self) -> pulumi.Input[str]:
"""
Metric type, `GAUGE` or `COUNTER`.
"""
return pulumi.get(self, "metric_type")
@metric_type.setter
def metric_type(self, value: pulumi.Input[str]):
pulumi.set(self, "metric_type", value)
@property
@pulumi.getter(name="ruleType")
def rule_type(self) -> pulumi.Input[str]:
"""
Rule type: `UTILIZATION` - This type means that the metric applies to one instance.
First, Instance Groups calculates the average metric value for each instance,
then averages the values for instances in one availability zone.
This type of metric must have the `instance_id` label. `WORKLOAD` - This type means that the metric applies to instances in one availability zone.
This type of metric must have the `zone_id` label.
"""
return pulumi.get(self, "rule_type")
@rule_type.setter
def rule_type(self, value: pulumi.Input[str]):
pulumi.set(self, "rule_type", value)
@property
@pulumi.getter
def target(self) -> pulumi.Input[float]:
"""
Target metric value level.
"""
return pulumi.get(self, "target")
@target.setter
def target(self, value: pulumi.Input[float]):
pulumi.set(self, "target", value)
@property
@pulumi.getter(name="folderId")
def folder_id(self) -> Optional[pulumi.Input[str]]:
"""
Folder ID of custom metric in Yandex Monitoring that should be used for scaling.
"""
return pulumi.get(self, "folder_id")
@folder_id.setter
def folder_id(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "folder_id", value)
@property
@pulumi.getter
def labels(self) -> Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]]:
"""
A map of labels of metric.
"""
return pulumi.get(self, "labels")
@labels.setter
def labels(self, value: Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]]):
pulumi.set(self, "labels", value)
@property
@pulumi.getter
def service(self) -> Optional[pulumi.Input[str]]:
"""
Service of custom metric in Yandex Monitoring that should be used for scaling.
"""
return pulumi.get(self, "service")
@service.setter
def service(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "service", value)
@pulumi.input_type
class ComputeInstanceNetworkInterfaceArgs:
def __init__(__self__, *,
subnet_id: pulumi.Input[str],
dns_records: Optional[pulumi.Input[Sequence[pulumi.Input['ComputeInstanceNetworkInterfaceDnsRecordArgs']]]] = None,
index: Optional[pulumi.Input[int]] = None,
ip_address: Optional[pulumi.Input[str]] = None,
ipv4: Optional[pulumi.Input[bool]] = None,
ipv6: Optional[pulumi.Input[bool]] = None,
ipv6_address: Optional[pulumi.Input[str]] = None,
ipv6_dns_records: Optional[pulumi.Input[Sequence[pulumi.Input['ComputeInstanceNetworkInterfaceIpv6DnsRecordArgs']]]] = None,
mac_address: Optional[pulumi.Input[str]] = None,
nat: Optional[pulumi.Input[bool]] = None,
nat_dns_records: Optional[pulumi.Input[Sequence[pulumi.Input['ComputeInstanceNetworkInterfaceNatDnsRecordArgs']]]] = None,
nat_ip_address: Optional[pulumi.Input[str]] = None,
nat_ip_version: Optional[pulumi.Input[str]] = None,
security_group_ids: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None):
"""
:param pulumi.Input[str] subnet_id: ID of the subnet to attach this
interface to. The subnet must exist in the same zone where this instance will be
created.
:param pulumi.Input[Sequence[pulumi.Input['ComputeInstanceNetworkInterfaceDnsRecordArgs']]] dns_records: List of configurations for creating ipv4 DNS records. The structure is documented below.
:param pulumi.Input[str] ip_address: The private IP address to assign to the instance. If
empty, the address will be automatically assigned from the specified subnet.
:param pulumi.Input[bool] ipv4: Allocate an IPv4 address for the interface. The default value is `true`.
:param pulumi.Input[bool] ipv6: If true, allocate an IPv6 address for the interface.
The address will be automatically assigned from the specified subnet.
:param pulumi.Input[str] ipv6_address: The private IPv6 address to assign to the instance.
:param pulumi.Input[Sequence[pulumi.Input['ComputeInstanceNetworkInterfaceIpv6DnsRecordArgs']]] ipv6_dns_records: List of configurations for creating ipv6 DNS records. The structure is documented below.
:param pulumi.Input[bool] nat: Provide a public address, for instance, to access the internet over NAT.
:param pulumi.Input[Sequence[pulumi.Input['ComputeInstanceNetworkInterfaceNatDnsRecordArgs']]] nat_dns_records: List of configurations for creating ipv4 NAT DNS records. The structure is documented below.
:param pulumi.Input[str] nat_ip_address: Provide a public address, for instance, to access the internet over NAT. Address should be already reserved in web UI.
:param pulumi.Input[Sequence[pulumi.Input[str]]] security_group_ids: Security group ids for network interface.
"""
pulumi.set(__self__, "subnet_id", subnet_id)
if dns_records is not None:
pulumi.set(__self__, "dns_records", dns_records)
if index is not None:
pulumi.set(__self__, "index", index)
if ip_address is not None:
pulumi.set(__self__, "ip_address", ip_address)
if ipv4 is not None:
pulumi.set(__self__, "ipv4", ipv4)
if ipv6 is not None:
pulumi.set(__self__, "ipv6", ipv6)
if ipv6_address is not None:
pulumi.set(__self__, "ipv6_address", ipv6_address)
if ipv6_dns_records is not None:
pulumi.set(__self__, "ipv6_dns_records", ipv6_dns_records)
if mac_address is not None:
pulumi.set(__self__, "mac_address", mac_address)
if nat is not None:
pulumi.set(__self__, "nat", nat)
if nat_dns_records is not None:
pulumi.set(__self__, "nat_dns_records", nat_dns_records)
if nat_ip_address is not None:
pulumi.set(__self__, "nat_ip_address", nat_ip_address)
if nat_ip_version is not None:
pulumi.set(__self__, "nat_ip_version", nat_ip_version)
if security_group_ids is not None:
pulumi.set(__self__, "security_group_ids", security_group_ids)
@property
@pulumi.getter(name="subnetId")
def subnet_id(self) -> pulumi.Input[str]:
"""
ID of the subnet to attach this
interface to. The subnet must exist in the same zone where this instance will be
created.
"""
return pulumi.get(self, "subnet_id")
@subnet_id.setter
def subnet_id(self, value: pulumi.Input[str]):
pulumi.set(self, "subnet_id", value)
@property
@pulumi.getter(name="dnsRecords")
def dns_records(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['ComputeInstanceNetworkInterfaceDnsRecordArgs']]]]:
"""
List of configurations for creating ipv4 DNS records. The structure is documented below.
"""
return pulumi.get(self, "dns_records")
@dns_records.setter
def dns_records(self, value: Optional[pulumi.Input[Sequence[pulumi.Input['ComputeInstanceNetworkInterfaceDnsRecordArgs']]]]):
pulumi.set(self, "dns_records", value)
@property
@pulumi.getter
def index(self) -> Optional[pulumi.Input[int]]:
return pulumi.get(self, "index")
@index.setter
def index(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "index", value)
@property
@pulumi.getter(name="ipAddress")
def ip_address(self) -> Optional[pulumi.Input[str]]:
"""
The private IP address to assign to the instance. If
empty, the address will be automatically assigned from the specified subnet.
"""
return pulumi.get(self, "ip_address")
@ip_address.setter
def ip_address(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "ip_address", value)
@property
@pulumi.getter
def ipv4(self) -> Optional[pulumi.Input[bool]]:
"""
Allocate an IPv4 address for the interface. The default value is `true`.
"""
return pulumi.get(self, "ipv4")
@ipv4.setter
def ipv4(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "ipv4", value)
@property
@pulumi.getter
def ipv6(self) -> Optional[pulumi.Input[bool]]:
"""
If true, allocate an IPv6 address for the interface.
The address will be automatically assigned from the specified subnet.
"""
return pulumi.get(self, "ipv6")
@ipv6.setter
def ipv6(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "ipv6", value)
@property
@pulumi.getter(name="ipv6Address")
def ipv6_address(self) -> Optional[pulumi.Input[str]]:
"""
The private IPv6 address to assign to the instance.
"""
return pulumi.get(self, "ipv6_address")
@ipv6_address.setter
def ipv6_address(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "ipv6_address", value)
@property
@pulumi.getter(name="ipv6DnsRecords")
def ipv6_dns_records(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['ComputeInstanceNetworkInterfaceIpv6DnsRecordArgs']]]]:
"""
List of configurations for creating ipv6 DNS records. The structure is documented below.
"""
return pulumi.get(self, "ipv6_dns_records")
@ipv6_dns_records.setter
def ipv6_dns_records(self, value: Optional[pulumi.Input[Sequence[pulumi.Input['ComputeInstanceNetworkInterfaceIpv6DnsRecordArgs']]]]):
pulumi.set(self, "ipv6_dns_records", value)
@property
@pulumi.getter(name="macAddress")
def mac_address(self) -> Optional[pulumi.Input[str]]:
return pulumi.get(self, "mac_address")
@mac_address.setter
def mac_address(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "mac_address", value)
@property
@pulumi.getter
def nat(self) -> Optional[pulumi.Input[bool]]:
"""
Provide a public address, for instance, to access the internet over NAT.
"""
return pulumi.get(self, "nat")
@nat.setter
def nat(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "nat", value)
@property
@pulumi.getter(name="natDnsRecords")
def nat_dns_records(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['ComputeInstanceNetworkInterfaceNatDnsRecordArgs']]]]:
"""
List of configurations for creating ipv4 NAT DNS records. The structure is documented below.
"""
return pulumi.get(self, "nat_dns_records")
@nat_dns_records.setter
def nat_dns_records(self, value: Optional[pulumi.Input[Sequence[pulumi.Input['ComputeInstanceNetworkInterfaceNatDnsRecordArgs']]]]):
pulumi.set(self, "nat_dns_records", value)
@property
@pulumi.getter(name="natIpAddress")
def nat_ip_address(self) -> Optional[pulumi.Input[str]]:
"""
Provide a public address, for instance, to access the internet over NAT. Address should be already reserved in web UI.
"""
return pulumi.get(self, "nat_ip_address")
@nat_ip_address.setter
def nat_ip_address(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "nat_ip_address", value)
@property
@pulumi.getter(name="natIpVersion")
def nat_ip_version(self) -> Optional[pulumi.Input[str]]:
return pulumi.get(self, "nat_ip_version")
@nat_ip_version.setter
def nat_ip_version(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "nat_ip_version", value)
@property
@pulumi.getter(name="securityGroupIds")
def security_group_ids(self) -> Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]:
"""
Security group ids for network interface.
"""
return pulumi.get(self, "security_group_ids")
@security_group_ids.setter
def security_group_ids(self, value: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]):
pulumi.set(self, "security_group_ids", value)
@pulumi.input_type
class ComputeInstanceNetworkInterfaceDnsRecordArgs:
def __init__(__self__, *,
fqdn: pulumi.Input[str],
dns_zone_id: Optional[pulumi.Input[str]] = None,
ptr: Optional[pulumi.Input[bool]] = None,
ttl: Optional[pulumi.Input[int]] = None):
"""
:param pulumi.Input[str] fqdn: DNS record FQDN (must have a dot at the end).
:param pulumi.Input[str] dns_zone_id: DNS zone ID (if not set, private zone used).
:param pulumi.Input[bool] ptr: When set to true, also create a PTR DNS record.
:param pulumi.Input[int] ttl: DNS record TTL. in seconds
"""
pulumi.set(__self__, "fqdn", fqdn)
if dns_zone_id is not None:
pulumi.set(__self__, "dns_zone_id", dns_zone_id)
if ptr is not None:
pulumi.set(__self__, "ptr", ptr)
if ttl is not None:
pulumi.set(__self__, "ttl", ttl)
@property
@pulumi.getter
def fqdn(self) -> pulumi.Input[str]:
"""
DNS record FQDN (must have a dot at the end).
"""
return pulumi.get(self, "fqdn")
@fqdn.setter
def fqdn(self, value: pulumi.Input[str]):
pulumi.set(self, "fqdn", value)
@property
@pulumi.getter(name="dnsZoneId")
def dns_zone_id(self) -> Optional[pulumi.Input[str]]:
"""
DNS zone ID (if not set, private zone used).
"""
return pulumi.get(self, "dns_zone_id")
@dns_zone_id.setter
def dns_zone_id(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "dns_zone_id", value)
@property
@pulumi.getter
def ptr(self) -> Optional[pulumi.Input[bool]]:
"""
When set to true, also create a PTR DNS record.
"""
return pulumi.get(self, "ptr")
@ptr.setter
def ptr(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "ptr", value)
@property
@pulumi.getter
def ttl(self) -> Optional[pulumi.Input[int]]:
"""
DNS record TTL. in seconds
"""
return pulumi.get(self, "ttl")
@ttl.setter
def ttl(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "ttl", value)
@pulumi.input_type
class ComputeInstanceNetworkInterfaceIpv6DnsRecordArgs:
def __init__(__self__, *,
fqdn: pulumi.Input[str],
dns_zone_id: Optional[pulumi.Input[str]] = None,
ptr: Optional[pulumi.Input[bool]] = None,
ttl: Optional[pulumi.Input[int]] = None):
"""
:param pulumi.Input[str] fqdn: DNS record FQDN (must have a dot at the end).
:param pulumi.Input[str] dns_zone_id: DNS zone ID (if not set, private zone used).
:param pulumi.Input[bool] ptr: When set to true, also create a PTR DNS record.
:param pulumi.Input[int] ttl: DNS record TTL. in seconds
"""
pulumi.set(__self__, "fqdn", fqdn)
if dns_zone_id is not None:
pulumi.set(__self__, "dns_zone_id", dns_zone_id)
if ptr is not None:
pulumi.set(__self__, "ptr", ptr)
if ttl is not None:
pulumi.set(__self__, "ttl", ttl)
@property
@pulumi.getter
def fqdn(self) -> pulumi.Input[str]:
"""
DNS record FQDN (must have a dot at the end).
"""
return pulumi.get(self, "fqdn")
@fqdn.setter
def fqdn(self, value: pulumi.Input[str]):
pulumi.set(self, "fqdn", value)
@property
@pulumi.getter(name="dnsZoneId")
def dns_zone_id(self) -> Optional[pulumi.Input[str]]:
"""
DNS zone ID (if not set, private zone used).
"""
return pulumi.get(self, "dns_zone_id")
@dns_zone_id.setter
def dns_zone_id(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "dns_zone_id", value)
@property
@pulumi.getter
def ptr(self) -> Optional[pulumi.Input[bool]]:
"""
When set to true, also create a PTR DNS record.
"""
return pulumi.get(self, "ptr")
@ptr.setter
def ptr(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "ptr", value)
@property
@pulumi.getter
def ttl(self) -> Optional[pulumi.Input[int]]:
"""
DNS record TTL. in seconds
"""
return pulumi.get(self, "ttl")
@ttl.setter
def ttl(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "ttl", value)
@pulumi.input_type
class ComputeInstanceNetworkInterfaceNatDnsRecordArgs:
def __init__(__self__, *,
fqdn: pulumi.Input[str],
dns_zone_id: Optional[pulumi.Input[str]] = None,
ptr: Optional[pulumi.Input[bool]] = None,
ttl: Optional[pulumi.Input[int]] = None):
"""
:param pulumi.Input[str] fqdn: DNS record FQDN (must have a dot at the end).
:param pulumi.Input[str] dns_zone_id: DNS zone ID (if not set, private zone used).
:param pulumi.Input[bool] ptr: When set to true, also create a PTR DNS record.
:param pulumi.Input[int] ttl: DNS record TTL. in seconds
"""
pulumi.set(__self__, "fqdn", fqdn)
if dns_zone_id is not None:
pulumi.set(__self__, "dns_zone_id", dns_zone_id)
if ptr is not None:
pulumi.set(__self__, "ptr", ptr)
if ttl is not None:
pulumi.set(__self__, "ttl", ttl)
@property
@pulumi.getter
def fqdn(self) -> pulumi.Input[str]:
"""
DNS record FQDN (must have a dot at the end).
"""
return pulumi.get(self, "fqdn")
@fqdn.setter
def fqdn(self, value: pulumi.Input[str]):
pulumi.set(self, "fqdn", value)
@property
@pulumi.getter(name="dnsZoneId")
def dns_zone_id(self) -> Optional[pulumi.Input[str]]:
"""
DNS zone ID (if not set, private zone used).
"""
return pulumi.get(self, "dns_zone_id")
@dns_zone_id.setter
def dns_zone_id(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "dns_zone_id", value)
@property
@pulumi.getter
def ptr(self) -> Optional[pulumi.Input[bool]]:
"""
When set to true, also create a PTR DNS record.
"""
return pulumi.get(self, "ptr")
@ptr.setter
def ptr(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "ptr", value)
@property
@pulumi.getter
def ttl(self) -> Optional[pulumi.Input[int]]:
"""
DNS record TTL. in seconds
"""
return pulumi.get(self, "ttl")
@ttl.setter
def ttl(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "ttl", value)
@pulumi.input_type
class ComputeInstancePlacementPolicyArgs:
def __init__(__self__, *,
placement_group_id: pulumi.Input[str]):
"""
:param pulumi.Input[str] placement_group_id: Specifies the id of the Placement Group to assign to the instance.
"""
pulumi.set(__self__, "placement_group_id", placement_group_id)
@property
@pulumi.getter(name="placementGroupId")
def placement_group_id(self) -> pulumi.Input[str]:
"""
Specifies the id of the Placement Group to assign to the instance.
"""
return pulumi.get(self, "placement_group_id")
@placement_group_id.setter
def placement_group_id(self, value: pulumi.Input[str]):
pulumi.set(self, "placement_group_id", value)
@pulumi.input_type
class ComputeInstanceResourcesArgs:
def __init__(__self__, *,
cores: pulumi.Input[int],
memory: pulumi.Input[float],
core_fraction: Optional[pulumi.Input[int]] = None,
gpus: Optional[pulumi.Input[int]] = None):
"""
:param pulumi.Input[int] cores: CPU cores for the instance.
:param pulumi.Input[float] memory: Memory size in GB.
:param pulumi.Input[int] core_fraction: If provided, specifies baseline performance for a core as a percent.
"""
pulumi.set(__self__, "cores", cores)
pulumi.set(__self__, "memory", memory)
if core_fraction is not None:
pulumi.set(__self__, "core_fraction", core_fraction)
if gpus is not None:
pulumi.set(__self__, "gpus", gpus)
@property
@pulumi.getter
def cores(self) -> pulumi.Input[int]:
"""
CPU cores for the instance.
"""
return pulumi.get(self, "cores")
@cores.setter
def cores(self, value: pulumi.Input[int]):
pulumi.set(self, "cores", value)
@property
@pulumi.getter
def memory(self) -> pulumi.Input[float]:
"""
Memory size in GB.
"""
return pulumi.get(self, "memory")
@memory.setter
def memory(self, value: pulumi.Input[float]):
pulumi.set(self, "memory", value)
@property
@pulumi.getter(name="coreFraction")
def core_fraction(self) -> Optional[pulumi.Input[int]]:
"""
If provided, specifies baseline performance for a core as a percent.
"""
return pulumi.get(self, "core_fraction")
@core_fraction.setter
def core_fraction(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "core_fraction", value)
@property
@pulumi.getter
def gpus(self) -> Optional[pulumi.Input[int]]:
return pulumi.get(self, "gpus")
@gpus.setter
def gpus(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "gpus", value)
@pulumi.input_type
class ComputeInstanceSchedulingPolicyArgs:
def __init__(__self__, *,
preemptible: Optional[pulumi.Input[bool]] = None):
"""
:param pulumi.Input[bool] preemptible: Specifies if the instance is preemptible. Defaults to false.
"""
if preemptible is not None:
pulumi.set(__self__, "preemptible", preemptible)
@property
@pulumi.getter
def preemptible(self) -> Optional[pulumi.Input[bool]]:
"""
Specifies if the instance is preemptible. Defaults to false.
"""
return pulumi.get(self, "preemptible")
@preemptible.setter
def preemptible(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "preemptible", value)
@pulumi.input_type
class ComputeInstanceSecondaryDiskArgs:
def __init__(__self__, *,
disk_id: pulumi.Input[str],
auto_delete: Optional[pulumi.Input[bool]] = None,
device_name: Optional[pulumi.Input[str]] = None,
mode: Optional[pulumi.Input[str]] = None):
"""
:param pulumi.Input[str] disk_id: ID of the disk that is attached to the instance.
:param pulumi.Input[bool] auto_delete: Whether the disk is auto-deleted when the instance
is deleted. The default value is false.
:param pulumi.Input[str] device_name: Name that can be used to access an attached disk
under `/dev/disk/by-id/`.
:param pulumi.Input[str] mode: Type of access to the disk resource. By default, a disk is attached in `READ_WRITE` mode.
"""
pulumi.set(__self__, "disk_id", disk_id)
if auto_delete is not None:
pulumi.set(__self__, "auto_delete", auto_delete)
if device_name is not None:
pulumi.set(__self__, "device_name", device_name)
if mode is not None:
pulumi.set(__self__, "mode", mode)
@property
@pulumi.getter(name="diskId")
def disk_id(self) -> pulumi.Input[str]:
"""
ID of the disk that is attached to the instance.
"""
return pulumi.get(self, "disk_id")
@disk_id.setter
def disk_id(self, value: pulumi.Input[str]):
pulumi.set(self, "disk_id", value)
@property
@pulumi.getter(name="autoDelete")
def auto_delete(self) -> Optional[pulumi.Input[bool]]:
"""
Whether the disk is auto-deleted when the instance
is deleted. The default value is false.
"""
return pulumi.get(self, "auto_delete")
@auto_delete.setter
def auto_delete(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "auto_delete", value)
@property
@pulumi.getter(name="deviceName")
def device_name(self) -> Optional[pulumi.Input[str]]:
"""
Name that can be used to access an attached disk
under `/dev/disk/by-id/`.
"""
return pulumi.get(self, "device_name")
@device_name.setter
def device_name(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "device_name", value)
@property
@pulumi.getter
def mode(self) -> Optional[pulumi.Input[str]]:
"""
Type of access to the disk resource. By default, a disk is attached in `READ_WRITE` mode.
"""
return pulumi.get(self, "mode")
@mode.setter
def mode(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "mode", value)
@pulumi.input_type
class DataprocClusterClusterConfigArgs:
def __init__(__self__, *,
subcluster_specs: pulumi.Input[Sequence[pulumi.Input['DataprocClusterClusterConfigSubclusterSpecArgs']]],
hadoop: Optional[pulumi.Input['DataprocClusterClusterConfigHadoopArgs']] = None,
version_id: Optional[pulumi.Input[str]] = None):
"""
:param pulumi.Input[Sequence[pulumi.Input['DataprocClusterClusterConfigSubclusterSpecArgs']]] subcluster_specs: Configuration of the Data Proc subcluster. The structure is documented below.
:param pulumi.Input['DataprocClusterClusterConfigHadoopArgs'] hadoop: Data Proc specific options. The structure is documented below.
:param pulumi.Input[str] version_id: Version of Data Proc image.
"""
pulumi.set(__self__, "subcluster_specs", subcluster_specs)
if hadoop is not None:
pulumi.set(__self__, "hadoop", hadoop)
if version_id is not None:
pulumi.set(__self__, "version_id", version_id)
@property
@pulumi.getter(name="subclusterSpecs")
def subcluster_specs(self) -> pulumi.Input[Sequence[pulumi.Input['DataprocClusterClusterConfigSubclusterSpecArgs']]]:
"""
Configuration of the Data Proc subcluster. The structure is documented below.
"""
return pulumi.get(self, "subcluster_specs")
@subcluster_specs.setter
def subcluster_specs(self, value: pulumi.Input[Sequence[pulumi.Input['DataprocClusterClusterConfigSubclusterSpecArgs']]]):
pulumi.set(self, "subcluster_specs", value)
@property
@pulumi.getter
def hadoop(self) -> Optional[pulumi.Input['DataprocClusterClusterConfigHadoopArgs']]:
"""
Data Proc specific options. The structure is documented below.
"""
return pulumi.get(self, "hadoop")
@hadoop.setter
def hadoop(self, value: Optional[pulumi.Input['DataprocClusterClusterConfigHadoopArgs']]):
pulumi.set(self, "hadoop", value)
@property
@pulumi.getter(name="versionId")
def version_id(self) -> Optional[pulumi.Input[str]]:
"""
Version of Data Proc image.
"""
return pulumi.get(self, "version_id")
@version_id.setter
def version_id(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "version_id", value)
@pulumi.input_type
class DataprocClusterClusterConfigHadoopArgs:
def __init__(__self__, *,
properties: Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]] = None,
services: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None,
ssh_public_keys: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None):
"""
:param pulumi.Input[Mapping[str, pulumi.Input[str]]] properties: A set of key/value pairs that are used to configure cluster services.
:param pulumi.Input[Sequence[pulumi.Input[str]]] services: List of services to run on Data Proc cluster.
:param pulumi.Input[Sequence[pulumi.Input[str]]] ssh_public_keys: List of SSH public keys to put to the hosts of the cluster. For information on how to connect to the cluster, see [the official documentation](https://cloud.yandex.com/docs/data-proc/operations/connect).
"""
if properties is not None:
pulumi.set(__self__, "properties", properties)
if services is not None:
pulumi.set(__self__, "services", services)
if ssh_public_keys is not None:
pulumi.set(__self__, "ssh_public_keys", ssh_public_keys)
@property
@pulumi.getter
def properties(self) -> Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]]:
"""
A set of key/value pairs that are used to configure cluster services.
"""
return pulumi.get(self, "properties")
@properties.setter
def properties(self, value: Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]]):
pulumi.set(self, "properties", value)
@property
@pulumi.getter
def services(self) -> Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]:
"""
List of services to run on Data Proc cluster.
"""
return pulumi.get(self, "services")
@services.setter
def services(self, value: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]):
pulumi.set(self, "services", value)
@property
@pulumi.getter(name="sshPublicKeys")
def ssh_public_keys(self) -> Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]:
"""
List of SSH public keys to put to the hosts of the cluster. For information on how to connect to the cluster, see [the official documentation](https://cloud.yandex.com/docs/data-proc/operations/connect).
"""
return pulumi.get(self, "ssh_public_keys")
@ssh_public_keys.setter
def ssh_public_keys(self, value: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]):
pulumi.set(self, "ssh_public_keys", value)
@pulumi.input_type
class DataprocClusterClusterConfigSubclusterSpecArgs:
def __init__(__self__, *,
hosts_count: pulumi.Input[int],
name: pulumi.Input[str],
resources: pulumi.Input['DataprocClusterClusterConfigSubclusterSpecResourcesArgs'],
role: pulumi.Input[str],
subnet_id: pulumi.Input[str],
autoscaling_config: Optional[pulumi.Input['DataprocClusterClusterConfigSubclusterSpecAutoscalingConfigArgs']] = None,
id: Optional[pulumi.Input[str]] = None):
"""
:param pulumi.Input[int] hosts_count: Number of hosts within Data Proc subcluster.
:param pulumi.Input[str] name: Name of the Data Proc subcluster.
:param pulumi.Input['DataprocClusterClusterConfigSubclusterSpecResourcesArgs'] resources: Resources allocated to each host of the Data Proc subcluster. The structure is documented below.
:param pulumi.Input[str] role: Role of the subcluster in the Data Proc cluster.
:param pulumi.Input[str] subnet_id: The ID of the subnet, to which hosts of the subcluster belong. Subnets of all the subclusters must belong to the same VPC network.
:param pulumi.Input['DataprocClusterClusterConfigSubclusterSpecAutoscalingConfigArgs'] autoscaling_config: Autoscaling configuration for compute subclusters.
:param pulumi.Input[str] id: (Computed) ID of a new Data Proc cluster.
"""
pulumi.set(__self__, "hosts_count", hosts_count)
pulumi.set(__self__, "name", name)
pulumi.set(__self__, "resources", resources)
pulumi.set(__self__, "role", role)
pulumi.set(__self__, "subnet_id", subnet_id)
if autoscaling_config is not None:
pulumi.set(__self__, "autoscaling_config", autoscaling_config)
if id is not None:
pulumi.set(__self__, "id", id)
@property
@pulumi.getter(name="hostsCount")
def hosts_count(self) -> pulumi.Input[int]:
"""
Number of hosts within Data Proc subcluster.
"""
return pulumi.get(self, "hosts_count")
@hosts_count.setter
def hosts_count(self, value: pulumi.Input[int]):
pulumi.set(self, "hosts_count", value)
@property
@pulumi.getter
def name(self) -> pulumi.Input[str]:
"""
Name of the Data Proc subcluster.
"""
return pulumi.get(self, "name")
@name.setter
def name(self, value: pulumi.Input[str]):
pulumi.set(self, "name", value)
@property
@pulumi.getter
def resources(self) -> pulumi.Input['DataprocClusterClusterConfigSubclusterSpecResourcesArgs']:
"""
Resources allocated to each host of the Data Proc subcluster. The structure is documented below.
"""
return pulumi.get(self, "resources")
@resources.setter
def resources(self, value: pulumi.Input['DataprocClusterClusterConfigSubclusterSpecResourcesArgs']):
pulumi.set(self, "resources", value)
@property
@pulumi.getter
def role(self) -> pulumi.Input[str]:
"""
Role of the subcluster in the Data Proc cluster.
"""
return pulumi.get(self, "role")
@role.setter
def role(self, value: pulumi.Input[str]):
pulumi.set(self, "role", value)
@property
@pulumi.getter(name="subnetId")
def subnet_id(self) -> pulumi.Input[str]:
"""
The ID of the subnet, to which hosts of the subcluster belong. Subnets of all the subclusters must belong to the same VPC network.
"""
return pulumi.get(self, "subnet_id")
@subnet_id.setter
def subnet_id(self, value: pulumi.Input[str]):
pulumi.set(self, "subnet_id", value)
@property
@pulumi.getter(name="autoscalingConfig")
def autoscaling_config(self) -> Optional[pulumi.Input['DataprocClusterClusterConfigSubclusterSpecAutoscalingConfigArgs']]:
"""
Autoscaling configuration for compute subclusters.
"""
return pulumi.get(self, "autoscaling_config")
@autoscaling_config.setter
def autoscaling_config(self, value: Optional[pulumi.Input['DataprocClusterClusterConfigSubclusterSpecAutoscalingConfigArgs']]):
pulumi.set(self, "autoscaling_config", value)
@property
@pulumi.getter
def id(self) -> Optional[pulumi.Input[str]]:
"""
(Computed) ID of a new Data Proc cluster.
"""
return pulumi.get(self, "id")
@id.setter
def id(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "id", value)
@pulumi.input_type
class DataprocClusterClusterConfigSubclusterSpecAutoscalingConfigArgs:
def __init__(__self__, *,
max_hosts_count: pulumi.Input[int],
cpu_utilization_target: Optional[pulumi.Input[float]] = None,
decommission_timeout: Optional[pulumi.Input[int]] = None,
measurement_duration: Optional[pulumi.Input[int]] = None,
preemptible: Optional[pulumi.Input[bool]] = None,
stabilization_duration: Optional[pulumi.Input[int]] = None,
warmup_duration: Optional[pulumi.Input[int]] = None):
"""
:param pulumi.Input[int] max_hosts_count: Maximum number of nodes in autoscaling subclusters.
:param pulumi.Input[float] cpu_utilization_target: Defines an autoscaling rule based on the average CPU utilization of the instance group. If not set default autoscaling metric will be used.
:param pulumi.Input[int] decommission_timeout: Timeout to gracefully decommission nodes during downscaling. In seconds.
:param pulumi.Input[int] measurement_duration: Time in seconds allotted for averaging metrics.
:param pulumi.Input[bool] preemptible: Bool flag -- whether to use preemptible compute instances. Preemptible instances are stopped at least once every 24 hours, and can be stopped at any time if their resources are needed by Compute. For more information, see [Preemptible Virtual Machines](https://cloud.yandex.com/docs/compute/concepts/preemptible-vm).
:param pulumi.Input[int] stabilization_duration: Minimum amount of time in seconds allotted for monitoring before Instance Groups can reduce the number of instances in the group. During this time, the group size doesn't decrease, even if the new metric values indicate that it should.
:param pulumi.Input[int] warmup_duration: The warmup time of the instance in seconds. During this time, traffic is sent to the instance, but instance metrics are not collected.
"""
pulumi.set(__self__, "max_hosts_count", max_hosts_count)
if cpu_utilization_target is not None:
pulumi.set(__self__, "cpu_utilization_target", cpu_utilization_target)
if decommission_timeout is not None:
pulumi.set(__self__, "decommission_timeout", decommission_timeout)
if measurement_duration is not None:
pulumi.set(__self__, "measurement_duration", measurement_duration)
if preemptible is not None:
pulumi.set(__self__, "preemptible", preemptible)
if stabilization_duration is not None:
pulumi.set(__self__, "stabilization_duration", stabilization_duration)
if warmup_duration is not None:
pulumi.set(__self__, "warmup_duration", warmup_duration)
@property
@pulumi.getter(name="maxHostsCount")
def max_hosts_count(self) -> pulumi.Input[int]:
"""
Maximum number of nodes in autoscaling subclusters.
"""
return pulumi.get(self, "max_hosts_count")
@max_hosts_count.setter
def max_hosts_count(self, value: pulumi.Input[int]):
pulumi.set(self, "max_hosts_count", value)
@property
@pulumi.getter(name="cpuUtilizationTarget")
def cpu_utilization_target(self) -> Optional[pulumi.Input[float]]:
"""
Defines an autoscaling rule based on the average CPU utilization of the instance group. If not set default autoscaling metric will be used.
"""
return pulumi.get(self, "cpu_utilization_target")
@cpu_utilization_target.setter
def cpu_utilization_target(self, value: Optional[pulumi.Input[float]]):
pulumi.set(self, "cpu_utilization_target", value)
@property
@pulumi.getter(name="decommissionTimeout")
def decommission_timeout(self) -> Optional[pulumi.Input[int]]:
"""
Timeout to gracefully decommission nodes during downscaling. In seconds.
"""
return pulumi.get(self, "decommission_timeout")
@decommission_timeout.setter
def decommission_timeout(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "decommission_timeout", value)
@property
@pulumi.getter(name="measurementDuration")
def measurement_duration(self) -> Optional[pulumi.Input[int]]:
"""
Time in seconds allotted for averaging metrics.
"""
return pulumi.get(self, "measurement_duration")
@measurement_duration.setter
def measurement_duration(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "measurement_duration", value)
@property
@pulumi.getter
def preemptible(self) -> Optional[pulumi.Input[bool]]:
"""
Bool flag -- whether to use preemptible compute instances. Preemptible instances are stopped at least once every 24 hours, and can be stopped at any time if their resources are needed by Compute. For more information, see [Preemptible Virtual Machines](https://cloud.yandex.com/docs/compute/concepts/preemptible-vm).
"""
return pulumi.get(self, "preemptible")
@preemptible.setter
def preemptible(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "preemptible", value)
@property
@pulumi.getter(name="stabilizationDuration")
def stabilization_duration(self) -> Optional[pulumi.Input[int]]:
"""
Minimum amount of time in seconds allotted for monitoring before Instance Groups can reduce the number of instances in the group. During this time, the group size doesn't decrease, even if the new metric values indicate that it should.
"""
return pulumi.get(self, "stabilization_duration")
@stabilization_duration.setter
def stabilization_duration(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "stabilization_duration", value)
@property
@pulumi.getter(name="warmupDuration")
def warmup_duration(self) -> Optional[pulumi.Input[int]]:
"""
The warmup time of the instance in seconds. During this time, traffic is sent to the instance, but instance metrics are not collected.
"""
return pulumi.get(self, "warmup_duration")
@warmup_duration.setter
def warmup_duration(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "warmup_duration", value)
@pulumi.input_type
class DataprocClusterClusterConfigSubclusterSpecResourcesArgs:
def __init__(__self__, *,
disk_size: pulumi.Input[int],
resource_preset_id: pulumi.Input[str],
disk_type_id: Optional[pulumi.Input[str]] = None):
"""
:param pulumi.Input[int] disk_size: Volume of the storage available to a host, in gigabytes.
:param pulumi.Input[str] resource_preset_id: The ID of the preset for computational resources available to a host. All available presets are listed in the [documentation](https://cloud.yandex.com/docs/data-proc/concepts/instance-types).
:param pulumi.Input[str] disk_type_id: Type of the storage of a host. One of `network-hdd` (default) or `network-ssd`.
"""
pulumi.set(__self__, "disk_size", disk_size)
pulumi.set(__self__, "resource_preset_id", resource_preset_id)
if disk_type_id is not None:
pulumi.set(__self__, "disk_type_id", disk_type_id)
@property
@pulumi.getter(name="diskSize")
def disk_size(self) -> pulumi.Input[int]:
"""
Volume of the storage available to a host, in gigabytes.
"""
return pulumi.get(self, "disk_size")
@disk_size.setter
def disk_size(self, value: pulumi.Input[int]):
pulumi.set(self, "disk_size", value)
@property
@pulumi.getter(name="resourcePresetId")
def resource_preset_id(self) -> pulumi.Input[str]:
"""
The ID of the preset for computational resources available to a host. All available presets are listed in the [documentation](https://cloud.yandex.com/docs/data-proc/concepts/instance-types).
"""
return pulumi.get(self, "resource_preset_id")
@resource_preset_id.setter
def resource_preset_id(self, value: pulumi.Input[str]):
pulumi.set(self, "resource_preset_id", value)
@property
@pulumi.getter(name="diskTypeId")
def disk_type_id(self) -> Optional[pulumi.Input[str]]:
"""
Type of the storage of a host. One of `network-hdd` (default) or `network-ssd`.
"""
return pulumi.get(self, "disk_type_id")
@disk_type_id.setter
def disk_type_id(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "disk_type_id", value)
@pulumi.input_type
class FunctionContentArgs:
def __init__(__self__, *,
zip_filename: pulumi.Input[str]):
pulumi.set(__self__, "zip_filename", zip_filename)
@property
@pulumi.getter(name="zipFilename")
def zip_filename(self) -> pulumi.Input[str]:
return pulumi.get(self, "zip_filename")
@zip_filename.setter
def zip_filename(self, value: pulumi.Input[str]):
pulumi.set(self, "zip_filename", value)
@pulumi.input_type
class FunctionPackageArgs:
def __init__(__self__, *,
bucket_name: pulumi.Input[str],
object_name: pulumi.Input[str],
sha256: Optional[pulumi.Input[str]] = None):
pulumi.set(__self__, "bucket_name", bucket_name)
pulumi.set(__self__, "object_name", object_name)
if sha256 is not None:
pulumi.set(__self__, "sha256", sha256)
@property
@pulumi.getter(name="bucketName")
def bucket_name(self) -> pulumi.Input[str]:
return pulumi.get(self, "bucket_name")
@bucket_name.setter
def bucket_name(self, value: pulumi.Input[str]):
pulumi.set(self, "bucket_name", value)
@property
@pulumi.getter(name="objectName")
def object_name(self) -> pulumi.Input[str]:
return pulumi.get(self, "object_name")
@object_name.setter
def object_name(self, value: pulumi.Input[str]):
pulumi.set(self, "object_name", value)
@property
@pulumi.getter
def sha256(self) -> Optional[pulumi.Input[str]]:
return pulumi.get(self, "sha256")
@sha256.setter
def sha256(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "sha256", value)
@pulumi.input_type
class FunctionScalingPolicyPolicyArgs:
def __init__(__self__, *,
tag: pulumi.Input[str],
zone_instances_limit: Optional[pulumi.Input[int]] = None,
zone_requests_limit: Optional[pulumi.Input[int]] = None):
pulumi.set(__self__, "tag", tag)
if zone_instances_limit is not None:
pulumi.set(__self__, "zone_instances_limit", zone_instances_limit)
if zone_requests_limit is not None:
pulumi.set(__self__, "zone_requests_limit", zone_requests_limit)
@property
@pulumi.getter
def tag(self) -> pulumi.Input[str]:
return pulumi.get(self, "tag")
@tag.setter
def tag(self, value: pulumi.Input[str]):
pulumi.set(self, "tag", value)
@property
@pulumi.getter(name="zoneInstancesLimit")
def zone_instances_limit(self) -> Optional[pulumi.Input[int]]:
return pulumi.get(self, "zone_instances_limit")
@zone_instances_limit.setter
def zone_instances_limit(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "zone_instances_limit", value)
@property
@pulumi.getter(name="zoneRequestsLimit")
def zone_requests_limit(self) -> Optional[pulumi.Input[int]]:
return pulumi.get(self, "zone_requests_limit")
@zone_requests_limit.setter
def zone_requests_limit(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "zone_requests_limit", value)
@pulumi.input_type
class FunctionTriggerDlqArgs:
def __init__(__self__, *,
queue_id: pulumi.Input[str],
service_account_id: pulumi.Input[str]):
pulumi.set(__self__, "queue_id", queue_id)
pulumi.set(__self__, "service_account_id", service_account_id)
@property
@pulumi.getter(name="queueId")
def queue_id(self) -> pulumi.Input[str]:
return pulumi.get(self, "queue_id")
@queue_id.setter
def queue_id(self, value: pulumi.Input[str]):
pulumi.set(self, "queue_id", value)
@property
@pulumi.getter(name="serviceAccountId")
def service_account_id(self) -> pulumi.Input[str]:
return pulumi.get(self, "service_account_id")
@service_account_id.setter
def service_account_id(self, value: pulumi.Input[str]):
pulumi.set(self, "service_account_id", value)
@pulumi.input_type
class FunctionTriggerFunctionArgs:
def __init__(__self__, *,
id: pulumi.Input[str],
retry_attempts: Optional[pulumi.Input[str]] = None,
retry_interval: Optional[pulumi.Input[str]] = None,
service_account_id: Optional[pulumi.Input[str]] = None,
tag: Optional[pulumi.Input[str]] = None):
pulumi.set(__self__, "id", id)
if retry_attempts is not None:
pulumi.set(__self__, "retry_attempts", retry_attempts)
if retry_interval is not None:
pulumi.set(__self__, "retry_interval", retry_interval)
if service_account_id is not None:
pulumi.set(__self__, "service_account_id", service_account_id)
if tag is not None:
pulumi.set(__self__, "tag", tag)
@property
@pulumi.getter
def id(self) -> pulumi.Input[str]:
return pulumi.get(self, "id")
@id.setter
def id(self, value: pulumi.Input[str]):
pulumi.set(self, "id", value)
@property
@pulumi.getter(name="retryAttempts")
def retry_attempts(self) -> Optional[pulumi.Input[str]]:
return pulumi.get(self, "retry_attempts")
@retry_attempts.setter
def retry_attempts(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "retry_attempts", value)
@property
@pulumi.getter(name="retryInterval")
def retry_interval(self) -> Optional[pulumi.Input[str]]:
return pulumi.get(self, "retry_interval")
@retry_interval.setter
def retry_interval(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "retry_interval", value)
@property
@pulumi.getter(name="serviceAccountId")
def service_account_id(self) -> Optional[pulumi.Input[str]]:
return pulumi.get(self, "service_account_id")
@service_account_id.setter
def service_account_id(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "service_account_id", value)
@property
@pulumi.getter
def tag(self) -> Optional[pulumi.Input[str]]:
return pulumi.get(self, "tag")
@tag.setter
def tag(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "tag", value)
@pulumi.input_type
class FunctionTriggerIotArgs:
def __init__(__self__, *,
registry_id: pulumi.Input[str],
device_id: Optional[pulumi.Input[str]] = None,
topic: Optional[pulumi.Input[str]] = None):
pulumi.set(__self__, "registry_id", registry_id)
if device_id is not None:
pulumi.set(__self__, "device_id", device_id)
if topic is not None:
pulumi.set(__self__, "topic", topic)
@property
@pulumi.getter(name="registryId")
def registry_id(self) -> pulumi.Input[str]:
return pulumi.get(self, "registry_id")
@registry_id.setter
def registry_id(self, value: pulumi.Input[str]):
pulumi.set(self, "registry_id", value)
@property
@pulumi.getter(name="deviceId")
def device_id(self) -> Optional[pulumi.Input[str]]:
return pulumi.get(self, "device_id")
@device_id.setter
def device_id(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "device_id", value)
@property
@pulumi.getter
def topic(self) -> Optional[pulumi.Input[str]]:
return pulumi.get(self, "topic")
@topic.setter
def topic(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "topic", value)
@pulumi.input_type
class FunctionTriggerLogGroupArgs:
def __init__(__self__, *,
batch_cutoff: pulumi.Input[str],
log_group_ids: pulumi.Input[Sequence[pulumi.Input[str]]],
batch_size: Optional[pulumi.Input[str]] = None):
pulumi.set(__self__, "batch_cutoff", batch_cutoff)
pulumi.set(__self__, "log_group_ids", log_group_ids)
if batch_size is not None:
pulumi.set(__self__, "batch_size", batch_size)
@property
@pulumi.getter(name="batchCutoff")
def batch_cutoff(self) -> pulumi.Input[str]:
return pulumi.get(self, "batch_cutoff")
@batch_cutoff.setter
def batch_cutoff(self, value: pulumi.Input[str]):
pulumi.set(self, "batch_cutoff", value)
@property
@pulumi.getter(name="logGroupIds")
def log_group_ids(self) -> pulumi.Input[Sequence[pulumi.Input[str]]]:
return pulumi.get(self, "log_group_ids")
@log_group_ids.setter
def log_group_ids(self, value: pulumi.Input[Sequence[pulumi.Input[str]]]):
pulumi.set(self, "log_group_ids", value)
@property
@pulumi.getter(name="batchSize")
def batch_size(self) -> Optional[pulumi.Input[str]]:
return pulumi.get(self, "batch_size")
@batch_size.setter
def batch_size(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "batch_size", value)
@pulumi.input_type
class FunctionTriggerLoggingArgs:
def __init__(__self__, *,
batch_cutoff: pulumi.Input[str],
group_id: pulumi.Input[str],
levels: pulumi.Input[Sequence[pulumi.Input[str]]],
resource_ids: pulumi.Input[Sequence[pulumi.Input[str]]],
resource_types: pulumi.Input[Sequence[pulumi.Input[str]]],
batch_size: Optional[pulumi.Input[str]] = None):
pulumi.set(__self__, "batch_cutoff", batch_cutoff)
pulumi.set(__self__, "group_id", group_id)
pulumi.set(__self__, "levels", levels)
pulumi.set(__self__, "resource_ids", resource_ids)
pulumi.set(__self__, "resource_types", resource_types)
if batch_size is not None:
pulumi.set(__self__, "batch_size", batch_size)
@property
@pulumi.getter(name="batchCutoff")
def batch_cutoff(self) -> pulumi.Input[str]:
return pulumi.get(self, "batch_cutoff")
@batch_cutoff.setter
def batch_cutoff(self, value: pulumi.Input[str]):
pulumi.set(self, "batch_cutoff", value)
@property
@pulumi.getter(name="groupId")
def group_id(self) -> pulumi.Input[str]:
return pulumi.get(self, "group_id")
@group_id.setter
def group_id(self, value: pulumi.Input[str]):
pulumi.set(self, "group_id", value)
@property
@pulumi.getter
def levels(self) -> pulumi.Input[Sequence[pulumi.Input[str]]]:
return pulumi.get(self, "levels")
@levels.setter
def levels(self, value: pulumi.Input[Sequence[pulumi.Input[str]]]):
pulumi.set(self, "levels", value)
@property
@pulumi.getter(name="resourceIds")
def resource_ids(self) -> pulumi.Input[Sequence[pulumi.Input[str]]]:
return pulumi.get(self, "resource_ids")
@resource_ids.setter
def resource_ids(self, value: pulumi.Input[Sequence[pulumi.Input[str]]]):
pulumi.set(self, "resource_ids", value)
@property
@pulumi.getter(name="resourceTypes")
def resource_types(self) -> pulumi.Input[Sequence[pulumi.Input[str]]]:
return pulumi.get(self, "resource_types")
@resource_types.setter
def resource_types(self, value: pulumi.Input[Sequence[pulumi.Input[str]]]):
pulumi.set(self, "resource_types", value)
@property
@pulumi.getter(name="batchSize")
def batch_size(self) -> Optional[pulumi.Input[str]]:
return pulumi.get(self, "batch_size")
@batch_size.setter
def batch_size(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "batch_size", value)
@pulumi.input_type
class FunctionTriggerMessageQueueArgs:
def __init__(__self__, *,
batch_cutoff: pulumi.Input[str],
queue_id: pulumi.Input[str],
service_account_id: pulumi.Input[str],
batch_size: Optional[pulumi.Input[str]] = None,
visibility_timeout: Optional[pulumi.Input[str]] = None):
pulumi.set(__self__, "batch_cutoff", batch_cutoff)
pulumi.set(__self__, "queue_id", queue_id)
pulumi.set(__self__, "service_account_id", service_account_id)
if batch_size is not None:
pulumi.set(__self__, "batch_size", batch_size)
if visibility_timeout is not None:
pulumi.set(__self__, "visibility_timeout", visibility_timeout)
@property
@pulumi.getter(name="batchCutoff")
def batch_cutoff(self) -> pulumi.Input[str]:
return pulumi.get(self, "batch_cutoff")
@batch_cutoff.setter
def batch_cutoff(self, value: pulumi.Input[str]):
pulumi.set(self, "batch_cutoff", value)
@property
@pulumi.getter(name="queueId")
def queue_id(self) -> pulumi.Input[str]:
return pulumi.get(self, "queue_id")
@queue_id.setter
def queue_id(self, value: pulumi.Input[str]):
pulumi.set(self, "queue_id", value)
@property
@pulumi.getter(name="serviceAccountId")
def service_account_id(self) -> pulumi.Input[str]:
return pulumi.get(self, "service_account_id")
@service_account_id.setter
def service_account_id(self, value: pulumi.Input[str]):
pulumi.set(self, "service_account_id", value)
@property
@pulumi.getter(name="batchSize")
def batch_size(self) -> Optional[pulumi.Input[str]]:
return pulumi.get(self, "batch_size")
@batch_size.setter
def batch_size(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "batch_size", value)
@property
@pulumi.getter(name="visibilityTimeout")
def visibility_timeout(self) -> Optional[pulumi.Input[str]]:
return pulumi.get(self, "visibility_timeout")
@visibility_timeout.setter
def visibility_timeout(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "visibility_timeout", value)
@pulumi.input_type
class FunctionTriggerObjectStorageArgs:
def __init__(__self__, *,
bucket_id: pulumi.Input[str],
create: Optional[pulumi.Input[bool]] = None,
delete: Optional[pulumi.Input[bool]] = None,
prefix: Optional[pulumi.Input[str]] = None,
suffix: Optional[pulumi.Input[str]] = None,
update: Optional[pulumi.Input[bool]] = None):
pulumi.set(__self__, "bucket_id", bucket_id)
if create is not None:
pulumi.set(__self__, "create", create)
if delete is not None:
pulumi.set(__self__, "delete", delete)
if prefix is not None:
pulumi.set(__self__, "prefix", prefix)
if suffix is not None:
pulumi.set(__self__, "suffix", suffix)
if update is not None:
pulumi.set(__self__, "update", update)
@property
@pulumi.getter(name="bucketId")
def bucket_id(self) -> pulumi.Input[str]:
return pulumi.get(self, "bucket_id")
@bucket_id.setter
def bucket_id(self, value: pulumi.Input[str]):
pulumi.set(self, "bucket_id", value)
@property
@pulumi.getter
def create(self) -> Optional[pulumi.Input[bool]]:
return pulumi.get(self, "create")
@create.setter
def create(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "create", value)
@property
@pulumi.getter
def delete(self) -> Optional[pulumi.Input[bool]]:
return pulumi.get(self, "delete")
@delete.setter
def delete(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "delete", value)
@property
@pulumi.getter
def prefix(self) -> Optional[pulumi.Input[str]]:
return pulumi.get(self, "prefix")
@prefix.setter
def prefix(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "prefix", value)
@property
@pulumi.getter
def suffix(self) -> Optional[pulumi.Input[str]]:
return pulumi.get(self, "suffix")
@suffix.setter
def suffix(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "suffix", value)
@property
@pulumi.getter
def update(self) -> Optional[pulumi.Input[bool]]:
return pulumi.get(self, "update")
@update.setter
def update(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "update", value)
@pulumi.input_type
class FunctionTriggerTimerArgs:
def __init__(__self__, *,
cron_expression: pulumi.Input[str]):
pulumi.set(__self__, "cron_expression", cron_expression)
@property
@pulumi.getter(name="cronExpression")
def cron_expression(self) -> pulumi.Input[str]:
return pulumi.get(self, "cron_expression")
@cron_expression.setter
def cron_expression(self, value: pulumi.Input[str]):
pulumi.set(self, "cron_expression", value)
@pulumi.input_type
class KubernetesClusterKmsProviderArgs:
def __init__(__self__, *,
key_id: Optional[pulumi.Input[str]] = None):
"""
:param pulumi.Input[str] key_id: KMS key ID.
"""
if key_id is not None:
pulumi.set(__self__, "key_id", key_id)
@property
@pulumi.getter(name="keyId")
def key_id(self) -> Optional[pulumi.Input[str]]:
"""
KMS key ID.
"""
return pulumi.get(self, "key_id")
@key_id.setter
def key_id(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "key_id", value)
@pulumi.input_type
class KubernetesClusterMasterArgs:
def __init__(__self__, *,
cluster_ca_certificate: Optional[pulumi.Input[str]] = None,
external_v4_address: Optional[pulumi.Input[str]] = None,
external_v4_endpoint: Optional[pulumi.Input[str]] = None,
internal_v4_address: Optional[pulumi.Input[str]] = None,
internal_v4_endpoint: Optional[pulumi.Input[str]] = None,
maintenance_policy: Optional[pulumi.Input['KubernetesClusterMasterMaintenancePolicyArgs']] = None,
public_ip: Optional[pulumi.Input[bool]] = None,
regional: Optional[pulumi.Input['KubernetesClusterMasterRegionalArgs']] = None,
security_group_ids: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None,
version: Optional[pulumi.Input[str]] = None,
version_info: Optional[pulumi.Input['KubernetesClusterMasterVersionInfoArgs']] = None,
zonal: Optional[pulumi.Input['KubernetesClusterMasterZonalArgs']] = None):
"""
:param pulumi.Input[str] cluster_ca_certificate: (Computed) PEM-encoded public certificate that is the root of trust for the Kubernetes cluster.
:param pulumi.Input[str] external_v4_address: (Computed) An IPv4 external network address that is assigned to the master.
:param pulumi.Input[str] external_v4_endpoint: (Computed) External endpoint that can be used to access Kubernetes cluster API from the internet (outside of the cloud).
:param pulumi.Input[str] internal_v4_address: (Computed) An IPv4 internal network address that is assigned to the master.
:param pulumi.Input[str] internal_v4_endpoint: (Computed) Internal endpoint that can be used to connect to the master from cloud networks.
:param pulumi.Input['KubernetesClusterMasterMaintenancePolicyArgs'] maintenance_policy: (Optional) (Computed) Maintenance policy for Kubernetes master.
If policy is omitted, automatic revision upgrades of the kubernetes master are enabled and could happen at any time.
Revision upgrades are performed only within the same minor version, e.g. 1.13.
Minor version upgrades (e.g. 1.13->1.14) should be performed manually. The structure is documented below.
:param pulumi.Input[bool] public_ip: (Optional) (Computed) Boolean flag. When `true`, Kubernetes master will have visible ipv4 address.
:param pulumi.Input['KubernetesClusterMasterRegionalArgs'] regional: (Optional) Initialize parameters for Regional Master (highly available master). The structure is documented below.
:param pulumi.Input[Sequence[pulumi.Input[str]]] security_group_ids: (Optional) List of security group IDs to which the Kubernetes cluster belongs.
:param pulumi.Input[str] version: (Optional) (Computed) Version of Kubernetes that will be used for master.
:param pulumi.Input['KubernetesClusterMasterVersionInfoArgs'] version_info: (Computed) Information about cluster version. The structure is documented below.
:param pulumi.Input['KubernetesClusterMasterZonalArgs'] zonal: (Optional) Initialize parameters for Zonal Master (single node master). The structure is documented below.
"""
if cluster_ca_certificate is not None:
pulumi.set(__self__, "cluster_ca_certificate", cluster_ca_certificate)
if external_v4_address is not None:
pulumi.set(__self__, "external_v4_address", external_v4_address)
if external_v4_endpoint is not None:
pulumi.set(__self__, "external_v4_endpoint", external_v4_endpoint)
if internal_v4_address is not None:
pulumi.set(__self__, "internal_v4_address", internal_v4_address)
if internal_v4_endpoint is not None:
pulumi.set(__self__, "internal_v4_endpoint", internal_v4_endpoint)
if maintenance_policy is not None:
pulumi.set(__self__, "maintenance_policy", maintenance_policy)
if public_ip is not None:
pulumi.set(__self__, "public_ip", public_ip)
if regional is not None:
pulumi.set(__self__, "regional", regional)
if security_group_ids is not None:
pulumi.set(__self__, "security_group_ids", security_group_ids)
if version is not None:
pulumi.set(__self__, "version", version)
if version_info is not None:
pulumi.set(__self__, "version_info", version_info)
if zonal is not None:
pulumi.set(__self__, "zonal", zonal)
@property
@pulumi.getter(name="clusterCaCertificate")
def cluster_ca_certificate(self) -> Optional[pulumi.Input[str]]:
"""
(Computed) PEM-encoded public certificate that is the root of trust for the Kubernetes cluster.
"""
return pulumi.get(self, "cluster_ca_certificate")
@cluster_ca_certificate.setter
def cluster_ca_certificate(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "cluster_ca_certificate", value)
@property
@pulumi.getter(name="externalV4Address")
def external_v4_address(self) -> Optional[pulumi.Input[str]]:
"""
(Computed) An IPv4 external network address that is assigned to the master.
"""
return pulumi.get(self, "external_v4_address")
@external_v4_address.setter
def external_v4_address(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "external_v4_address", value)
@property
@pulumi.getter(name="externalV4Endpoint")
def external_v4_endpoint(self) -> Optional[pulumi.Input[str]]:
"""
(Computed) External endpoint that can be used to access Kubernetes cluster API from the internet (outside of the cloud).
"""
return pulumi.get(self, "external_v4_endpoint")
@external_v4_endpoint.setter
def external_v4_endpoint(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "external_v4_endpoint", value)
@property
@pulumi.getter(name="internalV4Address")
def internal_v4_address(self) -> Optional[pulumi.Input[str]]:
"""
(Computed) An IPv4 internal network address that is assigned to the master.
"""
return pulumi.get(self, "internal_v4_address")
@internal_v4_address.setter
def internal_v4_address(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "internal_v4_address", value)
@property
@pulumi.getter(name="internalV4Endpoint")
def internal_v4_endpoint(self) -> Optional[pulumi.Input[str]]:
"""
(Computed) Internal endpoint that can be used to connect to the master from cloud networks.
"""
return pulumi.get(self, "internal_v4_endpoint")
@internal_v4_endpoint.setter
def internal_v4_endpoint(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "internal_v4_endpoint", value)
@property
@pulumi.getter(name="maintenancePolicy")
def maintenance_policy(self) -> Optional[pulumi.Input['KubernetesClusterMasterMaintenancePolicyArgs']]:
"""
(Optional) (Computed) Maintenance policy for Kubernetes master.
If policy is omitted, automatic revision upgrades of the kubernetes master are enabled and could happen at any time.
Revision upgrades are performed only within the same minor version, e.g. 1.13.
Minor version upgrades (e.g. 1.13->1.14) should be performed manually. The structure is documented below.
"""
return pulumi.get(self, "maintenance_policy")
@maintenance_policy.setter
def maintenance_policy(self, value: Optional[pulumi.Input['KubernetesClusterMasterMaintenancePolicyArgs']]):
pulumi.set(self, "maintenance_policy", value)
@property
@pulumi.getter(name="publicIp")
def public_ip(self) -> Optional[pulumi.Input[bool]]:
"""
(Optional) (Computed) Boolean flag. When `true`, Kubernetes master will have visible ipv4 address.
"""
return pulumi.get(self, "public_ip")
@public_ip.setter
def public_ip(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "public_ip", value)
@property
@pulumi.getter
def regional(self) -> Optional[pulumi.Input['KubernetesClusterMasterRegionalArgs']]:
"""
(Optional) Initialize parameters for Regional Master (highly available master). The structure is documented below.
"""
return pulumi.get(self, "regional")
@regional.setter
def regional(self, value: Optional[pulumi.Input['KubernetesClusterMasterRegionalArgs']]):
pulumi.set(self, "regional", value)
@property
@pulumi.getter(name="securityGroupIds")
def security_group_ids(self) -> Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]:
"""
(Optional) List of security group IDs to which the Kubernetes cluster belongs.
"""
return pulumi.get(self, "security_group_ids")
@security_group_ids.setter
def security_group_ids(self, value: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]):
pulumi.set(self, "security_group_ids", value)
@property
@pulumi.getter
def version(self) -> Optional[pulumi.Input[str]]:
"""
(Optional) (Computed) Version of Kubernetes that will be used for master.
"""
return pulumi.get(self, "version")
@version.setter
def version(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "version", value)
@property
@pulumi.getter(name="versionInfo")
def version_info(self) -> Optional[pulumi.Input['KubernetesClusterMasterVersionInfoArgs']]:
"""
(Computed) Information about cluster version. The structure is documented below.
"""
return pulumi.get(self, "version_info")
@version_info.setter
def version_info(self, value: Optional[pulumi.Input['KubernetesClusterMasterVersionInfoArgs']]):
pulumi.set(self, "version_info", value)
@property
@pulumi.getter
def zonal(self) -> Optional[pulumi.Input['KubernetesClusterMasterZonalArgs']]:
"""
(Optional) Initialize parameters for Zonal Master (single node master). The structure is documented below.
"""
return pulumi.get(self, "zonal")
@zonal.setter
def zonal(self, value: Optional[pulumi.Input['KubernetesClusterMasterZonalArgs']]):
pulumi.set(self, "zonal", value)
@pulumi.input_type
class KubernetesClusterMasterMaintenancePolicyArgs:
def __init__(__self__, *,
auto_upgrade: pulumi.Input[bool],
maintenance_windows: Optional[pulumi.Input[Sequence[pulumi.Input['KubernetesClusterMasterMaintenancePolicyMaintenanceWindowArgs']]]] = None):
"""
:param pulumi.Input[bool] auto_upgrade: (Required) Boolean flag that specifies if master can be upgraded automatically. When omitted, default value is TRUE.
:param pulumi.Input[Sequence[pulumi.Input['KubernetesClusterMasterMaintenancePolicyMaintenanceWindowArgs']]] maintenance_windows: (Optional) (Computed) This structure specifies maintenance window, when update for master is allowed. When omitted, it defaults to any time.
To specify time of day interval, for all days, one element should be provided, with two fields set, `start_time` and `duration`.
Please see `zonal_cluster_resource_name` config example.
"""
pulumi.set(__self__, "auto_upgrade", auto_upgrade)
if maintenance_windows is not None:
pulumi.set(__self__, "maintenance_windows", maintenance_windows)
@property
@pulumi.getter(name="autoUpgrade")
def auto_upgrade(self) -> pulumi.Input[bool]:
"""
(Required) Boolean flag that specifies if master can be upgraded automatically. When omitted, default value is TRUE.
"""
return pulumi.get(self, "auto_upgrade")
@auto_upgrade.setter
def auto_upgrade(self, value: pulumi.Input[bool]):
pulumi.set(self, "auto_upgrade", value)
@property
@pulumi.getter(name="maintenanceWindows")
def maintenance_windows(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['KubernetesClusterMasterMaintenancePolicyMaintenanceWindowArgs']]]]:
"""
(Optional) (Computed) This structure specifies maintenance window, when update for master is allowed. When omitted, it defaults to any time.
To specify time of day interval, for all days, one element should be provided, with two fields set, `start_time` and `duration`.
Please see `zonal_cluster_resource_name` config example.
"""
return pulumi.get(self, "maintenance_windows")
@maintenance_windows.setter
def maintenance_windows(self, value: Optional[pulumi.Input[Sequence[pulumi.Input['KubernetesClusterMasterMaintenancePolicyMaintenanceWindowArgs']]]]):
pulumi.set(self, "maintenance_windows", value)
@pulumi.input_type
class KubernetesClusterMasterMaintenancePolicyMaintenanceWindowArgs:
def __init__(__self__, *,
duration: pulumi.Input[str],
start_time: pulumi.Input[str],
day: Optional[pulumi.Input[str]] = None):
pulumi.set(__self__, "duration", duration)
pulumi.set(__self__, "start_time", start_time)
if day is not None:
pulumi.set(__self__, "day", day)
@property
@pulumi.getter
def duration(self) -> pulumi.Input[str]:
return pulumi.get(self, "duration")
@duration.setter
def duration(self, value: pulumi.Input[str]):
pulumi.set(self, "duration", value)
@property
@pulumi.getter(name="startTime")
def start_time(self) -> pulumi.Input[str]:
return pulumi.get(self, "start_time")
@start_time.setter
def start_time(self, value: pulumi.Input[str]):
pulumi.set(self, "start_time", value)
@property
@pulumi.getter
def day(self) -> Optional[pulumi.Input[str]]:
return pulumi.get(self, "day")
@day.setter
def day(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "day", value)
@pulumi.input_type
class KubernetesClusterMasterRegionalArgs:
def __init__(__self__, *,
region: pulumi.Input[str],
locations: Optional[pulumi.Input[Sequence[pulumi.Input['KubernetesClusterMasterRegionalLocationArgs']]]] = None):
"""
:param pulumi.Input[str] region: (Required) Name of availability region (e.g. "ru-central1"), where master instances will be allocated.
:param pulumi.Input[Sequence[pulumi.Input['KubernetesClusterMasterRegionalLocationArgs']]] locations: Array of locations, where master instances will be allocated. The structure is documented below.
"""
pulumi.set(__self__, "region", region)
if locations is not None:
pulumi.set(__self__, "locations", locations)
@property
@pulumi.getter
def region(self) -> pulumi.Input[str]:
"""
(Required) Name of availability region (e.g. "ru-central1"), where master instances will be allocated.
"""
return pulumi.get(self, "region")
@region.setter
def region(self, value: pulumi.Input[str]):
pulumi.set(self, "region", value)
@property
@pulumi.getter
def locations(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['KubernetesClusterMasterRegionalLocationArgs']]]]:
"""
Array of locations, where master instances will be allocated. The structure is documented below.
"""
return pulumi.get(self, "locations")
@locations.setter
def locations(self, value: Optional[pulumi.Input[Sequence[pulumi.Input['KubernetesClusterMasterRegionalLocationArgs']]]]):
pulumi.set(self, "locations", value)
@pulumi.input_type
class KubernetesClusterMasterRegionalLocationArgs:
def __init__(__self__, *,
subnet_id: Optional[pulumi.Input[str]] = None,
zone: Optional[pulumi.Input[str]] = None):
"""
:param pulumi.Input[str] subnet_id: (Optional) ID of the subnet.
:param pulumi.Input[str] zone: (Optional) ID of the availability zone.
"""
if subnet_id is not None:
pulumi.set(__self__, "subnet_id", subnet_id)
if zone is not None:
pulumi.set(__self__, "zone", zone)
@property
@pulumi.getter(name="subnetId")
def subnet_id(self) -> Optional[pulumi.Input[str]]:
"""
(Optional) ID of the subnet.
"""
return pulumi.get(self, "subnet_id")
@subnet_id.setter
def subnet_id(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "subnet_id", value)
@property
@pulumi.getter
def zone(self) -> Optional[pulumi.Input[str]]:
"""
(Optional) ID of the availability zone.
"""
return pulumi.get(self, "zone")
@zone.setter
def zone(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "zone", value)
@pulumi.input_type
class KubernetesClusterMasterVersionInfoArgs:
def __init__(__self__, *,
current_version: Optional[pulumi.Input[str]] = None,
new_revision_available: Optional[pulumi.Input[bool]] = None,
new_revision_summary: Optional[pulumi.Input[str]] = None,
version_deprecated: Optional[pulumi.Input[bool]] = None):
"""
:param pulumi.Input[str] current_version: Current Kubernetes version, major.minor (e.g. 1.15).
:param pulumi.Input[bool] new_revision_available: Boolean flag.
Newer revisions may include Kubernetes patches (e.g 1.15.1 > 1.15.2) as well
as some internal component updates - new features or bug fixes in yandex-specific
components either on the master or nodes.
:param pulumi.Input[str] new_revision_summary: Human readable description of the changes to be applied
when updating to the latest revision. Empty if new_revision_available is false.
:param pulumi.Input[bool] version_deprecated: Boolean flag. The current version is on the deprecation schedule,
component (master or node group) should be upgraded.
"""
if current_version is not None:
pulumi.set(__self__, "current_version", current_version)
if new_revision_available is not None:
pulumi.set(__self__, "new_revision_available", new_revision_available)
if new_revision_summary is not None:
pulumi.set(__self__, "new_revision_summary", new_revision_summary)
if version_deprecated is not None:
pulumi.set(__self__, "version_deprecated", version_deprecated)
@property
@pulumi.getter(name="currentVersion")
def current_version(self) -> Optional[pulumi.Input[str]]:
"""
Current Kubernetes version, major.minor (e.g. 1.15).
"""
return pulumi.get(self, "current_version")
@current_version.setter
def current_version(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "current_version", value)
@property
@pulumi.getter(name="newRevisionAvailable")
def new_revision_available(self) -> Optional[pulumi.Input[bool]]:
"""
Boolean flag.
Newer revisions may include Kubernetes patches (e.g 1.15.1 > 1.15.2) as well
as some internal component updates - new features or bug fixes in yandex-specific
components either on the master or nodes.
"""
return pulumi.get(self, "new_revision_available")
@new_revision_available.setter
def new_revision_available(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "new_revision_available", value)
@property
@pulumi.getter(name="newRevisionSummary")
def new_revision_summary(self) -> Optional[pulumi.Input[str]]:
"""
Human readable description of the changes to be applied
when updating to the latest revision. Empty if new_revision_available is false.
"""
return pulumi.get(self, "new_revision_summary")
@new_revision_summary.setter
def new_revision_summary(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "new_revision_summary", value)
@property
@pulumi.getter(name="versionDeprecated")
def version_deprecated(self) -> Optional[pulumi.Input[bool]]:
"""
Boolean flag. The current version is on the deprecation schedule,
component (master or node group) should be upgraded.
"""
return pulumi.get(self, "version_deprecated")
@version_deprecated.setter
def version_deprecated(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "version_deprecated", value)
@pulumi.input_type
class KubernetesClusterMasterZonalArgs:
def __init__(__self__, *,
subnet_id: Optional[pulumi.Input[str]] = None,
zone: Optional[pulumi.Input[str]] = None):
"""
:param pulumi.Input[str] subnet_id: (Optional) ID of the subnet.
:param pulumi.Input[str] zone: (Optional) ID of the availability zone.
"""
if subnet_id is not None:
pulumi.set(__self__, "subnet_id", subnet_id)
if zone is not None:
pulumi.set(__self__, "zone", zone)
@property
@pulumi.getter(name="subnetId")
def subnet_id(self) -> Optional[pulumi.Input[str]]:
"""
(Optional) ID of the subnet.
"""
return pulumi.get(self, "subnet_id")
@subnet_id.setter
def subnet_id(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "subnet_id", value)
@property
@pulumi.getter
def zone(self) -> Optional[pulumi.Input[str]]:
"""
(Optional) ID of the availability zone.
"""
return pulumi.get(self, "zone")
@zone.setter
def zone(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "zone", value)
@pulumi.input_type
class KubernetesClusterNetworkImplementationArgs:
def __init__(__self__, *,
cilium: Optional[pulumi.Input['KubernetesClusterNetworkImplementationCiliumArgs']] = None):
"""
:param pulumi.Input['KubernetesClusterNetworkImplementationCiliumArgs'] cilium: (Optional) Cilium network implementation configuration. No options exist.
"""
if cilium is not None:
pulumi.set(__self__, "cilium", cilium)
@property
@pulumi.getter
def cilium(self) -> Optional[pulumi.Input['KubernetesClusterNetworkImplementationCiliumArgs']]:
"""
(Optional) Cilium network implementation configuration. No options exist.
"""
return pulumi.get(self, "cilium")
@cilium.setter
def cilium(self, value: Optional[pulumi.Input['KubernetesClusterNetworkImplementationCiliumArgs']]):
pulumi.set(self, "cilium", value)
@pulumi.input_type
class KubernetesClusterNetworkImplementationCiliumArgs:
def __init__(__self__):
pass
@pulumi.input_type
class KubernetesNodeGroupAllocationPolicyArgs:
def __init__(__self__, *,
locations: Optional[pulumi.Input[Sequence[pulumi.Input['KubernetesNodeGroupAllocationPolicyLocationArgs']]]] = None):
"""
:param pulumi.Input[Sequence[pulumi.Input['KubernetesNodeGroupAllocationPolicyLocationArgs']]] locations: Repeated field, that specify subnets (zones), that will be used by node group compute instances. The structure is documented below.
"""
if locations is not None:
pulumi.set(__self__, "locations", locations)
@property
@pulumi.getter
def locations(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['KubernetesNodeGroupAllocationPolicyLocationArgs']]]]:
"""
Repeated field, that specify subnets (zones), that will be used by node group compute instances. The structure is documented below.
"""
return pulumi.get(self, "locations")
@locations.setter
def locations(self, value: Optional[pulumi.Input[Sequence[pulumi.Input['KubernetesNodeGroupAllocationPolicyLocationArgs']]]]):
pulumi.set(self, "locations", value)
@pulumi.input_type
class KubernetesNodeGroupAllocationPolicyLocationArgs:
def __init__(__self__, *,
subnet_id: Optional[pulumi.Input[str]] = None,
zone: Optional[pulumi.Input[str]] = None):
"""
:param pulumi.Input[str] subnet_id: ID of the subnet, that will be used by one compute instance in node group.
:param pulumi.Input[str] zone: ID of the availability zone where for one compute instance in node group.
"""
if subnet_id is not None:
warnings.warn("""The 'subnet_id' field has been deprecated. Please use 'subnet_ids under network_interface' instead.""", DeprecationWarning)
pulumi.log.warn("""subnet_id is deprecated: The 'subnet_id' field has been deprecated. Please use 'subnet_ids under network_interface' instead.""")
if subnet_id is not None:
pulumi.set(__self__, "subnet_id", subnet_id)
if zone is not None:
pulumi.set(__self__, "zone", zone)
@property
@pulumi.getter(name="subnetId")
def subnet_id(self) -> Optional[pulumi.Input[str]]:
"""
ID of the subnet, that will be used by one compute instance in node group.
"""
return pulumi.get(self, "subnet_id")
@subnet_id.setter
def subnet_id(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "subnet_id", value)
@property
@pulumi.getter
def zone(self) -> Optional[pulumi.Input[str]]:
"""
ID of the availability zone where for one compute instance in node group.
"""
return pulumi.get(self, "zone")
@zone.setter
def zone(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "zone", value)
@pulumi.input_type
class KubernetesNodeGroupDeployPolicyArgs:
def __init__(__self__, *,
max_expansion: pulumi.Input[int],
max_unavailable: pulumi.Input[int]):
"""
:param pulumi.Input[int] max_expansion: The maximum number of instances that can be temporarily allocated above the group's target size during the update.
:param pulumi.Input[int] max_unavailable: The maximum number of running instances that can be taken offline during update.
"""
pulumi.set(__self__, "max_expansion", max_expansion)
pulumi.set(__self__, "max_unavailable", max_unavailable)
@property
@pulumi.getter(name="maxExpansion")
def max_expansion(self) -> pulumi.Input[int]:
"""
The maximum number of instances that can be temporarily allocated above the group's target size during the update.
"""
return pulumi.get(self, "max_expansion")
@max_expansion.setter
def max_expansion(self, value: pulumi.Input[int]):
pulumi.set(self, "max_expansion", value)
@property
@pulumi.getter(name="maxUnavailable")
def max_unavailable(self) -> pulumi.Input[int]:
"""
The maximum number of running instances that can be taken offline during update.
"""
return pulumi.get(self, "max_unavailable")
@max_unavailable.setter
def max_unavailable(self, value: pulumi.Input[int]):
pulumi.set(self, "max_unavailable", value)
@pulumi.input_type
class KubernetesNodeGroupInstanceTemplateArgs:
def __init__(__self__, *,
boot_disk: Optional[pulumi.Input['KubernetesNodeGroupInstanceTemplateBootDiskArgs']] = None,
metadata: Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]] = None,
nat: Optional[pulumi.Input[bool]] = None,
network_acceleration_type: Optional[pulumi.Input[str]] = None,
network_interfaces: Optional[pulumi.Input[Sequence[pulumi.Input['KubernetesNodeGroupInstanceTemplateNetworkInterfaceArgs']]]] = None,
placement_policy: Optional[pulumi.Input['KubernetesNodeGroupInstanceTemplatePlacementPolicyArgs']] = None,
platform_id: Optional[pulumi.Input[str]] = None,
resources: Optional[pulumi.Input['KubernetesNodeGroupInstanceTemplateResourcesArgs']] = None,
scheduling_policy: Optional[pulumi.Input['KubernetesNodeGroupInstanceTemplateSchedulingPolicyArgs']] = None):
"""
:param pulumi.Input['KubernetesNodeGroupInstanceTemplateBootDiskArgs'] boot_disk: The specifications for boot disks that will be attached to the instance. The structure is documented below.
:param pulumi.Input[Mapping[str, pulumi.Input[str]]] metadata: The set of metadata `key:value` pairs assigned to this instance template. This includes custom metadata and predefined keys.
* `resources.0.memory` - The memory size allocated to the instance.
* `resources.0.cores` - Number of CPU cores allocated to the instance.
* `resources.0.core_fraction` - Baseline core performance as a percent.
* `resources.0.gpus` - Number of GPU cores allocated to the instance.
:param pulumi.Input[bool] nat: A public address that can be used to access the internet over NAT.
:param pulumi.Input[str] network_acceleration_type: Type of network acceleration. Values: `standard`, `software_accelerated`.
:param pulumi.Input[Sequence[pulumi.Input['KubernetesNodeGroupInstanceTemplateNetworkInterfaceArgs']]] network_interfaces: An array with the network interfaces that will be attached to the instance. The structure is documented below.
:param pulumi.Input['KubernetesNodeGroupInstanceTemplatePlacementPolicyArgs'] placement_policy: The placement policy configuration. The structure is documented below.
:param pulumi.Input[str] platform_id: The ID of the hardware platform configuration for the node group compute instances.
:param pulumi.Input['KubernetesNodeGroupInstanceTemplateSchedulingPolicyArgs'] scheduling_policy: The scheduling policy for the instances in node group. The structure is documented below.
"""
if boot_disk is not None:
pulumi.set(__self__, "boot_disk", boot_disk)
if metadata is not None:
pulumi.set(__self__, "metadata", metadata)
if nat is not None:
warnings.warn("""The 'nat' field has been deprecated. Please use 'nat under network_interface' instead.""", DeprecationWarning)
pulumi.log.warn("""nat is deprecated: The 'nat' field has been deprecated. Please use 'nat under network_interface' instead.""")
if nat is not None:
pulumi.set(__self__, "nat", nat)
if network_acceleration_type is not None:
pulumi.set(__self__, "network_acceleration_type", network_acceleration_type)
if network_interfaces is not None:
pulumi.set(__self__, "network_interfaces", network_interfaces)
if placement_policy is not None:
pulumi.set(__self__, "placement_policy", placement_policy)
if platform_id is not None:
pulumi.set(__self__, "platform_id", platform_id)
if resources is not None:
pulumi.set(__self__, "resources", resources)
if scheduling_policy is not None:
pulumi.set(__self__, "scheduling_policy", scheduling_policy)
@property
@pulumi.getter(name="bootDisk")
def boot_disk(self) -> Optional[pulumi.Input['KubernetesNodeGroupInstanceTemplateBootDiskArgs']]:
"""
The specifications for boot disks that will be attached to the instance. The structure is documented below.
"""
return pulumi.get(self, "boot_disk")
@boot_disk.setter
def boot_disk(self, value: Optional[pulumi.Input['KubernetesNodeGroupInstanceTemplateBootDiskArgs']]):
pulumi.set(self, "boot_disk", value)
@property
@pulumi.getter
def metadata(self) -> Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]]:
"""
The set of metadata `key:value` pairs assigned to this instance template. This includes custom metadata and predefined keys.
* `resources.0.memory` - The memory size allocated to the instance.
* `resources.0.cores` - Number of CPU cores allocated to the instance.
* `resources.0.core_fraction` - Baseline core performance as a percent.
* `resources.0.gpus` - Number of GPU cores allocated to the instance.
"""
return pulumi.get(self, "metadata")
@metadata.setter
def metadata(self, value: Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]]):
pulumi.set(self, "metadata", value)
@property
@pulumi.getter
def nat(self) -> Optional[pulumi.Input[bool]]:
"""
A public address that can be used to access the internet over NAT.
"""
return pulumi.get(self, "nat")
@nat.setter
def nat(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "nat", value)
@property
@pulumi.getter(name="networkAccelerationType")
def network_acceleration_type(self) -> Optional[pulumi.Input[str]]:
"""
Type of network acceleration. Values: `standard`, `software_accelerated`.
"""
return pulumi.get(self, "network_acceleration_type")
@network_acceleration_type.setter
def network_acceleration_type(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "network_acceleration_type", value)
@property
@pulumi.getter(name="networkInterfaces")
def network_interfaces(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['KubernetesNodeGroupInstanceTemplateNetworkInterfaceArgs']]]]:
"""
An array with the network interfaces that will be attached to the instance. The structure is documented below.
"""
return pulumi.get(self, "network_interfaces")
@network_interfaces.setter
def network_interfaces(self, value: Optional[pulumi.Input[Sequence[pulumi.Input['KubernetesNodeGroupInstanceTemplateNetworkInterfaceArgs']]]]):
pulumi.set(self, "network_interfaces", value)
@property
@pulumi.getter(name="placementPolicy")
def placement_policy(self) -> Optional[pulumi.Input['KubernetesNodeGroupInstanceTemplatePlacementPolicyArgs']]:
"""
The placement policy configuration. The structure is documented below.
"""
return pulumi.get(self, "placement_policy")
@placement_policy.setter
def placement_policy(self, value: Optional[pulumi.Input['KubernetesNodeGroupInstanceTemplatePlacementPolicyArgs']]):
pulumi.set(self, "placement_policy", value)
@property
@pulumi.getter(name="platformId")
def platform_id(self) -> Optional[pulumi.Input[str]]:
"""
The ID of the hardware platform configuration for the node group compute instances.
"""
return pulumi.get(self, "platform_id")
@platform_id.setter
def platform_id(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "platform_id", value)
@property
@pulumi.getter
def resources(self) -> Optional[pulumi.Input['KubernetesNodeGroupInstanceTemplateResourcesArgs']]:
return pulumi.get(self, "resources")
@resources.setter
def resources(self, value: Optional[pulumi.Input['KubernetesNodeGroupInstanceTemplateResourcesArgs']]):
pulumi.set(self, "resources", value)
@property
@pulumi.getter(name="schedulingPolicy")
def scheduling_policy(self) -> Optional[pulumi.Input['KubernetesNodeGroupInstanceTemplateSchedulingPolicyArgs']]:
"""
The scheduling policy for the instances in node group. The structure is documented below.
"""
return pulumi.get(self, "scheduling_policy")
@scheduling_policy.setter
def scheduling_policy(self, value: Optional[pulumi.Input['KubernetesNodeGroupInstanceTemplateSchedulingPolicyArgs']]):
pulumi.set(self, "scheduling_policy", value)
@pulumi.input_type
class KubernetesNodeGroupInstanceTemplateBootDiskArgs:
def __init__(__self__, *,
size: Optional[pulumi.Input[int]] = None,
type: Optional[pulumi.Input[str]] = None):
"""
:param pulumi.Input[int] size: The number of instances in the node group.
:param pulumi.Input[str] type: The disk type.
"""
if size is not None:
pulumi.set(__self__, "size", size)
if type is not None:
pulumi.set(__self__, "type", type)
@property
@pulumi.getter
def size(self) -> Optional[pulumi.Input[int]]:
"""
The number of instances in the node group.
"""
return pulumi.get(self, "size")
@size.setter
def size(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "size", value)
@property
@pulumi.getter
def type(self) -> Optional[pulumi.Input[str]]:
"""
The disk type.
"""
return pulumi.get(self, "type")
@type.setter
def type(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "type", value)
@pulumi.input_type
class KubernetesNodeGroupInstanceTemplateNetworkInterfaceArgs:
def __init__(__self__, *,
subnet_ids: pulumi.Input[Sequence[pulumi.Input[str]]],
ipv4: Optional[pulumi.Input[bool]] = None,
ipv6: Optional[pulumi.Input[bool]] = None,
nat: Optional[pulumi.Input[bool]] = None,
security_group_ids: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None):
"""
:param pulumi.Input[Sequence[pulumi.Input[str]]] subnet_ids: The IDs of the subnets.
:param pulumi.Input[bool] ipv4: Allocate an IPv4 address for the interface. The default value is `true`.
:param pulumi.Input[bool] ipv6: If true, allocate an IPv6 address for the interface. The address will be automatically assigned from the specified subnet.
:param pulumi.Input[bool] nat: A public address that can be used to access the internet over NAT.
:param pulumi.Input[Sequence[pulumi.Input[str]]] security_group_ids: Security group ids for network interface.
"""
pulumi.set(__self__, "subnet_ids", subnet_ids)
if ipv4 is not None:
pulumi.set(__self__, "ipv4", ipv4)
if ipv6 is not None:
pulumi.set(__self__, "ipv6", ipv6)
if nat is not None:
pulumi.set(__self__, "nat", nat)
if security_group_ids is not None:
pulumi.set(__self__, "security_group_ids", security_group_ids)
@property
@pulumi.getter(name="subnetIds")
def subnet_ids(self) -> pulumi.Input[Sequence[pulumi.Input[str]]]:
"""
The IDs of the subnets.
"""
return pulumi.get(self, "subnet_ids")
@subnet_ids.setter
def subnet_ids(self, value: pulumi.Input[Sequence[pulumi.Input[str]]]):
pulumi.set(self, "subnet_ids", value)
@property
@pulumi.getter
def ipv4(self) -> Optional[pulumi.Input[bool]]:
"""
Allocate an IPv4 address for the interface. The default value is `true`.
"""
return pulumi.get(self, "ipv4")
@ipv4.setter
def ipv4(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "ipv4", value)
@property
@pulumi.getter
def ipv6(self) -> Optional[pulumi.Input[bool]]:
"""
If true, allocate an IPv6 address for the interface. The address will be automatically assigned from the specified subnet.
"""
return pulumi.get(self, "ipv6")
@ipv6.setter
def ipv6(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "ipv6", value)
@property
@pulumi.getter
def nat(self) -> Optional[pulumi.Input[bool]]:
"""
A public address that can be used to access the internet over NAT.
"""
return pulumi.get(self, "nat")
@nat.setter
def nat(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "nat", value)
@property
@pulumi.getter(name="securityGroupIds")
def security_group_ids(self) -> Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]:
"""
Security group ids for network interface.
"""
return pulumi.get(self, "security_group_ids")
@security_group_ids.setter
def security_group_ids(self, value: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]):
pulumi.set(self, "security_group_ids", value)
@pulumi.input_type
class KubernetesNodeGroupInstanceTemplatePlacementPolicyArgs:
def __init__(__self__, *,
placement_group_id: pulumi.Input[str]):
"""
:param pulumi.Input[str] placement_group_id: Specifies the id of the Placement Group to assign to the instances.
"""
pulumi.set(__self__, "placement_group_id", placement_group_id)
@property
@pulumi.getter(name="placementGroupId")
def placement_group_id(self) -> pulumi.Input[str]:
"""
Specifies the id of the Placement Group to assign to the instances.
"""
return pulumi.get(self, "placement_group_id")
@placement_group_id.setter
def placement_group_id(self, value: pulumi.Input[str]):
pulumi.set(self, "placement_group_id", value)
@pulumi.input_type
class KubernetesNodeGroupInstanceTemplateResourcesArgs:
def __init__(__self__, *,
core_fraction: Optional[pulumi.Input[int]] = None,
cores: Optional[pulumi.Input[int]] = None,
gpus: Optional[pulumi.Input[int]] = None,
memory: Optional[pulumi.Input[float]] = None):
if core_fraction is not None:
pulumi.set(__self__, "core_fraction", core_fraction)
if cores is not None:
pulumi.set(__self__, "cores", cores)
if gpus is not None:
pulumi.set(__self__, "gpus", gpus)
if memory is not None:
pulumi.set(__self__, "memory", memory)
@property
@pulumi.getter(name="coreFraction")
def core_fraction(self) -> Optional[pulumi.Input[int]]:
return pulumi.get(self, "core_fraction")
@core_fraction.setter
def core_fraction(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "core_fraction", value)
@property
@pulumi.getter
def cores(self) -> Optional[pulumi.Input[int]]:
return pulumi.get(self, "cores")
@cores.setter
def cores(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "cores", value)
@property
@pulumi.getter
def gpus(self) -> Optional[pulumi.Input[int]]:
return pulumi.get(self, "gpus")
@gpus.setter
def gpus(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "gpus", value)
@property
@pulumi.getter
def memory(self) -> Optional[pulumi.Input[float]]:
return pulumi.get(self, "memory")
@memory.setter
def memory(self, value: Optional[pulumi.Input[float]]):
pulumi.set(self, "memory", value)
@pulumi.input_type
class KubernetesNodeGroupInstanceTemplateSchedulingPolicyArgs:
def __init__(__self__, *,
preemptible: Optional[pulumi.Input[bool]] = None):
"""
:param pulumi.Input[bool] preemptible: Specifies if the instance is preemptible. Defaults to false.
---
"""
if preemptible is not None:
pulumi.set(__self__, "preemptible", preemptible)
@property
@pulumi.getter
def preemptible(self) -> Optional[pulumi.Input[bool]]:
"""
Specifies if the instance is preemptible. Defaults to false.
---
"""
return pulumi.get(self, "preemptible")
@preemptible.setter
def preemptible(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "preemptible", value)
@pulumi.input_type
class KubernetesNodeGroupMaintenancePolicyArgs:
def __init__(__self__, *,
auto_repair: pulumi.Input[bool],
auto_upgrade: pulumi.Input[bool],
maintenance_windows: Optional[pulumi.Input[Sequence[pulumi.Input['KubernetesNodeGroupMaintenancePolicyMaintenanceWindowArgs']]]] = None):
"""
:param pulumi.Input[bool] auto_repair: Boolean flag that specifies if node group can be repaired automatically. When omitted, default value is TRUE.
:param pulumi.Input[bool] auto_upgrade: Boolean flag that specifies if node group can be upgraded automatically. When omitted, default value is TRUE.
:param pulumi.Input[Sequence[pulumi.Input['KubernetesNodeGroupMaintenancePolicyMaintenanceWindowArgs']]] maintenance_windows: (Computed) Set of day intervals, when maintenance is allowed for this node group. When omitted, it defaults to any time.
"""
pulumi.set(__self__, "auto_repair", auto_repair)
pulumi.set(__self__, "auto_upgrade", auto_upgrade)
if maintenance_windows is not None:
pulumi.set(__self__, "maintenance_windows", maintenance_windows)
@property
@pulumi.getter(name="autoRepair")
def auto_repair(self) -> pulumi.Input[bool]:
"""
Boolean flag that specifies if node group can be repaired automatically. When omitted, default value is TRUE.
"""
return pulumi.get(self, "auto_repair")
@auto_repair.setter
def auto_repair(self, value: pulumi.Input[bool]):
pulumi.set(self, "auto_repair", value)
@property
@pulumi.getter(name="autoUpgrade")
def auto_upgrade(self) -> pulumi.Input[bool]:
"""
Boolean flag that specifies if node group can be upgraded automatically. When omitted, default value is TRUE.
"""
return pulumi.get(self, "auto_upgrade")
@auto_upgrade.setter
def auto_upgrade(self, value: pulumi.Input[bool]):
pulumi.set(self, "auto_upgrade", value)
@property
@pulumi.getter(name="maintenanceWindows")
def maintenance_windows(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['KubernetesNodeGroupMaintenancePolicyMaintenanceWindowArgs']]]]:
"""
(Computed) Set of day intervals, when maintenance is allowed for this node group. When omitted, it defaults to any time.
"""
return pulumi.get(self, "maintenance_windows")
@maintenance_windows.setter
def maintenance_windows(self, value: Optional[pulumi.Input[Sequence[pulumi.Input['KubernetesNodeGroupMaintenancePolicyMaintenanceWindowArgs']]]]):
pulumi.set(self, "maintenance_windows", value)
@pulumi.input_type
class KubernetesNodeGroupMaintenancePolicyMaintenanceWindowArgs:
def __init__(__self__, *,
duration: pulumi.Input[str],
start_time: pulumi.Input[str],
day: Optional[pulumi.Input[str]] = None):
pulumi.set(__self__, "duration", duration)
pulumi.set(__self__, "start_time", start_time)
if day is not None:
pulumi.set(__self__, "day", day)
@property
@pulumi.getter
def duration(self) -> pulumi.Input[str]:
return pulumi.get(self, "duration")
@duration.setter
def duration(self, value: pulumi.Input[str]):
pulumi.set(self, "duration", value)
@property
@pulumi.getter(name="startTime")
def start_time(self) -> pulumi.Input[str]:
return pulumi.get(self, "start_time")
@start_time.setter
def start_time(self, value: pulumi.Input[str]):
pulumi.set(self, "start_time", value)
@property
@pulumi.getter
def day(self) -> Optional[pulumi.Input[str]]:
return pulumi.get(self, "day")
@day.setter
def day(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "day", value)
@pulumi.input_type
class KubernetesNodeGroupScalePolicyArgs:
def __init__(__self__, *,
auto_scale: Optional[pulumi.Input['KubernetesNodeGroupScalePolicyAutoScaleArgs']] = None,
fixed_scale: Optional[pulumi.Input['KubernetesNodeGroupScalePolicyFixedScaleArgs']] = None):
"""
:param pulumi.Input['KubernetesNodeGroupScalePolicyAutoScaleArgs'] auto_scale: Scale policy for an autoscaled node group. The structure is documented below.
:param pulumi.Input['KubernetesNodeGroupScalePolicyFixedScaleArgs'] fixed_scale: Scale policy for a fixed scale node group. The structure is documented below.
"""
if auto_scale is not None:
pulumi.set(__self__, "auto_scale", auto_scale)
if fixed_scale is not None:
pulumi.set(__self__, "fixed_scale", fixed_scale)
@property
@pulumi.getter(name="autoScale")
def auto_scale(self) -> Optional[pulumi.Input['KubernetesNodeGroupScalePolicyAutoScaleArgs']]:
"""
Scale policy for an autoscaled node group. The structure is documented below.
"""
return pulumi.get(self, "auto_scale")
@auto_scale.setter
def auto_scale(self, value: Optional[pulumi.Input['KubernetesNodeGroupScalePolicyAutoScaleArgs']]):
pulumi.set(self, "auto_scale", value)
@property
@pulumi.getter(name="fixedScale")
def fixed_scale(self) -> Optional[pulumi.Input['KubernetesNodeGroupScalePolicyFixedScaleArgs']]:
"""
Scale policy for a fixed scale node group. The structure is documented below.
"""
return pulumi.get(self, "fixed_scale")
@fixed_scale.setter
def fixed_scale(self, value: Optional[pulumi.Input['KubernetesNodeGroupScalePolicyFixedScaleArgs']]):
pulumi.set(self, "fixed_scale", value)
@pulumi.input_type
class KubernetesNodeGroupScalePolicyAutoScaleArgs:
def __init__(__self__, *,
initial: pulumi.Input[int],
max: pulumi.Input[int],
min: pulumi.Input[int]):
"""
:param pulumi.Input[int] initial: Initial number of instances in the node group.
:param pulumi.Input[int] max: Maximum number of instances in the node group.
:param pulumi.Input[int] min: Minimum number of instances in the node group.
"""
pulumi.set(__self__, "initial", initial)
pulumi.set(__self__, "max", max)
pulumi.set(__self__, "min", min)
@property
@pulumi.getter
def initial(self) -> pulumi.Input[int]:
"""
Initial number of instances in the node group.
"""
return pulumi.get(self, "initial")
@initial.setter
def initial(self, value: pulumi.Input[int]):
pulumi.set(self, "initial", value)
@property
@pulumi.getter
def max(self) -> pulumi.Input[int]:
"""
Maximum number of instances in the node group.
"""
return pulumi.get(self, "max")
@max.setter
def max(self, value: pulumi.Input[int]):
pulumi.set(self, "max", value)
@property
@pulumi.getter
def min(self) -> pulumi.Input[int]:
"""
Minimum number of instances in the node group.
"""
return pulumi.get(self, "min")
@min.setter
def min(self, value: pulumi.Input[int]):
pulumi.set(self, "min", value)
@pulumi.input_type
class KubernetesNodeGroupScalePolicyFixedScaleArgs:
def __init__(__self__, *,
size: Optional[pulumi.Input[int]] = None):
"""
:param pulumi.Input[int] size: The number of instances in the node group.
"""
if size is not None:
pulumi.set(__self__, "size", size)
@property
@pulumi.getter
def size(self) -> Optional[pulumi.Input[int]]:
"""
The number of instances in the node group.
"""
return pulumi.get(self, "size")
@size.setter
def size(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "size", value)
@pulumi.input_type
class KubernetesNodeGroupVersionInfoArgs:
def __init__(__self__, *,
current_version: Optional[pulumi.Input[str]] = None,
new_revision_available: Optional[pulumi.Input[bool]] = None,
new_revision_summary: Optional[pulumi.Input[str]] = None,
version_deprecated: Optional[pulumi.Input[bool]] = None):
"""
:param pulumi.Input[str] current_version: Current Kubernetes version, major.minor (e.g. 1.15).
:param pulumi.Input[bool] new_revision_available: True/false flag.
Newer revisions may include Kubernetes patches (e.g 1.15.1 > 1.15.2) as well
as some internal component updates - new features or bug fixes in yandex-specific
components either on the master or nodes.
:param pulumi.Input[str] new_revision_summary: Human readable description of the changes to be applied
when updating to the latest revision. Empty if new_revision_available is false.
:param pulumi.Input[bool] version_deprecated: True/false flag. The current version is on the deprecation schedule,
component (master or node group) should be upgraded.
"""
if current_version is not None:
pulumi.set(__self__, "current_version", current_version)
if new_revision_available is not None:
pulumi.set(__self__, "new_revision_available", new_revision_available)
if new_revision_summary is not None:
pulumi.set(__self__, "new_revision_summary", new_revision_summary)
if version_deprecated is not None:
pulumi.set(__self__, "version_deprecated", version_deprecated)
@property
@pulumi.getter(name="currentVersion")
def current_version(self) -> Optional[pulumi.Input[str]]:
"""
Current Kubernetes version, major.minor (e.g. 1.15).
"""
return pulumi.get(self, "current_version")
@current_version.setter
def current_version(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "current_version", value)
@property
@pulumi.getter(name="newRevisionAvailable")
def new_revision_available(self) -> Optional[pulumi.Input[bool]]:
"""
True/false flag.
Newer revisions may include Kubernetes patches (e.g 1.15.1 > 1.15.2) as well
as some internal component updates - new features or bug fixes in yandex-specific
components either on the master or nodes.
"""
return pulumi.get(self, "new_revision_available")
@new_revision_available.setter
def new_revision_available(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "new_revision_available", value)
@property
@pulumi.getter(name="newRevisionSummary")
def new_revision_summary(self) -> Optional[pulumi.Input[str]]:
"""
Human readable description of the changes to be applied
when updating to the latest revision. Empty if new_revision_available is false.
"""
return pulumi.get(self, "new_revision_summary")
@new_revision_summary.setter
def new_revision_summary(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "new_revision_summary", value)
@property
@pulumi.getter(name="versionDeprecated")
def version_deprecated(self) -> Optional[pulumi.Input[bool]]:
"""
True/false flag. The current version is on the deprecation schedule,
component (master or node group) should be upgraded.
"""
return pulumi.get(self, "version_deprecated")
@version_deprecated.setter
def version_deprecated(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "version_deprecated", value)
@pulumi.input_type
class LbNetworkLoadBalancerAttachedTargetGroupArgs:
def __init__(__self__, *,
healthchecks: pulumi.Input[Sequence[pulumi.Input['LbNetworkLoadBalancerAttachedTargetGroupHealthcheckArgs']]],
target_group_id: pulumi.Input[str]):
"""
:param pulumi.Input[Sequence[pulumi.Input['LbNetworkLoadBalancerAttachedTargetGroupHealthcheckArgs']]] healthchecks: A HealthCheck resource. The structure is documented below.
:param pulumi.Input[str] target_group_id: ID of the target group.
"""
pulumi.set(__self__, "healthchecks", healthchecks)
pulumi.set(__self__, "target_group_id", target_group_id)
@property
@pulumi.getter
def healthchecks(self) -> pulumi.Input[Sequence[pulumi.Input['LbNetworkLoadBalancerAttachedTargetGroupHealthcheckArgs']]]:
"""
A HealthCheck resource. The structure is documented below.
"""
return pulumi.get(self, "healthchecks")
@healthchecks.setter
def healthchecks(self, value: pulumi.Input[Sequence[pulumi.Input['LbNetworkLoadBalancerAttachedTargetGroupHealthcheckArgs']]]):
pulumi.set(self, "healthchecks", value)
@property
@pulumi.getter(name="targetGroupId")
def target_group_id(self) -> pulumi.Input[str]:
"""
ID of the target group.
"""
return pulumi.get(self, "target_group_id")
@target_group_id.setter
def target_group_id(self, value: pulumi.Input[str]):
pulumi.set(self, "target_group_id", value)
@pulumi.input_type
class LbNetworkLoadBalancerAttachedTargetGroupHealthcheckArgs:
def __init__(__self__, *,
name: pulumi.Input[str],
healthy_threshold: Optional[pulumi.Input[int]] = None,
http_options: Optional[pulumi.Input['LbNetworkLoadBalancerAttachedTargetGroupHealthcheckHttpOptionsArgs']] = None,
interval: Optional[pulumi.Input[int]] = None,
tcp_options: Optional[pulumi.Input['LbNetworkLoadBalancerAttachedTargetGroupHealthcheckTcpOptionsArgs']] = None,
timeout: Optional[pulumi.Input[int]] = None,
unhealthy_threshold: Optional[pulumi.Input[int]] = None):
"""
:param pulumi.Input[str] name: Name of the listener. The name must be unique for each listener on a single load balancer.
:param pulumi.Input[int] healthy_threshold: Number of successful health checks required in order to set the `HEALTHY` status for the target.
:param pulumi.Input['LbNetworkLoadBalancerAttachedTargetGroupHealthcheckHttpOptionsArgs'] http_options: Options for HTTP health check. The structure is documented below.
:param pulumi.Input[int] interval: The interval between health checks. The default is 2 seconds.
:param pulumi.Input['LbNetworkLoadBalancerAttachedTargetGroupHealthcheckTcpOptionsArgs'] tcp_options: Options for TCP health check. The structure is documented below.
:param pulumi.Input[int] timeout: Timeout for a target to return a response for the health check. The default is 1 second.
:param pulumi.Input[int] unhealthy_threshold: Number of failed health checks before changing the status to `UNHEALTHY`. The default is 2.
"""
pulumi.set(__self__, "name", name)
if healthy_threshold is not None:
pulumi.set(__self__, "healthy_threshold", healthy_threshold)
if http_options is not None:
pulumi.set(__self__, "http_options", http_options)
if interval is not None:
pulumi.set(__self__, "interval", interval)
if tcp_options is not None:
pulumi.set(__self__, "tcp_options", tcp_options)
if timeout is not None:
pulumi.set(__self__, "timeout", timeout)
if unhealthy_threshold is not None:
pulumi.set(__self__, "unhealthy_threshold", unhealthy_threshold)
@property
@pulumi.getter
def name(self) -> pulumi.Input[str]:
"""
Name of the listener. The name must be unique for each listener on a single load balancer.
"""
return pulumi.get(self, "name")
@name.setter
def name(self, value: pulumi.Input[str]):
pulumi.set(self, "name", value)
@property
@pulumi.getter(name="healthyThreshold")
def healthy_threshold(self) -> Optional[pulumi.Input[int]]:
"""
Number of successful health checks required in order to set the `HEALTHY` status for the target.
"""
return pulumi.get(self, "healthy_threshold")
@healthy_threshold.setter
def healthy_threshold(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "healthy_threshold", value)
@property
@pulumi.getter(name="httpOptions")
def http_options(self) -> Optional[pulumi.Input['LbNetworkLoadBalancerAttachedTargetGroupHealthcheckHttpOptionsArgs']]:
"""
Options for HTTP health check. The structure is documented below.
"""
return pulumi.get(self, "http_options")
@http_options.setter
def http_options(self, value: Optional[pulumi.Input['LbNetworkLoadBalancerAttachedTargetGroupHealthcheckHttpOptionsArgs']]):
pulumi.set(self, "http_options", value)
@property
@pulumi.getter
def interval(self) -> Optional[pulumi.Input[int]]:
"""
The interval between health checks. The default is 2 seconds.
"""
return pulumi.get(self, "interval")
@interval.setter
def interval(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "interval", value)
@property
@pulumi.getter(name="tcpOptions")
def tcp_options(self) -> Optional[pulumi.Input['LbNetworkLoadBalancerAttachedTargetGroupHealthcheckTcpOptionsArgs']]:
"""
Options for TCP health check. The structure is documented below.
"""
return pulumi.get(self, "tcp_options")
@tcp_options.setter
def tcp_options(self, value: Optional[pulumi.Input['LbNetworkLoadBalancerAttachedTargetGroupHealthcheckTcpOptionsArgs']]):
pulumi.set(self, "tcp_options", value)
@property
@pulumi.getter
def timeout(self) -> Optional[pulumi.Input[int]]:
"""
Timeout for a target to return a response for the health check. The default is 1 second.
"""
return pulumi.get(self, "timeout")
@timeout.setter
def timeout(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "timeout", value)
@property
@pulumi.getter(name="unhealthyThreshold")
def unhealthy_threshold(self) -> Optional[pulumi.Input[int]]:
"""
Number of failed health checks before changing the status to `UNHEALTHY`. The default is 2.
"""
return pulumi.get(self, "unhealthy_threshold")
@unhealthy_threshold.setter
def unhealthy_threshold(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "unhealthy_threshold", value)
@pulumi.input_type
class LbNetworkLoadBalancerAttachedTargetGroupHealthcheckHttpOptionsArgs:
def __init__(__self__, *,
port: pulumi.Input[int],
path: Optional[pulumi.Input[str]] = None):
"""
:param pulumi.Input[int] port: Port for incoming traffic.
:param pulumi.Input[str] path: URL path to set for health checking requests for every target in the target group. For example `/ping`. The default path is `/`.
"""
pulumi.set(__self__, "port", port)
if path is not None:
pulumi.set(__self__, "path", path)
@property
@pulumi.getter
def port(self) -> pulumi.Input[int]:
"""
Port for incoming traffic.
"""
return pulumi.get(self, "port")
@port.setter
def port(self, value: pulumi.Input[int]):
pulumi.set(self, "port", value)
@property
@pulumi.getter
def path(self) -> Optional[pulumi.Input[str]]:
"""
URL path to set for health checking requests for every target in the target group. For example `/ping`. The default path is `/`.
"""
return pulumi.get(self, "path")
@path.setter
def path(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "path", value)
@pulumi.input_type
class LbNetworkLoadBalancerAttachedTargetGroupHealthcheckTcpOptionsArgs:
def __init__(__self__, *,
port: pulumi.Input[int]):
"""
:param pulumi.Input[int] port: Port for incoming traffic.
"""
pulumi.set(__self__, "port", port)
@property
@pulumi.getter
def port(self) -> pulumi.Input[int]:
"""
Port for incoming traffic.
"""
return pulumi.get(self, "port")
@port.setter
def port(self, value: pulumi.Input[int]):
pulumi.set(self, "port", value)
@pulumi.input_type
class LbNetworkLoadBalancerListenerArgs:
def __init__(__self__, *,
name: pulumi.Input[str],
port: pulumi.Input[int],
external_address_spec: Optional[pulumi.Input['LbNetworkLoadBalancerListenerExternalAddressSpecArgs']] = None,
internal_address_spec: Optional[pulumi.Input['LbNetworkLoadBalancerListenerInternalAddressSpecArgs']] = None,
protocol: Optional[pulumi.Input[str]] = None,
target_port: Optional[pulumi.Input[int]] = None):
"""
:param pulumi.Input[str] name: Name of the listener. The name must be unique for each listener on a single load balancer.
:param pulumi.Input[int] port: Port for incoming traffic.
:param pulumi.Input['LbNetworkLoadBalancerListenerExternalAddressSpecArgs'] external_address_spec: External IP address specification. The structure is documented below.
:param pulumi.Input['LbNetworkLoadBalancerListenerInternalAddressSpecArgs'] internal_address_spec: Internal IP address specification. The structure is documented below.
:param pulumi.Input[str] protocol: Protocol for incoming traffic. TCP or UDP and the default is TCP.
:param pulumi.Input[int] target_port: Port of a target. The default is the same as listener's port.
"""
pulumi.set(__self__, "name", name)
pulumi.set(__self__, "port", port)
if external_address_spec is not None:
pulumi.set(__self__, "external_address_spec", external_address_spec)
if internal_address_spec is not None:
pulumi.set(__self__, "internal_address_spec", internal_address_spec)
if protocol is not None:
pulumi.set(__self__, "protocol", protocol)
if target_port is not None:
pulumi.set(__self__, "target_port", target_port)
@property
@pulumi.getter
def name(self) -> pulumi.Input[str]:
"""
Name of the listener. The name must be unique for each listener on a single load balancer.
"""
return pulumi.get(self, "name")
@name.setter
def name(self, value: pulumi.Input[str]):
pulumi.set(self, "name", value)
@property
@pulumi.getter
def port(self) -> pulumi.Input[int]:
"""
Port for incoming traffic.
"""
return pulumi.get(self, "port")
@port.setter
def port(self, value: pulumi.Input[int]):
pulumi.set(self, "port", value)
@property
@pulumi.getter(name="externalAddressSpec")
def external_address_spec(self) -> Optional[pulumi.Input['LbNetworkLoadBalancerListenerExternalAddressSpecArgs']]:
"""
External IP address specification. The structure is documented below.
"""
return pulumi.get(self, "external_address_spec")
@external_address_spec.setter
def external_address_spec(self, value: Optional[pulumi.Input['LbNetworkLoadBalancerListenerExternalAddressSpecArgs']]):
pulumi.set(self, "external_address_spec", value)
@property
@pulumi.getter(name="internalAddressSpec")
def internal_address_spec(self) -> Optional[pulumi.Input['LbNetworkLoadBalancerListenerInternalAddressSpecArgs']]:
"""
Internal IP address specification. The structure is documented below.
"""
return pulumi.get(self, "internal_address_spec")
@internal_address_spec.setter
def internal_address_spec(self, value: Optional[pulumi.Input['LbNetworkLoadBalancerListenerInternalAddressSpecArgs']]):
pulumi.set(self, "internal_address_spec", value)
@property
@pulumi.getter
def protocol(self) -> Optional[pulumi.Input[str]]:
"""
Protocol for incoming traffic. TCP or UDP and the default is TCP.
"""
return pulumi.get(self, "protocol")
@protocol.setter
def protocol(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "protocol", value)
@property
@pulumi.getter(name="targetPort")
def target_port(self) -> Optional[pulumi.Input[int]]:
"""
Port of a target. The default is the same as listener's port.
"""
return pulumi.get(self, "target_port")
@target_port.setter
def target_port(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "target_port", value)
@pulumi.input_type
class LbNetworkLoadBalancerListenerExternalAddressSpecArgs:
def __init__(__self__, *,
address: Optional[pulumi.Input[str]] = None,
ip_version: Optional[pulumi.Input[str]] = None):
"""
:param pulumi.Input[str] address: Internal IP address for a listener. Must belong to the subnet that is referenced in subnet_id. IP address will be allocated if it wasn't been set.
:param pulumi.Input[str] ip_version: IP version of the internal addresses that the load balancer works with. Must be one of ipv4 or ipv6. The default is ipv4.
"""
if address is not None:
pulumi.set(__self__, "address", address)
if ip_version is not None:
pulumi.set(__self__, "ip_version", ip_version)
@property
@pulumi.getter
def address(self) -> Optional[pulumi.Input[str]]:
"""
Internal IP address for a listener. Must belong to the subnet that is referenced in subnet_id. IP address will be allocated if it wasn't been set.
"""
return pulumi.get(self, "address")
@address.setter
def address(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "address", value)
@property
@pulumi.getter(name="ipVersion")
def ip_version(self) -> Optional[pulumi.Input[str]]:
"""
IP version of the internal addresses that the load balancer works with. Must be one of ipv4 or ipv6. The default is ipv4.
"""
return pulumi.get(self, "ip_version")
@ip_version.setter
def ip_version(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "ip_version", value)
@pulumi.input_type
class LbNetworkLoadBalancerListenerInternalAddressSpecArgs:
def __init__(__self__, *,
subnet_id: pulumi.Input[str],
address: Optional[pulumi.Input[str]] = None,
ip_version: Optional[pulumi.Input[str]] = None):
"""
:param pulumi.Input[str] subnet_id: ID of the subnet to which the internal IP address belongs.
:param pulumi.Input[str] address: Internal IP address for a listener. Must belong to the subnet that is referenced in subnet_id. IP address will be allocated if it wasn't been set.
:param pulumi.Input[str] ip_version: IP version of the internal addresses that the load balancer works with. Must be one of ipv4 or ipv6. The default is ipv4.
"""
pulumi.set(__self__, "subnet_id", subnet_id)
if address is not None:
pulumi.set(__self__, "address", address)
if ip_version is not None:
pulumi.set(__self__, "ip_version", ip_version)
@property
@pulumi.getter(name="subnetId")
def subnet_id(self) -> pulumi.Input[str]:
"""
ID of the subnet to which the internal IP address belongs.
"""
return pulumi.get(self, "subnet_id")
@subnet_id.setter
def subnet_id(self, value: pulumi.Input[str]):
pulumi.set(self, "subnet_id", value)
@property
@pulumi.getter
def address(self) -> Optional[pulumi.Input[str]]:
"""
Internal IP address for a listener. Must belong to the subnet that is referenced in subnet_id. IP address will be allocated if it wasn't been set.
"""
return pulumi.get(self, "address")
@address.setter
def address(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "address", value)
@property
@pulumi.getter(name="ipVersion")
def ip_version(self) -> Optional[pulumi.Input[str]]:
"""
IP version of the internal addresses that the load balancer works with. Must be one of ipv4 or ipv6. The default is ipv4.
"""
return pulumi.get(self, "ip_version")
@ip_version.setter
def ip_version(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "ip_version", value)
@pulumi.input_type
class LbTargetGroupTargetArgs:
def __init__(__self__, *,
address: pulumi.Input[str],
subnet_id: pulumi.Input[str]):
"""
:param pulumi.Input[str] address: IP address of the target.
:param pulumi.Input[str] subnet_id: ID of the subnet that targets are connected to.
All targets in the target group must be connected to the same subnet within a single availability zone.
"""
pulumi.set(__self__, "address", address)
pulumi.set(__self__, "subnet_id", subnet_id)
@property
@pulumi.getter
def address(self) -> pulumi.Input[str]:
"""
IP address of the target.
"""
return pulumi.get(self, "address")
@address.setter
def address(self, value: pulumi.Input[str]):
pulumi.set(self, "address", value)
@property
@pulumi.getter(name="subnetId")
def subnet_id(self) -> pulumi.Input[str]:
"""
ID of the subnet that targets are connected to.
All targets in the target group must be connected to the same subnet within a single availability zone.
"""
return pulumi.get(self, "subnet_id")
@subnet_id.setter
def subnet_id(self, value: pulumi.Input[str]):
pulumi.set(self, "subnet_id", value)
@pulumi.input_type
class MdbClickhouseClusterAccessArgs:
def __init__(__self__, *,
data_lens: Optional[pulumi.Input[bool]] = None,
metrika: Optional[pulumi.Input[bool]] = None,
serverless: Optional[pulumi.Input[bool]] = None,
web_sql: Optional[pulumi.Input[bool]] = None):
"""
:param pulumi.Input[bool] data_lens: Allow access for DataLens. Can be either `true` or `false`.
:param pulumi.Input[bool] metrika: Allow access for Yandex.Metrika. Can be either `true` or `false`.
:param pulumi.Input[bool] serverless: Allow access for Serverless. Can be either `true` or `false`.
:param pulumi.Input[bool] web_sql: Allow access for Web SQL. Can be either `true` or `false`.
"""
if data_lens is not None:
pulumi.set(__self__, "data_lens", data_lens)
if metrika is not None:
pulumi.set(__self__, "metrika", metrika)
if serverless is not None:
pulumi.set(__self__, "serverless", serverless)
if web_sql is not None:
pulumi.set(__self__, "web_sql", web_sql)
@property
@pulumi.getter(name="dataLens")
def data_lens(self) -> Optional[pulumi.Input[bool]]:
"""
Allow access for DataLens. Can be either `true` or `false`.
"""
return pulumi.get(self, "data_lens")
@data_lens.setter
def data_lens(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "data_lens", value)
@property
@pulumi.getter
def metrika(self) -> Optional[pulumi.Input[bool]]:
"""
Allow access for Yandex.Metrika. Can be either `true` or `false`.
"""
return pulumi.get(self, "metrika")
@metrika.setter
def metrika(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "metrika", value)
@property
@pulumi.getter
def serverless(self) -> Optional[pulumi.Input[bool]]:
"""
Allow access for Serverless. Can be either `true` or `false`.
"""
return pulumi.get(self, "serverless")
@serverless.setter
def serverless(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "serverless", value)
@property
@pulumi.getter(name="webSql")
def web_sql(self) -> Optional[pulumi.Input[bool]]:
"""
Allow access for Web SQL. Can be either `true` or `false`.
"""
return pulumi.get(self, "web_sql")
@web_sql.setter
def web_sql(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "web_sql", value)
@pulumi.input_type
class MdbClickhouseClusterBackupWindowStartArgs:
def __init__(__self__, *,
hours: Optional[pulumi.Input[int]] = None,
minutes: Optional[pulumi.Input[int]] = None):
"""
:param pulumi.Input[int] hours: The hour at which backup will be started.
:param pulumi.Input[int] minutes: The minute at which backup will be started.
"""
if hours is not None:
pulumi.set(__self__, "hours", hours)
if minutes is not None:
pulumi.set(__self__, "minutes", minutes)
@property
@pulumi.getter
def hours(self) -> Optional[pulumi.Input[int]]:
"""
The hour at which backup will be started.
"""
return pulumi.get(self, "hours")
@hours.setter
def hours(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "hours", value)
@property
@pulumi.getter
def minutes(self) -> Optional[pulumi.Input[int]]:
"""
The minute at which backup will be started.
"""
return pulumi.get(self, "minutes")
@minutes.setter
def minutes(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "minutes", value)
@pulumi.input_type
class MdbClickhouseClusterClickhouseArgs:
def __init__(__self__, *,
resources: pulumi.Input['MdbClickhouseClusterClickhouseResourcesArgs'],
config: Optional[pulumi.Input['MdbClickhouseClusterClickhouseConfigArgs']] = None):
"""
:param pulumi.Input['MdbClickhouseClusterClickhouseResourcesArgs'] resources: Resources allocated to hosts of the ZooKeeper subcluster. The structure is documented below.
:param pulumi.Input['MdbClickhouseClusterClickhouseConfigArgs'] config: Main ClickHouse cluster configuration.
"""
pulumi.set(__self__, "resources", resources)
if config is not None:
pulumi.set(__self__, "config", config)
@property
@pulumi.getter
def resources(self) -> pulumi.Input['MdbClickhouseClusterClickhouseResourcesArgs']:
"""
Resources allocated to hosts of the ZooKeeper subcluster. The structure is documented below.
"""
return pulumi.get(self, "resources")
@resources.setter
def resources(self, value: pulumi.Input['MdbClickhouseClusterClickhouseResourcesArgs']):
pulumi.set(self, "resources", value)
@property
@pulumi.getter
def config(self) -> Optional[pulumi.Input['MdbClickhouseClusterClickhouseConfigArgs']]:
"""
Main ClickHouse cluster configuration.
"""
return pulumi.get(self, "config")
@config.setter
def config(self, value: Optional[pulumi.Input['MdbClickhouseClusterClickhouseConfigArgs']]):
pulumi.set(self, "config", value)
@pulumi.input_type
class MdbClickhouseClusterClickhouseConfigArgs:
def __init__(__self__, *,
background_pool_size: Optional[pulumi.Input[int]] = None,
background_schedule_pool_size: Optional[pulumi.Input[int]] = None,
compressions: Optional[pulumi.Input[Sequence[pulumi.Input['MdbClickhouseClusterClickhouseConfigCompressionArgs']]]] = None,
geobase_uri: Optional[pulumi.Input[str]] = None,
graphite_rollups: Optional[pulumi.Input[Sequence[pulumi.Input['MdbClickhouseClusterClickhouseConfigGraphiteRollupArgs']]]] = None,
kafka: Optional[pulumi.Input['MdbClickhouseClusterClickhouseConfigKafkaArgs']] = None,
kafka_topics: Optional[pulumi.Input[Sequence[pulumi.Input['MdbClickhouseClusterClickhouseConfigKafkaTopicArgs']]]] = None,
keep_alive_timeout: Optional[pulumi.Input[int]] = None,
log_level: Optional[pulumi.Input[str]] = None,
mark_cache_size: Optional[pulumi.Input[int]] = None,
max_concurrent_queries: Optional[pulumi.Input[int]] = None,
max_connections: Optional[pulumi.Input[int]] = None,
max_partition_size_to_drop: Optional[pulumi.Input[int]] = None,
max_table_size_to_drop: Optional[pulumi.Input[int]] = None,
merge_tree: Optional[pulumi.Input['MdbClickhouseClusterClickhouseConfigMergeTreeArgs']] = None,
metric_log_enabled: Optional[pulumi.Input[bool]] = None,
metric_log_retention_size: Optional[pulumi.Input[int]] = None,
metric_log_retention_time: Optional[pulumi.Input[int]] = None,
part_log_retention_size: Optional[pulumi.Input[int]] = None,
part_log_retention_time: Optional[pulumi.Input[int]] = None,
query_log_retention_size: Optional[pulumi.Input[int]] = None,
query_log_retention_time: Optional[pulumi.Input[int]] = None,
query_thread_log_enabled: Optional[pulumi.Input[bool]] = None,
query_thread_log_retention_size: Optional[pulumi.Input[int]] = None,
query_thread_log_retention_time: Optional[pulumi.Input[int]] = None,
rabbitmq: Optional[pulumi.Input['MdbClickhouseClusterClickhouseConfigRabbitmqArgs']] = None,
text_log_enabled: Optional[pulumi.Input[bool]] = None,
text_log_level: Optional[pulumi.Input[str]] = None,
text_log_retention_size: Optional[pulumi.Input[int]] = None,
text_log_retention_time: Optional[pulumi.Input[int]] = None,
timezone: Optional[pulumi.Input[str]] = None,
trace_log_enabled: Optional[pulumi.Input[bool]] = None,
trace_log_retention_size: Optional[pulumi.Input[int]] = None,
trace_log_retention_time: Optional[pulumi.Input[int]] = None,
uncompressed_cache_size: Optional[pulumi.Input[int]] = None):
"""
:param pulumi.Input[Sequence[pulumi.Input['MdbClickhouseClusterClickhouseConfigCompressionArgs']]] compressions: Data compression configuration. The structure is documented below.
:param pulumi.Input[Sequence[pulumi.Input['MdbClickhouseClusterClickhouseConfigGraphiteRollupArgs']]] graphite_rollups: Graphite rollup configuration. The structure is documented below.
:param pulumi.Input['MdbClickhouseClusterClickhouseConfigKafkaArgs'] kafka: Kafka connection configuration. The structure is documented below.
:param pulumi.Input[Sequence[pulumi.Input['MdbClickhouseClusterClickhouseConfigKafkaTopicArgs']]] kafka_topics: Kafka topic connection configuration. The structure is documented below.
:param pulumi.Input['MdbClickhouseClusterClickhouseConfigMergeTreeArgs'] merge_tree: MergeTree engine configuration. The structure is documented below.
:param pulumi.Input['MdbClickhouseClusterClickhouseConfigRabbitmqArgs'] rabbitmq: RabbitMQ connection configuration. The structure is documented below.
"""
if background_pool_size is not None:
pulumi.set(__self__, "background_pool_size", background_pool_size)
if background_schedule_pool_size is not None:
pulumi.set(__self__, "background_schedule_pool_size", background_schedule_pool_size)
if compressions is not None:
pulumi.set(__self__, "compressions", compressions)
if geobase_uri is not None:
pulumi.set(__self__, "geobase_uri", geobase_uri)
if graphite_rollups is not None:
pulumi.set(__self__, "graphite_rollups", graphite_rollups)
if kafka is not None:
pulumi.set(__self__, "kafka", kafka)
if kafka_topics is not None:
pulumi.set(__self__, "kafka_topics", kafka_topics)
if keep_alive_timeout is not None:
pulumi.set(__self__, "keep_alive_timeout", keep_alive_timeout)
if log_level is not None:
pulumi.set(__self__, "log_level", log_level)
if mark_cache_size is not None:
pulumi.set(__self__, "mark_cache_size", mark_cache_size)
if max_concurrent_queries is not None:
pulumi.set(__self__, "max_concurrent_queries", max_concurrent_queries)
if max_connections is not None:
pulumi.set(__self__, "max_connections", max_connections)
if max_partition_size_to_drop is not None:
pulumi.set(__self__, "max_partition_size_to_drop", max_partition_size_to_drop)
if max_table_size_to_drop is not None:
pulumi.set(__self__, "max_table_size_to_drop", max_table_size_to_drop)
if merge_tree is not None:
pulumi.set(__self__, "merge_tree", merge_tree)
if metric_log_enabled is not None:
pulumi.set(__self__, "metric_log_enabled", metric_log_enabled)
if metric_log_retention_size is not None:
pulumi.set(__self__, "metric_log_retention_size", metric_log_retention_size)
if metric_log_retention_time is not None:
pulumi.set(__self__, "metric_log_retention_time", metric_log_retention_time)
if part_log_retention_size is not None:
pulumi.set(__self__, "part_log_retention_size", part_log_retention_size)
if part_log_retention_time is not None:
pulumi.set(__self__, "part_log_retention_time", part_log_retention_time)
if query_log_retention_size is not None:
pulumi.set(__self__, "query_log_retention_size", query_log_retention_size)
if query_log_retention_time is not None:
pulumi.set(__self__, "query_log_retention_time", query_log_retention_time)
if query_thread_log_enabled is not None:
pulumi.set(__self__, "query_thread_log_enabled", query_thread_log_enabled)
if query_thread_log_retention_size is not None:
pulumi.set(__self__, "query_thread_log_retention_size", query_thread_log_retention_size)
if query_thread_log_retention_time is not None:
pulumi.set(__self__, "query_thread_log_retention_time", query_thread_log_retention_time)
if rabbitmq is not None:
pulumi.set(__self__, "rabbitmq", rabbitmq)
if text_log_enabled is not None:
pulumi.set(__self__, "text_log_enabled", text_log_enabled)
if text_log_level is not None:
pulumi.set(__self__, "text_log_level", text_log_level)
if text_log_retention_size is not None:
pulumi.set(__self__, "text_log_retention_size", text_log_retention_size)
if text_log_retention_time is not None:
pulumi.set(__self__, "text_log_retention_time", text_log_retention_time)
if timezone is not None:
pulumi.set(__self__, "timezone", timezone)
if trace_log_enabled is not None:
pulumi.set(__self__, "trace_log_enabled", trace_log_enabled)
if trace_log_retention_size is not None:
pulumi.set(__self__, "trace_log_retention_size", trace_log_retention_size)
if trace_log_retention_time is not None:
pulumi.set(__self__, "trace_log_retention_time", trace_log_retention_time)
if uncompressed_cache_size is not None:
pulumi.set(__self__, "uncompressed_cache_size", uncompressed_cache_size)
@property
@pulumi.getter(name="backgroundPoolSize")
def background_pool_size(self) -> Optional[pulumi.Input[int]]:
return pulumi.get(self, "background_pool_size")
@background_pool_size.setter
def background_pool_size(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "background_pool_size", value)
@property
@pulumi.getter(name="backgroundSchedulePoolSize")
def background_schedule_pool_size(self) -> Optional[pulumi.Input[int]]:
return pulumi.get(self, "background_schedule_pool_size")
@background_schedule_pool_size.setter
def background_schedule_pool_size(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "background_schedule_pool_size", value)
@property
@pulumi.getter
def compressions(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['MdbClickhouseClusterClickhouseConfigCompressionArgs']]]]:
"""
Data compression configuration. The structure is documented below.
"""
return pulumi.get(self, "compressions")
@compressions.setter
def compressions(self, value: Optional[pulumi.Input[Sequence[pulumi.Input['MdbClickhouseClusterClickhouseConfigCompressionArgs']]]]):
pulumi.set(self, "compressions", value)
@property
@pulumi.getter(name="geobaseUri")
def geobase_uri(self) -> Optional[pulumi.Input[str]]:
return pulumi.get(self, "geobase_uri")
@geobase_uri.setter
def geobase_uri(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "geobase_uri", value)
@property
@pulumi.getter(name="graphiteRollups")
def graphite_rollups(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['MdbClickhouseClusterClickhouseConfigGraphiteRollupArgs']]]]:
"""
Graphite rollup configuration. The structure is documented below.
"""
return pulumi.get(self, "graphite_rollups")
@graphite_rollups.setter
def graphite_rollups(self, value: Optional[pulumi.Input[Sequence[pulumi.Input['MdbClickhouseClusterClickhouseConfigGraphiteRollupArgs']]]]):
pulumi.set(self, "graphite_rollups", value)
@property
@pulumi.getter
def kafka(self) -> Optional[pulumi.Input['MdbClickhouseClusterClickhouseConfigKafkaArgs']]:
"""
Kafka connection configuration. The structure is documented below.
"""
return pulumi.get(self, "kafka")
@kafka.setter
def kafka(self, value: Optional[pulumi.Input['MdbClickhouseClusterClickhouseConfigKafkaArgs']]):
pulumi.set(self, "kafka", value)
@property
@pulumi.getter(name="kafkaTopics")
def kafka_topics(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['MdbClickhouseClusterClickhouseConfigKafkaTopicArgs']]]]:
"""
Kafka topic connection configuration. The structure is documented below.
"""
return pulumi.get(self, "kafka_topics")
@kafka_topics.setter
def kafka_topics(self, value: Optional[pulumi.Input[Sequence[pulumi.Input['MdbClickhouseClusterClickhouseConfigKafkaTopicArgs']]]]):
pulumi.set(self, "kafka_topics", value)
@property
@pulumi.getter(name="keepAliveTimeout")
def keep_alive_timeout(self) -> Optional[pulumi.Input[int]]:
return pulumi.get(self, "keep_alive_timeout")
@keep_alive_timeout.setter
def keep_alive_timeout(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "keep_alive_timeout", value)
@property
@pulumi.getter(name="logLevel")
def log_level(self) -> Optional[pulumi.Input[str]]:
return pulumi.get(self, "log_level")
@log_level.setter
def log_level(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "log_level", value)
@property
@pulumi.getter(name="markCacheSize")
def mark_cache_size(self) -> Optional[pulumi.Input[int]]:
return pulumi.get(self, "mark_cache_size")
@mark_cache_size.setter
def mark_cache_size(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "mark_cache_size", value)
@property
@pulumi.getter(name="maxConcurrentQueries")
def max_concurrent_queries(self) -> Optional[pulumi.Input[int]]:
return pulumi.get(self, "max_concurrent_queries")
@max_concurrent_queries.setter
def max_concurrent_queries(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "max_concurrent_queries", value)
@property
@pulumi.getter(name="maxConnections")
def max_connections(self) -> Optional[pulumi.Input[int]]:
return pulumi.get(self, "max_connections")
@max_connections.setter
def max_connections(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "max_connections", value)
@property
@pulumi.getter(name="maxPartitionSizeToDrop")
def max_partition_size_to_drop(self) -> Optional[pulumi.Input[int]]:
return pulumi.get(self, "max_partition_size_to_drop")
@max_partition_size_to_drop.setter
def max_partition_size_to_drop(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "max_partition_size_to_drop", value)
@property
@pulumi.getter(name="maxTableSizeToDrop")
def max_table_size_to_drop(self) -> Optional[pulumi.Input[int]]:
return pulumi.get(self, "max_table_size_to_drop")
@max_table_size_to_drop.setter
def max_table_size_to_drop(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "max_table_size_to_drop", value)
@property
@pulumi.getter(name="mergeTree")
def merge_tree(self) -> Optional[pulumi.Input['MdbClickhouseClusterClickhouseConfigMergeTreeArgs']]:
"""
MergeTree engine configuration. The structure is documented below.
"""
return pulumi.get(self, "merge_tree")
@merge_tree.setter
def merge_tree(self, value: Optional[pulumi.Input['MdbClickhouseClusterClickhouseConfigMergeTreeArgs']]):
pulumi.set(self, "merge_tree", value)
@property
@pulumi.getter(name="metricLogEnabled")
def metric_log_enabled(self) -> Optional[pulumi.Input[bool]]:
return pulumi.get(self, "metric_log_enabled")
@metric_log_enabled.setter
def metric_log_enabled(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "metric_log_enabled", value)
@property
@pulumi.getter(name="metricLogRetentionSize")
def metric_log_retention_size(self) -> Optional[pulumi.Input[int]]:
return pulumi.get(self, "metric_log_retention_size")
@metric_log_retention_size.setter
def metric_log_retention_size(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "metric_log_retention_size", value)
@property
@pulumi.getter(name="metricLogRetentionTime")
def metric_log_retention_time(self) -> Optional[pulumi.Input[int]]:
return pulumi.get(self, "metric_log_retention_time")
@metric_log_retention_time.setter
def metric_log_retention_time(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "metric_log_retention_time", value)
@property
@pulumi.getter(name="partLogRetentionSize")
def part_log_retention_size(self) -> Optional[pulumi.Input[int]]:
return pulumi.get(self, "part_log_retention_size")
@part_log_retention_size.setter
def part_log_retention_size(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "part_log_retention_size", value)
@property
@pulumi.getter(name="partLogRetentionTime")
def part_log_retention_time(self) -> Optional[pulumi.Input[int]]:
return pulumi.get(self, "part_log_retention_time")
@part_log_retention_time.setter
def part_log_retention_time(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "part_log_retention_time", value)
@property
@pulumi.getter(name="queryLogRetentionSize")
def query_log_retention_size(self) -> Optional[pulumi.Input[int]]:
return pulumi.get(self, "query_log_retention_size")
@query_log_retention_size.setter
def query_log_retention_size(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "query_log_retention_size", value)
@property
@pulumi.getter(name="queryLogRetentionTime")
def query_log_retention_time(self) -> Optional[pulumi.Input[int]]:
return pulumi.get(self, "query_log_retention_time")
@query_log_retention_time.setter
def query_log_retention_time(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "query_log_retention_time", value)
@property
@pulumi.getter(name="queryThreadLogEnabled")
def query_thread_log_enabled(self) -> Optional[pulumi.Input[bool]]:
return pulumi.get(self, "query_thread_log_enabled")
@query_thread_log_enabled.setter
def query_thread_log_enabled(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "query_thread_log_enabled", value)
@property
@pulumi.getter(name="queryThreadLogRetentionSize")
def query_thread_log_retention_size(self) -> Optional[pulumi.Input[int]]:
return pulumi.get(self, "query_thread_log_retention_size")
@query_thread_log_retention_size.setter
def query_thread_log_retention_size(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "query_thread_log_retention_size", value)
@property
@pulumi.getter(name="queryThreadLogRetentionTime")
def query_thread_log_retention_time(self) -> Optional[pulumi.Input[int]]:
return pulumi.get(self, "query_thread_log_retention_time")
@query_thread_log_retention_time.setter
def query_thread_log_retention_time(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "query_thread_log_retention_time", value)
@property
@pulumi.getter
def rabbitmq(self) -> Optional[pulumi.Input['MdbClickhouseClusterClickhouseConfigRabbitmqArgs']]:
"""
RabbitMQ connection configuration. The structure is documented below.
"""
return pulumi.get(self, "rabbitmq")
@rabbitmq.setter
def rabbitmq(self, value: Optional[pulumi.Input['MdbClickhouseClusterClickhouseConfigRabbitmqArgs']]):
pulumi.set(self, "rabbitmq", value)
@property
@pulumi.getter(name="textLogEnabled")
def text_log_enabled(self) -> Optional[pulumi.Input[bool]]:
return pulumi.get(self, "text_log_enabled")
@text_log_enabled.setter
def text_log_enabled(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "text_log_enabled", value)
@property
@pulumi.getter(name="textLogLevel")
def text_log_level(self) -> Optional[pulumi.Input[str]]:
return pulumi.get(self, "text_log_level")
@text_log_level.setter
def text_log_level(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "text_log_level", value)
@property
@pulumi.getter(name="textLogRetentionSize")
def text_log_retention_size(self) -> Optional[pulumi.Input[int]]:
return pulumi.get(self, "text_log_retention_size")
@text_log_retention_size.setter
def text_log_retention_size(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "text_log_retention_size", value)
@property
@pulumi.getter(name="textLogRetentionTime")
def text_log_retention_time(self) -> Optional[pulumi.Input[int]]:
return pulumi.get(self, "text_log_retention_time")
@text_log_retention_time.setter
def text_log_retention_time(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "text_log_retention_time", value)
@property
@pulumi.getter
def timezone(self) -> Optional[pulumi.Input[str]]:
return pulumi.get(self, "timezone")
@timezone.setter
def timezone(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "timezone", value)
@property
@pulumi.getter(name="traceLogEnabled")
def trace_log_enabled(self) -> Optional[pulumi.Input[bool]]:
return pulumi.get(self, "trace_log_enabled")
@trace_log_enabled.setter
def trace_log_enabled(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "trace_log_enabled", value)
@property
@pulumi.getter(name="traceLogRetentionSize")
def trace_log_retention_size(self) -> Optional[pulumi.Input[int]]:
return pulumi.get(self, "trace_log_retention_size")
@trace_log_retention_size.setter
def trace_log_retention_size(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "trace_log_retention_size", value)
@property
@pulumi.getter(name="traceLogRetentionTime")
def trace_log_retention_time(self) -> Optional[pulumi.Input[int]]:
return pulumi.get(self, "trace_log_retention_time")
@trace_log_retention_time.setter
def trace_log_retention_time(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "trace_log_retention_time", value)
@property
@pulumi.getter(name="uncompressedCacheSize")
def uncompressed_cache_size(self) -> Optional[pulumi.Input[int]]:
return pulumi.get(self, "uncompressed_cache_size")
@uncompressed_cache_size.setter
def uncompressed_cache_size(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "uncompressed_cache_size", value)
@pulumi.input_type
class MdbClickhouseClusterClickhouseConfigCompressionArgs:
def __init__(__self__, *,
method: pulumi.Input[str],
min_part_size: pulumi.Input[int],
min_part_size_ratio: pulumi.Input[float]):
"""
:param pulumi.Input[str] method: Method: Compression method. Two methods are available: LZ4 and zstd.
:param pulumi.Input[int] min_part_size: Min part size: Minimum size (in bytes) of a data part in a table. ClickHouse only applies the rule to tables with data parts greater than or equal to the Min part size value.
:param pulumi.Input[float] min_part_size_ratio: Min part size ratio: Minimum table part size to total table size ratio. ClickHouse only applies the rule to tables in which this ratio is greater than or equal to the Min part size ratio value.
"""
pulumi.set(__self__, "method", method)
pulumi.set(__self__, "min_part_size", min_part_size)
pulumi.set(__self__, "min_part_size_ratio", min_part_size_ratio)
@property
@pulumi.getter
def method(self) -> pulumi.Input[str]:
"""
Method: Compression method. Two methods are available: LZ4 and zstd.
"""
return pulumi.get(self, "method")
@method.setter
def method(self, value: pulumi.Input[str]):
pulumi.set(self, "method", value)
@property
@pulumi.getter(name="minPartSize")
def min_part_size(self) -> pulumi.Input[int]:
"""
Min part size: Minimum size (in bytes) of a data part in a table. ClickHouse only applies the rule to tables with data parts greater than or equal to the Min part size value.
"""
return pulumi.get(self, "min_part_size")
@min_part_size.setter
def min_part_size(self, value: pulumi.Input[int]):
pulumi.set(self, "min_part_size", value)
@property
@pulumi.getter(name="minPartSizeRatio")
def min_part_size_ratio(self) -> pulumi.Input[float]:
"""
Min part size ratio: Minimum table part size to total table size ratio. ClickHouse only applies the rule to tables in which this ratio is greater than or equal to the Min part size ratio value.
"""
return pulumi.get(self, "min_part_size_ratio")
@min_part_size_ratio.setter
def min_part_size_ratio(self, value: pulumi.Input[float]):
pulumi.set(self, "min_part_size_ratio", value)
@pulumi.input_type
class MdbClickhouseClusterClickhouseConfigGraphiteRollupArgs:
def __init__(__self__, *,
name: pulumi.Input[str],
patterns: Optional[pulumi.Input[Sequence[pulumi.Input['MdbClickhouseClusterClickhouseConfigGraphiteRollupPatternArgs']]]] = None):
"""
:param pulumi.Input[str] name: Graphite rollup configuration name.
:param pulumi.Input[Sequence[pulumi.Input['MdbClickhouseClusterClickhouseConfigGraphiteRollupPatternArgs']]] patterns: Set of thinning rules.
"""
pulumi.set(__self__, "name", name)
if patterns is not None:
pulumi.set(__self__, "patterns", patterns)
@property
@pulumi.getter
def name(self) -> pulumi.Input[str]:
"""
Graphite rollup configuration name.
"""
return pulumi.get(self, "name")
@name.setter
def name(self, value: pulumi.Input[str]):
pulumi.set(self, "name", value)
@property
@pulumi.getter
def patterns(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['MdbClickhouseClusterClickhouseConfigGraphiteRollupPatternArgs']]]]:
"""
Set of thinning rules.
"""
return pulumi.get(self, "patterns")
@patterns.setter
def patterns(self, value: Optional[pulumi.Input[Sequence[pulumi.Input['MdbClickhouseClusterClickhouseConfigGraphiteRollupPatternArgs']]]]):
pulumi.set(self, "patterns", value)
@pulumi.input_type
class MdbClickhouseClusterClickhouseConfigGraphiteRollupPatternArgs:
def __init__(__self__, *,
function: pulumi.Input[str],
regexp: Optional[pulumi.Input[str]] = None,
retentions: Optional[pulumi.Input[Sequence[pulumi.Input['MdbClickhouseClusterClickhouseConfigGraphiteRollupPatternRetentionArgs']]]] = None):
"""
:param pulumi.Input[str] function: Aggregation function name.
:param pulumi.Input[str] regexp: Regular expression that the metric name must match.
:param pulumi.Input[Sequence[pulumi.Input['MdbClickhouseClusterClickhouseConfigGraphiteRollupPatternRetentionArgs']]] retentions: Retain parameters.
"""
pulumi.set(__self__, "function", function)
if regexp is not None:
pulumi.set(__self__, "regexp", regexp)
if retentions is not None:
pulumi.set(__self__, "retentions", retentions)
@property
@pulumi.getter
def function(self) -> pulumi.Input[str]:
"""
Aggregation function name.
"""
return pulumi.get(self, "function")
@function.setter
def function(self, value: pulumi.Input[str]):
pulumi.set(self, "function", value)
@property
@pulumi.getter
def regexp(self) -> Optional[pulumi.Input[str]]:
"""
Regular expression that the metric name must match.
"""
return pulumi.get(self, "regexp")
@regexp.setter
def regexp(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "regexp", value)
@property
@pulumi.getter
def retentions(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['MdbClickhouseClusterClickhouseConfigGraphiteRollupPatternRetentionArgs']]]]:
"""
Retain parameters.
"""
return pulumi.get(self, "retentions")
@retentions.setter
def retentions(self, value: Optional[pulumi.Input[Sequence[pulumi.Input['MdbClickhouseClusterClickhouseConfigGraphiteRollupPatternRetentionArgs']]]]):
pulumi.set(self, "retentions", value)
@pulumi.input_type
class MdbClickhouseClusterClickhouseConfigGraphiteRollupPatternRetentionArgs:
def __init__(__self__, *,
age: pulumi.Input[int],
precision: pulumi.Input[int]):
"""
:param pulumi.Input[int] age: Minimum data age in seconds.
:param pulumi.Input[int] precision: Accuracy of determining the age of the data in seconds.
"""
pulumi.set(__self__, "age", age)
pulumi.set(__self__, "precision", precision)
@property
@pulumi.getter
def age(self) -> pulumi.Input[int]:
"""
Minimum data age in seconds.
"""
return pulumi.get(self, "age")
@age.setter
def age(self, value: pulumi.Input[int]):
pulumi.set(self, "age", value)
@property
@pulumi.getter
def precision(self) -> pulumi.Input[int]:
"""
Accuracy of determining the age of the data in seconds.
"""
return pulumi.get(self, "precision")
@precision.setter
def precision(self, value: pulumi.Input[int]):
pulumi.set(self, "precision", value)
@pulumi.input_type
class MdbClickhouseClusterClickhouseConfigKafkaArgs:
def __init__(__self__, *,
sasl_mechanism: Optional[pulumi.Input[str]] = None,
sasl_password: Optional[pulumi.Input[str]] = None,
sasl_username: Optional[pulumi.Input[str]] = None,
security_protocol: Optional[pulumi.Input[str]] = None):
"""
:param pulumi.Input[str] sasl_mechanism: SASL mechanism used in kafka authentication.
:param pulumi.Input[str] sasl_password: <PASSWORD> on kafka server.
:param pulumi.Input[str] sasl_username: Username on kafka server.
:param pulumi.Input[str] security_protocol: Security protocol used to connect to kafka server.
"""
if sasl_mechanism is not None:
pulumi.set(__self__, "sasl_mechanism", sasl_mechanism)
if sasl_password is not None:
pulumi.set(__self__, "sasl_password", sasl_password)
if sasl_username is not None:
pulumi.set(__self__, "sasl_username", sasl_username)
if security_protocol is not None:
pulumi.set(__self__, "security_protocol", security_protocol)
@property
@pulumi.getter(name="saslMechanism")
def sasl_mechanism(self) -> Optional[pulumi.Input[str]]:
"""
SASL mechanism used in kafka authentication.
"""
return pulumi.get(self, "sasl_mechanism")
@sasl_mechanism.setter
def sasl_mechanism(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "sasl_mechanism", value)
@property
@pulumi.getter(name="saslPassword")
def sasl_password(self) -> Optional[pulumi.Input[str]]:
"""
User password on kafka server.
"""
return pulumi.get(self, "sasl_password")
@sasl_password.setter
def sasl_password(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "sasl_password", value)
@property
@pulumi.getter(name="saslUsername")
def sasl_username(self) -> Optional[pulumi.Input[str]]:
"""
Username on kafka server.
"""
return pulumi.get(self, "sasl_username")
@sasl_username.setter
def sasl_username(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "sasl_username", value)
@property
@pulumi.getter(name="securityProtocol")
def security_protocol(self) -> Optional[pulumi.Input[str]]:
"""
Security protocol used to connect to kafka server.
"""
return pulumi.get(self, "security_protocol")
@security_protocol.setter
def security_protocol(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "security_protocol", value)
@pulumi.input_type
class MdbClickhouseClusterClickhouseConfigKafkaTopicArgs:
def __init__(__self__, *,
name: pulumi.Input[str],
settings: Optional[pulumi.Input['MdbClickhouseClusterClickhouseConfigKafkaTopicSettingsArgs']] = None):
"""
:param pulumi.Input[str] name: Graphite rollup configuration name.
:param pulumi.Input['MdbClickhouseClusterClickhouseConfigKafkaTopicSettingsArgs'] settings: Kafka connection settngs sanem as `kafka` block.
"""
pulumi.set(__self__, "name", name)
if settings is not None:
pulumi.set(__self__, "settings", settings)
@property
@pulumi.getter
def name(self) -> pulumi.Input[str]:
"""
Graphite rollup configuration name.
"""
return pulumi.get(self, "name")
@name.setter
def name(self, value: pulumi.Input[str]):
pulumi.set(self, "name", value)
@property
@pulumi.getter
def settings(self) -> Optional[pulumi.Input['MdbClickhouseClusterClickhouseConfigKafkaTopicSettingsArgs']]:
"""
Kafka connection settngs sanem as `kafka` block.
"""
return pulumi.get(self, "settings")
@settings.setter
def settings(self, value: Optional[pulumi.Input['MdbClickhouseClusterClickhouseConfigKafkaTopicSettingsArgs']]):
pulumi.set(self, "settings", value)
@pulumi.input_type
class MdbClickhouseClusterClickhouseConfigKafkaTopicSettingsArgs:
def __init__(__self__, *,
sasl_mechanism: Optional[pulumi.Input[str]] = None,
sasl_password: Optional[pulumi.Input[str]] = None,
sasl_username: Optional[pulumi.Input[str]] = None,
security_protocol: Optional[pulumi.Input[str]] = None):
"""
:param pulumi.Input[str] sasl_mechanism: SASL mechanism used in kafka authentication.
:param pulumi.Input[str] sasl_password: <PASSWORD> on kafka server.
:param pulumi.Input[str] sasl_username: Username on kafka server.
:param pulumi.Input[str] security_protocol: Security protocol used to connect to kafka server.
"""
if sasl_mechanism is not None:
pulumi.set(__self__, "sasl_mechanism", sasl_mechanism)
if sasl_password is not None:
pulumi.set(__self__, "sasl_password", sasl_password)
if sasl_username is not None:
pulumi.set(__self__, "sasl_username", sasl_username)
if security_protocol is not None:
pulumi.set(__self__, "security_protocol", security_protocol)
@property
@pulumi.getter(name="saslMechanism")
def sasl_mechanism(self) -> Optional[pulumi.Input[str]]:
"""
SASL mechanism used in kafka authentication.
"""
return pulumi.get(self, "sasl_mechanism")
@sasl_mechanism.setter
def sasl_mechanism(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "sasl_mechanism", value)
@property
@pulumi.getter(name="saslPassword")
def sasl_password(self) -> Optional[pulumi.Input[str]]:
"""
User password on kafka server.
"""
return pulumi.get(self, "sasl_password")
@sasl_password.setter
def sasl_password(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "sasl_password", value)
@property
@pulumi.getter(name="saslUsername")
def sasl_username(self) -> Optional[pulumi.Input[str]]:
"""
Username on kafka server.
"""
return pulumi.get(self, "sasl_username")
@sasl_username.setter
def sasl_username(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "sasl_username", value)
@property
@pulumi.getter(name="securityProtocol")
def security_protocol(self) -> Optional[pulumi.Input[str]]:
"""
Security protocol used to connect to kafka server.
"""
return pulumi.get(self, "security_protocol")
@security_protocol.setter
def security_protocol(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "security_protocol", value)
@pulumi.input_type
class MdbClickhouseClusterClickhouseConfigMergeTreeArgs:
def __init__(__self__, *,
max_bytes_to_merge_at_min_space_in_pool: Optional[pulumi.Input[int]] = None,
max_replicated_merges_in_queue: Optional[pulumi.Input[int]] = None,
number_of_free_entries_in_pool_to_lower_max_size_of_merge: Optional[pulumi.Input[int]] = None,
parts_to_delay_insert: Optional[pulumi.Input[int]] = None,
parts_to_throw_insert: Optional[pulumi.Input[int]] = None,
replicated_deduplication_window: Optional[pulumi.Input[int]] = None,
replicated_deduplication_window_seconds: Optional[pulumi.Input[int]] = None):
"""
:param pulumi.Input[int] max_bytes_to_merge_at_min_space_in_pool: Max bytes to merge at min space in pool: Maximum total size of a data part to merge when the number of free threads in the background pool is minimum.
:param pulumi.Input[int] max_replicated_merges_in_queue: Max replicated merges in queue: Maximum number of merge tasks that can be in the ReplicatedMergeTree queue at the same time.
:param pulumi.Input[int] number_of_free_entries_in_pool_to_lower_max_size_of_merge: Number of free entries in pool to lower max size of merge: Threshold value of free entries in the pool. If the number of entries in the pool falls below this value, ClickHouse reduces the maximum size of a data part to merge. This helps handle small merges faster, rather than filling the pool with lengthy merges.
:param pulumi.Input[int] parts_to_delay_insert: Parts to delay insert: Number of active data parts in a table, on exceeding which ClickHouse starts artificially reduce the rate of inserting data into the table.
:param pulumi.Input[int] parts_to_throw_insert: Parts to throw insert: Threshold value of active data parts in a table, on exceeding which ClickHouse throws the 'Too many parts ...' exception.
:param pulumi.Input[int] replicated_deduplication_window: Replicated deduplication window: Number of recent hash blocks that ZooKeeper will store (the old ones will be deleted).
:param pulumi.Input[int] replicated_deduplication_window_seconds: Replicated deduplication window seconds: Time during which ZooKeeper stores the hash blocks (the old ones wil be deleted).
"""
if max_bytes_to_merge_at_min_space_in_pool is not None:
pulumi.set(__self__, "max_bytes_to_merge_at_min_space_in_pool", max_bytes_to_merge_at_min_space_in_pool)
if max_replicated_merges_in_queue is not None:
pulumi.set(__self__, "max_replicated_merges_in_queue", max_replicated_merges_in_queue)
if number_of_free_entries_in_pool_to_lower_max_size_of_merge is not None:
pulumi.set(__self__, "number_of_free_entries_in_pool_to_lower_max_size_of_merge", number_of_free_entries_in_pool_to_lower_max_size_of_merge)
if parts_to_delay_insert is not None:
pulumi.set(__self__, "parts_to_delay_insert", parts_to_delay_insert)
if parts_to_throw_insert is not None:
pulumi.set(__self__, "parts_to_throw_insert", parts_to_throw_insert)
if replicated_deduplication_window is not None:
pulumi.set(__self__, "replicated_deduplication_window", replicated_deduplication_window)
if replicated_deduplication_window_seconds is not None:
pulumi.set(__self__, "replicated_deduplication_window_seconds", replicated_deduplication_window_seconds)
@property
@pulumi.getter(name="maxBytesToMergeAtMinSpaceInPool")
def max_bytes_to_merge_at_min_space_in_pool(self) -> Optional[pulumi.Input[int]]:
"""
Max bytes to merge at min space in pool: Maximum total size of a data part to merge when the number of free threads in the background pool is minimum.
"""
return pulumi.get(self, "max_bytes_to_merge_at_min_space_in_pool")
@max_bytes_to_merge_at_min_space_in_pool.setter
def max_bytes_to_merge_at_min_space_in_pool(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "max_bytes_to_merge_at_min_space_in_pool", value)
@property
@pulumi.getter(name="maxReplicatedMergesInQueue")
def max_replicated_merges_in_queue(self) -> Optional[pulumi.Input[int]]:
"""
Max replicated merges in queue: Maximum number of merge tasks that can be in the ReplicatedMergeTree queue at the same time.
"""
return pulumi.get(self, "max_replicated_merges_in_queue")
@max_replicated_merges_in_queue.setter
def max_replicated_merges_in_queue(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "max_replicated_merges_in_queue", value)
@property
@pulumi.getter(name="numberOfFreeEntriesInPoolToLowerMaxSizeOfMerge")
def number_of_free_entries_in_pool_to_lower_max_size_of_merge(self) -> Optional[pulumi.Input[int]]:
"""
Number of free entries in pool to lower max size of merge: Threshold value of free entries in the pool. If the number of entries in the pool falls below this value, ClickHouse reduces the maximum size of a data part to merge. This helps handle small merges faster, rather than filling the pool with lengthy merges.
"""
return pulumi.get(self, "number_of_free_entries_in_pool_to_lower_max_size_of_merge")
@number_of_free_entries_in_pool_to_lower_max_size_of_merge.setter
def number_of_free_entries_in_pool_to_lower_max_size_of_merge(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "number_of_free_entries_in_pool_to_lower_max_size_of_merge", value)
@property
@pulumi.getter(name="partsToDelayInsert")
def parts_to_delay_insert(self) -> Optional[pulumi.Input[int]]:
"""
Parts to delay insert: Number of active data parts in a table, on exceeding which ClickHouse starts artificially reduce the rate of inserting data into the table.
"""
return pulumi.get(self, "parts_to_delay_insert")
@parts_to_delay_insert.setter
def parts_to_delay_insert(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "parts_to_delay_insert", value)
@property
@pulumi.getter(name="partsToThrowInsert")
def parts_to_throw_insert(self) -> Optional[pulumi.Input[int]]:
"""
Parts to throw insert: Threshold value of active data parts in a table, on exceeding which ClickHouse throws the 'Too many parts ...' exception.
"""
return pulumi.get(self, "parts_to_throw_insert")
@parts_to_throw_insert.setter
def parts_to_throw_insert(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "parts_to_throw_insert", value)
@property
@pulumi.getter(name="replicatedDeduplicationWindow")
def replicated_deduplication_window(self) -> Optional[pulumi.Input[int]]:
"""
Replicated deduplication window: Number of recent hash blocks that ZooKeeper will store (the old ones will be deleted).
"""
return pulumi.get(self, "replicated_deduplication_window")
@replicated_deduplication_window.setter
def replicated_deduplication_window(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "replicated_deduplication_window", value)
@property
@pulumi.getter(name="replicatedDeduplicationWindowSeconds")
def replicated_deduplication_window_seconds(self) -> Optional[pulumi.Input[int]]:
"""
Replicated deduplication window seconds: Time during which ZooKeeper stores the hash blocks (the old ones wil be deleted).
"""
return pulumi.get(self, "replicated_deduplication_window_seconds")
@replicated_deduplication_window_seconds.setter
def replicated_deduplication_window_seconds(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "replicated_deduplication_window_seconds", value)
@pulumi.input_type
class MdbClickhouseClusterClickhouseConfigRabbitmqArgs:
def __init__(__self__, *,
password: Optional[pulumi.Input[str]] = None,
username: Optional[pulumi.Input[str]] = None):
"""
:param pulumi.Input[str] password: <PASSWORD>.
:param pulumi.Input[str] username: RabbitMQ username.
"""
if password is not None:
pulumi.set(__self__, "password", password)
if username is not None:
pulumi.set(__self__, "username", username)
@property
@pulumi.getter
def password(self) -> Optional[pulumi.Input[str]]:
"""
RabbitMQ user password.
"""
return pulumi.get(self, "password")
@password.setter
def password(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "password", value)
@property
@pulumi.getter
def username(self) -> Optional[pulumi.Input[str]]:
"""
RabbitMQ username.
"""
return pulumi.get(self, "username")
@username.setter
def username(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "username", value)
@pulumi.input_type
class MdbClickhouseClusterClickhouseResourcesArgs:
def __init__(__self__, *,
disk_size: pulumi.Input[int],
disk_type_id: pulumi.Input[str],
resource_preset_id: pulumi.Input[str]):
"""
:param pulumi.Input[int] disk_size: Volume of the storage available to a ZooKeeper host, in gigabytes.
:param pulumi.Input[str] disk_type_id: Type of the storage of ZooKeeper hosts.
For more information see [the official documentation](https://cloud.yandex.com/docs/managed-clickhouse/concepts/storage).
"""
pulumi.set(__self__, "disk_size", disk_size)
pulumi.set(__self__, "disk_type_id", disk_type_id)
pulumi.set(__self__, "resource_preset_id", resource_preset_id)
@property
@pulumi.getter(name="diskSize")
def disk_size(self) -> pulumi.Input[int]:
"""
Volume of the storage available to a ZooKeeper host, in gigabytes.
"""
return pulumi.get(self, "disk_size")
@disk_size.setter
def disk_size(self, value: pulumi.Input[int]):
pulumi.set(self, "disk_size", value)
@property
@pulumi.getter(name="diskTypeId")
def disk_type_id(self) -> pulumi.Input[str]:
"""
Type of the storage of ZooKeeper hosts.
For more information see [the official documentation](https://cloud.yandex.com/docs/managed-clickhouse/concepts/storage).
"""
return pulumi.get(self, "disk_type_id")
@disk_type_id.setter
def disk_type_id(self, value: pulumi.Input[str]):
pulumi.set(self, "disk_type_id", value)
@property
@pulumi.getter(name="resourcePresetId")
def resource_preset_id(self) -> pulumi.Input[str]:
return pulumi.get(self, "resource_preset_id")
@resource_preset_id.setter
def resource_preset_id(self, value: pulumi.Input[str]):
pulumi.set(self, "resource_preset_id", value)
@pulumi.input_type
class MdbClickhouseClusterCloudStorageArgs:
def __init__(__self__, *,
enabled: pulumi.Input[bool]):
"""
:param pulumi.Input[bool] enabled: Whether to use Yandex Object Storage for storing ClickHouse data. Can be either `true` or `false`.
"""
pulumi.set(__self__, "enabled", enabled)
@property
@pulumi.getter
def enabled(self) -> pulumi.Input[bool]:
"""
Whether to use Yandex Object Storage for storing ClickHouse data. Can be either `true` or `false`.
"""
return pulumi.get(self, "enabled")
@enabled.setter
def enabled(self, value: pulumi.Input[bool]):
pulumi.set(self, "enabled", value)
@pulumi.input_type
class MdbClickhouseClusterDatabaseArgs:
def __init__(__self__, *,
name: pulumi.Input[str]):
"""
:param pulumi.Input[str] name: Graphite rollup configuration name.
"""
pulumi.set(__self__, "name", name)
@property
@pulumi.getter
def name(self) -> pulumi.Input[str]:
"""
Graphite rollup configuration name.
"""
return pulumi.get(self, "name")
@name.setter
def name(self, value: pulumi.Input[str]):
pulumi.set(self, "name", value)
@pulumi.input_type
class MdbClickhouseClusterFormatSchemaArgs:
def __init__(__self__, *,
name: pulumi.Input[str],
type: pulumi.Input[str],
uri: pulumi.Input[str]):
"""
:param pulumi.Input[str] name: Graphite rollup configuration name.
:param pulumi.Input[str] type: Type of maintenance window. Can be either `ANYTIME` or `WEEKLY`. A day and hour of window need to be specified with weekly window.
:param pulumi.Input[str] uri: Model file URL. You can only use models stored in Yandex Object Storage.
"""
pulumi.set(__self__, "name", name)
pulumi.set(__self__, "type", type)
pulumi.set(__self__, "uri", uri)
@property
@pulumi.getter
def name(self) -> pulumi.Input[str]:
"""
Graphite rollup configuration name.
"""
return pulumi.get(self, "name")
@name.setter
def name(self, value: pulumi.Input[str]):
pulumi.set(self, "name", value)
@property
@pulumi.getter
def type(self) -> pulumi.Input[str]:
"""
Type of maintenance window. Can be either `ANYTIME` or `WEEKLY`. A day and hour of window need to be specified with weekly window.
"""
return pulumi.get(self, "type")
@type.setter
def type(self, value: pulumi.Input[str]):
pulumi.set(self, "type", value)
@property
@pulumi.getter
def uri(self) -> pulumi.Input[str]:
"""
Model file URL. You can only use models stored in Yandex Object Storage.
"""
return pulumi.get(self, "uri")
@uri.setter
def uri(self, value: pulumi.Input[str]):
pulumi.set(self, "uri", value)
@pulumi.input_type
class MdbClickhouseClusterHostArgs:
def __init__(__self__, *,
type: pulumi.Input[str],
zone: pulumi.Input[str],
assign_public_ip: Optional[pulumi.Input[bool]] = None,
fqdn: Optional[pulumi.Input[str]] = None,
shard_name: Optional[pulumi.Input[str]] = None,
subnet_id: Optional[pulumi.Input[str]] = None):
"""
:param pulumi.Input[str] type: Type of maintenance window. Can be either `ANYTIME` or `WEEKLY`. A day and hour of window need to be specified with weekly window.
:param pulumi.Input[str] zone: The availability zone where the ClickHouse host will be created.
For more information see [the official documentation](https://cloud.yandex.com/docs/overview/concepts/geo-scope).
:param pulumi.Input[bool] assign_public_ip: Sets whether the host should get a public IP address on creation. Can be either `true` or `false`.
:param pulumi.Input[str] fqdn: The fully qualified domain name of the host.
:param pulumi.Input[str] shard_name: The name of the shard to which the host belongs.
:param pulumi.Input[str] subnet_id: The ID of the subnet, to which the host belongs. The subnet must be a part of the network to which the cluster belongs.
"""
pulumi.set(__self__, "type", type)
pulumi.set(__self__, "zone", zone)
if assign_public_ip is not None:
pulumi.set(__self__, "assign_public_ip", assign_public_ip)
if fqdn is not None:
pulumi.set(__self__, "fqdn", fqdn)
if shard_name is not None:
pulumi.set(__self__, "shard_name", shard_name)
if subnet_id is not None:
pulumi.set(__self__, "subnet_id", subnet_id)
@property
@pulumi.getter
def type(self) -> pulumi.Input[str]:
"""
Type of maintenance window. Can be either `ANYTIME` or `WEEKLY`. A day and hour of window need to be specified with weekly window.
"""
return pulumi.get(self, "type")
@type.setter
def type(self, value: pulumi.Input[str]):
pulumi.set(self, "type", value)
@property
@pulumi.getter
def zone(self) -> pulumi.Input[str]:
"""
The availability zone where the ClickHouse host will be created.
For more information see [the official documentation](https://cloud.yandex.com/docs/overview/concepts/geo-scope).
"""
return pulumi.get(self, "zone")
@zone.setter
def zone(self, value: pulumi.Input[str]):
pulumi.set(self, "zone", value)
@property
@pulumi.getter(name="assignPublicIp")
def assign_public_ip(self) -> Optional[pulumi.Input[bool]]:
"""
Sets whether the host should get a public IP address on creation. Can be either `true` or `false`.
"""
return pulumi.get(self, "assign_public_ip")
@assign_public_ip.setter
def assign_public_ip(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "assign_public_ip", value)
@property
@pulumi.getter
def fqdn(self) -> Optional[pulumi.Input[str]]:
"""
The fully qualified domain name of the host.
"""
return pulumi.get(self, "fqdn")
@fqdn.setter
def fqdn(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "fqdn", value)
@property
@pulumi.getter(name="shardName")
def shard_name(self) -> Optional[pulumi.Input[str]]:
"""
The name of the shard to which the host belongs.
"""
return pulumi.get(self, "shard_name")
@shard_name.setter
def shard_name(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "shard_name", value)
@property
@pulumi.getter(name="subnetId")
def subnet_id(self) -> Optional[pulumi.Input[str]]:
"""
The ID of the subnet, to which the host belongs. The subnet must be a part of the network to which the cluster belongs.
"""
return pulumi.get(self, "subnet_id")
@subnet_id.setter
def subnet_id(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "subnet_id", value)
@pulumi.input_type
class MdbClickhouseClusterMaintenanceWindowArgs:
def __init__(__self__, *,
type: pulumi.Input[str],
day: Optional[pulumi.Input[str]] = None,
hour: Optional[pulumi.Input[int]] = None):
"""
:param pulumi.Input[str] type: Type of maintenance window. Can be either `ANYTIME` or `WEEKLY`. A day and hour of window need to be specified with weekly window.
:param pulumi.Input[str] day: Day of week for maintenance window if window type is weekly. Possible values: `MON`, `TUE`, `WED`, `THU`, `FRI`, `SAT`, `SUN`.
:param pulumi.Input[int] hour: Hour of day in UTC time zone (1-24) for maintenance window if window type is weekly.
"""
pulumi.set(__self__, "type", type)
if day is not None:
pulumi.set(__self__, "day", day)
if hour is not None:
pulumi.set(__self__, "hour", hour)
@property
@pulumi.getter
def type(self) -> pulumi.Input[str]:
"""
Type of maintenance window. Can be either `ANYTIME` or `WEEKLY`. A day and hour of window need to be specified with weekly window.
"""
return pulumi.get(self, "type")
@type.setter
def type(self, value: pulumi.Input[str]):
pulumi.set(self, "type", value)
@property
@pulumi.getter
def day(self) -> Optional[pulumi.Input[str]]:
"""
Day of week for maintenance window if window type is weekly. Possible values: `MON`, `TUE`, `WED`, `THU`, `FRI`, `SAT`, `SUN`.
"""
return pulumi.get(self, "day")
@day.setter
def day(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "day", value)
@property
@pulumi.getter
def hour(self) -> Optional[pulumi.Input[int]]:
"""
Hour of day in UTC time zone (1-24) for maintenance window if window type is weekly.
"""
return pulumi.get(self, "hour")
@hour.setter
def hour(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "hour", value)
@pulumi.input_type
class MdbClickhouseClusterMlModelArgs:
def __init__(__self__, *,
name: pulumi.Input[str],
type: pulumi.Input[str],
uri: pulumi.Input[str]):
"""
:param pulumi.Input[str] name: Graphite rollup configuration name.
:param pulumi.Input[str] type: Type of maintenance window. Can be either `ANYTIME` or `WEEKLY`. A day and hour of window need to be specified with weekly window.
:param pulumi.Input[str] uri: Model file URL. You can only use models stored in Yandex Object Storage.
"""
pulumi.set(__self__, "name", name)
pulumi.set(__self__, "type", type)
pulumi.set(__self__, "uri", uri)
@property
@pulumi.getter
def name(self) -> pulumi.Input[str]:
"""
Graphite rollup configuration name.
"""
return pulumi.get(self, "name")
@name.setter
def name(self, value: pulumi.Input[str]):
pulumi.set(self, "name", value)
@property
@pulumi.getter
def type(self) -> pulumi.Input[str]:
"""
Type of maintenance window. Can be either `ANYTIME` or `WEEKLY`. A day and hour of window need to be specified with weekly window.
"""
return pulumi.get(self, "type")
@type.setter
def type(self, value: pulumi.Input[str]):
pulumi.set(self, "type", value)
@property
@pulumi.getter
def uri(self) -> pulumi.Input[str]:
"""
Model file URL. You can only use models stored in Yandex Object Storage.
"""
return pulumi.get(self, "uri")
@uri.setter
def uri(self, value: pulumi.Input[str]):
pulumi.set(self, "uri", value)
@pulumi.input_type
class MdbClickhouseClusterShardGroupArgs:
def __init__(__self__, *,
name: pulumi.Input[str],
shard_names: pulumi.Input[Sequence[pulumi.Input[str]]],
description: Optional[pulumi.Input[str]] = None):
"""
:param pulumi.Input[str] name: Graphite rollup configuration name.
:param pulumi.Input[Sequence[pulumi.Input[str]]] shard_names: List of shards names that belong to the shard group.
:param pulumi.Input[str] description: Description of the shard group.
"""
pulumi.set(__self__, "name", name)
pulumi.set(__self__, "shard_names", shard_names)
if description is not None:
pulumi.set(__self__, "description", description)
@property
@pulumi.getter
def name(self) -> pulumi.Input[str]:
"""
Graphite rollup configuration name.
"""
return pulumi.get(self, "name")
@name.setter
def name(self, value: pulumi.Input[str]):
pulumi.set(self, "name", value)
@property
@pulumi.getter(name="shardNames")
def shard_names(self) -> pulumi.Input[Sequence[pulumi.Input[str]]]:
"""
List of shards names that belong to the shard group.
"""
return pulumi.get(self, "shard_names")
@shard_names.setter
def shard_names(self, value: pulumi.Input[Sequence[pulumi.Input[str]]]):
pulumi.set(self, "shard_names", value)
@property
@pulumi.getter
def description(self) -> Optional[pulumi.Input[str]]:
"""
Description of the shard group.
"""
return pulumi.get(self, "description")
@description.setter
def description(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "description", value)
@pulumi.input_type
class MdbClickhouseClusterUserArgs:
def __init__(__self__, *,
name: pulumi.Input[str],
password: pulumi.Input[str],
permissions: Optional[pulumi.Input[Sequence[pulumi.Input['MdbClickhouseClusterUserPermissionArgs']]]] = None,
quotas: Optional[pulumi.Input[Sequence[pulumi.Input['MdbClickhouseClusterUserQuotaArgs']]]] = None,
settings: Optional[pulumi.Input['MdbClickhouseClusterUserSettingsArgs']] = None):
"""
:param pulumi.Input[str] name: Graphite rollup configuration name.
:param pulumi.Input[str] password: <PASSWORD>.
:param pulumi.Input[Sequence[pulumi.Input['MdbClickhouseClusterUserPermissionArgs']]] permissions: Set of permissions granted to the user. The structure is documented below.
:param pulumi.Input[Sequence[pulumi.Input['MdbClickhouseClusterUserQuotaArgs']]] quotas: Set of user quotas. The structure is documented below.
:param pulumi.Input['MdbClickhouseClusterUserSettingsArgs'] settings: Kafka connection settngs sanem as `kafka` block.
"""
pulumi.set(__self__, "name", name)
pulumi.set(__self__, "password", password)
if permissions is not None:
pulumi.set(__self__, "permissions", permissions)
if quotas is not None:
pulumi.set(__self__, "quotas", quotas)
if settings is not None:
pulumi.set(__self__, "settings", settings)
@property
@pulumi.getter
def name(self) -> pulumi.Input[str]:
"""
Graphite rollup configuration name.
"""
return pulumi.get(self, "name")
@name.setter
def name(self, value: pulumi.Input[str]):
pulumi.set(self, "name", value)
@property
@pulumi.getter
def password(self) -> pulumi.Input[str]:
"""
RabbitMQ user password.
"""
return pulumi.get(self, "password")
@password.setter
def password(self, value: pulumi.Input[str]):
pulumi.set(self, "password", value)
@property
@pulumi.getter
def permissions(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['MdbClickhouseClusterUserPermissionArgs']]]]:
"""
Set of permissions granted to the user. The structure is documented below.
"""
return pulumi.get(self, "permissions")
@permissions.setter
def permissions(self, value: Optional[pulumi.Input[Sequence[pulumi.Input['MdbClickhouseClusterUserPermissionArgs']]]]):
pulumi.set(self, "permissions", value)
@property
@pulumi.getter
def quotas(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['MdbClickhouseClusterUserQuotaArgs']]]]:
"""
Set of user quotas. The structure is documented below.
"""
return pulumi.get(self, "quotas")
@quotas.setter
def quotas(self, value: Optional[pulumi.Input[Sequence[pulumi.Input['MdbClickhouseClusterUserQuotaArgs']]]]):
pulumi.set(self, "quotas", value)
@property
@pulumi.getter
def settings(self) -> Optional[pulumi.Input['MdbClickhouseClusterUserSettingsArgs']]:
"""
Kafka connection settngs sanem as `kafka` block.
"""
return pulumi.get(self, "settings")
@settings.setter
def settings(self, value: Optional[pulumi.Input['MdbClickhouseClusterUserSettingsArgs']]):
pulumi.set(self, "settings", value)
@pulumi.input_type
class MdbClickhouseClusterUserPermissionArgs:
def __init__(__self__, *,
database_name: pulumi.Input[str]):
"""
:param pulumi.Input[str] database_name: The name of the database that the permission grants access to.
"""
pulumi.set(__self__, "database_name", database_name)
@property
@pulumi.getter(name="databaseName")
def database_name(self) -> pulumi.Input[str]:
"""
The name of the database that the permission grants access to.
"""
return pulumi.get(self, "database_name")
@database_name.setter
def database_name(self, value: pulumi.Input[str]):
pulumi.set(self, "database_name", value)
@pulumi.input_type
class MdbClickhouseClusterUserQuotaArgs:
def __init__(__self__, *,
interval_duration: pulumi.Input[int],
errors: Optional[pulumi.Input[int]] = None,
execution_time: Optional[pulumi.Input[int]] = None,
queries: Optional[pulumi.Input[int]] = None,
read_rows: Optional[pulumi.Input[int]] = None,
result_rows: Optional[pulumi.Input[int]] = None):
"""
:param pulumi.Input[int] interval_duration: Duration of interval for quota in milliseconds.
:param pulumi.Input[int] errors: The number of queries that threw exception.
:param pulumi.Input[int] execution_time: The total query execution time, in milliseconds (wall time).
:param pulumi.Input[int] queries: The total number of queries.
:param pulumi.Input[int] read_rows: The total number of source rows read from tables for running the query, on all remote servers.
:param pulumi.Input[int] result_rows: The total number of rows given as the result.
"""
pulumi.set(__self__, "interval_duration", interval_duration)
if errors is not None:
pulumi.set(__self__, "errors", errors)
if execution_time is not None:
pulumi.set(__self__, "execution_time", execution_time)
if queries is not None:
pulumi.set(__self__, "queries", queries)
if read_rows is not None:
pulumi.set(__self__, "read_rows", read_rows)
if result_rows is not None:
pulumi.set(__self__, "result_rows", result_rows)
@property
@pulumi.getter(name="intervalDuration")
def interval_duration(self) -> pulumi.Input[int]:
"""
Duration of interval for quota in milliseconds.
"""
return pulumi.get(self, "interval_duration")
@interval_duration.setter
def interval_duration(self, value: pulumi.Input[int]):
pulumi.set(self, "interval_duration", value)
@property
@pulumi.getter
def errors(self) -> Optional[pulumi.Input[int]]:
"""
The number of queries that threw exception.
"""
return pulumi.get(self, "errors")
@errors.setter
def errors(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "errors", value)
@property
@pulumi.getter(name="executionTime")
def execution_time(self) -> Optional[pulumi.Input[int]]:
"""
The total query execution time, in milliseconds (wall time).
"""
return pulumi.get(self, "execution_time")
@execution_time.setter
def execution_time(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "execution_time", value)
@property
@pulumi.getter
def queries(self) -> Optional[pulumi.Input[int]]:
"""
The total number of queries.
"""
return pulumi.get(self, "queries")
@queries.setter
def queries(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "queries", value)
@property
@pulumi.getter(name="readRows")
def read_rows(self) -> Optional[pulumi.Input[int]]:
"""
The total number of source rows read from tables for running the query, on all remote servers.
"""
return pulumi.get(self, "read_rows")
@read_rows.setter
def read_rows(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "read_rows", value)
@property
@pulumi.getter(name="resultRows")
def result_rows(self) -> Optional[pulumi.Input[int]]:
"""
The total number of rows given as the result.
"""
return pulumi.get(self, "result_rows")
@result_rows.setter
def result_rows(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "result_rows", value)
@pulumi.input_type
class MdbClickhouseClusterUserSettingsArgs:
def __init__(__self__, *,
add_http_cors_header: Optional[pulumi.Input[bool]] = None,
allow_ddl: Optional[pulumi.Input[bool]] = None,
compile: Optional[pulumi.Input[bool]] = None,
compile_expressions: Optional[pulumi.Input[bool]] = None,
connect_timeout: Optional[pulumi.Input[int]] = None,
count_distinct_implementation: Optional[pulumi.Input[str]] = None,
distinct_overflow_mode: Optional[pulumi.Input[str]] = None,
distributed_aggregation_memory_efficient: Optional[pulumi.Input[bool]] = None,
distributed_ddl_task_timeout: Optional[pulumi.Input[int]] = None,
distributed_product_mode: Optional[pulumi.Input[str]] = None,
empty_result_for_aggregation_by_empty_set: Optional[pulumi.Input[bool]] = None,
enable_http_compression: Optional[pulumi.Input[bool]] = None,
fallback_to_stale_replicas_for_distributed_queries: Optional[pulumi.Input[bool]] = None,
force_index_by_date: Optional[pulumi.Input[bool]] = None,
force_primary_key: Optional[pulumi.Input[bool]] = None,
group_by_overflow_mode: Optional[pulumi.Input[str]] = None,
group_by_two_level_threshold: Optional[pulumi.Input[int]] = None,
group_by_two_level_threshold_bytes: Optional[pulumi.Input[int]] = None,
http_connection_timeout: Optional[pulumi.Input[int]] = None,
http_headers_progress_interval: Optional[pulumi.Input[int]] = None,
http_receive_timeout: Optional[pulumi.Input[int]] = None,
http_send_timeout: Optional[pulumi.Input[int]] = None,
input_format_defaults_for_omitted_fields: Optional[pulumi.Input[bool]] = None,
input_format_values_interpret_expressions: Optional[pulumi.Input[bool]] = None,
insert_quorum: Optional[pulumi.Input[int]] = None,
insert_quorum_timeout: Optional[pulumi.Input[int]] = None,
join_overflow_mode: Optional[pulumi.Input[str]] = None,
join_use_nulls: Optional[pulumi.Input[bool]] = None,
joined_subquery_requires_alias: Optional[pulumi.Input[bool]] = None,
low_cardinality_allow_in_native_format: Optional[pulumi.Input[bool]] = None,
max_ast_depth: Optional[pulumi.Input[int]] = None,
max_ast_elements: Optional[pulumi.Input[int]] = None,
max_block_size: Optional[pulumi.Input[int]] = None,
max_bytes_before_external_group_by: Optional[pulumi.Input[int]] = None,
max_bytes_before_external_sort: Optional[pulumi.Input[int]] = None,
max_bytes_in_distinct: Optional[pulumi.Input[int]] = None,
max_bytes_in_join: Optional[pulumi.Input[int]] = None,
max_bytes_in_set: Optional[pulumi.Input[int]] = None,
max_bytes_to_read: Optional[pulumi.Input[int]] = None,
max_bytes_to_sort: Optional[pulumi.Input[int]] = None,
max_bytes_to_transfer: Optional[pulumi.Input[int]] = None,
max_columns_to_read: Optional[pulumi.Input[int]] = None,
max_execution_time: Optional[pulumi.Input[int]] = None,
max_expanded_ast_elements: Optional[pulumi.Input[int]] = None,
max_insert_block_size: Optional[pulumi.Input[int]] = None,
max_memory_usage: Optional[pulumi.Input[int]] = None,
max_memory_usage_for_user: Optional[pulumi.Input[int]] = None,
max_network_bandwidth: Optional[pulumi.Input[int]] = None,
max_network_bandwidth_for_user: Optional[pulumi.Input[int]] = None,
max_query_size: Optional[pulumi.Input[int]] = None,
max_replica_delay_for_distributed_queries: Optional[pulumi.Input[int]] = None,
max_result_bytes: Optional[pulumi.Input[int]] = None,
max_result_rows: Optional[pulumi.Input[int]] = None,
max_rows_in_distinct: Optional[pulumi.Input[int]] = None,
max_rows_in_join: Optional[pulumi.Input[int]] = None,
max_rows_in_set: Optional[pulumi.Input[int]] = None,
max_rows_to_group_by: Optional[pulumi.Input[int]] = None,
max_rows_to_read: Optional[pulumi.Input[int]] = None,
max_rows_to_sort: Optional[pulumi.Input[int]] = None,
max_rows_to_transfer: Optional[pulumi.Input[int]] = None,
max_temporary_columns: Optional[pulumi.Input[int]] = None,
max_temporary_non_const_columns: Optional[pulumi.Input[int]] = None,
max_threads: Optional[pulumi.Input[int]] = None,
merge_tree_max_bytes_to_use_cache: Optional[pulumi.Input[int]] = None,
merge_tree_max_rows_to_use_cache: Optional[pulumi.Input[int]] = None,
merge_tree_min_bytes_for_concurrent_read: Optional[pulumi.Input[int]] = None,
merge_tree_min_rows_for_concurrent_read: Optional[pulumi.Input[int]] = None,
min_bytes_to_use_direct_io: Optional[pulumi.Input[int]] = None,
min_count_to_compile: Optional[pulumi.Input[int]] = None,
min_count_to_compile_expression: Optional[pulumi.Input[int]] = None,
min_execution_speed: Optional[pulumi.Input[int]] = None,
min_execution_speed_bytes: Optional[pulumi.Input[int]] = None,
min_insert_block_size_bytes: Optional[pulumi.Input[int]] = None,
min_insert_block_size_rows: Optional[pulumi.Input[int]] = None,
output_format_json_quote64bit_integers: Optional[pulumi.Input[bool]] = None,
output_format_json_quote_denormals: Optional[pulumi.Input[bool]] = None,
priority: Optional[pulumi.Input[int]] = None,
quota_mode: Optional[pulumi.Input[str]] = None,
read_overflow_mode: Optional[pulumi.Input[str]] = None,
readonly: Optional[pulumi.Input[int]] = None,
receive_timeout: Optional[pulumi.Input[int]] = None,
replication_alter_partitions_sync: Optional[pulumi.Input[int]] = None,
result_overflow_mode: Optional[pulumi.Input[str]] = None,
select_sequential_consistency: Optional[pulumi.Input[bool]] = None,
send_progress_in_http_headers: Optional[pulumi.Input[bool]] = None,
send_timeout: Optional[pulumi.Input[int]] = None,
set_overflow_mode: Optional[pulumi.Input[str]] = None,
skip_unavailable_shards: Optional[pulumi.Input[bool]] = None,
sort_overflow_mode: Optional[pulumi.Input[str]] = None,
timeout_overflow_mode: Optional[pulumi.Input[str]] = None,
transfer_overflow_mode: Optional[pulumi.Input[str]] = None,
transform_null_in: Optional[pulumi.Input[bool]] = None,
use_uncompressed_cache: Optional[pulumi.Input[bool]] = None):
"""
:param pulumi.Input[bool] add_http_cors_header: Include CORS headers in HTTP responces.
:param pulumi.Input[bool] allow_ddl: Allows or denies DDL queries.
:param pulumi.Input[bool] compile: Enable compilation of queries.
:param pulumi.Input[bool] compile_expressions: Turn on expression compilation.
:param pulumi.Input[int] connect_timeout: Connect timeout in milliseconds on the socket used for communicating with the client.
:param pulumi.Input[str] count_distinct_implementation: Specifies which of the uniq* functions should be used to perform the COUNT(DISTINCT …) construction.
:param pulumi.Input[str] distinct_overflow_mode: Sets behaviour on overflow when using DISTINCT. Possible values:
:param pulumi.Input[bool] distributed_aggregation_memory_efficient: Determine the behavior of distributed subqueries.
:param pulumi.Input[int] distributed_ddl_task_timeout: Timeout for DDL queries, in milliseconds.
:param pulumi.Input[str] distributed_product_mode: Changes the behaviour of distributed subqueries.
:param pulumi.Input[bool] empty_result_for_aggregation_by_empty_set: Allows to retunr empty result.
:param pulumi.Input[bool] enable_http_compression: Enables or disables data compression in the response to an HTTP request.
:param pulumi.Input[bool] fallback_to_stale_replicas_for_distributed_queries: Forces a query to an out-of-date replica if updated data is not available.
:param pulumi.Input[bool] force_index_by_date: Disables query execution if the index can’t be used by date.
:param pulumi.Input[bool] force_primary_key: Disables query execution if indexing by the primary key is not possible.
:param pulumi.Input[str] group_by_overflow_mode: Sets behaviour on overflow while GROUP BY operation. Possible values:
:param pulumi.Input[int] group_by_two_level_threshold: Sets the threshold of the number of keys, after that the two-level aggregation should be used.
:param pulumi.Input[int] group_by_two_level_threshold_bytes: Sets the threshold of the number of bytes, after that the two-level aggregation should be used.
:param pulumi.Input[int] http_connection_timeout: Timeout for HTTP connection in milliseconds.
:param pulumi.Input[int] http_headers_progress_interval: Sets minimal interval between notifications about request process in HTTP header X-ClickHouse-Progress.
:param pulumi.Input[int] http_receive_timeout: Timeout for HTTP connection in milliseconds.
:param pulumi.Input[int] http_send_timeout: Timeout for HTTP connection in milliseconds.
:param pulumi.Input[bool] input_format_defaults_for_omitted_fields: When performing INSERT queries, replace omitted input column values with default values of the respective columns.
:param pulumi.Input[bool] input_format_values_interpret_expressions: Enables or disables the full SQL parser if the fast stream parser can’t parse the data.
:param pulumi.Input[int] insert_quorum: Enables the quorum writes.
:param pulumi.Input[int] insert_quorum_timeout: Write to a quorum timeout in milliseconds.
:param pulumi.Input[str] join_overflow_mode: Sets behaviour on overflow in JOIN. Possible values:
:param pulumi.Input[bool] join_use_nulls: Sets the type of JOIN behaviour. When merging tables, empty cells may appear. ClickHouse fills them differently based on this setting.
:param pulumi.Input[bool] joined_subquery_requires_alias: Require aliases for subselects and table functions in FROM that more than one table is present.
:param pulumi.Input[bool] low_cardinality_allow_in_native_format: Allows or restricts using the LowCardinality data type with the Native format.
:param pulumi.Input[int] max_ast_depth: Maximum abstract syntax tree depth.
:param pulumi.Input[int] max_ast_elements: Maximum abstract syntax tree elements.
:param pulumi.Input[int] max_block_size: A recommendation for what size of the block (in a count of rows) to load from tables.
:param pulumi.Input[int] max_bytes_before_external_group_by: Limit in bytes for using memoru for GROUP BY before using swap on disk.
:param pulumi.Input[int] max_bytes_before_external_sort: This setting is equivalent of the max_bytes_before_external_group_by setting, except for it is for sort operation (ORDER BY), not aggregation.
:param pulumi.Input[int] max_bytes_in_distinct: Limits the maximum size of a hash table in bytes (uncompressed data) when using DISTINCT.
:param pulumi.Input[int] max_bytes_in_join: Limit on maximum size of the hash table for JOIN, in bytes.
:param pulumi.Input[int] max_bytes_in_set: Limit on the number of bytes in the set resulting from the execution of the IN section.
:param pulumi.Input[int] max_bytes_to_read: Limits the maximum number of bytes (uncompressed data) that can be read from a table when running a query.
:param pulumi.Input[int] max_bytes_to_sort: Limits the maximum number of bytes (uncompressed data) that can be read from a table for sorting.
:param pulumi.Input[int] max_bytes_to_transfer: Limits the maximum number of bytes (uncompressed data) that can be passed to a remote server or saved in a temporary table when using GLOBAL IN.
:param pulumi.Input[int] max_columns_to_read: Limits the maximum number of columns that can be read from a table in a single query.
:param pulumi.Input[int] max_execution_time: Limits the maximum query execution time in milliseconds.
:param pulumi.Input[int] max_expanded_ast_elements: Maximum abstract syntax tree depth after after expansion of aliases.
:param pulumi.Input[int] max_insert_block_size: The size of blocks (in a count of rows) to form for insertion into a table.
:param pulumi.Input[int] max_memory_usage: Limits the maximum memory usage (in bytes) for processing queries on a single server.
:param pulumi.Input[int] max_memory_usage_for_user: Limits the maximum memory usage (in bytes) for processing of user's queries on a single server.
:param pulumi.Input[int] max_network_bandwidth: Limits the speed of the data exchange over the network in bytes per second.
:param pulumi.Input[int] max_network_bandwidth_for_user: Limits the speed of the data exchange over the network in bytes per second.
:param pulumi.Input[int] max_query_size: The maximum part of a query that can be taken to RAM for parsing with the SQL parser.
:param pulumi.Input[int] max_replica_delay_for_distributed_queries: Disables lagging replicas for distributed queries.
:param pulumi.Input[int] max_result_bytes: Limits the number of bytes in the result.
:param pulumi.Input[int] max_result_rows: Limits the number of rows in the result.
:param pulumi.Input[int] max_rows_in_distinct: Limits the maximum number of different rows when using DISTINCT.
:param pulumi.Input[int] max_rows_in_join: Limit on maximum size of the hash table for JOIN, in rows.
:param pulumi.Input[int] max_rows_in_set: Limit on the number of rows in the set resulting from the execution of the IN section.
:param pulumi.Input[int] max_rows_to_group_by: Limits the maximum number of unique keys received from aggregation function.
:param pulumi.Input[int] max_rows_to_read: Limits the maximum number of rows that can be read from a table when running a query.
:param pulumi.Input[int] max_rows_to_sort: Limits the maximum number of rows that can be read from a table for sorting.
:param pulumi.Input[int] max_rows_to_transfer: Limits the maximum number of rows that can be passed to a remote server or saved in a temporary table when using GLOBAL IN.
:param pulumi.Input[int] max_temporary_columns: Limits the maximum number of temporary columns that must be kept in RAM at the same time when running a query, including constant columns.
:param pulumi.Input[int] max_temporary_non_const_columns: Limits the maximum number of temporary columns that must be kept in RAM at the same time when running a query, excluding constant columns.
:param pulumi.Input[int] max_threads: The maximum number of query processing threads, excluding threads for retrieving data from remote servers.
:param pulumi.Input[int] merge_tree_max_bytes_to_use_cache: If ClickHouse should read more than merge_tree_max_bytes_to_use_cache bytes in one query, it doesn’t use the cache of uncompressed blocks.
:param pulumi.Input[int] merge_tree_max_rows_to_use_cache: If ClickHouse should read more than merge_tree_max_rows_to_use_cache rows in one query, it doesn’t use the cache of uncompressed blocks.
:param pulumi.Input[int] merge_tree_min_bytes_for_concurrent_read: If the number of bytes to read from one file of a MergeTree-engine table exceeds merge_tree_min_bytes_for_concurrent_read, then ClickHouse tries to concurrently read from this file in several threads.
:param pulumi.Input[int] merge_tree_min_rows_for_concurrent_read: If the number of rows to be read from a file of a MergeTree table exceeds merge_tree_min_rows_for_concurrent_read then ClickHouse tries to perform a concurrent reading from this file on several threads.
:param pulumi.Input[int] min_bytes_to_use_direct_io: The minimum data volume required for using direct I/O access to the storage disk.
:param pulumi.Input[int] min_count_to_compile: How many times to potentially use a compiled chunk of code before running compilation.
:param pulumi.Input[int] min_count_to_compile_expression: A query waits for expression compilation process to complete prior to continuing execution.
:param pulumi.Input[int] min_execution_speed: Minimal execution speed in rows per second.
:param pulumi.Input[int] min_execution_speed_bytes: Minimal execution speed in bytes per second.
:param pulumi.Input[int] min_insert_block_size_bytes: Sets the minimum number of bytes in the block which can be inserted into a table by an INSERT query.
:param pulumi.Input[int] min_insert_block_size_rows: Sets the minimum number of rows in the block which can be inserted into a table by an INSERT query.
:param pulumi.Input[bool] output_format_json_quote64bit_integers: If the value is true, integers appear in quotes when using JSON* Int64 and UInt64 formats (for compatibility with most JavaScript implementations); otherwise, integers are output without the quotes.
:param pulumi.Input[bool] output_format_json_quote_denormals: Enables +nan, -nan, +inf, -inf outputs in JSON output format.
:param pulumi.Input[int] priority: Query priority.
:param pulumi.Input[str] quota_mode: Quota accounting mode.
:param pulumi.Input[str] read_overflow_mode: Sets behaviour on overflow while read. Possible values:
:param pulumi.Input[int] readonly: Restricts permissions for reading data, write data and change settings queries.
:param pulumi.Input[int] receive_timeout: Receive timeout in milliseconds on the socket used for communicating with the client.
:param pulumi.Input[int] replication_alter_partitions_sync: For ALTER ... ATTACH|DETACH|DROP queries, you can use the replication_alter_partitions_sync setting to set up waiting.
:param pulumi.Input[str] result_overflow_mode: Sets behaviour on overflow in result. Possible values:
:param pulumi.Input[bool] select_sequential_consistency: Enables or disables sequential consistency for SELECT queries.
:param pulumi.Input[bool] send_progress_in_http_headers: Enables or disables X-ClickHouse-Progress HTTP response headers in clickhouse-server responses.
:param pulumi.Input[int] send_timeout: Send timeout in milliseconds on the socket used for communicating with the client.
:param pulumi.Input[str] set_overflow_mode: Sets behaviour on overflow in the set resulting. Possible values:
:param pulumi.Input[bool] skip_unavailable_shards: Enables or disables silently skipping of unavailable shards.
:param pulumi.Input[str] sort_overflow_mode: Sets behaviour on overflow while sort. Possible values:
:param pulumi.Input[str] timeout_overflow_mode: Sets behaviour on overflow. Possible values:
:param pulumi.Input[str] transfer_overflow_mode: Sets behaviour on overflow. Possible values:
:param pulumi.Input[bool] transform_null_in: Enables equality of NULL values for IN operator.
:param pulumi.Input[bool] use_uncompressed_cache: Whether to use a cache of uncompressed blocks.
"""
if add_http_cors_header is not None:
pulumi.set(__self__, "add_http_cors_header", add_http_cors_header)
if allow_ddl is not None:
pulumi.set(__self__, "allow_ddl", allow_ddl)
if compile is not None:
pulumi.set(__self__, "compile", compile)
if compile_expressions is not None:
pulumi.set(__self__, "compile_expressions", compile_expressions)
if connect_timeout is not None:
pulumi.set(__self__, "connect_timeout", connect_timeout)
if count_distinct_implementation is not None:
pulumi.set(__self__, "count_distinct_implementation", count_distinct_implementation)
if distinct_overflow_mode is not None:
pulumi.set(__self__, "distinct_overflow_mode", distinct_overflow_mode)
if distributed_aggregation_memory_efficient is not None:
pulumi.set(__self__, "distributed_aggregation_memory_efficient", distributed_aggregation_memory_efficient)
if distributed_ddl_task_timeout is not None:
pulumi.set(__self__, "distributed_ddl_task_timeout", distributed_ddl_task_timeout)
if distributed_product_mode is not None:
pulumi.set(__self__, "distributed_product_mode", distributed_product_mode)
if empty_result_for_aggregation_by_empty_set is not None:
pulumi.set(__self__, "empty_result_for_aggregation_by_empty_set", empty_result_for_aggregation_by_empty_set)
if enable_http_compression is not None:
pulumi.set(__self__, "enable_http_compression", enable_http_compression)
if fallback_to_stale_replicas_for_distributed_queries is not None:
pulumi.set(__self__, "fallback_to_stale_replicas_for_distributed_queries", fallback_to_stale_replicas_for_distributed_queries)
if force_index_by_date is not None:
pulumi.set(__self__, "force_index_by_date", force_index_by_date)
if force_primary_key is not None:
pulumi.set(__self__, "force_primary_key", force_primary_key)
if group_by_overflow_mode is not None:
pulumi.set(__self__, "group_by_overflow_mode", group_by_overflow_mode)
if group_by_two_level_threshold is not None:
pulumi.set(__self__, "group_by_two_level_threshold", group_by_two_level_threshold)
if group_by_two_level_threshold_bytes is not None:
pulumi.set(__self__, "group_by_two_level_threshold_bytes", group_by_two_level_threshold_bytes)
if http_connection_timeout is not None:
pulumi.set(__self__, "http_connection_timeout", http_connection_timeout)
if http_headers_progress_interval is not None:
pulumi.set(__self__, "http_headers_progress_interval", http_headers_progress_interval)
if http_receive_timeout is not None:
pulumi.set(__self__, "http_receive_timeout", http_receive_timeout)
if http_send_timeout is not None:
pulumi.set(__self__, "http_send_timeout", http_send_timeout)
if input_format_defaults_for_omitted_fields is not None:
pulumi.set(__self__, "input_format_defaults_for_omitted_fields", input_format_defaults_for_omitted_fields)
if input_format_values_interpret_expressions is not None:
pulumi.set(__self__, "input_format_values_interpret_expressions", input_format_values_interpret_expressions)
if insert_quorum is not None:
pulumi.set(__self__, "insert_quorum", insert_quorum)
if insert_quorum_timeout is not None:
pulumi.set(__self__, "insert_quorum_timeout", insert_quorum_timeout)
if join_overflow_mode is not None:
pulumi.set(__self__, "join_overflow_mode", join_overflow_mode)
if join_use_nulls is not None:
pulumi.set(__self__, "join_use_nulls", join_use_nulls)
if joined_subquery_requires_alias is not None:
pulumi.set(__self__, "joined_subquery_requires_alias", joined_subquery_requires_alias)
if low_cardinality_allow_in_native_format is not None:
pulumi.set(__self__, "low_cardinality_allow_in_native_format", low_cardinality_allow_in_native_format)
if max_ast_depth is not None:
pulumi.set(__self__, "max_ast_depth", max_ast_depth)
if max_ast_elements is not None:
pulumi.set(__self__, "max_ast_elements", max_ast_elements)
if max_block_size is not None:
pulumi.set(__self__, "max_block_size", max_block_size)
if max_bytes_before_external_group_by is not None:
pulumi.set(__self__, "max_bytes_before_external_group_by", max_bytes_before_external_group_by)
if max_bytes_before_external_sort is not None:
pulumi.set(__self__, "max_bytes_before_external_sort", max_bytes_before_external_sort)
if max_bytes_in_distinct is not None:
pulumi.set(__self__, "max_bytes_in_distinct", max_bytes_in_distinct)
if max_bytes_in_join is not None:
pulumi.set(__self__, "max_bytes_in_join", max_bytes_in_join)
if max_bytes_in_set is not None:
pulumi.set(__self__, "max_bytes_in_set", max_bytes_in_set)
if max_bytes_to_read is not None:
pulumi.set(__self__, "max_bytes_to_read", max_bytes_to_read)
if max_bytes_to_sort is not None:
pulumi.set(__self__, "max_bytes_to_sort", max_bytes_to_sort)
if max_bytes_to_transfer is not None:
pulumi.set(__self__, "max_bytes_to_transfer", max_bytes_to_transfer)
if max_columns_to_read is not None:
pulumi.set(__self__, "max_columns_to_read", max_columns_to_read)
if max_execution_time is not None:
pulumi.set(__self__, "max_execution_time", max_execution_time)
if max_expanded_ast_elements is not None:
pulumi.set(__self__, "max_expanded_ast_elements", max_expanded_ast_elements)
if max_insert_block_size is not None:
pulumi.set(__self__, "max_insert_block_size", max_insert_block_size)
if max_memory_usage is not None:
pulumi.set(__self__, "max_memory_usage", max_memory_usage)
if max_memory_usage_for_user is not None:
pulumi.set(__self__, "max_memory_usage_for_user", max_memory_usage_for_user)
if max_network_bandwidth is not None:
pulumi.set(__self__, "max_network_bandwidth", max_network_bandwidth)
if max_network_bandwidth_for_user is not None:
pulumi.set(__self__, "max_network_bandwidth_for_user", max_network_bandwidth_for_user)
if max_query_size is not None:
pulumi.set(__self__, "max_query_size", max_query_size)
if max_replica_delay_for_distributed_queries is not None:
pulumi.set(__self__, "max_replica_delay_for_distributed_queries", max_replica_delay_for_distributed_queries)
if max_result_bytes is not None:
pulumi.set(__self__, "max_result_bytes", max_result_bytes)
if max_result_rows is not None:
pulumi.set(__self__, "max_result_rows", max_result_rows)
if max_rows_in_distinct is not None:
pulumi.set(__self__, "max_rows_in_distinct", max_rows_in_distinct)
if max_rows_in_join is not None:
pulumi.set(__self__, "max_rows_in_join", max_rows_in_join)
if max_rows_in_set is not None:
pulumi.set(__self__, "max_rows_in_set", max_rows_in_set)
if max_rows_to_group_by is not None:
pulumi.set(__self__, "max_rows_to_group_by", max_rows_to_group_by)
if max_rows_to_read is not None:
pulumi.set(__self__, "max_rows_to_read", max_rows_to_read)
if max_rows_to_sort is not None:
pulumi.set(__self__, "max_rows_to_sort", max_rows_to_sort)
if max_rows_to_transfer is not None:
pulumi.set(__self__, "max_rows_to_transfer", max_rows_to_transfer)
if max_temporary_columns is not None:
pulumi.set(__self__, "max_temporary_columns", max_temporary_columns)
if max_temporary_non_const_columns is not None:
pulumi.set(__self__, "max_temporary_non_const_columns", max_temporary_non_const_columns)
if max_threads is not None:
pulumi.set(__self__, "max_threads", max_threads)
if merge_tree_max_bytes_to_use_cache is not None:
pulumi.set(__self__, "merge_tree_max_bytes_to_use_cache", merge_tree_max_bytes_to_use_cache)
if merge_tree_max_rows_to_use_cache is not None:
pulumi.set(__self__, "merge_tree_max_rows_to_use_cache", merge_tree_max_rows_to_use_cache)
if merge_tree_min_bytes_for_concurrent_read is not None:
pulumi.set(__self__, "merge_tree_min_bytes_for_concurrent_read", merge_tree_min_bytes_for_concurrent_read)
if merge_tree_min_rows_for_concurrent_read is not None:
pulumi.set(__self__, "merge_tree_min_rows_for_concurrent_read", merge_tree_min_rows_for_concurrent_read)
if min_bytes_to_use_direct_io is not None:
pulumi.set(__self__, "min_bytes_to_use_direct_io", min_bytes_to_use_direct_io)
if min_count_to_compile is not None:
pulumi.set(__self__, "min_count_to_compile", min_count_to_compile)
if min_count_to_compile_expression is not None:
pulumi.set(__self__, "min_count_to_compile_expression", min_count_to_compile_expression)
if min_execution_speed is not None:
pulumi.set(__self__, "min_execution_speed", min_execution_speed)
if min_execution_speed_bytes is not None:
pulumi.set(__self__, "min_execution_speed_bytes", min_execution_speed_bytes)
if min_insert_block_size_bytes is not None:
pulumi.set(__self__, "min_insert_block_size_bytes", min_insert_block_size_bytes)
if min_insert_block_size_rows is not None:
pulumi.set(__self__, "min_insert_block_size_rows", min_insert_block_size_rows)
if output_format_json_quote64bit_integers is not None:
pulumi.set(__self__, "output_format_json_quote64bit_integers", output_format_json_quote64bit_integers)
if output_format_json_quote_denormals is not None:
pulumi.set(__self__, "output_format_json_quote_denormals", output_format_json_quote_denormals)
if priority is not None:
pulumi.set(__self__, "priority", priority)
if quota_mode is not None:
pulumi.set(__self__, "quota_mode", quota_mode)
if read_overflow_mode is not None:
pulumi.set(__self__, "read_overflow_mode", read_overflow_mode)
if readonly is not None:
pulumi.set(__self__, "readonly", readonly)
if receive_timeout is not None:
pulumi.set(__self__, "receive_timeout", receive_timeout)
if replication_alter_partitions_sync is not None:
pulumi.set(__self__, "replication_alter_partitions_sync", replication_alter_partitions_sync)
if result_overflow_mode is not None:
pulumi.set(__self__, "result_overflow_mode", result_overflow_mode)
if select_sequential_consistency is not None:
pulumi.set(__self__, "select_sequential_consistency", select_sequential_consistency)
if send_progress_in_http_headers is not None:
pulumi.set(__self__, "send_progress_in_http_headers", send_progress_in_http_headers)
if send_timeout is not None:
pulumi.set(__self__, "send_timeout", send_timeout)
if set_overflow_mode is not None:
pulumi.set(__self__, "set_overflow_mode", set_overflow_mode)
if skip_unavailable_shards is not None:
pulumi.set(__self__, "skip_unavailable_shards", skip_unavailable_shards)
if sort_overflow_mode is not None:
pulumi.set(__self__, "sort_overflow_mode", sort_overflow_mode)
if timeout_overflow_mode is not None:
pulumi.set(__self__, "timeout_overflow_mode", timeout_overflow_mode)
if transfer_overflow_mode is not None:
pulumi.set(__self__, "transfer_overflow_mode", transfer_overflow_mode)
if transform_null_in is not None:
pulumi.set(__self__, "transform_null_in", transform_null_in)
if use_uncompressed_cache is not None:
pulumi.set(__self__, "use_uncompressed_cache", use_uncompressed_cache)
@property
@pulumi.getter(name="addHttpCorsHeader")
def add_http_cors_header(self) -> Optional[pulumi.Input[bool]]:
"""
Include CORS headers in HTTP responces.
"""
return pulumi.get(self, "add_http_cors_header")
@add_http_cors_header.setter
def add_http_cors_header(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "add_http_cors_header", value)
@property
@pulumi.getter(name="allowDdl")
def allow_ddl(self) -> Optional[pulumi.Input[bool]]:
"""
Allows or denies DDL queries.
"""
return pulumi.get(self, "allow_ddl")
@allow_ddl.setter
def allow_ddl(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "allow_ddl", value)
@property
@pulumi.getter
def compile(self) -> Optional[pulumi.Input[bool]]:
"""
Enable compilation of queries.
"""
return pulumi.get(self, "compile")
@compile.setter
def compile(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "compile", value)
@property
@pulumi.getter(name="compileExpressions")
def compile_expressions(self) -> Optional[pulumi.Input[bool]]:
"""
Turn on expression compilation.
"""
return pulumi.get(self, "compile_expressions")
@compile_expressions.setter
def compile_expressions(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "compile_expressions", value)
@property
@pulumi.getter(name="connectTimeout")
def connect_timeout(self) -> Optional[pulumi.Input[int]]:
"""
Connect timeout in milliseconds on the socket used for communicating with the client.
"""
return pulumi.get(self, "connect_timeout")
@connect_timeout.setter
def connect_timeout(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "connect_timeout", value)
@property
@pulumi.getter(name="countDistinctImplementation")
def count_distinct_implementation(self) -> Optional[pulumi.Input[str]]:
"""
Specifies which of the uniq* functions should be used to perform the COUNT(DISTINCT …) construction.
"""
return pulumi.get(self, "count_distinct_implementation")
@count_distinct_implementation.setter
def count_distinct_implementation(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "count_distinct_implementation", value)
@property
@pulumi.getter(name="distinctOverflowMode")
def distinct_overflow_mode(self) -> Optional[pulumi.Input[str]]:
"""
Sets behaviour on overflow when using DISTINCT. Possible values:
"""
return pulumi.get(self, "distinct_overflow_mode")
@distinct_overflow_mode.setter
def distinct_overflow_mode(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "distinct_overflow_mode", value)
@property
@pulumi.getter(name="distributedAggregationMemoryEfficient")
def distributed_aggregation_memory_efficient(self) -> Optional[pulumi.Input[bool]]:
"""
Determine the behavior of distributed subqueries.
"""
return pulumi.get(self, "distributed_aggregation_memory_efficient")
@distributed_aggregation_memory_efficient.setter
def distributed_aggregation_memory_efficient(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "distributed_aggregation_memory_efficient", value)
@property
@pulumi.getter(name="distributedDdlTaskTimeout")
def distributed_ddl_task_timeout(self) -> Optional[pulumi.Input[int]]:
"""
Timeout for DDL queries, in milliseconds.
"""
return pulumi.get(self, "distributed_ddl_task_timeout")
@distributed_ddl_task_timeout.setter
def distributed_ddl_task_timeout(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "distributed_ddl_task_timeout", value)
@property
@pulumi.getter(name="distributedProductMode")
def distributed_product_mode(self) -> Optional[pulumi.Input[str]]:
"""
Changes the behaviour of distributed subqueries.
"""
return pulumi.get(self, "distributed_product_mode")
@distributed_product_mode.setter
def distributed_product_mode(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "distributed_product_mode", value)
@property
@pulumi.getter(name="emptyResultForAggregationByEmptySet")
def empty_result_for_aggregation_by_empty_set(self) -> Optional[pulumi.Input[bool]]:
"""
Allows to retunr empty result.
"""
return pulumi.get(self, "empty_result_for_aggregation_by_empty_set")
@empty_result_for_aggregation_by_empty_set.setter
def empty_result_for_aggregation_by_empty_set(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "empty_result_for_aggregation_by_empty_set", value)
@property
@pulumi.getter(name="enableHttpCompression")
def enable_http_compression(self) -> Optional[pulumi.Input[bool]]:
"""
Enables or disables data compression in the response to an HTTP request.
"""
return pulumi.get(self, "enable_http_compression")
@enable_http_compression.setter
def enable_http_compression(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "enable_http_compression", value)
@property
@pulumi.getter(name="fallbackToStaleReplicasForDistributedQueries")
def fallback_to_stale_replicas_for_distributed_queries(self) -> Optional[pulumi.Input[bool]]:
"""
Forces a query to an out-of-date replica if updated data is not available.
"""
return pulumi.get(self, "fallback_to_stale_replicas_for_distributed_queries")
@fallback_to_stale_replicas_for_distributed_queries.setter
def fallback_to_stale_replicas_for_distributed_queries(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "fallback_to_stale_replicas_for_distributed_queries", value)
@property
@pulumi.getter(name="forceIndexByDate")
def force_index_by_date(self) -> Optional[pulumi.Input[bool]]:
"""
Disables query execution if the index can’t be used by date.
"""
return pulumi.get(self, "force_index_by_date")
@force_index_by_date.setter
def force_index_by_date(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "force_index_by_date", value)
@property
@pulumi.getter(name="forcePrimaryKey")
def force_primary_key(self) -> Optional[pulumi.Input[bool]]:
"""
Disables query execution if indexing by the primary key is not possible.
"""
return pulumi.get(self, "force_primary_key")
@force_primary_key.setter
def force_primary_key(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "force_primary_key", value)
@property
@pulumi.getter(name="groupByOverflowMode")
def group_by_overflow_mode(self) -> Optional[pulumi.Input[str]]:
"""
Sets behaviour on overflow while GROUP BY operation. Possible values:
"""
return pulumi.get(self, "group_by_overflow_mode")
@group_by_overflow_mode.setter
def group_by_overflow_mode(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "group_by_overflow_mode", value)
@property
@pulumi.getter(name="groupByTwoLevelThreshold")
def group_by_two_level_threshold(self) -> Optional[pulumi.Input[int]]:
"""
Sets the threshold of the number of keys, after that the two-level aggregation should be used.
"""
return pulumi.get(self, "group_by_two_level_threshold")
@group_by_two_level_threshold.setter
def group_by_two_level_threshold(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "group_by_two_level_threshold", value)
@property
@pulumi.getter(name="groupByTwoLevelThresholdBytes")
def group_by_two_level_threshold_bytes(self) -> Optional[pulumi.Input[int]]:
"""
Sets the threshold of the number of bytes, after that the two-level aggregation should be used.
"""
return pulumi.get(self, "group_by_two_level_threshold_bytes")
@group_by_two_level_threshold_bytes.setter
def group_by_two_level_threshold_bytes(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "group_by_two_level_threshold_bytes", value)
@property
@pulumi.getter(name="httpConnectionTimeout")
def http_connection_timeout(self) -> Optional[pulumi.Input[int]]:
"""
Timeout for HTTP connection in milliseconds.
"""
return pulumi.get(self, "http_connection_timeout")
@http_connection_timeout.setter
def http_connection_timeout(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "http_connection_timeout", value)
@property
@pulumi.getter(name="httpHeadersProgressInterval")
def http_headers_progress_interval(self) -> Optional[pulumi.Input[int]]:
"""
Sets minimal interval between notifications about request process in HTTP header X-ClickHouse-Progress.
"""
return pulumi.get(self, "http_headers_progress_interval")
@http_headers_progress_interval.setter
def http_headers_progress_interval(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "http_headers_progress_interval", value)
@property
@pulumi.getter(name="httpReceiveTimeout")
def http_receive_timeout(self) -> Optional[pulumi.Input[int]]:
"""
Timeout for HTTP connection in milliseconds.
"""
return pulumi.get(self, "http_receive_timeout")
@http_receive_timeout.setter
def http_receive_timeout(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "http_receive_timeout", value)
@property
@pulumi.getter(name="httpSendTimeout")
def http_send_timeout(self) -> Optional[pulumi.Input[int]]:
"""
Timeout for HTTP connection in milliseconds.
"""
return pulumi.get(self, "http_send_timeout")
@http_send_timeout.setter
def http_send_timeout(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "http_send_timeout", value)
@property
@pulumi.getter(name="inputFormatDefaultsForOmittedFields")
def input_format_defaults_for_omitted_fields(self) -> Optional[pulumi.Input[bool]]:
"""
When performing INSERT queries, replace omitted input column values with default values of the respective columns.
"""
return pulumi.get(self, "input_format_defaults_for_omitted_fields")
@input_format_defaults_for_omitted_fields.setter
def input_format_defaults_for_omitted_fields(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "input_format_defaults_for_omitted_fields", value)
@property
@pulumi.getter(name="inputFormatValuesInterpretExpressions")
def input_format_values_interpret_expressions(self) -> Optional[pulumi.Input[bool]]:
"""
Enables or disables the full SQL parser if the fast stream parser can’t parse the data.
"""
return pulumi.get(self, "input_format_values_interpret_expressions")
@input_format_values_interpret_expressions.setter
def input_format_values_interpret_expressions(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "input_format_values_interpret_expressions", value)
@property
@pulumi.getter(name="insertQuorum")
def insert_quorum(self) -> Optional[pulumi.Input[int]]:
"""
Enables the quorum writes.
"""
return pulumi.get(self, "insert_quorum")
@insert_quorum.setter
def insert_quorum(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "insert_quorum", value)
@property
@pulumi.getter(name="insertQuorumTimeout")
def insert_quorum_timeout(self) -> Optional[pulumi.Input[int]]:
"""
Write to a quorum timeout in milliseconds.
"""
return pulumi.get(self, "insert_quorum_timeout")
@insert_quorum_timeout.setter
def insert_quorum_timeout(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "insert_quorum_timeout", value)
@property
@pulumi.getter(name="joinOverflowMode")
def join_overflow_mode(self) -> Optional[pulumi.Input[str]]:
"""
Sets behaviour on overflow in JOIN. Possible values:
"""
return pulumi.get(self, "join_overflow_mode")
@join_overflow_mode.setter
def join_overflow_mode(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "join_overflow_mode", value)
@property
@pulumi.getter(name="joinUseNulls")
def join_use_nulls(self) -> Optional[pulumi.Input[bool]]:
"""
Sets the type of JOIN behaviour. When merging tables, empty cells may appear. ClickHouse fills them differently based on this setting.
"""
return pulumi.get(self, "join_use_nulls")
@join_use_nulls.setter
def join_use_nulls(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "join_use_nulls", value)
@property
@pulumi.getter(name="joinedSubqueryRequiresAlias")
def joined_subquery_requires_alias(self) -> Optional[pulumi.Input[bool]]:
"""
Require aliases for subselects and table functions in FROM that more than one table is present.
"""
return pulumi.get(self, "joined_subquery_requires_alias")
@joined_subquery_requires_alias.setter
def joined_subquery_requires_alias(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "joined_subquery_requires_alias", value)
@property
@pulumi.getter(name="lowCardinalityAllowInNativeFormat")
def low_cardinality_allow_in_native_format(self) -> Optional[pulumi.Input[bool]]:
"""
Allows or restricts using the LowCardinality data type with the Native format.
"""
return pulumi.get(self, "low_cardinality_allow_in_native_format")
@low_cardinality_allow_in_native_format.setter
def low_cardinality_allow_in_native_format(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "low_cardinality_allow_in_native_format", value)
@property
@pulumi.getter(name="maxAstDepth")
def max_ast_depth(self) -> Optional[pulumi.Input[int]]:
"""
Maximum abstract syntax tree depth.
"""
return pulumi.get(self, "max_ast_depth")
@max_ast_depth.setter
def max_ast_depth(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "max_ast_depth", value)
@property
@pulumi.getter(name="maxAstElements")
def max_ast_elements(self) -> Optional[pulumi.Input[int]]:
"""
Maximum abstract syntax tree elements.
"""
return pulumi.get(self, "max_ast_elements")
@max_ast_elements.setter
def max_ast_elements(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "max_ast_elements", value)
@property
@pulumi.getter(name="maxBlockSize")
def max_block_size(self) -> Optional[pulumi.Input[int]]:
"""
A recommendation for what size of the block (in a count of rows) to load from tables.
"""
return pulumi.get(self, "max_block_size")
@max_block_size.setter
def max_block_size(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "max_block_size", value)
@property
@pulumi.getter(name="maxBytesBeforeExternalGroupBy")
def max_bytes_before_external_group_by(self) -> Optional[pulumi.Input[int]]:
"""
Limit in bytes for using memoru for GROUP BY before using swap on disk.
"""
return pulumi.get(self, "max_bytes_before_external_group_by")
@max_bytes_before_external_group_by.setter
def max_bytes_before_external_group_by(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "max_bytes_before_external_group_by", value)
@property
@pulumi.getter(name="maxBytesBeforeExternalSort")
def max_bytes_before_external_sort(self) -> Optional[pulumi.Input[int]]:
"""
This setting is equivalent of the max_bytes_before_external_group_by setting, except for it is for sort operation (ORDER BY), not aggregation.
"""
return pulumi.get(self, "max_bytes_before_external_sort")
@max_bytes_before_external_sort.setter
def max_bytes_before_external_sort(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "max_bytes_before_external_sort", value)
@property
@pulumi.getter(name="maxBytesInDistinct")
def max_bytes_in_distinct(self) -> Optional[pulumi.Input[int]]:
"""
Limits the maximum size of a hash table in bytes (uncompressed data) when using DISTINCT.
"""
return pulumi.get(self, "max_bytes_in_distinct")
@max_bytes_in_distinct.setter
def max_bytes_in_distinct(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "max_bytes_in_distinct", value)
@property
@pulumi.getter(name="maxBytesInJoin")
def max_bytes_in_join(self) -> Optional[pulumi.Input[int]]:
"""
Limit on maximum size of the hash table for JOIN, in bytes.
"""
return pulumi.get(self, "max_bytes_in_join")
@max_bytes_in_join.setter
def max_bytes_in_join(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "max_bytes_in_join", value)
@property
@pulumi.getter(name="maxBytesInSet")
def max_bytes_in_set(self) -> Optional[pulumi.Input[int]]:
"""
Limit on the number of bytes in the set resulting from the execution of the IN section.
"""
return pulumi.get(self, "max_bytes_in_set")
@max_bytes_in_set.setter
def max_bytes_in_set(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "max_bytes_in_set", value)
@property
@pulumi.getter(name="maxBytesToRead")
def max_bytes_to_read(self) -> Optional[pulumi.Input[int]]:
"""
Limits the maximum number of bytes (uncompressed data) that can be read from a table when running a query.
"""
return pulumi.get(self, "max_bytes_to_read")
@max_bytes_to_read.setter
def max_bytes_to_read(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "max_bytes_to_read", value)
@property
@pulumi.getter(name="maxBytesToSort")
def max_bytes_to_sort(self) -> Optional[pulumi.Input[int]]:
"""
Limits the maximum number of bytes (uncompressed data) that can be read from a table for sorting.
"""
return pulumi.get(self, "max_bytes_to_sort")
@max_bytes_to_sort.setter
def max_bytes_to_sort(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "max_bytes_to_sort", value)
@property
@pulumi.getter(name="maxBytesToTransfer")
def max_bytes_to_transfer(self) -> Optional[pulumi.Input[int]]:
"""
Limits the maximum number of bytes (uncompressed data) that can be passed to a remote server or saved in a temporary table when using GLOBAL IN.
"""
return pulumi.get(self, "max_bytes_to_transfer")
@max_bytes_to_transfer.setter
def max_bytes_to_transfer(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "max_bytes_to_transfer", value)
@property
@pulumi.getter(name="maxColumnsToRead")
def max_columns_to_read(self) -> Optional[pulumi.Input[int]]:
"""
Limits the maximum number of columns that can be read from a table in a single query.
"""
return pulumi.get(self, "max_columns_to_read")
@max_columns_to_read.setter
def max_columns_to_read(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "max_columns_to_read", value)
@property
@pulumi.getter(name="maxExecutionTime")
def max_execution_time(self) -> Optional[pulumi.Input[int]]:
"""
Limits the maximum query execution time in milliseconds.
"""
return pulumi.get(self, "max_execution_time")
@max_execution_time.setter
def max_execution_time(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "max_execution_time", value)
@property
@pulumi.getter(name="maxExpandedAstElements")
def max_expanded_ast_elements(self) -> Optional[pulumi.Input[int]]:
"""
Maximum abstract syntax tree depth after after expansion of aliases.
"""
return pulumi.get(self, "max_expanded_ast_elements")
@max_expanded_ast_elements.setter
def max_expanded_ast_elements(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "max_expanded_ast_elements", value)
@property
@pulumi.getter(name="maxInsertBlockSize")
def max_insert_block_size(self) -> Optional[pulumi.Input[int]]:
"""
The size of blocks (in a count of rows) to form for insertion into a table.
"""
return pulumi.get(self, "max_insert_block_size")
@max_insert_block_size.setter
def max_insert_block_size(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "max_insert_block_size", value)
@property
@pulumi.getter(name="maxMemoryUsage")
def max_memory_usage(self) -> Optional[pulumi.Input[int]]:
"""
Limits the maximum memory usage (in bytes) for processing queries on a single server.
"""
return pulumi.get(self, "max_memory_usage")
@max_memory_usage.setter
def max_memory_usage(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "max_memory_usage", value)
@property
@pulumi.getter(name="maxMemoryUsageForUser")
def max_memory_usage_for_user(self) -> Optional[pulumi.Input[int]]:
"""
Limits the maximum memory usage (in bytes) for processing of user's queries on a single server.
"""
return pulumi.get(self, "max_memory_usage_for_user")
@max_memory_usage_for_user.setter
def max_memory_usage_for_user(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "max_memory_usage_for_user", value)
@property
@pulumi.getter(name="maxNetworkBandwidth")
def max_network_bandwidth(self) -> Optional[pulumi.Input[int]]:
"""
Limits the speed of the data exchange over the network in bytes per second.
"""
return pulumi.get(self, "max_network_bandwidth")
@max_network_bandwidth.setter
def max_network_bandwidth(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "max_network_bandwidth", value)
@property
@pulumi.getter(name="maxNetworkBandwidthForUser")
def max_network_bandwidth_for_user(self) -> Optional[pulumi.Input[int]]:
"""
Limits the speed of the data exchange over the network in bytes per second.
"""
return pulumi.get(self, "max_network_bandwidth_for_user")
@max_network_bandwidth_for_user.setter
def max_network_bandwidth_for_user(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "max_network_bandwidth_for_user", value)
@property
@pulumi.getter(name="maxQuerySize")
def max_query_size(self) -> Optional[pulumi.Input[int]]:
"""
The maximum part of a query that can be taken to RAM for parsing with the SQL parser.
"""
return pulumi.get(self, "max_query_size")
@max_query_size.setter
def max_query_size(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "max_query_size", value)
@property
@pulumi.getter(name="maxReplicaDelayForDistributedQueries")
def max_replica_delay_for_distributed_queries(self) -> Optional[pulumi.Input[int]]:
"""
Disables lagging replicas for distributed queries.
"""
return pulumi.get(self, "max_replica_delay_for_distributed_queries")
@max_replica_delay_for_distributed_queries.setter
def max_replica_delay_for_distributed_queries(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "max_replica_delay_for_distributed_queries", value)
@property
@pulumi.getter(name="maxResultBytes")
def max_result_bytes(self) -> Optional[pulumi.Input[int]]:
"""
Limits the number of bytes in the result.
"""
return pulumi.get(self, "max_result_bytes")
@max_result_bytes.setter
def max_result_bytes(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "max_result_bytes", value)
@property
@pulumi.getter(name="maxResultRows")
def max_result_rows(self) -> Optional[pulumi.Input[int]]:
"""
Limits the number of rows in the result.
"""
return pulumi.get(self, "max_result_rows")
@max_result_rows.setter
def max_result_rows(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "max_result_rows", value)
@property
@pulumi.getter(name="maxRowsInDistinct")
def max_rows_in_distinct(self) -> Optional[pulumi.Input[int]]:
"""
Limits the maximum number of different rows when using DISTINCT.
"""
return pulumi.get(self, "max_rows_in_distinct")
@max_rows_in_distinct.setter
def max_rows_in_distinct(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "max_rows_in_distinct", value)
@property
@pulumi.getter(name="maxRowsInJoin")
def max_rows_in_join(self) -> Optional[pulumi.Input[int]]:
"""
Limit on maximum size of the hash table for JOIN, in rows.
"""
return pulumi.get(self, "max_rows_in_join")
@max_rows_in_join.setter
def max_rows_in_join(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "max_rows_in_join", value)
@property
@pulumi.getter(name="maxRowsInSet")
def max_rows_in_set(self) -> Optional[pulumi.Input[int]]:
"""
Limit on the number of rows in the set resulting from the execution of the IN section.
"""
return pulumi.get(self, "max_rows_in_set")
@max_rows_in_set.setter
def max_rows_in_set(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "max_rows_in_set", value)
@property
@pulumi.getter(name="maxRowsToGroupBy")
def max_rows_to_group_by(self) -> Optional[pulumi.Input[int]]:
"""
Limits the maximum number of unique keys received from aggregation function.
"""
return pulumi.get(self, "max_rows_to_group_by")
@max_rows_to_group_by.setter
def max_rows_to_group_by(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "max_rows_to_group_by", value)
@property
@pulumi.getter(name="maxRowsToRead")
def max_rows_to_read(self) -> Optional[pulumi.Input[int]]:
"""
Limits the maximum number of rows that can be read from a table when running a query.
"""
return pulumi.get(self, "max_rows_to_read")
@max_rows_to_read.setter
def max_rows_to_read(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "max_rows_to_read", value)
@property
@pulumi.getter(name="maxRowsToSort")
def max_rows_to_sort(self) -> Optional[pulumi.Input[int]]:
"""
Limits the maximum number of rows that can be read from a table for sorting.
"""
return pulumi.get(self, "max_rows_to_sort")
@max_rows_to_sort.setter
def max_rows_to_sort(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "max_rows_to_sort", value)
@property
@pulumi.getter(name="maxRowsToTransfer")
def max_rows_to_transfer(self) -> Optional[pulumi.Input[int]]:
"""
Limits the maximum number of rows that can be passed to a remote server or saved in a temporary table when using GLOBAL IN.
"""
return pulumi.get(self, "max_rows_to_transfer")
@max_rows_to_transfer.setter
def max_rows_to_transfer(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "max_rows_to_transfer", value)
@property
@pulumi.getter(name="maxTemporaryColumns")
def max_temporary_columns(self) -> Optional[pulumi.Input[int]]:
"""
Limits the maximum number of temporary columns that must be kept in RAM at the same time when running a query, including constant columns.
"""
return pulumi.get(self, "max_temporary_columns")
@max_temporary_columns.setter
def max_temporary_columns(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "max_temporary_columns", value)
@property
@pulumi.getter(name="maxTemporaryNonConstColumns")
def max_temporary_non_const_columns(self) -> Optional[pulumi.Input[int]]:
"""
Limits the maximum number of temporary columns that must be kept in RAM at the same time when running a query, excluding constant columns.
"""
return pulumi.get(self, "max_temporary_non_const_columns")
@max_temporary_non_const_columns.setter
def max_temporary_non_const_columns(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "max_temporary_non_const_columns", value)
@property
@pulumi.getter(name="maxThreads")
def max_threads(self) -> Optional[pulumi.Input[int]]:
"""
The maximum number of query processing threads, excluding threads for retrieving data from remote servers.
"""
return pulumi.get(self, "max_threads")
@max_threads.setter
def max_threads(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "max_threads", value)
@property
@pulumi.getter(name="mergeTreeMaxBytesToUseCache")
def merge_tree_max_bytes_to_use_cache(self) -> Optional[pulumi.Input[int]]:
"""
If ClickHouse should read more than merge_tree_max_bytes_to_use_cache bytes in one query, it doesn’t use the cache of uncompressed blocks.
"""
return pulumi.get(self, "merge_tree_max_bytes_to_use_cache")
@merge_tree_max_bytes_to_use_cache.setter
def merge_tree_max_bytes_to_use_cache(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "merge_tree_max_bytes_to_use_cache", value)
@property
@pulumi.getter(name="mergeTreeMaxRowsToUseCache")
def merge_tree_max_rows_to_use_cache(self) -> Optional[pulumi.Input[int]]:
"""
If ClickHouse should read more than merge_tree_max_rows_to_use_cache rows in one query, it doesn’t use the cache of uncompressed blocks.
"""
return pulumi.get(self, "merge_tree_max_rows_to_use_cache")
@merge_tree_max_rows_to_use_cache.setter
def merge_tree_max_rows_to_use_cache(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "merge_tree_max_rows_to_use_cache", value)
@property
@pulumi.getter(name="mergeTreeMinBytesForConcurrentRead")
def merge_tree_min_bytes_for_concurrent_read(self) -> Optional[pulumi.Input[int]]:
"""
If the number of bytes to read from one file of a MergeTree-engine table exceeds merge_tree_min_bytes_for_concurrent_read, then ClickHouse tries to concurrently read from this file in several threads.
"""
return pulumi.get(self, "merge_tree_min_bytes_for_concurrent_read")
@merge_tree_min_bytes_for_concurrent_read.setter
def merge_tree_min_bytes_for_concurrent_read(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "merge_tree_min_bytes_for_concurrent_read", value)
@property
@pulumi.getter(name="mergeTreeMinRowsForConcurrentRead")
def merge_tree_min_rows_for_concurrent_read(self) -> Optional[pulumi.Input[int]]:
"""
If the number of rows to be read from a file of a MergeTree table exceeds merge_tree_min_rows_for_concurrent_read then ClickHouse tries to perform a concurrent reading from this file on several threads.
"""
return pulumi.get(self, "merge_tree_min_rows_for_concurrent_read")
@merge_tree_min_rows_for_concurrent_read.setter
def merge_tree_min_rows_for_concurrent_read(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "merge_tree_min_rows_for_concurrent_read", value)
@property
@pulumi.getter(name="minBytesToUseDirectIo")
def min_bytes_to_use_direct_io(self) -> Optional[pulumi.Input[int]]:
"""
The minimum data volume required for using direct I/O access to the storage disk.
"""
return pulumi.get(self, "min_bytes_to_use_direct_io")
@min_bytes_to_use_direct_io.setter
def min_bytes_to_use_direct_io(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "min_bytes_to_use_direct_io", value)
@property
@pulumi.getter(name="minCountToCompile")
def min_count_to_compile(self) -> Optional[pulumi.Input[int]]:
"""
How many times to potentially use a compiled chunk of code before running compilation.
"""
return pulumi.get(self, "min_count_to_compile")
@min_count_to_compile.setter
def min_count_to_compile(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "min_count_to_compile", value)
@property
@pulumi.getter(name="minCountToCompileExpression")
def min_count_to_compile_expression(self) -> Optional[pulumi.Input[int]]:
"""
A query waits for expression compilation process to complete prior to continuing execution.
"""
return pulumi.get(self, "min_count_to_compile_expression")
@min_count_to_compile_expression.setter
def min_count_to_compile_expression(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "min_count_to_compile_expression", value)
@property
@pulumi.getter(name="minExecutionSpeed")
def min_execution_speed(self) -> Optional[pulumi.Input[int]]:
"""
Minimal execution speed in rows per second.
"""
return pulumi.get(self, "min_execution_speed")
@min_execution_speed.setter
def min_execution_speed(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "min_execution_speed", value)
@property
@pulumi.getter(name="minExecutionSpeedBytes")
def min_execution_speed_bytes(self) -> Optional[pulumi.Input[int]]:
"""
Minimal execution speed in bytes per second.
"""
return pulumi.get(self, "min_execution_speed_bytes")
@min_execution_speed_bytes.setter
def min_execution_speed_bytes(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "min_execution_speed_bytes", value)
@property
@pulumi.getter(name="minInsertBlockSizeBytes")
def min_insert_block_size_bytes(self) -> Optional[pulumi.Input[int]]:
"""
Sets the minimum number of bytes in the block which can be inserted into a table by an INSERT query.
"""
return pulumi.get(self, "min_insert_block_size_bytes")
@min_insert_block_size_bytes.setter
def min_insert_block_size_bytes(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "min_insert_block_size_bytes", value)
@property
@pulumi.getter(name="minInsertBlockSizeRows")
def min_insert_block_size_rows(self) -> Optional[pulumi.Input[int]]:
"""
Sets the minimum number of rows in the block which can be inserted into a table by an INSERT query.
"""
return pulumi.get(self, "min_insert_block_size_rows")
@min_insert_block_size_rows.setter
def min_insert_block_size_rows(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "min_insert_block_size_rows", value)
@property
@pulumi.getter(name="outputFormatJsonQuote64bitIntegers")
def output_format_json_quote64bit_integers(self) -> Optional[pulumi.Input[bool]]:
"""
If the value is true, integers appear in quotes when using JSON* Int64 and UInt64 formats (for compatibility with most JavaScript implementations); otherwise, integers are output without the quotes.
"""
return pulumi.get(self, "output_format_json_quote64bit_integers")
@output_format_json_quote64bit_integers.setter
def output_format_json_quote64bit_integers(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "output_format_json_quote64bit_integers", value)
@property
@pulumi.getter(name="outputFormatJsonQuoteDenormals")
def output_format_json_quote_denormals(self) -> Optional[pulumi.Input[bool]]:
"""
Enables +nan, -nan, +inf, -inf outputs in JSON output format.
"""
return pulumi.get(self, "output_format_json_quote_denormals")
@output_format_json_quote_denormals.setter
def output_format_json_quote_denormals(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "output_format_json_quote_denormals", value)
@property
@pulumi.getter
def priority(self) -> Optional[pulumi.Input[int]]:
"""
Query priority.
"""
return pulumi.get(self, "priority")
@priority.setter
def priority(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "priority", value)
@property
@pulumi.getter(name="quotaMode")
def quota_mode(self) -> Optional[pulumi.Input[str]]:
"""
Quota accounting mode.
"""
return pulumi.get(self, "quota_mode")
@quota_mode.setter
def quota_mode(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "quota_mode", value)
@property
@pulumi.getter(name="readOverflowMode")
def read_overflow_mode(self) -> Optional[pulumi.Input[str]]:
"""
Sets behaviour on overflow while read. Possible values:
"""
return pulumi.get(self, "read_overflow_mode")
@read_overflow_mode.setter
def read_overflow_mode(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "read_overflow_mode", value)
@property
@pulumi.getter
def readonly(self) -> Optional[pulumi.Input[int]]:
"""
Restricts permissions for reading data, write data and change settings queries.
"""
return pulumi.get(self, "readonly")
@readonly.setter
def readonly(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "readonly", value)
@property
@pulumi.getter(name="receiveTimeout")
def receive_timeout(self) -> Optional[pulumi.Input[int]]:
"""
Receive timeout in milliseconds on the socket used for communicating with the client.
"""
return pulumi.get(self, "receive_timeout")
@receive_timeout.setter
def receive_timeout(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "receive_timeout", value)
@property
@pulumi.getter(name="replicationAlterPartitionsSync")
def replication_alter_partitions_sync(self) -> Optional[pulumi.Input[int]]:
"""
For ALTER ... ATTACH|DETACH|DROP queries, you can use the replication_alter_partitions_sync setting to set up waiting.
"""
return pulumi.get(self, "replication_alter_partitions_sync")
@replication_alter_partitions_sync.setter
def replication_alter_partitions_sync(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "replication_alter_partitions_sync", value)
@property
@pulumi.getter(name="resultOverflowMode")
def result_overflow_mode(self) -> Optional[pulumi.Input[str]]:
"""
Sets behaviour on overflow in result. Possible values:
"""
return pulumi.get(self, "result_overflow_mode")
@result_overflow_mode.setter
def result_overflow_mode(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "result_overflow_mode", value)
@property
@pulumi.getter(name="selectSequentialConsistency")
def select_sequential_consistency(self) -> Optional[pulumi.Input[bool]]:
"""
Enables or disables sequential consistency for SELECT queries.
"""
return pulumi.get(self, "select_sequential_consistency")
@select_sequential_consistency.setter
def select_sequential_consistency(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "select_sequential_consistency", value)
@property
@pulumi.getter(name="sendProgressInHttpHeaders")
def send_progress_in_http_headers(self) -> Optional[pulumi.Input[bool]]:
"""
Enables or disables X-ClickHouse-Progress HTTP response headers in clickhouse-server responses.
"""
return pulumi.get(self, "send_progress_in_http_headers")
@send_progress_in_http_headers.setter
def send_progress_in_http_headers(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "send_progress_in_http_headers", value)
@property
@pulumi.getter(name="sendTimeout")
def send_timeout(self) -> Optional[pulumi.Input[int]]:
"""
Send timeout in milliseconds on the socket used for communicating with the client.
"""
return pulumi.get(self, "send_timeout")
@send_timeout.setter
def send_timeout(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "send_timeout", value)
@property
@pulumi.getter(name="setOverflowMode")
def set_overflow_mode(self) -> Optional[pulumi.Input[str]]:
"""
Sets behaviour on overflow in the set resulting. Possible values:
"""
return pulumi.get(self, "set_overflow_mode")
@set_overflow_mode.setter
def set_overflow_mode(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "set_overflow_mode", value)
@property
@pulumi.getter(name="skipUnavailableShards")
def skip_unavailable_shards(self) -> Optional[pulumi.Input[bool]]:
"""
Enables or disables silently skipping of unavailable shards.
"""
return pulumi.get(self, "skip_unavailable_shards")
@skip_unavailable_shards.setter
def skip_unavailable_shards(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "skip_unavailable_shards", value)
@property
@pulumi.getter(name="sortOverflowMode")
def sort_overflow_mode(self) -> Optional[pulumi.Input[str]]:
"""
Sets behaviour on overflow while sort. Possible values:
"""
return pulumi.get(self, "sort_overflow_mode")
@sort_overflow_mode.setter
def sort_overflow_mode(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "sort_overflow_mode", value)
@property
@pulumi.getter(name="timeoutOverflowMode")
def timeout_overflow_mode(self) -> Optional[pulumi.Input[str]]:
"""
Sets behaviour on overflow. Possible values:
"""
return pulumi.get(self, "timeout_overflow_mode")
@timeout_overflow_mode.setter
def timeout_overflow_mode(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "timeout_overflow_mode", value)
@property
@pulumi.getter(name="transferOverflowMode")
def transfer_overflow_mode(self) -> Optional[pulumi.Input[str]]:
"""
Sets behaviour on overflow. Possible values:
"""
return pulumi.get(self, "transfer_overflow_mode")
@transfer_overflow_mode.setter
def transfer_overflow_mode(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "transfer_overflow_mode", value)
@property
@pulumi.getter(name="transformNullIn")
def transform_null_in(self) -> Optional[pulumi.Input[bool]]:
"""
Enables equality of NULL values for IN operator.
"""
return pulumi.get(self, "transform_null_in")
@transform_null_in.setter
def transform_null_in(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "transform_null_in", value)
@property
@pulumi.getter(name="useUncompressedCache")
def use_uncompressed_cache(self) -> Optional[pulumi.Input[bool]]:
"""
Whether to use a cache of uncompressed blocks.
"""
return pulumi.get(self, "use_uncompressed_cache")
@use_uncompressed_cache.setter
def use_uncompressed_cache(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "use_uncompressed_cache", value)
@pulumi.input_type
class MdbClickhouseClusterZookeeperArgs:
def __init__(__self__, *,
resources: Optional[pulumi.Input['MdbClickhouseClusterZookeeperResourcesArgs']] = None):
"""
:param pulumi.Input['MdbClickhouseClusterZookeeperResourcesArgs'] resources: Resources allocated to hosts of the ZooKeeper subcluster. The structure is documented below.
"""
if resources is not None:
pulumi.set(__self__, "resources", resources)
@property
@pulumi.getter
def resources(self) -> Optional[pulumi.Input['MdbClickhouseClusterZookeeperResourcesArgs']]:
"""
Resources allocated to hosts of the ZooKeeper subcluster. The structure is documented below.
"""
return pulumi.get(self, "resources")
@resources.setter
def resources(self, value: Optional[pulumi.Input['MdbClickhouseClusterZookeeperResourcesArgs']]):
pulumi.set(self, "resources", value)
@pulumi.input_type
class MdbClickhouseClusterZookeeperResourcesArgs:
def __init__(__self__, *,
disk_size: Optional[pulumi.Input[int]] = None,
disk_type_id: Optional[pulumi.Input[str]] = None,
resource_preset_id: Optional[pulumi.Input[str]] = None):
"""
:param pulumi.Input[int] disk_size: Volume of the storage available to a ZooKeeper host, in gigabytes.
:param pulumi.Input[str] disk_type_id: Type of the storage of ZooKeeper hosts.
For more information see [the official documentation](https://cloud.yandex.com/docs/managed-clickhouse/concepts/storage).
"""
if disk_size is not None:
pulumi.set(__self__, "disk_size", disk_size)
if disk_type_id is not None:
pulumi.set(__self__, "disk_type_id", disk_type_id)
if resource_preset_id is not None:
pulumi.set(__self__, "resource_preset_id", resource_preset_id)
@property
@pulumi.getter(name="diskSize")
def disk_size(self) -> Optional[pulumi.Input[int]]:
"""
Volume of the storage available to a ZooKeeper host, in gigabytes.
"""
return pulumi.get(self, "disk_size")
@disk_size.setter
def disk_size(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "disk_size", value)
@property
@pulumi.getter(name="diskTypeId")
def disk_type_id(self) -> Optional[pulumi.Input[str]]:
"""
Type of the storage of ZooKeeper hosts.
For more information see [the official documentation](https://cloud.yandex.com/docs/managed-clickhouse/concepts/storage).
"""
return pulumi.get(self, "disk_type_id")
@disk_type_id.setter
def disk_type_id(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "disk_type_id", value)
@property
@pulumi.getter(name="resourcePresetId")
def resource_preset_id(self) -> Optional[pulumi.Input[str]]:
return pulumi.get(self, "resource_preset_id")
@resource_preset_id.setter
def resource_preset_id(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "resource_preset_id", value)
@pulumi.input_type
class MdbElasticSearchClusterConfigArgs:
def __init__(__self__, *,
admin_password: pulumi.Input[str],
data_node: pulumi.Input['MdbElasticSearchClusterConfigDataNodeArgs'],
edition: Optional[pulumi.Input[str]] = None,
master_node: Optional[pulumi.Input['MdbElasticSearchClusterConfigMasterNodeArgs']] = None,
plugins: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None,
version: Optional[pulumi.Input[str]] = None):
"""
:param pulumi.Input[str] admin_password: Password for admin user of Elasticsearch.
:param pulumi.Input['MdbElasticSearchClusterConfigDataNodeArgs'] data_node: Configuration for Elasticsearch data nodes subcluster. The structure is documented below.
:param pulumi.Input[str] edition: Edition of Elasticsearch. For more information, see [the official documentation](https://cloud.yandex.com/en-ru/docs/managed-elasticsearch/concepts/es-editions).
:param pulumi.Input['MdbElasticSearchClusterConfigMasterNodeArgs'] master_node: Configuration for Elasticsearch master nodes subcluster. The structure is documented below.
:param pulumi.Input[Sequence[pulumi.Input[str]]] plugins: A set of Elasticsearch plugins to install.
:param pulumi.Input[str] version: Version of Elasticsearch.
"""
pulumi.set(__self__, "admin_password", admin_password)
pulumi.set(__self__, "data_node", data_node)
if edition is not None:
pulumi.set(__self__, "edition", edition)
if master_node is not None:
pulumi.set(__self__, "master_node", master_node)
if plugins is not None:
pulumi.set(__self__, "plugins", plugins)
if version is not None:
pulumi.set(__self__, "version", version)
@property
@pulumi.getter(name="adminPassword")
def admin_password(self) -> pulumi.Input[str]:
"""
Password for admin user of Elasticsearch.
"""
return pulumi.get(self, "admin_password")
@admin_password.setter
def admin_password(self, value: pulumi.Input[str]):
pulumi.set(self, "admin_password", value)
@property
@pulumi.getter(name="dataNode")
def data_node(self) -> pulumi.Input['MdbElasticSearchClusterConfigDataNodeArgs']:
"""
Configuration for Elasticsearch data nodes subcluster. The structure is documented below.
"""
return pulumi.get(self, "data_node")
@data_node.setter
def data_node(self, value: pulumi.Input['MdbElasticSearchClusterConfigDataNodeArgs']):
pulumi.set(self, "data_node", value)
@property
@pulumi.getter
def edition(self) -> Optional[pulumi.Input[str]]:
"""
Edition of Elasticsearch. For more information, see [the official documentation](https://cloud.yandex.com/en-ru/docs/managed-elasticsearch/concepts/es-editions).
"""
return pulumi.get(self, "edition")
@edition.setter
def edition(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "edition", value)
@property
@pulumi.getter(name="masterNode")
def master_node(self) -> Optional[pulumi.Input['MdbElasticSearchClusterConfigMasterNodeArgs']]:
"""
Configuration for Elasticsearch master nodes subcluster. The structure is documented below.
"""
return pulumi.get(self, "master_node")
@master_node.setter
def master_node(self, value: Optional[pulumi.Input['MdbElasticSearchClusterConfigMasterNodeArgs']]):
pulumi.set(self, "master_node", value)
@property
@pulumi.getter
def plugins(self) -> Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]:
"""
A set of Elasticsearch plugins to install.
"""
return pulumi.get(self, "plugins")
@plugins.setter
def plugins(self, value: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]):
pulumi.set(self, "plugins", value)
@property
@pulumi.getter
def version(self) -> Optional[pulumi.Input[str]]:
"""
Version of Elasticsearch.
"""
return pulumi.get(self, "version")
@version.setter
def version(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "version", value)
@pulumi.input_type
class MdbElasticSearchClusterConfigDataNodeArgs:
def __init__(__self__, *,
resources: pulumi.Input['MdbElasticSearchClusterConfigDataNodeResourcesArgs']):
"""
:param pulumi.Input['MdbElasticSearchClusterConfigDataNodeResourcesArgs'] resources: Resources allocated to hosts of the Elasticsearch master nodes subcluster. The structure is documented below.
"""
pulumi.set(__self__, "resources", resources)
@property
@pulumi.getter
def resources(self) -> pulumi.Input['MdbElasticSearchClusterConfigDataNodeResourcesArgs']:
"""
Resources allocated to hosts of the Elasticsearch master nodes subcluster. The structure is documented below.
"""
return pulumi.get(self, "resources")
@resources.setter
def resources(self, value: pulumi.Input['MdbElasticSearchClusterConfigDataNodeResourcesArgs']):
pulumi.set(self, "resources", value)
@pulumi.input_type
class MdbElasticSearchClusterConfigDataNodeResourcesArgs:
def __init__(__self__, *,
disk_size: pulumi.Input[int],
disk_type_id: pulumi.Input[str],
resource_preset_id: pulumi.Input[str]):
"""
:param pulumi.Input[int] disk_size: Volume of the storage available to a host, in gigabytes.
:param pulumi.Input[str] disk_type_id: Type of the storage of Elasticsearch hosts.
"""
pulumi.set(__self__, "disk_size", disk_size)
pulumi.set(__self__, "disk_type_id", disk_type_id)
pulumi.set(__self__, "resource_preset_id", resource_preset_id)
@property
@pulumi.getter(name="diskSize")
def disk_size(self) -> pulumi.Input[int]:
"""
Volume of the storage available to a host, in gigabytes.
"""
return pulumi.get(self, "disk_size")
@disk_size.setter
def disk_size(self, value: pulumi.Input[int]):
pulumi.set(self, "disk_size", value)
@property
@pulumi.getter(name="diskTypeId")
def disk_type_id(self) -> pulumi.Input[str]:
"""
Type of the storage of Elasticsearch hosts.
"""
return pulumi.get(self, "disk_type_id")
@disk_type_id.setter
def disk_type_id(self, value: pulumi.Input[str]):
pulumi.set(self, "disk_type_id", value)
@property
@pulumi.getter(name="resourcePresetId")
def resource_preset_id(self) -> pulumi.Input[str]:
return pulumi.get(self, "resource_preset_id")
@resource_preset_id.setter
def resource_preset_id(self, value: pulumi.Input[str]):
pulumi.set(self, "resource_preset_id", value)
@pulumi.input_type
class MdbElasticSearchClusterConfigMasterNodeArgs:
def __init__(__self__, *,
resources: pulumi.Input['MdbElasticSearchClusterConfigMasterNodeResourcesArgs']):
"""
:param pulumi.Input['MdbElasticSearchClusterConfigMasterNodeResourcesArgs'] resources: Resources allocated to hosts of the Elasticsearch master nodes subcluster. The structure is documented below.
"""
pulumi.set(__self__, "resources", resources)
@property
@pulumi.getter
def resources(self) -> pulumi.Input['MdbElasticSearchClusterConfigMasterNodeResourcesArgs']:
"""
Resources allocated to hosts of the Elasticsearch master nodes subcluster. The structure is documented below.
"""
return pulumi.get(self, "resources")
@resources.setter
def resources(self, value: pulumi.Input['MdbElasticSearchClusterConfigMasterNodeResourcesArgs']):
pulumi.set(self, "resources", value)
@pulumi.input_type
class MdbElasticSearchClusterConfigMasterNodeResourcesArgs:
def __init__(__self__, *,
disk_size: pulumi.Input[int],
disk_type_id: pulumi.Input[str],
resource_preset_id: pulumi.Input[str]):
"""
:param pulumi.Input[int] disk_size: Volume of the storage available to a host, in gigabytes.
:param pulumi.Input[str] disk_type_id: Type of the storage of Elasticsearch hosts.
"""
pulumi.set(__self__, "disk_size", disk_size)
pulumi.set(__self__, "disk_type_id", disk_type_id)
pulumi.set(__self__, "resource_preset_id", resource_preset_id)
@property
@pulumi.getter(name="diskSize")
def disk_size(self) -> pulumi.Input[int]:
"""
Volume of the storage available to a host, in gigabytes.
"""
return pulumi.get(self, "disk_size")
@disk_size.setter
def disk_size(self, value: pulumi.Input[int]):
pulumi.set(self, "disk_size", value)
@property
@pulumi.getter(name="diskTypeId")
def disk_type_id(self) -> pulumi.Input[str]:
"""
Type of the storage of Elasticsearch hosts.
"""
return pulumi.get(self, "disk_type_id")
@disk_type_id.setter
def disk_type_id(self, value: pulumi.Input[str]):
pulumi.set(self, "disk_type_id", value)
@property
@pulumi.getter(name="resourcePresetId")
def resource_preset_id(self) -> pulumi.Input[str]:
return pulumi.get(self, "resource_preset_id")
@resource_preset_id.setter
def resource_preset_id(self, value: pulumi.Input[str]):
pulumi.set(self, "resource_preset_id", value)
@pulumi.input_type
class MdbElasticSearchClusterHostArgs:
def __init__(__self__, *,
name: pulumi.Input[str],
type: pulumi.Input[str],
zone: pulumi.Input[str],
assign_public_ip: Optional[pulumi.Input[bool]] = None,
fqdn: Optional[pulumi.Input[str]] = None,
subnet_id: Optional[pulumi.Input[str]] = None):
"""
:param pulumi.Input[str] name: User defined host name.
:param pulumi.Input[str] type: The type of the host to be deployed. Can be either `DATA_NODE` or `MASTER_NODE`.
:param pulumi.Input[str] zone: The availability zone where the Elasticsearch host will be created.
For more information see [the official documentation](https://cloud.yandex.com/docs/overview/concepts/geo-scope).
:param pulumi.Input[bool] assign_public_ip: Sets whether the host should get a public IP address on creation. Can be either `true` or `false`.
:param pulumi.Input[str] fqdn: The fully qualified domain name of the host.
:param pulumi.Input[str] subnet_id: The ID of the subnet, to which the host belongs. The subnet must
be a part of the network to which the cluster belongs.
"""
pulumi.set(__self__, "name", name)
pulumi.set(__self__, "type", type)
pulumi.set(__self__, "zone", zone)
if assign_public_ip is not None:
pulumi.set(__self__, "assign_public_ip", assign_public_ip)
if fqdn is not None:
pulumi.set(__self__, "fqdn", fqdn)
if subnet_id is not None:
pulumi.set(__self__, "subnet_id", subnet_id)
@property
@pulumi.getter
def name(self) -> pulumi.Input[str]:
"""
User defined host name.
"""
return pulumi.get(self, "name")
@name.setter
def name(self, value: pulumi.Input[str]):
pulumi.set(self, "name", value)
@property
@pulumi.getter
def type(self) -> pulumi.Input[str]:
"""
The type of the host to be deployed. Can be either `DATA_NODE` or `MASTER_NODE`.
"""
return pulumi.get(self, "type")
@type.setter
def type(self, value: pulumi.Input[str]):
pulumi.set(self, "type", value)
@property
@pulumi.getter
def zone(self) -> pulumi.Input[str]:
"""
The availability zone where the Elasticsearch host will be created.
For more information see [the official documentation](https://cloud.yandex.com/docs/overview/concepts/geo-scope).
"""
return pulumi.get(self, "zone")
@zone.setter
def zone(self, value: pulumi.Input[str]):
pulumi.set(self, "zone", value)
@property
@pulumi.getter(name="assignPublicIp")
def assign_public_ip(self) -> Optional[pulumi.Input[bool]]:
"""
Sets whether the host should get a public IP address on creation. Can be either `true` or `false`.
"""
return pulumi.get(self, "assign_public_ip")
@assign_public_ip.setter
def assign_public_ip(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "assign_public_ip", value)
@property
@pulumi.getter
def fqdn(self) -> Optional[pulumi.Input[str]]:
"""
The fully qualified domain name of the host.
"""
return pulumi.get(self, "fqdn")
@fqdn.setter
def fqdn(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "fqdn", value)
@property
@pulumi.getter(name="subnetId")
def subnet_id(self) -> Optional[pulumi.Input[str]]:
"""
The ID of the subnet, to which the host belongs. The subnet must
be a part of the network to which the cluster belongs.
"""
return pulumi.get(self, "subnet_id")
@subnet_id.setter
def subnet_id(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "subnet_id", value)
@pulumi.input_type
class MdbGreenplumClusterAccessArgs:
def __init__(__self__, *,
data_lens: Optional[pulumi.Input[bool]] = None,
web_sql: Optional[pulumi.Input[bool]] = None):
"""
:param pulumi.Input[bool] data_lens: Allow access for [Yandex DataLens](https://cloud.yandex.com/services/datalens).
:param pulumi.Input[bool] web_sql: Allows access for SQL queries in the management console
"""
if data_lens is not None:
pulumi.set(__self__, "data_lens", data_lens)
if web_sql is not None:
pulumi.set(__self__, "web_sql", web_sql)
@property
@pulumi.getter(name="dataLens")
def data_lens(self) -> Optional[pulumi.Input[bool]]:
"""
Allow access for [Yandex DataLens](https://cloud.yandex.com/services/datalens).
"""
return pulumi.get(self, "data_lens")
@data_lens.setter
def data_lens(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "data_lens", value)
@property
@pulumi.getter(name="webSql")
def web_sql(self) -> Optional[pulumi.Input[bool]]:
"""
Allows access for SQL queries in the management console
"""
return pulumi.get(self, "web_sql")
@web_sql.setter
def web_sql(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "web_sql", value)
@pulumi.input_type
class MdbGreenplumClusterBackupWindowStartArgs:
def __init__(__self__, *,
hours: Optional[pulumi.Input[int]] = None,
minutes: Optional[pulumi.Input[int]] = None):
"""
:param pulumi.Input[int] hours: The hour at which backup will be started (UTC).
:param pulumi.Input[int] minutes: The minute at which backup will be started (UTC).
"""
if hours is not None:
pulumi.set(__self__, "hours", hours)
if minutes is not None:
pulumi.set(__self__, "minutes", minutes)
@property
@pulumi.getter
def hours(self) -> Optional[pulumi.Input[int]]:
"""
The hour at which backup will be started (UTC).
"""
return pulumi.get(self, "hours")
@hours.setter
def hours(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "hours", value)
@property
@pulumi.getter
def minutes(self) -> Optional[pulumi.Input[int]]:
"""
The minute at which backup will be started (UTC).
"""
return pulumi.get(self, "minutes")
@minutes.setter
def minutes(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "minutes", value)
@pulumi.input_type
class MdbGreenplumClusterMasterHostArgs:
def __init__(__self__, *,
assign_public_ip: Optional[pulumi.Input[bool]] = None,
fqdn: Optional[pulumi.Input[str]] = None):
"""
:param pulumi.Input[bool] assign_public_ip: Sets whether the master hosts should get a public IP address on creation. Changing this parameter for an existing host is not supported at the moment.
:param pulumi.Input[str] fqdn: (Computed) The fully qualified domain name of the host.
"""
if assign_public_ip is not None:
pulumi.set(__self__, "assign_public_ip", assign_public_ip)
if fqdn is not None:
pulumi.set(__self__, "fqdn", fqdn)
@property
@pulumi.getter(name="assignPublicIp")
def assign_public_ip(self) -> Optional[pulumi.Input[bool]]:
"""
Sets whether the master hosts should get a public IP address on creation. Changing this parameter for an existing host is not supported at the moment.
"""
return pulumi.get(self, "assign_public_ip")
@assign_public_ip.setter
def assign_public_ip(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "assign_public_ip", value)
@property
@pulumi.getter
def fqdn(self) -> Optional[pulumi.Input[str]]:
"""
(Computed) The fully qualified domain name of the host.
"""
return pulumi.get(self, "fqdn")
@fqdn.setter
def fqdn(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "fqdn", value)
@pulumi.input_type
class MdbGreenplumClusterMasterSubclusterArgs:
def __init__(__self__, *,
resources: pulumi.Input['MdbGreenplumClusterMasterSubclusterResourcesArgs']):
"""
:param pulumi.Input['MdbGreenplumClusterMasterSubclusterResourcesArgs'] resources: Resources allocated to hosts for segment subcluster of the Greenplum cluster. The structure is documented below.
"""
pulumi.set(__self__, "resources", resources)
@property
@pulumi.getter
def resources(self) -> pulumi.Input['MdbGreenplumClusterMasterSubclusterResourcesArgs']:
"""
Resources allocated to hosts for segment subcluster of the Greenplum cluster. The structure is documented below.
"""
return pulumi.get(self, "resources")
@resources.setter
def resources(self, value: pulumi.Input['MdbGreenplumClusterMasterSubclusterResourcesArgs']):
pulumi.set(self, "resources", value)
@pulumi.input_type
class MdbGreenplumClusterMasterSubclusterResourcesArgs:
def __init__(__self__, *,
disk_size: pulumi.Input[int],
disk_type_id: pulumi.Input[str],
resource_preset_id: pulumi.Input[str]):
pulumi.set(__self__, "disk_size", disk_size)
pulumi.set(__self__, "disk_type_id", disk_type_id)
pulumi.set(__self__, "resource_preset_id", resource_preset_id)
@property
@pulumi.getter(name="diskSize")
def disk_size(self) -> pulumi.Input[int]:
return pulumi.get(self, "disk_size")
@disk_size.setter
def disk_size(self, value: pulumi.Input[int]):
pulumi.set(self, "disk_size", value)
@property
@pulumi.getter(name="diskTypeId")
def disk_type_id(self) -> pulumi.Input[str]:
return pulumi.get(self, "disk_type_id")
@disk_type_id.setter
def disk_type_id(self, value: pulumi.Input[str]):
pulumi.set(self, "disk_type_id", value)
@property
@pulumi.getter(name="resourcePresetId")
def resource_preset_id(self) -> pulumi.Input[str]:
return pulumi.get(self, "resource_preset_id")
@resource_preset_id.setter
def resource_preset_id(self, value: pulumi.Input[str]):
pulumi.set(self, "resource_preset_id", value)
@pulumi.input_type
class MdbGreenplumClusterSegmentHostArgs:
def __init__(__self__, *,
fqdn: Optional[pulumi.Input[str]] = None):
"""
:param pulumi.Input[str] fqdn: (Computed) The fully qualified domain name of the host.
"""
if fqdn is not None:
pulumi.set(__self__, "fqdn", fqdn)
@property
@pulumi.getter
def fqdn(self) -> Optional[pulumi.Input[str]]:
"""
(Computed) The fully qualified domain name of the host.
"""
return pulumi.get(self, "fqdn")
@fqdn.setter
def fqdn(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "fqdn", value)
@pulumi.input_type
class MdbGreenplumClusterSegmentSubclusterArgs:
def __init__(__self__, *,
resources: pulumi.Input['MdbGreenplumClusterSegmentSubclusterResourcesArgs']):
"""
:param pulumi.Input['MdbGreenplumClusterSegmentSubclusterResourcesArgs'] resources: Resources allocated to hosts for segment subcluster of the Greenplum cluster. The structure is documented below.
"""
pulumi.set(__self__, "resources", resources)
@property
@pulumi.getter
def resources(self) -> pulumi.Input['MdbGreenplumClusterSegmentSubclusterResourcesArgs']:
"""
Resources allocated to hosts for segment subcluster of the Greenplum cluster. The structure is documented below.
"""
return pulumi.get(self, "resources")
@resources.setter
def resources(self, value: pulumi.Input['MdbGreenplumClusterSegmentSubclusterResourcesArgs']):
pulumi.set(self, "resources", value)
@pulumi.input_type
class MdbGreenplumClusterSegmentSubclusterResourcesArgs:
def __init__(__self__, *,
disk_size: pulumi.Input[int],
disk_type_id: pulumi.Input[str],
resource_preset_id: pulumi.Input[str]):
pulumi.set(__self__, "disk_size", disk_size)
pulumi.set(__self__, "disk_type_id", disk_type_id)
pulumi.set(__self__, "resource_preset_id", resource_preset_id)
@property
@pulumi.getter(name="diskSize")
def disk_size(self) -> pulumi.Input[int]:
return pulumi.get(self, "disk_size")
@disk_size.setter
def disk_size(self, value: pulumi.Input[int]):
pulumi.set(self, "disk_size", value)
@property
@pulumi.getter(name="diskTypeId")
def disk_type_id(self) -> pulumi.Input[str]:
return pulumi.get(self, "disk_type_id")
@disk_type_id.setter
def disk_type_id(self, value: pulumi.Input[str]):
pulumi.set(self, "disk_type_id", value)
@property
@pulumi.getter(name="resourcePresetId")
def resource_preset_id(self) -> pulumi.Input[str]:
return pulumi.get(self, "resource_preset_id")
@resource_preset_id.setter
def resource_preset_id(self, value: pulumi.Input[str]):
pulumi.set(self, "resource_preset_id", value)
@pulumi.input_type
class MdbKafkaClusterConfigArgs:
def __init__(__self__, *,
kafka: pulumi.Input['MdbKafkaClusterConfigKafkaArgs'],
version: pulumi.Input[str],
zones: pulumi.Input[Sequence[pulumi.Input[str]]],
assign_public_ip: Optional[pulumi.Input[bool]] = None,
brokers_count: Optional[pulumi.Input[int]] = None,
schema_registry: Optional[pulumi.Input[bool]] = None,
unmanaged_topics: Optional[pulumi.Input[bool]] = None,
zookeeper: Optional[pulumi.Input['MdbKafkaClusterConfigZookeeperArgs']] = None):
"""
:param pulumi.Input['MdbKafkaClusterConfigKafkaArgs'] kafka: Configuration of the Kafka subcluster. The structure is documented below.
:param pulumi.Input[str] version: Version of the Kafka server software.
:param pulumi.Input[Sequence[pulumi.Input[str]]] zones: List of availability zones.
:param pulumi.Input[bool] assign_public_ip: Determines whether each broker will be assigned a public IP address. The default is `false`.
:param pulumi.Input[int] brokers_count: Count of brokers per availability zone. The default is `1`.
:param pulumi.Input[bool] schema_registry: Enables managed schema registry on cluster. The default is `false`.
:param pulumi.Input[bool] unmanaged_topics: Allows to use Kafka AdminAPI to manage topics. The default is `false`.
:param pulumi.Input['MdbKafkaClusterConfigZookeeperArgs'] zookeeper: Configuration of the ZooKeeper subcluster. The structure is documented below.
"""
pulumi.set(__self__, "kafka", kafka)
pulumi.set(__self__, "version", version)
pulumi.set(__self__, "zones", zones)
if assign_public_ip is not None:
pulumi.set(__self__, "assign_public_ip", assign_public_ip)
if brokers_count is not None:
pulumi.set(__self__, "brokers_count", brokers_count)
if schema_registry is not None:
pulumi.set(__self__, "schema_registry", schema_registry)
if unmanaged_topics is not None:
pulumi.set(__self__, "unmanaged_topics", unmanaged_topics)
if zookeeper is not None:
pulumi.set(__self__, "zookeeper", zookeeper)
@property
@pulumi.getter
def kafka(self) -> pulumi.Input['MdbKafkaClusterConfigKafkaArgs']:
"""
Configuration of the Kafka subcluster. The structure is documented below.
"""
return pulumi.get(self, "kafka")
@kafka.setter
def kafka(self, value: pulumi.Input['MdbKafkaClusterConfigKafkaArgs']):
pulumi.set(self, "kafka", value)
@property
@pulumi.getter
def version(self) -> pulumi.Input[str]:
"""
Version of the Kafka server software.
"""
return pulumi.get(self, "version")
@version.setter
def version(self, value: pulumi.Input[str]):
pulumi.set(self, "version", value)
@property
@pulumi.getter
def zones(self) -> pulumi.Input[Sequence[pulumi.Input[str]]]:
"""
List of availability zones.
"""
return pulumi.get(self, "zones")
@zones.setter
def zones(self, value: pulumi.Input[Sequence[pulumi.Input[str]]]):
pulumi.set(self, "zones", value)
@property
@pulumi.getter(name="assignPublicIp")
def assign_public_ip(self) -> Optional[pulumi.Input[bool]]:
"""
Determines whether each broker will be assigned a public IP address. The default is `false`.
"""
return pulumi.get(self, "assign_public_ip")
@assign_public_ip.setter
def assign_public_ip(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "assign_public_ip", value)
@property
@pulumi.getter(name="brokersCount")
def brokers_count(self) -> Optional[pulumi.Input[int]]:
"""
Count of brokers per availability zone. The default is `1`.
"""
return pulumi.get(self, "brokers_count")
@brokers_count.setter
def brokers_count(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "brokers_count", value)
@property
@pulumi.getter(name="schemaRegistry")
def schema_registry(self) -> Optional[pulumi.Input[bool]]:
"""
Enables managed schema registry on cluster. The default is `false`.
"""
return pulumi.get(self, "schema_registry")
@schema_registry.setter
def schema_registry(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "schema_registry", value)
@property
@pulumi.getter(name="unmanagedTopics")
def unmanaged_topics(self) -> Optional[pulumi.Input[bool]]:
"""
Allows to use Kafka AdminAPI to manage topics. The default is `false`.
"""
return pulumi.get(self, "unmanaged_topics")
@unmanaged_topics.setter
def unmanaged_topics(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "unmanaged_topics", value)
@property
@pulumi.getter
def zookeeper(self) -> Optional[pulumi.Input['MdbKafkaClusterConfigZookeeperArgs']]:
"""
Configuration of the ZooKeeper subcluster. The structure is documented below.
"""
return pulumi.get(self, "zookeeper")
@zookeeper.setter
def zookeeper(self, value: Optional[pulumi.Input['MdbKafkaClusterConfigZookeeperArgs']]):
pulumi.set(self, "zookeeper", value)
@pulumi.input_type
class MdbKafkaClusterConfigKafkaArgs:
def __init__(__self__, *,
resources: pulumi.Input['MdbKafkaClusterConfigKafkaResourcesArgs'],
kafka_config: Optional[pulumi.Input['MdbKafkaClusterConfigKafkaKafkaConfigArgs']] = None):
"""
:param pulumi.Input['MdbKafkaClusterConfigKafkaResourcesArgs'] resources: Resources allocated to hosts of the ZooKeeper subcluster. The structure is documented below.
:param pulumi.Input['MdbKafkaClusterConfigKafkaKafkaConfigArgs'] kafka_config: User-defined settings for the Kafka cluster. The structure is documented below.
"""
pulumi.set(__self__, "resources", resources)
if kafka_config is not None:
pulumi.set(__self__, "kafka_config", kafka_config)
@property
@pulumi.getter
def resources(self) -> pulumi.Input['MdbKafkaClusterConfigKafkaResourcesArgs']:
"""
Resources allocated to hosts of the ZooKeeper subcluster. The structure is documented below.
"""
return pulumi.get(self, "resources")
@resources.setter
def resources(self, value: pulumi.Input['MdbKafkaClusterConfigKafkaResourcesArgs']):
pulumi.set(self, "resources", value)
@property
@pulumi.getter(name="kafkaConfig")
def kafka_config(self) -> Optional[pulumi.Input['MdbKafkaClusterConfigKafkaKafkaConfigArgs']]:
"""
User-defined settings for the Kafka cluster. The structure is documented below.
"""
return pulumi.get(self, "kafka_config")
@kafka_config.setter
def kafka_config(self, value: Optional[pulumi.Input['MdbKafkaClusterConfigKafkaKafkaConfigArgs']]):
pulumi.set(self, "kafka_config", value)
@pulumi.input_type
class MdbKafkaClusterConfigKafkaKafkaConfigArgs:
def __init__(__self__, *,
auto_create_topics_enable: Optional[pulumi.Input[bool]] = None,
compression_type: Optional[pulumi.Input[str]] = None,
default_replication_factor: Optional[pulumi.Input[str]] = None,
log_flush_interval_messages: Optional[pulumi.Input[str]] = None,
log_flush_interval_ms: Optional[pulumi.Input[str]] = None,
log_flush_scheduler_interval_ms: Optional[pulumi.Input[str]] = None,
log_preallocate: Optional[pulumi.Input[bool]] = None,
log_retention_bytes: Optional[pulumi.Input[str]] = None,
log_retention_hours: Optional[pulumi.Input[str]] = None,
log_retention_minutes: Optional[pulumi.Input[str]] = None,
log_retention_ms: Optional[pulumi.Input[str]] = None,
log_segment_bytes: Optional[pulumi.Input[str]] = None,
num_partitions: Optional[pulumi.Input[str]] = None,
socket_receive_buffer_bytes: Optional[pulumi.Input[str]] = None,
socket_send_buffer_bytes: Optional[pulumi.Input[str]] = None):
if auto_create_topics_enable is not None:
pulumi.set(__self__, "auto_create_topics_enable", auto_create_topics_enable)
if compression_type is not None:
pulumi.set(__self__, "compression_type", compression_type)
if default_replication_factor is not None:
pulumi.set(__self__, "default_replication_factor", default_replication_factor)
if log_flush_interval_messages is not None:
pulumi.set(__self__, "log_flush_interval_messages", log_flush_interval_messages)
if log_flush_interval_ms is not None:
pulumi.set(__self__, "log_flush_interval_ms", log_flush_interval_ms)
if log_flush_scheduler_interval_ms is not None:
pulumi.set(__self__, "log_flush_scheduler_interval_ms", log_flush_scheduler_interval_ms)
if log_preallocate is not None:
pulumi.set(__self__, "log_preallocate", log_preallocate)
if log_retention_bytes is not None:
pulumi.set(__self__, "log_retention_bytes", log_retention_bytes)
if log_retention_hours is not None:
pulumi.set(__self__, "log_retention_hours", log_retention_hours)
if log_retention_minutes is not None:
pulumi.set(__self__, "log_retention_minutes", log_retention_minutes)
if log_retention_ms is not None:
pulumi.set(__self__, "log_retention_ms", log_retention_ms)
if log_segment_bytes is not None:
pulumi.set(__self__, "log_segment_bytes", log_segment_bytes)
if num_partitions is not None:
pulumi.set(__self__, "num_partitions", num_partitions)
if socket_receive_buffer_bytes is not None:
pulumi.set(__self__, "socket_receive_buffer_bytes", socket_receive_buffer_bytes)
if socket_send_buffer_bytes is not None:
pulumi.set(__self__, "socket_send_buffer_bytes", socket_send_buffer_bytes)
@property
@pulumi.getter(name="autoCreateTopicsEnable")
def auto_create_topics_enable(self) -> Optional[pulumi.Input[bool]]:
return pulumi.get(self, "auto_create_topics_enable")
@auto_create_topics_enable.setter
def auto_create_topics_enable(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "auto_create_topics_enable", value)
@property
@pulumi.getter(name="compressionType")
def compression_type(self) -> Optional[pulumi.Input[str]]:
return pulumi.get(self, "compression_type")
@compression_type.setter
def compression_type(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "compression_type", value)
@property
@pulumi.getter(name="defaultReplicationFactor")
def default_replication_factor(self) -> Optional[pulumi.Input[str]]:
return pulumi.get(self, "default_replication_factor")
@default_replication_factor.setter
def default_replication_factor(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "default_replication_factor", value)
@property
@pulumi.getter(name="logFlushIntervalMessages")
def log_flush_interval_messages(self) -> Optional[pulumi.Input[str]]:
return pulumi.get(self, "log_flush_interval_messages")
@log_flush_interval_messages.setter
def log_flush_interval_messages(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "log_flush_interval_messages", value)
@property
@pulumi.getter(name="logFlushIntervalMs")
def log_flush_interval_ms(self) -> Optional[pulumi.Input[str]]:
return pulumi.get(self, "log_flush_interval_ms")
@log_flush_interval_ms.setter
def log_flush_interval_ms(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "log_flush_interval_ms", value)
@property
@pulumi.getter(name="logFlushSchedulerIntervalMs")
def log_flush_scheduler_interval_ms(self) -> Optional[pulumi.Input[str]]:
return pulumi.get(self, "log_flush_scheduler_interval_ms")
@log_flush_scheduler_interval_ms.setter
def log_flush_scheduler_interval_ms(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "log_flush_scheduler_interval_ms", value)
@property
@pulumi.getter(name="logPreallocate")
def log_preallocate(self) -> Optional[pulumi.Input[bool]]:
return pulumi.get(self, "log_preallocate")
@log_preallocate.setter
def log_preallocate(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "log_preallocate", value)
@property
@pulumi.getter(name="logRetentionBytes")
def log_retention_bytes(self) -> Optional[pulumi.Input[str]]:
return pulumi.get(self, "log_retention_bytes")
@log_retention_bytes.setter
def log_retention_bytes(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "log_retention_bytes", value)
@property
@pulumi.getter(name="logRetentionHours")
def log_retention_hours(self) -> Optional[pulumi.Input[str]]:
return pulumi.get(self, "log_retention_hours")
@log_retention_hours.setter
def log_retention_hours(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "log_retention_hours", value)
@property
@pulumi.getter(name="logRetentionMinutes")
def log_retention_minutes(self) -> Optional[pulumi.Input[str]]:
return pulumi.get(self, "log_retention_minutes")
@log_retention_minutes.setter
def log_retention_minutes(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "log_retention_minutes", value)
@property
@pulumi.getter(name="logRetentionMs")
def log_retention_ms(self) -> Optional[pulumi.Input[str]]:
return pulumi.get(self, "log_retention_ms")
@log_retention_ms.setter
def log_retention_ms(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "log_retention_ms", value)
@property
@pulumi.getter(name="logSegmentBytes")
def log_segment_bytes(self) -> Optional[pulumi.Input[str]]:
return pulumi.get(self, "log_segment_bytes")
@log_segment_bytes.setter
def log_segment_bytes(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "log_segment_bytes", value)
@property
@pulumi.getter(name="numPartitions")
def num_partitions(self) -> Optional[pulumi.Input[str]]:
return pulumi.get(self, "num_partitions")
@num_partitions.setter
def num_partitions(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "num_partitions", value)
@property
@pulumi.getter(name="socketReceiveBufferBytes")
def socket_receive_buffer_bytes(self) -> Optional[pulumi.Input[str]]:
return pulumi.get(self, "socket_receive_buffer_bytes")
@socket_receive_buffer_bytes.setter
def socket_receive_buffer_bytes(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "socket_receive_buffer_bytes", value)
@property
@pulumi.getter(name="socketSendBufferBytes")
def socket_send_buffer_bytes(self) -> Optional[pulumi.Input[str]]:
return pulumi.get(self, "socket_send_buffer_bytes")
@socket_send_buffer_bytes.setter
def socket_send_buffer_bytes(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "socket_send_buffer_bytes", value)
@pulumi.input_type
class MdbKafkaClusterConfigKafkaResourcesArgs:
def __init__(__self__, *,
disk_size: pulumi.Input[int],
disk_type_id: pulumi.Input[str],
resource_preset_id: pulumi.Input[str]):
"""
:param pulumi.Input[int] disk_size: Volume of the storage available to a ZooKeeper host, in gigabytes.
:param pulumi.Input[str] disk_type_id: Type of the storage of ZooKeeper hosts.
For more information see [the official documentation](https://cloud.yandex.com/docs/managed-kafka/concepts/storage).
"""
pulumi.set(__self__, "disk_size", disk_size)
pulumi.set(__self__, "disk_type_id", disk_type_id)
pulumi.set(__self__, "resource_preset_id", resource_preset_id)
@property
@pulumi.getter(name="diskSize")
def disk_size(self) -> pulumi.Input[int]:
"""
Volume of the storage available to a ZooKeeper host, in gigabytes.
"""
return pulumi.get(self, "disk_size")
@disk_size.setter
def disk_size(self, value: pulumi.Input[int]):
pulumi.set(self, "disk_size", value)
@property
@pulumi.getter(name="diskTypeId")
def disk_type_id(self) -> pulumi.Input[str]:
"""
Type of the storage of ZooKeeper hosts.
For more information see [the official documentation](https://cloud.yandex.com/docs/managed-kafka/concepts/storage).
"""
return pulumi.get(self, "disk_type_id")
@disk_type_id.setter
def disk_type_id(self, value: pulumi.Input[str]):
pulumi.set(self, "disk_type_id", value)
@property
@pulumi.getter(name="resourcePresetId")
def resource_preset_id(self) -> pulumi.Input[str]:
return pulumi.get(self, "resource_preset_id")
@resource_preset_id.setter
def resource_preset_id(self, value: pulumi.Input[str]):
pulumi.set(self, "resource_preset_id", value)
@pulumi.input_type
class MdbKafkaClusterConfigZookeeperArgs:
def __init__(__self__, *,
resources: Optional[pulumi.Input['MdbKafkaClusterConfigZookeeperResourcesArgs']] = None):
"""
:param pulumi.Input['MdbKafkaClusterConfigZookeeperResourcesArgs'] resources: Resources allocated to hosts of the ZooKeeper subcluster. The structure is documented below.
"""
if resources is not None:
pulumi.set(__self__, "resources", resources)
@property
@pulumi.getter
def resources(self) -> Optional[pulumi.Input['MdbKafkaClusterConfigZookeeperResourcesArgs']]:
"""
Resources allocated to hosts of the ZooKeeper subcluster. The structure is documented below.
"""
return pulumi.get(self, "resources")
@resources.setter
def resources(self, value: Optional[pulumi.Input['MdbKafkaClusterConfigZookeeperResourcesArgs']]):
pulumi.set(self, "resources", value)
@pulumi.input_type
class MdbKafkaClusterConfigZookeeperResourcesArgs:
def __init__(__self__, *,
disk_size: Optional[pulumi.Input[int]] = None,
disk_type_id: Optional[pulumi.Input[str]] = None,
resource_preset_id: Optional[pulumi.Input[str]] = None):
"""
:param pulumi.Input[int] disk_size: Volume of the storage available to a ZooKeeper host, in gigabytes.
:param pulumi.Input[str] disk_type_id: Type of the storage of ZooKeeper hosts.
For more information see [the official documentation](https://cloud.yandex.com/docs/managed-kafka/concepts/storage).
"""
if disk_size is not None:
pulumi.set(__self__, "disk_size", disk_size)
if disk_type_id is not None:
pulumi.set(__self__, "disk_type_id", disk_type_id)
if resource_preset_id is not None:
pulumi.set(__self__, "resource_preset_id", resource_preset_id)
@property
@pulumi.getter(name="diskSize")
def disk_size(self) -> Optional[pulumi.Input[int]]:
"""
Volume of the storage available to a ZooKeeper host, in gigabytes.
"""
return pulumi.get(self, "disk_size")
@disk_size.setter
def disk_size(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "disk_size", value)
@property
@pulumi.getter(name="diskTypeId")
def disk_type_id(self) -> Optional[pulumi.Input[str]]:
"""
Type of the storage of ZooKeeper hosts.
For more information see [the official documentation](https://cloud.yandex.com/docs/managed-kafka/concepts/storage).
"""
return pulumi.get(self, "disk_type_id")
@disk_type_id.setter
def disk_type_id(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "disk_type_id", value)
@property
@pulumi.getter(name="resourcePresetId")
def resource_preset_id(self) -> Optional[pulumi.Input[str]]:
return pulumi.get(self, "resource_preset_id")
@resource_preset_id.setter
def resource_preset_id(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "resource_preset_id", value)
@pulumi.input_type
class MdbKafkaClusterHostArgs:
def __init__(__self__, *,
assign_public_ip: Optional[pulumi.Input[bool]] = None,
health: Optional[pulumi.Input[str]] = None,
name: Optional[pulumi.Input[str]] = None,
role: Optional[pulumi.Input[str]] = None,
subnet_id: Optional[pulumi.Input[str]] = None,
zone_id: Optional[pulumi.Input[str]] = None):
"""
:param pulumi.Input[bool] assign_public_ip: Determines whether each broker will be assigned a public IP address. The default is `false`.
:param pulumi.Input[str] health: Health of the host.
:param pulumi.Input[str] name: The name of the topic.
:param pulumi.Input[str] role: The role type to grant to the topic.
:param pulumi.Input[str] subnet_id: The ID of the subnet, to which the host belongs.
:param pulumi.Input[str] zone_id: The availability zone where the Kafka host was created.
"""
if assign_public_ip is not None:
pulumi.set(__self__, "assign_public_ip", assign_public_ip)
if health is not None:
pulumi.set(__self__, "health", health)
if name is not None:
pulumi.set(__self__, "name", name)
if role is not None:
pulumi.set(__self__, "role", role)
if subnet_id is not None:
pulumi.set(__self__, "subnet_id", subnet_id)
if zone_id is not None:
pulumi.set(__self__, "zone_id", zone_id)
@property
@pulumi.getter(name="assignPublicIp")
def assign_public_ip(self) -> Optional[pulumi.Input[bool]]:
"""
Determines whether each broker will be assigned a public IP address. The default is `false`.
"""
return pulumi.get(self, "assign_public_ip")
@assign_public_ip.setter
def assign_public_ip(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "assign_public_ip", value)
@property
@pulumi.getter
def health(self) -> Optional[pulumi.Input[str]]:
"""
Health of the host.
"""
return pulumi.get(self, "health")
@health.setter
def health(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "health", value)
@property
@pulumi.getter
def name(self) -> Optional[pulumi.Input[str]]:
"""
The name of the topic.
"""
return pulumi.get(self, "name")
@name.setter
def name(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "name", value)
@property
@pulumi.getter
def role(self) -> Optional[pulumi.Input[str]]:
"""
The role type to grant to the topic.
"""
return pulumi.get(self, "role")
@role.setter
def role(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "role", value)
@property
@pulumi.getter(name="subnetId")
def subnet_id(self) -> Optional[pulumi.Input[str]]:
"""
The ID of the subnet, to which the host belongs.
"""
return pulumi.get(self, "subnet_id")
@subnet_id.setter
def subnet_id(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "subnet_id", value)
@property
@pulumi.getter(name="zoneId")
def zone_id(self) -> Optional[pulumi.Input[str]]:
"""
The availability zone where the Kafka host was created.
"""
return pulumi.get(self, "zone_id")
@zone_id.setter
def zone_id(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "zone_id", value)
@pulumi.input_type
class MdbKafkaClusterTopicArgs:
def __init__(__self__, *,
name: pulumi.Input[str],
partitions: pulumi.Input[int],
replication_factor: pulumi.Input[int],
topic_config: Optional[pulumi.Input['MdbKafkaClusterTopicTopicConfigArgs']] = None):
"""
:param pulumi.Input[str] name: The name of the topic.
:param pulumi.Input[int] partitions: The number of the topic's partitions.
:param pulumi.Input[int] replication_factor: Amount of data copies (replicas) for the topic in the cluster.
:param pulumi.Input['MdbKafkaClusterTopicTopicConfigArgs'] topic_config: User-defined settings for the topic. The structure is documented below.
"""
pulumi.set(__self__, "name", name)
pulumi.set(__self__, "partitions", partitions)
pulumi.set(__self__, "replication_factor", replication_factor)
if topic_config is not None:
pulumi.set(__self__, "topic_config", topic_config)
@property
@pulumi.getter
def name(self) -> pulumi.Input[str]:
"""
The name of the topic.
"""
return pulumi.get(self, "name")
@name.setter
def name(self, value: pulumi.Input[str]):
pulumi.set(self, "name", value)
@property
@pulumi.getter
def partitions(self) -> pulumi.Input[int]:
"""
The number of the topic's partitions.
"""
return pulumi.get(self, "partitions")
@partitions.setter
def partitions(self, value: pulumi.Input[int]):
pulumi.set(self, "partitions", value)
@property
@pulumi.getter(name="replicationFactor")
def replication_factor(self) -> pulumi.Input[int]:
"""
Amount of data copies (replicas) for the topic in the cluster.
"""
return pulumi.get(self, "replication_factor")
@replication_factor.setter
def replication_factor(self, value: pulumi.Input[int]):
pulumi.set(self, "replication_factor", value)
@property
@pulumi.getter(name="topicConfig")
def topic_config(self) -> Optional[pulumi.Input['MdbKafkaClusterTopicTopicConfigArgs']]:
"""
User-defined settings for the topic. The structure is documented below.
"""
return pulumi.get(self, "topic_config")
@topic_config.setter
def topic_config(self, value: Optional[pulumi.Input['MdbKafkaClusterTopicTopicConfigArgs']]):
pulumi.set(self, "topic_config", value)
@pulumi.input_type
class MdbKafkaClusterTopicTopicConfigArgs:
def __init__(__self__, *,
cleanup_policy: Optional[pulumi.Input[str]] = None,
compression_type: Optional[pulumi.Input[str]] = None,
delete_retention_ms: Optional[pulumi.Input[str]] = None,
file_delete_delay_ms: Optional[pulumi.Input[str]] = None,
flush_messages: Optional[pulumi.Input[str]] = None,
flush_ms: Optional[pulumi.Input[str]] = None,
max_message_bytes: Optional[pulumi.Input[str]] = None,
min_compaction_lag_ms: Optional[pulumi.Input[str]] = None,
min_insync_replicas: Optional[pulumi.Input[str]] = None,
preallocate: Optional[pulumi.Input[bool]] = None,
retention_bytes: Optional[pulumi.Input[str]] = None,
retention_ms: Optional[pulumi.Input[str]] = None,
segment_bytes: Optional[pulumi.Input[str]] = None):
if cleanup_policy is not None:
pulumi.set(__self__, "cleanup_policy", cleanup_policy)
if compression_type is not None:
pulumi.set(__self__, "compression_type", compression_type)
if delete_retention_ms is not None:
pulumi.set(__self__, "delete_retention_ms", delete_retention_ms)
if file_delete_delay_ms is not None:
pulumi.set(__self__, "file_delete_delay_ms", file_delete_delay_ms)
if flush_messages is not None:
pulumi.set(__self__, "flush_messages", flush_messages)
if flush_ms is not None:
pulumi.set(__self__, "flush_ms", flush_ms)
if max_message_bytes is not None:
pulumi.set(__self__, "max_message_bytes", max_message_bytes)
if min_compaction_lag_ms is not None:
pulumi.set(__self__, "min_compaction_lag_ms", min_compaction_lag_ms)
if min_insync_replicas is not None:
pulumi.set(__self__, "min_insync_replicas", min_insync_replicas)
if preallocate is not None:
pulumi.set(__self__, "preallocate", preallocate)
if retention_bytes is not None:
pulumi.set(__self__, "retention_bytes", retention_bytes)
if retention_ms is not None:
pulumi.set(__self__, "retention_ms", retention_ms)
if segment_bytes is not None:
pulumi.set(__self__, "segment_bytes", segment_bytes)
@property
@pulumi.getter(name="cleanupPolicy")
def cleanup_policy(self) -> Optional[pulumi.Input[str]]:
return pulumi.get(self, "cleanup_policy")
@cleanup_policy.setter
def cleanup_policy(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "cleanup_policy", value)
@property
@pulumi.getter(name="compressionType")
def compression_type(self) -> Optional[pulumi.Input[str]]:
return pulumi.get(self, "compression_type")
@compression_type.setter
def compression_type(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "compression_type", value)
@property
@pulumi.getter(name="deleteRetentionMs")
def delete_retention_ms(self) -> Optional[pulumi.Input[str]]:
return pulumi.get(self, "delete_retention_ms")
@delete_retention_ms.setter
def delete_retention_ms(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "delete_retention_ms", value)
@property
@pulumi.getter(name="fileDeleteDelayMs")
def file_delete_delay_ms(self) -> Optional[pulumi.Input[str]]:
return pulumi.get(self, "file_delete_delay_ms")
@file_delete_delay_ms.setter
def file_delete_delay_ms(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "file_delete_delay_ms", value)
@property
@pulumi.getter(name="flushMessages")
def flush_messages(self) -> Optional[pulumi.Input[str]]:
return pulumi.get(self, "flush_messages")
@flush_messages.setter
def flush_messages(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "flush_messages", value)
@property
@pulumi.getter(name="flushMs")
def flush_ms(self) -> Optional[pulumi.Input[str]]:
return pulumi.get(self, "flush_ms")
@flush_ms.setter
def flush_ms(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "flush_ms", value)
@property
@pulumi.getter(name="maxMessageBytes")
def max_message_bytes(self) -> Optional[pulumi.Input[str]]:
return pulumi.get(self, "max_message_bytes")
@max_message_bytes.setter
def max_message_bytes(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "max_message_bytes", value)
@property
@pulumi.getter(name="minCompactionLagMs")
def min_compaction_lag_ms(self) -> Optional[pulumi.Input[str]]:
return pulumi.get(self, "min_compaction_lag_ms")
@min_compaction_lag_ms.setter
def min_compaction_lag_ms(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "min_compaction_lag_ms", value)
@property
@pulumi.getter(name="minInsyncReplicas")
def min_insync_replicas(self) -> Optional[pulumi.Input[str]]:
return pulumi.get(self, "min_insync_replicas")
@min_insync_replicas.setter
def min_insync_replicas(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "min_insync_replicas", value)
@property
@pulumi.getter
def preallocate(self) -> Optional[pulumi.Input[bool]]:
return pulumi.get(self, "preallocate")
@preallocate.setter
def preallocate(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "preallocate", value)
@property
@pulumi.getter(name="retentionBytes")
def retention_bytes(self) -> Optional[pulumi.Input[str]]:
return pulumi.get(self, "retention_bytes")
@retention_bytes.setter
def retention_bytes(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "retention_bytes", value)
@property
@pulumi.getter(name="retentionMs")
def retention_ms(self) -> Optional[pulumi.Input[str]]:
return pulumi.get(self, "retention_ms")
@retention_ms.setter
def retention_ms(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "retention_ms", value)
@property
@pulumi.getter(name="segmentBytes")
def segment_bytes(self) -> Optional[pulumi.Input[str]]:
return pulumi.get(self, "segment_bytes")
@segment_bytes.setter
def segment_bytes(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "segment_bytes", value)
@pulumi.input_type
class MdbKafkaClusterUserArgs:
def __init__(__self__, *,
name: pulumi.Input[str],
password: pulumi.Input[str],
permissions: Optional[pulumi.Input[Sequence[pulumi.Input['MdbKafkaClusterUserPermissionArgs']]]] = None):
"""
:param pulumi.Input[str] name: The name of the topic.
:param pulumi.Input[str] password: The <PASSWORD> the <PASSWORD>.
:param pulumi.Input[Sequence[pulumi.Input['MdbKafkaClusterUserPermissionArgs']]] permissions: Set of permissions granted to the user. The structure is documented below.
"""
pulumi.set(__self__, "name", name)
pulumi.set(__self__, "password", password)
if permissions is not None:
pulumi.set(__self__, "permissions", permissions)
@property
@pulumi.getter
def name(self) -> pulumi.Input[str]:
"""
The name of the topic.
"""
return pulumi.get(self, "name")
@name.setter
def name(self, value: pulumi.Input[str]):
pulumi.set(self, "name", value)
@property
@pulumi.getter
def password(self) -> pulumi.Input[str]:
"""
The password of the user.
"""
return pulumi.get(self, "password")
@password.setter
def password(self, value: pulumi.Input[str]):
pulumi.set(self, "password", value)
@property
@pulumi.getter
def permissions(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['MdbKafkaClusterUserPermissionArgs']]]]:
"""
Set of permissions granted to the user. The structure is documented below.
"""
return pulumi.get(self, "permissions")
@permissions.setter
def permissions(self, value: Optional[pulumi.Input[Sequence[pulumi.Input['MdbKafkaClusterUserPermissionArgs']]]]):
pulumi.set(self, "permissions", value)
@pulumi.input_type
class MdbKafkaClusterUserPermissionArgs:
def __init__(__self__, *,
role: pulumi.Input[str],
topic_name: pulumi.Input[str]):
"""
:param pulumi.Input[str] role: The role type to grant to the topic.
:param pulumi.Input[str] topic_name: The name of the topic that the permission grants access to.
"""
pulumi.set(__self__, "role", role)
pulumi.set(__self__, "topic_name", topic_name)
@property
@pulumi.getter
def role(self) -> pulumi.Input[str]:
"""
The role type to grant to the topic.
"""
return pulumi.get(self, "role")
@role.setter
def role(self, value: pulumi.Input[str]):
pulumi.set(self, "role", value)
@property
@pulumi.getter(name="topicName")
def topic_name(self) -> pulumi.Input[str]:
"""
The name of the topic that the permission grants access to.
"""
return pulumi.get(self, "topic_name")
@topic_name.setter
def topic_name(self, value: pulumi.Input[str]):
pulumi.set(self, "topic_name", value)
@pulumi.input_type
class MdbKafkaTopicTopicConfigArgs:
def __init__(__self__, *,
cleanup_policy: Optional[pulumi.Input[str]] = None,
compression_type: Optional[pulumi.Input[str]] = None,
delete_retention_ms: Optional[pulumi.Input[str]] = None,
file_delete_delay_ms: Optional[pulumi.Input[str]] = None,
flush_messages: Optional[pulumi.Input[str]] = None,
flush_ms: Optional[pulumi.Input[str]] = None,
max_message_bytes: Optional[pulumi.Input[str]] = None,
min_compaction_lag_ms: Optional[pulumi.Input[str]] = None,
min_insync_replicas: Optional[pulumi.Input[str]] = None,
preallocate: Optional[pulumi.Input[bool]] = None,
retention_bytes: Optional[pulumi.Input[str]] = None,
retention_ms: Optional[pulumi.Input[str]] = None,
segment_bytes: Optional[pulumi.Input[str]] = None):
if cleanup_policy is not None:
pulumi.set(__self__, "cleanup_policy", cleanup_policy)
if compression_type is not None:
pulumi.set(__self__, "compression_type", compression_type)
if delete_retention_ms is not None:
pulumi.set(__self__, "delete_retention_ms", delete_retention_ms)
if file_delete_delay_ms is not None:
pulumi.set(__self__, "file_delete_delay_ms", file_delete_delay_ms)
if flush_messages is not None:
pulumi.set(__self__, "flush_messages", flush_messages)
if flush_ms is not None:
pulumi.set(__self__, "flush_ms", flush_ms)
if max_message_bytes is not None:
pulumi.set(__self__, "max_message_bytes", max_message_bytes)
if min_compaction_lag_ms is not None:
pulumi.set(__self__, "min_compaction_lag_ms", min_compaction_lag_ms)
if min_insync_replicas is not None:
pulumi.set(__self__, "min_insync_replicas", min_insync_replicas)
if preallocate is not None:
pulumi.set(__self__, "preallocate", preallocate)
if retention_bytes is not None:
pulumi.set(__self__, "retention_bytes", retention_bytes)
if retention_ms is not None:
pulumi.set(__self__, "retention_ms", retention_ms)
if segment_bytes is not None:
pulumi.set(__self__, "segment_bytes", segment_bytes)
@property
@pulumi.getter(name="cleanupPolicy")
def cleanup_policy(self) -> Optional[pulumi.Input[str]]:
return pulumi.get(self, "cleanup_policy")
@cleanup_policy.setter
def cleanup_policy(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "cleanup_policy", value)
@property
@pulumi.getter(name="compressionType")
def compression_type(self) -> Optional[pulumi.Input[str]]:
return pulumi.get(self, "compression_type")
@compression_type.setter
def compression_type(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "compression_type", value)
@property
@pulumi.getter(name="deleteRetentionMs")
def delete_retention_ms(self) -> Optional[pulumi.Input[str]]:
return pulumi.get(self, "delete_retention_ms")
@delete_retention_ms.setter
def delete_retention_ms(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "delete_retention_ms", value)
@property
@pulumi.getter(name="fileDeleteDelayMs")
def file_delete_delay_ms(self) -> Optional[pulumi.Input[str]]:
return pulumi.get(self, "file_delete_delay_ms")
@file_delete_delay_ms.setter
def file_delete_delay_ms(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "file_delete_delay_ms", value)
@property
@pulumi.getter(name="flushMessages")
def flush_messages(self) -> Optional[pulumi.Input[str]]:
return pulumi.get(self, "flush_messages")
@flush_messages.setter
def flush_messages(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "flush_messages", value)
@property
@pulumi.getter(name="flushMs")
def flush_ms(self) -> Optional[pulumi.Input[str]]:
return pulumi.get(self, "flush_ms")
@flush_ms.setter
def flush_ms(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "flush_ms", value)
@property
@pulumi.getter(name="maxMessageBytes")
def max_message_bytes(self) -> Optional[pulumi.Input[str]]:
return pulumi.get(self, "max_message_bytes")
@max_message_bytes.setter
def max_message_bytes(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "max_message_bytes", value)
@property
@pulumi.getter(name="minCompactionLagMs")
def min_compaction_lag_ms(self) -> Optional[pulumi.Input[str]]:
return pulumi.get(self, "min_compaction_lag_ms")
@min_compaction_lag_ms.setter
def min_compaction_lag_ms(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "min_compaction_lag_ms", value)
@property
@pulumi.getter(name="minInsyncReplicas")
def min_insync_replicas(self) -> Optional[pulumi.Input[str]]:
return pulumi.get(self, "min_insync_replicas")
@min_insync_replicas.setter
def min_insync_replicas(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "min_insync_replicas", value)
@property
@pulumi.getter
def preallocate(self) -> Optional[pulumi.Input[bool]]:
return pulumi.get(self, "preallocate")
@preallocate.setter
def preallocate(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "preallocate", value)
@property
@pulumi.getter(name="retentionBytes")
def retention_bytes(self) -> Optional[pulumi.Input[str]]:
return pulumi.get(self, "retention_bytes")
@retention_bytes.setter
def retention_bytes(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "retention_bytes", value)
@property
@pulumi.getter(name="retentionMs")
def retention_ms(self) -> Optional[pulumi.Input[str]]:
return pulumi.get(self, "retention_ms")
@retention_ms.setter
def retention_ms(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "retention_ms", value)
@property
@pulumi.getter(name="segmentBytes")
def segment_bytes(self) -> Optional[pulumi.Input[str]]:
return pulumi.get(self, "segment_bytes")
@segment_bytes.setter
def segment_bytes(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "segment_bytes", value)
@pulumi.input_type
class MdbMongodbClusterClusterConfigArgs:
def __init__(__self__, *,
version: pulumi.Input[str],
access: Optional[pulumi.Input['MdbMongodbClusterClusterConfigAccessArgs']] = None,
backup_window_start: Optional[pulumi.Input['MdbMongodbClusterClusterConfigBackupWindowStartArgs']] = None,
feature_compatibility_version: Optional[pulumi.Input[str]] = None):
"""
:param pulumi.Input[str] version: Version of MongoDB (either 5.0, 4.4, 4.2 or 4.0).
:param pulumi.Input['MdbMongodbClusterClusterConfigAccessArgs'] access: Shows whether cluster has access to data lens. The structure is documented below.
:param pulumi.Input['MdbMongodbClusterClusterConfigBackupWindowStartArgs'] backup_window_start: Time to start the daily backup, in the UTC timezone. The structure is documented below.
:param pulumi.Input[str] feature_compatibility_version: Feature compatibility version of MongoDB. If not provided version is taken. Can be either `5.0`, `4.4`, `4.2` and `4.0`.
"""
pulumi.set(__self__, "version", version)
if access is not None:
pulumi.set(__self__, "access", access)
if backup_window_start is not None:
pulumi.set(__self__, "backup_window_start", backup_window_start)
if feature_compatibility_version is not None:
pulumi.set(__self__, "feature_compatibility_version", feature_compatibility_version)
@property
@pulumi.getter
def version(self) -> pulumi.Input[str]:
"""
Version of MongoDB (either 5.0, 4.4, 4.2 or 4.0).
"""
return pulumi.get(self, "version")
@version.setter
def version(self, value: pulumi.Input[str]):
pulumi.set(self, "version", value)
@property
@pulumi.getter
def access(self) -> Optional[pulumi.Input['MdbMongodbClusterClusterConfigAccessArgs']]:
"""
Shows whether cluster has access to data lens. The structure is documented below.
"""
return pulumi.get(self, "access")
@access.setter
def access(self, value: Optional[pulumi.Input['MdbMongodbClusterClusterConfigAccessArgs']]):
pulumi.set(self, "access", value)
@property
@pulumi.getter(name="backupWindowStart")
def backup_window_start(self) -> Optional[pulumi.Input['MdbMongodbClusterClusterConfigBackupWindowStartArgs']]:
"""
Time to start the daily backup, in the UTC timezone. The structure is documented below.
"""
return pulumi.get(self, "backup_window_start")
@backup_window_start.setter
def backup_window_start(self, value: Optional[pulumi.Input['MdbMongodbClusterClusterConfigBackupWindowStartArgs']]):
pulumi.set(self, "backup_window_start", value)
@property
@pulumi.getter(name="featureCompatibilityVersion")
def feature_compatibility_version(self) -> Optional[pulumi.Input[str]]:
"""
Feature compatibility version of MongoDB. If not provided version is taken. Can be either `5.0`, `4.4`, `4.2` and `4.0`.
"""
return pulumi.get(self, "feature_compatibility_version")
@feature_compatibility_version.setter
def feature_compatibility_version(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "feature_compatibility_version", value)
@pulumi.input_type
class MdbMongodbClusterClusterConfigAccessArgs:
def __init__(__self__, *,
data_lens: Optional[pulumi.Input[bool]] = None):
"""
:param pulumi.Input[bool] data_lens: Allow access for DataLens.
"""
if data_lens is not None:
pulumi.set(__self__, "data_lens", data_lens)
@property
@pulumi.getter(name="dataLens")
def data_lens(self) -> Optional[pulumi.Input[bool]]:
"""
Allow access for DataLens.
"""
return pulumi.get(self, "data_lens")
@data_lens.setter
def data_lens(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "data_lens", value)
@pulumi.input_type
class MdbMongodbClusterClusterConfigBackupWindowStartArgs:
def __init__(__self__, *,
hours: Optional[pulumi.Input[int]] = None,
minutes: Optional[pulumi.Input[int]] = None):
"""
:param pulumi.Input[int] hours: The hour at which backup will be started.
:param pulumi.Input[int] minutes: The minute at which backup will be started.
"""
if hours is not None:
pulumi.set(__self__, "hours", hours)
if minutes is not None:
pulumi.set(__self__, "minutes", minutes)
@property
@pulumi.getter
def hours(self) -> Optional[pulumi.Input[int]]:
"""
The hour at which backup will be started.
"""
return pulumi.get(self, "hours")
@hours.setter
def hours(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "hours", value)
@property
@pulumi.getter
def minutes(self) -> Optional[pulumi.Input[int]]:
"""
The minute at which backup will be started.
"""
return pulumi.get(self, "minutes")
@minutes.setter
def minutes(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "minutes", value)
@pulumi.input_type
class MdbMongodbClusterDatabaseArgs:
def __init__(__self__, *,
name: pulumi.Input[str]):
"""
:param pulumi.Input[str] name: The fully qualified domain name of the host. Computed on server side.
"""
pulumi.set(__self__, "name", name)
@property
@pulumi.getter
def name(self) -> pulumi.Input[str]:
"""
The fully qualified domain name of the host. Computed on server side.
"""
return pulumi.get(self, "name")
@name.setter
def name(self, value: pulumi.Input[str]):
pulumi.set(self, "name", value)
@pulumi.input_type
class MdbMongodbClusterHostArgs:
def __init__(__self__, *,
subnet_id: pulumi.Input[str],
zone_id: pulumi.Input[str],
assign_public_ip: Optional[pulumi.Input[bool]] = None,
health: Optional[pulumi.Input[str]] = None,
name: Optional[pulumi.Input[str]] = None,
role: Optional[pulumi.Input[str]] = None,
shard_name: Optional[pulumi.Input[str]] = None,
type: Optional[pulumi.Input[str]] = None):
"""
:param pulumi.Input[str] subnet_id: The ID of the subnet, to which the host belongs. The subnet must
be a part of the network to which the cluster belongs.
:param pulumi.Input[str] zone_id: The availability zone where the MongoDB host will be created.
For more information see [the official documentation](https://cloud.yandex.com/docs/overview/concepts/geo-scope).
:param pulumi.Input[bool] assign_public_ip: -(Optional) Should this host have assigned public IP assigned. Can be either `true` or `false`.
:param pulumi.Input[str] health: The health of the host.
:param pulumi.Input[str] name: The fully qualified domain name of the host. Computed on server side.
:param pulumi.Input[str] role: The role of the cluster (either PRIMARY or SECONDARY).
:param pulumi.Input[str] shard_name: The name of the shard to which the host belongs.
:param pulumi.Input[str] type: Type of maintenance window. Can be either `ANYTIME` or `WEEKLY`. A day and hour of window need to be specified with weekly window.
"""
pulumi.set(__self__, "subnet_id", subnet_id)
pulumi.set(__self__, "zone_id", zone_id)
if assign_public_ip is not None:
pulumi.set(__self__, "assign_public_ip", assign_public_ip)
if health is not None:
pulumi.set(__self__, "health", health)
if name is not None:
pulumi.set(__self__, "name", name)
if role is not None:
pulumi.set(__self__, "role", role)
if shard_name is not None:
pulumi.set(__self__, "shard_name", shard_name)
if type is not None:
pulumi.set(__self__, "type", type)
@property
@pulumi.getter(name="subnetId")
def subnet_id(self) -> pulumi.Input[str]:
"""
The ID of the subnet, to which the host belongs. The subnet must
be a part of the network to which the cluster belongs.
"""
return pulumi.get(self, "subnet_id")
@subnet_id.setter
def subnet_id(self, value: pulumi.Input[str]):
pulumi.set(self, "subnet_id", value)
@property
@pulumi.getter(name="zoneId")
def zone_id(self) -> pulumi.Input[str]:
"""
The availability zone where the MongoDB host will be created.
For more information see [the official documentation](https://cloud.yandex.com/docs/overview/concepts/geo-scope).
"""
return pulumi.get(self, "zone_id")
@zone_id.setter
def zone_id(self, value: pulumi.Input[str]):
pulumi.set(self, "zone_id", value)
@property
@pulumi.getter(name="assignPublicIp")
def assign_public_ip(self) -> Optional[pulumi.Input[bool]]:
"""
-(Optional) Should this host have assigned public IP assigned. Can be either `true` or `false`.
"""
return pulumi.get(self, "assign_public_ip")
@assign_public_ip.setter
def assign_public_ip(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "assign_public_ip", value)
@property
@pulumi.getter
def health(self) -> Optional[pulumi.Input[str]]:
"""
The health of the host.
"""
return pulumi.get(self, "health")
@health.setter
def health(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "health", value)
@property
@pulumi.getter
def name(self) -> Optional[pulumi.Input[str]]:
"""
The fully qualified domain name of the host. Computed on server side.
"""
return pulumi.get(self, "name")
@name.setter
def name(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "name", value)
@property
@pulumi.getter
def role(self) -> Optional[pulumi.Input[str]]:
"""
The role of the cluster (either PRIMARY or SECONDARY).
"""
return pulumi.get(self, "role")
@role.setter
def role(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "role", value)
@property
@pulumi.getter(name="shardName")
def shard_name(self) -> Optional[pulumi.Input[str]]:
"""
The name of the shard to which the host belongs.
"""
return pulumi.get(self, "shard_name")
@shard_name.setter
def shard_name(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "shard_name", value)
@property
@pulumi.getter
def type(self) -> Optional[pulumi.Input[str]]:
"""
Type of maintenance window. Can be either `ANYTIME` or `WEEKLY`. A day and hour of window need to be specified with weekly window.
"""
return pulumi.get(self, "type")
@type.setter
def type(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "type", value)
@pulumi.input_type
class MdbMongodbClusterMaintenanceWindowArgs:
def __init__(__self__, *,
type: pulumi.Input[str],
day: Optional[pulumi.Input[str]] = None,
hour: Optional[pulumi.Input[int]] = None):
"""
:param pulumi.Input[str] type: Type of maintenance window. Can be either `ANYTIME` or `WEEKLY`. A day and hour of window need to be specified with weekly window.
:param pulumi.Input[str] day: Day of week for maintenance window if window type is weekly. Possible values: `MON`, `TUE`, `WED`, `THU`, `FRI`, `SAT`, `SUN`.
:param pulumi.Input[int] hour: Hour of day in UTC time zone (1-24) for maintenance window if window type is weekly.
"""
pulumi.set(__self__, "type", type)
if day is not None:
pulumi.set(__self__, "day", day)
if hour is not None:
pulumi.set(__self__, "hour", hour)
@property
@pulumi.getter
def type(self) -> pulumi.Input[str]:
"""
Type of maintenance window. Can be either `ANYTIME` or `WEEKLY`. A day and hour of window need to be specified with weekly window.
"""
return pulumi.get(self, "type")
@type.setter
def type(self, value: pulumi.Input[str]):
pulumi.set(self, "type", value)
@property
@pulumi.getter
def day(self) -> Optional[pulumi.Input[str]]:
"""
Day of week for maintenance window if window type is weekly. Possible values: `MON`, `TUE`, `WED`, `THU`, `FRI`, `SAT`, `SUN`.
"""
return pulumi.get(self, "day")
@day.setter
def day(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "day", value)
@property
@pulumi.getter
def hour(self) -> Optional[pulumi.Input[int]]:
"""
Hour of day in UTC time zone (1-24) for maintenance window if window type is weekly.
"""
return pulumi.get(self, "hour")
@hour.setter
def hour(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "hour", value)
@pulumi.input_type
class MdbMongodbClusterResourcesArgs:
def __init__(__self__, *,
disk_size: pulumi.Input[int],
disk_type_id: pulumi.Input[str],
resource_preset_id: pulumi.Input[str]):
"""
:param pulumi.Input[int] disk_size: Volume of the storage available to a MongoDB host, in gigabytes.
:param pulumi.Input[str] disk_type_id: Type of the storage of MongoDB hosts.
For more information see [the official documentation](https://cloud.yandex.com/docs/managed-clickhouse/concepts/storage).
"""
pulumi.set(__self__, "disk_size", disk_size)
pulumi.set(__self__, "disk_type_id", disk_type_id)
pulumi.set(__self__, "resource_preset_id", resource_preset_id)
@property
@pulumi.getter(name="diskSize")
def disk_size(self) -> pulumi.Input[int]:
"""
Volume of the storage available to a MongoDB host, in gigabytes.
"""
return pulumi.get(self, "disk_size")
@disk_size.setter
def disk_size(self, value: pulumi.Input[int]):
pulumi.set(self, "disk_size", value)
@property
@pulumi.getter(name="diskTypeId")
def disk_type_id(self) -> pulumi.Input[str]:
"""
Type of the storage of MongoDB hosts.
For more information see [the official documentation](https://cloud.yandex.com/docs/managed-clickhouse/concepts/storage).
"""
return pulumi.get(self, "disk_type_id")
@disk_type_id.setter
def disk_type_id(self, value: pulumi.Input[str]):
pulumi.set(self, "disk_type_id", value)
@property
@pulumi.getter(name="resourcePresetId")
def resource_preset_id(self) -> pulumi.Input[str]:
return pulumi.get(self, "resource_preset_id")
@resource_preset_id.setter
def resource_preset_id(self, value: pulumi.Input[str]):
pulumi.set(self, "resource_preset_id", value)
@pulumi.input_type
class MdbMongodbClusterUserArgs:
def __init__(__self__, *,
name: pulumi.Input[str],
password: pulumi.Input[str],
permissions: Optional[pulumi.Input[Sequence[pulumi.Input['MdbMongodbClusterUserPermissionArgs']]]] = None):
"""
:param pulumi.Input[str] name: The fully qualified domain name of the host. Computed on server side.
:param pulumi.Input[str] password: <PASSWORD>.
:param pulumi.Input[Sequence[pulumi.Input['MdbMongodbClusterUserPermissionArgs']]] permissions: Set of permissions granted to the user. The structure is documented below.
"""
pulumi.set(__self__, "name", name)
pulumi.set(__self__, "password", password)
if permissions is not None:
pulumi.set(__self__, "permissions", permissions)
@property
@pulumi.getter
def name(self) -> pulumi.Input[str]:
"""
The fully qualified domain name of the host. Computed on server side.
"""
return pulumi.get(self, "name")
@name.setter
def name(self, value: pulumi.Input[str]):
pulumi.set(self, "name", value)
@property
@pulumi.getter
def password(self) -> pulumi.Input[str]:
"""
The password of the user.
"""
return pulumi.get(self, "password")
@password.setter
def password(self, value: pulumi.Input[str]):
pulumi.set(self, "password", value)
@property
@pulumi.getter
def permissions(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['MdbMongodbClusterUserPermissionArgs']]]]:
"""
Set of permissions granted to the user. The structure is documented below.
"""
return pulumi.get(self, "permissions")
@permissions.setter
def permissions(self, value: Optional[pulumi.Input[Sequence[pulumi.Input['MdbMongodbClusterUserPermissionArgs']]]]):
pulumi.set(self, "permissions", value)
@pulumi.input_type
class MdbMongodbClusterUserPermissionArgs:
def __init__(__self__, *,
database_name: pulumi.Input[str],
roles: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None):
"""
:param pulumi.Input[str] database_name: The name of the database that the permission grants access to.
"""
pulumi.set(__self__, "database_name", database_name)
if roles is not None:
pulumi.set(__self__, "roles", roles)
@property
@pulumi.getter(name="databaseName")
def database_name(self) -> pulumi.Input[str]:
"""
The name of the database that the permission grants access to.
"""
return pulumi.get(self, "database_name")
@database_name.setter
def database_name(self, value: pulumi.Input[str]):
pulumi.set(self, "database_name", value)
@property
@pulumi.getter
def roles(self) -> Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]:
return pulumi.get(self, "roles")
@roles.setter
def roles(self, value: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]):
pulumi.set(self, "roles", value)
@pulumi.input_type
class MdbMysqlClusterAccessArgs:
def __init__(__self__, *,
data_lens: Optional[pulumi.Input[bool]] = None,
web_sql: Optional[pulumi.Input[bool]] = None):
"""
:param pulumi.Input[bool] data_lens: Allow access for [Yandex DataLens](https://cloud.yandex.com/services/datalens).
:param pulumi.Input[bool] web_sql: Allows access for [SQL queries in the management console](https://cloud.yandex.com/docs/managed-mysql/operations/web-sql-query).
"""
if data_lens is not None:
pulumi.set(__self__, "data_lens", data_lens)
if web_sql is not None:
pulumi.set(__self__, "web_sql", web_sql)
@property
@pulumi.getter(name="dataLens")
def data_lens(self) -> Optional[pulumi.Input[bool]]:
"""
Allow access for [Yandex DataLens](https://cloud.yandex.com/services/datalens).
"""
return pulumi.get(self, "data_lens")
@data_lens.setter
def data_lens(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "data_lens", value)
@property
@pulumi.getter(name="webSql")
def web_sql(self) -> Optional[pulumi.Input[bool]]:
"""
Allows access for [SQL queries in the management console](https://cloud.yandex.com/docs/managed-mysql/operations/web-sql-query).
"""
return pulumi.get(self, "web_sql")
@web_sql.setter
def web_sql(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "web_sql", value)
@pulumi.input_type
class MdbMysqlClusterBackupWindowStartArgs:
def __init__(__self__, *,
hours: Optional[pulumi.Input[int]] = None,
minutes: Optional[pulumi.Input[int]] = None):
"""
:param pulumi.Input[int] hours: The hour at which backup will be started.
:param pulumi.Input[int] minutes: The minute at which backup will be started.
"""
if hours is not None:
pulumi.set(__self__, "hours", hours)
if minutes is not None:
pulumi.set(__self__, "minutes", minutes)
@property
@pulumi.getter
def hours(self) -> Optional[pulumi.Input[int]]:
"""
The hour at which backup will be started.
"""
return pulumi.get(self, "hours")
@hours.setter
def hours(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "hours", value)
@property
@pulumi.getter
def minutes(self) -> Optional[pulumi.Input[int]]:
"""
The minute at which backup will be started.
"""
return pulumi.get(self, "minutes")
@minutes.setter
def minutes(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "minutes", value)
@pulumi.input_type
class MdbMysqlClusterDatabaseArgs:
def __init__(__self__, *,
name: pulumi.Input[str]):
"""
:param pulumi.Input[str] name: Host state name. It should be set for all hosts or unset for all hosts. This field can be used by another host, to select which host will be its replication source. Please refer to `replication_source_name` parameter.
"""
pulumi.set(__self__, "name", name)
@property
@pulumi.getter
def name(self) -> pulumi.Input[str]:
"""
Host state name. It should be set for all hosts or unset for all hosts. This field can be used by another host, to select which host will be its replication source. Please refer to `replication_source_name` parameter.
"""
return pulumi.get(self, "name")
@name.setter
def name(self, value: pulumi.Input[str]):
pulumi.set(self, "name", value)
@pulumi.input_type
class MdbMysqlClusterHostArgs:
def __init__(__self__, *,
zone: pulumi.Input[str],
assign_public_ip: Optional[pulumi.Input[bool]] = None,
fqdn: Optional[pulumi.Input[str]] = None,
name: Optional[pulumi.Input[str]] = None,
replication_source: Optional[pulumi.Input[str]] = None,
replication_source_name: Optional[pulumi.Input[str]] = None,
subnet_id: Optional[pulumi.Input[str]] = None):
"""
:param pulumi.Input[str] zone: The availability zone where the MySQL host will be created.
:param pulumi.Input[bool] assign_public_ip: Sets whether the host should get a public IP address. It can be changed on the fly only when `name` is set.
:param pulumi.Input[str] fqdn: The fully qualified domain name of the host.
:param pulumi.Input[str] name: Host state name. It should be set for all hosts or unset for all hosts. This field can be used by another host, to select which host will be its replication source. Please refer to `replication_source_name` parameter.
:param pulumi.Input[str] replication_source: Host replication source (fqdn), when replication_source is empty then host is in HA group.
:param pulumi.Input[str] replication_source_name: Host replication source name points to host's `name` from which this host should replicate. When not set then host in HA group. It works only when `name` is set.
:param pulumi.Input[str] subnet_id: The ID of the subnet, to which the host belongs. The subnet must be a part of the network to which the cluster belongs.
"""
pulumi.set(__self__, "zone", zone)
if assign_public_ip is not None:
pulumi.set(__self__, "assign_public_ip", assign_public_ip)
if fqdn is not None:
pulumi.set(__self__, "fqdn", fqdn)
if name is not None:
pulumi.set(__self__, "name", name)
if replication_source is not None:
pulumi.set(__self__, "replication_source", replication_source)
if replication_source_name is not None:
pulumi.set(__self__, "replication_source_name", replication_source_name)
if subnet_id is not None:
pulumi.set(__self__, "subnet_id", subnet_id)
@property
@pulumi.getter
def zone(self) -> pulumi.Input[str]:
"""
The availability zone where the MySQL host will be created.
"""
return pulumi.get(self, "zone")
@zone.setter
def zone(self, value: pulumi.Input[str]):
pulumi.set(self, "zone", value)
@property
@pulumi.getter(name="assignPublicIp")
def assign_public_ip(self) -> Optional[pulumi.Input[bool]]:
"""
Sets whether the host should get a public IP address. It can be changed on the fly only when `name` is set.
"""
return pulumi.get(self, "assign_public_ip")
@assign_public_ip.setter
def assign_public_ip(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "assign_public_ip", value)
@property
@pulumi.getter
def fqdn(self) -> Optional[pulumi.Input[str]]:
"""
The fully qualified domain name of the host.
"""
return pulumi.get(self, "fqdn")
@fqdn.setter
def fqdn(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "fqdn", value)
@property
@pulumi.getter
def name(self) -> Optional[pulumi.Input[str]]:
"""
Host state name. It should be set for all hosts or unset for all hosts. This field can be used by another host, to select which host will be its replication source. Please refer to `replication_source_name` parameter.
"""
return pulumi.get(self, "name")
@name.setter
def name(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "name", value)
@property
@pulumi.getter(name="replicationSource")
def replication_source(self) -> Optional[pulumi.Input[str]]:
"""
Host replication source (fqdn), when replication_source is empty then host is in HA group.
"""
return pulumi.get(self, "replication_source")
@replication_source.setter
def replication_source(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "replication_source", value)
@property
@pulumi.getter(name="replicationSourceName")
def replication_source_name(self) -> Optional[pulumi.Input[str]]:
"""
Host replication source name points to host's `name` from which this host should replicate. When not set then host in HA group. It works only when `name` is set.
"""
return pulumi.get(self, "replication_source_name")
@replication_source_name.setter
def replication_source_name(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "replication_source_name", value)
@property
@pulumi.getter(name="subnetId")
def subnet_id(self) -> Optional[pulumi.Input[str]]:
"""
The ID of the subnet, to which the host belongs. The subnet must be a part of the network to which the cluster belongs.
"""
return pulumi.get(self, "subnet_id")
@subnet_id.setter
def subnet_id(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "subnet_id", value)
@pulumi.input_type
class MdbMysqlClusterMaintenanceWindowArgs:
def __init__(__self__, *,
type: pulumi.Input[str],
day: Optional[pulumi.Input[str]] = None,
hour: Optional[pulumi.Input[int]] = None):
"""
:param pulumi.Input[str] type: Type of maintenance window. Can be either `ANYTIME` or `WEEKLY`. A day and hour of window need to be specified with weekly window.
:param pulumi.Input[str] day: Day of the week (in `DDD` format). Allowed values: "MON", "TUE", "WED", "THU", "FRI", "SAT", "SUN"
:param pulumi.Input[int] hour: Hour of the day in UTC (in `HH` format). Allowed value is between 0 and 23.
"""
pulumi.set(__self__, "type", type)
if day is not None:
pulumi.set(__self__, "day", day)
if hour is not None:
pulumi.set(__self__, "hour", hour)
@property
@pulumi.getter
def type(self) -> pulumi.Input[str]:
"""
Type of maintenance window. Can be either `ANYTIME` or `WEEKLY`. A day and hour of window need to be specified with weekly window.
"""
return pulumi.get(self, "type")
@type.setter
def type(self, value: pulumi.Input[str]):
pulumi.set(self, "type", value)
@property
@pulumi.getter
def day(self) -> Optional[pulumi.Input[str]]:
"""
Day of the week (in `DDD` format). Allowed values: "MON", "TUE", "WED", "THU", "FRI", "SAT", "SUN"
"""
return pulumi.get(self, "day")
@day.setter
def day(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "day", value)
@property
@pulumi.getter
def hour(self) -> Optional[pulumi.Input[int]]:
"""
Hour of the day in UTC (in `HH` format). Allowed value is between 0 and 23.
"""
return pulumi.get(self, "hour")
@hour.setter
def hour(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "hour", value)
@pulumi.input_type
class MdbMysqlClusterResourcesArgs:
def __init__(__self__, *,
disk_size: pulumi.Input[int],
disk_type_id: pulumi.Input[str],
resource_preset_id: pulumi.Input[str]):
"""
:param pulumi.Input[int] disk_size: Volume of the storage available to a MySQL host, in gigabytes.
:param pulumi.Input[str] disk_type_id: Type of the storage of MySQL hosts.
"""
pulumi.set(__self__, "disk_size", disk_size)
pulumi.set(__self__, "disk_type_id", disk_type_id)
pulumi.set(__self__, "resource_preset_id", resource_preset_id)
@property
@pulumi.getter(name="diskSize")
def disk_size(self) -> pulumi.Input[int]:
"""
Volume of the storage available to a MySQL host, in gigabytes.
"""
return pulumi.get(self, "disk_size")
@disk_size.setter
def disk_size(self, value: pulumi.Input[int]):
pulumi.set(self, "disk_size", value)
@property
@pulumi.getter(name="diskTypeId")
def disk_type_id(self) -> pulumi.Input[str]:
"""
Type of the storage of MySQL hosts.
"""
return pulumi.get(self, "disk_type_id")
@disk_type_id.setter
def disk_type_id(self, value: pulumi.Input[str]):
pulumi.set(self, "disk_type_id", value)
@property
@pulumi.getter(name="resourcePresetId")
def resource_preset_id(self) -> pulumi.Input[str]:
return pulumi.get(self, "resource_preset_id")
@resource_preset_id.setter
def resource_preset_id(self, value: pulumi.Input[str]):
pulumi.set(self, "resource_preset_id", value)
@pulumi.input_type
class MdbMysqlClusterRestoreArgs:
def __init__(__self__, *,
backup_id: pulumi.Input[str],
time: Optional[pulumi.Input[str]] = None):
"""
:param pulumi.Input[str] backup_id: Backup ID. The cluster will be created from the specified backup. [How to get a list of MySQL backups](https://cloud.yandex.com/docs/managed-mysql/operations/cluster-backups).
:param pulumi.Input[str] time: Timestamp of the moment to which the MySQL cluster should be restored. (Format: "2006-01-02T15:04:05" - UTC). When not set, current time is used.
"""
pulumi.set(__self__, "backup_id", backup_id)
if time is not None:
pulumi.set(__self__, "time", time)
@property
@pulumi.getter(name="backupId")
def backup_id(self) -> pulumi.Input[str]:
"""
Backup ID. The cluster will be created from the specified backup. [How to get a list of MySQL backups](https://cloud.yandex.com/docs/managed-mysql/operations/cluster-backups).
"""
return pulumi.get(self, "backup_id")
@backup_id.setter
def backup_id(self, value: pulumi.Input[str]):
pulumi.set(self, "backup_id", value)
@property
@pulumi.getter
def time(self) -> Optional[pulumi.Input[str]]:
"""
Timestamp of the moment to which the MySQL cluster should be restored. (Format: "2006-01-02T15:04:05" - UTC). When not set, current time is used.
"""
return pulumi.get(self, "time")
@time.setter
def time(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "time", value)
@pulumi.input_type
class MdbMysqlClusterUserArgs:
def __init__(__self__, *,
name: pulumi.Input[str],
password: pulumi.Input[str],
authentication_plugin: Optional[pulumi.Input[str]] = None,
connection_limits: Optional[pulumi.Input['MdbMysqlClusterUserConnectionLimitsArgs']] = None,
global_permissions: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None,
permissions: Optional[pulumi.Input[Sequence[pulumi.Input['MdbMysqlClusterUserPermissionArgs']]]] = None):
"""
:param pulumi.Input[str] name: Host state name. It should be set for all hosts or unset for all hosts. This field can be used by another host, to select which host will be its replication source. Please refer to `replication_source_name` parameter.
:param pulumi.Input[str] password: <PASSWORD>.
:param pulumi.Input[str] authentication_plugin: Authentication plugin. Allowed values: `MYSQL_NATIVE_PASSWORD`, `CACHING_SHA2_PASSWORD`, `<PASSWORD>` (for version 5.7 `MYSQL_NATIVE_PASSWORD`, `<PASSWORD>`)
:param pulumi.Input['MdbMysqlClusterUserConnectionLimitsArgs'] connection_limits: User's connection limits. The structure is documented below.
If the attribute is not specified there will be no changes.
:param pulumi.Input[Sequence[pulumi.Input[str]]] global_permissions: List user's global permissions
Allowed permissions: `REPLICATION_CLIENT`, `REPLICATION_SLAVE`, `PROCESS` for clear list use empty list.
If the attribute is not specified there will be no changes.
:param pulumi.Input[Sequence[pulumi.Input['MdbMysqlClusterUserPermissionArgs']]] permissions: Set of permissions granted to the user. The structure is documented below.
"""
pulumi.set(__self__, "name", name)
pulumi.set(__self__, "password", password)
if authentication_plugin is not None:
pulumi.set(__self__, "authentication_plugin", authentication_plugin)
if connection_limits is not None:
pulumi.set(__self__, "connection_limits", connection_limits)
if global_permissions is not None:
pulumi.set(__self__, "global_permissions", global_permissions)
if permissions is not None:
pulumi.set(__self__, "permissions", permissions)
@property
@pulumi.getter
def name(self) -> pulumi.Input[str]:
"""
Host state name. It should be set for all hosts or unset for all hosts. This field can be used by another host, to select which host will be its replication source. Please refer to `replication_source_name` parameter.
"""
return pulumi.get(self, "name")
@name.setter
def name(self, value: pulumi.Input[str]):
pulumi.set(self, "name", value)
@property
@pulumi.getter
def password(self) -> pulumi.Input[str]:
"""
The password of the user.
"""
return pulumi.get(self, "password")
@password.setter
def password(self, value: pulumi.Input[str]):
pulumi.set(self, "password", value)
@property
@pulumi.getter(name="authenticationPlugin")
def authentication_plugin(self) -> Optional[pulumi.Input[str]]:
"""
Authentication plugin. Allowed values: `MYSQL_NATIVE_PASSWORD`, `CACHING_SHA2_PASSWORD`, `SHA256_PASSWORD` (for version 5.7 `MYSQL_NATIVE_PASSWORD`, `SHA256_PASSWORD`)
"""
return pulumi.get(self, "authentication_plugin")
@authentication_plugin.setter
def authentication_plugin(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "authentication_plugin", value)
@property
@pulumi.getter(name="connectionLimits")
def connection_limits(self) -> Optional[pulumi.Input['MdbMysqlClusterUserConnectionLimitsArgs']]:
"""
User's connection limits. The structure is documented below.
If the attribute is not specified there will be no changes.
"""
return pulumi.get(self, "connection_limits")
@connection_limits.setter
def connection_limits(self, value: Optional[pulumi.Input['MdbMysqlClusterUserConnectionLimitsArgs']]):
pulumi.set(self, "connection_limits", value)
@property
@pulumi.getter(name="globalPermissions")
def global_permissions(self) -> Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]:
"""
List user's global permissions
Allowed permissions: `REPLICATION_CLIENT`, `REPLICATION_SLAVE`, `PROCESS` for clear list use empty list.
If the attribute is not specified there will be no changes.
"""
return pulumi.get(self, "global_permissions")
@global_permissions.setter
def global_permissions(self, value: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]):
pulumi.set(self, "global_permissions", value)
@property
@pulumi.getter
def permissions(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['MdbMysqlClusterUserPermissionArgs']]]]:
"""
Set of permissions granted to the user. The structure is documented below.
"""
return pulumi.get(self, "permissions")
@permissions.setter
def permissions(self, value: Optional[pulumi.Input[Sequence[pulumi.Input['MdbMysqlClusterUserPermissionArgs']]]]):
pulumi.set(self, "permissions", value)
@pulumi.input_type
class MdbMysqlClusterUserConnectionLimitsArgs:
def __init__(__self__, *,
max_connections_per_hour: Optional[pulumi.Input[int]] = None,
max_questions_per_hour: Optional[pulumi.Input[int]] = None,
max_updates_per_hour: Optional[pulumi.Input[int]] = None,
max_user_connections: Optional[pulumi.Input[int]] = None):
"""
:param pulumi.Input[int] max_connections_per_hour: Max connections per hour.
:param pulumi.Input[int] max_questions_per_hour: Max questions per hour.
:param pulumi.Input[int] max_updates_per_hour: Max updates per hour.
:param pulumi.Input[int] max_user_connections: Max user connections.
"""
if max_connections_per_hour is not None:
pulumi.set(__self__, "max_connections_per_hour", max_connections_per_hour)
if max_questions_per_hour is not None:
pulumi.set(__self__, "max_questions_per_hour", max_questions_per_hour)
if max_updates_per_hour is not None:
pulumi.set(__self__, "max_updates_per_hour", max_updates_per_hour)
if max_user_connections is not None:
pulumi.set(__self__, "max_user_connections", max_user_connections)
@property
@pulumi.getter(name="maxConnectionsPerHour")
def max_connections_per_hour(self) -> Optional[pulumi.Input[int]]:
"""
Max connections per hour.
"""
return pulumi.get(self, "max_connections_per_hour")
@max_connections_per_hour.setter
def max_connections_per_hour(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "max_connections_per_hour", value)
@property
@pulumi.getter(name="maxQuestionsPerHour")
def max_questions_per_hour(self) -> Optional[pulumi.Input[int]]:
"""
Max questions per hour.
"""
return pulumi.get(self, "max_questions_per_hour")
@max_questions_per_hour.setter
def max_questions_per_hour(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "max_questions_per_hour", value)
@property
@pulumi.getter(name="maxUpdatesPerHour")
def max_updates_per_hour(self) -> Optional[pulumi.Input[int]]:
"""
Max updates per hour.
"""
return pulumi.get(self, "max_updates_per_hour")
@max_updates_per_hour.setter
def max_updates_per_hour(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "max_updates_per_hour", value)
@property
@pulumi.getter(name="maxUserConnections")
def max_user_connections(self) -> Optional[pulumi.Input[int]]:
"""
Max user connections.
"""
return pulumi.get(self, "max_user_connections")
@max_user_connections.setter
def max_user_connections(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "max_user_connections", value)
@pulumi.input_type
class MdbMysqlClusterUserPermissionArgs:
def __init__(__self__, *,
database_name: pulumi.Input[str],
roles: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None):
"""
:param pulumi.Input[str] database_name: The name of the database that the permission grants access to.
:param pulumi.Input[Sequence[pulumi.Input[str]]] roles: List user's roles in the database.
Allowed roles: `ALL`,`ALTER`,`ALTER_ROUTINE`,`CREATE`,`CREATE_ROUTINE`,`CREATE_TEMPORARY_TABLES`,
`CREATE_VIEW`,`DELETE`,`DROP`,`EVENT`,`EXECUTE`,`INDEX`,`INSERT`,`LOCK_TABLES`,`SELECT`,`SHOW_VIEW`,`TRIGGER`,`UPDATE`.
"""
pulumi.set(__self__, "database_name", database_name)
if roles is not None:
pulumi.set(__self__, "roles", roles)
@property
@pulumi.getter(name="databaseName")
def database_name(self) -> pulumi.Input[str]:
"""
The name of the database that the permission grants access to.
"""
return pulumi.get(self, "database_name")
@database_name.setter
def database_name(self, value: pulumi.Input[str]):
pulumi.set(self, "database_name", value)
@property
@pulumi.getter
def roles(self) -> Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]:
"""
List user's roles in the database.
Allowed roles: `ALL`,`ALTER`,`ALTER_ROUTINE`,`CREATE`,`CREATE_ROUTINE`,`CREATE_TEMPORARY_TABLES`,
`CREATE_VIEW`,`DELETE`,`DROP`,`EVENT`,`EXECUTE`,`INDEX`,`INSERT`,`LOCK_TABLES`,`SELECT`,`SHOW_VIEW`,`TRIGGER`,`UPDATE`.
"""
return pulumi.get(self, "roles")
@roles.setter
def roles(self, value: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]):
pulumi.set(self, "roles", value)
@pulumi.input_type
class MdbRedisClusterConfigArgs:
def __init__(__self__, *,
password: pulumi.Input[str],
version: pulumi.Input[str],
databases: Optional[pulumi.Input[int]] = None,
maxmemory_policy: Optional[pulumi.Input[str]] = None,
notify_keyspace_events: Optional[pulumi.Input[str]] = None,
slowlog_log_slower_than: Optional[pulumi.Input[int]] = None,
slowlog_max_len: Optional[pulumi.Input[int]] = None,
timeout: Optional[pulumi.Input[int]] = None):
"""
:param pulumi.Input[str] password: Password for the Redis cluster.
:param pulumi.Input[str] version: Version of Redis (5.0, 6.0 or 6.2).
:param pulumi.Input[int] databases: Number of databases (changing requires redis-server restart).
:param pulumi.Input[str] maxmemory_policy: Redis key eviction policy for a dataset that reaches maximum memory.
Can be any of the listed in [the official RedisDB documentation](https://docs.redislabs.com/latest/rs/administering/database-operations/eviction-policy/).
:param pulumi.Input[str] notify_keyspace_events: Select the events that Redis will notify among a set of classes.
:param pulumi.Input[int] slowlog_log_slower_than: Log slow queries below this number in microseconds.
:param pulumi.Input[int] slowlog_max_len: Slow queries log length.
:param pulumi.Input[int] timeout: Close the connection after a client is idle for N seconds.
"""
pulumi.set(__self__, "password", password)
pulumi.set(__self__, "version", version)
if databases is not None:
pulumi.set(__self__, "databases", databases)
if maxmemory_policy is not None:
pulumi.set(__self__, "maxmemory_policy", maxmemory_policy)
if notify_keyspace_events is not None:
pulumi.set(__self__, "notify_keyspace_events", notify_keyspace_events)
if slowlog_log_slower_than is not None:
pulumi.set(__self__, "slowlog_log_slower_than", slowlog_log_slower_than)
if slowlog_max_len is not None:
pulumi.set(__self__, "slowlog_max_len", slowlog_max_len)
if timeout is not None:
pulumi.set(__self__, "timeout", timeout)
@property
@pulumi.getter
def password(self) -> pulumi.Input[str]:
"""
Password for the Redis cluster.
"""
return pulumi.get(self, "password")
@password.setter
def password(self, value: pulumi.Input[str]):
pulumi.set(self, "password", value)
@property
@pulumi.getter
def version(self) -> pulumi.Input[str]:
"""
Version of Redis (5.0, 6.0 or 6.2).
"""
return pulumi.get(self, "version")
@version.setter
def version(self, value: pulumi.Input[str]):
pulumi.set(self, "version", value)
@property
@pulumi.getter
def databases(self) -> Optional[pulumi.Input[int]]:
"""
Number of databases (changing requires redis-server restart).
"""
return pulumi.get(self, "databases")
@databases.setter
def databases(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "databases", value)
@property
@pulumi.getter(name="maxmemoryPolicy")
def maxmemory_policy(self) -> Optional[pulumi.Input[str]]:
"""
Redis key eviction policy for a dataset that reaches maximum memory.
Can be any of the listed in [the official RedisDB documentation](https://docs.redislabs.com/latest/rs/administering/database-operations/eviction-policy/).
"""
return pulumi.get(self, "maxmemory_policy")
@maxmemory_policy.setter
def maxmemory_policy(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "maxmemory_policy", value)
@property
@pulumi.getter(name="notifyKeyspaceEvents")
def notify_keyspace_events(self) -> Optional[pulumi.Input[str]]:
"""
Select the events that Redis will notify among a set of classes.
"""
return pulumi.get(self, "notify_keyspace_events")
@notify_keyspace_events.setter
def notify_keyspace_events(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "notify_keyspace_events", value)
@property
@pulumi.getter(name="slowlogLogSlowerThan")
def slowlog_log_slower_than(self) -> Optional[pulumi.Input[int]]:
"""
Log slow queries below this number in microseconds.
"""
return pulumi.get(self, "slowlog_log_slower_than")
@slowlog_log_slower_than.setter
def slowlog_log_slower_than(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "slowlog_log_slower_than", value)
@property
@pulumi.getter(name="slowlogMaxLen")
def slowlog_max_len(self) -> Optional[pulumi.Input[int]]:
"""
Slow queries log length.
"""
return pulumi.get(self, "slowlog_max_len")
@slowlog_max_len.setter
def slowlog_max_len(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "slowlog_max_len", value)
@property
@pulumi.getter
def timeout(self) -> Optional[pulumi.Input[int]]:
"""
Close the connection after a client is idle for N seconds.
"""
return pulumi.get(self, "timeout")
@timeout.setter
def timeout(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "timeout", value)
@pulumi.input_type
class MdbRedisClusterHostArgs:
def __init__(__self__, *,
zone: pulumi.Input[str],
fqdn: Optional[pulumi.Input[str]] = None,
shard_name: Optional[pulumi.Input[str]] = None,
subnet_id: Optional[pulumi.Input[str]] = None):
"""
:param pulumi.Input[str] zone: The availability zone where the Redis host will be created.
For more information see [the official documentation](https://cloud.yandex.com/docs/overview/concepts/geo-scope).
:param pulumi.Input[str] fqdn: The fully qualified domain name of the host.
:param pulumi.Input[str] shard_name: The name of the shard to which the host belongs.
:param pulumi.Input[str] subnet_id: The ID of the subnet, to which the host belongs. The subnet must
be a part of the network to which the cluster belongs.
"""
pulumi.set(__self__, "zone", zone)
if fqdn is not None:
pulumi.set(__self__, "fqdn", fqdn)
if shard_name is not None:
pulumi.set(__self__, "shard_name", shard_name)
if subnet_id is not None:
pulumi.set(__self__, "subnet_id", subnet_id)
@property
@pulumi.getter
def zone(self) -> pulumi.Input[str]:
"""
The availability zone where the Redis host will be created.
For more information see [the official documentation](https://cloud.yandex.com/docs/overview/concepts/geo-scope).
"""
return pulumi.get(self, "zone")
@zone.setter
def zone(self, value: pulumi.Input[str]):
pulumi.set(self, "zone", value)
@property
@pulumi.getter
def fqdn(self) -> Optional[pulumi.Input[str]]:
"""
The fully qualified domain name of the host.
"""
return pulumi.get(self, "fqdn")
@fqdn.setter
def fqdn(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "fqdn", value)
@property
@pulumi.getter(name="shardName")
def shard_name(self) -> Optional[pulumi.Input[str]]:
"""
The name of the shard to which the host belongs.
"""
return pulumi.get(self, "shard_name")
@shard_name.setter
def shard_name(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "shard_name", value)
@property
@pulumi.getter(name="subnetId")
def subnet_id(self) -> Optional[pulumi.Input[str]]:
"""
The ID of the subnet, to which the host belongs. The subnet must
be a part of the network to which the cluster belongs.
"""
return pulumi.get(self, "subnet_id")
@subnet_id.setter
def subnet_id(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "subnet_id", value)
@pulumi.input_type
class MdbRedisClusterMaintenanceWindowArgs:
def __init__(__self__, *,
type: pulumi.Input[str],
day: Optional[pulumi.Input[str]] = None,
hour: Optional[pulumi.Input[int]] = None):
"""
:param pulumi.Input[str] type: Type of maintenance window. Can be either `ANYTIME` or `WEEKLY`. A day and hour of window need to be specified with weekly window.
:param pulumi.Input[str] day: Day of week for maintenance window if window type is weekly. Possible values: `MON`, `TUE`, `WED`, `THU`, `FRI`, `SAT`, `SUN`.
:param pulumi.Input[int] hour: Hour of day in UTC time zone (1-24) for maintenance window if window type is weekly.
"""
pulumi.set(__self__, "type", type)
if day is not None:
pulumi.set(__self__, "day", day)
if hour is not None:
pulumi.set(__self__, "hour", hour)
@property
@pulumi.getter
def type(self) -> pulumi.Input[str]:
"""
Type of maintenance window. Can be either `ANYTIME` or `WEEKLY`. A day and hour of window need to be specified with weekly window.
"""
return pulumi.get(self, "type")
@type.setter
def type(self, value: pulumi.Input[str]):
pulumi.set(self, "type", value)
@property
@pulumi.getter
def day(self) -> Optional[pulumi.Input[str]]:
"""
Day of week for maintenance window if window type is weekly. Possible values: `MON`, `TUE`, `WED`, `THU`, `FRI`, `SAT`, `SUN`.
"""
return pulumi.get(self, "day")
@day.setter
def day(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "day", value)
@property
@pulumi.getter
def hour(self) -> Optional[pulumi.Input[int]]:
"""
Hour of day in UTC time zone (1-24) for maintenance window if window type is weekly.
"""
return pulumi.get(self, "hour")
@hour.setter
def hour(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "hour", value)
@pulumi.input_type
class MdbRedisClusterResourcesArgs:
def __init__(__self__, *,
disk_size: pulumi.Input[int],
resource_preset_id: pulumi.Input[str],
disk_type_id: Optional[pulumi.Input[str]] = None):
"""
:param pulumi.Input[int] disk_size: Volume of the storage available to a host, in gigabytes.
:param pulumi.Input[str] disk_type_id: Type of the storage of Redis hosts - environment default is used if missing.
"""
pulumi.set(__self__, "disk_size", disk_size)
pulumi.set(__self__, "resource_preset_id", resource_preset_id)
if disk_type_id is not None:
pulumi.set(__self__, "disk_type_id", disk_type_id)
@property
@pulumi.getter(name="diskSize")
def disk_size(self) -> pulumi.Input[int]:
"""
Volume of the storage available to a host, in gigabytes.
"""
return pulumi.get(self, "disk_size")
@disk_size.setter
def disk_size(self, value: pulumi.Input[int]):
pulumi.set(self, "disk_size", value)
@property
@pulumi.getter(name="resourcePresetId")
def resource_preset_id(self) -> pulumi.Input[str]:
return pulumi.get(self, "resource_preset_id")
@resource_preset_id.setter
def resource_preset_id(self, value: pulumi.Input[str]):
pulumi.set(self, "resource_preset_id", value)
@property
@pulumi.getter(name="diskTypeId")
def disk_type_id(self) -> Optional[pulumi.Input[str]]:
"""
Type of the storage of Redis hosts - environment default is used if missing.
"""
return pulumi.get(self, "disk_type_id")
@disk_type_id.setter
def disk_type_id(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "disk_type_id", value)
@pulumi.input_type
class MdbSqlServerClusterBackupWindowStartArgs:
def __init__(__self__, *,
hours: Optional[pulumi.Input[int]] = None,
minutes: Optional[pulumi.Input[int]] = None):
"""
:param pulumi.Input[int] hours: The hour at which backup will be started.
:param pulumi.Input[int] minutes: The minute at which backup will be started.
"""
if hours is not None:
pulumi.set(__self__, "hours", hours)
if minutes is not None:
pulumi.set(__self__, "minutes", minutes)
@property
@pulumi.getter
def hours(self) -> Optional[pulumi.Input[int]]:
"""
The hour at which backup will be started.
"""
return pulumi.get(self, "hours")
@hours.setter
def hours(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "hours", value)
@property
@pulumi.getter
def minutes(self) -> Optional[pulumi.Input[int]]:
"""
The minute at which backup will be started.
"""
return pulumi.get(self, "minutes")
@minutes.setter
def minutes(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "minutes", value)
@pulumi.input_type
class MdbSqlServerClusterDatabaseArgs:
def __init__(__self__, *,
name: pulumi.Input[str]):
"""
:param pulumi.Input[str] name: The name of the database.
"""
pulumi.set(__self__, "name", name)
@property
@pulumi.getter
def name(self) -> pulumi.Input[str]:
"""
The name of the database.
"""
return pulumi.get(self, "name")
@name.setter
def name(self, value: pulumi.Input[str]):
pulumi.set(self, "name", value)
@pulumi.input_type
class MdbSqlServerClusterHostArgs:
def __init__(__self__, *,
zone: pulumi.Input[str],
assign_public_ip: Optional[pulumi.Input[bool]] = None,
fqdn: Optional[pulumi.Input[str]] = None,
subnet_id: Optional[pulumi.Input[str]] = None):
"""
:param pulumi.Input[str] zone: The availability zone where the SQLServer host will be created.
:param pulumi.Input[bool] assign_public_ip: Sets whether the host should get a public IP address on creation. Changing this parameter for an existing host is not supported at the moment
:param pulumi.Input[str] fqdn: The fully qualified domain name of the host.
:param pulumi.Input[str] subnet_id: The ID of the subnet, to which the host belongs. The subnet must be a part of the network to which the cluster belongs.
"""
pulumi.set(__self__, "zone", zone)
if assign_public_ip is not None:
pulumi.set(__self__, "assign_public_ip", assign_public_ip)
if fqdn is not None:
pulumi.set(__self__, "fqdn", fqdn)
if subnet_id is not None:
pulumi.set(__self__, "subnet_id", subnet_id)
@property
@pulumi.getter
def zone(self) -> pulumi.Input[str]:
"""
The availability zone where the SQLServer host will be created.
"""
return pulumi.get(self, "zone")
@zone.setter
def zone(self, value: pulumi.Input[str]):
pulumi.set(self, "zone", value)
@property
@pulumi.getter(name="assignPublicIp")
def assign_public_ip(self) -> Optional[pulumi.Input[bool]]:
"""
Sets whether the host should get a public IP address on creation. Changing this parameter for an existing host is not supported at the moment
"""
return pulumi.get(self, "assign_public_ip")
@assign_public_ip.setter
def assign_public_ip(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "assign_public_ip", value)
@property
@pulumi.getter
def fqdn(self) -> Optional[pulumi.Input[str]]:
"""
The fully qualified domain name of the host.
"""
return pulumi.get(self, "fqdn")
@fqdn.setter
def fqdn(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "fqdn", value)
@property
@pulumi.getter(name="subnetId")
def subnet_id(self) -> Optional[pulumi.Input[str]]:
"""
The ID of the subnet, to which the host belongs. The subnet must be a part of the network to which the cluster belongs.
"""
return pulumi.get(self, "subnet_id")
@subnet_id.setter
def subnet_id(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "subnet_id", value)
@pulumi.input_type
class MdbSqlServerClusterResourcesArgs:
def __init__(__self__, *,
disk_size: pulumi.Input[int],
disk_type_id: pulumi.Input[str],
resource_preset_id: pulumi.Input[str]):
"""
:param pulumi.Input[int] disk_size: Volume of the storage available to a SQLServer host, in gigabytes.
:param pulumi.Input[str] disk_type_id: Type of the storage of SQLServer hosts.
"""
pulumi.set(__self__, "disk_size", disk_size)
pulumi.set(__self__, "disk_type_id", disk_type_id)
pulumi.set(__self__, "resource_preset_id", resource_preset_id)
@property
@pulumi.getter(name="diskSize")
def disk_size(self) -> pulumi.Input[int]:
"""
Volume of the storage available to a SQLServer host, in gigabytes.
"""
return pulumi.get(self, "disk_size")
@disk_size.setter
def disk_size(self, value: pulumi.Input[int]):
pulumi.set(self, "disk_size", value)
@property
@pulumi.getter(name="diskTypeId")
def disk_type_id(self) -> pulumi.Input[str]:
"""
Type of the storage of SQLServer hosts.
"""
return pulumi.get(self, "disk_type_id")
@disk_type_id.setter
def disk_type_id(self, value: pulumi.Input[str]):
pulumi.set(self, "disk_type_id", value)
@property
@pulumi.getter(name="resourcePresetId")
def resource_preset_id(self) -> pulumi.Input[str]:
return pulumi.get(self, "resource_preset_id")
@resource_preset_id.setter
def resource_preset_id(self, value: pulumi.Input[str]):
pulumi.set(self, "resource_preset_id", value)
@pulumi.input_type
class MdbSqlServerClusterUserArgs:
def __init__(__self__, *,
name: pulumi.Input[str],
password: pulumi.Input[str],
permissions: Optional[pulumi.Input[Sequence[pulumi.Input['MdbSqlServerClusterUserPermissionArgs']]]] = None):
"""
:param pulumi.Input[str] name: The name of the database.
:param pulumi.Input[str] password: <PASSWORD>.
:param pulumi.Input[Sequence[pulumi.Input['MdbSqlServerClusterUserPermissionArgs']]] permissions: Set of permissions granted to the user. The structure is documented below.
"""
pulumi.set(__self__, "name", name)
pulumi.set(__self__, "password", password)
if permissions is not None:
pulumi.set(__self__, "permissions", permissions)
@property
@pulumi.getter
def name(self) -> pulumi.Input[str]:
"""
The name of the database.
"""
return pulumi.get(self, "name")
@name.setter
def name(self, value: pulumi.Input[str]):
pulumi.set(self, "name", value)
@property
@pulumi.getter
def password(self) -> pulumi.Input[str]:
"""
The password of the user.
"""
return pulumi.get(self, "password")
@password.setter
def password(self, value: pulumi.Input[str]):
pulumi.set(self, "password", value)
@property
@pulumi.getter
def permissions(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['MdbSqlServerClusterUserPermissionArgs']]]]:
"""
Set of permissions granted to the user. The structure is documented below.
"""
return pulumi.get(self, "permissions")
@permissions.setter
def permissions(self, value: Optional[pulumi.Input[Sequence[pulumi.Input['MdbSqlServerClusterUserPermissionArgs']]]]):
pulumi.set(self, "permissions", value)
@pulumi.input_type
class MdbSqlServerClusterUserPermissionArgs:
def __init__(__self__, *,
database_name: pulumi.Input[str],
roles: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None):
"""
:param pulumi.Input[str] database_name: The name of the database that the permission grants access to.
:param pulumi.Input[Sequence[pulumi.Input[str]]] roles: List user's roles in the database.
Allowed roles: `OWNER`, `SECURITYADMIN`, `ACCESSADMIN`, `BACKUPOPERATOR`, `DDLADMIN`, `DATAWRITER`, `DATAREADER`, `DENYDATAWRITER`, `DENYDATAREADER`.
"""
pulumi.set(__self__, "database_name", database_name)
if roles is not None:
pulumi.set(__self__, "roles", roles)
@property
@pulumi.getter(name="databaseName")
def database_name(self) -> pulumi.Input[str]:
"""
The name of the database that the permission grants access to.
"""
return pulumi.get(self, "database_name")
@database_name.setter
def database_name(self, value: pulumi.Input[str]):
pulumi.set(self, "database_name", value)
@property
@pulumi.getter
def roles(self) -> Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]:
"""
List user's roles in the database.
Allowed roles: `OWNER`, `SECURITYADMIN`, `ACCESSADMIN`, `BACKUPOPERATOR`, `DDLADMIN`, `DATAWRITER`, `DATAREADER`, `DENYDATAWRITER`, `DENYDATAREADER`.
"""
return pulumi.get(self, "roles")
@roles.setter
def roles(self, value: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]):
pulumi.set(self, "roles", value)
@pulumi.input_type
class StorageBucketCorsRuleArgs:
def __init__(__self__, *,
allowed_methods: pulumi.Input[Sequence[pulumi.Input[str]]],
allowed_origins: pulumi.Input[Sequence[pulumi.Input[str]]],
allowed_headers: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None,
expose_headers: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None,
max_age_seconds: Optional[pulumi.Input[int]] = None):
"""
:param pulumi.Input[Sequence[pulumi.Input[str]]] allowed_methods: Specifies which methods are allowed. Can be `GET`, `PUT`, `POST`, `DELETE` or `HEAD`.
:param pulumi.Input[Sequence[pulumi.Input[str]]] allowed_origins: Specifies which origins are allowed.
:param pulumi.Input[Sequence[pulumi.Input[str]]] allowed_headers: Specifies which headers are allowed.
:param pulumi.Input[Sequence[pulumi.Input[str]]] expose_headers: Specifies expose header in the response.
:param pulumi.Input[int] max_age_seconds: Specifies time in seconds that browser can cache the response for a preflight request.
"""
pulumi.set(__self__, "allowed_methods", allowed_methods)
pulumi.set(__self__, "allowed_origins", allowed_origins)
if allowed_headers is not None:
pulumi.set(__self__, "allowed_headers", allowed_headers)
if expose_headers is not None:
pulumi.set(__self__, "expose_headers", expose_headers)
if max_age_seconds is not None:
pulumi.set(__self__, "max_age_seconds", max_age_seconds)
@property
@pulumi.getter(name="allowedMethods")
def allowed_methods(self) -> pulumi.Input[Sequence[pulumi.Input[str]]]:
"""
Specifies which methods are allowed. Can be `GET`, `PUT`, `POST`, `DELETE` or `HEAD`.
"""
return pulumi.get(self, "allowed_methods")
@allowed_methods.setter
def allowed_methods(self, value: pulumi.Input[Sequence[pulumi.Input[str]]]):
pulumi.set(self, "allowed_methods", value)
@property
@pulumi.getter(name="allowedOrigins")
def allowed_origins(self) -> pulumi.Input[Sequence[pulumi.Input[str]]]:
"""
Specifies which origins are allowed.
"""
return pulumi.get(self, "allowed_origins")
@allowed_origins.setter
def allowed_origins(self, value: pulumi.Input[Sequence[pulumi.Input[str]]]):
pulumi.set(self, "allowed_origins", value)
@property
@pulumi.getter(name="allowedHeaders")
def allowed_headers(self) -> Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]:
"""
Specifies which headers are allowed.
"""
return pulumi.get(self, "allowed_headers")
@allowed_headers.setter
def allowed_headers(self, value: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]):
pulumi.set(self, "allowed_headers", value)
@property
@pulumi.getter(name="exposeHeaders")
def expose_headers(self) -> Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]:
"""
Specifies expose header in the response.
"""
return pulumi.get(self, "expose_headers")
@expose_headers.setter
def expose_headers(self, value: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]):
pulumi.set(self, "expose_headers", value)
@property
@pulumi.getter(name="maxAgeSeconds")
def max_age_seconds(self) -> Optional[pulumi.Input[int]]:
"""
Specifies time in seconds that browser can cache the response for a preflight request.
"""
return pulumi.get(self, "max_age_seconds")
@max_age_seconds.setter
def max_age_seconds(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "max_age_seconds", value)
@pulumi.input_type
class StorageBucketGrantArgs:
def __init__(__self__, *,
permissions: pulumi.Input[Sequence[pulumi.Input[str]]],
type: pulumi.Input[str],
id: Optional[pulumi.Input[str]] = None,
uri: Optional[pulumi.Input[str]] = None):
"""
:param pulumi.Input[str] id: Unique identifier for the rule. Must be less than or equal to 255 characters in length.
"""
pulumi.set(__self__, "permissions", permissions)
pulumi.set(__self__, "type", type)
if id is not None:
pulumi.set(__self__, "id", id)
if uri is not None:
pulumi.set(__self__, "uri", uri)
@property
@pulumi.getter
def permissions(self) -> pulumi.Input[Sequence[pulumi.Input[str]]]:
return pulumi.get(self, "permissions")
@permissions.setter
def permissions(self, value: pulumi.Input[Sequence[pulumi.Input[str]]]):
pulumi.set(self, "permissions", value)
@property
@pulumi.getter
def type(self) -> pulumi.Input[str]:
return pulumi.get(self, "type")
@type.setter
def type(self, value: pulumi.Input[str]):
pulumi.set(self, "type", value)
@property
@pulumi.getter
def id(self) -> Optional[pulumi.Input[str]]:
"""
Unique identifier for the rule. Must be less than or equal to 255 characters in length.
"""
return pulumi.get(self, "id")
@id.setter
def id(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "id", value)
@property
@pulumi.getter
def uri(self) -> Optional[pulumi.Input[str]]:
return pulumi.get(self, "uri")
@uri.setter
def uri(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "uri", value)
@pulumi.input_type
class StorageBucketLifecycleRuleArgs:
def __init__(__self__, *,
enabled: pulumi.Input[bool],
abort_incomplete_multipart_upload_days: Optional[pulumi.Input[int]] = None,
expiration: Optional[pulumi.Input['StorageBucketLifecycleRuleExpirationArgs']] = None,
id: Optional[pulumi.Input[str]] = None,
noncurrent_version_expiration: Optional[pulumi.Input['StorageBucketLifecycleRuleNoncurrentVersionExpirationArgs']] = None,
noncurrent_version_transitions: Optional[pulumi.Input[Sequence[pulumi.Input['StorageBucketLifecycleRuleNoncurrentVersionTransitionArgs']]]] = None,
prefix: Optional[pulumi.Input[str]] = None,
transitions: Optional[pulumi.Input[Sequence[pulumi.Input['StorageBucketLifecycleRuleTransitionArgs']]]] = None):
"""
:param pulumi.Input[bool] enabled: Specifies lifecycle rule status.
:param pulumi.Input[int] abort_incomplete_multipart_upload_days: Specifies the number of days after initiating a multipart upload when the multipart upload must be completed.
:param pulumi.Input['StorageBucketLifecycleRuleExpirationArgs'] expiration: Specifies a period in the object's expire (documented below).
:param pulumi.Input[str] id: Unique identifier for the rule. Must be less than or equal to 255 characters in length.
:param pulumi.Input['StorageBucketLifecycleRuleNoncurrentVersionExpirationArgs'] noncurrent_version_expiration: Specifies when noncurrent object versions expire (documented below).
:param pulumi.Input[Sequence[pulumi.Input['StorageBucketLifecycleRuleNoncurrentVersionTransitionArgs']]] noncurrent_version_transitions: Specifies when noncurrent object versions transitions (documented below).
:param pulumi.Input[str] prefix: Object key prefix identifying one or more objects to which the rule applies.
:param pulumi.Input[Sequence[pulumi.Input['StorageBucketLifecycleRuleTransitionArgs']]] transitions: Specifies a period in the object's transitions (documented below).
"""
pulumi.set(__self__, "enabled", enabled)
if abort_incomplete_multipart_upload_days is not None:
pulumi.set(__self__, "abort_incomplete_multipart_upload_days", abort_incomplete_multipart_upload_days)
if expiration is not None:
pulumi.set(__self__, "expiration", expiration)
if id is not None:
pulumi.set(__self__, "id", id)
if noncurrent_version_expiration is not None:
pulumi.set(__self__, "noncurrent_version_expiration", noncurrent_version_expiration)
if noncurrent_version_transitions is not None:
pulumi.set(__self__, "noncurrent_version_transitions", noncurrent_version_transitions)
if prefix is not None:
pulumi.set(__self__, "prefix", prefix)
if transitions is not None:
pulumi.set(__self__, "transitions", transitions)
@property
@pulumi.getter
def enabled(self) -> pulumi.Input[bool]:
"""
Specifies lifecycle rule status.
"""
return pulumi.get(self, "enabled")
@enabled.setter
def enabled(self, value: pulumi.Input[bool]):
pulumi.set(self, "enabled", value)
@property
@pulumi.getter(name="abortIncompleteMultipartUploadDays")
def abort_incomplete_multipart_upload_days(self) -> Optional[pulumi.Input[int]]:
"""
Specifies the number of days after initiating a multipart upload when the multipart upload must be completed.
"""
return pulumi.get(self, "abort_incomplete_multipart_upload_days")
@abort_incomplete_multipart_upload_days.setter
def abort_incomplete_multipart_upload_days(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "abort_incomplete_multipart_upload_days", value)
@property
@pulumi.getter
def expiration(self) -> Optional[pulumi.Input['StorageBucketLifecycleRuleExpirationArgs']]:
"""
Specifies a period in the object's expire (documented below).
"""
return pulumi.get(self, "expiration")
@expiration.setter
def expiration(self, value: Optional[pulumi.Input['StorageBucketLifecycleRuleExpirationArgs']]):
pulumi.set(self, "expiration", value)
@property
@pulumi.getter
def id(self) -> Optional[pulumi.Input[str]]:
"""
Unique identifier for the rule. Must be less than or equal to 255 characters in length.
"""
return pulumi.get(self, "id")
@id.setter
def id(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "id", value)
@property
@pulumi.getter(name="noncurrentVersionExpiration")
def noncurrent_version_expiration(self) -> Optional[pulumi.Input['StorageBucketLifecycleRuleNoncurrentVersionExpirationArgs']]:
"""
Specifies when noncurrent object versions expire (documented below).
"""
return pulumi.get(self, "noncurrent_version_expiration")
@noncurrent_version_expiration.setter
def noncurrent_version_expiration(self, value: Optional[pulumi.Input['StorageBucketLifecycleRuleNoncurrentVersionExpirationArgs']]):
pulumi.set(self, "noncurrent_version_expiration", value)
@property
@pulumi.getter(name="noncurrentVersionTransitions")
def noncurrent_version_transitions(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['StorageBucketLifecycleRuleNoncurrentVersionTransitionArgs']]]]:
"""
Specifies when noncurrent object versions transitions (documented below).
"""
return pulumi.get(self, "noncurrent_version_transitions")
@noncurrent_version_transitions.setter
def noncurrent_version_transitions(self, value: Optional[pulumi.Input[Sequence[pulumi.Input['StorageBucketLifecycleRuleNoncurrentVersionTransitionArgs']]]]):
pulumi.set(self, "noncurrent_version_transitions", value)
@property
@pulumi.getter
def prefix(self) -> Optional[pulumi.Input[str]]:
"""
Object key prefix identifying one or more objects to which the rule applies.
"""
return pulumi.get(self, "prefix")
@prefix.setter
def prefix(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "prefix", value)
@property
@pulumi.getter
def transitions(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['StorageBucketLifecycleRuleTransitionArgs']]]]:
"""
Specifies a period in the object's transitions (documented below).
"""
return pulumi.get(self, "transitions")
@transitions.setter
def transitions(self, value: Optional[pulumi.Input[Sequence[pulumi.Input['StorageBucketLifecycleRuleTransitionArgs']]]]):
pulumi.set(self, "transitions", value)
@pulumi.input_type
class StorageBucketLifecycleRuleExpirationArgs:
def __init__(__self__, *,
date: Optional[pulumi.Input[str]] = None,
days: Optional[pulumi.Input[int]] = None,
expired_object_delete_marker: Optional[pulumi.Input[bool]] = None):
"""
:param pulumi.Input[str] date: Specifies the date after which you want the corresponding action to take effect.
:param pulumi.Input[int] days: Specifies the number of days after object creation when the specific rule action takes effect.
:param pulumi.Input[bool] expired_object_delete_marker: On a versioned bucket (versioning-enabled or versioning-suspended bucket), you can add this element in the lifecycle configuration to direct Object Storage to delete expired object delete markers.
"""
if date is not None:
pulumi.set(__self__, "date", date)
if days is not None:
pulumi.set(__self__, "days", days)
if expired_object_delete_marker is not None:
pulumi.set(__self__, "expired_object_delete_marker", expired_object_delete_marker)
@property
@pulumi.getter
def date(self) -> Optional[pulumi.Input[str]]:
"""
Specifies the date after which you want the corresponding action to take effect.
"""
return pulumi.get(self, "date")
@date.setter
def date(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "date", value)
@property
@pulumi.getter
def days(self) -> Optional[pulumi.Input[int]]:
"""
Specifies the number of days after object creation when the specific rule action takes effect.
"""
return pulumi.get(self, "days")
@days.setter
def days(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "days", value)
@property
@pulumi.getter(name="expiredObjectDeleteMarker")
def expired_object_delete_marker(self) -> Optional[pulumi.Input[bool]]:
"""
On a versioned bucket (versioning-enabled or versioning-suspended bucket), you can add this element in the lifecycle configuration to direct Object Storage to delete expired object delete markers.
"""
return pulumi.get(self, "expired_object_delete_marker")
@expired_object_delete_marker.setter
def expired_object_delete_marker(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "expired_object_delete_marker", value)
@pulumi.input_type
class StorageBucketLifecycleRuleNoncurrentVersionExpirationArgs:
def __init__(__self__, *,
days: Optional[pulumi.Input[int]] = None):
"""
:param pulumi.Input[int] days: Specifies the number of days noncurrent object versions expire.
"""
if days is not None:
pulumi.set(__self__, "days", days)
@property
@pulumi.getter
def days(self) -> Optional[pulumi.Input[int]]:
"""
Specifies the number of days noncurrent object versions expire.
"""
return pulumi.get(self, "days")
@days.setter
def days(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "days", value)
@pulumi.input_type
class StorageBucketLifecycleRuleNoncurrentVersionTransitionArgs:
def __init__(__self__, *,
storage_class: pulumi.Input[str],
days: Optional[pulumi.Input[int]] = None):
"""
:param pulumi.Input[str] storage_class: Specifies the storage class to which you want the noncurrent object versions to transition. Can only be `COLD` or `STANDARD_IA`.
:param pulumi.Input[int] days: Specifies the number of days noncurrent object versions transition.
"""
pulumi.set(__self__, "storage_class", storage_class)
if days is not None:
pulumi.set(__self__, "days", days)
@property
@pulumi.getter(name="storageClass")
def storage_class(self) -> pulumi.Input[str]:
"""
Specifies the storage class to which you want the noncurrent object versions to transition. Can only be `COLD` or `STANDARD_IA`.
"""
return pulumi.get(self, "storage_class")
@storage_class.setter
def storage_class(self, value: pulumi.Input[str]):
pulumi.set(self, "storage_class", value)
@property
@pulumi.getter
def days(self) -> Optional[pulumi.Input[int]]:
"""
Specifies the number of days noncurrent object versions transition.
"""
return pulumi.get(self, "days")
@days.setter
def days(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "days", value)
@pulumi.input_type
class StorageBucketLifecycleRuleTransitionArgs:
def __init__(__self__, *,
storage_class: pulumi.Input[str],
date: Optional[pulumi.Input[str]] = None,
days: Optional[pulumi.Input[int]] = None):
"""
:param pulumi.Input[str] storage_class: Specifies the storage class to which you want the object to transition. Can only be `COLD` or `STANDARD_IA`.
:param pulumi.Input[str] date: Specifies the date after which you want the corresponding action to take effect.
:param pulumi.Input[int] days: Specifies the number of days after object creation when the specific rule action takes effect.
"""
pulumi.set(__self__, "storage_class", storage_class)
if date is not None:
pulumi.set(__self__, "date", date)
if days is not None:
pulumi.set(__self__, "days", days)
@property
@pulumi.getter(name="storageClass")
def storage_class(self) -> pulumi.Input[str]:
"""
Specifies the storage class to which you want the object to transition. Can only be `COLD` or `STANDARD_IA`.
"""
return pulumi.get(self, "storage_class")
@storage_class.setter
def storage_class(self, value: pulumi.Input[str]):
pulumi.set(self, "storage_class", value)
@property
@pulumi.getter
def date(self) -> Optional[pulumi.Input[str]]:
"""
Specifies the date after which you want the corresponding action to take effect.
"""
return pulumi.get(self, "date")
@date.setter
def date(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "date", value)
@property
@pulumi.getter
def days(self) -> Optional[pulumi.Input[int]]:
"""
Specifies the number of days after object creation when the specific rule action takes effect.
"""
return pulumi.get(self, "days")
@days.setter
def days(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "days", value)
@pulumi.input_type
class StorageBucketLoggingArgs:
def __init__(__self__, *,
target_bucket: pulumi.Input[str],
target_prefix: Optional[pulumi.Input[str]] = None):
"""
:param pulumi.Input[str] target_bucket: The name of the bucket that will receive the log objects.
:param pulumi.Input[str] target_prefix: To specify a key prefix for log objects.
"""
pulumi.set(__self__, "target_bucket", target_bucket)
if target_prefix is not None:
pulumi.set(__self__, "target_prefix", target_prefix)
@property
@pulumi.getter(name="targetBucket")
def target_bucket(self) -> pulumi.Input[str]:
"""
The name of the bucket that will receive the log objects.
"""
return pulumi.get(self, "target_bucket")
@target_bucket.setter
def target_bucket(self, value: pulumi.Input[str]):
pulumi.set(self, "target_bucket", value)
@property
@pulumi.getter(name="targetPrefix")
def target_prefix(self) -> Optional[pulumi.Input[str]]:
"""
To specify a key prefix for log objects.
"""
return pulumi.get(self, "target_prefix")
@target_prefix.setter
def target_prefix(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "target_prefix", value)
@pulumi.input_type
class StorageBucketServerSideEncryptionConfigurationArgs:
def __init__(__self__, *,
rule: pulumi.Input['StorageBucketServerSideEncryptionConfigurationRuleArgs']):
"""
:param pulumi.Input['StorageBucketServerSideEncryptionConfigurationRuleArgs'] rule: A single object for server-side encryption by default configuration. (documented below)
"""
pulumi.set(__self__, "rule", rule)
@property
@pulumi.getter
def rule(self) -> pulumi.Input['StorageBucketServerSideEncryptionConfigurationRuleArgs']:
"""
A single object for server-side encryption by default configuration. (documented below)
"""
return pulumi.get(self, "rule")
@rule.setter
def rule(self, value: pulumi.Input['StorageBucketServerSideEncryptionConfigurationRuleArgs']):
pulumi.set(self, "rule", value)
@pulumi.input_type
class StorageBucketServerSideEncryptionConfigurationRuleArgs:
def __init__(__self__, *,
apply_server_side_encryption_by_default: pulumi.Input['StorageBucketServerSideEncryptionConfigurationRuleApplyServerSideEncryptionByDefaultArgs']):
"""
:param pulumi.Input['StorageBucketServerSideEncryptionConfigurationRuleApplyServerSideEncryptionByDefaultArgs'] apply_server_side_encryption_by_default: A single object for setting server-side encryption by default. (documented below)
"""
pulumi.set(__self__, "apply_server_side_encryption_by_default", apply_server_side_encryption_by_default)
@property
@pulumi.getter(name="applyServerSideEncryptionByDefault")
def apply_server_side_encryption_by_default(self) -> pulumi.Input['StorageBucketServerSideEncryptionConfigurationRuleApplyServerSideEncryptionByDefaultArgs']:
"""
A single object for setting server-side encryption by default. (documented below)
"""
return pulumi.get(self, "apply_server_side_encryption_by_default")
@apply_server_side_encryption_by_default.setter
def apply_server_side_encryption_by_default(self, value: pulumi.Input['StorageBucketServerSideEncryptionConfigurationRuleApplyServerSideEncryptionByDefaultArgs']):
pulumi.set(self, "apply_server_side_encryption_by_default", value)
@pulumi.input_type
class StorageBucketServerSideEncryptionConfigurationRuleApplyServerSideEncryptionByDefaultArgs:
def __init__(__self__, *,
kms_master_key_id: pulumi.Input[str],
sse_algorithm: pulumi.Input[str]):
"""
:param pulumi.Input[str] kms_master_key_id: The KMS master key ID used for the SSE-KMS encryption.
:param pulumi.Input[str] sse_algorithm: The server-side encryption algorithm to use. Single valid value is `aws:kms`
"""
pulumi.set(__self__, "kms_master_key_id", kms_master_key_id)
pulumi.set(__self__, "sse_algorithm", sse_algorithm)
@property
@pulumi.getter(name="kmsMasterKeyId")
def kms_master_key_id(self) -> pulumi.Input[str]:
"""
The KMS master key ID used for the SSE-KMS encryption.
"""
return pulumi.get(self, "kms_master_key_id")
@kms_master_key_id.setter
def kms_master_key_id(self, value: pulumi.Input[str]):
pulumi.set(self, "kms_master_key_id", value)
@property
@pulumi.getter(name="sseAlgorithm")
def sse_algorithm(self) -> pulumi.Input[str]:
"""
The server-side encryption algorithm to use. Single valid value is `aws:kms`
"""
return pulumi.get(self, "sse_algorithm")
@sse_algorithm.setter
def sse_algorithm(self, value: pulumi.Input[str]):
pulumi.set(self, "sse_algorithm", value)
@pulumi.input_type
class StorageBucketVersioningArgs:
def __init__(__self__, *,
enabled: Optional[pulumi.Input[bool]] = None):
"""
:param pulumi.Input[bool] enabled: Enable versioning. Once you version-enable a bucket, it can never return to an unversioned state. You can, however, suspend versioning on that bucket.
"""
if enabled is not None:
pulumi.set(__self__, "enabled", enabled)
@property
@pulumi.getter
def enabled(self) -> Optional[pulumi.Input[bool]]:
"""
Enable versioning. Once you version-enable a bucket, it can never return to an unversioned state. You can, however, suspend versioning on that bucket.
"""
return pulumi.get(self, "enabled")
@enabled.setter
def enabled(self, value: Optional[pulumi.Input[bool]]):
pulumi.set(self, "enabled", value)
@pulumi.input_type
class StorageBucketWebsiteArgs:
def __init__(__self__, *,
error_document: Optional[pulumi.Input[str]] = None,
index_document: Optional[pulumi.Input[str]] = None,
redirect_all_requests_to: Optional[pulumi.Input[str]] = None,
routing_rules: Optional[pulumi.Input[str]] = None):
"""
:param pulumi.Input[str] error_document: An absolute path to the document to return in case of a 4XX error.
:param pulumi.Input[str] index_document: Storage returns this index document when requests are made to the root domain or any of the subfolders.
:param pulumi.Input[str] redirect_all_requests_to: A hostname to redirect all website requests for this bucket to. Hostname can optionally be prefixed with a protocol (`http://` or `https://`) to use when redirecting requests. The default is the protocol that is used in the original request.
:param pulumi.Input[str] routing_rules: A json array containing [routing rules](https://cloud.yandex.com/docs/storage/s3/api-ref/hosting/upload#request-scheme) describing redirect behavior and when redirects are applied.
"""
if error_document is not None:
pulumi.set(__self__, "error_document", error_document)
if index_document is not None:
pulumi.set(__self__, "index_document", index_document)
if redirect_all_requests_to is not None:
pulumi.set(__self__, "redirect_all_requests_to", redirect_all_requests_to)
if routing_rules is not None:
pulumi.set(__self__, "routing_rules", routing_rules)
@property
@pulumi.getter(name="errorDocument")
def error_document(self) -> Optional[pulumi.Input[str]]:
"""
An absolute path to the document to return in case of a 4XX error.
"""
return pulumi.get(self, "error_document")
@error_document.setter
def error_document(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "error_document", value)
@property
@pulumi.getter(name="indexDocument")
def index_document(self) -> Optional[pulumi.Input[str]]:
"""
Storage returns this index document when requests are made to the root domain or any of the subfolders.
"""
return pulumi.get(self, "index_document")
@index_document.setter
def index_document(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "index_document", value)
@property
@pulumi.getter(name="redirectAllRequestsTo")
def redirect_all_requests_to(self) -> Optional[pulumi.Input[str]]:
"""
A hostname to redirect all website requests for this bucket to. Hostname can optionally be prefixed with a protocol (`http://` or `https://`) to use when redirecting requests. The default is the protocol that is used in the original request.
"""
return pulumi.get(self, "redirect_all_requests_to")
@redirect_all_requests_to.setter
def redirect_all_requests_to(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "redirect_all_requests_to", value)
@property
@pulumi.getter(name="routingRules")
def routing_rules(self) -> Optional[pulumi.Input[str]]:
"""
A json array containing [routing rules](https://cloud.yandex.com/docs/storage/s3/api-ref/hosting/upload#request-scheme) describing redirect behavior and when redirects are applied.
"""
return pulumi.get(self, "routing_rules")
@routing_rules.setter
def routing_rules(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "routing_rules", value)
@pulumi.input_type
class VpcAddressExternalIpv4AddressArgs:
def __init__(__self__, *,
address: Optional[pulumi.Input[str]] = None,
ddos_protection_provider: Optional[pulumi.Input[str]] = None,
outgoing_smtp_capability: Optional[pulumi.Input[str]] = None,
zone_id: Optional[pulumi.Input[str]] = None):
"""
:param pulumi.Input[str] address: - Desired IP.
:param pulumi.Input[str] ddos_protection_provider: Enable DDOS protection. Possible values are: "qrator"
:param pulumi.Input[str] outgoing_smtp_capability: Wanted outgoing smtp capability.
:param pulumi.Input[str] zone_id: - Zone for allocating address.
"""
if address is not None:
pulumi.set(__self__, "address", address)
if ddos_protection_provider is not None:
pulumi.set(__self__, "ddos_protection_provider", ddos_protection_provider)
if outgoing_smtp_capability is not None:
pulumi.set(__self__, "outgoing_smtp_capability", outgoing_smtp_capability)
if zone_id is not None:
pulumi.set(__self__, "zone_id", zone_id)
@property
@pulumi.getter
def address(self) -> Optional[pulumi.Input[str]]:
"""
- Desired IP.
"""
return pulumi.get(self, "address")
@address.setter
def address(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "address", value)
@property
@pulumi.getter(name="ddosProtectionProvider")
def ddos_protection_provider(self) -> Optional[pulumi.Input[str]]:
"""
Enable DDOS protection. Possible values are: "qrator"
"""
return pulumi.get(self, "ddos_protection_provider")
@ddos_protection_provider.setter
def ddos_protection_provider(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "ddos_protection_provider", value)
@property
@pulumi.getter(name="outgoingSmtpCapability")
def outgoing_smtp_capability(self) -> Optional[pulumi.Input[str]]:
"""
Wanted outgoing smtp capability.
"""
return pulumi.get(self, "outgoing_smtp_capability")
@outgoing_smtp_capability.setter
def outgoing_smtp_capability(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "outgoing_smtp_capability", value)
@property
@pulumi.getter(name="zoneId")
def zone_id(self) -> Optional[pulumi.Input[str]]:
"""
- Zone for allocating address.
"""
return pulumi.get(self, "zone_id")
@zone_id.setter
def zone_id(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "zone_id", value)
@pulumi.input_type
class VpcDefaultSecurityGroupEgressArgs:
def __init__(__self__, *,
protocol: pulumi.Input[str],
description: Optional[pulumi.Input[str]] = None,
from_port: Optional[pulumi.Input[int]] = None,
id: Optional[pulumi.Input[str]] = None,
labels: Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]] = None,
port: Optional[pulumi.Input[int]] = None,
predefined_target: Optional[pulumi.Input[str]] = None,
security_group_id: Optional[pulumi.Input[str]] = None,
to_port: Optional[pulumi.Input[int]] = None,
v4_cidr_blocks: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None,
v6_cidr_blocks: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None):
"""
:param pulumi.Input[str] description: Description of the security group.
:param pulumi.Input[str] id: Id of the security group.
:param pulumi.Input[Mapping[str, pulumi.Input[str]]] labels: Labels to assign to this security group.
"""
pulumi.set(__self__, "protocol", protocol)
if description is not None:
pulumi.set(__self__, "description", description)
if from_port is not None:
pulumi.set(__self__, "from_port", from_port)
if id is not None:
pulumi.set(__self__, "id", id)
if labels is not None:
pulumi.set(__self__, "labels", labels)
if port is not None:
pulumi.set(__self__, "port", port)
if predefined_target is not None:
pulumi.set(__self__, "predefined_target", predefined_target)
if security_group_id is not None:
pulumi.set(__self__, "security_group_id", security_group_id)
if to_port is not None:
pulumi.set(__self__, "to_port", to_port)
if v4_cidr_blocks is not None:
pulumi.set(__self__, "v4_cidr_blocks", v4_cidr_blocks)
if v6_cidr_blocks is not None:
pulumi.set(__self__, "v6_cidr_blocks", v6_cidr_blocks)
@property
@pulumi.getter
def protocol(self) -> pulumi.Input[str]:
return pulumi.get(self, "protocol")
@protocol.setter
def protocol(self, value: pulumi.Input[str]):
pulumi.set(self, "protocol", value)
@property
@pulumi.getter
def description(self) -> Optional[pulumi.Input[str]]:
"""
Description of the security group.
"""
return pulumi.get(self, "description")
@description.setter
def description(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "description", value)
@property
@pulumi.getter(name="fromPort")
def from_port(self) -> Optional[pulumi.Input[int]]:
return pulumi.get(self, "from_port")
@from_port.setter
def from_port(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "from_port", value)
@property
@pulumi.getter
def id(self) -> Optional[pulumi.Input[str]]:
"""
Id of the security group.
"""
return pulumi.get(self, "id")
@id.setter
def id(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "id", value)
@property
@pulumi.getter
def labels(self) -> Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]]:
"""
Labels to assign to this security group.
"""
return pulumi.get(self, "labels")
@labels.setter
def labels(self, value: Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]]):
pulumi.set(self, "labels", value)
@property
@pulumi.getter
def port(self) -> Optional[pulumi.Input[int]]:
return pulumi.get(self, "port")
@port.setter
def port(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "port", value)
@property
@pulumi.getter(name="predefinedTarget")
def predefined_target(self) -> Optional[pulumi.Input[str]]:
return pulumi.get(self, "predefined_target")
@predefined_target.setter
def predefined_target(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "predefined_target", value)
@property
@pulumi.getter(name="securityGroupId")
def security_group_id(self) -> Optional[pulumi.Input[str]]:
return pulumi.get(self, "security_group_id")
@security_group_id.setter
def security_group_id(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "security_group_id", value)
@property
@pulumi.getter(name="toPort")
def to_port(self) -> Optional[pulumi.Input[int]]:
return pulumi.get(self, "to_port")
@to_port.setter
def to_port(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "to_port", value)
@property
@pulumi.getter(name="v4CidrBlocks")
def v4_cidr_blocks(self) -> Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]:
return pulumi.get(self, "v4_cidr_blocks")
@v4_cidr_blocks.setter
def v4_cidr_blocks(self, value: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]):
pulumi.set(self, "v4_cidr_blocks", value)
@property
@pulumi.getter(name="v6CidrBlocks")
def v6_cidr_blocks(self) -> Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]:
return pulumi.get(self, "v6_cidr_blocks")
@v6_cidr_blocks.setter
def v6_cidr_blocks(self, value: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]):
pulumi.set(self, "v6_cidr_blocks", value)
@pulumi.input_type
class VpcDefaultSecurityGroupIngressArgs:
def __init__(__self__, *,
protocol: pulumi.Input[str],
description: Optional[pulumi.Input[str]] = None,
from_port: Optional[pulumi.Input[int]] = None,
id: Optional[pulumi.Input[str]] = None,
labels: Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]] = None,
port: Optional[pulumi.Input[int]] = None,
predefined_target: Optional[pulumi.Input[str]] = None,
security_group_id: Optional[pulumi.Input[str]] = None,
to_port: Optional[pulumi.Input[int]] = None,
v4_cidr_blocks: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None,
v6_cidr_blocks: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None):
"""
:param pulumi.Input[str] description: Description of the security group.
:param pulumi.Input[str] id: Id of the security group.
:param pulumi.Input[Mapping[str, pulumi.Input[str]]] labels: Labels to assign to this security group.
"""
pulumi.set(__self__, "protocol", protocol)
if description is not None:
pulumi.set(__self__, "description", description)
if from_port is not None:
pulumi.set(__self__, "from_port", from_port)
if id is not None:
pulumi.set(__self__, "id", id)
if labels is not None:
pulumi.set(__self__, "labels", labels)
if port is not None:
pulumi.set(__self__, "port", port)
if predefined_target is not None:
pulumi.set(__self__, "predefined_target", predefined_target)
if security_group_id is not None:
pulumi.set(__self__, "security_group_id", security_group_id)
if to_port is not None:
pulumi.set(__self__, "to_port", to_port)
if v4_cidr_blocks is not None:
pulumi.set(__self__, "v4_cidr_blocks", v4_cidr_blocks)
if v6_cidr_blocks is not None:
pulumi.set(__self__, "v6_cidr_blocks", v6_cidr_blocks)
@property
@pulumi.getter
def protocol(self) -> pulumi.Input[str]:
return pulumi.get(self, "protocol")
@protocol.setter
def protocol(self, value: pulumi.Input[str]):
pulumi.set(self, "protocol", value)
@property
@pulumi.getter
def description(self) -> Optional[pulumi.Input[str]]:
"""
Description of the security group.
"""
return pulumi.get(self, "description")
@description.setter
def description(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "description", value)
@property
@pulumi.getter(name="fromPort")
def from_port(self) -> Optional[pulumi.Input[int]]:
return pulumi.get(self, "from_port")
@from_port.setter
def from_port(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "from_port", value)
@property
@pulumi.getter
def id(self) -> Optional[pulumi.Input[str]]:
"""
Id of the security group.
"""
return pulumi.get(self, "id")
@id.setter
def id(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "id", value)
@property
@pulumi.getter
def labels(self) -> Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]]:
"""
Labels to assign to this security group.
"""
return pulumi.get(self, "labels")
@labels.setter
def labels(self, value: Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]]):
pulumi.set(self, "labels", value)
@property
@pulumi.getter
def port(self) -> Optional[pulumi.Input[int]]:
return pulumi.get(self, "port")
@port.setter
def port(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "port", value)
@property
@pulumi.getter(name="predefinedTarget")
def predefined_target(self) -> Optional[pulumi.Input[str]]:
return pulumi.get(self, "predefined_target")
@predefined_target.setter
def predefined_target(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "predefined_target", value)
@property
@pulumi.getter(name="securityGroupId")
def security_group_id(self) -> Optional[pulumi.Input[str]]:
return pulumi.get(self, "security_group_id")
@security_group_id.setter
def security_group_id(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "security_group_id", value)
@property
@pulumi.getter(name="toPort")
def to_port(self) -> Optional[pulumi.Input[int]]:
return pulumi.get(self, "to_port")
@to_port.setter
def to_port(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "to_port", value)
@property
@pulumi.getter(name="v4CidrBlocks")
def v4_cidr_blocks(self) -> Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]:
return pulumi.get(self, "v4_cidr_blocks")
@v4_cidr_blocks.setter
def v4_cidr_blocks(self, value: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]):
pulumi.set(self, "v4_cidr_blocks", value)
@property
@pulumi.getter(name="v6CidrBlocks")
def v6_cidr_blocks(self) -> Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]:
return pulumi.get(self, "v6_cidr_blocks")
@v6_cidr_blocks.setter
def v6_cidr_blocks(self, value: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]):
pulumi.set(self, "v6_cidr_blocks", value)
@pulumi.input_type
class VpcRouteTableStaticRouteArgs:
def __init__(__self__, *,
destination_prefix: Optional[pulumi.Input[str]] = None,
next_hop_address: Optional[pulumi.Input[str]] = None):
"""
:param pulumi.Input[str] destination_prefix: Route prefix in CIDR notation.
:param pulumi.Input[str] next_hop_address: Address of the next hop.
"""
if destination_prefix is not None:
pulumi.set(__self__, "destination_prefix", destination_prefix)
if next_hop_address is not None:
pulumi.set(__self__, "next_hop_address", next_hop_address)
@property
@pulumi.getter(name="destinationPrefix")
def destination_prefix(self) -> Optional[pulumi.Input[str]]:
"""
Route prefix in CIDR notation.
"""
return pulumi.get(self, "destination_prefix")
@destination_prefix.setter
def destination_prefix(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "destination_prefix", value)
@property
@pulumi.getter(name="nextHopAddress")
def next_hop_address(self) -> Optional[pulumi.Input[str]]:
"""
Address of the next hop.
"""
return pulumi.get(self, "next_hop_address")
@next_hop_address.setter
def next_hop_address(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "next_hop_address", value)
@pulumi.input_type
class VpcSecurityGroupEgressArgs:
def __init__(__self__, *,
protocol: pulumi.Input[str],
description: Optional[pulumi.Input[str]] = None,
from_port: Optional[pulumi.Input[int]] = None,
id: Optional[pulumi.Input[str]] = None,
labels: Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]] = None,
port: Optional[pulumi.Input[int]] = None,
predefined_target: Optional[pulumi.Input[str]] = None,
security_group_id: Optional[pulumi.Input[str]] = None,
to_port: Optional[pulumi.Input[int]] = None,
v4_cidr_blocks: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None,
v6_cidr_blocks: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None):
"""
:param pulumi.Input[str] description: Description of the security group.
:param pulumi.Input[str] id: Id of the rule.
:param pulumi.Input[Mapping[str, pulumi.Input[str]]] labels: Labels to assign to this security group.
"""
pulumi.set(__self__, "protocol", protocol)
if description is not None:
pulumi.set(__self__, "description", description)
if from_port is not None:
pulumi.set(__self__, "from_port", from_port)
if id is not None:
pulumi.set(__self__, "id", id)
if labels is not None:
pulumi.set(__self__, "labels", labels)
if port is not None:
pulumi.set(__self__, "port", port)
if predefined_target is not None:
pulumi.set(__self__, "predefined_target", predefined_target)
if security_group_id is not None:
pulumi.set(__self__, "security_group_id", security_group_id)
if to_port is not None:
pulumi.set(__self__, "to_port", to_port)
if v4_cidr_blocks is not None:
pulumi.set(__self__, "v4_cidr_blocks", v4_cidr_blocks)
if v6_cidr_blocks is not None:
pulumi.set(__self__, "v6_cidr_blocks", v6_cidr_blocks)
@property
@pulumi.getter
def protocol(self) -> pulumi.Input[str]:
return pulumi.get(self, "protocol")
@protocol.setter
def protocol(self, value: pulumi.Input[str]):
pulumi.set(self, "protocol", value)
@property
@pulumi.getter
def description(self) -> Optional[pulumi.Input[str]]:
"""
Description of the security group.
"""
return pulumi.get(self, "description")
@description.setter
def description(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "description", value)
@property
@pulumi.getter(name="fromPort")
def from_port(self) -> Optional[pulumi.Input[int]]:
return pulumi.get(self, "from_port")
@from_port.setter
def from_port(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "from_port", value)
@property
@pulumi.getter
def id(self) -> Optional[pulumi.Input[str]]:
"""
Id of the rule.
"""
return pulumi.get(self, "id")
@id.setter
def id(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "id", value)
@property
@pulumi.getter
def labels(self) -> Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]]:
"""
Labels to assign to this security group.
"""
return pulumi.get(self, "labels")
@labels.setter
def labels(self, value: Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]]):
pulumi.set(self, "labels", value)
@property
@pulumi.getter
def port(self) -> Optional[pulumi.Input[int]]:
return pulumi.get(self, "port")
@port.setter
def port(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "port", value)
@property
@pulumi.getter(name="predefinedTarget")
def predefined_target(self) -> Optional[pulumi.Input[str]]:
return pulumi.get(self, "predefined_target")
@predefined_target.setter
def predefined_target(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "predefined_target", value)
@property
@pulumi.getter(name="securityGroupId")
def security_group_id(self) -> Optional[pulumi.Input[str]]:
return pulumi.get(self, "security_group_id")
@security_group_id.setter
def security_group_id(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "security_group_id", value)
@property
@pulumi.getter(name="toPort")
def to_port(self) -> Optional[pulumi.Input[int]]:
return pulumi.get(self, "to_port")
@to_port.setter
def to_port(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "to_port", value)
@property
@pulumi.getter(name="v4CidrBlocks")
def v4_cidr_blocks(self) -> Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]:
return pulumi.get(self, "v4_cidr_blocks")
@v4_cidr_blocks.setter
def v4_cidr_blocks(self, value: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]):
pulumi.set(self, "v4_cidr_blocks", value)
@property
@pulumi.getter(name="v6CidrBlocks")
def v6_cidr_blocks(self) -> Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]:
return pulumi.get(self, "v6_cidr_blocks")
@v6_cidr_blocks.setter
def v6_cidr_blocks(self, value: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]):
pulumi.set(self, "v6_cidr_blocks", value)
@pulumi.input_type
class VpcSecurityGroupIngressArgs:
def __init__(__self__, *,
protocol: pulumi.Input[str],
description: Optional[pulumi.Input[str]] = None,
from_port: Optional[pulumi.Input[int]] = None,
id: Optional[pulumi.Input[str]] = None,
labels: Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]] = None,
port: Optional[pulumi.Input[int]] = None,
predefined_target: Optional[pulumi.Input[str]] = None,
security_group_id: Optional[pulumi.Input[str]] = None,
to_port: Optional[pulumi.Input[int]] = None,
v4_cidr_blocks: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None,
v6_cidr_blocks: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None):
"""
:param pulumi.Input[str] description: Description of the security group.
:param pulumi.Input[str] id: Id of the rule.
:param pulumi.Input[Mapping[str, pulumi.Input[str]]] labels: Labels to assign to this security group.
"""
pulumi.set(__self__, "protocol", protocol)
if description is not None:
pulumi.set(__self__, "description", description)
if from_port is not None:
pulumi.set(__self__, "from_port", from_port)
if id is not None:
pulumi.set(__self__, "id", id)
if labels is not None:
pulumi.set(__self__, "labels", labels)
if port is not None:
pulumi.set(__self__, "port", port)
if predefined_target is not None:
pulumi.set(__self__, "predefined_target", predefined_target)
if security_group_id is not None:
pulumi.set(__self__, "security_group_id", security_group_id)
if to_port is not None:
pulumi.set(__self__, "to_port", to_port)
if v4_cidr_blocks is not None:
pulumi.set(__self__, "v4_cidr_blocks", v4_cidr_blocks)
if v6_cidr_blocks is not None:
pulumi.set(__self__, "v6_cidr_blocks", v6_cidr_blocks)
@property
@pulumi.getter
def protocol(self) -> pulumi.Input[str]:
return pulumi.get(self, "protocol")
@protocol.setter
def protocol(self, value: pulumi.Input[str]):
pulumi.set(self, "protocol", value)
@property
@pulumi.getter
def description(self) -> Optional[pulumi.Input[str]]:
"""
Description of the security group.
"""
return pulumi.get(self, "description")
@description.setter
def description(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "description", value)
@property
@pulumi.getter(name="fromPort")
def from_port(self) -> Optional[pulumi.Input[int]]:
return pulumi.get(self, "from_port")
@from_port.setter
def from_port(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "from_port", value)
@property
@pulumi.getter
def id(self) -> Optional[pulumi.Input[str]]:
"""
Id of the rule.
"""
return pulumi.get(self, "id")
@id.setter
def id(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "id", value)
@property
@pulumi.getter
def labels(self) -> Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]]:
"""
Labels to assign to this security group.
"""
return pulumi.get(self, "labels")
@labels.setter
def labels(self, value: Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]]):
pulumi.set(self, "labels", value)
@property
@pulumi.getter
def port(self) -> Optional[pulumi.Input[int]]:
return pulumi.get(self, "port")
@port.setter
def port(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "port", value)
@property
@pulumi.getter(name="predefinedTarget")
def predefined_target(self) -> Optional[pulumi.Input[str]]:
return pulumi.get(self, "predefined_target")
@predefined_target.setter
def predefined_target(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "predefined_target", value)
@property
@pulumi.getter(name="securityGroupId")
def security_group_id(self) -> Optional[pulumi.Input[str]]:
return pulumi.get(self, "security_group_id")
@security_group_id.setter
def security_group_id(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "security_group_id", value)
@property
@pulumi.getter(name="toPort")
def to_port(self) -> Optional[pulumi.Input[int]]:
return pulumi.get(self, "to_port")
@to_port.setter
def to_port(self, value: Optional[pulumi.Input[int]]):
pulumi.set(self, "to_port", value)
@property
@pulumi.getter(name="v4CidrBlocks")
def v4_cidr_blocks(self) -> Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]:
return pulumi.get(self, "v4_cidr_blocks")
@v4_cidr_blocks.setter
def v4_cidr_blocks(self, value: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]):
pulumi.set(self, "v4_cidr_blocks", value)
@property
@pulumi.getter(name="v6CidrBlocks")
def v6_cidr_blocks(self) -> Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]:
return pulumi.get(self, "v6_cidr_blocks")
@v6_cidr_blocks.setter
def v6_cidr_blocks(self, value: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]):
pulumi.set(self, "v6_cidr_blocks", value)
@pulumi.input_type
class VpcSubnetDhcpOptionsArgs:
def __init__(__self__, *,
domain_name: Optional[pulumi.Input[str]] = None,
domain_name_servers: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None,
ntp_servers: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None):
"""
:param pulumi.Input[str] domain_name: Domain name.
:param pulumi.Input[Sequence[pulumi.Input[str]]] domain_name_servers: Domain name server IP addresses.
:param pulumi.Input[Sequence[pulumi.Input[str]]] ntp_servers: NTP server IP addresses.
"""
if domain_name is not None:
pulumi.set(__self__, "domain_name", domain_name)
if domain_name_servers is not None:
pulumi.set(__self__, "domain_name_servers", domain_name_servers)
if ntp_servers is not None:
pulumi.set(__self__, "ntp_servers", ntp_servers)
@property
@pulumi.getter(name="domainName")
def domain_name(self) -> Optional[pulumi.Input[str]]:
"""
Domain name.
"""
return pulumi.get(self, "domain_name")
@domain_name.setter
def domain_name(self, value: Optional[pulumi.Input[str]]):
pulumi.set(self, "domain_name", value)
@property
@pulumi.getter(name="domainNameServers")
def domain_name_servers(self) -> Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]:
"""
Domain name server IP addresses.
"""
return pulumi.get(self, "domain_name_servers")
@domain_name_servers.setter
def domain_name_servers(self, value: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]):
pulumi.set(self, "domain_name_servers", value)
@property
@pulumi.getter(name="ntpServers")
def ntp_servers(self) -> Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]:
"""
NTP server IP addresses.
"""
return pulumi.get(self, "ntp_servers")
@ntp_servers.setter
def ntp_servers(self, value: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]):
pulumi.set(self, "ntp_servers", value)
@pulumi.input_type
class YdbDatabaseDedicatedLocationArgs:
def __init__(__self__, *,
region: Optional[pulumi.Input['YdbDatabaseDedicatedLocationRegionArgs']] = None):
"""
:param pulumi.Input['YdbDatabaseDedicatedLocationRegionArgs'] region: Region for the Yandex Database cluster.
The structure is documented below.
"""
if region is not None:
pulumi.set(__self__, "region", region)
@property
@pulumi.getter
def region(self) -> Optional[pulumi.Input['YdbDatabaseDedicatedLocationRegionArgs']]:
"""
Region for the Yandex Database cluster.
The structure is documented below.
"""
return pulumi.get(self, "region")
@region.setter
def region(self, value: Optional[pulumi.Input['YdbDatabaseDedicatedLocationRegionArgs']]):
pulumi.set(self, "region", value)
@pulumi.input_type
class YdbDatabaseDedicatedLocationRegionArgs:
def __init__(__self__, *,
id: pulumi.Input[str]):
"""
:param pulumi.Input[str] id: Region ID for the Yandex Database cluster.
"""
pulumi.set(__self__, "id", id)
@property
@pulumi.getter
def id(self) -> pulumi.Input[str]:
"""
Region ID for the Yandex Database cluster.
"""
return pulumi.get(self, "id")
@id.setter
def id(self, value: pulumi.Input[str]):
pulumi.set(self, "id", value)
@pulumi.input_type
class YdbDatabaseDedicatedScalePolicyArgs:
def __init__(__self__, *,
fixed_scale: pulumi.Input['YdbDatabaseDedicatedScalePolicyFixedScaleArgs']):
"""
:param pulumi.Input['YdbDatabaseDedicatedScalePolicyFixedScaleArgs'] fixed_scale: Fixed scaling policy for the Yandex Database cluster.
The structure is documented below.
"""
pulumi.set(__self__, "fixed_scale", fixed_scale)
@property
@pulumi.getter(name="fixedScale")
def fixed_scale(self) -> pulumi.Input['YdbDatabaseDedicatedScalePolicyFixedScaleArgs']:
"""
Fixed scaling policy for the Yandex Database cluster.
The structure is documented below.
"""
return pulumi.get(self, "fixed_scale")
@fixed_scale.setter
def fixed_scale(self, value: pulumi.Input['YdbDatabaseDedicatedScalePolicyFixedScaleArgs']):
pulumi.set(self, "fixed_scale", value)
@pulumi.input_type
class YdbDatabaseDedicatedScalePolicyFixedScaleArgs:
def __init__(__self__, *,
size: pulumi.Input[int]):
"""
:param pulumi.Input[int] size: Number of instances for the Yandex Database cluster.
"""
pulumi.set(__self__, "size", size)
@property
@pulumi.getter
def size(self) -> pulumi.Input[int]:
"""
Number of instances for the Yandex Database cluster.
"""
return pulumi.get(self, "size")
@size.setter
def size(self, value: pulumi.Input[int]):
pulumi.set(self, "size", value)
@pulumi.input_type
class YdbDatabaseDedicatedStorageConfigArgs:
def __init__(__self__, *,
group_count: pulumi.Input[int],
storage_type_id: pulumi.Input[str]):
"""
:param pulumi.Input[int] group_count: Amount of storage groups of selected type for the Yandex Database cluster.
:param pulumi.Input[str] storage_type_id: Storage type ID for the Yandex Database cluster.
Available presets can be obtained via `yc ydb storage-type list` command.
"""
pulumi.set(__self__, "group_count", group_count)
pulumi.set(__self__, "storage_type_id", storage_type_id)
@property
@pulumi.getter(name="groupCount")
def group_count(self) -> pulumi.Input[int]:
"""
Amount of storage groups of selected type for the Yandex Database cluster.
"""
return pulumi.get(self, "group_count")
@group_count.setter
def group_count(self, value: pulumi.Input[int]):
pulumi.set(self, "group_count", value)
@property
@pulumi.getter(name="storageTypeId")
def storage_type_id(self) -> pulumi.Input[str]:
"""
Storage type ID for the Yandex Database cluster.
Available presets can be obtained via `yc ydb storage-type list` command.
"""
return pulumi.get(self, "storage_type_id")
@storage_type_id.setter
def storage_type_id(self, value: pulumi.Input[str]):
pulumi.set(self, "storage_type_id", value)
@pulumi.input_type
class GetAlbBackendGroupGrpcBackendArgs:
def __init__(__self__, *,
healthcheck: 'GetAlbBackendGroupGrpcBackendHealthcheckArgs',
load_balancing_config: 'GetAlbBackendGroupGrpcBackendLoadBalancingConfigArgs',
name: str,
port: int,
target_group_ids: Sequence[str],
tls: 'GetAlbBackendGroupGrpcBackendTlsArgs',
weight: int):
"""
:param 'GetAlbBackendGroupGrpcBackendHealthcheckArgs' healthcheck: Healthcheck specification that will be used by this backend. Structure is documented below.
:param 'GetAlbBackendGroupGrpcBackendLoadBalancingConfigArgs' load_balancing_config: Load Balancing Config specification that will be used by this backend. Structure is documented below.
:param str name: - Name of the Backend Group.
:param int port: Port for incoming traffic.
:param Sequence[str] target_group_ids: References target groups for the backend.
:param 'GetAlbBackendGroupGrpcBackendTlsArgs' tls: Tls specification that will be used by this backend. Structure is documented below.
:param int weight: Weight of the backend. Traffic will be split between backends of the same BackendGroup according to their weights.
"""
pulumi.set(__self__, "healthcheck", healthcheck)
pulumi.set(__self__, "load_balancing_config", load_balancing_config)
pulumi.set(__self__, "name", name)
pulumi.set(__self__, "port", port)
pulumi.set(__self__, "target_group_ids", target_group_ids)
pulumi.set(__self__, "tls", tls)
pulumi.set(__self__, "weight", weight)
@property
@pulumi.getter
def healthcheck(self) -> 'GetAlbBackendGroupGrpcBackendHealthcheckArgs':
"""
Healthcheck specification that will be used by this backend. Structure is documented below.
"""
return pulumi.get(self, "healthcheck")
@healthcheck.setter
def healthcheck(self, value: 'GetAlbBackendGroupGrpcBackendHealthcheckArgs'):
pulumi.set(self, "healthcheck", value)
@property
@pulumi.getter(name="loadBalancingConfig")
def load_balancing_config(self) -> 'GetAlbBackendGroupGrpcBackendLoadBalancingConfigArgs':
"""
Load Balancing Config specification that will be used by this backend. Structure is documented below.
"""
return pulumi.get(self, "load_balancing_config")
@load_balancing_config.setter
def load_balancing_config(self, value: 'GetAlbBackendGroupGrpcBackendLoadBalancingConfigArgs'):
pulumi.set(self, "load_balancing_config", value)
@property
@pulumi.getter
def name(self) -> str:
"""
- Name of the Backend Group.
"""
return pulumi.get(self, "name")
@name.setter
def name(self, value: str):
pulumi.set(self, "name", value)
@property
@pulumi.getter
def port(self) -> int:
"""
Port for incoming traffic.
"""
return pulumi.get(self, "port")
@port.setter
def port(self, value: int):
pulumi.set(self, "port", value)
@property
@pulumi.getter(name="targetGroupIds")
def target_group_ids(self) -> Sequence[str]:
"""
References target groups for the backend.
"""
return pulumi.get(self, "target_group_ids")
@target_group_ids.setter
def target_group_ids(self, value: Sequence[str]):
pulumi.set(self, "target_group_ids", value)
@property
@pulumi.getter
def tls(self) -> 'GetAlbBackendGroupGrpcBackendTlsArgs':
"""
Tls specification that will be used by this backend. Structure is documented below.
"""
return pulumi.get(self, "tls")
@tls.setter
def tls(self, value: 'GetAlbBackendGroupGrpcBackendTlsArgs'):
pulumi.set(self, "tls", value)
@property
@pulumi.getter
def weight(self) -> int:
"""
Weight of the backend. Traffic will be split between backends of the same BackendGroup according to their weights.
"""
return pulumi.get(self, "weight")
@weight.setter
def weight(self, value: int):
pulumi.set(self, "weight", value)
@pulumi.input_type
class GetAlbBackendGroupGrpcBackendHealthcheckArgs:
def __init__(__self__, *,
grpc_healthcheck: 'GetAlbBackendGroupGrpcBackendHealthcheckGrpcHealthcheckArgs',
healthcheck_port: int,
healthy_threshold: int,
http_healthcheck: 'GetAlbBackendGroupGrpcBackendHealthcheckHttpHealthcheckArgs',
interval: str,
interval_jitter_percent: float,
stream_healthcheck: 'GetAlbBackendGroupGrpcBackendHealthcheckStreamHealthcheckArgs',
timeout: str,
unhealthy_threshold: int):
"""
:param 'GetAlbBackendGroupGrpcBackendHealthcheckGrpcHealthcheckArgs' grpc_healthcheck: Grpc Healthcheck specification that will be used by this healthcheck. Structure is documented below.
:param int healthcheck_port: Optional alternative port for health checking.
:param int healthy_threshold: Number of consecutive successful health checks required to promote endpoint into the healthy state. 0 means 1. Note that during startup, only a single successful health check is required to mark a host healthy.
:param 'GetAlbBackendGroupGrpcBackendHealthcheckHttpHealthcheckArgs' http_healthcheck: Http Healthcheck specification that will be used by this healthcheck. Structure is documented below.
:param str interval: Interval between health checks.
:param float interval_jitter_percent: An optional jitter amount as a percentage of interval. If specified, during every interval value of (interval_ms * interval_jitter_percent / 100) will be added to the wait time.
:param 'GetAlbBackendGroupGrpcBackendHealthcheckStreamHealthcheckArgs' stream_healthcheck: Stream Healthcheck specification that will be used by this healthcheck. Structure is documented below.
:param str timeout: Time to wait for a health check response.
:param int unhealthy_threshold: Number of consecutive failed health checks required to demote endpoint into the unhealthy state. 0 means 1. Note that for HTTP health checks, a single 503 immediately makes endpoint unhealthy.
"""
pulumi.set(__self__, "grpc_healthcheck", grpc_healthcheck)
pulumi.set(__self__, "healthcheck_port", healthcheck_port)
pulumi.set(__self__, "healthy_threshold", healthy_threshold)
pulumi.set(__self__, "http_healthcheck", http_healthcheck)
pulumi.set(__self__, "interval", interval)
pulumi.set(__self__, "interval_jitter_percent", interval_jitter_percent)
pulumi.set(__self__, "stream_healthcheck", stream_healthcheck)
pulumi.set(__self__, "timeout", timeout)
pulumi.set(__self__, "unhealthy_threshold", unhealthy_threshold)
@property
@pulumi.getter(name="grpcHealthcheck")
def grpc_healthcheck(self) -> 'GetAlbBackendGroupGrpcBackendHealthcheckGrpcHealthcheckArgs':
"""
Grpc Healthcheck specification that will be used by this healthcheck. Structure is documented below.
"""
return pulumi.get(self, "grpc_healthcheck")
@grpc_healthcheck.setter
def grpc_healthcheck(self, value: 'GetAlbBackendGroupGrpcBackendHealthcheckGrpcHealthcheckArgs'):
pulumi.set(self, "grpc_healthcheck", value)
@property
@pulumi.getter(name="healthcheckPort")
def healthcheck_port(self) -> int:
"""
Optional alternative port for health checking.
"""
return pulumi.get(self, "healthcheck_port")
@healthcheck_port.setter
def healthcheck_port(self, value: int):
pulumi.set(self, "healthcheck_port", value)
@property
@pulumi.getter(name="healthyThreshold")
def healthy_threshold(self) -> int:
"""
Number of consecutive successful health checks required to promote endpoint into the healthy state. 0 means 1. Note that during startup, only a single successful health check is required to mark a host healthy.
"""
return pulumi.get(self, "healthy_threshold")
@healthy_threshold.setter
def healthy_threshold(self, value: int):
pulumi.set(self, "healthy_threshold", value)
@property
@pulumi.getter(name="httpHealthcheck")
def http_healthcheck(self) -> 'GetAlbBackendGroupGrpcBackendHealthcheckHttpHealthcheckArgs':
"""
Http Healthcheck specification that will be used by this healthcheck. Structure is documented below.
"""
return pulumi.get(self, "http_healthcheck")
@http_healthcheck.setter
def http_healthcheck(self, value: 'GetAlbBackendGroupGrpcBackendHealthcheckHttpHealthcheckArgs'):
pulumi.set(self, "http_healthcheck", value)
@property
@pulumi.getter
def interval(self) -> str:
"""
Interval between health checks.
"""
return pulumi.get(self, "interval")
@interval.setter
def interval(self, value: str):
pulumi.set(self, "interval", value)
@property
@pulumi.getter(name="intervalJitterPercent")
def interval_jitter_percent(self) -> float:
"""
An optional jitter amount as a percentage of interval. If specified, during every interval value of (interval_ms * interval_jitter_percent / 100) will be added to the wait time.
"""
return pulumi.get(self, "interval_jitter_percent")
@interval_jitter_percent.setter
def interval_jitter_percent(self, value: float):
pulumi.set(self, "interval_jitter_percent", value)
@property
@pulumi.getter(name="streamHealthcheck")
def stream_healthcheck(self) -> 'GetAlbBackendGroupGrpcBackendHealthcheckStreamHealthcheckArgs':
"""
Stream Healthcheck specification that will be used by this healthcheck. Structure is documented below.
"""
return pulumi.get(self, "stream_healthcheck")
@stream_healthcheck.setter
def stream_healthcheck(self, value: 'GetAlbBackendGroupGrpcBackendHealthcheckStreamHealthcheckArgs'):
pulumi.set(self, "stream_healthcheck", value)
@property
@pulumi.getter
def timeout(self) -> str:
"""
Time to wait for a health check response.
"""
return pulumi.get(self, "timeout")
@timeout.setter
def timeout(self, value: str):
pulumi.set(self, "timeout", value)
@property
@pulumi.getter(name="unhealthyThreshold")
def unhealthy_threshold(self) -> int:
"""
Number of consecutive failed health checks required to demote endpoint into the unhealthy state. 0 means 1. Note that for HTTP health checks, a single 503 immediately makes endpoint unhealthy.
"""
return pulumi.get(self, "unhealthy_threshold")
@unhealthy_threshold.setter
def unhealthy_threshold(self, value: int):
pulumi.set(self, "unhealthy_threshold", value)
@pulumi.input_type
class GetAlbBackendGroupGrpcBackendHealthcheckGrpcHealthcheckArgs:
def __init__(__self__, *,
service_name: str):
"""
:param str service_name: Optional service name for grpc.health.v1.HealthCheckRequest message.
"""
pulumi.set(__self__, "service_name", service_name)
@property
@pulumi.getter(name="serviceName")
def service_name(self) -> str:
"""
Optional service name for grpc.health.v1.HealthCheckRequest message.
"""
return pulumi.get(self, "service_name")
@service_name.setter
def service_name(self, value: str):
pulumi.set(self, "service_name", value)
@pulumi.input_type
class GetAlbBackendGroupGrpcBackendHealthcheckHttpHealthcheckArgs:
def __init__(__self__, *,
host: str,
http2: bool,
path: str):
"""
:param str host: Optional "Host" HTTP header value.
:param bool http2: If set, health checks will use HTTP2.
:param str path: HTTP path.
"""
pulumi.set(__self__, "host", host)
pulumi.set(__self__, "http2", http2)
pulumi.set(__self__, "path", path)
@property
@pulumi.getter
def host(self) -> str:
"""
Optional "Host" HTTP header value.
"""
return pulumi.get(self, "host")
@host.setter
def host(self, value: str):
pulumi.set(self, "host", value)
@property
@pulumi.getter
def http2(self) -> bool:
"""
If set, health checks will use HTTP2.
"""
return pulumi.get(self, "http2")
@http2.setter
def http2(self, value: bool):
pulumi.set(self, "http2", value)
@property
@pulumi.getter
def path(self) -> str:
"""
HTTP path.
"""
return pulumi.get(self, "path")
@path.setter
def path(self, value: str):
pulumi.set(self, "path", value)
@pulumi.input_type
class GetAlbBackendGroupGrpcBackendHealthcheckStreamHealthcheckArgs:
def __init__(__self__, *,
receive: str,
send: str):
"""
:param str receive: Optional text to search in reply.
:param str send: Optional message to send. If empty, it's a connect-only health check.
"""
pulumi.set(__self__, "receive", receive)
pulumi.set(__self__, "send", send)
@property
@pulumi.getter
def receive(self) -> str:
"""
Optional text to search in reply.
"""
return pulumi.get(self, "receive")
@receive.setter
def receive(self, value: str):
pulumi.set(self, "receive", value)
@property
@pulumi.getter
def send(self) -> str:
"""
Optional message to send. If empty, it's a connect-only health check.
"""
return pulumi.get(self, "send")
@send.setter
def send(self, value: str):
pulumi.set(self, "send", value)
@pulumi.input_type
class GetAlbBackendGroupGrpcBackendLoadBalancingConfigArgs:
def __init__(__self__, *,
locality_aware_routing_percent: int,
panic_threshold: int,
strict_locality: bool):
"""
:param int locality_aware_routing_percent: Percent of traffic to be sent to the same availability zone. The rest will be equally divided between other zones.
:param int panic_threshold: If percentage of healthy hosts in the backend is lower than panic_threshold, traffic will be routed to all backends no matter what the health status is. This helps to avoid healthy backends overloading when everything is bad. Zero means no panic threshold.
:param bool strict_locality: If set, will route requests only to the same availability zone. Balancer won't know about endpoints in other zones.
"""
pulumi.set(__self__, "locality_aware_routing_percent", locality_aware_routing_percent)
pulumi.set(__self__, "panic_threshold", panic_threshold)
pulumi.set(__self__, "strict_locality", strict_locality)
@property
@pulumi.getter(name="localityAwareRoutingPercent")
def locality_aware_routing_percent(self) -> int:
"""
Percent of traffic to be sent to the same availability zone. The rest will be equally divided between other zones.
"""
return pulumi.get(self, "locality_aware_routing_percent")
@locality_aware_routing_percent.setter
def locality_aware_routing_percent(self, value: int):
pulumi.set(self, "locality_aware_routing_percent", value)
@property
@pulumi.getter(name="panicThreshold")
def panic_threshold(self) -> int:
"""
If percentage of healthy hosts in the backend is lower than panic_threshold, traffic will be routed to all backends no matter what the health status is. This helps to avoid healthy backends overloading when everything is bad. Zero means no panic threshold.
"""
return pulumi.get(self, "panic_threshold")
@panic_threshold.setter
def panic_threshold(self, value: int):
pulumi.set(self, "panic_threshold", value)
@property
@pulumi.getter(name="strictLocality")
def strict_locality(self) -> bool:
"""
If set, will route requests only to the same availability zone. Balancer won't know about endpoints in other zones.
"""
return pulumi.get(self, "strict_locality")
@strict_locality.setter
def strict_locality(self, value: bool):
pulumi.set(self, "strict_locality", value)
@pulumi.input_type
class GetAlbBackendGroupGrpcBackendTlsArgs:
def __init__(__self__, *,
sni: str,
validation_context: 'GetAlbBackendGroupGrpcBackendTlsValidationContextArgs'):
"""
:param str sni: [SNI](https://en.wikipedia.org/wiki/Server_Name_Indication) string for TLS connections.
* `validation_context.0.trusted_ca_id` - Trusted CA certificate ID in the Certificate Manager.
* `validation_context.0.trusted_ca_bytes` - PEM-encoded trusted CA certificate chain.
"""
pulumi.set(__self__, "sni", sni)
pulumi.set(__self__, "validation_context", validation_context)
@property
@pulumi.getter
def sni(self) -> str:
"""
[SNI](https://en.wikipedia.org/wiki/Server_Name_Indication) string for TLS connections.
* `validation_context.0.trusted_ca_id` - Trusted CA certificate ID in the Certificate Manager.
* `validation_context.0.trusted_ca_bytes` - PEM-encoded trusted CA certificate chain.
"""
return pulumi.get(self, "sni")
@sni.setter
def sni(self, value: str):
pulumi.set(self, "sni", value)
@property
@pulumi.getter(name="validationContext")
def validation_context(self) -> 'GetAlbBackendGroupGrpcBackendTlsValidationContextArgs':
return pulumi.get(self, "validation_context")
@validation_context.setter
def validation_context(self, value: 'GetAlbBackendGroupGrpcBackendTlsValidationContextArgs'):
pulumi.set(self, "validation_context", value)
@pulumi.input_type
class GetAlbBackendGroupGrpcBackendTlsValidationContextArgs:
def __init__(__self__, *,
trusted_ca_bytes: str,
trusted_ca_id: str):
pulumi.set(__self__, "trusted_ca_bytes", trusted_ca_bytes)
pulumi.set(__self__, "trusted_ca_id", trusted_ca_id)
@property
@pulumi.getter(name="trustedCaBytes")
def trusted_ca_bytes(self) -> str:
return pulumi.get(self, "trusted_ca_bytes")
@trusted_ca_bytes.setter
def trusted_ca_bytes(self, value: str):
pulumi.set(self, "trusted_ca_bytes", value)
@property
@pulumi.getter(name="trustedCaId")
def trusted_ca_id(self) -> str:
return pulumi.get(self, "trusted_ca_id")
@trusted_ca_id.setter
def trusted_ca_id(self, value: str):
pulumi.set(self, "trusted_ca_id", value)
@pulumi.input_type
class GetAlbBackendGroupHttpBackendArgs:
def __init__(__self__, *,
healthcheck: 'GetAlbBackendGroupHttpBackendHealthcheckArgs',
http2: bool,
load_balancing_config: 'GetAlbBackendGroupHttpBackendLoadBalancingConfigArgs',
name: str,
port: int,
target_group_ids: Sequence[str],
tls: 'GetAlbBackendGroupHttpBackendTlsArgs',
weight: int):
"""
:param 'GetAlbBackendGroupHttpBackendHealthcheckArgs' healthcheck: Healthcheck specification that will be used by this backend. Structure is documented below.
:param bool http2: If set, health checks will use HTTP2.
:param 'GetAlbBackendGroupHttpBackendLoadBalancingConfigArgs' load_balancing_config: Load Balancing Config specification that will be used by this backend. Structure is documented below.
:param str name: - Name of the Backend Group.
:param int port: Port for incoming traffic.
:param Sequence[str] target_group_ids: References target groups for the backend.
:param 'GetAlbBackendGroupHttpBackendTlsArgs' tls: Tls specification that will be used by this backend. Structure is documented below.
:param int weight: Weight of the backend. Traffic will be split between backends of the same BackendGroup according to their weights.
"""
pulumi.set(__self__, "healthcheck", healthcheck)
pulumi.set(__self__, "http2", http2)
pulumi.set(__self__, "load_balancing_config", load_balancing_config)
pulumi.set(__self__, "name", name)
pulumi.set(__self__, "port", port)
pulumi.set(__self__, "target_group_ids", target_group_ids)
pulumi.set(__self__, "tls", tls)
pulumi.set(__self__, "weight", weight)
@property
@pulumi.getter
def healthcheck(self) -> 'GetAlbBackendGroupHttpBackendHealthcheckArgs':
"""
Healthcheck specification that will be used by this backend. Structure is documented below.
"""
return pulumi.get(self, "healthcheck")
@healthcheck.setter
def healthcheck(self, value: 'GetAlbBackendGroupHttpBackendHealthcheckArgs'):
pulumi.set(self, "healthcheck", value)
@property
@pulumi.getter
def http2(self) -> bool:
"""
If set, health checks will use HTTP2.
"""
return pulumi.get(self, "http2")
@http2.setter
def http2(self, value: bool):
pulumi.set(self, "http2", value)
@property
@pulumi.getter(name="loadBalancingConfig")
def load_balancing_config(self) -> 'GetAlbBackendGroupHttpBackendLoadBalancingConfigArgs':
"""
Load Balancing Config specification that will be used by this backend. Structure is documented below.
"""
return pulumi.get(self, "load_balancing_config")
@load_balancing_config.setter
def load_balancing_config(self, value: 'GetAlbBackendGroupHttpBackendLoadBalancingConfigArgs'):
pulumi.set(self, "load_balancing_config", value)
@property
@pulumi.getter
def name(self) -> str:
"""
- Name of the Backend Group.
"""
return pulumi.get(self, "name")
@name.setter
def name(self, value: str):
pulumi.set(self, "name", value)
@property
@pulumi.getter
def port(self) -> int:
"""
Port for incoming traffic.
"""
return pulumi.get(self, "port")
@port.setter
def port(self, value: int):
pulumi.set(self, "port", value)
@property
@pulumi.getter(name="targetGroupIds")
def target_group_ids(self) -> Sequence[str]:
"""
References target groups for the backend.
"""
return pulumi.get(self, "target_group_ids")
@target_group_ids.setter
def target_group_ids(self, value: Sequence[str]):
pulumi.set(self, "target_group_ids", value)
@property
@pulumi.getter
def tls(self) -> 'GetAlbBackendGroupHttpBackendTlsArgs':
"""
Tls specification that will be used by this backend. Structure is documented below.
"""
return pulumi.get(self, "tls")
@tls.setter
def tls(self, value: 'GetAlbBackendGroupHttpBackendTlsArgs'):
pulumi.set(self, "tls", value)
@property
@pulumi.getter
def weight(self) -> int:
"""
Weight of the backend. Traffic will be split between backends of the same BackendGroup according to their weights.
"""
return pulumi.get(self, "weight")
@weight.setter
def weight(self, value: int):
pulumi.set(self, "weight", value)
@pulumi.input_type
class GetAlbBackendGroupHttpBackendHealthcheckArgs:
def __init__(__self__, *,
grpc_healthcheck: 'GetAlbBackendGroupHttpBackendHealthcheckGrpcHealthcheckArgs',
healthcheck_port: int,
healthy_threshold: int,
http_healthcheck: 'GetAlbBackendGroupHttpBackendHealthcheckHttpHealthcheckArgs',
interval: str,
interval_jitter_percent: float,
stream_healthcheck: 'GetAlbBackendGroupHttpBackendHealthcheckStreamHealthcheckArgs',
timeout: str,
unhealthy_threshold: int):
"""
:param 'GetAlbBackendGroupHttpBackendHealthcheckGrpcHealthcheckArgs' grpc_healthcheck: Grpc Healthcheck specification that will be used by this healthcheck. Structure is documented below.
:param int healthcheck_port: Optional alternative port for health checking.
:param int healthy_threshold: Number of consecutive successful health checks required to promote endpoint into the healthy state. 0 means 1. Note that during startup, only a single successful health check is required to mark a host healthy.
:param 'GetAlbBackendGroupHttpBackendHealthcheckHttpHealthcheckArgs' http_healthcheck: Http Healthcheck specification that will be used by this healthcheck. Structure is documented below.
:param str interval: Interval between health checks.
:param float interval_jitter_percent: An optional jitter amount as a percentage of interval. If specified, during every interval value of (interval_ms * interval_jitter_percent / 100) will be added to the wait time.
:param 'GetAlbBackendGroupHttpBackendHealthcheckStreamHealthcheckArgs' stream_healthcheck: Stream Healthcheck specification that will be used by this healthcheck. Structure is documented below.
:param str timeout: Time to wait for a health check response.
:param int unhealthy_threshold: Number of consecutive failed health checks required to demote endpoint into the unhealthy state. 0 means 1. Note that for HTTP health checks, a single 503 immediately makes endpoint unhealthy.
"""
pulumi.set(__self__, "grpc_healthcheck", grpc_healthcheck)
pulumi.set(__self__, "healthcheck_port", healthcheck_port)
pulumi.set(__self__, "healthy_threshold", healthy_threshold)
pulumi.set(__self__, "http_healthcheck", http_healthcheck)
pulumi.set(__self__, "interval", interval)
pulumi.set(__self__, "interval_jitter_percent", interval_jitter_percent)
pulumi.set(__self__, "stream_healthcheck", stream_healthcheck)
pulumi.set(__self__, "timeout", timeout)
pulumi.set(__self__, "unhealthy_threshold", unhealthy_threshold)
@property
@pulumi.getter(name="grpcHealthcheck")
def grpc_healthcheck(self) -> 'GetAlbBackendGroupHttpBackendHealthcheckGrpcHealthcheckArgs':
"""
Grpc Healthcheck specification that will be used by this healthcheck. Structure is documented below.
"""
return pulumi.get(self, "grpc_healthcheck")
@grpc_healthcheck.setter
def grpc_healthcheck(self, value: 'GetAlbBackendGroupHttpBackendHealthcheckGrpcHealthcheckArgs'):
pulumi.set(self, "grpc_healthcheck", value)
@property
@pulumi.getter(name="healthcheckPort")
def healthcheck_port(self) -> int:
"""
Optional alternative port for health checking.
"""
return pulumi.get(self, "healthcheck_port")
@healthcheck_port.setter
def healthcheck_port(self, value: int):
pulumi.set(self, "healthcheck_port", value)
@property
@pulumi.getter(name="healthyThreshold")
def healthy_threshold(self) -> int:
"""
Number of consecutive successful health checks required to promote endpoint into the healthy state. 0 means 1. Note that during startup, only a single successful health check is required to mark a host healthy.
"""
return pulumi.get(self, "healthy_threshold")
@healthy_threshold.setter
def healthy_threshold(self, value: int):
pulumi.set(self, "healthy_threshold", value)
@property
@pulumi.getter(name="httpHealthcheck")
def http_healthcheck(self) -> 'GetAlbBackendGroupHttpBackendHealthcheckHttpHealthcheckArgs':
"""
Http Healthcheck specification that will be used by this healthcheck. Structure is documented below.
"""
return pulumi.get(self, "http_healthcheck")
@http_healthcheck.setter
def http_healthcheck(self, value: 'GetAlbBackendGroupHttpBackendHealthcheckHttpHealthcheckArgs'):
pulumi.set(self, "http_healthcheck", value)
@property
@pulumi.getter
def interval(self) -> str:
"""
Interval between health checks.
"""
return pulumi.get(self, "interval")
@interval.setter
def interval(self, value: str):
pulumi.set(self, "interval", value)
@property
@pulumi.getter(name="intervalJitterPercent")
def interval_jitter_percent(self) -> float:
"""
An optional jitter amount as a percentage of interval. If specified, during every interval value of (interval_ms * interval_jitter_percent / 100) will be added to the wait time.
"""
return pulumi.get(self, "interval_jitter_percent")
@interval_jitter_percent.setter
def interval_jitter_percent(self, value: float):
pulumi.set(self, "interval_jitter_percent", value)
@property
@pulumi.getter(name="streamHealthcheck")
def stream_healthcheck(self) -> 'GetAlbBackendGroupHttpBackendHealthcheckStreamHealthcheckArgs':
"""
Stream Healthcheck specification that will be used by this healthcheck. Structure is documented below.
"""
return pulumi.get(self, "stream_healthcheck")
@stream_healthcheck.setter
def stream_healthcheck(self, value: 'GetAlbBackendGroupHttpBackendHealthcheckStreamHealthcheckArgs'):
pulumi.set(self, "stream_healthcheck", value)
@property
@pulumi.getter
def timeout(self) -> str:
"""
Time to wait for a health check response.
"""
return pulumi.get(self, "timeout")
@timeout.setter
def timeout(self, value: str):
pulumi.set(self, "timeout", value)
@property
@pulumi.getter(name="unhealthyThreshold")
def unhealthy_threshold(self) -> int:
"""
Number of consecutive failed health checks required to demote endpoint into the unhealthy state. 0 means 1. Note that for HTTP health checks, a single 503 immediately makes endpoint unhealthy.
"""
return pulumi.get(self, "unhealthy_threshold")
@unhealthy_threshold.setter
def unhealthy_threshold(self, value: int):
pulumi.set(self, "unhealthy_threshold", value)
@pulumi.input_type
class GetAlbBackendGroupHttpBackendHealthcheckGrpcHealthcheckArgs:
def __init__(__self__, *,
service_name: str):
"""
:param str service_name: Optional service name for grpc.health.v1.HealthCheckRequest message.
"""
pulumi.set(__self__, "service_name", service_name)
@property
@pulumi.getter(name="serviceName")
def service_name(self) -> str:
"""
Optional service name for grpc.health.v1.HealthCheckRequest message.
"""
return pulumi.get(self, "service_name")
@service_name.setter
def service_name(self, value: str):
pulumi.set(self, "service_name", value)
@pulumi.input_type
class GetAlbBackendGroupHttpBackendHealthcheckHttpHealthcheckArgs:
def __init__(__self__, *,
host: str,
http2: bool,
path: str):
"""
:param str host: Optional "Host" HTTP header value.
:param bool http2: If set, health checks will use HTTP2.
:param str path: HTTP path.
"""
pulumi.set(__self__, "host", host)
pulumi.set(__self__, "http2", http2)
pulumi.set(__self__, "path", path)
@property
@pulumi.getter
def host(self) -> str:
"""
Optional "Host" HTTP header value.
"""
return pulumi.get(self, "host")
@host.setter
def host(self, value: str):
pulumi.set(self, "host", value)
@property
@pulumi.getter
def http2(self) -> bool:
"""
If set, health checks will use HTTP2.
"""
return pulumi.get(self, "http2")
@http2.setter
def http2(self, value: bool):
pulumi.set(self, "http2", value)
@property
@pulumi.getter
def path(self) -> str:
"""
HTTP path.
"""
return pulumi.get(self, "path")
@path.setter
def path(self, value: str):
pulumi.set(self, "path", value)
@pulumi.input_type
class GetAlbBackendGroupHttpBackendHealthcheckStreamHealthcheckArgs:
def __init__(__self__, *,
receive: str,
send: str):
"""
:param str receive: Optional text to search in reply.
:param str send: Optional message to send. If empty, it's a connect-only health check.
"""
pulumi.set(__self__, "receive", receive)
pulumi.set(__self__, "send", send)
@property
@pulumi.getter
def receive(self) -> str:
"""
Optional text to search in reply.
"""
return pulumi.get(self, "receive")
@receive.setter
def receive(self, value: str):
pulumi.set(self, "receive", value)
@property
@pulumi.getter
def send(self) -> str:
"""
Optional message to send. If empty, it's a connect-only health check.
"""
return pulumi.get(self, "send")
@send.setter
def send(self, value: str):
pulumi.set(self, "send", value)
@pulumi.input_type
class GetAlbBackendGroupHttpBackendLoadBalancingConfigArgs:
def __init__(__self__, *,
locality_aware_routing_percent: int,
panic_threshold: int,
strict_locality: bool):
"""
:param int locality_aware_routing_percent: Percent of traffic to be sent to the same availability zone. The rest will be equally divided between other zones.
:param int panic_threshold: If percentage of healthy hosts in the backend is lower than panic_threshold, traffic will be routed to all backends no matter what the health status is. This helps to avoid healthy backends overloading when everything is bad. Zero means no panic threshold.
:param bool strict_locality: If set, will route requests only to the same availability zone. Balancer won't know about endpoints in other zones.
"""
pulumi.set(__self__, "locality_aware_routing_percent", locality_aware_routing_percent)
pulumi.set(__self__, "panic_threshold", panic_threshold)
pulumi.set(__self__, "strict_locality", strict_locality)
@property
@pulumi.getter(name="localityAwareRoutingPercent")
def locality_aware_routing_percent(self) -> int:
"""
Percent of traffic to be sent to the same availability zone. The rest will be equally divided between other zones.
"""
return pulumi.get(self, "locality_aware_routing_percent")
@locality_aware_routing_percent.setter
def locality_aware_routing_percent(self, value: int):
pulumi.set(self, "locality_aware_routing_percent", value)
@property
@pulumi.getter(name="panicThreshold")
def panic_threshold(self) -> int:
"""
If percentage of healthy hosts in the backend is lower than panic_threshold, traffic will be routed to all backends no matter what the health status is. This helps to avoid healthy backends overloading when everything is bad. Zero means no panic threshold.
"""
return pulumi.get(self, "panic_threshold")
@panic_threshold.setter
def panic_threshold(self, value: int):
pulumi.set(self, "panic_threshold", value)
@property
@pulumi.getter(name="strictLocality")
def strict_locality(self) -> bool:
"""
If set, will route requests only to the same availability zone. Balancer won't know about endpoints in other zones.
"""
return pulumi.get(self, "strict_locality")
@strict_locality.setter
def strict_locality(self, value: bool):
pulumi.set(self, "strict_locality", value)
@pulumi.input_type
class GetAlbBackendGroupHttpBackendTlsArgs:
def __init__(__self__, *,
sni: str,
validation_context: 'GetAlbBackendGroupHttpBackendTlsValidationContextArgs'):
"""
:param str sni: [SNI](https://en.wikipedia.org/wiki/Server_Name_Indication) string for TLS connections.
* `validation_context.0.trusted_ca_id` - Trusted CA certificate ID in the Certificate Manager.
* `validation_context.0.trusted_ca_bytes` - PEM-encoded trusted CA certificate chain.
"""
pulumi.set(__self__, "sni", sni)
pulumi.set(__self__, "validation_context", validation_context)
@property
@pulumi.getter
def sni(self) -> str:
"""
[SNI](https://en.wikipedia.org/wiki/Server_Name_Indication) string for TLS connections.
* `validation_context.0.trusted_ca_id` - Trusted CA certificate ID in the Certificate Manager.
* `validation_context.0.trusted_ca_bytes` - PEM-encoded trusted CA certificate chain.
"""
return pulumi.get(self, "sni")
@sni.setter
def sni(self, value: str):
pulumi.set(self, "sni", value)
@property
@pulumi.getter(name="validationContext")
def validation_context(self) -> 'GetAlbBackendGroupHttpBackendTlsValidationContextArgs':
return pulumi.get(self, "validation_context")
@validation_context.setter
def validation_context(self, value: 'GetAlbBackendGroupHttpBackendTlsValidationContextArgs'):
pulumi.set(self, "validation_context", value)
@pulumi.input_type
class GetAlbBackendGroupHttpBackendTlsValidationContextArgs:
def __init__(__self__, *,
trusted_ca_bytes: str,
trusted_ca_id: str):
pulumi.set(__self__, "trusted_ca_bytes", trusted_ca_bytes)
pulumi.set(__self__, "trusted_ca_id", trusted_ca_id)
@property
@pulumi.getter(name="trustedCaBytes")
def trusted_ca_bytes(self) -> str:
return pulumi.get(self, "trusted_ca_bytes")
@trusted_ca_bytes.setter
def trusted_ca_bytes(self, value: str):
pulumi.set(self, "trusted_ca_bytes", value)
@property
@pulumi.getter(name="trustedCaId")
def trusted_ca_id(self) -> str:
return pulumi.get(self, "trusted_ca_id")
@trusted_ca_id.setter
def trusted_ca_id(self, value: str):
pulumi.set(self, "trusted_ca_id", value)
@pulumi.input_type
class GetComputeDiskDiskPlacementPolicyArgs:
def __init__(__self__, *,
disk_placement_group_id: str):
pulumi.set(__self__, "disk_placement_group_id", disk_placement_group_id)
@property
@pulumi.getter(name="diskPlacementGroupId")
def disk_placement_group_id(self) -> str:
return pulumi.get(self, "disk_placement_group_id")
@disk_placement_group_id.setter
def disk_placement_group_id(self, value: str):
pulumi.set(self, "disk_placement_group_id", value)
@pulumi.input_type
class GetComputeInstancePlacementPolicyArgs:
def __init__(__self__, *,
placement_group_id: str):
pulumi.set(__self__, "placement_group_id", placement_group_id)
@property
@pulumi.getter(name="placementGroupId")
def placement_group_id(self) -> str:
return pulumi.get(self, "placement_group_id")
@placement_group_id.setter
def placement_group_id(self, value: str):
pulumi.set(self, "placement_group_id", value)
@pulumi.input_type
class GetComputeInstanceSchedulingPolicyArgs:
def __init__(__self__, *,
preemptible: Optional[bool] = None):
"""
:param bool preemptible: (Optional) Specifies if the instance is preemptible. Defaults to false.
"""
if preemptible is not None:
pulumi.set(__self__, "preemptible", preemptible)
@property
@pulumi.getter
def preemptible(self) -> Optional[bool]:
"""
(Optional) Specifies if the instance is preemptible. Defaults to false.
"""
return pulumi.get(self, "preemptible")
@preemptible.setter
def preemptible(self, value: Optional[bool]):
pulumi.set(self, "preemptible", value)
@pulumi.input_type
class GetFunctionScalingPolicyPolicyArgs:
def __init__(__self__, *,
tag: str,
zone_instances_limit: Optional[int] = None,
zone_requests_limit: Optional[int] = None):
pulumi.set(__self__, "tag", tag)
if zone_instances_limit is not None:
pulumi.set(__self__, "zone_instances_limit", zone_instances_limit)
if zone_requests_limit is not None:
pulumi.set(__self__, "zone_requests_limit", zone_requests_limit)
@property
@pulumi.getter
def tag(self) -> str:
return pulumi.get(self, "tag")
@tag.setter
def tag(self, value: str):
pulumi.set(self, "tag", value)
@property
@pulumi.getter(name="zoneInstancesLimit")
def zone_instances_limit(self) -> Optional[int]:
return pulumi.get(self, "zone_instances_limit")
@zone_instances_limit.setter
def zone_instances_limit(self, value: Optional[int]):
pulumi.set(self, "zone_instances_limit", value)
@property
@pulumi.getter(name="zoneRequestsLimit")
def zone_requests_limit(self) -> Optional[int]:
return pulumi.get(self, "zone_requests_limit")
@zone_requests_limit.setter
def zone_requests_limit(self, value: Optional[int]):
pulumi.set(self, "zone_requests_limit", value)
@pulumi.input_type
class GetIamPolicyBindingArgs:
def __init__(__self__, *,
members: Sequence[str],
role: str):
"""
:param Sequence[str] members: An array of identities that will be granted the privilege in the `role`.
Each entry can have one of the following values:
* **userAccount:{user_id}**: A unique user ID that represents a specific Yandex account.
* **serviceAccount:{service_account_id}**: A unique service account ID.
:param str role: The role/permission that will be granted to the members.
See the [IAM Roles] documentation for a complete list of roles.
"""
pulumi.set(__self__, "members", members)
pulumi.set(__self__, "role", role)
@property
@pulumi.getter
def members(self) -> Sequence[str]:
"""
An array of identities that will be granted the privilege in the `role`.
Each entry can have one of the following values:
* **userAccount:{user_id}**: A unique user ID that represents a specific Yandex account.
* **serviceAccount:{service_account_id}**: A unique service account ID.
"""
return pulumi.get(self, "members")
@members.setter
def members(self, value: Sequence[str]):
pulumi.set(self, "members", value)
@property
@pulumi.getter
def role(self) -> str:
"""
The role/permission that will be granted to the members.
See the [IAM Roles] documentation for a complete list of roles.
"""
return pulumi.get(self, "role")
@role.setter
def role(self, value: str):
pulumi.set(self, "role", value)
@pulumi.input_type
class GetMdbClickhouseClusterCloudStorageArgs:
def __init__(__self__, *,
enabled: bool):
"""
:param bool enabled: (Required) Whether to use Yandex Object Storage for storing ClickHouse data. Can be either `true` or `false`.
"""
pulumi.set(__self__, "enabled", enabled)
@property
@pulumi.getter
def enabled(self) -> bool:
"""
(Required) Whether to use Yandex Object Storage for storing ClickHouse data. Can be either `true` or `false`.
"""
return pulumi.get(self, "enabled")
@enabled.setter
def enabled(self, value: bool):
pulumi.set(self, "enabled", value)
@pulumi.input_type
class GetMdbKafkaClusterConfigArgs:
def __init__(__self__, *,
kafka: 'GetMdbKafkaClusterConfigKafkaArgs',
version: str,
zones: Sequence[str],
zookeeper: 'GetMdbKafkaClusterConfigZookeeperArgs',
assign_public_ip: Optional[bool] = None,
brokers_count: Optional[int] = None,
schema_registry: Optional[bool] = None,
unmanaged_topics: Optional[bool] = None):
"""
:param 'GetMdbKafkaClusterConfigKafkaArgs' kafka: (Optional) Configuration of the Kafka subcluster. The structure is documented below.
:param str version: (Required) Version of the Kafka server software.
:param Sequence[str] zones: (Optional) List of availability zones.
:param 'GetMdbKafkaClusterConfigZookeeperArgs' zookeeper: (Optional) Configuration of the ZooKeeper subcluster. The structure is documented below.
:param bool assign_public_ip: The flag that defines whether a public IP address is assigned to the node.
:param int brokers_count: (Optional) Count of brokers per availability zone.
:param bool schema_registry: (Optional) Enables managed schema registry on cluster. Can be either `true` or `false`.
:param bool unmanaged_topics: (Optional) Allows to use Kafka AdminAPI to manage topics. Can be either `true` or `false`.
"""
pulumi.set(__self__, "kafka", kafka)
pulumi.set(__self__, "version", version)
pulumi.set(__self__, "zones", zones)
pulumi.set(__self__, "zookeeper", zookeeper)
if assign_public_ip is not None:
pulumi.set(__self__, "assign_public_ip", assign_public_ip)
if brokers_count is not None:
pulumi.set(__self__, "brokers_count", brokers_count)
if schema_registry is not None:
pulumi.set(__self__, "schema_registry", schema_registry)
if unmanaged_topics is not None:
pulumi.set(__self__, "unmanaged_topics", unmanaged_topics)
@property
@pulumi.getter
def kafka(self) -> 'GetMdbKafkaClusterConfigKafkaArgs':
"""
(Optional) Configuration of the Kafka subcluster. The structure is documented below.
"""
return pulumi.get(self, "kafka")
@kafka.setter
def kafka(self, value: 'GetMdbKafkaClusterConfigKafkaArgs'):
pulumi.set(self, "kafka", value)
@property
@pulumi.getter
def version(self) -> str:
"""
(Required) Version of the Kafka server software.
"""
return pulumi.get(self, "version")
@version.setter
def version(self, value: str):
pulumi.set(self, "version", value)
@property
@pulumi.getter
def zones(self) -> Sequence[str]:
"""
(Optional) List of availability zones.
"""
return pulumi.get(self, "zones")
@zones.setter
def zones(self, value: Sequence[str]):
pulumi.set(self, "zones", value)
@property
@pulumi.getter
def zookeeper(self) -> 'GetMdbKafkaClusterConfigZookeeperArgs':
"""
(Optional) Configuration of the ZooKeeper subcluster. The structure is documented below.
"""
return pulumi.get(self, "zookeeper")
@zookeeper.setter
def zookeeper(self, value: 'GetMdbKafkaClusterConfigZookeeperArgs'):
pulumi.set(self, "zookeeper", value)
@property
@pulumi.getter(name="assignPublicIp")
def assign_public_ip(self) -> Optional[bool]:
"""
The flag that defines whether a public IP address is assigned to the node.
"""
return pulumi.get(self, "assign_public_ip")
@assign_public_ip.setter
def assign_public_ip(self, value: Optional[bool]):
pulumi.set(self, "assign_public_ip", value)
@property
@pulumi.getter(name="brokersCount")
def brokers_count(self) -> Optional[int]:
"""
(Optional) Count of brokers per availability zone.
"""
return pulumi.get(self, "brokers_count")
@brokers_count.setter
def brokers_count(self, value: Optional[int]):
pulumi.set(self, "brokers_count", value)
@property
@pulumi.getter(name="schemaRegistry")
def schema_registry(self) -> Optional[bool]:
"""
(Optional) Enables managed schema registry on cluster. Can be either `true` or `false`.
"""
return pulumi.get(self, "schema_registry")
@schema_registry.setter
def schema_registry(self, value: Optional[bool]):
pulumi.set(self, "schema_registry", value)
@property
@pulumi.getter(name="unmanagedTopics")
def unmanaged_topics(self) -> Optional[bool]:
"""
(Optional) Allows to use Kafka AdminAPI to manage topics. Can be either `true` or `false`.
"""
return pulumi.get(self, "unmanaged_topics")
@unmanaged_topics.setter
def unmanaged_topics(self, value: Optional[bool]):
pulumi.set(self, "unmanaged_topics", value)
@pulumi.input_type
class GetMdbKafkaClusterConfigKafkaArgs:
def __init__(__self__, *,
resources: 'GetMdbKafkaClusterConfigKafkaResourcesArgs',
kafka_config: Optional['GetMdbKafkaClusterConfigKafkaKafkaConfigArgs'] = None):
"""
:param 'GetMdbKafkaClusterConfigKafkaResourcesArgs' resources: (Optional) Resources allocated to hosts of the ZooKeeper subcluster. The structure is documented below.
:param 'GetMdbKafkaClusterConfigKafkaKafkaConfigArgs' kafka_config: (Optional) User-defined settings for the Kafka cluster. The structure is documented below.
"""
pulumi.set(__self__, "resources", resources)
if kafka_config is not None:
pulumi.set(__self__, "kafka_config", kafka_config)
@property
@pulumi.getter
def resources(self) -> 'GetMdbKafkaClusterConfigKafkaResourcesArgs':
"""
(Optional) Resources allocated to hosts of the ZooKeeper subcluster. The structure is documented below.
"""
return pulumi.get(self, "resources")
@resources.setter
def resources(self, value: 'GetMdbKafkaClusterConfigKafkaResourcesArgs'):
pulumi.set(self, "resources", value)
@property
@pulumi.getter(name="kafkaConfig")
def kafka_config(self) -> Optional['GetMdbKafkaClusterConfigKafkaKafkaConfigArgs']:
"""
(Optional) User-defined settings for the Kafka cluster. The structure is documented below.
"""
return pulumi.get(self, "kafka_config")
@kafka_config.setter
def kafka_config(self, value: Optional['GetMdbKafkaClusterConfigKafkaKafkaConfigArgs']):
pulumi.set(self, "kafka_config", value)
@pulumi.input_type
class GetMdbKafkaClusterConfigKafkaKafkaConfigArgs:
def __init__(__self__, *,
auto_create_topics_enable: Optional[bool] = None,
compression_type: Optional[str] = None,
default_replication_factor: Optional[str] = None,
log_flush_interval_messages: Optional[str] = None,
log_flush_interval_ms: Optional[str] = None,
log_flush_scheduler_interval_ms: Optional[str] = None,
log_preallocate: Optional[bool] = None,
log_retention_bytes: Optional[str] = None,
log_retention_hours: Optional[str] = None,
log_retention_minutes: Optional[str] = None,
log_retention_ms: Optional[str] = None,
log_segment_bytes: Optional[str] = None,
num_partitions: Optional[str] = None,
socket_receive_buffer_bytes: Optional[str] = None,
socket_send_buffer_bytes: Optional[str] = None):
if auto_create_topics_enable is not None:
pulumi.set(__self__, "auto_create_topics_enable", auto_create_topics_enable)
if compression_type is not None:
pulumi.set(__self__, "compression_type", compression_type)
if default_replication_factor is not None:
pulumi.set(__self__, "default_replication_factor", default_replication_factor)
if log_flush_interval_messages is not None:
pulumi.set(__self__, "log_flush_interval_messages", log_flush_interval_messages)
if log_flush_interval_ms is not None:
pulumi.set(__self__, "log_flush_interval_ms", log_flush_interval_ms)
if log_flush_scheduler_interval_ms is not None:
pulumi.set(__self__, "log_flush_scheduler_interval_ms", log_flush_scheduler_interval_ms)
if log_preallocate is not None:
pulumi.set(__self__, "log_preallocate", log_preallocate)
if log_retention_bytes is not None:
pulumi.set(__self__, "log_retention_bytes", log_retention_bytes)
if log_retention_hours is not None:
pulumi.set(__self__, "log_retention_hours", log_retention_hours)
if log_retention_minutes is not None:
pulumi.set(__self__, "log_retention_minutes", log_retention_minutes)
if log_retention_ms is not None:
pulumi.set(__self__, "log_retention_ms", log_retention_ms)
if log_segment_bytes is not None:
pulumi.set(__self__, "log_segment_bytes", log_segment_bytes)
if num_partitions is not None:
pulumi.set(__self__, "num_partitions", num_partitions)
if socket_receive_buffer_bytes is not None:
pulumi.set(__self__, "socket_receive_buffer_bytes", socket_receive_buffer_bytes)
if socket_send_buffer_bytes is not None:
pulumi.set(__self__, "socket_send_buffer_bytes", socket_send_buffer_bytes)
@property
@pulumi.getter(name="autoCreateTopicsEnable")
def auto_create_topics_enable(self) -> Optional[bool]:
return pulumi.get(self, "auto_create_topics_enable")
@auto_create_topics_enable.setter
def auto_create_topics_enable(self, value: Optional[bool]):
pulumi.set(self, "auto_create_topics_enable", value)
@property
@pulumi.getter(name="compressionType")
def compression_type(self) -> Optional[str]:
return pulumi.get(self, "compression_type")
@compression_type.setter
def compression_type(self, value: Optional[str]):
pulumi.set(self, "compression_type", value)
@property
@pulumi.getter(name="defaultReplicationFactor")
def default_replication_factor(self) -> Optional[str]:
return pulumi.get(self, "default_replication_factor")
@default_replication_factor.setter
def default_replication_factor(self, value: Optional[str]):
pulumi.set(self, "default_replication_factor", value)
@property
@pulumi.getter(name="logFlushIntervalMessages")
def log_flush_interval_messages(self) -> Optional[str]:
return pulumi.get(self, "log_flush_interval_messages")
@log_flush_interval_messages.setter
def log_flush_interval_messages(self, value: Optional[str]):
pulumi.set(self, "log_flush_interval_messages", value)
@property
@pulumi.getter(name="logFlushIntervalMs")
def log_flush_interval_ms(self) -> Optional[str]:
return pulumi.get(self, "log_flush_interval_ms")
@log_flush_interval_ms.setter
def log_flush_interval_ms(self, value: Optional[str]):
pulumi.set(self, "log_flush_interval_ms", value)
@property
@pulumi.getter(name="logFlushSchedulerIntervalMs")
def log_flush_scheduler_interval_ms(self) -> Optional[str]:
return pulumi.get(self, "log_flush_scheduler_interval_ms")
@log_flush_scheduler_interval_ms.setter
def log_flush_scheduler_interval_ms(self, value: Optional[str]):
pulumi.set(self, "log_flush_scheduler_interval_ms", value)
@property
@pulumi.getter(name="logPreallocate")
def log_preallocate(self) -> Optional[bool]:
return pulumi.get(self, "log_preallocate")
@log_preallocate.setter
def log_preallocate(self, value: Optional[bool]):
pulumi.set(self, "log_preallocate", value)
@property
@pulumi.getter(name="logRetentionBytes")
def log_retention_bytes(self) -> Optional[str]:
return pulumi.get(self, "log_retention_bytes")
@log_retention_bytes.setter
def log_retention_bytes(self, value: Optional[str]):
pulumi.set(self, "log_retention_bytes", value)
@property
@pulumi.getter(name="logRetentionHours")
def log_retention_hours(self) -> Optional[str]:
return pulumi.get(self, "log_retention_hours")
@log_retention_hours.setter
def log_retention_hours(self, value: Optional[str]):
pulumi.set(self, "log_retention_hours", value)
@property
@pulumi.getter(name="logRetentionMinutes")
def log_retention_minutes(self) -> Optional[str]:
return pulumi.get(self, "log_retention_minutes")
@log_retention_minutes.setter
def log_retention_minutes(self, value: Optional[str]):
pulumi.set(self, "log_retention_minutes", value)
@property
@pulumi.getter(name="logRetentionMs")
def log_retention_ms(self) -> Optional[str]:
return pulumi.get(self, "log_retention_ms")
@log_retention_ms.setter
def log_retention_ms(self, value: Optional[str]):
pulumi.set(self, "log_retention_ms", value)
@property
@pulumi.getter(name="logSegmentBytes")
def log_segment_bytes(self) -> Optional[str]:
return pulumi.get(self, "log_segment_bytes")
@log_segment_bytes.setter
def log_segment_bytes(self, value: Optional[str]):
pulumi.set(self, "log_segment_bytes", value)
@property
@pulumi.getter(name="numPartitions")
def num_partitions(self) -> Optional[str]:
return pulumi.get(self, "num_partitions")
@num_partitions.setter
def num_partitions(self, value: Optional[str]):
pulumi.set(self, "num_partitions", value)
@property
@pulumi.getter(name="socketReceiveBufferBytes")
def socket_receive_buffer_bytes(self) -> Optional[str]:
return pulumi.get(self, "socket_receive_buffer_bytes")
@socket_receive_buffer_bytes.setter
def socket_receive_buffer_bytes(self, value: Optional[str]):
pulumi.set(self, "socket_receive_buffer_bytes", value)
@property
@pulumi.getter(name="socketSendBufferBytes")
def socket_send_buffer_bytes(self) -> Optional[str]:
return pulumi.get(self, "socket_send_buffer_bytes")
@socket_send_buffer_bytes.setter
def socket_send_buffer_bytes(self, value: Optional[str]):
pulumi.set(self, "socket_send_buffer_bytes", value)
@pulumi.input_type
class GetMdbKafkaClusterConfigKafkaResourcesArgs:
def __init__(__self__, *,
disk_size: int,
disk_type_id: str,
resource_preset_id: str):
"""
:param int disk_size: (Optional) Volume of the storage available to a ZooKeeper host, in gigabytes.
:param str disk_type_id: (Optional) Type of the storage of ZooKeeper hosts.
For more information see [the official documentation](https://cloud.yandex.com/docs/managed-kafka/concepts/storage).
"""
pulumi.set(__self__, "disk_size", disk_size)
pulumi.set(__self__, "disk_type_id", disk_type_id)
pulumi.set(__self__, "resource_preset_id", resource_preset_id)
@property
@pulumi.getter(name="diskSize")
def disk_size(self) -> int:
"""
(Optional) Volume of the storage available to a ZooKeeper host, in gigabytes.
"""
return pulumi.get(self, "disk_size")
@disk_size.setter
def disk_size(self, value: int):
pulumi.set(self, "disk_size", value)
@property
@pulumi.getter(name="diskTypeId")
def disk_type_id(self) -> str:
"""
(Optional) Type of the storage of ZooKeeper hosts.
For more information see [the official documentation](https://cloud.yandex.com/docs/managed-kafka/concepts/storage).
"""
return pulumi.get(self, "disk_type_id")
@disk_type_id.setter
def disk_type_id(self, value: str):
pulumi.set(self, "disk_type_id", value)
@property
@pulumi.getter(name="resourcePresetId")
def resource_preset_id(self) -> str:
return pulumi.get(self, "resource_preset_id")
@resource_preset_id.setter
def resource_preset_id(self, value: str):
pulumi.set(self, "resource_preset_id", value)
@pulumi.input_type
class GetMdbKafkaClusterConfigZookeeperArgs:
def __init__(__self__, *,
resources: 'GetMdbKafkaClusterConfigZookeeperResourcesArgs'):
"""
:param 'GetMdbKafkaClusterConfigZookeeperResourcesArgs' resources: (Optional) Resources allocated to hosts of the ZooKeeper subcluster. The structure is documented below.
"""
pulumi.set(__self__, "resources", resources)
@property
@pulumi.getter
def resources(self) -> 'GetMdbKafkaClusterConfigZookeeperResourcesArgs':
"""
(Optional) Resources allocated to hosts of the ZooKeeper subcluster. The structure is documented below.
"""
return pulumi.get(self, "resources")
@resources.setter
def resources(self, value: 'GetMdbKafkaClusterConfigZookeeperResourcesArgs'):
pulumi.set(self, "resources", value)
@pulumi.input_type
class GetMdbKafkaClusterConfigZookeeperResourcesArgs:
def __init__(__self__, *,
disk_size: int,
disk_type_id: str,
resource_preset_id: str):
"""
:param int disk_size: (Optional) Volume of the storage available to a ZooKeeper host, in gigabytes.
:param str disk_type_id: (Optional) Type of the storage of ZooKeeper hosts.
For more information see [the official documentation](https://cloud.yandex.com/docs/managed-kafka/concepts/storage).
"""
pulumi.set(__self__, "disk_size", disk_size)
pulumi.set(__self__, "disk_type_id", disk_type_id)
pulumi.set(__self__, "resource_preset_id", resource_preset_id)
@property
@pulumi.getter(name="diskSize")
def disk_size(self) -> int:
"""
(Optional) Volume of the storage available to a ZooKeeper host, in gigabytes.
"""
return pulumi.get(self, "disk_size")
@disk_size.setter
def disk_size(self, value: int):
pulumi.set(self, "disk_size", value)
@property
@pulumi.getter(name="diskTypeId")
def disk_type_id(self) -> str:
"""
(Optional) Type of the storage of ZooKeeper hosts.
For more information see [the official documentation](https://cloud.yandex.com/docs/managed-kafka/concepts/storage).
"""
return pulumi.get(self, "disk_type_id")
@disk_type_id.setter
def disk_type_id(self, value: str):
pulumi.set(self, "disk_type_id", value)
@property
@pulumi.getter(name="resourcePresetId")
def resource_preset_id(self) -> str:
return pulumi.get(self, "resource_preset_id")
@resource_preset_id.setter
def resource_preset_id(self, value: str):
pulumi.set(self, "resource_preset_id", value)
@pulumi.input_type
class GetMdbKafkaClusterTopicArgs:
def __init__(__self__, *,
cluster_id: str,
name: str,
partitions: int,
replication_factor: int,
topic_config: Optional['GetMdbKafkaClusterTopicTopicConfigArgs'] = None):
"""
:param str cluster_id: The ID of the Kafka cluster.
:param str name: The name of the Kafka cluster.
:param int partitions: (Required) The number of the topic's partitions.
:param int replication_factor: (Required) Amount of data copies (replicas) for the topic in the cluster.
:param 'GetMdbKafkaClusterTopicTopicConfigArgs' topic_config: (Required) User-defined settings for the topic. The structure is documented below.
"""
pulumi.set(__self__, "cluster_id", cluster_id)
pulumi.set(__self__, "name", name)
pulumi.set(__self__, "partitions", partitions)
pulumi.set(__self__, "replication_factor", replication_factor)
if topic_config is not None:
pulumi.set(__self__, "topic_config", topic_config)
@property
@pulumi.getter(name="clusterId")
def cluster_id(self) -> str:
"""
The ID of the Kafka cluster.
"""
return pulumi.get(self, "cluster_id")
@cluster_id.setter
def cluster_id(self, value: str):
pulumi.set(self, "cluster_id", value)
@property
@pulumi.getter
def name(self) -> str:
"""
The name of the Kafka cluster.
"""
return pulumi.get(self, "name")
@name.setter
def name(self, value: str):
pulumi.set(self, "name", value)
@property
@pulumi.getter
def partitions(self) -> int:
"""
(Required) The number of the topic's partitions.
"""
return pulumi.get(self, "partitions")
@partitions.setter
def partitions(self, value: int):
pulumi.set(self, "partitions", value)
@property
@pulumi.getter(name="replicationFactor")
def replication_factor(self) -> int:
"""
(Required) Amount of data copies (replicas) for the topic in the cluster.
"""
return pulumi.get(self, "replication_factor")
@replication_factor.setter
def replication_factor(self, value: int):
pulumi.set(self, "replication_factor", value)
@property
@pulumi.getter(name="topicConfig")
def topic_config(self) -> Optional['GetMdbKafkaClusterTopicTopicConfigArgs']:
"""
(Required) User-defined settings for the topic. The structure is documented below.
"""
return pulumi.get(self, "topic_config")
@topic_config.setter
def topic_config(self, value: Optional['GetMdbKafkaClusterTopicTopicConfigArgs']):
pulumi.set(self, "topic_config", value)
@pulumi.input_type
class GetMdbKafkaClusterTopicTopicConfigArgs:
def __init__(__self__, *,
cleanup_policy: Optional[str] = None,
compression_type: Optional[str] = None,
delete_retention_ms: Optional[str] = None,
file_delete_delay_ms: Optional[str] = None,
flush_messages: Optional[str] = None,
flush_ms: Optional[str] = None,
max_message_bytes: Optional[str] = None,
min_compaction_lag_ms: Optional[str] = None,
min_insync_replicas: Optional[str] = None,
preallocate: Optional[bool] = None,
retention_bytes: Optional[str] = None,
retention_ms: Optional[str] = None,
segment_bytes: Optional[str] = None):
if cleanup_policy is not None:
pulumi.set(__self__, "cleanup_policy", cleanup_policy)
if compression_type is not None:
pulumi.set(__self__, "compression_type", compression_type)
if delete_retention_ms is not None:
pulumi.set(__self__, "delete_retention_ms", delete_retention_ms)
if file_delete_delay_ms is not None:
pulumi.set(__self__, "file_delete_delay_ms", file_delete_delay_ms)
if flush_messages is not None:
pulumi.set(__self__, "flush_messages", flush_messages)
if flush_ms is not None:
pulumi.set(__self__, "flush_ms", flush_ms)
if max_message_bytes is not None:
pulumi.set(__self__, "max_message_bytes", max_message_bytes)
if min_compaction_lag_ms is not None:
pulumi.set(__self__, "min_compaction_lag_ms", min_compaction_lag_ms)
if min_insync_replicas is not None:
pulumi.set(__self__, "min_insync_replicas", min_insync_replicas)
if preallocate is not None:
pulumi.set(__self__, "preallocate", preallocate)
if retention_bytes is not None:
pulumi.set(__self__, "retention_bytes", retention_bytes)
if retention_ms is not None:
pulumi.set(__self__, "retention_ms", retention_ms)
if segment_bytes is not None:
pulumi.set(__self__, "segment_bytes", segment_bytes)
@property
@pulumi.getter(name="cleanupPolicy")
def cleanup_policy(self) -> Optional[str]:
return pulumi.get(self, "cleanup_policy")
@cleanup_policy.setter
def cleanup_policy(self, value: Optional[str]):
pulumi.set(self, "cleanup_policy", value)
@property
@pulumi.getter(name="compressionType")
def compression_type(self) -> Optional[str]:
return pulumi.get(self, "compression_type")
@compression_type.setter
def compression_type(self, value: Optional[str]):
pulumi.set(self, "compression_type", value)
@property
@pulumi.getter(name="deleteRetentionMs")
def delete_retention_ms(self) -> Optional[str]:
return pulumi.get(self, "delete_retention_ms")
@delete_retention_ms.setter
def delete_retention_ms(self, value: Optional[str]):
pulumi.set(self, "delete_retention_ms", value)
@property
@pulumi.getter(name="fileDeleteDelayMs")
def file_delete_delay_ms(self) -> Optional[str]:
return pulumi.get(self, "file_delete_delay_ms")
@file_delete_delay_ms.setter
def file_delete_delay_ms(self, value: Optional[str]):
pulumi.set(self, "file_delete_delay_ms", value)
@property
@pulumi.getter(name="flushMessages")
def flush_messages(self) -> Optional[str]:
return pulumi.get(self, "flush_messages")
@flush_messages.setter
def flush_messages(self, value: Optional[str]):
pulumi.set(self, "flush_messages", value)
@property
@pulumi.getter(name="flushMs")
def flush_ms(self) -> Optional[str]:
return pulumi.get(self, "flush_ms")
@flush_ms.setter
def flush_ms(self, value: Optional[str]):
pulumi.set(self, "flush_ms", value)
@property
@pulumi.getter(name="maxMessageBytes")
def max_message_bytes(self) -> Optional[str]:
return pulumi.get(self, "max_message_bytes")
@max_message_bytes.setter
def max_message_bytes(self, value: Optional[str]):
pulumi.set(self, "max_message_bytes", value)
@property
@pulumi.getter(name="minCompactionLagMs")
def min_compaction_lag_ms(self) -> Optional[str]:
return pulumi.get(self, "min_compaction_lag_ms")
@min_compaction_lag_ms.setter
def min_compaction_lag_ms(self, value: Optional[str]):
pulumi.set(self, "min_compaction_lag_ms", value)
@property
@pulumi.getter(name="minInsyncReplicas")
def min_insync_replicas(self) -> Optional[str]:
return pulumi.get(self, "min_insync_replicas")
@min_insync_replicas.setter
def min_insync_replicas(self, value: Optional[str]):
pulumi.set(self, "min_insync_replicas", value)
@property
@pulumi.getter
def preallocate(self) -> Optional[bool]:
return pulumi.get(self, "preallocate")
@preallocate.setter
def preallocate(self, value: Optional[bool]):
pulumi.set(self, "preallocate", value)
@property
@pulumi.getter(name="retentionBytes")
def retention_bytes(self) -> Optional[str]:
return pulumi.get(self, "retention_bytes")
@retention_bytes.setter
def retention_bytes(self, value: Optional[str]):
pulumi.set(self, "retention_bytes", value)
@property
@pulumi.getter(name="retentionMs")
def retention_ms(self) -> Optional[str]:
return pulumi.get(self, "retention_ms")
@retention_ms.setter
def retention_ms(self, value: Optional[str]):
pulumi.set(self, "retention_ms", value)
@property
@pulumi.getter(name="segmentBytes")
def segment_bytes(self) -> Optional[str]:
return pulumi.get(self, "segment_bytes")
@segment_bytes.setter
def segment_bytes(self, value: Optional[str]):
pulumi.set(self, "segment_bytes", value)
@pulumi.input_type
class GetMdbKafkaClusterUserArgs:
def __init__(__self__, *,
name: str,
password: str,
permissions: Optional[Sequence['GetMdbKafkaClusterUserPermissionArgs']] = None):
"""
:param str name: The name of the Kafka cluster.
:param str password: (Required) The password of the user.
:param Sequence['GetMdbKafkaClusterUserPermissionArgs'] permissions: (Optional) Set of permissions granted to the user. The structure is documented below.
"""
pulumi.set(__self__, "name", name)
pulumi.set(__self__, "password", password)
if permissions is not None:
pulumi.set(__self__, "permissions", permissions)
@property
@pulumi.getter
def name(self) -> str:
"""
The name of the Kafka cluster.
"""
return pulumi.get(self, "name")
@name.setter
def name(self, value: str):
pulumi.set(self, "name", value)
@property
@pulumi.getter
def password(self) -> str:
"""
(Required) The password of the user.
"""
return pulumi.get(self, "password")
@password.setter
def password(self, value: str):
pulumi.set(self, "password", value)
@property
@pulumi.getter
def permissions(self) -> Optional[Sequence['GetMdbKafkaClusterUserPermissionArgs']]:
"""
(Optional) Set of permissions granted to the user. The structure is documented below.
"""
return pulumi.get(self, "permissions")
@permissions.setter
def permissions(self, value: Optional[Sequence['GetMdbKafkaClusterUserPermissionArgs']]):
pulumi.set(self, "permissions", value)
@pulumi.input_type
class GetMdbKafkaClusterUserPermissionArgs:
def __init__(__self__, *,
role: str,
topic_name: str):
"""
:param str role: Role of the host in the cluster.
:param str topic_name: (Required) The name of the topic that the permission grants access to.
"""
pulumi.set(__self__, "role", role)
pulumi.set(__self__, "topic_name", topic_name)
@property
@pulumi.getter
def role(self) -> str:
"""
Role of the host in the cluster.
"""
return pulumi.get(self, "role")
@role.setter
def role(self, value: str):
pulumi.set(self, "role", value)
@property
@pulumi.getter(name="topicName")
def topic_name(self) -> str:
"""
(Required) The name of the topic that the permission grants access to.
"""
return pulumi.get(self, "topic_name")
@topic_name.setter
def topic_name(self, value: str):
pulumi.set(self, "topic_name", value)
@pulumi.input_type
class GetMdbMysqlClusterAccessArgs:
def __init__(__self__, *,
data_lens: bool,
web_sql: bool):
"""
:param bool data_lens: Allow access for [Yandex DataLens](https://cloud.yandex.com/services/datalens).
:param bool web_sql: Allows access for [SQL queries in the management console](https://cloud.yandex.com/docs/managed-mysql/operations/web-sql-query).
"""
pulumi.set(__self__, "data_lens", data_lens)
pulumi.set(__self__, "web_sql", web_sql)
@property
@pulumi.getter(name="dataLens")
def data_lens(self) -> bool:
"""
Allow access for [Yandex DataLens](https://cloud.yandex.com/services/datalens).
"""
return pulumi.get(self, "data_lens")
@data_lens.setter
def data_lens(self, value: bool):
pulumi.set(self, "data_lens", value)
@property
@pulumi.getter(name="webSql")
def web_sql(self) -> bool:
"""
Allows access for [SQL queries in the management console](https://cloud.yandex.com/docs/managed-mysql/operations/web-sql-query).
"""
return pulumi.get(self, "web_sql")
@web_sql.setter
def web_sql(self, value: bool):
pulumi.set(self, "web_sql", value)
| [
"pulumi.get",
"pulumi.getter",
"pulumi.set",
"warnings.warn",
"pulumi.log.warn"
] | [((16397, 16433), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""targetGroupIds"""'}), "(name='targetGroupIds')\n", (16410, 16433), False, 'import pulumi\n'), ((17299, 17340), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""loadBalancingConfig"""'}), "(name='loadBalancingConfig')\n", (17312, 17340), False, 'import pulumi\n'), ((23050, 23087), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""grpcHealthcheck"""'}), "(name='grpcHealthcheck')\n", (23063, 23087), False, 'import pulumi\n'), ((23616, 23653), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""healthcheckPort"""'}), "(name='healthcheckPort')\n", (23629, 23653), False, 'import pulumi\n'), ((24018, 24056), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""healthyThreshold"""'}), "(name='healthyThreshold')\n", (24031, 24056), False, 'import pulumi\n'), ((24590, 24627), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""httpHealthcheck"""'}), "(name='httpHealthcheck')\n", (24603, 24627), False, 'import pulumi\n'), ((25156, 25199), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""intervalJitterPercent"""'}), "(name='intervalJitterPercent')\n", (25169, 25199), False, 'import pulumi\n'), ((25734, 25773), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""streamHealthcheck"""'}), "(name='streamHealthcheck')\n", (25747, 25773), False, 'import pulumi\n'), ((26318, 26358), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""unhealthyThreshold"""'}), "(name='unhealthyThreshold')\n", (26331, 26358), False, 'import pulumi\n'), ((27298, 27331), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""serviceName"""'}), "(name='serviceName')\n", (27311, 27331), False, 'import pulumi\n'), ((31878, 31927), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""localityAwareRoutingPercent"""'}), "(name='localityAwareRoutingPercent')\n", (31891, 31927), False, 'import pulumi\n'), ((32430, 32466), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""panicThreshold"""'}), "(name='panicThreshold')\n", (32443, 32466), False, 'import pulumi\n'), ((33037, 33073), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""strictLocality"""'}), "(name='strictLocality')\n", (33050, 33073), False, 'import pulumi\n'), ((34924, 34963), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""validationContext"""'}), "(name='validationContext')\n", (34937, 34963), False, 'import pulumi\n'), ((35821, 35857), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""trustedCaBytes"""'}), "(name='trustedCaBytes')\n", (35834, 35857), False, 'import pulumi\n'), ((36143, 36176), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""trustedCaId"""'}), "(name='trustedCaId')\n", (36156, 36176), False, 'import pulumi\n'), ((39114, 39150), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""targetGroupIds"""'}), "(name='targetGroupIds')\n", (39127, 39150), False, 'import pulumi\n'), ((40332, 40373), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""loadBalancingConfig"""'}), "(name='loadBalancingConfig')\n", (40345, 40373), False, 'import pulumi\n'), ((46083, 46120), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""grpcHealthcheck"""'}), "(name='grpcHealthcheck')\n", (46096, 46120), False, 'import pulumi\n'), ((46649, 46686), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""healthcheckPort"""'}), "(name='healthcheckPort')\n", (46662, 46686), False, 'import pulumi\n'), ((47051, 47089), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""healthyThreshold"""'}), "(name='healthyThreshold')\n", (47064, 47089), False, 'import pulumi\n'), ((47623, 47660), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""httpHealthcheck"""'}), "(name='httpHealthcheck')\n", (47636, 47660), False, 'import pulumi\n'), ((48189, 48232), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""intervalJitterPercent"""'}), "(name='intervalJitterPercent')\n", (48202, 48232), False, 'import pulumi\n'), ((48767, 48806), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""streamHealthcheck"""'}), "(name='streamHealthcheck')\n", (48780, 48806), False, 'import pulumi\n'), ((49351, 49391), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""unhealthyThreshold"""'}), "(name='unhealthyThreshold')\n", (49364, 49391), False, 'import pulumi\n'), ((50331, 50364), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""serviceName"""'}), "(name='serviceName')\n", (50344, 50364), False, 'import pulumi\n'), ((54911, 54960), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""localityAwareRoutingPercent"""'}), "(name='localityAwareRoutingPercent')\n", (54924, 54960), False, 'import pulumi\n'), ((55463, 55499), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""panicThreshold"""'}), "(name='panicThreshold')\n", (55476, 55499), False, 'import pulumi\n'), ((56070, 56106), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""strictLocality"""'}), "(name='strictLocality')\n", (56083, 56106), False, 'import pulumi\n'), ((57957, 57996), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""validationContext"""'}), "(name='validationContext')\n", (57970, 57996), False, 'import pulumi\n'), ((58854, 58890), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""trustedCaBytes"""'}), "(name='trustedCaBytes')\n", (58867, 58890), False, 'import pulumi\n'), ((59176, 59209), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""trustedCaId"""'}), "(name='trustedCaId')\n", (59189, 59209), False, 'import pulumi\n'), ((61207, 61237), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""subnetId"""'}), "(name='subnetId')\n", (61220, 61237), False, 'import pulumi\n'), ((61550, 61578), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""zoneId"""'}), "(name='zoneId')\n", (61563, 61578), False, 'import pulumi\n'), ((61875, 61911), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""disableTraffic"""'}), "(name='disableTraffic')\n", (61888, 61911), False, 'import pulumi\n'), ((67963, 68004), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""externalIpv4Address"""'}), "(name='externalIpv4Address')\n", (67976, 68004), False, 'import pulumi\n'), ((68525, 68566), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""externalIpv6Address"""'}), "(name='externalIpv6Address')\n", (68538, 68566), False, 'import pulumi\n'), ((69087, 69128), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""internalIpv4Address"""'}), "(name='internalIpv4Address')\n", (69100, 69128), False, 'import pulumi\n'), ((72058, 72088), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""subnetId"""'}), "(name='subnetId')\n", (72071, 72088), False, 'import pulumi\n'), ((75174, 75207), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""allowHttp10"""'}), "(name='allowHttp10')\n", (75187, 75207), False, 'import pulumi\n'), ((75569, 75603), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""http2Options"""'}), "(name='http2Options')\n", (75582, 75603), False, 'import pulumi\n'), ((76091, 76125), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""httpRouterId"""'}), "(name='httpRouterId')\n", (76104, 76125), False, 'import pulumi\n'), ((76885, 76927), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxConcurrentStreams"""'}), "(name='maxConcurrentStreams')\n", (76898, 76927), False, 'import pulumi\n'), ((77585, 77618), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""httpToHttps"""'}), "(name='httpToHttps')\n", (77598, 77618), False, 'import pulumi\n'), ((78711, 78747), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""defaultHandler"""'}), "(name='defaultHandler')\n", (78724, 78747), False, 'import pulumi\n'), ((79183, 79216), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""sniHandlers"""'}), "(name='sniHandlers')\n", (79196, 79216), False, 'import pulumi\n'), ((80640, 80676), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""certificateIds"""'}), "(name='certificateIds')\n", (80653, 80676), False, 'import pulumi\n'), ((81231, 81264), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""httpHandler"""'}), "(name='httpHandler')\n", (81244, 81264), False, 'import pulumi\n'), ((82823, 82856), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""allowHttp10"""'}), "(name='allowHttp10')\n", (82836, 82856), False, 'import pulumi\n'), ((83218, 83252), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""http2Options"""'}), "(name='http2Options')\n", (83231, 83252), False, 'import pulumi\n'), ((83774, 83808), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""httpRouterId"""'}), "(name='httpRouterId')\n", (83787, 83808), False, 'import pulumi\n'), ((84585, 84627), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxConcurrentStreams"""'}), "(name='maxConcurrentStreams')\n", (84598, 84627), False, 'import pulumi\n'), ((86525, 86558), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""serverNames"""'}), "(name='serverNames')\n", (86538, 86558), False, 'import pulumi\n'), ((87861, 87897), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""certificateIds"""'}), "(name='certificateIds')\n", (87874, 87897), False, 'import pulumi\n'), ((88452, 88485), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""httpHandler"""'}), "(name='httpHandler')\n", (88465, 88485), False, 'import pulumi\n'), ((90059, 90092), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""allowHttp10"""'}), "(name='allowHttp10')\n", (90072, 90092), False, 'import pulumi\n'), ((90454, 90488), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""http2Options"""'}), "(name='http2Options')\n", (90467, 90488), False, 'import pulumi\n'), ((91016, 91050), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""httpRouterId"""'}), "(name='httpRouterId')\n", (91029, 91050), False, 'import pulumi\n'), ((91830, 91872), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxConcurrentStreams"""'}), "(name='maxConcurrentStreams')\n", (91843, 91872), False, 'import pulumi\n'), ((92851, 92882), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""ipAddress"""'}), "(name='ipAddress')\n", (92864, 92882), False, 'import pulumi\n'), ((93176, 93206), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""subnetId"""'}), "(name='subnetId')\n", (93189, 93206), False, 'import pulumi\n'), ((99666, 99697), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""grpcRoute"""'}), "(name='grpcRoute')\n", (99679, 99697), False, 'import pulumi\n'), ((100103, 100134), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""httpRoute"""'}), "(name='httpRoute')\n", (100116, 100134), False, 'import pulumi\n'), ((102219, 102252), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""grpcMatches"""'}), "(name='grpcMatches')\n", (102232, 102252), False, 'import pulumi\n'), ((102743, 102780), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""grpcRouteAction"""'}), "(name='grpcRouteAction')\n", (102756, 102780), False, 'import pulumi\n'), ((103258, 103304), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""grpcStatusResponseAction"""'}), "(name='grpcStatusResponseAction')\n", (103271, 103304), False, 'import pulumi\n'), ((107411, 107447), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""backendGroupId"""'}), "(name='backendGroupId')\n", (107424, 107447), False, 'import pulumi\n'), ((107778, 107815), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""autoHostRewrite"""'}), "(name='autoHostRewrite')\n", (107791, 107815), False, 'import pulumi\n'), ((108181, 108214), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""hostRewrite"""'}), "(name='hostRewrite')\n", (108194, 108214), False, 'import pulumi\n'), ((108536, 108569), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""idleTimeout"""'}), "(name='idleTimeout')\n", (108549, 108569), False, 'import pulumi\n'), ((109205, 109237), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxTimeout"""'}), "(name='maxTimeout')\n", (109218, 109237), False, 'import pulumi\n'), ((112321, 112363), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""directResponseAction"""'}), "(name='directResponseAction')\n", (112334, 112363), False, 'import pulumi\n'), ((112881, 112914), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""httpMatches"""'}), "(name='httpMatches')\n", (112894, 112914), False, 'import pulumi\n'), ((113405, 113442), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""httpRouteAction"""'}), "(name='httpRouteAction')\n", (113418, 113442), False, 'import pulumi\n'), ((113920, 113956), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""redirectAction"""'}), "(name='redirectAction')\n", (113933, 113956), False, 'import pulumi\n'), ((116553, 116586), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""httpMethods"""'}), "(name='httpMethods')\n", (116566, 116586), False, 'import pulumi\n'), ((120667, 120703), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""backendGroupId"""'}), "(name='backendGroupId')\n", (120680, 120703), False, 'import pulumi\n'), ((121034, 121071), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""autoHostRewrite"""'}), "(name='autoHostRewrite')\n", (121047, 121071), False, 'import pulumi\n'), ((121437, 121470), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""hostRewrite"""'}), "(name='hostRewrite')\n", (121450, 121470), False, 'import pulumi\n'), ((121792, 121825), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""idleTimeout"""'}), "(name='idleTimeout')\n", (121805, 121825), False, 'import pulumi\n'), ((122461, 122496), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""prefixRewrite"""'}), "(name='prefixRewrite')\n", (122474, 122496), False, 'import pulumi\n'), ((123298, 123332), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""upgradeTypes"""'}), "(name='upgradeTypes')\n", (123311, 123332), False, 'import pulumi\n'), ((125925, 125958), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""removeQuery"""'}), "(name='removeQuery')\n", (125938, 125958), False, 'import pulumi\n'), ((126226, 126259), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""replaceHost"""'}), "(name='replaceHost')\n", (126239, 126259), False, 'import pulumi\n'), ((126576, 126609), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""replacePath"""'}), "(name='replacePath')\n", (126589, 126609), False, 'import pulumi\n'), ((126921, 126954), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""replacePort"""'}), "(name='replacePort')\n", (126934, 126954), False, 'import pulumi\n'), ((127267, 127302), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""replacePrefix"""'}), "(name='replacePrefix')\n", (127280, 127302), False, 'import pulumi\n'), ((127790, 127825), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""replaceScheme"""'}), "(name='replaceScheme')\n", (127803, 127825), False, 'import pulumi\n'), ((128253, 128287), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""responseCode"""'}), "(name='responseCode')\n", (128266, 128287), False, 'import pulumi\n'), ((129105, 129147), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""diskPlacementGroupId"""'}), "(name='diskPlacementGroupId')\n", (129118, 129147), False, 'import pulumi\n'), ((131245, 131277), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""autoDelete"""'}), "(name='autoDelete')\n", (131258, 131277), False, 'import pulumi\n'), ((131671, 131703), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""deviceName"""'}), "(name='deviceName')\n", (131684, 131703), False, 'import pulumi\n'), ((132079, 132107), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""diskId"""'}), "(name='diskId')\n", (132092, 132107), False, 'import pulumi\n'), ((132429, 132467), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""initializeParams"""'}), "(name='initializeParams')\n", (132442, 132467), False, 'import pulumi\n'), ((135175, 135204), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""imageId"""'}), "(name='imageId')\n", (135188, 135204), False, 'import pulumi\n'), ((136114, 136146), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""snapshotId"""'}), "(name='snapshotId')\n", (136127, 136146), False, 'import pulumi\n'), ((139380, 139427), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxOpeningTrafficDuration"""'}), "(name='maxOpeningTrafficDuration')\n", (139393, 139427), False, 'import pulumi\n'), ((139980, 140015), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""statusMessage"""'}), "(name='statusMessage')\n", (139993, 140015), False, 'import pulumi\n'), ((140359, 140403), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""targetGroupDescription"""'}), "(name='targetGroupDescription')\n", (140372, 140403), False, 'import pulumi\n'), ((140796, 140831), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""targetGroupId"""'}), "(name='targetGroupId')\n", (140809, 140831), False, 'import pulumi\n'), ((141112, 141151), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""targetGroupLabels"""'}), "(name='targetGroupLabels')\n", (141125, 141151), False, 'import pulumi\n'), ((141572, 141609), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""targetGroupName"""'}), "(name='targetGroupName')\n", (141585, 141609), False, 'import pulumi\n'), ((144258, 144292), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxExpansion"""'}), "(name='maxExpansion')\n", (144271, 144292), False, 'import pulumi\n'), ((144706, 144742), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxUnavailable"""'}), "(name='maxUnavailable')\n", (144719, 144742), False, 'import pulumi\n'), ((145174, 145207), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxCreating"""'}), "(name='maxCreating')\n", (145187, 145207), False, 'import pulumi\n'), ((145575, 145608), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxDeleting"""'}), "(name='maxDeleting')\n", (145588, 145608), False, 'import pulumi\n'), ((145976, 146013), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""startupDuration"""'}), "(name='startupDuration')\n", (145989, 146013), False, 'import pulumi\n'), ((149271, 149309), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""healthyThreshold"""'}), "(name='healthyThreshold')\n", (149284, 149309), False, 'import pulumi\n'), ((149720, 149753), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""httpOptions"""'}), "(name='httpOptions')\n", (149733, 149753), False, 'import pulumi\n'), ((150588, 150620), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""tcpOptions"""'}), "(name='tcpOptions')\n", (150601, 150620), False, 'import pulumi\n'), ((151475, 151515), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""unhealthyThreshold"""'}), "(name='unhealthyThreshold')\n", (151488, 151515), False, 'import pulumi\n'), ((156547, 156579), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""instanceId"""'}), "(name='instanceId')\n", (156560, 156579), False, 'import pulumi\n'), ((157621, 157660), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""networkInterfaces"""'}), "(name='networkInterfaces')\n", (157634, 157660), False, 'import pulumi\n'), ((158579, 158616), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""statusChangedAt"""'}), "(name='statusChangedAt')\n", (158592, 158616), False, 'import pulumi\n'), ((158907, 158942), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""statusMessage"""'}), "(name='statusMessage')\n", (158920, 158942), False, 'import pulumi\n'), ((159286, 159314), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""zoneId"""'}), "(name='zoneId')\n", (159299, 159314), False, 'import pulumi\n'), ((162561, 162592), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""ipAddress"""'}), "(name='ipAddress')\n", (162574, 162592), False, 'import pulumi\n'), ((163482, 163515), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""ipv6Address"""'}), "(name='ipv6Address')\n", (163495, 163515), False, 'import pulumi\n'), ((163845, 163877), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""macAddress"""'}), "(name='macAddress')\n", (163858, 163877), False, 'import pulumi\n'), ((164509, 164543), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""natIpAddress"""'}), "(name='natIpAddress')\n", (164522, 164543), False, 'import pulumi\n'), ((164942, 164976), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""natIpVersion"""'}), "(name='natIpVersion')\n", (164955, 164976), False, 'import pulumi\n'), ((165323, 165353), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""subnetId"""'}), "(name='subnetId')\n", (165336, 165353), False, 'import pulumi\n'), ((171891, 171921), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""bootDisk"""'}), "(name='bootDisk')\n", (171904, 171921), False, 'import pulumi\n'), ((172356, 172395), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""networkInterfaces"""'}), "(name='networkInterfaces')\n", (172369, 172395), False, 'import pulumi\n'), ((176343, 176380), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""networkSettings"""'}), "(name='networkSettings')\n", (176356, 176380), False, 'import pulumi\n'), ((176927, 176964), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""placementPolicy"""'}), "(name='placementPolicy')\n", (176940, 176964), False, 'import pulumi\n'), ((177461, 177493), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""platformId"""'}), "(name='platformId')\n", (177474, 177493), False, 'import pulumi\n'), ((177880, 177918), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""schedulingPolicy"""'}), "(name='schedulingPolicy')\n", (177893, 177918), False, 'import pulumi\n'), ((178423, 178459), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""secondaryDisks"""'}), "(name='secondaryDisks')\n", (178436, 178459), False, 'import pulumi\n'), ((179002, 179040), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""serviceAccountId"""'}), "(name='serviceAccountId')\n", (179015, 179040), False, 'import pulumi\n'), ((180797, 180829), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""deviceName"""'}), "(name='deviceName')\n", (180810, 180829), False, 'import pulumi\n'), ((181195, 181223), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""diskId"""'}), "(name='diskId')\n", (181208, 181223), False, 'import pulumi\n'), ((181543, 181581), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""initializeParams"""'}), "(name='initializeParams')\n", (181556, 181581), False, 'import pulumi\n'), ((184165, 184194), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""imageId"""'}), "(name='imageId')\n", (184178, 184194), False, 'import pulumi\n'), ((184835, 184867), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""snapshotId"""'}), "(name='snapshotId')\n", (184848, 184867), False, 'import pulumi\n'), ((189299, 189331), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""dnsRecords"""'}), "(name='dnsRecords')\n", (189312, 189331), False, 'import pulumi\n'), ((189857, 189888), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""ipAddress"""'}), "(name='ipAddress')\n", (189870, 189888), False, 'import pulumi\n'), ((190778, 190811), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""ipv6Address"""'}), "(name='ipv6Address')\n", (190791, 190811), False, 'import pulumi\n'), ((191141, 191177), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""ipv6DnsRecords"""'}), "(name='ipv6DnsRecords')\n", (191154, 191177), False, 'import pulumi\n'), ((192029, 192064), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""natDnsRecords"""'}), "(name='natDnsRecords')\n", (192042, 192064), False, 'import pulumi\n'), ((192620, 192654), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""natIpAddress"""'}), "(name='natIpAddress')\n", (192633, 192654), False, 'import pulumi\n'), ((193053, 193084), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""networkId"""'}), "(name='networkId')\n", (193066, 193084), False, 'import pulumi\n'), ((193395, 193433), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""securityGroupIds"""'}), "(name='securityGroupIds')\n", (193408, 193433), False, 'import pulumi\n'), ((193851, 193882), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""subnetIds"""'}), "(name='subnetIds')\n", (193864, 193882), False, 'import pulumi\n'), ((195537, 195568), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""dnsZoneId"""'}), "(name='dnsZoneId')\n", (195550, 195568), False, 'import pulumi\n'), ((197774, 197805), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""dnsZoneId"""'}), "(name='dnsZoneId')\n", (197787, 197805), False, 'import pulumi\n'), ((200010, 200041), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""dnsZoneId"""'}), "(name='dnsZoneId')\n", (200023, 200041), False, 'import pulumi\n'), ((202087, 202125), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""placementGroupId"""'}), "(name='placementGroupId')\n", (202100, 202125), False, 'import pulumi\n'), ((203966, 204000), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""coreFraction"""'}), "(name='coreFraction')\n", (203979, 204000), False, 'import pulumi\n'), ((206770, 206802), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""deviceName"""'}), "(name='deviceName')\n", (206783, 206802), False, 'import pulumi\n'), ((207168, 207196), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""diskId"""'}), "(name='diskId')\n", (207181, 207196), False, 'import pulumi\n'), ((207516, 207554), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""initializeParams"""'}), "(name='initializeParams')\n", (207529, 207554), False, 'import pulumi\n'), ((210153, 210182), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""imageId"""'}), "(name='imageId')\n", (210166, 210182), False, 'import pulumi\n'), ((210823, 210855), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""snapshotId"""'}), "(name='snapshotId')\n", (210836, 210855), False, 'import pulumi\n'), ((213476, 213523), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxOpeningTrafficDuration"""'}), "(name='maxOpeningTrafficDuration')\n", (213489, 213523), False, 'import pulumi\n'), ((214076, 214111), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""statusMessage"""'}), "(name='statusMessage')\n", (214089, 214111), False, 'import pulumi\n'), ((214455, 214499), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""targetGroupDescription"""'}), "(name='targetGroupDescription')\n", (214468, 214499), False, 'import pulumi\n'), ((214892, 214927), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""targetGroupId"""'}), "(name='targetGroupId')\n", (214905, 214927), False, 'import pulumi\n'), ((215208, 215247), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""targetGroupLabels"""'}), "(name='targetGroupLabels')\n", (215221, 215247), False, 'import pulumi\n'), ((215668, 215705), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""targetGroupName"""'}), "(name='targetGroupName')\n", (215681, 215705), False, 'import pulumi\n'), ((217380, 217411), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""autoScale"""'}), "(name='autoScale')\n", (217393, 217411), False, 'import pulumi\n'), ((217867, 217899), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""fixedScale"""'}), "(name='fixedScale')\n", (217880, 217899), False, 'import pulumi\n'), ((218363, 218398), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""testAutoScale"""'}), "(name='testAutoScale')\n", (218376, 218398), False, 'import pulumi\n'), ((221933, 221966), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""initialSize"""'}), "(name='initialSize')\n", (221946, 221966), False, 'import pulumi\n'), ((222299, 222340), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""measurementDuration"""'}), "(name='measurementDuration')\n", (222312, 222340), False, 'import pulumi\n'), ((222908, 222950), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""cpuUtilizationTarget"""'}), "(name='cpuUtilizationTarget')\n", (222921, 222950), False, 'import pulumi\n'), ((223325, 223358), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""customRules"""'}), "(name='customRules')\n", (223338, 223358), False, 'import pulumi\n'), ((223869, 223898), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxSize"""'}), "(name='maxSize')\n", (223882, 223898), False, 'import pulumi\n'), ((224229, 224262), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""minZoneSize"""'}), "(name='minZoneSize')\n", (224242, 224262), False, 'import pulumi\n'), ((224635, 224678), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""stabilizationDuration"""'}), "(name='stabilizationDuration')\n", (224648, 224678), False, 'import pulumi\n'), ((225304, 225340), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""warmupDuration"""'}), "(name='warmupDuration')\n", (225317, 225340), False, 'import pulumi\n'), ((227920, 227952), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""metricName"""'}), "(name='metricName')\n", (227933, 227952), False, 'import pulumi\n'), ((228245, 228277), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""metricType"""'}), "(name='metricType')\n", (228258, 228277), False, 'import pulumi\n'), ((228585, 228615), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""ruleType"""'}), "(name='ruleType')\n", (228598, 228615), False, 'import pulumi\n'), ((229627, 229657), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""folderId"""'}), "(name='folderId')\n", (229640, 229657), False, 'import pulumi\n'), ((234360, 234393), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""initialSize"""'}), "(name='initialSize')\n", (234373, 234393), False, 'import pulumi\n'), ((234726, 234767), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""measurementDuration"""'}), "(name='measurementDuration')\n", (234739, 234767), False, 'import pulumi\n'), ((235335, 235377), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""cpuUtilizationTarget"""'}), "(name='cpuUtilizationTarget')\n", (235348, 235377), False, 'import pulumi\n'), ((235752, 235785), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""customRules"""'}), "(name='customRules')\n", (235765, 235785), False, 'import pulumi\n'), ((236304, 236333), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxSize"""'}), "(name='maxSize')\n", (236317, 236333), False, 'import pulumi\n'), ((236664, 236697), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""minZoneSize"""'}), "(name='minZoneSize')\n", (236677, 236697), False, 'import pulumi\n'), ((237070, 237113), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""stabilizationDuration"""'}), "(name='stabilizationDuration')\n", (237083, 237113), False, 'import pulumi\n'), ((237739, 237775), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""warmupDuration"""'}), "(name='warmupDuration')\n", (237752, 237775), False, 'import pulumi\n'), ((240359, 240391), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""metricName"""'}), "(name='metricName')\n", (240372, 240391), False, 'import pulumi\n'), ((240684, 240716), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""metricType"""'}), "(name='metricType')\n", (240697, 240716), False, 'import pulumi\n'), ((241024, 241054), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""ruleType"""'}), "(name='ruleType')\n", (241037, 241054), False, 'import pulumi\n'), ((242066, 242096), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""folderId"""'}), "(name='folderId')\n", (242079, 242096), False, 'import pulumi\n'), ((247539, 247569), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""subnetId"""'}), "(name='subnetId')\n", (247552, 247569), False, 'import pulumi\n'), ((247970, 248002), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""dnsRecords"""'}), "(name='dnsRecords')\n", (247983, 248002), False, 'import pulumi\n'), ((248762, 248793), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""ipAddress"""'}), "(name='ipAddress')\n", (248775, 248793), False, 'import pulumi\n'), ((249969, 250002), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""ipv6Address"""'}), "(name='ipv6Address')\n", (249982, 250002), False, 'import pulumi\n'), ((250352, 250388), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""ipv6DnsRecords"""'}), "(name='ipv6DnsRecords')\n", (250365, 250388), False, 'import pulumi\n'), ((250937, 250969), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""macAddress"""'}), "(name='macAddress')\n", (250950, 250969), False, 'import pulumi\n'), ((251571, 251606), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""natDnsRecords"""'}), "(name='natDnsRecords')\n", (251584, 251606), False, 'import pulumi\n'), ((252152, 252186), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""natIpAddress"""'}), "(name='natIpAddress')\n", (252165, 252186), False, 'import pulumi\n'), ((252613, 252647), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""natIpVersion"""'}), "(name='natIpVersion')\n", (252626, 252647), False, 'import pulumi\n'), ((252923, 252961), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""securityGroupIds"""'}), "(name='securityGroupIds')\n", (252936, 252961), False, 'import pulumi\n'), ((254643, 254674), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""dnsZoneId"""'}), "(name='dnsZoneId')\n", (254656, 254674), False, 'import pulumi\n'), ((256889, 256920), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""dnsZoneId"""'}), "(name='dnsZoneId')\n", (256902, 256920), False, 'import pulumi\n'), ((259134, 259165), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""dnsZoneId"""'}), "(name='dnsZoneId')\n", (259147, 259165), False, 'import pulumi\n'), ((260477, 260515), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""placementGroupId"""'}), "(name='placementGroupId')\n", (260490, 260515), False, 'import pulumi\n'), ((262304, 262338), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""coreFraction"""'}), "(name='coreFraction')\n", (262317, 262338), False, 'import pulumi\n'), ((264890, 264918), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""diskId"""'}), "(name='diskId')\n", (264903, 264918), False, 'import pulumi\n'), ((265220, 265252), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""autoDelete"""'}), "(name='autoDelete')\n", (265233, 265252), False, 'import pulumi\n'), ((265646, 265678), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""deviceName"""'}), "(name='deviceName')\n", (265659, 265678), False, 'import pulumi\n'), ((267472, 267509), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""subclusterSpecs"""'}), "(name='subclusterSpecs')\n", (267485, 267509), False, 'import pulumi\n'), ((268441, 268472), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""versionId"""'}), "(name='versionId')\n", (268454, 268472), False, 'import pulumi\n'), ((270821, 270856), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""sshPublicKeys"""'}), "(name='sshPublicKeys')\n", (270834, 270856), False, 'import pulumi\n'), ((273328, 273360), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""hostsCount"""'}), "(name='hostsCount')\n", (273341, 273360), False, 'import pulumi\n'), ((274744, 274774), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""subnetId"""'}), "(name='subnetId')\n", (274757, 274774), False, 'import pulumi\n'), ((275168, 275207), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""autoscalingConfig"""'}), "(name='autoscalingConfig')\n", (275181, 275207), False, 'import pulumi\n'), ((278809, 278844), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxHostsCount"""'}), "(name='maxHostsCount')\n", (278822, 278844), False, 'import pulumi\n'), ((279189, 279231), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""cpuUtilizationTarget"""'}), "(name='cpuUtilizationTarget')\n", (279202, 279231), False, 'import pulumi\n'), ((279723, 279764), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""decommissionTimeout"""'}), "(name='decommissionTimeout')\n", (279736, 279764), False, 'import pulumi\n'), ((280175, 280216), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""measurementDuration"""'}), "(name='measurementDuration')\n", (280188, 280216), False, 'import pulumi\n'), ((281227, 281270), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""stabilizationDuration"""'}), "(name='stabilizationDuration')\n", (281240, 281270), False, 'import pulumi\n'), ((281854, 281890), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""warmupDuration"""'}), "(name='warmupDuration')\n", (281867, 281890), False, 'import pulumi\n'), ((283344, 283374), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""diskSize"""'}), "(name='diskSize')\n", (283357, 283374), False, 'import pulumi\n'), ((283694, 283732), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""resourcePresetId"""'}), "(name='resourcePresetId')\n", (283707, 283732), False, 'import pulumi\n'), ((284232, 284264), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""diskTypeId"""'}), "(name='diskTypeId')\n", (284245, 284264), False, 'import pulumi\n'), ((284830, 284863), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""zipFilename"""'}), "(name='zipFilename')\n", (284843, 284863), False, 'import pulumi\n'), ((285543, 285575), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""bucketName"""'}), "(name='bucketName')\n", (285556, 285575), False, 'import pulumi\n'), ((285816, 285848), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""objectName"""'}), "(name='objectName')\n", (285829, 285848), False, 'import pulumi\n'), ((287119, 287159), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""zoneInstancesLimit"""'}), "(name='zoneInstancesLimit')\n", (287132, 287159), False, 'import pulumi\n'), ((287465, 287504), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""zoneRequestsLimit"""'}), "(name='zoneRequestsLimit')\n", (287478, 287504), False, 'import pulumi\n'), ((288111, 288140), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""queueId"""'}), "(name='queueId')\n", (288124, 288140), False, 'import pulumi\n'), ((288366, 288404), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""serviceAccountId"""'}), "(name='serviceAccountId')\n", (288379, 288404), False, 'import pulumi\n'), ((289727, 289762), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""retryAttempts"""'}), "(name='retryAttempts')\n", (289740, 289762), False, 'import pulumi\n'), ((290038, 290073), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""retryInterval"""'}), "(name='retryInterval')\n", (290051, 290073), False, 'import pulumi\n'), ((290349, 290387), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""serviceAccountId"""'}), "(name='serviceAccountId')\n", (290362, 290387), False, 'import pulumi\n'), ((291399, 291431), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""registryId"""'}), "(name='registryId')\n", (291412, 291431), False, 'import pulumi\n'), ((291672, 291702), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""deviceId"""'}), "(name='deviceId')\n", (291685, 291702), False, 'import pulumi\n'), ((292688, 292721), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""batchCutoff"""'}), "(name='batchCutoff')\n", (292701, 292721), False, 'import pulumi\n'), ((292967, 293000), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""logGroupIds"""'}), "(name='logGroupIds')\n", (292980, 293000), False, 'import pulumi\n'), ((293299, 293330), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""batchSize"""'}), "(name='batchSize')\n", (293312, 293330), False, 'import pulumi\n'), ((294424, 294457), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""batchCutoff"""'}), "(name='batchCutoff')\n", (294437, 294457), False, 'import pulumi\n'), ((294703, 294732), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""groupId"""'}), "(name='groupId')\n", (294716, 294732), False, 'import pulumi\n'), ((295235, 295268), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""resourceIds"""'}), "(name='resourceIds')\n", (295248, 295268), False, 'import pulumi\n'), ((295562, 295597), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""resourceTypes"""'}), "(name='resourceTypes')\n", (295575, 295597), False, 'import pulumi\n'), ((295901, 295932), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""batchSize"""'}), "(name='batchSize')\n", (295914, 295932), False, 'import pulumi\n'), ((296962, 296995), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""batchCutoff"""'}), "(name='batchCutoff')\n", (296975, 296995), False, 'import pulumi\n'), ((297241, 297270), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""queueId"""'}), "(name='queueId')\n", (297254, 297270), False, 'import pulumi\n'), ((297496, 297534), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""serviceAccountId"""'}), "(name='serviceAccountId')\n", (297509, 297534), False, 'import pulumi\n'), ((297810, 297841), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""batchSize"""'}), "(name='batchSize')\n", (297823, 297841), False, 'import pulumi\n'), ((298097, 298136), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""visibilityTimeout"""'}), "(name='visibilityTimeout')\n", (298110, 298136), False, 'import pulumi\n'), ((299342, 299372), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""bucketId"""'}), "(name='bucketId')\n", (299355, 299372), False, 'import pulumi\n'), ((301056, 301092), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""cronExpression"""'}), "(name='cronExpression')\n", (301069, 301092), False, 'import pulumi\n'), ((301665, 301692), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""keyId"""'}), "(name='keyId')\n", (301678, 301692), False, 'import pulumi\n'), ((306583, 306625), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""clusterCaCertificate"""'}), "(name='clusterCaCertificate')\n", (306596, 306625), False, 'import pulumi\n'), ((307069, 307108), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""externalV4Address"""'}), "(name='externalV4Address')\n", (307082, 307108), False, 'import pulumi\n'), ((307517, 307557), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""externalV4Endpoint"""'}), "(name='externalV4Endpoint')\n", (307530, 307557), False, 'import pulumi\n'), ((308016, 308055), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""internalV4Address"""'}), "(name='internalV4Address')\n", (308029, 308055), False, 'import pulumi\n'), ((308464, 308504), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""internalV4Endpoint"""'}), "(name='internalV4Endpoint')\n", (308477, 308504), False, 'import pulumi\n'), ((308934, 308973), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maintenancePolicy"""'}), "(name='maintenancePolicy')\n", (308947, 308973), False, 'import pulumi\n'), ((309777, 309807), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""publicIp"""'}), "(name='publicIp')\n", (309790, 309807), False, 'import pulumi\n'), ((310665, 310703), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""securityGroupIds"""'}), "(name='securityGroupIds')\n", (310678, 310703), False, 'import pulumi\n'), ((311518, 311551), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""versionInfo"""'}), "(name='versionInfo')\n", (311531, 311551), False, 'import pulumi\n'), ((313626, 313659), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""autoUpgrade"""'}), "(name='autoUpgrade')\n", (313639, 313659), False, 'import pulumi\n'), ((314056, 314096), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maintenanceWindows"""'}), "(name='maintenanceWindows')\n", (314069, 314096), False, 'import pulumi\n'), ((315631, 315662), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""startTime"""'}), "(name='startTime')\n", (315644, 315662), False, 'import pulumi\n'), ((318372, 318402), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""subnetId"""'}), "(name='subnetId')\n", (318385, 318402), False, 'import pulumi\n'), ((320740, 320776), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""currentVersion"""'}), "(name='currentVersion')\n", (320753, 320776), False, 'import pulumi\n'), ((321142, 321184), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""newRevisionAvailable"""'}), "(name='newRevisionAvailable')\n", (321155, 321184), False, 'import pulumi\n'), ((321773, 321813), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""newRevisionSummary"""'}), "(name='newRevisionSummary')\n", (321786, 321813), False, 'import pulumi\n'), ((322295, 322334), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""versionDeprecated"""'}), "(name='versionDeprecated')\n", (322308, 322334), False, 'import pulumi\n'), ((323349, 323379), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""subnetId"""'}), "(name='subnetId')\n", (323362, 323379), False, 'import pulumi\n'), ((327209, 327239), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""subnetId"""'}), "(name='subnetId')\n", (327222, 327239), False, 'import pulumi\n'), ((328585, 328619), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxExpansion"""'}), "(name='maxExpansion')\n", (328598, 328619), False, 'import pulumi\n'), ((329017, 329053), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxUnavailable"""'}), "(name='maxUnavailable')\n", (329030, 329053), False, 'import pulumi\n'), ((333419, 333449), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""bootDisk"""'}), "(name='bootDisk')\n", (333432, 333449), False, 'import pulumi\n'), ((335052, 335097), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""networkAccelerationType"""'}), "(name='networkAccelerationType')\n", (335065, 335097), False, 'import pulumi\n'), ((335534, 335573), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""networkInterfaces"""'}), "(name='networkInterfaces')\n", (335547, 335573), False, 'import pulumi\n'), ((336168, 336205), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""placementPolicy"""'}), "(name='placementPolicy')\n", (336181, 336205), False, 'import pulumi\n'), ((336700, 336732), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""platformId"""'}), "(name='platformId')\n", (336713, 336732), False, 'import pulumi\n'), ((337467, 337505), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""schedulingPolicy"""'}), "(name='schedulingPolicy')\n", (337480, 337505), False, 'import pulumi\n'), ((340642, 340673), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""subnetIds"""'}), "(name='subnetIds')\n", (340655, 340673), False, 'import pulumi\n'), ((342090, 342128), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""securityGroupIds"""'}), "(name='securityGroupIds')\n", (342103, 342128), False, 'import pulumi\n'), ((342932, 342970), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""placementGroupId"""'}), "(name='placementGroupId')\n", (342945, 342970), False, 'import pulumi\n'), ((344044, 344078), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""coreFraction"""'}), "(name='coreFraction')\n", (344057, 344078), False, 'import pulumi\n'), ((347082, 347114), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""autoRepair"""'}), "(name='autoRepair')\n", (347095, 347114), False, 'import pulumi\n'), ((347499, 347532), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""autoUpgrade"""'}), "(name='autoUpgrade')\n", (347512, 347532), False, 'import pulumi\n'), ((347922, 347962), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maintenanceWindows"""'}), "(name='maintenanceWindows')\n", (347935, 347962), False, 'import pulumi\n'), ((349263, 349294), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""startTime"""'}), "(name='startTime')\n", (349276, 349294), False, 'import pulumi\n'), ((350621, 350652), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""autoScale"""'}), "(name='autoScale')\n", (350634, 350652), False, 'import pulumi\n'), ((351102, 351134), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""fixedScale"""'}), "(name='fixedScale')\n", (351115, 351134), False, 'import pulumi\n'), ((355510, 355546), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""currentVersion"""'}), "(name='currentVersion')\n", (355523, 355546), False, 'import pulumi\n'), ((355912, 355954), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""newRevisionAvailable"""'}), "(name='newRevisionAvailable')\n", (355925, 355954), False, 'import pulumi\n'), ((356546, 356586), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""newRevisionSummary"""'}), "(name='newRevisionSummary')\n", (356559, 356586), False, 'import pulumi\n'), ((357068, 357107), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""versionDeprecated"""'}), "(name='versionDeprecated')\n", (357081, 357107), False, 'import pulumi\n'), ((358764, 358799), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""targetGroupId"""'}), "(name='targetGroupId')\n", (358777, 358799), False, 'import pulumi\n'), ((361835, 361873), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""healthyThreshold"""'}), "(name='healthyThreshold')\n", (361848, 361873), False, 'import pulumi\n'), ((362293, 362326), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""httpOptions"""'}), "(name='httpOptions')\n", (362306, 362326), False, 'import pulumi\n'), ((363173, 363205), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""tcpOptions"""'}), "(name='tcpOptions')\n", (363186, 363205), False, 'import pulumi\n'), ((364066, 364106), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""unhealthyThreshold"""'}), "(name='unhealthyThreshold')\n", (364079, 364106), False, 'import pulumi\n'), ((368887, 368928), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""externalAddressSpec"""'}), "(name='externalAddressSpec')\n", (368900, 368928), False, 'import pulumi\n'), ((369443, 369484), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""internalAddressSpec"""'}), "(name='internalAddressSpec')\n", (369456, 369484), False, 'import pulumi\n'), ((370356, 370388), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""targetPort"""'}), "(name='targetPort')\n", (370369, 370388), False, 'import pulumi\n'), ((371974, 372005), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""ipVersion"""'}), "(name='ipVersion')\n", (371987, 372005), False, 'import pulumi\n'), ((373416, 373446), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""subnetId"""'}), "(name='subnetId')\n", (373429, 373446), False, 'import pulumi\n'), ((374201, 374232), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""ipVersion"""'}), "(name='ipVersion')\n", (374214, 374232), False, 'import pulumi\n'), ((375515, 375545), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""subnetId"""'}), "(name='subnetId')\n", (375528, 375545), False, 'import pulumi\n'), ((377119, 377149), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""dataLens"""'}), "(name='dataLens')\n", (377132, 377149), False, 'import pulumi\n'), ((378213, 378241), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""webSql"""'}), "(name='webSql')\n", (378226, 378241), False, 'import pulumi\n'), ((389775, 389815), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""backgroundPoolSize"""'}), "(name='backgroundPoolSize')\n", (389788, 389815), False, 'import pulumi\n'), ((390121, 390169), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""backgroundSchedulePoolSize"""'}), "(name='backgroundSchedulePoolSize')\n", (390134, 390169), False, 'import pulumi\n'), ((391046, 391078), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""geobaseUri"""'}), "(name='geobaseUri')\n", (391059, 391078), False, 'import pulumi\n'), ((391339, 391376), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""graphiteRollups"""'}), "(name='graphiteRollups')\n", (391352, 391376), False, 'import pulumi\n'), ((392345, 392378), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""kafkaTopics"""'}), "(name='kafkaTopics')\n", (392358, 392378), False, 'import pulumi\n'), ((392895, 392933), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""keepAliveTimeout"""'}), "(name='keepAliveTimeout')\n", (392908, 392933), False, 'import pulumi\n'), ((393229, 393259), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""logLevel"""'}), "(name='logLevel')\n", (393242, 393259), False, 'import pulumi\n'), ((393510, 393545), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""markCacheSize"""'}), "(name='markCacheSize')\n", (393523, 393545), False, 'import pulumi\n'), ((393826, 393868), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxConcurrentQueries"""'}), "(name='maxConcurrentQueries')\n", (393839, 393868), False, 'import pulumi\n'), ((394184, 394220), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxConnections"""'}), "(name='maxConnections')\n", (394197, 394220), False, 'import pulumi\n'), ((394501, 394545), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxPartitionSizeToDrop"""'}), "(name='maxPartitionSizeToDrop')\n", (394514, 394545), False, 'import pulumi\n'), ((394881, 394921), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxTableSizeToDrop"""'}), "(name='maxTableSizeToDrop')\n", (394894, 394921), False, 'import pulumi\n'), ((395237, 395268), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""mergeTree"""'}), "(name='mergeTree')\n", (395250, 395268), False, 'import pulumi\n'), ((395719, 395757), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""metricLogEnabled"""'}), "(name='metricLogEnabled')\n", (395732, 395757), False, 'import pulumi\n'), ((396055, 396099), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""metricLogRetentionSize"""'}), "(name='metricLogRetentionSize')\n", (396068, 396099), False, 'import pulumi\n'), ((396430, 396474), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""metricLogRetentionTime"""'}), "(name='metricLogRetentionTime')\n", (396443, 396474), False, 'import pulumi\n'), ((396805, 396847), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""partLogRetentionSize"""'}), "(name='partLogRetentionSize')\n", (396818, 396847), False, 'import pulumi\n'), ((397168, 397210), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""partLogRetentionTime"""'}), "(name='partLogRetentionTime')\n", (397181, 397210), False, 'import pulumi\n'), ((397531, 397574), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""queryLogRetentionSize"""'}), "(name='queryLogRetentionSize')\n", (397544, 397574), False, 'import pulumi\n'), ((397900, 397943), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""queryLogRetentionTime"""'}), "(name='queryLogRetentionTime')\n", (397913, 397943), False, 'import pulumi\n'), ((398269, 398312), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""queryThreadLogEnabled"""'}), "(name='queryThreadLogEnabled')\n", (398282, 398312), False, 'import pulumi\n'), ((398640, 398689), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""queryThreadLogRetentionSize"""'}), "(name='queryThreadLogRetentionSize')\n", (398653, 398689), False, 'import pulumi\n'), ((399050, 399099), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""queryThreadLogRetentionTime"""'}), "(name='queryThreadLogRetentionTime')\n", (399063, 399099), False, 'import pulumi\n'), ((399915, 399951), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""textLogEnabled"""'}), "(name='textLogEnabled')\n", (399928, 399951), False, 'import pulumi\n'), ((400239, 400273), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""textLogLevel"""'}), "(name='textLogLevel')\n", (400252, 400273), False, 'import pulumi\n'), ((400549, 400591), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""textLogRetentionSize"""'}), "(name='textLogRetentionSize')\n", (400562, 400591), False, 'import pulumi\n'), ((400912, 400954), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""textLogRetentionTime"""'}), "(name='textLogRetentionTime')\n", (400925, 400954), False, 'import pulumi\n'), ((401534, 401571), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""traceLogEnabled"""'}), "(name='traceLogEnabled')\n", (401547, 401571), False, 'import pulumi\n'), ((401864, 401907), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""traceLogRetentionSize"""'}), "(name='traceLogRetentionSize')\n", (401877, 401907), False, 'import pulumi\n'), ((402233, 402276), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""traceLogRetentionTime"""'}), "(name='traceLogRetentionTime')\n", (402246, 402276), False, 'import pulumi\n'), ((402602, 402645), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""uncompressedCacheSize"""'}), "(name='uncompressedCacheSize')\n", (402615, 402645), False, 'import pulumi\n'), ((404349, 404382), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""minPartSize"""'}), "(name='minPartSize')\n", (404362, 404382), False, 'import pulumi\n'), ((404840, 404878), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""minPartSizeRatio"""'}), "(name='minPartSizeRatio')\n", (404853, 404878), False, 'import pulumi\n'), ((411225, 411260), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""saslMechanism"""'}), "(name='saslMechanism')\n", (411238, 411260), False, 'import pulumi\n'), ((411613, 411647), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""saslPassword"""'}), "(name='saslPassword')\n", (411626, 411647), False, 'import pulumi\n'), ((411981, 412015), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""saslUsername"""'}), "(name='saslUsername')\n", (411994, 412015), False, 'import pulumi\n'), ((412344, 412382), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""securityProtocol"""'}), "(name='securityProtocol')\n", (412357, 412382), False, 'import pulumi\n'), ((415341, 415376), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""saslMechanism"""'}), "(name='saslMechanism')\n", (415354, 415376), False, 'import pulumi\n'), ((415729, 415763), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""saslPassword"""'}), "(name='saslPassword')\n", (415742, 415763), False, 'import pulumi\n'), ((416097, 416131), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""saslUsername"""'}), "(name='saslUsername')\n", (416110, 416131), False, 'import pulumi\n'), ((416460, 416498), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""securityProtocol"""'}), "(name='securityProtocol')\n", (416473, 416498), False, 'import pulumi\n'), ((420415, 420468), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxBytesToMergeAtMinSpaceInPool"""'}), "(name='maxBytesToMergeAtMinSpaceInPool')\n", (420428, 420468), False, 'import pulumi\n'), ((421052, 421100), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxReplicatedMergesInQueue"""'}), "(name='maxReplicatedMergesInQueue')\n", (421065, 421100), False, 'import pulumi\n'), ((421613, 421681), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""numberOfFreeEntriesInPoolToLowerMaxSizeOfMerge"""'}), "(name='numberOfFreeEntriesInPoolToLowerMaxSizeOfMerge')\n", (421626, 421681), False, 'import pulumi\n'), ((422519, 422559), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""partsToDelayInsert"""'}), "(name='partsToDelayInsert')\n", (422532, 422559), False, 'import pulumi\n'), ((423065, 423105), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""partsToThrowInsert"""'}), "(name='partsToThrowInsert')\n", (423078, 423105), False, 'import pulumi\n'), ((423593, 423644), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""replicatedDeduplicationWindow"""'}), "(name='replicatedDeduplicationWindow')\n", (423606, 423644), False, 'import pulumi\n'), ((424157, 424215), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""replicatedDeduplicationWindowSeconds"""'}), "(name='replicatedDeduplicationWindowSeconds')\n", (424170, 424215), False, 'import pulumi\n'), ((426745, 426775), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""diskSize"""'}), "(name='diskSize')\n", (426758, 426775), False, 'import pulumi\n'), ((427105, 427137), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""diskTypeId"""'}), "(name='diskTypeId')\n", (427118, 427137), False, 'import pulumi\n'), ((427585, 427623), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""resourcePresetId"""'}), "(name='resourcePresetId')\n", (427598, 427623), False, 'import pulumi\n'), ((433533, 433569), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""assignPublicIp"""'}), "(name='assignPublicIp')\n", (433546, 433569), False, 'import pulumi\n'), ((434304, 434335), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""shardName"""'}), "(name='shardName')\n", (434317, 434335), False, 'import pulumi\n'), ((434672, 434702), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""subnetId"""'}), "(name='subnetId')\n", (434685, 434702), False, 'import pulumi\n'), ((439941, 439973), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""shardNames"""'}), "(name='shardNames')\n", (439954, 439973), False, 'import pulumi\n'), ((444467, 444501), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""databaseName"""'}), "(name='databaseName')\n", (444480, 444501), False, 'import pulumi\n'), ((446468, 446506), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""intervalDuration"""'}), "(name='intervalDuration')\n", (446481, 446506), False, 'import pulumi\n'), ((447182, 447217), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""executionTime"""'}), "(name='executionTime')\n", (447195, 447217), False, 'import pulumi\n'), ((447901, 447931), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""readRows"""'}), "(name='readRows')\n", (447914, 447931), False, 'import pulumi\n'), ((448309, 448341), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""resultRows"""'}), "(name='resultRows')\n", (448322, 448341), False, 'import pulumi\n'), ((481080, 481119), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""addHttpCorsHeader"""'}), "(name='addHttpCorsHeader')\n", (481093, 481119), False, 'import pulumi\n'), ((481499, 481529), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""allowDdl"""'}), "(name='allowDdl')\n", (481512, 481529), False, 'import pulumi\n'), ((482163, 482203), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""compileExpressions"""'}), "(name='compileExpressions')\n", (482176, 482203), False, 'import pulumi\n'), ((482570, 482606), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""connectTimeout"""'}), "(name='connectTimeout')\n", (482583, 482606), False, 'import pulumi\n'), ((483005, 483054), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""countDistinctImplementation"""'}), "(name='countDistinctImplementation')\n", (483018, 483054), False, 'import pulumi\n'), ((483538, 483580), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""distinctOverflowMode"""'}), "(name='distinctOverflowMode')\n", (483551, 483580), False, 'import pulumi\n'), ((483993, 484052), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""distributedAggregationMemoryEfficient"""'}), "(name='distributedAggregationMemoryEfficient')\n", (484006, 484052), False, 'import pulumi\n'), ((484542, 484589), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""distributedDdlTaskTimeout"""'}), "(name='distributedDdlTaskTimeout')\n", (484555, 484589), False, 'import pulumi\n'), ((485009, 485053), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""distributedProductMode"""'}), "(name='distributedProductMode')\n", (485022, 485053), False, 'import pulumi\n'), ((485460, 485517), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""emptyResultForAggregationByEmptySet"""'}), "(name='emptyResultForAggregationByEmptySet')\n", (485473, 485517), False, 'import pulumi\n'), ((485993, 486036), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""enableHttpCompression"""'}), "(name='enableHttpCompression')\n", (486006, 486036), False, 'import pulumi\n'), ((486464, 486530), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""fallbackToStaleReplicasForDistributedQueries"""'}), "(name='fallbackToStaleReplicasForDistributedQueries')\n", (486477, 486530), False, 'import pulumi\n'), ((487095, 487133), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""forceIndexByDate"""'}), "(name='forceIndexByDate')\n", (487108, 487133), False, 'import pulumi\n'), ((487529, 487566), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""forcePrimaryKey"""'}), "(name='forcePrimaryKey')\n", (487542, 487566), False, 'import pulumi\n'), ((487964, 488005), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""groupByOverflowMode"""'}), "(name='groupByOverflowMode')\n", (487977, 488005), False, 'import pulumi\n'), ((488423, 488469), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""groupByTwoLevelThreshold"""'}), "(name='groupByTwoLevelThreshold')\n", (488436, 488469), False, 'import pulumi\n'), ((488942, 488993), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""groupByTwoLevelThresholdBytes"""'}), "(name='groupByTwoLevelThresholdBytes')\n", (488955, 488993), False, 'import pulumi\n'), ((489497, 489540), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""httpConnectionTimeout"""'}), "(name='httpConnectionTimeout')\n", (489510, 489540), False, 'import pulumi\n'), ((489938, 489987), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""httpHeadersProgressInterval"""'}), "(name='httpHeadersProgressInterval')\n", (489951, 489987), False, 'import pulumi\n'), ((490479, 490519), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""httpReceiveTimeout"""'}), "(name='httpReceiveTimeout')\n", (490492, 490519), False, 'import pulumi\n'), ((490902, 490939), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""httpSendTimeout"""'}), "(name='httpSendTimeout')\n", (490915, 490939), False, 'import pulumi\n'), ((491307, 491364), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""inputFormatDefaultsForOmittedFields"""'}), "(name='inputFormatDefaultsForOmittedFields')\n", (491320, 491364), False, 'import pulumi\n'), ((491919, 491978), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""inputFormatValuesInterpretExpressions"""'}), "(name='inputFormatValuesInterpretExpressions')\n", (491932, 491978), False, 'import pulumi\n'), ((492511, 492545), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""insertQuorum"""'}), "(name='insertQuorum')\n", (492524, 492545), False, 'import pulumi\n'), ((492875, 492916), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""insertQuorumTimeout"""'}), "(name='insertQuorumTimeout')\n", (492888, 492916), False, 'import pulumi\n'), ((493302, 493340), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""joinOverflowMode"""'}), "(name='joinOverflowMode')\n", (493315, 493340), False, 'import pulumi\n'), ((493721, 493755), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""joinUseNulls"""'}), "(name='joinUseNulls')\n", (493734, 493755), False, 'import pulumi\n'), ((494200, 494249), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""joinedSubqueryRequiresAlias"""'}), "(name='joinedSubqueryRequiresAlias')\n", (494213, 494249), False, 'import pulumi\n'), ((494735, 494790), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""lowCardinalityAllowInNativeFormat"""'}), "(name='lowCardinalityAllowInNativeFormat')\n", (494748, 494790), False, 'import pulumi\n'), ((495299, 495332), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxAstDepth"""'}), "(name='maxAstDepth')\n", (495312, 495332), False, 'import pulumi\n'), ((495671, 495707), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxAstElements"""'}), "(name='maxAstElements')\n", (495684, 495707), False, 'import pulumi\n'), ((496064, 496098), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxBlockSize"""'}), "(name='maxBlockSize')\n", (496077, 496098), False, 'import pulumi\n'), ((496492, 496543), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxBytesBeforeExternalGroupBy"""'}), "(name='maxBytesBeforeExternalGroupBy')\n", (496505, 496543), False, 'import pulumi\n'), ((497023, 497071), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxBytesBeforeExternalSort"""'}), "(name='maxBytesBeforeExternalSort')\n", (497036, 497071), False, 'import pulumi\n'), ((497602, 497642), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxBytesInDistinct"""'}), "(name='maxBytesInDistinct')\n", (497615, 497642), False, 'import pulumi\n'), ((498075, 498111), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxBytesInJoin"""'}), "(name='maxBytesInJoin')\n", (498088, 498111), False, 'import pulumi\n'), ((498494, 498529), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxBytesInSet"""'}), "(name='maxBytesInSet')\n", (498507, 498529), False, 'import pulumi\n'), ((498935, 498971), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxBytesToRead"""'}), "(name='maxBytesToRead')\n", (498948, 498971), False, 'import pulumi\n'), ((499401, 499437), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxBytesToSort"""'}), "(name='maxBytesToSort')\n", (499414, 499437), False, 'import pulumi\n'), ((499858, 499898), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxBytesToTransfer"""'}), "(name='maxBytesToTransfer')\n", (499871, 499898), False, 'import pulumi\n'), ((500386, 500424), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxColumnsToRead"""'}), "(name='maxColumnsToRead')\n", (500399, 500424), False, 'import pulumi\n'), ((500843, 500881), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxExecutionTime"""'}), "(name='maxExecutionTime')\n", (500856, 500881), False, 'import pulumi\n'), ((501266, 501310), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxExpandedAstElements"""'}), "(name='maxExpandedAstElements')\n", (501279, 501310), False, 'import pulumi\n'), ((501742, 501782), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxInsertBlockSize"""'}), "(name='maxInsertBlockSize')\n", (501755, 501782), False, 'import pulumi\n'), ((502201, 502237), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxMemoryUsage"""'}), "(name='maxMemoryUsage')\n", (502214, 502237), False, 'import pulumi\n'), ((502641, 502684), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxMemoryUsageForUser"""'}), "(name='maxMemoryUsageForUser')\n", (502654, 502684), False, 'import pulumi\n'), ((503143, 503184), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxNetworkBandwidth"""'}), "(name='maxNetworkBandwidth')\n", (503156, 503184), False, 'import pulumi\n'), ((503603, 503651), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxNetworkBandwidthForUser"""'}), "(name='maxNetworkBandwidthForUser')\n", (503616, 503651), False, 'import pulumi\n'), ((504115, 504149), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxQuerySize"""'}), "(name='maxQuerySize')\n", (504128, 504149), False, 'import pulumi\n'), ((504543, 504601), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxReplicaDelayForDistributedQueries"""'}), "(name='maxReplicaDelayForDistributedQueries')\n", (504556, 504601), False, 'import pulumi\n'), ((505095, 505131), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxResultBytes"""'}), "(name='maxResultBytes')\n", (505108, 505131), False, 'import pulumi\n'), ((505491, 505526), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxResultRows"""'}), "(name='maxResultRows')\n", (505504, 505526), False, 'import pulumi\n'), ((505880, 505919), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxRowsInDistinct"""'}), "(name='maxRowsInDistinct')\n", (505893, 505919), False, 'import pulumi\n'), ((506322, 506357), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxRowsInJoin"""'}), "(name='maxRowsInJoin')\n", (506335, 506357), False, 'import pulumi\n'), ((506734, 506768), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxRowsInSet"""'}), "(name='maxRowsInSet')\n", (506747, 506768), False, 'import pulumi\n'), ((507168, 507206), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxRowsToGroupBy"""'}), "(name='maxRowsToGroupBy')\n", (507181, 507206), False, 'import pulumi\n'), ((507621, 507656), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxRowsToRead"""'}), "(name='maxRowsToRead')\n", (507634, 507656), False, 'import pulumi\n'), ((508060, 508095), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxRowsToSort"""'}), "(name='maxRowsToSort')\n", (508073, 508095), False, 'import pulumi\n'), ((508490, 508529), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxRowsToTransfer"""'}), "(name='maxRowsToTransfer')\n", (508503, 508529), False, 'import pulumi\n'), ((508991, 509032), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxTemporaryColumns"""'}), "(name='maxTemporaryColumns')\n", (509004, 509032), False, 'import pulumi\n'), ((509514, 509563), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxTemporaryNonConstColumns"""'}), "(name='maxTemporaryNonConstColumns')\n", (509527, 509563), False, 'import pulumi\n'), ((510095, 510127), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxThreads"""'}), "(name='maxThreads')\n", (510108, 510127), False, 'import pulumi\n'), ((510527, 510576), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""mergeTreeMaxBytesToUseCache"""'}), "(name='mergeTreeMaxBytesToUseCache')\n", (510540, 510576), False, 'import pulumi\n'), ((511118, 511166), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""mergeTreeMaxRowsToUseCache"""'}), "(name='mergeTreeMaxRowsToUseCache')\n", (511131, 511166), False, 'import pulumi\n'), ((511701, 511757), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""mergeTreeMinBytesForConcurrentRead"""'}), "(name='mergeTreeMinBytesForConcurrentRead')\n", (511714, 511757), False, 'import pulumi\n'), ((512396, 512451), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""mergeTreeMinRowsForConcurrentRead"""'}), "(name='mergeTreeMinRowsForConcurrentRead')\n", (512409, 512451), False, 'import pulumi\n'), ((513087, 513130), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""minBytesToUseDirectIo"""'}), "(name='minBytesToUseDirectIo')\n", (513100, 513130), False, 'import pulumi\n'), ((513580, 513619), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""minCountToCompile"""'}), "(name='minCountToCompile')\n", (513593, 513619), False, 'import pulumi\n'), ((514044, 514093), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""minCountToCompileExpression"""'}), "(name='minCountToCompileExpression')\n", (514057, 514093), False, 'import pulumi\n'), ((514578, 514617), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""minExecutionSpeed"""'}), "(name='minExecutionSpeed')\n", (514591, 514617), False, 'import pulumi\n'), ((514994, 515038), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""minExecutionSpeedBytes"""'}), "(name='minExecutionSpeedBytes')\n", (515007, 515038), False, 'import pulumi\n'), ((515446, 515491), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""minInsertBlockSizeBytes"""'}), "(name='minInsertBlockSizeBytes')\n", (515459, 515491), False, 'import pulumi\n'), ((515965, 516009), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""minInsertBlockSizeRows"""'}), "(name='minInsertBlockSizeRows')\n", (515978, 516009), False, 'import pulumi\n'), ((516477, 516533), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""outputFormatJsonQuote64bitIntegers"""'}), "(name='outputFormatJsonQuote64bitIntegers')\n", (516490, 516533), False, 'import pulumi\n'), ((517162, 517214), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""outputFormatJsonQuoteDenormals"""'}), "(name='outputFormatJsonQuoteDenormals')\n", (517175, 517214), False, 'import pulumi\n'), ((517993, 518024), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""quotaMode"""'}), "(name='quotaMode')\n", (518006, 518024), False, 'import pulumi\n'), ((518335, 518373), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""readOverflowMode"""'}), "(name='readOverflowMode')\n", (518348, 518373), False, 'import pulumi\n'), ((519128, 519164), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""receiveTimeout"""'}), "(name='receiveTimeout')\n", (519141, 519164), False, 'import pulumi\n'), ((519563, 519615), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""replicationAlterPartitionsSync"""'}), "(name='replicationAlterPartitionsSync')\n", (519576, 519615), False, 'import pulumi\n'), ((520137, 520177), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""resultOverflowMode"""'}), "(name='resultOverflowMode')\n", (520150, 520177), False, 'import pulumi\n'), ((520570, 520619), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""selectSequentialConsistency"""'}), "(name='selectSequentialConsistency')\n", (520583, 520619), False, 'import pulumi\n'), ((521067, 521114), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""sendProgressInHttpHeaders"""'}), "(name='sendProgressInHttpHeaders')\n", (521080, 521114), False, 'import pulumi\n'), ((521595, 521628), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""sendTimeout"""'}), "(name='sendTimeout')\n", (521608, 521628), False, 'import pulumi\n'), ((522009, 522046), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""setOverflowMode"""'}), "(name='setOverflowMode')\n", (522022, 522046), False, 'import pulumi\n'), ((522435, 522478), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""skipUnavailableShards"""'}), "(name='skipUnavailableShards')\n", (522448, 522478), False, 'import pulumi\n'), ((522894, 522932), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""sortOverflowMode"""'}), "(name='sortOverflowMode')\n", (522907, 522932), False, 'import pulumi\n'), ((523316, 523357), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""timeoutOverflowMode"""'}), "(name='timeoutOverflowMode')\n", (523329, 523357), False, 'import pulumi\n'), ((523745, 523787), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""transferOverflowMode"""'}), "(name='transferOverflowMode')\n", (523758, 523787), False, 'import pulumi\n'), ((524180, 524217), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""transformNullIn"""'}), "(name='transformNullIn')\n", (524193, 524217), False, 'import pulumi\n'), ((524591, 524633), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""useUncompressedCache"""'}), "(name='useUncompressedCache')\n", (524604, 524633), False, 'import pulumi\n'), ((526966, 526996), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""diskSize"""'}), "(name='diskSize')\n", (526979, 526996), False, 'import pulumi\n'), ((527346, 527378), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""diskTypeId"""'}), "(name='diskTypeId')\n", (527359, 527378), False, 'import pulumi\n'), ((527846, 527884), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""resourcePresetId"""'}), "(name='resourcePresetId')\n", (527859, 527884), False, 'import pulumi\n'), ((530048, 530083), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""adminPassword"""'}), "(name='adminPassword')\n", (530061, 530083), False, 'import pulumi\n'), ((530413, 530443), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""dataNode"""'}), "(name='dataNode')\n", (530426, 530443), False, 'import pulumi\n'), ((531324, 531356), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""masterNode"""'}), "(name='masterNode')\n", (531337, 531356), False, 'import pulumi\n'), ((534137, 534167), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""diskSize"""'}), "(name='diskSize')\n", (534150, 534167), False, 'import pulumi\n'), ((534487, 534519), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""diskTypeId"""'}), "(name='diskTypeId')\n", (534500, 534519), False, 'import pulumi\n'), ((534841, 534879), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""resourcePresetId"""'}), "(name='resourcePresetId')\n", (534854, 534879), False, 'import pulumi\n'), ((536790, 536820), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""diskSize"""'}), "(name='diskSize')\n", (536803, 536820), False, 'import pulumi\n'), ((537140, 537172), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""diskTypeId"""'}), "(name='diskTypeId')\n", (537153, 537172), False, 'import pulumi\n'), ((537494, 537532), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""resourcePresetId"""'}), "(name='resourcePresetId')\n", (537507, 537532), False, 'import pulumi\n'), ((540533, 540569), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""assignPublicIp"""'}), "(name='assignPublicIp')\n", (540546, 540569), False, 'import pulumi\n'), ((541304, 541334), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""subnetId"""'}), "(name='subnetId')\n", (541317, 541334), False, 'import pulumi\n'), ((542386, 542416), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""dataLens"""'}), "(name='dataLens')\n", (542399, 542416), False, 'import pulumi\n'), ((542781, 542809), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""webSql"""'}), "(name='webSql')\n", (542794, 542809), False, 'import pulumi\n'), ((545124, 545160), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""assignPublicIp"""'}), "(name='assignPublicIp')\n", (545137, 545160), False, 'import pulumi\n'), ((547359, 547389), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""diskSize"""'}), "(name='diskSize')\n", (547372, 547389), False, 'import pulumi\n'), ((547620, 547652), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""diskTypeId"""'}), "(name='diskTypeId')\n", (547633, 547652), False, 'import pulumi\n'), ((547898, 547936), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""resourcePresetId"""'}), "(name='resourcePresetId')\n", (547911, 547936), False, 'import pulumi\n'), ((550294, 550324), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""diskSize"""'}), "(name='diskSize')\n", (550307, 550324), False, 'import pulumi\n'), ((550555, 550587), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""diskTypeId"""'}), "(name='diskTypeId')\n", (550568, 550587), False, 'import pulumi\n'), ((550833, 550871), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""resourcePresetId"""'}), "(name='resourcePresetId')\n", (550846, 550871), False, 'import pulumi\n'), ((554475, 554511), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""assignPublicIp"""'}), "(name='assignPublicIp')\n", (554488, 554511), False, 'import pulumi\n'), ((554924, 554958), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""brokersCount"""'}), "(name='brokersCount')\n", (554937, 554958), False, 'import pulumi\n'), ((555321, 555357), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""schemaRegistry"""'}), "(name='schemaRegistry')\n", (555334, 555357), False, 'import pulumi\n'), ((555740, 555777), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""unmanagedTopics"""'}), "(name='unmanagedTopics')\n", (555753, 555777), False, 'import pulumi\n'), ((557854, 557887), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""kafkaConfig"""'}), "(name='kafkaConfig')\n", (557867, 557887), False, 'import pulumi\n'), ((561505, 561549), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""autoCreateTopicsEnable"""'}), "(name='autoCreateTopicsEnable')\n", (561518, 561549), False, 'import pulumi\n'), ((561882, 561919), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""compressionType"""'}), "(name='compressionType')\n", (561895, 561919), False, 'import pulumi\n'), ((562205, 562251), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""defaultReplicationFactor"""'}), "(name='defaultReplicationFactor')\n", (562218, 562251), False, 'import pulumi\n'), ((562587, 562633), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""logFlushIntervalMessages"""'}), "(name='logFlushIntervalMessages')\n", (562600, 562633), False, 'import pulumi\n'), ((562974, 563014), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""logFlushIntervalMs"""'}), "(name='logFlushIntervalMs')\n", (562987, 563014), False, 'import pulumi\n'), ((563325, 563374), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""logFlushSchedulerIntervalMs"""'}), "(name='logFlushSchedulerIntervalMs')\n", (563338, 563374), False, 'import pulumi\n'), ((563735, 563771), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""logPreallocate"""'}), "(name='logPreallocate')\n", (563748, 563771), False, 'import pulumi\n'), ((564054, 564093), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""logRetentionBytes"""'}), "(name='logRetentionBytes')\n", (564067, 564093), False, 'import pulumi\n'), ((564394, 564433), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""logRetentionHours"""'}), "(name='logRetentionHours')\n", (564407, 564433), False, 'import pulumi\n'), ((564734, 564775), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""logRetentionMinutes"""'}), "(name='logRetentionMinutes')\n", (564747, 564775), False, 'import pulumi\n'), ((565086, 565122), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""logRetentionMs"""'}), "(name='logRetentionMs')\n", (565099, 565122), False, 'import pulumi\n'), ((565408, 565445), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""logSegmentBytes"""'}), "(name='logSegmentBytes')\n", (565421, 565445), False, 'import pulumi\n'), ((565736, 565771), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""numPartitions"""'}), "(name='numPartitions')\n", (565749, 565771), False, 'import pulumi\n'), ((566047, 566093), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""socketReceiveBufferBytes"""'}), "(name='socketReceiveBufferBytes')\n", (566060, 566093), False, 'import pulumi\n'), ((566434, 566477), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""socketSendBufferBytes"""'}), "(name='socketSendBufferBytes')\n", (566447, 566477), False, 'import pulumi\n'), ((567592, 567622), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""diskSize"""'}), "(name='diskSize')\n", (567605, 567622), False, 'import pulumi\n'), ((567952, 567984), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""diskTypeId"""'}), "(name='diskTypeId')\n", (567965, 567984), False, 'import pulumi\n'), ((568427, 568465), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""resourcePresetId"""'}), "(name='resourcePresetId')\n", (568440, 568465), False, 'import pulumi\n'), ((570678, 570708), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""diskSize"""'}), "(name='diskSize')\n", (570691, 570708), False, 'import pulumi\n'), ((571058, 571090), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""diskTypeId"""'}), "(name='diskTypeId')\n", (571071, 571090), False, 'import pulumi\n'), ((571553, 571591), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""resourcePresetId"""'}), "(name='resourcePresetId')\n", (571566, 571591), False, 'import pulumi\n'), ((573428, 573464), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""assignPublicIp"""'}), "(name='assignPublicIp')\n", (573441, 573464), False, 'import pulumi\n'), ((574780, 574810), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""subnetId"""'}), "(name='subnetId')\n", (574793, 574810), False, 'import pulumi\n'), ((575142, 575170), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""zoneId"""'}), "(name='zoneId')\n", (575155, 575170), False, 'import pulumi\n'), ((577130, 577169), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""replicationFactor"""'}), "(name='replicationFactor')\n", (577143, 577169), False, 'import pulumi\n'), ((577540, 577573), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""topicConfig"""'}), "(name='topicConfig')\n", (577553, 577573), False, 'import pulumi\n'), ((580450, 580485), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""cleanupPolicy"""'}), "(name='cleanupPolicy')\n", (580463, 580485), False, 'import pulumi\n'), ((580761, 580798), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""compressionType"""'}), "(name='compressionType')\n", (580774, 580798), False, 'import pulumi\n'), ((581084, 581123), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""deleteRetentionMs"""'}), "(name='deleteRetentionMs')\n", (581097, 581123), False, 'import pulumi\n'), ((581424, 581463), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""fileDeleteDelayMs"""'}), "(name='fileDeleteDelayMs')\n", (581437, 581463), False, 'import pulumi\n'), ((581769, 581804), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""flushMessages"""'}), "(name='flushMessages')\n", (581782, 581804), False, 'import pulumi\n'), ((582080, 582109), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""flushMs"""'}), "(name='flushMs')\n", (582093, 582109), False, 'import pulumi\n'), ((582355, 582392), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxMessageBytes"""'}), "(name='maxMessageBytes')\n", (582368, 582392), False, 'import pulumi\n'), ((582683, 582723), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""minCompactionLagMs"""'}), "(name='minCompactionLagMs')\n", (582696, 582723), False, 'import pulumi\n'), ((583034, 583073), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""minInsyncReplicas"""'}), "(name='minInsyncReplicas')\n", (583047, 583073), False, 'import pulumi\n'), ((583650, 583686), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""retentionBytes"""'}), "(name='retentionBytes')\n", (583663, 583686), False, 'import pulumi\n'), ((583967, 584000), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""retentionMs"""'}), "(name='retentionMs')\n", (583980, 584000), False, 'import pulumi\n'), ((584266, 584300), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""segmentBytes"""'}), "(name='segmentBytes')\n", (584279, 584300), False, 'import pulumi\n'), ((587230, 587261), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""topicName"""'}), "(name='topicName')\n", (587243, 587261), False, 'import pulumi\n'), ((590021, 590056), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""cleanupPolicy"""'}), "(name='cleanupPolicy')\n", (590034, 590056), False, 'import pulumi\n'), ((590332, 590369), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""compressionType"""'}), "(name='compressionType')\n", (590345, 590369), False, 'import pulumi\n'), ((590655, 590694), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""deleteRetentionMs"""'}), "(name='deleteRetentionMs')\n", (590668, 590694), False, 'import pulumi\n'), ((590995, 591034), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""fileDeleteDelayMs"""'}), "(name='fileDeleteDelayMs')\n", (591008, 591034), False, 'import pulumi\n'), ((591340, 591375), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""flushMessages"""'}), "(name='flushMessages')\n", (591353, 591375), False, 'import pulumi\n'), ((591651, 591680), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""flushMs"""'}), "(name='flushMs')\n", (591664, 591680), False, 'import pulumi\n'), ((591926, 591963), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxMessageBytes"""'}), "(name='maxMessageBytes')\n", (591939, 591963), False, 'import pulumi\n'), ((592254, 592294), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""minCompactionLagMs"""'}), "(name='minCompactionLagMs')\n", (592267, 592294), False, 'import pulumi\n'), ((592605, 592644), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""minInsyncReplicas"""'}), "(name='minInsyncReplicas')\n", (592618, 592644), False, 'import pulumi\n'), ((593221, 593257), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""retentionBytes"""'}), "(name='retentionBytes')\n", (593234, 593257), False, 'import pulumi\n'), ((593538, 593571), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""retentionMs"""'}), "(name='retentionMs')\n", (593551, 593571), False, 'import pulumi\n'), ((593837, 593871), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""segmentBytes"""'}), "(name='segmentBytes')\n", (593850, 593871), False, 'import pulumi\n'), ((596404, 596443), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""backupWindowStart"""'}), "(name='backupWindowStart')\n", (596417, 596443), False, 'import pulumi\n'), ((596964, 597013), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""featureCompatibilityVersion"""'}), "(name='featureCompatibilityVersion')\n", (596977, 597013), False, 'import pulumi\n'), ((597869, 597899), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""dataLens"""'}), "(name='dataLens')\n", (597882, 597899), False, 'import pulumi\n'), ((602363, 602393), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""subnetId"""'}), "(name='subnetId')\n", (602376, 602393), False, 'import pulumi\n'), ((602784, 602812), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""zoneId"""'}), "(name='zoneId')\n", (602797, 602812), False, 'import pulumi\n'), ((603249, 603285), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""assignPublicIp"""'}), "(name='assignPublicIp')\n", (603262, 603285), False, 'import pulumi\n'), ((604674, 604705), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""shardName"""'}), "(name='shardName')\n", (604687, 604705), False, 'import pulumi\n'), ((608288, 608318), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""diskSize"""'}), "(name='diskSize')\n", (608301, 608318), False, 'import pulumi\n'), ((608646, 608678), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""diskTypeId"""'}), "(name='diskTypeId')\n", (608659, 608678), False, 'import pulumi\n'), ((609124, 609162), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""resourcePresetId"""'}), "(name='resourcePresetId')\n", (609137, 609162), False, 'import pulumi\n'), ((611913, 611947), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""databaseName"""'}), "(name='databaseName')\n", (611926, 611947), False, 'import pulumi\n'), ((613295, 613325), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""dataLens"""'}), "(name='dataLens')\n", (613308, 613325), False, 'import pulumi\n'), ((613690, 613718), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""webSql"""'}), "(name='webSql')\n", (613703, 613718), False, 'import pulumi\n'), ((618908, 618944), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""assignPublicIp"""'}), "(name='assignPublicIp')\n", (618921, 618944), False, 'import pulumi\n'), ((620177, 620216), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""replicationSource"""'}), "(name='replicationSource')\n", (620190, 620216), False, 'import pulumi\n'), ((620635, 620678), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""replicationSourceName"""'}), "(name='replicationSourceName')\n", (620648, 620678), False, 'import pulumi\n'), ((621193, 621223), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""subnetId"""'}), "(name='subnetId')\n", (621206, 621223), False, 'import pulumi\n'), ((624251, 624281), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""diskSize"""'}), "(name='diskSize')\n", (624264, 624281), False, 'import pulumi\n'), ((624607, 624639), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""diskTypeId"""'}), "(name='diskTypeId')\n", (624620, 624639), False, 'import pulumi\n'), ((624953, 624991), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""resourcePresetId"""'}), "(name='resourcePresetId')\n", (624966, 624991), False, 'import pulumi\n'), ((626017, 626047), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""backupId"""'}), "(name='backupId')\n", (626030, 626047), False, 'import pulumi\n'), ((630058, 630100), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""authenticationPlugin"""'}), "(name='authenticationPlugin')\n", (630071, 630100), False, 'import pulumi\n'), ((630611, 630649), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""connectionLimits"""'}), "(name='connectionLimits')\n", (630624, 630649), False, 'import pulumi\n'), ((631177, 631216), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""globalPermissions"""'}), "(name='globalPermissions')\n", (631190, 631216), False, 'import pulumi\n'), ((633566, 633609), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxConnectionsPerHour"""'}), "(name='maxConnectionsPerHour')\n", (633579, 633609), False, 'import pulumi\n'), ((633993, 634034), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxQuestionsPerHour"""'}), "(name='maxQuestionsPerHour')\n", (634006, 634034), False, 'import pulumi\n'), ((634406, 634445), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxUpdatesPerHour"""'}), "(name='maxUpdatesPerHour')\n", (634419, 634445), False, 'import pulumi\n'), ((634805, 634845), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxUserConnections"""'}), "(name='maxUserConnections')\n", (634818, 634845), False, 'import pulumi\n'), ((636055, 636089), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""databaseName"""'}), "(name='databaseName')\n", (636068, 636089), False, 'import pulumi\n'), ((640330, 640367), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxmemoryPolicy"""'}), "(name='maxmemoryPolicy')\n", (640343, 640367), False, 'import pulumi\n'), ((640917, 640959), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""notifyKeyspaceEvents"""'}), "(name='notifyKeyspaceEvents')\n", (640930, 640959), False, 'import pulumi\n'), ((641372, 641414), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""slowlogLogSlowerThan"""'}), "(name='slowlogLogSlowerThan')\n", (641385, 641414), False, 'import pulumi\n'), ((641819, 641854), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""slowlogMaxLen"""'}), "(name='slowlogMaxLen')\n", (641832, 641854), False, 'import pulumi\n'), ((644512, 644543), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""shardName"""'}), "(name='shardName')\n", (644525, 644543), False, 'import pulumi\n'), ((644880, 644910), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""subnetId"""'}), "(name='subnetId')\n", (644893, 644910), False, 'import pulumi\n'), ((648113, 648143), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""diskSize"""'}), "(name='diskSize')\n", (648126, 648143), False, 'import pulumi\n'), ((648463, 648501), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""resourcePresetId"""'}), "(name='resourcePresetId')\n", (648476, 648501), False, 'import pulumi\n'), ((648777, 648809), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""diskTypeId"""'}), "(name='diskTypeId')\n", (648790, 648809), False, 'import pulumi\n'), ((652482, 652518), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""assignPublicIp"""'}), "(name='assignPublicIp')\n", (652495, 652518), False, 'import pulumi\n'), ((653296, 653326), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""subnetId"""'}), "(name='subnetId')\n", (653309, 653326), False, 'import pulumi\n'), ((654379, 654409), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""diskSize"""'}), "(name='diskSize')\n", (654392, 654409), False, 'import pulumi\n'), ((654739, 654771), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""diskTypeId"""'}), "(name='diskTypeId')\n", (654752, 654771), False, 'import pulumi\n'), ((655089, 655127), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""resourcePresetId"""'}), "(name='resourcePresetId')\n", (655102, 655127), False, 'import pulumi\n'), ((658066, 658100), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""databaseName"""'}), "(name='databaseName')\n", (658079, 658100), False, 'import pulumi\n'), ((660570, 660606), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""allowedMethods"""'}), "(name='allowedMethods')\n", (660583, 660606), False, 'import pulumi\n'), ((661033, 661069), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""allowedOrigins"""'}), "(name='allowedOrigins')\n", (661046, 661069), False, 'import pulumi\n'), ((661447, 661483), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""allowedHeaders"""'}), "(name='allowedHeaders')\n", (661460, 661483), False, 'import pulumi\n'), ((661881, 661916), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""exposeHeaders"""'}), "(name='exposeHeaders')\n", (661894, 661916), False, 'import pulumi\n'), ((662313, 662348), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxAgeSeconds"""'}), "(name='maxAgeSeconds')\n", (662326, 662348), False, 'import pulumi\n'), ((667873, 667929), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""abortIncompleteMultipartUploadDays"""'}), "(name='abortIncompleteMultipartUploadDays')\n", (667886, 667929), False, 'import pulumi\n'), ((669257, 669306), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""noncurrentVersionExpiration"""'}), "(name='noncurrentVersionExpiration')\n", (669270, 669306), False, 'import pulumi\n'), ((669870, 669920), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""noncurrentVersionTransitions"""'}), "(name='noncurrentVersionTransitions')\n", (669883, 669920), False, 'import pulumi\n'), ((673258, 673305), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""expiredObjectDeleteMarker"""'}), "(name='expiredObjectDeleteMarker')\n", (673271, 673305), False, 'import pulumi\n'), ((675268, 675302), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""storageClass"""'}), "(name='storageClass')\n", (675281, 675302), False, 'import pulumi\n'), ((676970, 677004), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""storageClass"""'}), "(name='storageClass')\n", (676983, 677004), False, 'import pulumi\n'), ((678700, 678734), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""targetBucket"""'}), "(name='targetBucket')\n", (678713, 678734), False, 'import pulumi\n'), ((679075, 679109), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""targetPrefix"""'}), "(name='targetPrefix')\n", (679088, 679109), False, 'import pulumi\n'), ((681008, 681064), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""applyServerSideEncryptionByDefault"""'}), "(name='applyServerSideEncryptionByDefault')\n", (681021, 681064), False, 'import pulumi\n'), ((682373, 682409), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""kmsMasterKeyId"""'}), "(name='kmsMasterKeyId')\n", (682386, 682409), False, 'import pulumi\n'), ((682767, 682801), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""sseAlgorithm"""'}), "(name='sseAlgorithm')\n", (682780, 682801), False, 'import pulumi\n'), ((685696, 685731), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""errorDocument"""'}), "(name='errorDocument')\n", (685709, 685731), False, 'import pulumi\n'), ((686106, 686141), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""indexDocument"""'}), "(name='indexDocument')\n", (686119, 686141), False, 'import pulumi\n'), ((686553, 686596), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""redirectAllRequestsTo"""'}), "(name='redirectAllRequestsTo')\n", (686566, 686596), False, 'import pulumi\n'), ((687196, 687230), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""routingRules"""'}), "(name='routingRules')\n", (687209, 687230), False, 'import pulumi\n'), ((689189, 689233), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""ddosProtectionProvider"""'}), "(name='ddosProtectionProvider')\n", (689202, 689233), False, 'import pulumi\n'), ((689645, 689689), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""outgoingSmtpCapability"""'}), "(name='outgoingSmtpCapability')\n", (689658, 689689), False, 'import pulumi\n'), ((690080, 690108), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""zoneId"""'}), "(name='zoneId')\n", (690093, 690108), False, 'import pulumi\n'), ((693130, 693160), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""fromPort"""'}), "(name='fromPort')\n", (693143, 693160), False, 'import pulumi\n'), ((694315, 694353), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""predefinedTarget"""'}), "(name='predefinedTarget')\n", (694328, 694353), False, 'import pulumi\n'), ((694644, 694681), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""securityGroupId"""'}), "(name='securityGroupId')\n", (694657, 694681), False, 'import pulumi\n'), ((694972, 695000), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""toPort"""'}), "(name='toPort')\n", (694985, 695000), False, 'import pulumi\n'), ((695241, 695275), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""v4CidrBlocks"""'}), "(name='v4CidrBlocks')\n", (695254, 695275), False, 'import pulumi\n'), ((695599, 695633), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""v6CidrBlocks"""'}), "(name='v6CidrBlocks')\n", (695612, 695633), False, 'import pulumi\n'), ((698676, 698706), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""fromPort"""'}), "(name='fromPort')\n", (698689, 698706), False, 'import pulumi\n'), ((699861, 699899), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""predefinedTarget"""'}), "(name='predefinedTarget')\n", (699874, 699899), False, 'import pulumi\n'), ((700190, 700227), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""securityGroupId"""'}), "(name='securityGroupId')\n", (700203, 700227), False, 'import pulumi\n'), ((700518, 700546), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""toPort"""'}), "(name='toPort')\n", (700531, 700546), False, 'import pulumi\n'), ((700787, 700821), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""v4CidrBlocks"""'}), "(name='v4CidrBlocks')\n", (700800, 700821), False, 'import pulumi\n'), ((701145, 701179), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""v6CidrBlocks"""'}), "(name='v6CidrBlocks')\n", (701158, 701179), False, 'import pulumi\n'), ((702149, 702188), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""destinationPrefix"""'}), "(name='destinationPrefix')\n", (702162, 702188), False, 'import pulumi\n'), ((702547, 702583), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""nextHopAddress"""'}), "(name='nextHopAddress')\n", (702560, 702583), False, 'import pulumi\n'), ((705627, 705657), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""fromPort"""'}), "(name='fromPort')\n", (705640, 705657), False, 'import pulumi\n'), ((706802, 706840), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""predefinedTarget"""'}), "(name='predefinedTarget')\n", (706815, 706840), False, 'import pulumi\n'), ((707131, 707168), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""securityGroupId"""'}), "(name='securityGroupId')\n", (707144, 707168), False, 'import pulumi\n'), ((707459, 707487), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""toPort"""'}), "(name='toPort')\n", (707472, 707487), False, 'import pulumi\n'), ((707728, 707762), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""v4CidrBlocks"""'}), "(name='v4CidrBlocks')\n", (707741, 707762), False, 'import pulumi\n'), ((708086, 708120), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""v6CidrBlocks"""'}), "(name='v6CidrBlocks')\n", (708099, 708120), False, 'import pulumi\n'), ((711146, 711176), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""fromPort"""'}), "(name='fromPort')\n", (711159, 711176), False, 'import pulumi\n'), ((712321, 712359), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""predefinedTarget"""'}), "(name='predefinedTarget')\n", (712334, 712359), False, 'import pulumi\n'), ((712650, 712687), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""securityGroupId"""'}), "(name='securityGroupId')\n", (712663, 712687), False, 'import pulumi\n'), ((712978, 713006), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""toPort"""'}), "(name='toPort')\n", (712991, 713006), False, 'import pulumi\n'), ((713247, 713281), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""v4CidrBlocks"""'}), "(name='v4CidrBlocks')\n", (713260, 713281), False, 'import pulumi\n'), ((713605, 713639), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""v6CidrBlocks"""'}), "(name='v6CidrBlocks')\n", (713618, 713639), False, 'import pulumi\n'), ((714905, 714937), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""domainName"""'}), "(name='domainName')\n", (714918, 714937), False, 'import pulumi\n'), ((715243, 715282), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""domainNameServers"""'}), "(name='domainNameServers')\n", (715256, 715282), False, 'import pulumi\n'), ((715696, 715728), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""ntpServers"""'}), "(name='ntpServers')\n", (715709, 715728), False, 'import pulumi\n'), ((718024, 718056), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""fixedScale"""'}), "(name='fixedScale')\n", (718037, 718056), False, 'import pulumi\n'), ((719778, 719810), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""groupCount"""'}), "(name='groupCount')\n", (719791, 719810), False, 'import pulumi\n'), ((720158, 720193), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""storageTypeId"""'}), "(name='storageTypeId')\n", (720171, 720193), False, 'import pulumi\n'), ((722759, 722800), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""loadBalancingConfig"""'}), "(name='loadBalancingConfig')\n", (722772, 722800), False, 'import pulumi\n'), ((723801, 723837), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""targetGroupIds"""'}), "(name='targetGroupIds')\n", (723814, 723837), False, 'import pulumi\n'), ((727661, 727698), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""grpcHealthcheck"""'}), "(name='grpcHealthcheck')\n", (727674, 727698), False, 'import pulumi\n'), ((728185, 728222), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""healthcheckPort"""'}), "(name='healthcheckPort')\n", (728198, 728222), False, 'import pulumi\n'), ((728539, 728577), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""healthyThreshold"""'}), "(name='healthyThreshold')\n", (728552, 728577), False, 'import pulumi\n'), ((729063, 729100), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""httpHealthcheck"""'}), "(name='httpHealthcheck')\n", (729076, 729100), False, 'import pulumi\n'), ((729862, 729905), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""intervalJitterPercent"""'}), "(name='intervalJitterPercent')\n", (729875, 729905), False, 'import pulumi\n'), ((730392, 730431), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""streamHealthcheck"""'}), "(name='streamHealthcheck')\n", (730405, 730431), False, 'import pulumi\n'), ((731214, 731254), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""unhealthyThreshold"""'}), "(name='unhealthyThreshold')\n", (731227, 731254), False, 'import pulumi\n'), ((732072, 732105), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""serviceName"""'}), "(name='serviceName')\n", (732085, 732105), False, 'import pulumi\n'), ((735830, 735879), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""localityAwareRoutingPercent"""'}), "(name='localityAwareRoutingPercent')\n", (735843, 735879), False, 'import pulumi\n'), ((736334, 736370), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""panicThreshold"""'}), "(name='panicThreshold')\n", (736347, 736370), False, 'import pulumi\n'), ((736893, 736929), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""strictLocality"""'}), "(name='strictLocality')\n", (736906, 736929), False, 'import pulumi\n'), ((738491, 738530), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""validationContext"""'}), "(name='validationContext')\n", (738504, 738530), False, 'import pulumi\n'), ((739200, 739236), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""trustedCaBytes"""'}), "(name='trustedCaBytes')\n", (739213, 739236), False, 'import pulumi\n'), ((739474, 739507), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""trustedCaId"""'}), "(name='trustedCaId')\n", (739487, 739507), False, 'import pulumi\n'), ((742280, 742321), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""loadBalancingConfig"""'}), "(name='loadBalancingConfig')\n", (742293, 742321), False, 'import pulumi\n'), ((743322, 743358), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""targetGroupIds"""'}), "(name='targetGroupIds')\n", (743335, 743358), False, 'import pulumi\n'), ((747182, 747219), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""grpcHealthcheck"""'}), "(name='grpcHealthcheck')\n", (747195, 747219), False, 'import pulumi\n'), ((747706, 747743), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""healthcheckPort"""'}), "(name='healthcheckPort')\n", (747719, 747743), False, 'import pulumi\n'), ((748060, 748098), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""healthyThreshold"""'}), "(name='healthyThreshold')\n", (748073, 748098), False, 'import pulumi\n'), ((748584, 748621), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""httpHealthcheck"""'}), "(name='httpHealthcheck')\n", (748597, 748621), False, 'import pulumi\n'), ((749383, 749426), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""intervalJitterPercent"""'}), "(name='intervalJitterPercent')\n", (749396, 749426), False, 'import pulumi\n'), ((749913, 749952), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""streamHealthcheck"""'}), "(name='streamHealthcheck')\n", (749926, 749952), False, 'import pulumi\n'), ((750735, 750775), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""unhealthyThreshold"""'}), "(name='unhealthyThreshold')\n", (750748, 750775), False, 'import pulumi\n'), ((751593, 751626), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""serviceName"""'}), "(name='serviceName')\n", (751606, 751626), False, 'import pulumi\n'), ((755351, 755400), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""localityAwareRoutingPercent"""'}), "(name='localityAwareRoutingPercent')\n", (755364, 755400), False, 'import pulumi\n'), ((755855, 755891), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""panicThreshold"""'}), "(name='panicThreshold')\n", (755868, 755891), False, 'import pulumi\n'), ((756414, 756450), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""strictLocality"""'}), "(name='strictLocality')\n", (756427, 756450), False, 'import pulumi\n'), ((758012, 758051), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""validationContext"""'}), "(name='validationContext')\n", (758025, 758051), False, 'import pulumi\n'), ((758721, 758757), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""trustedCaBytes"""'}), "(name='trustedCaBytes')\n", (758734, 758757), False, 'import pulumi\n'), ((758995, 759028), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""trustedCaId"""'}), "(name='trustedCaId')\n", (759008, 759028), False, 'import pulumi\n'), ((759476, 759518), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""diskPlacementGroupId"""'}), "(name='diskPlacementGroupId')\n", (759489, 759518), False, 'import pulumi\n'), ((760001, 760039), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""placementGroupId"""'}), "(name='placementGroupId')\n", (760014, 760039), False, 'import pulumi\n'), ((761730, 761770), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""zoneInstancesLimit"""'}), "(name='zoneInstancesLimit')\n", (761743, 761770), False, 'import pulumi\n'), ((762048, 762087), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""zoneRequestsLimit"""'}), "(name='zoneRequestsLimit')\n", (762061, 762087), False, 'import pulumi\n'), ((768194, 768230), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""assignPublicIp"""'}), "(name='assignPublicIp')\n", (768207, 768230), False, 'import pulumi\n'), ((768597, 768631), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""brokersCount"""'}), "(name='brokersCount')\n", (768610, 768631), False, 'import pulumi\n'), ((768957, 768993), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""schemaRegistry"""'}), "(name='schemaRegistry')\n", (768970, 768993), False, 'import pulumi\n'), ((769368, 769405), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""unmanagedTopics"""'}), "(name='unmanagedTopics')\n", (769381, 769405), False, 'import pulumi\n'), ((771004, 771037), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""kafkaConfig"""'}), "(name='kafkaConfig')\n", (771017, 771037), False, 'import pulumi\n'), ((774437, 774481), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""autoCreateTopicsEnable"""'}), "(name='autoCreateTopicsEnable')\n", (774450, 774481), False, 'import pulumi\n'), ((774786, 774823), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""compressionType"""'}), "(name='compressionType')\n", (774799, 774823), False, 'import pulumi\n'), ((775081, 775127), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""defaultReplicationFactor"""'}), "(name='defaultReplicationFactor')\n", (775094, 775127), False, 'import pulumi\n'), ((775435, 775481), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""logFlushIntervalMessages"""'}), "(name='logFlushIntervalMessages')\n", (775448, 775481), False, 'import pulumi\n'), ((775794, 775834), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""logFlushIntervalMs"""'}), "(name='logFlushIntervalMs')\n", (775807, 775834), False, 'import pulumi\n'), ((776117, 776166), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""logFlushSchedulerIntervalMs"""'}), "(name='logFlushSchedulerIntervalMs')\n", (776130, 776166), False, 'import pulumi\n'), ((776499, 776535), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""logPreallocate"""'}), "(name='logPreallocate')\n", (776512, 776535), False, 'import pulumi\n'), ((776790, 776829), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""logRetentionBytes"""'}), "(name='logRetentionBytes')\n", (776803, 776829), False, 'import pulumi\n'), ((777102, 777141), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""logRetentionHours"""'}), "(name='logRetentionHours')\n", (777115, 777141), False, 'import pulumi\n'), ((777414, 777455), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""logRetentionMinutes"""'}), "(name='logRetentionMinutes')\n", (777427, 777455), False, 'import pulumi\n'), ((777738, 777774), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""logRetentionMs"""'}), "(name='logRetentionMs')\n", (777751, 777774), False, 'import pulumi\n'), ((778032, 778069), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""logSegmentBytes"""'}), "(name='logSegmentBytes')\n", (778045, 778069), False, 'import pulumi\n'), ((778332, 778367), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""numPartitions"""'}), "(name='numPartitions')\n", (778345, 778367), False, 'import pulumi\n'), ((778615, 778661), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""socketReceiveBufferBytes"""'}), "(name='socketReceiveBufferBytes')\n", (778628, 778661), False, 'import pulumi\n'), ((778974, 779017), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""socketSendBufferBytes"""'}), "(name='socketSendBufferBytes')\n", (778987, 779017), False, 'import pulumi\n'), ((780059, 780089), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""diskSize"""'}), "(name='diskSize')\n", (780072, 780089), False, 'import pulumi\n'), ((780402, 780434), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""diskTypeId"""'}), "(name='diskTypeId')\n", (780415, 780434), False, 'import pulumi\n'), ((780860, 780898), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""resourcePresetId"""'}), "(name='resourcePresetId')\n", (780873, 780898), False, 'import pulumi\n'), ((782767, 782797), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""diskSize"""'}), "(name='diskSize')\n", (782780, 782797), False, 'import pulumi\n'), ((783110, 783142), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""diskTypeId"""'}), "(name='diskTypeId')\n", (783123, 783142), False, 'import pulumi\n'), ((783568, 783606), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""resourcePresetId"""'}), "(name='resourcePresetId')\n", (783581, 783606), False, 'import pulumi\n'), ((784979, 785010), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""clusterId"""'}), "(name='clusterId')\n", (784992, 785010), False, 'import pulumi\n'), ((785835, 785874), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""replicationFactor"""'}), "(name='replicationFactor')\n", (785848, 785874), False, 'import pulumi\n'), ((786228, 786261), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""topicConfig"""'}), "(name='topicConfig')\n", (786241, 786261), False, 'import pulumi\n'), ((788948, 788983), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""cleanupPolicy"""'}), "(name='cleanupPolicy')\n", (788961, 788983), False, 'import pulumi\n'), ((789231, 789268), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""compressionType"""'}), "(name='compressionType')\n", (789244, 789268), False, 'import pulumi\n'), ((789526, 789565), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""deleteRetentionMs"""'}), "(name='deleteRetentionMs')\n", (789539, 789565), False, 'import pulumi\n'), ((789838, 789877), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""fileDeleteDelayMs"""'}), "(name='fileDeleteDelayMs')\n", (789851, 789877), False, 'import pulumi\n'), ((790155, 790190), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""flushMessages"""'}), "(name='flushMessages')\n", (790168, 790190), False, 'import pulumi\n'), ((790438, 790467), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""flushMs"""'}), "(name='flushMs')\n", (790451, 790467), False, 'import pulumi\n'), ((790685, 790722), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxMessageBytes"""'}), "(name='maxMessageBytes')\n", (790698, 790722), False, 'import pulumi\n'), ((790985, 791025), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""minCompactionLagMs"""'}), "(name='minCompactionLagMs')\n", (790998, 791025), False, 'import pulumi\n'), ((791308, 791347), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""minInsyncReplicas"""'}), "(name='minInsyncReplicas')\n", (791321, 791347), False, 'import pulumi\n'), ((791868, 791904), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""retentionBytes"""'}), "(name='retentionBytes')\n", (791881, 791904), False, 'import pulumi\n'), ((792157, 792190), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""retentionMs"""'}), "(name='retentionMs')\n", (792170, 792190), False, 'import pulumi\n'), ((792428, 792462), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""segmentBytes"""'}), "(name='segmentBytes')\n", (792441, 792462), False, 'import pulumi\n'), ((795132, 795163), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""topicName"""'}), "(name='topicName')\n", (795145, 795163), False, 'import pulumi\n'), ((796023, 796053), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""dataLens"""'}), "(name='dataLens')\n", (796036, 796053), False, 'import pulumi\n'), ((796370, 796398), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""webSql"""'}), "(name='webSql')\n", (796383, 796398), False, 'import pulumi\n'), ((15548, 15582), 'pulumi.set', 'pulumi.set', (['__self__', '"""name"""', 'name'], {}), "(__self__, 'name', name)\n", (15558, 15582), False, 'import pulumi\n'), ((15591, 15649), 'pulumi.set', 'pulumi.set', (['__self__', '"""target_group_ids"""', 'target_group_ids'], {}), "(__self__, 'target_group_ids', target_group_ids)\n", (15601, 15649), False, 'import pulumi\n'), ((16248, 16272), 'pulumi.get', 'pulumi.get', (['self', '"""name"""'], {}), "(self, 'name')\n", (16258, 16272), False, 'import pulumi\n'), ((16345, 16376), 'pulumi.set', 'pulumi.set', (['self', '"""name"""', 'value'], {}), "(self, 'name', value)\n", (16355, 16376), False, 'import pulumi\n'), ((16600, 16636), 'pulumi.get', 'pulumi.get', (['self', '"""target_group_ids"""'], {}), "(self, 'target_group_ids')\n", (16610, 16636), False, 'import pulumi\n'), ((16757, 16800), 'pulumi.set', 'pulumi.set', (['self', '"""target_group_ids"""', 'value'], {}), "(self, 'target_group_ids', value)\n", (16767, 16800), False, 'import pulumi\n'), ((17072, 17103), 'pulumi.get', 'pulumi.get', (['self', '"""healthcheck"""'], {}), "(self, 'healthcheck')\n", (17082, 17103), False, 'import pulumi\n'), ((17240, 17278), 'pulumi.set', 'pulumi.set', (['self', '"""healthcheck"""', 'value'], {}), "(self, 'healthcheck', value)\n", (17250, 17278), False, 'import pulumi\n'), ((17606, 17647), 'pulumi.get', 'pulumi.get', (['self', '"""load_balancing_config"""'], {}), "(self, 'load_balancing_config')\n", (17616, 17647), False, 'import pulumi\n'), ((17812, 17860), 'pulumi.set', 'pulumi.set', (['self', '"""load_balancing_config"""', 'value'], {}), "(self, 'load_balancing_config', value)\n", (17822, 17860), False, 'import pulumi\n'), ((18020, 18044), 'pulumi.get', 'pulumi.get', (['self', '"""port"""'], {}), "(self, 'port')\n", (18030, 18044), False, 'import pulumi\n'), ((18127, 18158), 'pulumi.set', 'pulumi.set', (['self', '"""port"""', 'value'], {}), "(self, 'port', value)\n", (18137, 18158), False, 'import pulumi\n'), ((18406, 18429), 'pulumi.get', 'pulumi.get', (['self', '"""tls"""'], {}), "(self, 'tls')\n", (18416, 18429), False, 'import pulumi\n'), ((18542, 18572), 'pulumi.set', 'pulumi.set', (['self', '"""tls"""', 'value'], {}), "(self, 'tls', value)\n", (18552, 18572), False, 'import pulumi\n'), ((18822, 18848), 'pulumi.get', 'pulumi.get', (['self', '"""weight"""'], {}), "(self, 'weight')\n", (18832, 18848), False, 'import pulumi\n'), ((18935, 18968), 'pulumi.set', 'pulumi.set', (['self', '"""weight"""', 'value'], {}), "(self, 'weight', value)\n", (18945, 18968), False, 'import pulumi\n'), ((21504, 21546), 'pulumi.set', 'pulumi.set', (['__self__', '"""interval"""', 'interval'], {}), "(__self__, 'interval', interval)\n", (21514, 21546), False, 'import pulumi\n'), ((21555, 21595), 'pulumi.set', 'pulumi.set', (['__self__', '"""timeout"""', 'timeout'], {}), "(__self__, 'timeout', timeout)\n", (21565, 21595), False, 'import pulumi\n'), ((22577, 22605), 'pulumi.get', 'pulumi.get', (['self', '"""interval"""'], {}), "(self, 'interval')\n", (22587, 22605), False, 'import pulumi\n'), ((22686, 22721), 'pulumi.set', 'pulumi.set', (['self', '"""interval"""', 'value'], {}), "(self, 'interval', value)\n", (22696, 22721), False, 'import pulumi\n'), ((22889, 22916), 'pulumi.get', 'pulumi.get', (['self', '"""timeout"""'], {}), "(self, 'timeout')\n", (22899, 22916), False, 'import pulumi\n'), ((22995, 23029), 'pulumi.set', 'pulumi.set', (['self', '"""timeout"""', 'value'], {}), "(self, 'timeout', value)\n", (23005, 23029), False, 'import pulumi\n'), ((23354, 23390), 'pulumi.get', 'pulumi.get', (['self', '"""grpc_healthcheck"""'], {}), "(self, 'grpc_healthcheck')\n", (23364, 23390), False, 'import pulumi\n'), ((23552, 23595), 'pulumi.set', 'pulumi.set', (['self', '"""grpc_healthcheck"""', 'value'], {}), "(self, 'grpc_healthcheck', value)\n", (23562, 23595), False, 'import pulumi\n'), ((23811, 23847), 'pulumi.get', 'pulumi.get', (['self', '"""healthcheck_port"""'], {}), "(self, 'healthcheck_port')\n", (23821, 23847), False, 'import pulumi\n'), ((23954, 23997), 'pulumi.set', 'pulumi.set', (['self', '"""healthcheck_port"""', 'value'], {}), "(self, 'healthcheck_port', value)\n", (23964, 23997), False, 'import pulumi\n'), ((24379, 24416), 'pulumi.get', 'pulumi.get', (['self', '"""healthy_threshold"""'], {}), "(self, 'healthy_threshold')\n", (24389, 24416), False, 'import pulumi\n'), ((24525, 24569), 'pulumi.set', 'pulumi.set', (['self', '"""healthy_threshold"""', 'value'], {}), "(self, 'healthy_threshold', value)\n", (24535, 24569), False, 'import pulumi\n'), ((24894, 24930), 'pulumi.get', 'pulumi.get', (['self', '"""http_healthcheck"""'], {}), "(self, 'http_healthcheck')\n", (24904, 24930), False, 'import pulumi\n'), ((25092, 25135), 'pulumi.set', 'pulumi.set', (['self', '"""http_healthcheck"""', 'value'], {}), "(self, 'http_healthcheck', value)\n", (25102, 25135), False, 'import pulumi\n'), ((25497, 25540), 'pulumi.get', 'pulumi.get', (['self', '"""interval_jitter_percent"""'], {}), "(self, 'interval_jitter_percent')\n", (25507, 25540), False, 'import pulumi\n'), ((25663, 25713), 'pulumi.set', 'pulumi.set', (['self', '"""interval_jitter_percent"""', 'value'], {}), "(self, 'interval_jitter_percent', value)\n", (25673, 25713), False, 'import pulumi\n'), ((26046, 26084), 'pulumi.get', 'pulumi.get', (['self', '"""stream_healthcheck"""'], {}), "(self, 'stream_healthcheck')\n", (26056, 26084), False, 'import pulumi\n'), ((26252, 26297), 'pulumi.set', 'pulumi.set', (['self', '"""stream_healthcheck"""', 'value'], {}), "(self, 'stream_healthcheck', value)\n", (26262, 26297), False, 'import pulumi\n'), ((26665, 26704), 'pulumi.get', 'pulumi.get', (['self', '"""unhealthy_threshold"""'], {}), "(self, 'unhealthy_threshold')\n", (26675, 26704), False, 'import pulumi\n'), ((26817, 26863), 'pulumi.set', 'pulumi.set', (['self', '"""unhealthy_threshold"""', 'value'], {}), "(self, 'unhealthy_threshold', value)\n", (26827, 26863), False, 'import pulumi\n'), ((27498, 27530), 'pulumi.get', 'pulumi.get', (['self', '"""service_name"""'], {}), "(self, 'service_name')\n", (27508, 27530), False, 'import pulumi\n'), ((27629, 27668), 'pulumi.set', 'pulumi.set', (['self', '"""service_name"""', 'value'], {}), "(self, 'service_name', value)\n", (27639, 27668), False, 'import pulumi\n'), ((28173, 28207), 'pulumi.set', 'pulumi.set', (['__self__', '"""path"""', 'path'], {}), "(__self__, 'path', path)\n", (28183, 28207), False, 'import pulumi\n'), ((28496, 28520), 'pulumi.get', 'pulumi.get', (['self', '"""path"""'], {}), "(self, 'path')\n", (28506, 28520), False, 'import pulumi\n'), ((28593, 28624), 'pulumi.set', 'pulumi.set', (['self', '"""path"""', 'value'], {}), "(self, 'path', value)\n", (28603, 28624), False, 'import pulumi\n'), ((28783, 28807), 'pulumi.get', 'pulumi.get', (['self', '"""host"""'], {}), "(self, 'host')\n", (28793, 28807), False, 'import pulumi\n'), ((28890, 28921), 'pulumi.set', 'pulumi.set', (['self', '"""host"""', 'value'], {}), "(self, 'host', value)\n", (28900, 28921), False, 'import pulumi\n'), ((29094, 29119), 'pulumi.get', 'pulumi.get', (['self', '"""http2"""'], {}), "(self, 'http2')\n", (29104, 29119), False, 'import pulumi\n'), ((29205, 29237), 'pulumi.set', 'pulumi.set', (['self', '"""http2"""', 'value'], {}), "(self, 'http2', value)\n", (29215, 29237), False, 'import pulumi\n'), ((29989, 30016), 'pulumi.get', 'pulumi.get', (['self', '"""receive"""'], {}), "(self, 'receive')\n", (29999, 30016), False, 'import pulumi\n'), ((30105, 30139), 'pulumi.set', 'pulumi.set', (['self', '"""receive"""', 'value'], {}), "(self, 'receive', value)\n", (30115, 30139), False, 'import pulumi\n'), ((30333, 30357), 'pulumi.get', 'pulumi.get', (['self', '"""send"""'], {}), "(self, 'send')\n", (30343, 30357), False, 'import pulumi\n'), ((30440, 30471), 'pulumi.set', 'pulumi.set', (['self', '"""send"""', 'value'], {}), "(self, 'send', value)\n", (30450, 30471), False, 'import pulumi\n'), ((32167, 32217), 'pulumi.get', 'pulumi.get', (['self', '"""locality_aware_routing_percent"""'], {}), "(self, 'locality_aware_routing_percent')\n", (32177, 32217), False, 'import pulumi\n'), ((32352, 32409), 'pulumi.set', 'pulumi.set', (['self', '"""locality_aware_routing_percent"""', 'value'], {}), "(self, 'locality_aware_routing_percent', value)\n", (32362, 32409), False, 'import pulumi\n'), ((32834, 32869), 'pulumi.get', 'pulumi.get', (['self', '"""panic_threshold"""'], {}), "(self, 'panic_threshold')\n", (32844, 32869), False, 'import pulumi\n'), ((32974, 33016), 'pulumi.set', 'pulumi.set', (['self', '"""panic_threshold"""', 'value'], {}), "(self, 'panic_threshold', value)\n", (32984, 33016), False, 'import pulumi\n'), ((33300, 33335), 'pulumi.get', 'pulumi.get', (['self', '"""strict_locality"""'], {}), "(self, 'strict_locality')\n", (33310, 33335), False, 'import pulumi\n'), ((33441, 33483), 'pulumi.set', 'pulumi.set', (['self', '"""strict_locality"""', 'value'], {}), "(self, 'strict_locality', value)\n", (33451, 33483), False, 'import pulumi\n'), ((34769, 34792), 'pulumi.get', 'pulumi.get', (['self', '"""sni"""'], {}), "(self, 'sni')\n", (34779, 34792), False, 'import pulumi\n'), ((34873, 34903), 'pulumi.set', 'pulumi.set', (['self', '"""sni"""', 'value'], {}), "(self, 'sni', value)\n", (34883, 34903), False, 'import pulumi\n'), ((35093, 35131), 'pulumi.get', 'pulumi.get', (['self', '"""validation_context"""'], {}), "(self, 'validation_context')\n", (35103, 35131), False, 'import pulumi\n'), ((35291, 35336), 'pulumi.set', 'pulumi.set', (['self', '"""validation_context"""', 'value'], {}), "(self, 'validation_context', value)\n", (35301, 35336), False, 'import pulumi\n'), ((35936, 35972), 'pulumi.get', 'pulumi.get', (['self', '"""trusted_ca_bytes"""'], {}), "(self, 'trusted_ca_bytes')\n", (35946, 35972), False, 'import pulumi\n'), ((36079, 36122), 'pulumi.set', 'pulumi.set', (['self', '"""trusted_ca_bytes"""', 'value'], {}), "(self, 'trusted_ca_bytes', value)\n", (36089, 36122), False, 'import pulumi\n'), ((36252, 36285), 'pulumi.get', 'pulumi.get', (['self', '"""trusted_ca_id"""'], {}), "(self, 'trusted_ca_id')\n", (36262, 36285), False, 'import pulumi\n'), ((36386, 36426), 'pulumi.set', 'pulumi.set', (['self', '"""trusted_ca_id"""', 'value'], {}), "(self, 'trusted_ca_id', value)\n", (36396, 36426), False, 'import pulumi\n'), ((38186, 38220), 'pulumi.set', 'pulumi.set', (['__self__', '"""name"""', 'name'], {}), "(__self__, 'name', name)\n", (38196, 38220), False, 'import pulumi\n'), ((38229, 38287), 'pulumi.set', 'pulumi.set', (['__self__', '"""target_group_ids"""', 'target_group_ids'], {}), "(__self__, 'target_group_ids', target_group_ids)\n", (38239, 38287), False, 'import pulumi\n'), ((38965, 38989), 'pulumi.get', 'pulumi.get', (['self', '"""name"""'], {}), "(self, 'name')\n", (38975, 38989), False, 'import pulumi\n'), ((39062, 39093), 'pulumi.set', 'pulumi.set', (['self', '"""name"""', 'value'], {}), "(self, 'name', value)\n", (39072, 39093), False, 'import pulumi\n'), ((39317, 39353), 'pulumi.get', 'pulumi.get', (['self', '"""target_group_ids"""'], {}), "(self, 'target_group_ids')\n", (39327, 39353), False, 'import pulumi\n'), ((39474, 39517), 'pulumi.set', 'pulumi.set', (['self', '"""target_group_ids"""', 'value'], {}), "(self, 'target_group_ids', value)\n", (39484, 39517), False, 'import pulumi\n'), ((39789, 39820), 'pulumi.get', 'pulumi.get', (['self', '"""healthcheck"""'], {}), "(self, 'healthcheck')\n", (39799, 39820), False, 'import pulumi\n'), ((39957, 39995), 'pulumi.set', 'pulumi.set', (['self', '"""healthcheck"""', 'value'], {}), "(self, 'healthcheck', value)\n", (39967, 39995), False, 'import pulumi\n'), ((40168, 40193), 'pulumi.get', 'pulumi.get', (['self', '"""http2"""'], {}), "(self, 'http2')\n", (40178, 40193), False, 'import pulumi\n'), ((40279, 40311), 'pulumi.set', 'pulumi.set', (['self', '"""http2"""', 'value'], {}), "(self, 'http2', value)\n", (40289, 40311), False, 'import pulumi\n'), ((40639, 40680), 'pulumi.get', 'pulumi.get', (['self', '"""load_balancing_config"""'], {}), "(self, 'load_balancing_config')\n", (40649, 40680), False, 'import pulumi\n'), ((40845, 40893), 'pulumi.set', 'pulumi.set', (['self', '"""load_balancing_config"""', 'value'], {}), "(self, 'load_balancing_config', value)\n", (40855, 40893), False, 'import pulumi\n'), ((41053, 41077), 'pulumi.get', 'pulumi.get', (['self', '"""port"""'], {}), "(self, 'port')\n", (41063, 41077), False, 'import pulumi\n'), ((41160, 41191), 'pulumi.set', 'pulumi.set', (['self', '"""port"""', 'value'], {}), "(self, 'port', value)\n", (41170, 41191), False, 'import pulumi\n'), ((41439, 41462), 'pulumi.get', 'pulumi.get', (['self', '"""tls"""'], {}), "(self, 'tls')\n", (41449, 41462), False, 'import pulumi\n'), ((41575, 41605), 'pulumi.set', 'pulumi.set', (['self', '"""tls"""', 'value'], {}), "(self, 'tls', value)\n", (41585, 41605), False, 'import pulumi\n'), ((41855, 41881), 'pulumi.get', 'pulumi.get', (['self', '"""weight"""'], {}), "(self, 'weight')\n", (41865, 41881), False, 'import pulumi\n'), ((41968, 42001), 'pulumi.set', 'pulumi.set', (['self', '"""weight"""', 'value'], {}), "(self, 'weight', value)\n", (41978, 42001), False, 'import pulumi\n'), ((44537, 44579), 'pulumi.set', 'pulumi.set', (['__self__', '"""interval"""', 'interval'], {}), "(__self__, 'interval', interval)\n", (44547, 44579), False, 'import pulumi\n'), ((44588, 44628), 'pulumi.set', 'pulumi.set', (['__self__', '"""timeout"""', 'timeout'], {}), "(__self__, 'timeout', timeout)\n", (44598, 44628), False, 'import pulumi\n'), ((45610, 45638), 'pulumi.get', 'pulumi.get', (['self', '"""interval"""'], {}), "(self, 'interval')\n", (45620, 45638), False, 'import pulumi\n'), ((45719, 45754), 'pulumi.set', 'pulumi.set', (['self', '"""interval"""', 'value'], {}), "(self, 'interval', value)\n", (45729, 45754), False, 'import pulumi\n'), ((45922, 45949), 'pulumi.get', 'pulumi.get', (['self', '"""timeout"""'], {}), "(self, 'timeout')\n", (45932, 45949), False, 'import pulumi\n'), ((46028, 46062), 'pulumi.set', 'pulumi.set', (['self', '"""timeout"""', 'value'], {}), "(self, 'timeout', value)\n", (46038, 46062), False, 'import pulumi\n'), ((46387, 46423), 'pulumi.get', 'pulumi.get', (['self', '"""grpc_healthcheck"""'], {}), "(self, 'grpc_healthcheck')\n", (46397, 46423), False, 'import pulumi\n'), ((46585, 46628), 'pulumi.set', 'pulumi.set', (['self', '"""grpc_healthcheck"""', 'value'], {}), "(self, 'grpc_healthcheck', value)\n", (46595, 46628), False, 'import pulumi\n'), ((46844, 46880), 'pulumi.get', 'pulumi.get', (['self', '"""healthcheck_port"""'], {}), "(self, 'healthcheck_port')\n", (46854, 46880), False, 'import pulumi\n'), ((46987, 47030), 'pulumi.set', 'pulumi.set', (['self', '"""healthcheck_port"""', 'value'], {}), "(self, 'healthcheck_port', value)\n", (46997, 47030), False, 'import pulumi\n'), ((47412, 47449), 'pulumi.get', 'pulumi.get', (['self', '"""healthy_threshold"""'], {}), "(self, 'healthy_threshold')\n", (47422, 47449), False, 'import pulumi\n'), ((47558, 47602), 'pulumi.set', 'pulumi.set', (['self', '"""healthy_threshold"""', 'value'], {}), "(self, 'healthy_threshold', value)\n", (47568, 47602), False, 'import pulumi\n'), ((47927, 47963), 'pulumi.get', 'pulumi.get', (['self', '"""http_healthcheck"""'], {}), "(self, 'http_healthcheck')\n", (47937, 47963), False, 'import pulumi\n'), ((48125, 48168), 'pulumi.set', 'pulumi.set', (['self', '"""http_healthcheck"""', 'value'], {}), "(self, 'http_healthcheck', value)\n", (48135, 48168), False, 'import pulumi\n'), ((48530, 48573), 'pulumi.get', 'pulumi.get', (['self', '"""interval_jitter_percent"""'], {}), "(self, 'interval_jitter_percent')\n", (48540, 48573), False, 'import pulumi\n'), ((48696, 48746), 'pulumi.set', 'pulumi.set', (['self', '"""interval_jitter_percent"""', 'value'], {}), "(self, 'interval_jitter_percent', value)\n", (48706, 48746), False, 'import pulumi\n'), ((49079, 49117), 'pulumi.get', 'pulumi.get', (['self', '"""stream_healthcheck"""'], {}), "(self, 'stream_healthcheck')\n", (49089, 49117), False, 'import pulumi\n'), ((49285, 49330), 'pulumi.set', 'pulumi.set', (['self', '"""stream_healthcheck"""', 'value'], {}), "(self, 'stream_healthcheck', value)\n", (49295, 49330), False, 'import pulumi\n'), ((49698, 49737), 'pulumi.get', 'pulumi.get', (['self', '"""unhealthy_threshold"""'], {}), "(self, 'unhealthy_threshold')\n", (49708, 49737), False, 'import pulumi\n'), ((49850, 49896), 'pulumi.set', 'pulumi.set', (['self', '"""unhealthy_threshold"""', 'value'], {}), "(self, 'unhealthy_threshold', value)\n", (49860, 49896), False, 'import pulumi\n'), ((50531, 50563), 'pulumi.get', 'pulumi.get', (['self', '"""service_name"""'], {}), "(self, 'service_name')\n", (50541, 50563), False, 'import pulumi\n'), ((50662, 50701), 'pulumi.set', 'pulumi.set', (['self', '"""service_name"""', 'value'], {}), "(self, 'service_name', value)\n", (50672, 50701), False, 'import pulumi\n'), ((51206, 51240), 'pulumi.set', 'pulumi.set', (['__self__', '"""path"""', 'path'], {}), "(__self__, 'path', path)\n", (51216, 51240), False, 'import pulumi\n'), ((51529, 51553), 'pulumi.get', 'pulumi.get', (['self', '"""path"""'], {}), "(self, 'path')\n", (51539, 51553), False, 'import pulumi\n'), ((51626, 51657), 'pulumi.set', 'pulumi.set', (['self', '"""path"""', 'value'], {}), "(self, 'path', value)\n", (51636, 51657), False, 'import pulumi\n'), ((51816, 51840), 'pulumi.get', 'pulumi.get', (['self', '"""host"""'], {}), "(self, 'host')\n", (51826, 51840), False, 'import pulumi\n'), ((51923, 51954), 'pulumi.set', 'pulumi.set', (['self', '"""host"""', 'value'], {}), "(self, 'host', value)\n", (51933, 51954), False, 'import pulumi\n'), ((52127, 52152), 'pulumi.get', 'pulumi.get', (['self', '"""http2"""'], {}), "(self, 'http2')\n", (52137, 52152), False, 'import pulumi\n'), ((52238, 52270), 'pulumi.set', 'pulumi.set', (['self', '"""http2"""', 'value'], {}), "(self, 'http2', value)\n", (52248, 52270), False, 'import pulumi\n'), ((53022, 53049), 'pulumi.get', 'pulumi.get', (['self', '"""receive"""'], {}), "(self, 'receive')\n", (53032, 53049), False, 'import pulumi\n'), ((53138, 53172), 'pulumi.set', 'pulumi.set', (['self', '"""receive"""', 'value'], {}), "(self, 'receive', value)\n", (53148, 53172), False, 'import pulumi\n'), ((53366, 53390), 'pulumi.get', 'pulumi.get', (['self', '"""send"""'], {}), "(self, 'send')\n", (53376, 53390), False, 'import pulumi\n'), ((53473, 53504), 'pulumi.set', 'pulumi.set', (['self', '"""send"""', 'value'], {}), "(self, 'send', value)\n", (53483, 53504), False, 'import pulumi\n'), ((55200, 55250), 'pulumi.get', 'pulumi.get', (['self', '"""locality_aware_routing_percent"""'], {}), "(self, 'locality_aware_routing_percent')\n", (55210, 55250), False, 'import pulumi\n'), ((55385, 55442), 'pulumi.set', 'pulumi.set', (['self', '"""locality_aware_routing_percent"""', 'value'], {}), "(self, 'locality_aware_routing_percent', value)\n", (55395, 55442), False, 'import pulumi\n'), ((55867, 55902), 'pulumi.get', 'pulumi.get', (['self', '"""panic_threshold"""'], {}), "(self, 'panic_threshold')\n", (55877, 55902), False, 'import pulumi\n'), ((56007, 56049), 'pulumi.set', 'pulumi.set', (['self', '"""panic_threshold"""', 'value'], {}), "(self, 'panic_threshold', value)\n", (56017, 56049), False, 'import pulumi\n'), ((56333, 56368), 'pulumi.get', 'pulumi.get', (['self', '"""strict_locality"""'], {}), "(self, 'strict_locality')\n", (56343, 56368), False, 'import pulumi\n'), ((56474, 56516), 'pulumi.set', 'pulumi.set', (['self', '"""strict_locality"""', 'value'], {}), "(self, 'strict_locality', value)\n", (56484, 56516), False, 'import pulumi\n'), ((57802, 57825), 'pulumi.get', 'pulumi.get', (['self', '"""sni"""'], {}), "(self, 'sni')\n", (57812, 57825), False, 'import pulumi\n'), ((57906, 57936), 'pulumi.set', 'pulumi.set', (['self', '"""sni"""', 'value'], {}), "(self, 'sni', value)\n", (57916, 57936), False, 'import pulumi\n'), ((58126, 58164), 'pulumi.get', 'pulumi.get', (['self', '"""validation_context"""'], {}), "(self, 'validation_context')\n", (58136, 58164), False, 'import pulumi\n'), ((58324, 58369), 'pulumi.set', 'pulumi.set', (['self', '"""validation_context"""', 'value'], {}), "(self, 'validation_context', value)\n", (58334, 58369), False, 'import pulumi\n'), ((58969, 59005), 'pulumi.get', 'pulumi.get', (['self', '"""trusted_ca_bytes"""'], {}), "(self, 'trusted_ca_bytes')\n", (58979, 59005), False, 'import pulumi\n'), ((59112, 59155), 'pulumi.set', 'pulumi.set', (['self', '"""trusted_ca_bytes"""', 'value'], {}), "(self, 'trusted_ca_bytes', value)\n", (59122, 59155), False, 'import pulumi\n'), ((59285, 59318), 'pulumi.get', 'pulumi.get', (['self', '"""trusted_ca_id"""'], {}), "(self, 'trusted_ca_id')\n", (59295, 59318), False, 'import pulumi\n'), ((59419, 59459), 'pulumi.set', 'pulumi.set', (['self', '"""trusted_ca_id"""', 'value'], {}), "(self, 'trusted_ca_id', value)\n", (59429, 59459), False, 'import pulumi\n'), ((59870, 59914), 'pulumi.set', 'pulumi.set', (['__self__', '"""locations"""', 'locations'], {}), "(__self__, 'locations', locations)\n", (59880, 59914), False, 'import pulumi\n'), ((60168, 60197), 'pulumi.get', 'pulumi.get', (['self', '"""locations"""'], {}), "(self, 'locations')\n", (60178, 60197), False, 'import pulumi\n'), ((60346, 60382), 'pulumi.set', 'pulumi.set', (['self', '"""locations"""', 'value'], {}), "(self, 'locations', value)\n", (60356, 60382), False, 'import pulumi\n'), ((60984, 61028), 'pulumi.set', 'pulumi.set', (['__self__', '"""subnet_id"""', 'subnet_id'], {}), "(__self__, 'subnet_id', subnet_id)\n", (60994, 61028), False, 'import pulumi\n'), ((61037, 61077), 'pulumi.set', 'pulumi.set', (['__self__', '"""zone_id"""', 'zone_id'], {}), "(__self__, 'zone_id', zone_id)\n", (61047, 61077), False, 'import pulumi\n'), ((61381, 61410), 'pulumi.get', 'pulumi.get', (['self', '"""subnet_id"""'], {}), "(self, 'subnet_id')\n", (61391, 61410), False, 'import pulumi\n'), ((61493, 61529), 'pulumi.set', 'pulumi.set', (['self', '"""subnet_id"""', 'value'], {}), "(self, 'subnet_id', value)\n", (61503, 61529), False, 'import pulumi\n'), ((61714, 61741), 'pulumi.get', 'pulumi.get', (['self', '"""zone_id"""'], {}), "(self, 'zone_id')\n", (61724, 61741), False, 'import pulumi\n'), ((61820, 61854), 'pulumi.set', 'pulumi.set', (['self', '"""zone_id"""', 'value'], {}), "(self, 'zone_id', value)\n", (61830, 61854), False, 'import pulumi\n'), ((62094, 62129), 'pulumi.get', 'pulumi.get', (['self', '"""disable_traffic"""'], {}), "(self, 'disable_traffic')\n", (62104, 62129), False, 'import pulumi\n'), ((62235, 62277), 'pulumi.set', 'pulumi.set', (['self', '"""disable_traffic"""', 'value'], {}), "(self, 'disable_traffic', value)\n", (62245, 62277), False, 'import pulumi\n'), ((63242, 63276), 'pulumi.set', 'pulumi.set', (['__self__', '"""name"""', 'name'], {}), "(__self__, 'name', name)\n", (63252, 63276), False, 'import pulumi\n'), ((63658, 63682), 'pulumi.get', 'pulumi.get', (['self', '"""name"""'], {}), "(self, 'name')\n", (63668, 63682), False, 'import pulumi\n'), ((63755, 63786), 'pulumi.set', 'pulumi.set', (['self', '"""name"""', 'value'], {}), "(self, 'name', value)\n", (63765, 63786), False, 'import pulumi\n'), ((64074, 64103), 'pulumi.get', 'pulumi.get', (['self', '"""endpoints"""'], {}), "(self, 'endpoints')\n", (64084, 64103), False, 'import pulumi\n'), ((64254, 64290), 'pulumi.set', 'pulumi.set', (['self', '"""endpoints"""', 'value'], {}), "(self, 'endpoints', value)\n", (64264, 64290), False, 'import pulumi\n'), ((64512, 64536), 'pulumi.get', 'pulumi.get', (['self', '"""http"""'], {}), "(self, 'http')\n", (64522, 64536), False, 'import pulumi\n'), ((64649, 64680), 'pulumi.set', 'pulumi.set', (['self', '"""http"""', 'value'], {}), "(self, 'http', value)\n", (64659, 64680), False, 'import pulumi\n'), ((64899, 64922), 'pulumi.get', 'pulumi.get', (['self', '"""tls"""'], {}), "(self, 'tls')\n", (64909, 64922), False, 'import pulumi\n'), ((65032, 65062), 'pulumi.set', 'pulumi.set', (['self', '"""tls"""', 'value'], {}), "(self, 'tls', value)\n", (65042, 65062), False, 'import pulumi\n'), ((65624, 65668), 'pulumi.set', 'pulumi.set', (['__self__', '"""addresses"""', 'addresses'], {}), "(__self__, 'addresses', addresses)\n", (65634, 65668), False, 'import pulumi\n'), ((65677, 65713), 'pulumi.set', 'pulumi.set', (['__self__', '"""ports"""', 'ports'], {}), "(__self__, 'ports', ports)\n", (65687, 65713), False, 'import pulumi\n'), ((65956, 65985), 'pulumi.get', 'pulumi.get', (['self', '"""addresses"""'], {}), "(self, 'addresses')\n", (65966, 65985), False, 'import pulumi\n'), ((66133, 66169), 'pulumi.set', 'pulumi.set', (['self', '"""addresses"""', 'value'], {}), "(self, 'addresses', value)\n", (66143, 66169), False, 'import pulumi\n'), ((66349, 66374), 'pulumi.get', 'pulumi.get', (['self', '"""ports"""'], {}), "(self, 'ports')\n", (66359, 66374), False, 'import pulumi\n'), ((66473, 66505), 'pulumi.set', 'pulumi.set', (['self', '"""ports"""', 'value'], {}), "(self, 'ports', value)\n", (66483, 66505), False, 'import pulumi\n'), ((68238, 68279), 'pulumi.get', 'pulumi.get', (['self', '"""external_ipv4_address"""'], {}), "(self, 'external_ipv4_address')\n", (68248, 68279), False, 'import pulumi\n'), ((68456, 68504), 'pulumi.set', 'pulumi.set', (['self', '"""external_ipv4_address"""', 'value'], {}), "(self, 'external_ipv4_address', value)\n", (68466, 68504), False, 'import pulumi\n'), ((68800, 68841), 'pulumi.get', 'pulumi.get', (['self', '"""external_ipv6_address"""'], {}), "(self, 'external_ipv6_address')\n", (68810, 68841), False, 'import pulumi\n'), ((69018, 69066), 'pulumi.set', 'pulumi.set', (['self', '"""external_ipv6_address"""', 'value'], {}), "(self, 'external_ipv6_address', value)\n", (69028, 69066), False, 'import pulumi\n'), ((69362, 69403), 'pulumi.get', 'pulumi.get', (['self', '"""internal_ipv4_address"""'], {}), "(self, 'internal_ipv4_address')\n", (69372, 69403), False, 'import pulumi\n'), ((69580, 69628), 'pulumi.set', 'pulumi.set', (['self', '"""internal_ipv4_address"""', 'value'], {}), "(self, 'internal_ipv4_address', value)\n", (69590, 69628), False, 'import pulumi\n'), ((70198, 70225), 'pulumi.get', 'pulumi.get', (['self', '"""address"""'], {}), "(self, 'address')\n", (70208, 70225), False, 'import pulumi\n'), ((70314, 70348), 'pulumi.set', 'pulumi.set', (['self', '"""address"""', 'value'], {}), "(self, 'address', value)\n", (70324, 70348), False, 'import pulumi\n'), ((70918, 70945), 'pulumi.get', 'pulumi.get', (['self', '"""address"""'], {}), "(self, 'address')\n", (70928, 70945), False, 'import pulumi\n'), ((71034, 71068), 'pulumi.set', 'pulumi.set', (['self', '"""address"""', 'value'], {}), "(self, 'address', value)\n", (71044, 71068), False, 'import pulumi\n'), ((71887, 71914), 'pulumi.get', 'pulumi.get', (['self', '"""address"""'], {}), "(self, 'address')\n", (71897, 71914), False, 'import pulumi\n'), ((72003, 72037), 'pulumi.set', 'pulumi.set', (['self', '"""address"""', 'value'], {}), "(self, 'address', value)\n", (72013, 72037), False, 'import pulumi\n'), ((72242, 72271), 'pulumi.get', 'pulumi.get', (['self', '"""subnet_id"""'], {}), "(self, 'subnet_id')\n", (72252, 72271), False, 'import pulumi\n'), ((72364, 72400), 'pulumi.set', 'pulumi.set', (['self', '"""subnet_id"""', 'value'], {}), "(self, 'subnet_id', value)\n", (72374, 72400), False, 'import pulumi\n'), ((73469, 73496), 'pulumi.get', 'pulumi.get', (['self', '"""handler"""'], {}), "(self, 'handler')\n", (73479, 73496), False, 'import pulumi\n'), ((73622, 73656), 'pulumi.set', 'pulumi.set', (['self', '"""handler"""', 'value'], {}), "(self, 'handler', value)\n", (73632, 73656), False, 'import pulumi\n'), ((73912, 73941), 'pulumi.get', 'pulumi.get', (['self', '"""redirects"""'], {}), "(self, 'redirects')\n", (73922, 73941), False, 'import pulumi\n'), ((74073, 74109), 'pulumi.set', 'pulumi.set', (['self', '"""redirects"""', 'value'], {}), "(self, 'redirects', value)\n", (74083, 74109), False, 'import pulumi\n'), ((75377, 75409), 'pulumi.get', 'pulumi.get', (['self', '"""allow_http10"""'], {}), "(self, 'allow_http10')\n", (75387, 75409), False, 'import pulumi\n'), ((75509, 75548), 'pulumi.set', 'pulumi.set', (['self', '"""allow_http10"""', 'value'], {}), "(self, 'allow_http10', value)\n", (75519, 75548), False, 'import pulumi\n'), ((75847, 75880), 'pulumi.get', 'pulumi.get', (['self', '"""http2_options"""'], {}), "(self, 'http2_options')\n", (75857, 75880), False, 'import pulumi\n'), ((76030, 76070), 'pulumi.set', 'pulumi.set', (['self', '"""http2_options"""', 'value'], {}), "(self, 'http2_options', value)\n", (76040, 76070), False, 'import pulumi\n'), ((76250, 76284), 'pulumi.get', 'pulumi.get', (['self', '"""http_router_id"""'], {}), "(self, 'http_router_id')\n", (76260, 76284), False, 'import pulumi\n'), ((76387, 76428), 'pulumi.set', 'pulumi.set', (['self', '"""http_router_id"""', 'value'], {}), "(self, 'http_router_id', value)\n", (76397, 76428), False, 'import pulumi\n'), ((77082, 77124), 'pulumi.get', 'pulumi.get', (['self', '"""max_concurrent_streams"""'], {}), "(self, 'max_concurrent_streams')\n", (77092, 77124), False, 'import pulumi\n'), ((77243, 77292), 'pulumi.set', 'pulumi.set', (['self', '"""max_concurrent_streams"""', 'value'], {}), "(self, 'max_concurrent_streams', value)\n", (77253, 77292), False, 'import pulumi\n'), ((77695, 77728), 'pulumi.get', 'pulumi.get', (['self', '"""http_to_https"""'], {}), "(self, 'http_to_https')\n", (77705, 77728), False, 'import pulumi\n'), ((77830, 77870), 'pulumi.set', 'pulumi.set', (['self', '"""http_to_https"""', 'value'], {}), "(self, 'http_to_https', value)\n", (77840, 77870), False, 'import pulumi\n'), ((78534, 78590), 'pulumi.set', 'pulumi.set', (['__self__', '"""default_handler"""', 'default_handler'], {}), "(__self__, 'default_handler', default_handler)\n", (78544, 78590), False, 'import pulumi\n'), ((78947, 78982), 'pulumi.get', 'pulumi.get', (['self', '"""default_handler"""'], {}), "(self, 'default_handler')\n", (78957, 78982), False, 'import pulumi\n'), ((79120, 79162), 'pulumi.set', 'pulumi.set', (['self', '"""default_handler"""', 'value'], {}), "(self, 'default_handler', value)\n", (79130, 79162), False, 'import pulumi\n'), ((79441, 79473), 'pulumi.get', 'pulumi.get', (['self', '"""sni_handlers"""'], {}), "(self, 'sni_handlers')\n", (79451, 79473), False, 'import pulumi\n'), ((79635, 79674), 'pulumi.set', 'pulumi.set', (['self', '"""sni_handlers"""', 'value'], {}), "(self, 'sni_handlers', value)\n", (79645, 79674), False, 'import pulumi\n'), ((80463, 80519), 'pulumi.set', 'pulumi.set', (['__self__', '"""certificate_ids"""', 'certificate_ids'], {}), "(__self__, 'certificate_ids', certificate_ids)\n", (80473, 80519), False, 'import pulumi\n'), ((81014, 81049), 'pulumi.get', 'pulumi.get', (['self', '"""certificate_ids"""'], {}), "(self, 'certificate_ids')\n", (81024, 81049), False, 'import pulumi\n'), ((81168, 81210), 'pulumi.set', 'pulumi.set', (['self', '"""certificate_ids"""', 'value'], {}), "(self, 'certificate_ids', value)\n", (81178, 81210), False, 'import pulumi\n'), ((81483, 81515), 'pulumi.get', 'pulumi.get', (['self', '"""http_handler"""'], {}), "(self, 'http_handler')\n", (81493, 81515), False, 'import pulumi\n'), ((81668, 81707), 'pulumi.set', 'pulumi.set', (['self', '"""http_handler"""', 'value'], {}), "(self, 'http_handler', value)\n", (81678, 81707), False, 'import pulumi\n'), ((83026, 83058), 'pulumi.get', 'pulumi.get', (['self', '"""allow_http10"""'], {}), "(self, 'allow_http10')\n", (83036, 83058), False, 'import pulumi\n'), ((83158, 83197), 'pulumi.set', 'pulumi.set', (['self', '"""allow_http10"""', 'value'], {}), "(self, 'allow_http10', value)\n", (83168, 83197), False, 'import pulumi\n'), ((83513, 83546), 'pulumi.get', 'pulumi.get', (['self', '"""http2_options"""'], {}), "(self, 'http2_options')\n", (83523, 83546), False, 'import pulumi\n'), ((83713, 83753), 'pulumi.set', 'pulumi.set', (['self', '"""http2_options"""', 'value'], {}), "(self, 'http2_options', value)\n", (83723, 83753), False, 'import pulumi\n'), ((83933, 83967), 'pulumi.get', 'pulumi.get', (['self', '"""http_router_id"""'], {}), "(self, 'http_router_id')\n", (83943, 83967), False, 'import pulumi\n'), ((84070, 84111), 'pulumi.set', 'pulumi.set', (['self', '"""http_router_id"""', 'value'], {}), "(self, 'http_router_id', value)\n", (84080, 84111), False, 'import pulumi\n'), ((84782, 84824), 'pulumi.get', 'pulumi.get', (['self', '"""max_concurrent_streams"""'], {}), "(self, 'max_concurrent_streams')\n", (84792, 84824), False, 'import pulumi\n'), ((84943, 84992), 'pulumi.set', 'pulumi.set', (['self', '"""max_concurrent_streams"""', 'value'], {}), "(self, 'max_concurrent_streams', value)\n", (84953, 84992), False, 'import pulumi\n'), ((85653, 85693), 'pulumi.set', 'pulumi.set', (['__self__', '"""handler"""', 'handler'], {}), "(__self__, 'handler', handler)\n", (85663, 85693), False, 'import pulumi\n'), ((85702, 85736), 'pulumi.set', 'pulumi.set', (['__self__', '"""name"""', 'name'], {}), "(__self__, 'name', name)\n", (85712, 85736), False, 'import pulumi\n'), ((85745, 85795), 'pulumi.set', 'pulumi.set', (['__self__', '"""server_names"""', 'server_names'], {}), "(__self__, 'server_names', server_names)\n", (85755, 85795), False, 'import pulumi\n'), ((86048, 86075), 'pulumi.get', 'pulumi.get', (['self', '"""handler"""'], {}), "(self, 'handler')\n", (86058, 86075), False, 'import pulumi\n'), ((86200, 86234), 'pulumi.set', 'pulumi.set', (['self', '"""handler"""', 'value'], {}), "(self, 'handler', value)\n", (86210, 86234), False, 'import pulumi\n'), ((86376, 86400), 'pulumi.get', 'pulumi.get', (['self', '"""name"""'], {}), "(self, 'name')\n", (86386, 86400), False, 'import pulumi\n'), ((86473, 86504), 'pulumi.set', 'pulumi.set', (['self', '"""name"""', 'value'], {}), "(self, 'name', value)\n", (86483, 86504), False, 'import pulumi\n'), ((86702, 86734), 'pulumi.get', 'pulumi.get', (['self', '"""server_names"""'], {}), "(self, 'server_names')\n", (86712, 86734), False, 'import pulumi\n'), ((86847, 86886), 'pulumi.set', 'pulumi.set', (['self', '"""server_names"""', 'value'], {}), "(self, 'server_names', value)\n", (86857, 86886), False, 'import pulumi\n'), ((87684, 87740), 'pulumi.set', 'pulumi.set', (['__self__', '"""certificate_ids"""', 'certificate_ids'], {}), "(__self__, 'certificate_ids', certificate_ids)\n", (87694, 87740), False, 'import pulumi\n'), ((88235, 88270), 'pulumi.get', 'pulumi.get', (['self', '"""certificate_ids"""'], {}), "(self, 'certificate_ids')\n", (88245, 88270), False, 'import pulumi\n'), ((88389, 88431), 'pulumi.set', 'pulumi.set', (['self', '"""certificate_ids"""', 'value'], {}), "(self, 'certificate_ids', value)\n", (88399, 88431), False, 'import pulumi\n'), ((88707, 88739), 'pulumi.get', 'pulumi.get', (['self', '"""http_handler"""'], {}), "(self, 'http_handler')\n", (88717, 88739), False, 'import pulumi\n'), ((88895, 88934), 'pulumi.set', 'pulumi.set', (['self', '"""http_handler"""', 'value'], {}), "(self, 'http_handler', value)\n", (88905, 88934), False, 'import pulumi\n'), ((90262, 90294), 'pulumi.get', 'pulumi.get', (['self', '"""allow_http10"""'], {}), "(self, 'allow_http10')\n", (90272, 90294), False, 'import pulumi\n'), ((90394, 90433), 'pulumi.set', 'pulumi.set', (['self', '"""allow_http10"""', 'value'], {}), "(self, 'allow_http10', value)\n", (90404, 90433), False, 'import pulumi\n'), ((90752, 90785), 'pulumi.get', 'pulumi.get', (['self', '"""http2_options"""'], {}), "(self, 'http2_options')\n", (90762, 90785), False, 'import pulumi\n'), ((90955, 90995), 'pulumi.set', 'pulumi.set', (['self', '"""http2_options"""', 'value'], {}), "(self, 'http2_options', value)\n", (90965, 90995), False, 'import pulumi\n'), ((91175, 91209), 'pulumi.get', 'pulumi.get', (['self', '"""http_router_id"""'], {}), "(self, 'http_router_id')\n", (91185, 91209), False, 'import pulumi\n'), ((91312, 91353), 'pulumi.set', 'pulumi.set', (['self', '"""http_router_id"""', 'value'], {}), "(self, 'http_router_id', value)\n", (91322, 91353), False, 'import pulumi\n'), ((92027, 92069), 'pulumi.get', 'pulumi.get', (['self', '"""max_concurrent_streams"""'], {}), "(self, 'max_concurrent_streams')\n", (92037, 92069), False, 'import pulumi\n'), ((92188, 92237), 'pulumi.set', 'pulumi.set', (['self', '"""max_concurrent_streams"""', 'value'], {}), "(self, 'max_concurrent_streams', value)\n", (92198, 92237), False, 'import pulumi\n'), ((92731, 92777), 'pulumi.set', 'pulumi.set', (['__self__', '"""ip_address"""', 'ip_address'], {}), "(__self__, 'ip_address', ip_address)\n", (92741, 92777), False, 'import pulumi\n'), ((92786, 92830), 'pulumi.set', 'pulumi.set', (['__self__', '"""subnet_id"""', 'subnet_id'], {}), "(__self__, 'subnet_id', subnet_id)\n", (92796, 92830), False, 'import pulumi\n'), ((93003, 93033), 'pulumi.get', 'pulumi.get', (['self', '"""ip_address"""'], {}), "(self, 'ip_address')\n", (93013, 93033), False, 'import pulumi\n'), ((93118, 93155), 'pulumi.set', 'pulumi.set', (['self', '"""ip_address"""', 'value'], {}), "(self, 'ip_address', value)\n", (93128, 93155), False, 'import pulumi\n'), ((93460, 93489), 'pulumi.get', 'pulumi.get', (['self', '"""subnet_id"""'], {}), "(self, 'subnet_id')\n", (93470, 93489), False, 'import pulumi\n'), ((93572, 93608), 'pulumi.set', 'pulumi.set', (['self', '"""subnet_id"""', 'value'], {}), "(self, 'subnet_id', value)\n", (93582, 93608), False, 'import pulumi\n'), ((94921, 94947), 'pulumi.get', 'pulumi.get', (['self', '"""append"""'], {}), "(self, 'append')\n", (94931, 94947), False, 'import pulumi\n'), ((95034, 95067), 'pulumi.set', 'pulumi.set', (['self', '"""append"""', 'value'], {}), "(self, 'append', value)\n", (95044, 95067), False, 'import pulumi\n'), ((95219, 95243), 'pulumi.get', 'pulumi.get', (['self', '"""name"""'], {}), "(self, 'name')\n", (95229, 95243), False, 'import pulumi\n'), ((95326, 95357), 'pulumi.set', 'pulumi.set', (['self', '"""name"""', 'value'], {}), "(self, 'name', value)\n", (95336, 95357), False, 'import pulumi\n'), ((95520, 95546), 'pulumi.get', 'pulumi.get', (['self', '"""remove"""'], {}), "(self, 'remove')\n", (95530, 95546), False, 'import pulumi\n'), ((95634, 95667), 'pulumi.set', 'pulumi.set', (['self', '"""remove"""', 'value'], {}), "(self, 'remove', value)\n", (95644, 95667), False, 'import pulumi\n'), ((96004, 96031), 'pulumi.get', 'pulumi.get', (['self', '"""replace"""'], {}), "(self, 'replace')\n", (96014, 96031), False, 'import pulumi\n'), ((96120, 96154), 'pulumi.set', 'pulumi.set', (['self', '"""replace"""', 'value'], {}), "(self, 'replace', value)\n", (96130, 96154), False, 'import pulumi\n'), ((97468, 97494), 'pulumi.get', 'pulumi.get', (['self', '"""append"""'], {}), "(self, 'append')\n", (97478, 97494), False, 'import pulumi\n'), ((97581, 97614), 'pulumi.set', 'pulumi.set', (['self', '"""append"""', 'value'], {}), "(self, 'append', value)\n", (97591, 97614), False, 'import pulumi\n'), ((97766, 97790), 'pulumi.get', 'pulumi.get', (['self', '"""name"""'], {}), "(self, 'name')\n", (97776, 97790), False, 'import pulumi\n'), ((97873, 97904), 'pulumi.set', 'pulumi.set', (['self', '"""name"""', 'value'], {}), "(self, 'name', value)\n", (97883, 97904), False, 'import pulumi\n'), ((98067, 98093), 'pulumi.get', 'pulumi.get', (['self', '"""remove"""'], {}), "(self, 'remove')\n", (98077, 98093), False, 'import pulumi\n'), ((98181, 98214), 'pulumi.set', 'pulumi.set', (['self', '"""remove"""', 'value'], {}), "(self, 'remove', value)\n", (98191, 98214), False, 'import pulumi\n'), ((98551, 98578), 'pulumi.get', 'pulumi.get', (['self', '"""replace"""'], {}), "(self, 'replace')\n", (98561, 98578), False, 'import pulumi\n'), ((98667, 98701), 'pulumi.set', 'pulumi.set', (['self', '"""replace"""', 'value'], {}), "(self, 'replace', value)\n", (98677, 98701), False, 'import pulumi\n'), ((99889, 99919), 'pulumi.get', 'pulumi.get', (['self', '"""grpc_route"""'], {}), "(self, 'grpc_route')\n", (99899, 99919), False, 'import pulumi\n'), ((100045, 100082), 'pulumi.set', 'pulumi.set', (['self', '"""grpc_route"""', 'value'], {}), "(self, 'grpc_route', value)\n", (100055, 100082), False, 'import pulumi\n'), ((100326, 100356), 'pulumi.get', 'pulumi.get', (['self', '"""http_route"""'], {}), "(self, 'http_route')\n", (100336, 100356), False, 'import pulumi\n'), ((100482, 100519), 'pulumi.set', 'pulumi.set', (['self', '"""http_route"""', 'value'], {}), "(self, 'http_route', value)\n", (100492, 100519), False, 'import pulumi\n'), ((100671, 100695), 'pulumi.get', 'pulumi.get', (['self', '"""name"""'], {}), "(self, 'name')\n", (100681, 100695), False, 'import pulumi\n'), ((100778, 100809), 'pulumi.set', 'pulumi.set', (['self', '"""name"""', 'value'], {}), "(self, 'name', value)\n", (100788, 100809), False, 'import pulumi\n'), ((102488, 102520), 'pulumi.get', 'pulumi.get', (['self', '"""grpc_matches"""'], {}), "(self, 'grpc_matches')\n", (102498, 102520), False, 'import pulumi\n'), ((102683, 102722), 'pulumi.set', 'pulumi.set', (['self', '"""grpc_matches"""', 'value'], {}), "(self, 'grpc_matches', value)\n", (102693, 102722), False, 'import pulumi\n'), ((103001, 103038), 'pulumi.get', 'pulumi.get', (['self', '"""grpc_route_action"""'], {}), "(self, 'grpc_route_action')\n", (103011, 103038), False, 'import pulumi\n'), ((103193, 103237), 'pulumi.set', 'pulumi.set', (['self', '"""grpc_route_action"""', 'value'], {}), "(self, 'grpc_route_action', value)\n", (103203, 103237), False, 'import pulumi\n'), ((103554, 103601), 'pulumi.get', 'pulumi.get', (['self', '"""grpc_status_response_action"""'], {}), "(self, 'grpc_status_response_action')\n", (103564, 103601), False, 'import pulumi\n'), ((103785, 103839), 'pulumi.set', 'pulumi.set', (['self', '"""grpc_status_response_action"""', 'value'], {}), "(self, 'grpc_status_response_action', value)\n", (103795, 103839), False, 'import pulumi\n'), ((104565, 104589), 'pulumi.get', 'pulumi.get', (['self', '"""fqmn"""'], {}), "(self, 'fqmn')\n", (104575, 104589), False, 'import pulumi\n'), ((104716, 104747), 'pulumi.set', 'pulumi.set', (['self', '"""fqmn"""', 'value'], {}), "(self, 'fqmn', value)\n", (104726, 104747), False, 'import pulumi\n'), ((105236, 105261), 'pulumi.get', 'pulumi.get', (['self', '"""exact"""'], {}), "(self, 'exact')\n", (105246, 105261), False, 'import pulumi\n'), ((105346, 105378), 'pulumi.set', 'pulumi.set', (['self', '"""exact"""', 'value'], {}), "(self, 'exact', value)\n", (105356, 105378), False, 'import pulumi\n'), ((105481, 105507), 'pulumi.get', 'pulumi.get', (['self', '"""prefix"""'], {}), "(self, 'prefix')\n", (105491, 105507), False, 'import pulumi\n'), ((105594, 105627), 'pulumi.set', 'pulumi.set', (['self', '"""prefix"""', 'value'], {}), "(self, 'prefix', value)\n", (105604, 105627), False, 'import pulumi\n'), ((106920, 106978), 'pulumi.set', 'pulumi.set', (['__self__', '"""backend_group_id"""', 'backend_group_id'], {}), "(__self__, 'backend_group_id', backend_group_id)\n", (106930, 106978), False, 'import pulumi\n'), ((107581, 107617), 'pulumi.get', 'pulumi.get', (['self', '"""backend_group_id"""'], {}), "(self, 'backend_group_id')\n", (107591, 107617), False, 'import pulumi\n'), ((107714, 107757), 'pulumi.set', 'pulumi.set', (['self', '"""backend_group_id"""', 'value'], {}), "(self, 'backend_group_id', value)\n", (107724, 107757), False, 'import pulumi\n'), ((107969, 108006), 'pulumi.get', 'pulumi.get', (['self', '"""auto_host_rewrite"""'], {}), "(self, 'auto_host_rewrite')\n", (107979, 108006), False, 'import pulumi\n'), ((108116, 108160), 'pulumi.set', 'pulumi.set', (['self', '"""auto_host_rewrite"""', 'value'], {}), "(self, 'auto_host_rewrite', value)\n", (108126, 108160), False, 'import pulumi\n'), ((108345, 108377), 'pulumi.get', 'pulumi.get', (['self', '"""host_rewrite"""'], {}), "(self, 'host_rewrite')\n", (108355, 108377), False, 'import pulumi\n'), ((108476, 108515), 'pulumi.set', 'pulumi.set', (['self', '"""host_rewrite"""', 'value'], {}), "(self, 'host_rewrite', value)\n", (108486, 108515), False, 'import pulumi\n'), ((109014, 109046), 'pulumi.get', 'pulumi.get', (['self', '"""idle_timeout"""'], {}), "(self, 'idle_timeout')\n", (109024, 109046), False, 'import pulumi\n'), ((109145, 109184), 'pulumi.set', 'pulumi.set', (['self', '"""idle_timeout"""', 'value'], {}), "(self, 'idle_timeout', value)\n", (109155, 109184), False, 'import pulumi\n'), ((109461, 109492), 'pulumi.get', 'pulumi.get', (['self', '"""max_timeout"""'], {}), "(self, 'max_timeout')\n", (109471, 109492), False, 'import pulumi\n'), ((109589, 109627), 'pulumi.set', 'pulumi.set', (['self', '"""max_timeout"""', 'value'], {}), "(self, 'max_timeout', value)\n", (109599, 109627), False, 'import pulumi\n'), ((110423, 110449), 'pulumi.get', 'pulumi.get', (['self', '"""status"""'], {}), "(self, 'status')\n", (110433, 110449), False, 'import pulumi\n'), ((110536, 110569), 'pulumi.set', 'pulumi.set', (['self', '"""status"""', 'value'], {}), "(self, 'status', value)\n", (110546, 110569), False, 'import pulumi\n'), ((112599, 112641), 'pulumi.get', 'pulumi.get', (['self', '"""direct_response_action"""'], {}), "(self, 'direct_response_action')\n", (112609, 112641), False, 'import pulumi\n'), ((112811, 112860), 'pulumi.set', 'pulumi.set', (['self', '"""direct_response_action"""', 'value'], {}), "(self, 'direct_response_action', value)\n", (112821, 112860), False, 'import pulumi\n'), ((113150, 113182), 'pulumi.get', 'pulumi.get', (['self', '"""http_matches"""'], {}), "(self, 'http_matches')\n", (113160, 113182), False, 'import pulumi\n'), ((113345, 113384), 'pulumi.set', 'pulumi.set', (['self', '"""http_matches"""', 'value'], {}), "(self, 'http_matches', value)\n", (113355, 113384), False, 'import pulumi\n'), ((113663, 113700), 'pulumi.get', 'pulumi.get', (['self', '"""http_route_action"""'], {}), "(self, 'http_route_action')\n", (113673, 113700), False, 'import pulumi\n'), ((113855, 113899), 'pulumi.set', 'pulumi.set', (['self', '"""http_route_action"""', 'value'], {}), "(self, 'http_route_action', value)\n", (113865, 113899), False, 'import pulumi\n'), ((114172, 114207), 'pulumi.get', 'pulumi.get', (['self', '"""redirect_action"""'], {}), "(self, 'redirect_action')\n", (114182, 114207), False, 'import pulumi\n'), ((114357, 114399), 'pulumi.set', 'pulumi.set', (['self', '"""redirect_action"""', 'value'], {}), "(self, 'redirect_action', value)\n", (114367, 114399), False, 'import pulumi\n'), ((115238, 115262), 'pulumi.get', 'pulumi.get', (['self', '"""body"""'], {}), "(self, 'body')\n", (115248, 115262), False, 'import pulumi\n'), ((115345, 115376), 'pulumi.set', 'pulumi.set', (['self', '"""body"""', 'value'], {}), "(self, 'body', value)\n", (115355, 115376), False, 'import pulumi\n'), ((115676, 115702), 'pulumi.get', 'pulumi.get', (['self', '"""status"""'], {}), "(self, 'status')\n", (115686, 115702), False, 'import pulumi\n'), ((115789, 115822), 'pulumi.set', 'pulumi.set', (['self', '"""status"""', 'value'], {}), "(self, 'status', value)\n", (115799, 115822), False, 'import pulumi\n'), ((116729, 116761), 'pulumi.get', 'pulumi.get', (['self', '"""http_methods"""'], {}), "(self, 'http_methods')\n", (116739, 116761), False, 'import pulumi\n'), ((116870, 116909), 'pulumi.set', 'pulumi.set', (['self', '"""http_methods"""', 'value'], {}), "(self, 'http_methods', value)\n", (116880, 116909), False, 'import pulumi\n'), ((117149, 117173), 'pulumi.get', 'pulumi.get', (['self', '"""path"""'], {}), "(self, 'path')\n", (117159, 117173), False, 'import pulumi\n'), ((117300, 117331), 'pulumi.set', 'pulumi.set', (['self', '"""path"""', 'value'], {}), "(self, 'path', value)\n", (117310, 117331), False, 'import pulumi\n'), ((117820, 117845), 'pulumi.get', 'pulumi.get', (['self', '"""exact"""'], {}), "(self, 'exact')\n", (117830, 117845), False, 'import pulumi\n'), ((117930, 117962), 'pulumi.set', 'pulumi.set', (['self', '"""exact"""', 'value'], {}), "(self, 'exact', value)\n", (117940, 117962), False, 'import pulumi\n'), ((118065, 118091), 'pulumi.get', 'pulumi.get', (['self', '"""prefix"""'], {}), "(self, 'prefix')\n", (118075, 118091), False, 'import pulumi\n'), ((118178, 118211), 'pulumi.set', 'pulumi.set', (['self', '"""prefix"""', 'value'], {}), "(self, 'prefix', value)\n", (118188, 118211), False, 'import pulumi\n'), ((119979, 120037), 'pulumi.set', 'pulumi.set', (['__self__', '"""backend_group_id"""', 'backend_group_id'], {}), "(__self__, 'backend_group_id', backend_group_id)\n", (119989, 120037), False, 'import pulumi\n'), ((120837, 120873), 'pulumi.get', 'pulumi.get', (['self', '"""backend_group_id"""'], {}), "(self, 'backend_group_id')\n", (120847, 120873), False, 'import pulumi\n'), ((120970, 121013), 'pulumi.set', 'pulumi.set', (['self', '"""backend_group_id"""', 'value'], {}), "(self, 'backend_group_id', value)\n", (120980, 121013), False, 'import pulumi\n'), ((121225, 121262), 'pulumi.get', 'pulumi.get', (['self', '"""auto_host_rewrite"""'], {}), "(self, 'auto_host_rewrite')\n", (121235, 121262), False, 'import pulumi\n'), ((121372, 121416), 'pulumi.set', 'pulumi.set', (['self', '"""auto_host_rewrite"""', 'value'], {}), "(self, 'auto_host_rewrite', value)\n", (121382, 121416), False, 'import pulumi\n'), ((121601, 121633), 'pulumi.get', 'pulumi.get', (['self', '"""host_rewrite"""'], {}), "(self, 'host_rewrite')\n", (121611, 121633), False, 'import pulumi\n'), ((121732, 121771), 'pulumi.set', 'pulumi.set', (['self', '"""host_rewrite"""', 'value'], {}), "(self, 'host_rewrite', value)\n", (121742, 121771), False, 'import pulumi\n'), ((122270, 122302), 'pulumi.get', 'pulumi.get', (['self', '"""idle_timeout"""'], {}), "(self, 'idle_timeout')\n", (122280, 122302), False, 'import pulumi\n'), ((122401, 122440), 'pulumi.set', 'pulumi.set', (['self', '"""idle_timeout"""', 'value'], {}), "(self, 'idle_timeout', value)\n", (122411, 122440), False, 'import pulumi\n'), ((122671, 122705), 'pulumi.get', 'pulumi.get', (['self', '"""prefix_rewrite"""'], {}), "(self, 'prefix_rewrite')\n", (122681, 122705), False, 'import pulumi\n'), ((122808, 122849), 'pulumi.set', 'pulumi.set', (['self', '"""prefix_rewrite"""', 'value'], {}), "(self, 'prefix_rewrite', value)\n", (122818, 122849), False, 'import pulumi\n'), ((123127, 123154), 'pulumi.get', 'pulumi.get', (['self', '"""timeout"""'], {}), "(self, 'timeout')\n", (123137, 123154), False, 'import pulumi\n'), ((123243, 123277), 'pulumi.set', 'pulumi.set', (['self', '"""timeout"""', 'value'], {}), "(self, 'timeout', value)\n", (123253, 123277), False, 'import pulumi\n'), ((123568, 123601), 'pulumi.get', 'pulumi.get', (['self', '"""upgrade_types"""'], {}), "(self, 'upgrade_types')\n", (123578, 123601), False, 'import pulumi\n'), ((123726, 123766), 'pulumi.set', 'pulumi.set', (['self', '"""upgrade_types"""', 'value'], {}), "(self, 'upgrade_types', value)\n", (123736, 123766), False, 'import pulumi\n'), ((126034, 126066), 'pulumi.get', 'pulumi.get', (['self', '"""remove_query"""'], {}), "(self, 'remove_query')\n", (126044, 126066), False, 'import pulumi\n'), ((126166, 126205), 'pulumi.set', 'pulumi.set', (['self', '"""remove_query"""', 'value'], {}), "(self, 'remove_query', value)\n", (126176, 126205), False, 'import pulumi\n'), ((126385, 126417), 'pulumi.get', 'pulumi.get', (['self', '"""replace_host"""'], {}), "(self, 'replace_host')\n", (126395, 126417), False, 'import pulumi\n'), ((126516, 126555), 'pulumi.set', 'pulumi.set', (['self', '"""replace_host"""', 'value'], {}), "(self, 'replace_host', value)\n", (126526, 126555), False, 'import pulumi\n'), ((126730, 126762), 'pulumi.get', 'pulumi.get', (['self', '"""replace_path"""'], {}), "(self, 'replace_path')\n", (126740, 126762), False, 'import pulumi\n'), ((126861, 126900), 'pulumi.set', 'pulumi.set', (['self', '"""replace_path"""', 'value'], {}), "(self, 'replace_path', value)\n", (126871, 126900), False, 'import pulumi\n'), ((127076, 127108), 'pulumi.get', 'pulumi.get', (['self', '"""replace_port"""'], {}), "(self, 'replace_port')\n", (127086, 127108), False, 'import pulumi\n'), ((127207, 127246), 'pulumi.set', 'pulumi.set', (['self', '"""replace_port"""', 'value'], {}), "(self, 'replace_port', value)\n", (127217, 127246), False, 'import pulumi\n'), ((127591, 127625), 'pulumi.get', 'pulumi.get', (['self', '"""replace_prefix"""'], {}), "(self, 'replace_prefix')\n", (127601, 127625), False, 'import pulumi\n'), ((127728, 127769), 'pulumi.set', 'pulumi.set', (['self', '"""replace_prefix"""', 'value'], {}), "(self, 'replace_prefix', value)\n", (127738, 127769), False, 'import pulumi\n'), ((128054, 128088), 'pulumi.get', 'pulumi.get', (['self', '"""replace_scheme"""'], {}), "(self, 'replace_scheme')\n", (128064, 128088), False, 'import pulumi\n'), ((128191, 128232), 'pulumi.set', 'pulumi.set', (['self', '"""replace_scheme"""', 'value'], {}), "(self, 'replace_scheme', value)\n", (128201, 128232), False, 'import pulumi\n'), ((128557, 128590), 'pulumi.get', 'pulumi.get', (['self', '"""response_code"""'], {}), "(self, 'response_code')\n", (128567, 128590), False, 'import pulumi\n'), ((128691, 128731), 'pulumi.set', 'pulumi.set', (['self', '"""response_code"""', 'value'], {}), "(self, 'response_code', value)\n", (128701, 128731), False, 'import pulumi\n'), ((129012, 129084), 'pulumi.set', 'pulumi.set', (['__self__', '"""disk_placement_group_id"""', 'disk_placement_group_id'], {}), "(__self__, 'disk_placement_group_id', disk_placement_group_id)\n", (129022, 129084), False, 'import pulumi\n'), ((129290, 129333), 'pulumi.get', 'pulumi.get', (['self', '"""disk_placement_group_id"""'], {}), "(self, 'disk_placement_group_id')\n", (129300, 129333), False, 'import pulumi\n'), ((129444, 129494), 'pulumi.set', 'pulumi.set', (['self', '"""disk_placement_group_id"""', 'value'], {}), "(self, 'disk_placement_group_id', value)\n", (129454, 129494), False, 'import pulumi\n'), ((131483, 131514), 'pulumi.get', 'pulumi.get', (['self', '"""auto_delete"""'], {}), "(self, 'auto_delete')\n", (131493, 131514), False, 'import pulumi\n'), ((131612, 131650), 'pulumi.set', 'pulumi.set', (['self', '"""auto_delete"""', 'value'], {}), "(self, 'auto_delete', value)\n", (131622, 131650), False, 'import pulumi\n'), ((131892, 131923), 'pulumi.get', 'pulumi.get', (['self', '"""device_name"""'], {}), "(self, 'device_name')\n", (131902, 131923), False, 'import pulumi\n'), ((132020, 132058), 'pulumi.set', 'pulumi.set', (['self', '"""device_name"""', 'value'], {}), "(self, 'device_name', value)\n", (132030, 132058), False, 'import pulumi\n'), ((132258, 132285), 'pulumi.get', 'pulumi.get', (['self', '"""disk_id"""'], {}), "(self, 'disk_id')\n", (132268, 132285), False, 'import pulumi\n'), ((132374, 132408), 'pulumi.set', 'pulumi.set', (['self', '"""disk_id"""', 'value'], {}), "(self, 'disk_id', value)\n", (132384, 132408), False, 'import pulumi\n'), ((132792, 132829), 'pulumi.get', 'pulumi.get', (['self', '"""initialize_params"""'], {}), "(self, 'initialize_params')\n", (132802, 132829), False, 'import pulumi\n'), ((132980, 133024), 'pulumi.set', 'pulumi.set', (['self', '"""initialize_params"""', 'value'], {}), "(self, 'initialize_params', value)\n", (132990, 133024), False, 'import pulumi\n'), ((133247, 133271), 'pulumi.get', 'pulumi.get', (['self', '"""mode"""'], {}), "(self, 'mode')\n", (133257, 133271), False, 'import pulumi\n'), ((133354, 133385), 'pulumi.set', 'pulumi.set', (['self', '"""mode"""', 'value'], {}), "(self, 'mode', value)\n", (133364, 133385), False, 'import pulumi\n'), ((134988, 135019), 'pulumi.get', 'pulumi.get', (['self', '"""description"""'], {}), "(self, 'description')\n", (134998, 135019), False, 'import pulumi\n'), ((135116, 135154), 'pulumi.set', 'pulumi.set', (['self', '"""description"""', 'value'], {}), "(self, 'description', value)\n", (135126, 135154), False, 'import pulumi\n'), ((135350, 135378), 'pulumi.get', 'pulumi.get', (['self', '"""image_id"""'], {}), "(self, 'image_id')\n", (135360, 135378), False, 'import pulumi\n'), ((135469, 135504), 'pulumi.set', 'pulumi.set', (['self', '"""image_id"""', 'value'], {}), "(self, 'image_id', value)\n", (135479, 135504), False, 'import pulumi\n'), ((135660, 135684), 'pulumi.get', 'pulumi.get', (['self', '"""name"""'], {}), "(self, 'name')\n", (135670, 135684), False, 'import pulumi\n'), ((135767, 135798), 'pulumi.set', 'pulumi.set', (['self', '"""name"""', 'value'], {}), "(self, 'name', value)\n", (135777, 135798), False, 'import pulumi\n'), ((135955, 135979), 'pulumi.get', 'pulumi.get', (['self', '"""size"""'], {}), "(self, 'size')\n", (135965, 135979), False, 'import pulumi\n'), ((136062, 136093), 'pulumi.set', 'pulumi.set', (['self', '"""size"""', 'value'], {}), "(self, 'size', value)\n", (136072, 136093), False, 'import pulumi\n'), ((136293, 136324), 'pulumi.get', 'pulumi.get', (['self', '"""snapshot_id"""'], {}), "(self, 'snapshot_id')\n", (136303, 136324), False, 'import pulumi\n'), ((136421, 136459), 'pulumi.set', 'pulumi.set', (['self', '"""snapshot_id"""', 'value'], {}), "(self, 'snapshot_id', value)\n", (136431, 136459), False, 'import pulumi\n'), ((136603, 136627), 'pulumi.get', 'pulumi.get', (['self', '"""type"""'], {}), "(self, 'type')\n", (136613, 136627), False, 'import pulumi\n'), ((136710, 136741), 'pulumi.set', 'pulumi.set', (['self', '"""type"""', 'value'], {}), "(self, 'type', value)\n", (136720, 136741), False, 'import pulumi\n'), ((137035, 137071), 'pulumi.set', 'pulumi.set', (['__self__', '"""zones"""', 'zones'], {}), "(__self__, 'zones', zones)\n", (137045, 137071), False, 'import pulumi\n'), ((137249, 137274), 'pulumi.get', 'pulumi.get', (['self', '"""zones"""'], {}), "(self, 'zones')\n", (137259, 137274), False, 'import pulumi\n'), ((137373, 137405), 'pulumi.set', 'pulumi.set', (['self', '"""zones"""', 'value'], {}), "(self, 'zones', value)\n", (137383, 137405), False, 'import pulumi\n'), ((139725, 139773), 'pulumi.get', 'pulumi.get', (['self', '"""max_opening_traffic_duration"""'], {}), "(self, 'max_opening_traffic_duration')\n", (139735, 139773), False, 'import pulumi\n'), ((139904, 139959), 'pulumi.set', 'pulumi.set', (['self', '"""max_opening_traffic_duration"""', 'value'], {}), "(self, 'max_opening_traffic_duration', value)\n", (139914, 139959), False, 'import pulumi\n'), ((140160, 140194), 'pulumi.get', 'pulumi.get', (['self', '"""status_message"""'], {}), "(self, 'status_message')\n", (140170, 140194), False, 'import pulumi\n'), ((140297, 140338), 'pulumi.set', 'pulumi.set', (['self', '"""status_message"""', 'value'], {}), "(self, 'status_message', value)\n", (140307, 140338), False, 'import pulumi\n'), ((140557, 140601), 'pulumi.get', 'pulumi.get', (['self', '"""target_group_description"""'], {}), "(self, 'target_group_description')\n", (140567, 140601), False, 'import pulumi\n'), ((140724, 140775), 'pulumi.set', 'pulumi.set', (['self', '"""target_group_description"""', 'value'], {}), "(self, 'target_group_description', value)\n", (140734, 140775), False, 'import pulumi\n'), ((140909, 140944), 'pulumi.get', 'pulumi.get', (['self', '"""target_group_id"""'], {}), "(self, 'target_group_id')\n", (140919, 140944), False, 'import pulumi\n'), ((141049, 141091), 'pulumi.set', 'pulumi.set', (['self', '"""target_group_id"""', 'value'], {}), "(self, 'target_group_id', value)\n", (141059, 141091), False, 'import pulumi\n'), ((141325, 141364), 'pulumi.get', 'pulumi.get', (['self', '"""target_group_labels"""'], {}), "(self, 'target_group_labels')\n", (141335, 141364), False, 'import pulumi\n'), ((141505, 141551), 'pulumi.set', 'pulumi.set', (['self', '"""target_group_labels"""', 'value'], {}), "(self, 'target_group_labels', value)\n", (141515, 141551), False, 'import pulumi\n'), ((141751, 141788), 'pulumi.get', 'pulumi.get', (['self', '"""target_group_name"""'], {}), "(self, 'target_group_name')\n", (141761, 141788), False, 'import pulumi\n'), ((141897, 141941), 'pulumi.set', 'pulumi.set', (['self', '"""target_group_name"""', 'value'], {}), "(self, 'target_group_name', value)\n", (141907, 141941), False, 'import pulumi\n'), ((143720, 143772), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_expansion"""', 'max_expansion'], {}), "(__self__, 'max_expansion', max_expansion)\n", (143730, 143772), False, 'import pulumi\n'), ((143781, 143837), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_unavailable"""', 'max_unavailable'], {}), "(__self__, 'max_unavailable', max_unavailable)\n", (143791, 143837), False, 'import pulumi\n'), ((144521, 144554), 'pulumi.get', 'pulumi.get', (['self', '"""max_expansion"""'], {}), "(self, 'max_expansion')\n", (144531, 144554), False, 'import pulumi\n'), ((144645, 144685), 'pulumi.set', 'pulumi.set', (['self', '"""max_expansion"""', 'value'], {}), "(self, 'max_expansion', value)\n", (144655, 144685), False, 'import pulumi\n'), ((144981, 145016), 'pulumi.get', 'pulumi.get', (['self', '"""max_unavailable"""'], {}), "(self, 'max_unavailable')\n", (144991, 145016), False, 'import pulumi\n'), ((145111, 145153), 'pulumi.set', 'pulumi.set', (['self', '"""max_unavailable"""', 'value'], {}), "(self, 'max_unavailable', value)\n", (145121, 145153), False, 'import pulumi\n'), ((145384, 145416), 'pulumi.get', 'pulumi.get', (['self', '"""max_creating"""'], {}), "(self, 'max_creating')\n", (145394, 145416), False, 'import pulumi\n'), ((145515, 145554), 'pulumi.set', 'pulumi.set', (['self', '"""max_creating"""', 'value'], {}), "(self, 'max_creating', value)\n", (145525, 145554), False, 'import pulumi\n'), ((145785, 145817), 'pulumi.get', 'pulumi.get', (['self', '"""max_deleting"""'], {}), "(self, 'max_deleting')\n", (145795, 145817), False, 'import pulumi\n'), ((145916, 145955), 'pulumi.set', 'pulumi.set', (['self', '"""max_deleting"""', 'value'], {}), "(self, 'max_deleting', value)\n", (145926, 145955), False, 'import pulumi\n'), ((146356, 146392), 'pulumi.get', 'pulumi.get', (['self', '"""startup_duration"""'], {}), "(self, 'startup_duration')\n", (146366, 146392), False, 'import pulumi\n'), ((146499, 146542), 'pulumi.set', 'pulumi.set', (['self', '"""startup_duration"""', 'value'], {}), "(self, 'startup_duration', value)\n", (146509, 146542), False, 'import pulumi\n'), ((146987, 147015), 'pulumi.get', 'pulumi.get', (['self', '"""strategy"""'], {}), "(self, 'strategy')\n", (146997, 147015), False, 'import pulumi\n'), ((147106, 147141), 'pulumi.set', 'pulumi.set', (['self', '"""strategy"""', 'value'], {}), "(self, 'strategy', value)\n", (147116, 147141), False, 'import pulumi\n'), ((149509, 149546), 'pulumi.get', 'pulumi.get', (['self', '"""healthy_threshold"""'], {}), "(self, 'healthy_threshold')\n", (149519, 149546), False, 'import pulumi\n'), ((149655, 149699), 'pulumi.set', 'pulumi.set', (['self', '"""healthy_threshold"""', 'value'], {}), "(self, 'healthy_threshold', value)\n", (149665, 149699), False, 'import pulumi\n'), ((149983, 150015), 'pulumi.get', 'pulumi.get', (['self', '"""http_options"""'], {}), "(self, 'http_options')\n", (149993, 150015), False, 'import pulumi\n'), ((150182, 150221), 'pulumi.set', 'pulumi.set', (['self', '"""http_options"""', 'value'], {}), "(self, 'http_options', value)\n", (150192, 150221), False, 'import pulumi\n'), ((150413, 150441), 'pulumi.get', 'pulumi.get', (['self', '"""interval"""'], {}), "(self, 'interval')\n", (150423, 150441), False, 'import pulumi\n'), ((150532, 150567), 'pulumi.set', 'pulumi.set', (['self', '"""interval"""', 'value'], {}), "(self, 'interval', value)\n", (150542, 150567), False, 'import pulumi\n'), ((150847, 150878), 'pulumi.get', 'pulumi.get', (['self', '"""tcp_options"""'], {}), "(self, 'tcp_options')\n", (150857, 150878), False, 'import pulumi\n'), ((151042, 151080), 'pulumi.set', 'pulumi.set', (['self', '"""tcp_options"""', 'value'], {}), "(self, 'tcp_options', value)\n", (151052, 151080), False, 'import pulumi\n'), ((151304, 151331), 'pulumi.get', 'pulumi.get', (['self', '"""timeout"""'], {}), "(self, 'timeout')\n", (151314, 151331), False, 'import pulumi\n'), ((151420, 151454), 'pulumi.set', 'pulumi.set', (['self', '"""timeout"""', 'value'], {}), "(self, 'timeout', value)\n", (151430, 151454), False, 'import pulumi\n'), ((151715, 151754), 'pulumi.get', 'pulumi.get', (['self', '"""unhealthy_threshold"""'], {}), "(self, 'unhealthy_threshold')\n", (151725, 151754), False, 'import pulumi\n'), ((151867, 151913), 'pulumi.set', 'pulumi.set', (['self', '"""unhealthy_threshold"""', 'value'], {}), "(self, 'unhealthy_threshold', value)\n", (151877, 151913), False, 'import pulumi\n'), ((152295, 152329), 'pulumi.set', 'pulumi.set', (['__self__', '"""path"""', 'path'], {}), "(__self__, 'path', path)\n", (152305, 152329), False, 'import pulumi\n'), ((152338, 152372), 'pulumi.set', 'pulumi.set', (['__self__', '"""port"""', 'port'], {}), "(__self__, 'port', port)\n", (152348, 152372), False, 'import pulumi\n'), ((152540, 152564), 'pulumi.get', 'pulumi.get', (['self', '"""path"""'], {}), "(self, 'path')\n", (152550, 152564), False, 'import pulumi\n'), ((152637, 152668), 'pulumi.set', 'pulumi.set', (['self', '"""path"""', 'value'], {}), "(self, 'path', value)\n", (152647, 152668), False, 'import pulumi\n'), ((152828, 152852), 'pulumi.get', 'pulumi.get', (['self', '"""port"""'], {}), "(self, 'port')\n", (152838, 152852), False, 'import pulumi\n'), ((152925, 152956), 'pulumi.set', 'pulumi.set', (['self', '"""port"""', 'value'], {}), "(self, 'port', value)\n", (152935, 152956), False, 'import pulumi\n'), ((153211, 153245), 'pulumi.set', 'pulumi.set', (['__self__', '"""port"""', 'port'], {}), "(__self__, 'port', port)\n", (153221, 153245), False, 'import pulumi\n'), ((153405, 153429), 'pulumi.get', 'pulumi.get', (['self', '"""port"""'], {}), "(self, 'port')\n", (153415, 153429), False, 'import pulumi\n'), ((153502, 153533), 'pulumi.set', 'pulumi.set', (['self', '"""port"""', 'value'], {}), "(self, 'port', value)\n", (153512, 153533), False, 'import pulumi\n'), ((156388, 156412), 'pulumi.get', 'pulumi.get', (['self', '"""fqdn"""'], {}), "(self, 'fqdn')\n", (156398, 156412), False, 'import pulumi\n'), ((156495, 156526), 'pulumi.set', 'pulumi.set', (['self', '"""fqdn"""', 'value'], {}), "(self, 'fqdn', value)\n", (156505, 156526), False, 'import pulumi\n'), ((156709, 156740), 'pulumi.get', 'pulumi.get', (['self', '"""instance_id"""'], {}), "(self, 'instance_id')\n", (156719, 156740), False, 'import pulumi\n'), ((156837, 156875), 'pulumi.set', 'pulumi.set', (['self', '"""instance_id"""', 'value'], {}), "(self, 'instance_id', value)\n", (156847, 156875), False, 'import pulumi\n'), ((157462, 157486), 'pulumi.get', 'pulumi.get', (['self', '"""name"""'], {}), "(self, 'name')\n", (157472, 157486), False, 'import pulumi\n'), ((157569, 157600), 'pulumi.set', 'pulumi.set', (['self', '"""name"""', 'value'], {}), "(self, 'name', value)\n", (157579, 157600), False, 'import pulumi\n'), ((157984, 158022), 'pulumi.get', 'pulumi.get', (['self', '"""network_interfaces"""'], {}), "(self, 'network_interfaces')\n", (157994, 158022), False, 'import pulumi\n'), ((158204, 158249), 'pulumi.set', 'pulumi.set', (['self', '"""network_interfaces"""', 'value'], {}), "(self, 'network_interfaces', value)\n", (158214, 158249), False, 'import pulumi\n'), ((158412, 158438), 'pulumi.get', 'pulumi.get', (['self', '"""status"""'], {}), "(self, 'status')\n", (158422, 158438), False, 'import pulumi\n'), ((158525, 158558), 'pulumi.set', 'pulumi.set', (['self', '"""status"""', 'value'], {}), "(self, 'status', value)\n", (158535, 158558), False, 'import pulumi\n'), ((158696, 158733), 'pulumi.get', 'pulumi.get', (['self', '"""status_changed_at"""'], {}), "(self, 'status_changed_at')\n", (158706, 158733), False, 'import pulumi\n'), ((158842, 158886), 'pulumi.set', 'pulumi.set', (['self', '"""status_changed_at"""', 'value'], {}), "(self, 'status_changed_at', value)\n", (158852, 158886), False, 'import pulumi\n'), ((159087, 159121), 'pulumi.get', 'pulumi.get', (['self', '"""status_message"""'], {}), "(self, 'status_message')\n", (159097, 159121), False, 'import pulumi\n'), ((159224, 159265), 'pulumi.set', 'pulumi.set', (['self', '"""status_message"""', 'value'], {}), "(self, 'status_message', value)\n", (159234, 159265), False, 'import pulumi\n'), ((159476, 159503), 'pulumi.get', 'pulumi.get', (['self', '"""zone_id"""'], {}), "(self, 'zone_id')\n", (159486, 159503), False, 'import pulumi\n'), ((159592, 159626), 'pulumi.set', 'pulumi.set', (['self', '"""zone_id"""', 'value'], {}), "(self, 'zone_id', value)\n", (159602, 159626), False, 'import pulumi\n'), ((162398, 162423), 'pulumi.get', 'pulumi.get', (['self', '"""index"""'], {}), "(self, 'index')\n", (162408, 162423), False, 'import pulumi\n'), ((162508, 162540), 'pulumi.set', 'pulumi.set', (['self', '"""index"""', 'value'], {}), "(self, 'index', value)\n", (162518, 162540), False, 'import pulumi\n'), ((162727, 162757), 'pulumi.get', 'pulumi.get', (['self', '"""ip_address"""'], {}), "(self, 'ip_address')\n", (162737, 162757), False, 'import pulumi\n'), ((162852, 162889), 'pulumi.set', 'pulumi.set', (['self', '"""ip_address"""', 'value'], {}), "(self, 'ip_address', value)\n", (162862, 162889), False, 'import pulumi\n'), ((163081, 163105), 'pulumi.get', 'pulumi.get', (['self', '"""ipv4"""'], {}), "(self, 'ipv4')\n", (163091, 163105), False, 'import pulumi\n'), ((163189, 163220), 'pulumi.set', 'pulumi.set', (['self', '"""ipv4"""', 'value'], {}), "(self, 'ipv4', value)\n", (163199, 163220), False, 'import pulumi\n'), ((163322, 163346), 'pulumi.get', 'pulumi.get', (['self', '"""ipv6"""'], {}), "(self, 'ipv6')\n", (163332, 163346), False, 'import pulumi\n'), ((163430, 163461), 'pulumi.set', 'pulumi.set', (['self', '"""ipv6"""', 'value'], {}), "(self, 'ipv6', value)\n", (163440, 163461), False, 'import pulumi\n'), ((163654, 163686), 'pulumi.get', 'pulumi.get', (['self', '"""ipv6_address"""'], {}), "(self, 'ipv6_address')\n", (163664, 163686), False, 'import pulumi\n'), ((163785, 163824), 'pulumi.set', 'pulumi.set', (['self', '"""ipv6_address"""', 'value'], {}), "(self, 'ipv6_address', value)\n", (163795, 163824), False, 'import pulumi\n'), ((164034, 164065), 'pulumi.get', 'pulumi.get', (['self', '"""mac_address"""'], {}), "(self, 'mac_address')\n", (164044, 164065), False, 'import pulumi\n'), ((164162, 164200), 'pulumi.set', 'pulumi.set', (['self', '"""mac_address"""', 'value'], {}), "(self, 'mac_address', value)\n", (164172, 164200), False, 'import pulumi\n'), ((164353, 164376), 'pulumi.get', 'pulumi.get', (['self', '"""nat"""'], {}), "(self, 'nat')\n", (164363, 164376), False, 'import pulumi\n'), ((164458, 164488), 'pulumi.set', 'pulumi.set', (['self', '"""nat"""', 'value'], {}), "(self, 'nat', value)\n", (164468, 164488), False, 'import pulumi\n'), ((164743, 164777), 'pulumi.get', 'pulumi.get', (['self', '"""nat_ip_address"""'], {}), "(self, 'nat_ip_address')\n", (164753, 164777), False, 'import pulumi\n'), ((164880, 164921), 'pulumi.set', 'pulumi.set', (['self', '"""nat_ip_address"""', 'value'], {}), "(self, 'nat_ip_address', value)\n", (164890, 164921), False, 'import pulumi\n'), ((165124, 165158), 'pulumi.get', 'pulumi.get', (['self', '"""nat_ip_version"""'], {}), "(self, 'nat_ip_version')\n", (165134, 165158), False, 'import pulumi\n'), ((165261, 165302), 'pulumi.set', 'pulumi.set', (['self', '"""nat_ip_version"""', 'value'], {}), "(self, 'nat_ip_version', value)\n", (165271, 165302), False, 'import pulumi\n'), ((165580, 165609), 'pulumi.get', 'pulumi.get', (['self', '"""subnet_id"""'], {}), "(self, 'subnet_id')\n", (165590, 165609), False, 'import pulumi\n'), ((165702, 165738), 'pulumi.set', 'pulumi.set', (['self', '"""subnet_id"""', 'value'], {}), "(self, 'subnet_id', value)\n", (165712, 165738), False, 'import pulumi\n'), ((170608, 170652), 'pulumi.set', 'pulumi.set', (['__self__', '"""boot_disk"""', 'boot_disk'], {}), "(__self__, 'boot_disk', boot_disk)\n", (170618, 170652), False, 'import pulumi\n'), ((170661, 170723), 'pulumi.set', 'pulumi.set', (['__self__', '"""network_interfaces"""', 'network_interfaces'], {}), "(__self__, 'network_interfaces', network_interfaces)\n", (170671, 170723), False, 'import pulumi\n'), ((170732, 170776), 'pulumi.set', 'pulumi.set', (['__self__', '"""resources"""', 'resources'], {}), "(__self__, 'resources', resources)\n", (170742, 170776), False, 'import pulumi\n'), ((172140, 172169), 'pulumi.get', 'pulumi.get', (['self', '"""boot_disk"""'], {}), "(self, 'boot_disk')\n", (172150, 172169), False, 'import pulumi\n'), ((172299, 172335), 'pulumi.set', 'pulumi.set', (['self', '"""boot_disk"""', 'value'], {}), "(self, 'boot_disk', value)\n", (172309, 172335), False, 'import pulumi\n'), ((172717, 172755), 'pulumi.get', 'pulumi.get', (['self', '"""network_interfaces"""'], {}), "(self, 'network_interfaces')\n", (172727, 172755), False, 'import pulumi\n'), ((172935, 172980), 'pulumi.set', 'pulumi.set', (['self', '"""network_interfaces"""', 'value'], {}), "(self, 'network_interfaces', value)\n", (172945, 172980), False, 'import pulumi\n'), ((173241, 173270), 'pulumi.get', 'pulumi.get', (['self', '"""resources"""'], {}), "(self, 'resources')\n", (173251, 173270), False, 'import pulumi\n'), ((173401, 173437), 'pulumi.set', 'pulumi.set', (['self', '"""resources"""', 'value'], {}), "(self, 'resources', value)\n", (173411, 173437), False, 'import pulumi\n'), ((173609, 173640), 'pulumi.get', 'pulumi.get', (['self', '"""description"""'], {}), "(self, 'description')\n", (173619, 173640), False, 'import pulumi\n'), ((173737, 173775), 'pulumi.set', 'pulumi.set', (['self', '"""description"""', 'value'], {}), "(self, 'description', value)\n", (173747, 173775), False, 'import pulumi\n'), ((174654, 174682), 'pulumi.get', 'pulumi.get', (['self', '"""hostname"""'], {}), "(self, 'hostname')\n", (174664, 174682), False, 'import pulumi\n'), ((174773, 174808), 'pulumi.set', 'pulumi.set', (['self', '"""hostname"""', 'value'], {}), "(self, 'hostname', value)\n", (174783, 174808), False, 'import pulumi\n'), ((174998, 175024), 'pulumi.get', 'pulumi.get', (['self', '"""labels"""'], {}), "(self, 'labels')\n", (175008, 175024), False, 'import pulumi\n'), ((175139, 175172), 'pulumi.set', 'pulumi.set', (['self', '"""labels"""', 'value'], {}), "(self, 'labels', value)\n", (175149, 175172), False, 'import pulumi\n'), ((175415, 175443), 'pulumi.get', 'pulumi.get', (['self', '"""metadata"""'], {}), "(self, 'metadata')\n", (175425, 175443), False, 'import pulumi\n'), ((175562, 175597), 'pulumi.set', 'pulumi.set', (['self', '"""metadata"""', 'value'], {}), "(self, 'metadata', value)\n", (175572, 175597), False, 'import pulumi\n'), ((176184, 176208), 'pulumi.get', 'pulumi.get', (['self', '"""name"""'], {}), "(self, 'name')\n", (176194, 176208), False, 'import pulumi\n'), ((176291, 176322), 'pulumi.set', 'pulumi.set', (['self', '"""name"""', 'value'], {}), "(self, 'name', value)\n", (176301, 176322), False, 'import pulumi\n'), ((176643, 176679), 'pulumi.get', 'pulumi.get', (['self', '"""network_settings"""'], {}), "(self, 'network_settings')\n", (176653, 176679), False, 'import pulumi\n'), ((176863, 176906), 'pulumi.set', 'pulumi.set', (['self', '"""network_settings"""', 'value'], {}), "(self, 'network_settings', value)\n", (176873, 176906), False, 'import pulumi\n'), ((177200, 177236), 'pulumi.get', 'pulumi.get', (['self', '"""placement_policy"""'], {}), "(self, 'placement_policy')\n", (177210, 177236), False, 'import pulumi\n'), ((177397, 177440), 'pulumi.set', 'pulumi.set', (['self', '"""placement_policy"""', 'value'], {}), "(self, 'placement_policy', value)\n", (177407, 177440), False, 'import pulumi\n'), ((177693, 177724), 'pulumi.get', 'pulumi.get', (['self', '"""platform_id"""'], {}), "(self, 'platform_id')\n", (177703, 177724), False, 'import pulumi\n'), ((177821, 177859), 'pulumi.set', 'pulumi.set', (['self', '"""platform_id"""', 'value'], {}), "(self, 'platform_id', value)\n", (177831, 177859), False, 'import pulumi\n'), ((178157, 178194), 'pulumi.get', 'pulumi.get', (['self', '"""scheduling_policy"""'], {}), "(self, 'scheduling_policy')\n", (178167, 178194), False, 'import pulumi\n'), ((178358, 178402), 'pulumi.set', 'pulumi.set', (['self', '"""scheduling_policy"""', 'value'], {}), "(self, 'scheduling_policy', value)\n", (178368, 178402), False, 'import pulumi\n'), ((178723, 178758), 'pulumi.get', 'pulumi.get', (['self', '"""secondary_disks"""'], {}), "(self, 'secondary_disks')\n", (178733, 178758), False, 'import pulumi\n'), ((178939, 178981), 'pulumi.set', 'pulumi.set', (['self', '"""secondary_disks"""', 'value'], {}), "(self, 'secondary_disks', value)\n", (178949, 178981), False, 'import pulumi\n'), ((179213, 179251), 'pulumi.get', 'pulumi.get', (['self', '"""service_account_id"""'], {}), "(self, 'service_account_id')\n", (179223, 179251), False, 'import pulumi\n'), ((179362, 179407), 'pulumi.set', 'pulumi.set', (['self', '"""service_account_id"""', 'value'], {}), "(self, 'service_account_id', value)\n", (179372, 179407), False, 'import pulumi\n'), ((181008, 181039), 'pulumi.get', 'pulumi.get', (['self', '"""device_name"""'], {}), "(self, 'device_name')\n", (181018, 181039), False, 'import pulumi\n'), ((181136, 181174), 'pulumi.set', 'pulumi.set', (['self', '"""device_name"""', 'value'], {}), "(self, 'device_name', value)\n", (181146, 181174), False, 'import pulumi\n'), ((181372, 181399), 'pulumi.get', 'pulumi.get', (['self', '"""disk_id"""'], {}), "(self, 'disk_id')\n", (181382, 181399), False, 'import pulumi\n'), ((181488, 181522), 'pulumi.set', 'pulumi.set', (['self', '"""disk_id"""', 'value'], {}), "(self, 'disk_id', value)\n", (181498, 181522), False, 'import pulumi\n'), ((181846, 181883), 'pulumi.get', 'pulumi.get', (['self', '"""initialize_params"""'], {}), "(self, 'initialize_params')\n", (181856, 181883), False, 'import pulumi\n'), ((182055, 182099), 'pulumi.set', 'pulumi.set', (['self', '"""initialize_params"""', 'value'], {}), "(self, 'initialize_params', value)\n", (182065, 182099), False, 'import pulumi\n'), ((182322, 182346), 'pulumi.get', 'pulumi.get', (['self', '"""mode"""'], {}), "(self, 'mode')\n", (182332, 182346), False, 'import pulumi\n'), ((182429, 182460), 'pulumi.set', 'pulumi.set', (['self', '"""mode"""', 'value'], {}), "(self, 'mode', value)\n", (182439, 182460), False, 'import pulumi\n'), ((183978, 184009), 'pulumi.get', 'pulumi.get', (['self', '"""description"""'], {}), "(self, 'description')\n", (183988, 184009), False, 'import pulumi\n'), ((184106, 184144), 'pulumi.set', 'pulumi.set', (['self', '"""description"""', 'value'], {}), "(self, 'description', value)\n", (184116, 184144), False, 'import pulumi\n'), ((184342, 184370), 'pulumi.get', 'pulumi.get', (['self', '"""image_id"""'], {}), "(self, 'image_id')\n", (184352, 184370), False, 'import pulumi\n'), ((184461, 184496), 'pulumi.set', 'pulumi.set', (['self', '"""image_id"""', 'value'], {}), "(self, 'image_id', value)\n", (184471, 184496), False, 'import pulumi\n'), ((184676, 184700), 'pulumi.get', 'pulumi.get', (['self', '"""size"""'], {}), "(self, 'size')\n", (184686, 184700), False, 'import pulumi\n'), ((184783, 184814), 'pulumi.set', 'pulumi.set', (['self', '"""size"""', 'value'], {}), "(self, 'size', value)\n", (184793, 184814), False, 'import pulumi\n'), ((185016, 185047), 'pulumi.get', 'pulumi.get', (['self', '"""snapshot_id"""'], {}), "(self, 'snapshot_id')\n", (185026, 185047), False, 'import pulumi\n'), ((185144, 185182), 'pulumi.set', 'pulumi.set', (['self', '"""snapshot_id"""', 'value'], {}), "(self, 'snapshot_id', value)\n", (185154, 185182), False, 'import pulumi\n'), ((185386, 185410), 'pulumi.get', 'pulumi.get', (['self', '"""type"""'], {}), "(self, 'type')\n", (185396, 185410), False, 'import pulumi\n'), ((185493, 185524), 'pulumi.set', 'pulumi.set', (['self', '"""type"""', 'value'], {}), "(self, 'type', value)\n", (185503, 185524), False, 'import pulumi\n'), ((189582, 189613), 'pulumi.get', 'pulumi.get', (['self', '"""dns_records"""'], {}), "(self, 'dns_records')\n", (189592, 189613), False, 'import pulumi\n'), ((189798, 189836), 'pulumi.set', 'pulumi.set', (['self', '"""dns_records"""', 'value'], {}), "(self, 'dns_records', value)\n", (189808, 189836), False, 'import pulumi\n'), ((190023, 190053), 'pulumi.get', 'pulumi.get', (['self', '"""ip_address"""'], {}), "(self, 'ip_address')\n", (190033, 190053), False, 'import pulumi\n'), ((190148, 190185), 'pulumi.set', 'pulumi.set', (['self', '"""ip_address"""', 'value'], {}), "(self, 'ip_address', value)\n", (190158, 190185), False, 'import pulumi\n'), ((190377, 190401), 'pulumi.get', 'pulumi.get', (['self', '"""ipv4"""'], {}), "(self, 'ipv4')\n", (190387, 190401), False, 'import pulumi\n'), ((190485, 190516), 'pulumi.set', 'pulumi.set', (['self', '"""ipv4"""', 'value'], {}), "(self, 'ipv4', value)\n", (190495, 190516), False, 'import pulumi\n'), ((190618, 190642), 'pulumi.get', 'pulumi.get', (['self', '"""ipv6"""'], {}), "(self, 'ipv6')\n", (190628, 190642), False, 'import pulumi\n'), ((190726, 190757), 'pulumi.set', 'pulumi.set', (['self', '"""ipv6"""', 'value'], {}), "(self, 'ipv6', value)\n", (190736, 190757), False, 'import pulumi\n'), ((190950, 190982), 'pulumi.get', 'pulumi.get', (['self', '"""ipv6_address"""'], {}), "(self, 'ipv6_address')\n", (190960, 190982), False, 'import pulumi\n'), ((191081, 191120), 'pulumi.set', 'pulumi.set', (['self', '"""ipv6_address"""', 'value'], {}), "(self, 'ipv6_address', value)\n", (191091, 191120), False, 'import pulumi\n'), ((191442, 191478), 'pulumi.get', 'pulumi.get', (['self', '"""ipv6_dns_records"""'], {}), "(self, 'ipv6_dns_records')\n", (191452, 191478), False, 'import pulumi\n'), ((191677, 191720), 'pulumi.set', 'pulumi.set', (['self', '"""ipv6_dns_records"""', 'value'], {}), "(self, 'ipv6_dns_records', value)\n", (191687, 191720), False, 'import pulumi\n'), ((191873, 191896), 'pulumi.get', 'pulumi.get', (['self', '"""nat"""'], {}), "(self, 'nat')\n", (191883, 191896), False, 'import pulumi\n'), ((191978, 192008), 'pulumi.set', 'pulumi.set', (['self', '"""nat"""', 'value'], {}), "(self, 'nat', value)\n", (191988, 192008), False, 'import pulumi\n'), ((192326, 192361), 'pulumi.get', 'pulumi.get', (['self', '"""nat_dns_records"""'], {}), "(self, 'nat_dns_records')\n", (192336, 192361), False, 'import pulumi\n'), ((192557, 192599), 'pulumi.set', 'pulumi.set', (['self', '"""nat_dns_records"""', 'value'], {}), "(self, 'nat_dns_records', value)\n", (192567, 192599), False, 'import pulumi\n'), ((192854, 192888), 'pulumi.get', 'pulumi.get', (['self', '"""nat_ip_address"""'], {}), "(self, 'nat_ip_address')\n", (192864, 192888), False, 'import pulumi\n'), ((192991, 193032), 'pulumi.set', 'pulumi.set', (['self', '"""nat_ip_address"""', 'value'], {}), "(self, 'nat_ip_address', value)\n", (193001, 193032), False, 'import pulumi\n'), ((193212, 193242), 'pulumi.get', 'pulumi.get', (['self', '"""network_id"""'], {}), "(self, 'network_id')\n", (193222, 193242), False, 'import pulumi\n'), ((193337, 193374), 'pulumi.set', 'pulumi.set', (['self', '"""network_id"""', 'value'], {}), "(self, 'network_id', value)\n", (193347, 193374), False, 'import pulumi\n'), ((193612, 193650), 'pulumi.get', 'pulumi.get', (['self', '"""security_group_ids"""'], {}), "(self, 'security_group_ids')\n", (193622, 193650), False, 'import pulumi\n'), ((193785, 193830), 'pulumi.set', 'pulumi.set', (['self', '"""security_group_ids"""', 'value'], {}), "(self, 'security_group_ids', value)\n", (193795, 193830), False, 'import pulumi\n'), ((194062, 194092), 'pulumi.get', 'pulumi.get', (['self', '"""subnet_ids"""'], {}), "(self, 'subnet_ids')\n", (194072, 194092), False, 'import pulumi\n'), ((194211, 194248), 'pulumi.set', 'pulumi.set', (['self', '"""subnet_ids"""', 'value'], {}), "(self, 'subnet_ids', value)\n", (194221, 194248), False, 'import pulumi\n'), ((194944, 194978), 'pulumi.set', 'pulumi.set', (['__self__', '"""fqdn"""', 'fqdn'], {}), "(__self__, 'fqdn', fqdn)\n", (194954, 194978), False, 'import pulumi\n'), ((195388, 195412), 'pulumi.get', 'pulumi.get', (['self', '"""fqdn"""'], {}), "(self, 'fqdn')\n", (195398, 195412), False, 'import pulumi\n'), ((195485, 195516), 'pulumi.set', 'pulumi.set', (['self', '"""fqdn"""', 'value'], {}), "(self, 'fqdn', value)\n", (195495, 195516), False, 'import pulumi\n'), ((195719, 195750), 'pulumi.get', 'pulumi.get', (['self', '"""dns_zone_id"""'], {}), "(self, 'dns_zone_id')\n", (195729, 195750), False, 'import pulumi\n'), ((195847, 195885), 'pulumi.set', 'pulumi.set', (['self', '"""dns_zone_id"""', 'value'], {}), "(self, 'dns_zone_id', value)\n", (195857, 195885), False, 'import pulumi\n'), ((196064, 196087), 'pulumi.get', 'pulumi.get', (['self', '"""ptr"""'], {}), "(self, 'ptr')\n", (196074, 196087), False, 'import pulumi\n'), ((196169, 196199), 'pulumi.set', 'pulumi.set', (['self', '"""ptr"""', 'value'], {}), "(self, 'ptr', value)\n", (196179, 196199), False, 'import pulumi\n'), ((196347, 196370), 'pulumi.get', 'pulumi.get', (['self', '"""ttl"""'], {}), "(self, 'ttl')\n", (196357, 196370), False, 'import pulumi\n'), ((196451, 196481), 'pulumi.set', 'pulumi.set', (['self', '"""ttl"""', 'value'], {}), "(self, 'ttl', value)\n", (196461, 196481), False, 'import pulumi\n'), ((197181, 197215), 'pulumi.set', 'pulumi.set', (['__self__', '"""fqdn"""', 'fqdn'], {}), "(__self__, 'fqdn', fqdn)\n", (197191, 197215), False, 'import pulumi\n'), ((197625, 197649), 'pulumi.get', 'pulumi.get', (['self', '"""fqdn"""'], {}), "(self, 'fqdn')\n", (197635, 197649), False, 'import pulumi\n'), ((197722, 197753), 'pulumi.set', 'pulumi.set', (['self', '"""fqdn"""', 'value'], {}), "(self, 'fqdn', value)\n", (197732, 197753), False, 'import pulumi\n'), ((197956, 197987), 'pulumi.get', 'pulumi.get', (['self', '"""dns_zone_id"""'], {}), "(self, 'dns_zone_id')\n", (197966, 197987), False, 'import pulumi\n'), ((198084, 198122), 'pulumi.set', 'pulumi.set', (['self', '"""dns_zone_id"""', 'value'], {}), "(self, 'dns_zone_id', value)\n", (198094, 198122), False, 'import pulumi\n'), ((198301, 198324), 'pulumi.get', 'pulumi.get', (['self', '"""ptr"""'], {}), "(self, 'ptr')\n", (198311, 198324), False, 'import pulumi\n'), ((198406, 198436), 'pulumi.set', 'pulumi.set', (['self', '"""ptr"""', 'value'], {}), "(self, 'ptr', value)\n", (198416, 198436), False, 'import pulumi\n'), ((198584, 198607), 'pulumi.get', 'pulumi.get', (['self', '"""ttl"""'], {}), "(self, 'ttl')\n", (198594, 198607), False, 'import pulumi\n'), ((198688, 198718), 'pulumi.set', 'pulumi.set', (['self', '"""ttl"""', 'value'], {}), "(self, 'ttl', value)\n", (198698, 198718), False, 'import pulumi\n'), ((199417, 199451), 'pulumi.set', 'pulumi.set', (['__self__', '"""fqdn"""', 'fqdn'], {}), "(__self__, 'fqdn', fqdn)\n", (199427, 199451), False, 'import pulumi\n'), ((199861, 199885), 'pulumi.get', 'pulumi.get', (['self', '"""fqdn"""'], {}), "(self, 'fqdn')\n", (199871, 199885), False, 'import pulumi\n'), ((199958, 199989), 'pulumi.set', 'pulumi.set', (['self', '"""fqdn"""', 'value'], {}), "(self, 'fqdn', value)\n", (199968, 199989), False, 'import pulumi\n'), ((200192, 200223), 'pulumi.get', 'pulumi.get', (['self', '"""dns_zone_id"""'], {}), "(self, 'dns_zone_id')\n", (200202, 200223), False, 'import pulumi\n'), ((200320, 200358), 'pulumi.set', 'pulumi.set', (['self', '"""dns_zone_id"""', 'value'], {}), "(self, 'dns_zone_id', value)\n", (200330, 200358), False, 'import pulumi\n'), ((200537, 200560), 'pulumi.get', 'pulumi.get', (['self', '"""ptr"""'], {}), "(self, 'ptr')\n", (200547, 200560), False, 'import pulumi\n'), ((200642, 200672), 'pulumi.set', 'pulumi.set', (['self', '"""ptr"""', 'value'], {}), "(self, 'ptr', value)\n", (200652, 200672), False, 'import pulumi\n'), ((200820, 200843), 'pulumi.get', 'pulumi.get', (['self', '"""ttl"""'], {}), "(self, 'ttl')\n", (200830, 200843), False, 'import pulumi\n'), ((200924, 200954), 'pulumi.set', 'pulumi.set', (['self', '"""ttl"""', 'value'], {}), "(self, 'ttl', value)\n", (200934, 200954), False, 'import pulumi\n'), ((201541, 201565), 'pulumi.get', 'pulumi.get', (['self', '"""type"""'], {}), "(self, 'type')\n", (201551, 201565), False, 'import pulumi\n'), ((201648, 201679), 'pulumi.set', 'pulumi.set', (['self', '"""type"""', 'value'], {}), "(self, 'type', value)\n", (201658, 201679), False, 'import pulumi\n'), ((202004, 202066), 'pulumi.set', 'pulumi.set', (['__self__', '"""placement_group_id"""', 'placement_group_id'], {}), "(__self__, 'placement_group_id', placement_group_id)\n", (202014, 202066), False, 'import pulumi\n'), ((202296, 202334), 'pulumi.get', 'pulumi.get', (['self', '"""placement_group_id"""'], {}), "(self, 'placement_group_id')\n", (202306, 202334), False, 'import pulumi\n'), ((202435, 202480), 'pulumi.set', 'pulumi.set', (['self', '"""placement_group_id"""', 'value'], {}), "(self, 'placement_group_id', value)\n", (202445, 202480), False, 'import pulumi\n'), ((203097, 203133), 'pulumi.set', 'pulumi.set', (['__self__', '"""cores"""', 'cores'], {}), "(__self__, 'cores', cores)\n", (203107, 203133), False, 'import pulumi\n'), ((203142, 203180), 'pulumi.set', 'pulumi.set', (['__self__', '"""memory"""', 'memory'], {}), "(__self__, 'memory', memory)\n", (203152, 203180), False, 'import pulumi\n'), ((203525, 203550), 'pulumi.get', 'pulumi.get', (['self', '"""cores"""'], {}), "(self, 'cores')\n", (203535, 203550), False, 'import pulumi\n'), ((203625, 203657), 'pulumi.set', 'pulumi.set', (['self', '"""cores"""', 'value'], {}), "(self, 'cores', value)\n", (203635, 203657), False, 'import pulumi\n'), ((203807, 203833), 'pulumi.get', 'pulumi.get', (['self', '"""memory"""'], {}), "(self, 'memory')\n", (203817, 203833), False, 'import pulumi\n'), ((203912, 203945), 'pulumi.set', 'pulumi.set', (['self', '"""memory"""', 'value'], {}), "(self, 'memory', value)\n", (203922, 203945), False, 'import pulumi\n'), ((204171, 204204), 'pulumi.get', 'pulumi.get', (['self', '"""core_fraction"""'], {}), "(self, 'core_fraction')\n", (204181, 204204), False, 'import pulumi\n'), ((204305, 204345), 'pulumi.set', 'pulumi.set', (['self', '"""core_fraction"""', 'value'], {}), "(self, 'core_fraction', value)\n", (204315, 204345), False, 'import pulumi\n'), ((204446, 204470), 'pulumi.get', 'pulumi.get', (['self', '"""gpus"""'], {}), "(self, 'gpus')\n", (204456, 204470), False, 'import pulumi\n'), ((204553, 204584), 'pulumi.set', 'pulumi.set', (['self', '"""gpus"""', 'value'], {}), "(self, 'gpus', value)\n", (204563, 204584), False, 'import pulumi\n'), ((205198, 205229), 'pulumi.get', 'pulumi.get', (['self', '"""preemptible"""'], {}), "(self, 'preemptible')\n", (205208, 205229), False, 'import pulumi\n'), ((205327, 205365), 'pulumi.set', 'pulumi.set', (['self', '"""preemptible"""', 'value'], {}), "(self, 'preemptible', value)\n", (205337, 205365), False, 'import pulumi\n'), ((206981, 207012), 'pulumi.get', 'pulumi.get', (['self', '"""device_name"""'], {}), "(self, 'device_name')\n", (206991, 207012), False, 'import pulumi\n'), ((207109, 207147), 'pulumi.set', 'pulumi.set', (['self', '"""device_name"""', 'value'], {}), "(self, 'device_name', value)\n", (207119, 207147), False, 'import pulumi\n'), ((207345, 207372), 'pulumi.get', 'pulumi.get', (['self', '"""disk_id"""'], {}), "(self, 'disk_id')\n", (207355, 207372), False, 'import pulumi\n'), ((207461, 207495), 'pulumi.set', 'pulumi.set', (['self', '"""disk_id"""', 'value'], {}), "(self, 'disk_id', value)\n", (207471, 207495), False, 'import pulumi\n'), ((207824, 207861), 'pulumi.get', 'pulumi.get', (['self', '"""initialize_params"""'], {}), "(self, 'initialize_params')\n", (207834, 207861), False, 'import pulumi\n'), ((208038, 208082), 'pulumi.set', 'pulumi.set', (['self', '"""initialize_params"""', 'value'], {}), "(self, 'initialize_params', value)\n", (208048, 208082), False, 'import pulumi\n'), ((208305, 208329), 'pulumi.get', 'pulumi.get', (['self', '"""mode"""'], {}), "(self, 'mode')\n", (208315, 208329), False, 'import pulumi\n'), ((208412, 208443), 'pulumi.set', 'pulumi.set', (['self', '"""mode"""', 'value'], {}), "(self, 'mode', value)\n", (208422, 208443), False, 'import pulumi\n'), ((209966, 209997), 'pulumi.get', 'pulumi.get', (['self', '"""description"""'], {}), "(self, 'description')\n", (209976, 209997), False, 'import pulumi\n'), ((210094, 210132), 'pulumi.set', 'pulumi.set', (['self', '"""description"""', 'value'], {}), "(self, 'description', value)\n", (210104, 210132), False, 'import pulumi\n'), ((210330, 210358), 'pulumi.get', 'pulumi.get', (['self', '"""image_id"""'], {}), "(self, 'image_id')\n", (210340, 210358), False, 'import pulumi\n'), ((210449, 210484), 'pulumi.set', 'pulumi.set', (['self', '"""image_id"""', 'value'], {}), "(self, 'image_id', value)\n", (210459, 210484), False, 'import pulumi\n'), ((210664, 210688), 'pulumi.get', 'pulumi.get', (['self', '"""size"""'], {}), "(self, 'size')\n", (210674, 210688), False, 'import pulumi\n'), ((210771, 210802), 'pulumi.set', 'pulumi.set', (['self', '"""size"""', 'value'], {}), "(self, 'size', value)\n", (210781, 210802), False, 'import pulumi\n'), ((211004, 211035), 'pulumi.get', 'pulumi.get', (['self', '"""snapshot_id"""'], {}), "(self, 'snapshot_id')\n", (211014, 211035), False, 'import pulumi\n'), ((211132, 211170), 'pulumi.set', 'pulumi.set', (['self', '"""snapshot_id"""', 'value'], {}), "(self, 'snapshot_id', value)\n", (211142, 211170), False, 'import pulumi\n'), ((211374, 211398), 'pulumi.get', 'pulumi.get', (['self', '"""type"""'], {}), "(self, 'type')\n", (211384, 211398), False, 'import pulumi\n'), ((211481, 211512), 'pulumi.set', 'pulumi.set', (['self', '"""type"""', 'value'], {}), "(self, 'type', value)\n", (211491, 211512), False, 'import pulumi\n'), ((213821, 213869), 'pulumi.get', 'pulumi.get', (['self', '"""max_opening_traffic_duration"""'], {}), "(self, 'max_opening_traffic_duration')\n", (213831, 213869), False, 'import pulumi\n'), ((214000, 214055), 'pulumi.set', 'pulumi.set', (['self', '"""max_opening_traffic_duration"""', 'value'], {}), "(self, 'max_opening_traffic_duration', value)\n", (214010, 214055), False, 'import pulumi\n'), ((214256, 214290), 'pulumi.get', 'pulumi.get', (['self', '"""status_message"""'], {}), "(self, 'status_message')\n", (214266, 214290), False, 'import pulumi\n'), ((214393, 214434), 'pulumi.set', 'pulumi.set', (['self', '"""status_message"""', 'value'], {}), "(self, 'status_message', value)\n", (214403, 214434), False, 'import pulumi\n'), ((214653, 214697), 'pulumi.get', 'pulumi.get', (['self', '"""target_group_description"""'], {}), "(self, 'target_group_description')\n", (214663, 214697), False, 'import pulumi\n'), ((214820, 214871), 'pulumi.set', 'pulumi.set', (['self', '"""target_group_description"""', 'value'], {}), "(self, 'target_group_description', value)\n", (214830, 214871), False, 'import pulumi\n'), ((215005, 215040), 'pulumi.get', 'pulumi.get', (['self', '"""target_group_id"""'], {}), "(self, 'target_group_id')\n", (215015, 215040), False, 'import pulumi\n'), ((215145, 215187), 'pulumi.set', 'pulumi.set', (['self', '"""target_group_id"""', 'value'], {}), "(self, 'target_group_id', value)\n", (215155, 215187), False, 'import pulumi\n'), ((215421, 215460), 'pulumi.get', 'pulumi.get', (['self', '"""target_group_labels"""'], {}), "(self, 'target_group_labels')\n", (215431, 215460), False, 'import pulumi\n'), ((215601, 215647), 'pulumi.set', 'pulumi.set', (['self', '"""target_group_labels"""', 'value'], {}), "(self, 'target_group_labels', value)\n", (215611, 215647), False, 'import pulumi\n'), ((215847, 215884), 'pulumi.get', 'pulumi.get', (['self', '"""target_group_name"""'], {}), "(self, 'target_group_name')\n", (215857, 215884), False, 'import pulumi\n'), ((215993, 216037), 'pulumi.set', 'pulumi.set', (['self', '"""target_group_name"""', 'value'], {}), "(self, 'target_group_name', value)\n", (216003, 216037), False, 'import pulumi\n'), ((217641, 217671), 'pulumi.get', 'pulumi.get', (['self', '"""auto_scale"""'], {}), "(self, 'auto_scale')\n", (217651, 217671), False, 'import pulumi\n'), ((217809, 217846), 'pulumi.set', 'pulumi.set', (['self', '"""auto_scale"""', 'value'], {}), "(self, 'auto_scale', value)\n", (217819, 217846), False, 'import pulumi\n'), ((218132, 218163), 'pulumi.get', 'pulumi.get', (['self', '"""fixed_scale"""'], {}), "(self, 'fixed_scale')\n", (218142, 218163), False, 'import pulumi\n'), ((218304, 218342), 'pulumi.set', 'pulumi.set', (['self', '"""fixed_scale"""', 'value'], {}), "(self, 'fixed_scale', value)\n", (218314, 218342), False, 'import pulumi\n'), ((218683, 218718), 'pulumi.get', 'pulumi.get', (['self', '"""test_auto_scale"""'], {}), "(self, 'test_auto_scale')\n", (218693, 218718), False, 'import pulumi\n'), ((218870, 218912), 'pulumi.set', 'pulumi.set', (['self', '"""test_auto_scale"""', 'value'], {}), "(self, 'test_auto_scale', value)\n", (218880, 218912), False, 'import pulumi\n'), ((221127, 221177), 'pulumi.set', 'pulumi.set', (['__self__', '"""initial_size"""', 'initial_size'], {}), "(__self__, 'initial_size', initial_size)\n", (221137, 221177), False, 'import pulumi\n'), ((221186, 221252), 'pulumi.set', 'pulumi.set', (['__self__', '"""measurement_duration"""', 'measurement_duration'], {}), "(__self__, 'measurement_duration', measurement_duration)\n", (221196, 221252), False, 'import pulumi\n'), ((222118, 222150), 'pulumi.get', 'pulumi.get', (['self', '"""initial_size"""'], {}), "(self, 'initial_size')\n", (222128, 222150), False, 'import pulumi\n'), ((222239, 222278), 'pulumi.set', 'pulumi.set', (['self', '"""initial_size"""', 'value'], {}), "(self, 'initial_size', value)\n", (222249, 222278), False, 'import pulumi\n'), ((222695, 222735), 'pulumi.get', 'pulumi.get', (['self', '"""measurement_duration"""'], {}), "(self, 'measurement_duration')\n", (222705, 222735), False, 'import pulumi\n'), ((222840, 222887), 'pulumi.set', 'pulumi.set', (['self', '"""measurement_duration"""', 'value'], {}), "(self, 'measurement_duration', value)\n", (222850, 222887), False, 'import pulumi\n'), ((223092, 223134), 'pulumi.get', 'pulumi.get', (['self', '"""cpu_utilization_target"""'], {}), "(self, 'cpu_utilization_target')\n", (223102, 223134), False, 'import pulumi\n'), ((223255, 223304), 'pulumi.set', 'pulumi.set', (['self', '"""cpu_utilization_target"""', 'value'], {}), "(self, 'cpu_utilization_target', value)\n", (223265, 223304), False, 'import pulumi\n'), ((223601, 223633), 'pulumi.get', 'pulumi.get', (['self', '"""custom_rules"""'], {}), "(self, 'custom_rules')\n", (223611, 223633), False, 'import pulumi\n'), ((223809, 223848), 'pulumi.set', 'pulumi.set', (['self', '"""custom_rules"""', 'value'], {}), "(self, 'custom_rules', value)\n", (223819, 223848), False, 'import pulumi\n'), ((224054, 224082), 'pulumi.get', 'pulumi.get', (['self', '"""max_size"""'], {}), "(self, 'max_size')\n", (224064, 224082), False, 'import pulumi\n'), ((224173, 224208), 'pulumi.set', 'pulumi.set', (['self', '"""max_size"""', 'value'], {}), "(self, 'max_size', value)\n", (224183, 224208), False, 'import pulumi\n'), ((224440, 224473), 'pulumi.get', 'pulumi.get', (['self', '"""min_zone_size"""'], {}), "(self, 'min_zone_size')\n", (224450, 224473), False, 'import pulumi\n'), ((224574, 224614), 'pulumi.set', 'pulumi.set', (['self', '"""min_zone_size"""', 'value'], {}), "(self, 'min_zone_size', value)\n", (224584, 224614), False, 'import pulumi\n'), ((225073, 225115), 'pulumi.get', 'pulumi.get', (['self', '"""stabilization_duration"""'], {}), "(self, 'stabilization_duration')\n", (225083, 225115), False, 'import pulumi\n'), ((225234, 225283), 'pulumi.set', 'pulumi.set', (['self', '"""stabilization_duration"""', 'value'], {}), "(self, 'stabilization_duration', value)\n", (225244, 225283), False, 'import pulumi\n'), ((225613, 225648), 'pulumi.get', 'pulumi.get', (['self', '"""warmup_duration"""'], {}), "(self, 'warmup_duration')\n", (225623, 225648), False, 'import pulumi\n'), ((225753, 225795), 'pulumi.set', 'pulumi.set', (['self', '"""warmup_duration"""', 'value'], {}), "(self, 'warmup_duration', value)\n", (225763, 225795), False, 'import pulumi\n'), ((227436, 227484), 'pulumi.set', 'pulumi.set', (['__self__', '"""metric_name"""', 'metric_name'], {}), "(__self__, 'metric_name', metric_name)\n", (227446, 227484), False, 'import pulumi\n'), ((227493, 227541), 'pulumi.set', 'pulumi.set', (['__self__', '"""metric_type"""', 'metric_type'], {}), "(__self__, 'metric_type', metric_type)\n", (227503, 227541), False, 'import pulumi\n'), ((227550, 227594), 'pulumi.set', 'pulumi.set', (['__self__', '"""rule_type"""', 'rule_type'], {}), "(__self__, 'rule_type', rule_type)\n", (227560, 227594), False, 'import pulumi\n'), ((227603, 227641), 'pulumi.set', 'pulumi.set', (['__self__', '"""target"""', 'target'], {}), "(__self__, 'target', target)\n", (227613, 227641), False, 'import pulumi\n'), ((228068, 228099), 'pulumi.get', 'pulumi.get', (['self', '"""metric_name"""'], {}), "(self, 'metric_name')\n", (228078, 228099), False, 'import pulumi\n'), ((228186, 228224), 'pulumi.set', 'pulumi.set', (['self', '"""metric_name"""', 'value'], {}), "(self, 'metric_name', value)\n", (228196, 228224), False, 'import pulumi\n'), ((228408, 228439), 'pulumi.get', 'pulumi.get', (['self', '"""metric_type"""'], {}), "(self, 'metric_type')\n", (228418, 228439), False, 'import pulumi\n'), ((228526, 228564), 'pulumi.set', 'pulumi.set', (['self', '"""metric_type"""', 'value'], {}), "(self, 'metric_type', value)\n", (228536, 228564), False, 'import pulumi\n'), ((229166, 229195), 'pulumi.get', 'pulumi.get', (['self', '"""rule_type"""'], {}), "(self, 'rule_type')\n", (229176, 229195), False, 'import pulumi\n'), ((229278, 229314), 'pulumi.set', 'pulumi.set', (['self', '"""rule_type"""', 'value'], {}), "(self, 'rule_type', value)\n", (229288, 229314), False, 'import pulumi\n'), ((229468, 229494), 'pulumi.get', 'pulumi.get', (['self', '"""target"""'], {}), "(self, 'target')\n", (229478, 229494), False, 'import pulumi\n'), ((229573, 229606), 'pulumi.set', 'pulumi.set', (['self', '"""target"""', 'value'], {}), "(self, 'target', value)\n", (229583, 229606), False, 'import pulumi\n'), ((229842, 229871), 'pulumi.get', 'pulumi.get', (['self', '"""folder_id"""'], {}), "(self, 'folder_id')\n", (229852, 229871), False, 'import pulumi\n'), ((229964, 230000), 'pulumi.set', 'pulumi.set', (['self', '"""folder_id"""', 'value'], {}), "(self, 'folder_id', value)\n", (229974, 230000), False, 'import pulumi\n'), ((230190, 230216), 'pulumi.get', 'pulumi.get', (['self', '"""labels"""'], {}), "(self, 'labels')\n", (230200, 230216), False, 'import pulumi\n'), ((230331, 230364), 'pulumi.set', 'pulumi.set', (['self', '"""labels"""', 'value'], {}), "(self, 'labels', value)\n", (230341, 230364), False, 'import pulumi\n'), ((230579, 230606), 'pulumi.get', 'pulumi.get', (['self', '"""service"""'], {}), "(self, 'service')\n", (230589, 230606), False, 'import pulumi\n'), ((230695, 230729), 'pulumi.set', 'pulumi.set', (['self', '"""service"""', 'value'], {}), "(self, 'service', value)\n", (230705, 230729), False, 'import pulumi\n'), ((230995, 231029), 'pulumi.set', 'pulumi.set', (['__self__', '"""size"""', 'size'], {}), "(__self__, 'size', size)\n", (231005, 231029), False, 'import pulumi\n'), ((231199, 231223), 'pulumi.get', 'pulumi.get', (['self', '"""size"""'], {}), "(self, 'size')\n", (231209, 231223), False, 'import pulumi\n'), ((231296, 231327), 'pulumi.set', 'pulumi.set', (['self', '"""size"""', 'value'], {}), "(self, 'size', value)\n", (231306, 231327), False, 'import pulumi\n'), ((233554, 233604), 'pulumi.set', 'pulumi.set', (['__self__', '"""initial_size"""', 'initial_size'], {}), "(__self__, 'initial_size', initial_size)\n", (233564, 233604), False, 'import pulumi\n'), ((233613, 233679), 'pulumi.set', 'pulumi.set', (['__self__', '"""measurement_duration"""', 'measurement_duration'], {}), "(__self__, 'measurement_duration', measurement_duration)\n", (233623, 233679), False, 'import pulumi\n'), ((234545, 234577), 'pulumi.get', 'pulumi.get', (['self', '"""initial_size"""'], {}), "(self, 'initial_size')\n", (234555, 234577), False, 'import pulumi\n'), ((234666, 234705), 'pulumi.set', 'pulumi.set', (['self', '"""initial_size"""', 'value'], {}), "(self, 'initial_size', value)\n", (234676, 234705), False, 'import pulumi\n'), ((235122, 235162), 'pulumi.get', 'pulumi.get', (['self', '"""measurement_duration"""'], {}), "(self, 'measurement_duration')\n", (235132, 235162), False, 'import pulumi\n'), ((235267, 235314), 'pulumi.set', 'pulumi.set', (['self', '"""measurement_duration"""', 'value'], {}), "(self, 'measurement_duration', value)\n", (235277, 235314), False, 'import pulumi\n'), ((235519, 235561), 'pulumi.get', 'pulumi.get', (['self', '"""cpu_utilization_target"""'], {}), "(self, 'cpu_utilization_target')\n", (235529, 235561), False, 'import pulumi\n'), ((235682, 235731), 'pulumi.set', 'pulumi.set', (['self', '"""cpu_utilization_target"""', 'value'], {}), "(self, 'cpu_utilization_target', value)\n", (235692, 235731), False, 'import pulumi\n'), ((236032, 236064), 'pulumi.get', 'pulumi.get', (['self', '"""custom_rules"""'], {}), "(self, 'custom_rules')\n", (236042, 236064), False, 'import pulumi\n'), ((236244, 236283), 'pulumi.set', 'pulumi.set', (['self', '"""custom_rules"""', 'value'], {}), "(self, 'custom_rules', value)\n", (236254, 236283), False, 'import pulumi\n'), ((236489, 236517), 'pulumi.get', 'pulumi.get', (['self', '"""max_size"""'], {}), "(self, 'max_size')\n", (236499, 236517), False, 'import pulumi\n'), ((236608, 236643), 'pulumi.set', 'pulumi.set', (['self', '"""max_size"""', 'value'], {}), "(self, 'max_size', value)\n", (236618, 236643), False, 'import pulumi\n'), ((236875, 236908), 'pulumi.get', 'pulumi.get', (['self', '"""min_zone_size"""'], {}), "(self, 'min_zone_size')\n", (236885, 236908), False, 'import pulumi\n'), ((237009, 237049), 'pulumi.set', 'pulumi.set', (['self', '"""min_zone_size"""', 'value'], {}), "(self, 'min_zone_size', value)\n", (237019, 237049), False, 'import pulumi\n'), ((237508, 237550), 'pulumi.get', 'pulumi.get', (['self', '"""stabilization_duration"""'], {}), "(self, 'stabilization_duration')\n", (237518, 237550), False, 'import pulumi\n'), ((237669, 237718), 'pulumi.set', 'pulumi.set', (['self', '"""stabilization_duration"""', 'value'], {}), "(self, 'stabilization_duration', value)\n", (237679, 237718), False, 'import pulumi\n'), ((238048, 238083), 'pulumi.get', 'pulumi.get', (['self', '"""warmup_duration"""'], {}), "(self, 'warmup_duration')\n", (238058, 238083), False, 'import pulumi\n'), ((238188, 238230), 'pulumi.set', 'pulumi.set', (['self', '"""warmup_duration"""', 'value'], {}), "(self, 'warmup_duration', value)\n", (238198, 238230), False, 'import pulumi\n'), ((239875, 239923), 'pulumi.set', 'pulumi.set', (['__self__', '"""metric_name"""', 'metric_name'], {}), "(__self__, 'metric_name', metric_name)\n", (239885, 239923), False, 'import pulumi\n'), ((239932, 239980), 'pulumi.set', 'pulumi.set', (['__self__', '"""metric_type"""', 'metric_type'], {}), "(__self__, 'metric_type', metric_type)\n", (239942, 239980), False, 'import pulumi\n'), ((239989, 240033), 'pulumi.set', 'pulumi.set', (['__self__', '"""rule_type"""', 'rule_type'], {}), "(__self__, 'rule_type', rule_type)\n", (239999, 240033), False, 'import pulumi\n'), ((240042, 240080), 'pulumi.set', 'pulumi.set', (['__self__', '"""target"""', 'target'], {}), "(__self__, 'target', target)\n", (240052, 240080), False, 'import pulumi\n'), ((240507, 240538), 'pulumi.get', 'pulumi.get', (['self', '"""metric_name"""'], {}), "(self, 'metric_name')\n", (240517, 240538), False, 'import pulumi\n'), ((240625, 240663), 'pulumi.set', 'pulumi.set', (['self', '"""metric_name"""', 'value'], {}), "(self, 'metric_name', value)\n", (240635, 240663), False, 'import pulumi\n'), ((240847, 240878), 'pulumi.get', 'pulumi.get', (['self', '"""metric_type"""'], {}), "(self, 'metric_type')\n", (240857, 240878), False, 'import pulumi\n'), ((240965, 241003), 'pulumi.set', 'pulumi.set', (['self', '"""metric_type"""', 'value'], {}), "(self, 'metric_type', value)\n", (240975, 241003), False, 'import pulumi\n'), ((241605, 241634), 'pulumi.get', 'pulumi.get', (['self', '"""rule_type"""'], {}), "(self, 'rule_type')\n", (241615, 241634), False, 'import pulumi\n'), ((241717, 241753), 'pulumi.set', 'pulumi.set', (['self', '"""rule_type"""', 'value'], {}), "(self, 'rule_type', value)\n", (241727, 241753), False, 'import pulumi\n'), ((241907, 241933), 'pulumi.get', 'pulumi.get', (['self', '"""target"""'], {}), "(self, 'target')\n", (241917, 241933), False, 'import pulumi\n'), ((242012, 242045), 'pulumi.set', 'pulumi.set', (['self', '"""target"""', 'value'], {}), "(self, 'target', value)\n", (242022, 242045), False, 'import pulumi\n'), ((242281, 242310), 'pulumi.get', 'pulumi.get', (['self', '"""folder_id"""'], {}), "(self, 'folder_id')\n", (242291, 242310), False, 'import pulumi\n'), ((242403, 242439), 'pulumi.set', 'pulumi.set', (['self', '"""folder_id"""', 'value'], {}), "(self, 'folder_id', value)\n", (242413, 242439), False, 'import pulumi\n'), ((242629, 242655), 'pulumi.get', 'pulumi.get', (['self', '"""labels"""'], {}), "(self, 'labels')\n", (242639, 242655), False, 'import pulumi\n'), ((242770, 242803), 'pulumi.set', 'pulumi.set', (['self', '"""labels"""', 'value'], {}), "(self, 'labels', value)\n", (242780, 242803), False, 'import pulumi\n'), ((243018, 243045), 'pulumi.get', 'pulumi.get', (['self', '"""service"""'], {}), "(self, 'service')\n", (243028, 243045), False, 'import pulumi\n'), ((243134, 243168), 'pulumi.set', 'pulumi.set', (['self', '"""service"""', 'value'], {}), "(self, 'service', value)\n", (243144, 243168), False, 'import pulumi\n'), ((246231, 246275), 'pulumi.set', 'pulumi.set', (['__self__', '"""subnet_id"""', 'subnet_id'], {}), "(__self__, 'subnet_id', subnet_id)\n", (246241, 246275), False, 'import pulumi\n'), ((247801, 247830), 'pulumi.get', 'pulumi.get', (['self', '"""subnet_id"""'], {}), "(self, 'subnet_id')\n", (247811, 247830), False, 'import pulumi\n'), ((247913, 247949), 'pulumi.set', 'pulumi.set', (['self', '"""subnet_id"""', 'value'], {}), "(self, 'subnet_id', value)\n", (247923, 247949), False, 'import pulumi\n'), ((248264, 248295), 'pulumi.get', 'pulumi.get', (['self', '"""dns_records"""'], {}), "(self, 'dns_records')\n", (248274, 248295), False, 'import pulumi\n'), ((248459, 248497), 'pulumi.set', 'pulumi.set', (['self', '"""dns_records"""', 'value'], {}), "(self, 'dns_records', value)\n", (248469, 248497), False, 'import pulumi\n'), ((248599, 248624), 'pulumi.get', 'pulumi.get', (['self', '"""index"""'], {}), "(self, 'index')\n", (248609, 248624), False, 'import pulumi\n'), ((248709, 248741), 'pulumi.set', 'pulumi.set', (['self', '"""index"""', 'value'], {}), "(self, 'index', value)\n", (248719, 248741), False, 'import pulumi\n'), ((249036, 249066), 'pulumi.get', 'pulumi.get', (['self', '"""ip_address"""'], {}), "(self, 'ip_address')\n", (249046, 249066), False, 'import pulumi\n'), ((249161, 249198), 'pulumi.set', 'pulumi.set', (['self', '"""ip_address"""', 'value'], {}), "(self, 'ip_address', value)\n", (249171, 249198), False, 'import pulumi\n'), ((249405, 249429), 'pulumi.get', 'pulumi.get', (['self', '"""ipv4"""'], {}), "(self, 'ipv4')\n", (249415, 249429), False, 'import pulumi\n'), ((249513, 249544), 'pulumi.set', 'pulumi.set', (['self', '"""ipv4"""', 'value'], {}), "(self, 'ipv4', value)\n", (249523, 249544), False, 'import pulumi\n'), ((249809, 249833), 'pulumi.get', 'pulumi.get', (['self', '"""ipv6"""'], {}), "(self, 'ipv6')\n", (249819, 249833), False, 'import pulumi\n'), ((249917, 249948), 'pulumi.set', 'pulumi.set', (['self', '"""ipv6"""', 'value'], {}), "(self, 'ipv6', value)\n", (249927, 249948), False, 'import pulumi\n'), ((250161, 250193), 'pulumi.get', 'pulumi.get', (['self', '"""ipv6_address"""'], {}), "(self, 'ipv6_address')\n", (250171, 250193), False, 'import pulumi\n'), ((250292, 250331), 'pulumi.set', 'pulumi.set', (['self', '"""ipv6_address"""', 'value'], {}), "(self, 'ipv6_address', value)\n", (250302, 250331), False, 'import pulumi\n'), ((250659, 250695), 'pulumi.get', 'pulumi.get', (['self', '"""ipv6_dns_records"""'], {}), "(self, 'ipv6_dns_records')\n", (250669, 250695), False, 'import pulumi\n'), ((250873, 250916), 'pulumi.set', 'pulumi.set', (['self', '"""ipv6_dns_records"""', 'value'], {}), "(self, 'ipv6_dns_records', value)\n", (250883, 250916), False, 'import pulumi\n'), ((251043, 251074), 'pulumi.get', 'pulumi.get', (['self', '"""mac_address"""'], {}), "(self, 'mac_address')\n", (251053, 251074), False, 'import pulumi\n'), ((251171, 251209), 'pulumi.set', 'pulumi.set', (['self', '"""mac_address"""', 'value'], {}), "(self, 'mac_address', value)\n", (251181, 251209), False, 'import pulumi\n'), ((251415, 251438), 'pulumi.get', 'pulumi.get', (['self', '"""nat"""'], {}), "(self, 'nat')\n", (251425, 251438), False, 'import pulumi\n'), ((251520, 251550), 'pulumi.set', 'pulumi.set', (['self', '"""nat"""', 'value'], {}), "(self, 'nat', value)\n", (251530, 251550), False, 'import pulumi\n'), ((251879, 251914), 'pulumi.get', 'pulumi.get', (['self', '"""nat_dns_records"""'], {}), "(self, 'nat_dns_records')\n", (251889, 251914), False, 'import pulumi\n'), ((252089, 252131), 'pulumi.set', 'pulumi.set', (['self', '"""nat_dns_records"""', 'value'], {}), "(self, 'nat_dns_records', value)\n", (252099, 252131), False, 'import pulumi\n'), ((252414, 252448), 'pulumi.get', 'pulumi.get', (['self', '"""nat_ip_address"""'], {}), "(self, 'nat_ip_address')\n", (252424, 252448), False, 'import pulumi\n'), ((252551, 252592), 'pulumi.set', 'pulumi.set', (['self', '"""nat_ip_address"""', 'value'], {}), "(self, 'nat_ip_address', value)\n", (252561, 252592), False, 'import pulumi\n'), ((252724, 252758), 'pulumi.get', 'pulumi.get', (['self', '"""nat_ip_version"""'], {}), "(self, 'nat_ip_version')\n", (252734, 252758), False, 'import pulumi\n'), ((252861, 252902), 'pulumi.set', 'pulumi.set', (['self', '"""nat_ip_version"""', 'value'], {}), "(self, 'nat_ip_version', value)\n", (252871, 252902), False, 'import pulumi\n'), ((253140, 253178), 'pulumi.get', 'pulumi.get', (['self', '"""security_group_ids"""'], {}), "(self, 'security_group_ids')\n", (253150, 253178), False, 'import pulumi\n'), ((253313, 253358), 'pulumi.set', 'pulumi.set', (['self', '"""security_group_ids"""', 'value'], {}), "(self, 'security_group_ids', value)\n", (253323, 253358), False, 'import pulumi\n'), ((254048, 254082), 'pulumi.set', 'pulumi.set', (['__self__', '"""fqdn"""', 'fqdn'], {}), "(__self__, 'fqdn', fqdn)\n", (254058, 254082), False, 'import pulumi\n'), ((254494, 254518), 'pulumi.get', 'pulumi.get', (['self', '"""fqdn"""'], {}), "(self, 'fqdn')\n", (254504, 254518), False, 'import pulumi\n'), ((254591, 254622), 'pulumi.set', 'pulumi.set', (['self', '"""fqdn"""', 'value'], {}), "(self, 'fqdn', value)\n", (254601, 254622), False, 'import pulumi\n'), ((254825, 254856), 'pulumi.get', 'pulumi.get', (['self', '"""dns_zone_id"""'], {}), "(self, 'dns_zone_id')\n", (254835, 254856), False, 'import pulumi\n'), ((254953, 254991), 'pulumi.set', 'pulumi.set', (['self', '"""dns_zone_id"""', 'value'], {}), "(self, 'dns_zone_id', value)\n", (254963, 254991), False, 'import pulumi\n'), ((255172, 255195), 'pulumi.get', 'pulumi.get', (['self', '"""ptr"""'], {}), "(self, 'ptr')\n", (255182, 255195), False, 'import pulumi\n'), ((255277, 255307), 'pulumi.set', 'pulumi.set', (['self', '"""ptr"""', 'value'], {}), "(self, 'ptr', value)\n", (255287, 255307), False, 'import pulumi\n'), ((255466, 255489), 'pulumi.get', 'pulumi.get', (['self', '"""ttl"""'], {}), "(self, 'ttl')\n", (255476, 255489), False, 'import pulumi\n'), ((255570, 255600), 'pulumi.set', 'pulumi.set', (['self', '"""ttl"""', 'value'], {}), "(self, 'ttl', value)\n", (255580, 255600), False, 'import pulumi\n'), ((256294, 256328), 'pulumi.set', 'pulumi.set', (['__self__', '"""fqdn"""', 'fqdn'], {}), "(__self__, 'fqdn', fqdn)\n", (256304, 256328), False, 'import pulumi\n'), ((256740, 256764), 'pulumi.get', 'pulumi.get', (['self', '"""fqdn"""'], {}), "(self, 'fqdn')\n", (256750, 256764), False, 'import pulumi\n'), ((256837, 256868), 'pulumi.set', 'pulumi.set', (['self', '"""fqdn"""', 'value'], {}), "(self, 'fqdn', value)\n", (256847, 256868), False, 'import pulumi\n'), ((257071, 257102), 'pulumi.get', 'pulumi.get', (['self', '"""dns_zone_id"""'], {}), "(self, 'dns_zone_id')\n", (257081, 257102), False, 'import pulumi\n'), ((257199, 257237), 'pulumi.set', 'pulumi.set', (['self', '"""dns_zone_id"""', 'value'], {}), "(self, 'dns_zone_id', value)\n", (257209, 257237), False, 'import pulumi\n'), ((257418, 257441), 'pulumi.get', 'pulumi.get', (['self', '"""ptr"""'], {}), "(self, 'ptr')\n", (257428, 257441), False, 'import pulumi\n'), ((257523, 257553), 'pulumi.set', 'pulumi.set', (['self', '"""ptr"""', 'value'], {}), "(self, 'ptr', value)\n", (257533, 257553), False, 'import pulumi\n'), ((257712, 257735), 'pulumi.get', 'pulumi.get', (['self', '"""ttl"""'], {}), "(self, 'ttl')\n", (257722, 257735), False, 'import pulumi\n'), ((257816, 257846), 'pulumi.set', 'pulumi.set', (['self', '"""ttl"""', 'value'], {}), "(self, 'ttl', value)\n", (257826, 257846), False, 'import pulumi\n'), ((258539, 258573), 'pulumi.set', 'pulumi.set', (['__self__', '"""fqdn"""', 'fqdn'], {}), "(__self__, 'fqdn', fqdn)\n", (258549, 258573), False, 'import pulumi\n'), ((258985, 259009), 'pulumi.get', 'pulumi.get', (['self', '"""fqdn"""'], {}), "(self, 'fqdn')\n", (258995, 259009), False, 'import pulumi\n'), ((259082, 259113), 'pulumi.set', 'pulumi.set', (['self', '"""fqdn"""', 'value'], {}), "(self, 'fqdn', value)\n", (259092, 259113), False, 'import pulumi\n'), ((259316, 259347), 'pulumi.get', 'pulumi.get', (['self', '"""dns_zone_id"""'], {}), "(self, 'dns_zone_id')\n", (259326, 259347), False, 'import pulumi\n'), ((259444, 259482), 'pulumi.set', 'pulumi.set', (['self', '"""dns_zone_id"""', 'value'], {}), "(self, 'dns_zone_id', value)\n", (259454, 259482), False, 'import pulumi\n'), ((259663, 259686), 'pulumi.get', 'pulumi.get', (['self', '"""ptr"""'], {}), "(self, 'ptr')\n", (259673, 259686), False, 'import pulumi\n'), ((259768, 259798), 'pulumi.set', 'pulumi.set', (['self', '"""ptr"""', 'value'], {}), "(self, 'ptr', value)\n", (259778, 259798), False, 'import pulumi\n'), ((259957, 259980), 'pulumi.get', 'pulumi.get', (['self', '"""ttl"""'], {}), "(self, 'ttl')\n", (259967, 259980), False, 'import pulumi\n'), ((260061, 260091), 'pulumi.set', 'pulumi.set', (['self', '"""ttl"""', 'value'], {}), "(self, 'ttl', value)\n", (260071, 260091), False, 'import pulumi\n'), ((260394, 260456), 'pulumi.set', 'pulumi.set', (['__self__', '"""placement_group_id"""', 'placement_group_id'], {}), "(__self__, 'placement_group_id', placement_group_id)\n", (260404, 260456), False, 'import pulumi\n'), ((260685, 260723), 'pulumi.get', 'pulumi.get', (['self', '"""placement_group_id"""'], {}), "(self, 'placement_group_id')\n", (260695, 260723), False, 'import pulumi\n'), ((260824, 260869), 'pulumi.set', 'pulumi.set', (['self', '"""placement_group_id"""', 'value'], {}), "(self, 'placement_group_id', value)\n", (260834, 260869), False, 'import pulumi\n'), ((261453, 261489), 'pulumi.set', 'pulumi.set', (['__self__', '"""cores"""', 'cores'], {}), "(__self__, 'cores', cores)\n", (261463, 261489), False, 'import pulumi\n'), ((261498, 261536), 'pulumi.set', 'pulumi.set', (['__self__', '"""memory"""', 'memory'], {}), "(__self__, 'memory', memory)\n", (261508, 261536), False, 'import pulumi\n'), ((261867, 261892), 'pulumi.get', 'pulumi.get', (['self', '"""cores"""'], {}), "(self, 'cores')\n", (261877, 261892), False, 'import pulumi\n'), ((261967, 261999), 'pulumi.set', 'pulumi.set', (['self', '"""cores"""', 'value'], {}), "(self, 'cores', value)\n", (261977, 261999), False, 'import pulumi\n'), ((262145, 262171), 'pulumi.get', 'pulumi.get', (['self', '"""memory"""'], {}), "(self, 'memory')\n", (262155, 262171), False, 'import pulumi\n'), ((262250, 262283), 'pulumi.set', 'pulumi.set', (['self', '"""memory"""', 'value'], {}), "(self, 'memory', value)\n", (262260, 262283), False, 'import pulumi\n'), ((262515, 262548), 'pulumi.get', 'pulumi.get', (['self', '"""core_fraction"""'], {}), "(self, 'core_fraction')\n", (262525, 262548), False, 'import pulumi\n'), ((262649, 262689), 'pulumi.set', 'pulumi.set', (['self', '"""core_fraction"""', 'value'], {}), "(self, 'core_fraction', value)\n", (262659, 262689), False, 'import pulumi\n'), ((262790, 262814), 'pulumi.get', 'pulumi.get', (['self', '"""gpus"""'], {}), "(self, 'gpus')\n", (262800, 262814), False, 'import pulumi\n'), ((262897, 262928), 'pulumi.set', 'pulumi.set', (['self', '"""gpus"""', 'value'], {}), "(self, 'gpus', value)\n", (262907, 262928), False, 'import pulumi\n'), ((263521, 263552), 'pulumi.get', 'pulumi.get', (['self', '"""preemptible"""'], {}), "(self, 'preemptible')\n", (263531, 263552), False, 'import pulumi\n'), ((263650, 263688), 'pulumi.set', 'pulumi.set', (['self', '"""preemptible"""', 'value'], {}), "(self, 'preemptible', value)\n", (263660, 263688), False, 'import pulumi\n'), ((264559, 264599), 'pulumi.set', 'pulumi.set', (['__self__', '"""disk_id"""', 'disk_id'], {}), "(__self__, 'disk_id', disk_id)\n", (264569, 264599), False, 'import pulumi\n'), ((265059, 265086), 'pulumi.get', 'pulumi.get', (['self', '"""disk_id"""'], {}), "(self, 'disk_id')\n", (265069, 265086), False, 'import pulumi\n'), ((265165, 265199), 'pulumi.set', 'pulumi.set', (['self', '"""disk_id"""', 'value'], {}), "(self, 'disk_id', value)\n", (265175, 265199), False, 'import pulumi\n'), ((265458, 265489), 'pulumi.get', 'pulumi.get', (['self', '"""auto_delete"""'], {}), "(self, 'auto_delete')\n", (265468, 265489), False, 'import pulumi\n'), ((265587, 265625), 'pulumi.set', 'pulumi.set', (['self', '"""auto_delete"""', 'value'], {}), "(self, 'auto_delete', value)\n", (265597, 265625), False, 'import pulumi\n'), ((265867, 265898), 'pulumi.get', 'pulumi.get', (['self', '"""device_name"""'], {}), "(self, 'device_name')\n", (265877, 265898), False, 'import pulumi\n'), ((265995, 266033), 'pulumi.set', 'pulumi.set', (['self', '"""device_name"""', 'value'], {}), "(self, 'device_name', value)\n", (266005, 266033), False, 'import pulumi\n'), ((266256, 266280), 'pulumi.get', 'pulumi.get', (['self', '"""mode"""'], {}), "(self, 'mode')\n", (266266, 266280), False, 'import pulumi\n'), ((266363, 266394), 'pulumi.set', 'pulumi.set', (['self', '"""mode"""', 'value'], {}), "(self, 'mode', value)\n", (266373, 266394), False, 'import pulumi\n'), ((267217, 267275), 'pulumi.set', 'pulumi.set', (['__self__', '"""subcluster_specs"""', 'subcluster_specs'], {}), "(__self__, 'subcluster_specs', subcluster_specs)\n", (267227, 267275), False, 'import pulumi\n'), ((267757, 267793), 'pulumi.get', 'pulumi.get', (['self', '"""subcluster_specs"""'], {}), "(self, 'subcluster_specs')\n", (267767, 267793), False, 'import pulumi\n'), ((267959, 268002), 'pulumi.set', 'pulumi.set', (['self', '"""subcluster_specs"""', 'value'], {}), "(self, 'subcluster_specs', value)\n", (267969, 268002), False, 'import pulumi\n'), ((268237, 268263), 'pulumi.get', 'pulumi.get', (['self', '"""hadoop"""'], {}), "(self, 'hadoop')\n", (268247, 268263), False, 'import pulumi\n'), ((268387, 268420), 'pulumi.set', 'pulumi.set', (['self', '"""hadoop"""', 'value'], {}), "(self, 'hadoop', value)\n", (268397, 268420), False, 'import pulumi\n'), ((268605, 268635), 'pulumi.get', 'pulumi.get', (['self', '"""version_id"""'], {}), "(self, 'version_id')\n", (268615, 268635), False, 'import pulumi\n'), ((268730, 268767), 'pulumi.set', 'pulumi.set', (['self', '"""version_id"""', 'value'], {}), "(self, 'version_id', value)\n", (268740, 268767), False, 'import pulumi\n'), ((270225, 270255), 'pulumi.get', 'pulumi.get', (['self', '"""properties"""'], {}), "(self, 'properties')\n", (270235, 270255), False, 'import pulumi\n'), ((270378, 270415), 'pulumi.set', 'pulumi.set', (['self', '"""properties"""', 'value'], {}), "(self, 'properties', value)\n", (270388, 270415), False, 'import pulumi\n'), ((270622, 270650), 'pulumi.get', 'pulumi.get', (['self', '"""services"""'], {}), "(self, 'services')\n", (270632, 270650), False, 'import pulumi\n'), ((270765, 270800), 'pulumi.set', 'pulumi.set', (['self', '"""services"""', 'value'], {}), "(self, 'services', value)\n", (270775, 270800), False, 'import pulumi\n'), ((271194, 271229), 'pulumi.get', 'pulumi.get', (['self', '"""ssh_public_keys"""'], {}), "(self, 'ssh_public_keys')\n", (271204, 271229), False, 'import pulumi\n'), ((271358, 271400), 'pulumi.set', 'pulumi.set', (['self', '"""ssh_public_keys"""', 'value'], {}), "(self, 'ssh_public_keys', value)\n", (271368, 271400), False, 'import pulumi\n'), ((272879, 272927), 'pulumi.set', 'pulumi.set', (['__self__', '"""hosts_count"""', 'hosts_count'], {}), "(__self__, 'hosts_count', hosts_count)\n", (272889, 272927), False, 'import pulumi\n'), ((272936, 272970), 'pulumi.set', 'pulumi.set', (['__self__', '"""name"""', 'name'], {}), "(__self__, 'name', name)\n", (272946, 272970), False, 'import pulumi\n'), ((272979, 273023), 'pulumi.set', 'pulumi.set', (['__self__', '"""resources"""', 'resources'], {}), "(__self__, 'resources', resources)\n", (272989, 273023), False, 'import pulumi\n'), ((273032, 273066), 'pulumi.set', 'pulumi.set', (['__self__', '"""role"""', 'role'], {}), "(__self__, 'role', role)\n", (273042, 273066), False, 'import pulumi\n'), ((273075, 273119), 'pulumi.set', 'pulumi.set', (['__self__', '"""subnet_id"""', 'subnet_id'], {}), "(__self__, 'subnet_id', subnet_id)\n", (273085, 273119), False, 'import pulumi\n'), ((273501, 273532), 'pulumi.get', 'pulumi.get', (['self', '"""hosts_count"""'], {}), "(self, 'hosts_count')\n", (273511, 273532), False, 'import pulumi\n'), ((273619, 273657), 'pulumi.set', 'pulumi.set', (['self', '"""hosts_count"""', 'value'], {}), "(self, 'hosts_count', value)\n", (273629, 273657), False, 'import pulumi\n'), ((273814, 273838), 'pulumi.get', 'pulumi.get', (['self', '"""name"""'], {}), "(self, 'name')\n", (273824, 273838), False, 'import pulumi\n'), ((273911, 273942), 'pulumi.set', 'pulumi.set', (['self', '"""name"""', 'value'], {}), "(self, 'name', value)\n", (273921, 273942), False, 'import pulumi\n'), ((274221, 274250), 'pulumi.get', 'pulumi.get', (['self', '"""resources"""'], {}), "(self, 'resources')\n", (274231, 274250), False, 'import pulumi\n'), ((274387, 274423), 'pulumi.set', 'pulumi.set', (['self', '"""resources"""', 'value'], {}), "(self, 'resources', value)\n", (274397, 274423), False, 'import pulumi\n'), ((274595, 274619), 'pulumi.get', 'pulumi.get', (['self', '"""role"""'], {}), "(self, 'role')\n", (274605, 274619), False, 'import pulumi\n'), ((274692, 274723), 'pulumi.set', 'pulumi.set', (['self', '"""role"""', 'value'], {}), "(self, 'role', value)\n", (274702, 274723), False, 'import pulumi\n'), ((274999, 275028), 'pulumi.get', 'pulumi.get', (['self', '"""subnet_id"""'], {}), "(self, 'subnet_id')\n", (275009, 275028), False, 'import pulumi\n'), ((275111, 275147), 'pulumi.set', 'pulumi.set', (['self', '"""subnet_id"""', 'value'], {}), "(self, 'subnet_id', value)\n", (275121, 275147), False, 'import pulumi\n'), ((275433, 275471), 'pulumi.get', 'pulumi.get', (['self', '"""autoscaling_config"""'], {}), "(self, 'autoscaling_config')\n", (275443, 275471), False, 'import pulumi\n'), ((275644, 275689), 'pulumi.set', 'pulumi.set', (['self', '"""autoscaling_config"""', 'value'], {}), "(self, 'autoscaling_config', value)\n", (275654, 275689), False, 'import pulumi\n'), ((275862, 275884), 'pulumi.get', 'pulumi.get', (['self', '"""id"""'], {}), "(self, 'id')\n", (275872, 275884), False, 'import pulumi\n'), ((275963, 275992), 'pulumi.set', 'pulumi.set', (['self', '"""id"""', 'value'], {}), "(self, 'id', value)\n", (275973, 275992), False, 'import pulumi\n'), ((278018, 278074), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_hosts_count"""', 'max_hosts_count'], {}), "(__self__, 'max_hosts_count', max_hosts_count)\n", (278028, 278074), False, 'import pulumi\n'), ((278996, 279031), 'pulumi.get', 'pulumi.get', (['self', '"""max_hosts_count"""'], {}), "(self, 'max_hosts_count')\n", (279006, 279031), False, 'import pulumi\n'), ((279126, 279168), 'pulumi.set', 'pulumi.set', (['self', '"""max_hosts_count"""', 'value'], {}), "(self, 'max_hosts_count', value)\n", (279136, 279168), False, 'import pulumi\n'), ((279490, 279532), 'pulumi.get', 'pulumi.get', (['self', '"""cpu_utilization_target"""'], {}), "(self, 'cpu_utilization_target')\n", (279500, 279532), False, 'import pulumi\n'), ((279653, 279702), 'pulumi.set', 'pulumi.set', (['self', '"""cpu_utilization_target"""', 'value'], {}), "(self, 'cpu_utilization_target', value)\n", (279663, 279702), False, 'import pulumi\n'), ((279952, 279992), 'pulumi.get', 'pulumi.get', (['self', '"""decommission_timeout"""'], {}), "(self, 'decommission_timeout')\n", (279962, 279992), False, 'import pulumi\n'), ((280107, 280154), 'pulumi.set', 'pulumi.set', (['self', '"""decommission_timeout"""', 'value'], {}), "(self, 'decommission_timeout', value)\n", (280117, 280154), False, 'import pulumi\n'), ((280379, 280419), 'pulumi.get', 'pulumi.get', (['self', '"""measurement_duration"""'], {}), "(self, 'measurement_duration')\n", (280389, 280419), False, 'import pulumi\n'), ((280534, 280581), 'pulumi.set', 'pulumi.set', (['self', '"""measurement_duration"""', 'value'], {}), "(self, 'measurement_duration', value)\n", (280544, 280581), False, 'import pulumi\n'), ((281039, 281070), 'pulumi.get', 'pulumi.get', (['self', '"""preemptible"""'], {}), "(self, 'preemptible')\n", (281049, 281070), False, 'import pulumi\n'), ((281168, 281206), 'pulumi.set', 'pulumi.set', (['self', '"""preemptible"""', 'value'], {}), "(self, 'preemptible', value)\n", (281178, 281206), False, 'import pulumi\n'), ((281623, 281665), 'pulumi.get', 'pulumi.get', (['self', '"""stabilization_duration"""'], {}), "(self, 'stabilization_duration')\n", (281633, 281665), False, 'import pulumi\n'), ((281784, 281833), 'pulumi.set', 'pulumi.set', (['self', '"""stabilization_duration"""', 'value'], {}), "(self, 'stabilization_duration', value)\n", (281794, 281833), False, 'import pulumi\n'), ((282135, 282170), 'pulumi.get', 'pulumi.get', (['self', '"""warmup_duration"""'], {}), "(self, 'warmup_duration')\n", (282145, 282170), False, 'import pulumi\n'), ((282275, 282317), 'pulumi.set', 'pulumi.set', (['self', '"""warmup_duration"""', 'value'], {}), "(self, 'warmup_duration', value)\n", (282285, 282317), False, 'import pulumi\n'), ((283108, 283152), 'pulumi.set', 'pulumi.set', (['__self__', '"""disk_size"""', 'disk_size'], {}), "(__self__, 'disk_size', disk_size)\n", (283118, 283152), False, 'import pulumi\n'), ((283161, 283223), 'pulumi.set', 'pulumi.set', (['__self__', '"""resource_preset_id"""', 'resource_preset_id'], {}), "(__self__, 'resource_preset_id', resource_preset_id)\n", (283171, 283223), False, 'import pulumi\n'), ((283525, 283554), 'pulumi.get', 'pulumi.get', (['self', '"""disk_size"""'], {}), "(self, 'disk_size')\n", (283535, 283554), False, 'import pulumi\n'), ((283637, 283673), 'pulumi.set', 'pulumi.set', (['self', '"""disk_size"""', 'value'], {}), "(self, 'disk_size', value)\n", (283647, 283673), False, 'import pulumi\n'), ((284027, 284065), 'pulumi.get', 'pulumi.get', (['self', '"""resource_preset_id"""'], {}), "(self, 'resource_preset_id')\n", (284037, 284065), False, 'import pulumi\n'), ((284166, 284211), 'pulumi.set', 'pulumi.set', (['self', '"""resource_preset_id"""', 'value'], {}), "(self, 'resource_preset_id', value)\n", (284176, 284211), False, 'import pulumi\n'), ((284451, 284483), 'pulumi.get', 'pulumi.get', (['self', '"""disk_type_id"""'], {}), "(self, 'disk_type_id')\n", (284461, 284483), False, 'import pulumi\n'), ((284582, 284621), 'pulumi.set', 'pulumi.set', (['self', '"""disk_type_id"""', 'value'], {}), "(self, 'disk_type_id', value)\n", (284592, 284621), False, 'import pulumi\n'), ((284759, 284809), 'pulumi.set', 'pulumi.set', (['__self__', '"""zip_filename"""', 'zip_filename'], {}), "(__self__, 'zip_filename', zip_filename)\n", (284769, 284809), False, 'import pulumi\n'), ((284928, 284960), 'pulumi.get', 'pulumi.get', (['self', '"""zip_filename"""'], {}), "(self, 'zip_filename')\n", (284938, 284960), False, 'import pulumi\n'), ((285049, 285088), 'pulumi.set', 'pulumi.set', (['self', '"""zip_filename"""', 'value'], {}), "(self, 'zip_filename', value)\n", (285059, 285088), False, 'import pulumi\n'), ((285335, 285383), 'pulumi.set', 'pulumi.set', (['__self__', '"""bucket_name"""', 'bucket_name'], {}), "(__self__, 'bucket_name', bucket_name)\n", (285345, 285383), False, 'import pulumi\n'), ((285392, 285440), 'pulumi.set', 'pulumi.set', (['__self__', '"""object_name"""', 'object_name'], {}), "(__self__, 'object_name', object_name)\n", (285402, 285440), False, 'import pulumi\n'), ((285639, 285670), 'pulumi.get', 'pulumi.get', (['self', '"""bucket_name"""'], {}), "(self, 'bucket_name')\n", (285649, 285670), False, 'import pulumi\n'), ((285757, 285795), 'pulumi.set', 'pulumi.set', (['self', '"""bucket_name"""', 'value'], {}), "(self, 'bucket_name', value)\n", (285767, 285795), False, 'import pulumi\n'), ((285912, 285943), 'pulumi.get', 'pulumi.get', (['self', '"""object_name"""'], {}), "(self, 'object_name')\n", (285922, 285943), False, 'import pulumi\n'), ((286030, 286068), 'pulumi.set', 'pulumi.set', (['self', '"""object_name"""', 'value'], {}), "(self, 'object_name', value)\n", (286040, 286068), False, 'import pulumi\n'), ((286171, 286197), 'pulumi.get', 'pulumi.get', (['self', '"""sha256"""'], {}), "(self, 'sha256')\n", (286181, 286197), False, 'import pulumi\n'), ((286284, 286317), 'pulumi.set', 'pulumi.set', (['self', '"""sha256"""', 'value'], {}), "(self, 'sha256', value)\n", (286294, 286317), False, 'import pulumi\n'), ((286607, 286639), 'pulumi.set', 'pulumi.set', (['__self__', '"""tag"""', 'tag'], {}), "(__self__, 'tag', tag)\n", (286617, 286639), False, 'import pulumi\n'), ((286974, 286997), 'pulumi.get', 'pulumi.get', (['self', '"""tag"""'], {}), "(self, 'tag')\n", (286984, 286997), False, 'import pulumi\n'), ((287068, 287098), 'pulumi.set', 'pulumi.set', (['self', '"""tag"""', 'value'], {}), "(self, 'tag', value)\n", (287078, 287098), False, 'import pulumi\n'), ((287242, 287282), 'pulumi.get', 'pulumi.get', (['self', '"""zone_instances_limit"""'], {}), "(self, 'zone_instances_limit')\n", (287252, 287282), False, 'import pulumi\n'), ((287397, 287444), 'pulumi.set', 'pulumi.set', (['self', '"""zone_instances_limit"""', 'value'], {}), "(self, 'zone_instances_limit', value)\n", (287407, 287444), False, 'import pulumi\n'), ((287586, 287625), 'pulumi.get', 'pulumi.get', (['self', '"""zone_requests_limit"""'], {}), "(self, 'zone_requests_limit')\n", (287596, 287625), False, 'import pulumi\n'), ((287738, 287784), 'pulumi.set', 'pulumi.set', (['self', '"""zone_requests_limit"""', 'value'], {}), "(self, 'zone_requests_limit', value)\n", (287748, 287784), False, 'import pulumi\n'), ((287977, 288019), 'pulumi.set', 'pulumi.set', (['__self__', '"""queue_id"""', 'queue_id'], {}), "(__self__, 'queue_id', queue_id)\n", (287987, 288019), False, 'import pulumi\n'), ((288028, 288090), 'pulumi.set', 'pulumi.set', (['__self__', '"""service_account_id"""', 'service_account_id'], {}), "(__self__, 'service_account_id', service_account_id)\n", (288038, 288090), False, 'import pulumi\n'), ((288201, 288229), 'pulumi.get', 'pulumi.get', (['self', '"""queue_id"""'], {}), "(self, 'queue_id')\n", (288211, 288229), False, 'import pulumi\n'), ((288310, 288345), 'pulumi.set', 'pulumi.set', (['self', '"""queue_id"""', 'value'], {}), "(self, 'queue_id', value)\n", (288320, 288345), False, 'import pulumi\n'), ((288475, 288513), 'pulumi.get', 'pulumi.get', (['self', '"""service_account_id"""'], {}), "(self, 'service_account_id')\n", (288485, 288513), False, 'import pulumi\n'), ((288614, 288659), 'pulumi.set', 'pulumi.set', (['self', '"""service_account_id"""', 'value'], {}), "(self, 'service_account_id', value)\n", (288624, 288659), False, 'import pulumi\n'), ((289064, 289094), 'pulumi.set', 'pulumi.set', (['__self__', '"""id"""', 'id'], {}), "(__self__, 'id', id)\n", (289074, 289094), False, 'import pulumi\n'), ((289586, 289608), 'pulumi.get', 'pulumi.get', (['self', '"""id"""'], {}), "(self, 'id')\n", (289596, 289608), False, 'import pulumi\n'), ((289677, 289706), 'pulumi.set', 'pulumi.set', (['self', '"""id"""', 'value'], {}), "(self, 'id', value)\n", (289687, 289706), False, 'import pulumi\n'), ((289839, 289873), 'pulumi.get', 'pulumi.get', (['self', '"""retry_attempts"""'], {}), "(self, 'retry_attempts')\n", (289849, 289873), False, 'import pulumi\n'), ((289976, 290017), 'pulumi.set', 'pulumi.set', (['self', '"""retry_attempts"""', 'value'], {}), "(self, 'retry_attempts', value)\n", (289986, 290017), False, 'import pulumi\n'), ((290150, 290184), 'pulumi.get', 'pulumi.get', (['self', '"""retry_interval"""'], {}), "(self, 'retry_interval')\n", (290160, 290184), False, 'import pulumi\n'), ((290287, 290328), 'pulumi.set', 'pulumi.set', (['self', '"""retry_interval"""', 'value'], {}), "(self, 'retry_interval', value)\n", (290297, 290328), False, 'import pulumi\n'), ((290468, 290506), 'pulumi.get', 'pulumi.get', (['self', '"""service_account_id"""'], {}), "(self, 'service_account_id')\n", (290478, 290506), False, 'import pulumi\n'), ((290617, 290662), 'pulumi.set', 'pulumi.set', (['self', '"""service_account_id"""', 'value'], {}), "(self, 'service_account_id', value)\n", (290627, 290662), False, 'import pulumi\n'), ((290762, 290785), 'pulumi.get', 'pulumi.get', (['self', '"""tag"""'], {}), "(self, 'tag')\n", (290772, 290785), False, 'import pulumi\n'), ((290866, 290896), 'pulumi.set', 'pulumi.set', (['self', '"""tag"""', 'value'], {}), "(self, 'tag', value)\n", (290876, 290896), False, 'import pulumi\n'), ((291160, 291208), 'pulumi.set', 'pulumi.set', (['__self__', '"""registry_id"""', 'registry_id'], {}), "(__self__, 'registry_id', registry_id)\n", (291170, 291208), False, 'import pulumi\n'), ((291495, 291526), 'pulumi.get', 'pulumi.get', (['self', '"""registry_id"""'], {}), "(self, 'registry_id')\n", (291505, 291526), False, 'import pulumi\n'), ((291613, 291651), 'pulumi.set', 'pulumi.set', (['self', '"""registry_id"""', 'value'], {}), "(self, 'registry_id', value)\n", (291623, 291651), False, 'import pulumi\n'), ((291774, 291803), 'pulumi.get', 'pulumi.get', (['self', '"""device_id"""'], {}), "(self, 'device_id')\n", (291784, 291803), False, 'import pulumi\n'), ((291896, 291932), 'pulumi.set', 'pulumi.set', (['self', '"""device_id"""', 'value'], {}), "(self, 'device_id', value)\n", (291906, 291932), False, 'import pulumi\n'), ((292034, 292059), 'pulumi.get', 'pulumi.get', (['self', '"""topic"""'], {}), "(self, 'topic')\n", (292044, 292059), False, 'import pulumi\n'), ((292144, 292176), 'pulumi.set', 'pulumi.set', (['self', '"""topic"""', 'value'], {}), "(self, 'topic', value)\n", (292154, 292176), False, 'import pulumi\n'), ((292462, 292512), 'pulumi.set', 'pulumi.set', (['__self__', '"""batch_cutoff"""', 'batch_cutoff'], {}), "(__self__, 'batch_cutoff', batch_cutoff)\n", (292472, 292512), False, 'import pulumi\n'), ((292521, 292573), 'pulumi.set', 'pulumi.set', (['__self__', '"""log_group_ids"""', 'log_group_ids'], {}), "(__self__, 'log_group_ids', log_group_ids)\n", (292531, 292573), False, 'import pulumi\n'), ((292786, 292818), 'pulumi.get', 'pulumi.get', (['self', '"""batch_cutoff"""'], {}), "(self, 'batch_cutoff')\n", (292796, 292818), False, 'import pulumi\n'), ((292907, 292946), 'pulumi.set', 'pulumi.set', (['self', '"""batch_cutoff"""', 'value'], {}), "(self, 'batch_cutoff', value)\n", (292917, 292946), False, 'import pulumi\n'), ((293090, 293123), 'pulumi.get', 'pulumi.get', (['self', '"""log_group_ids"""'], {}), "(self, 'log_group_ids')\n", (293100, 293123), False, 'import pulumi\n'), ((293238, 293278), 'pulumi.set', 'pulumi.set', (['self', '"""log_group_ids"""', 'value'], {}), "(self, 'log_group_ids', value)\n", (293248, 293278), False, 'import pulumi\n'), ((293403, 293433), 'pulumi.get', 'pulumi.get', (['self', '"""batch_size"""'], {}), "(self, 'batch_size')\n", (293413, 293433), False, 'import pulumi\n'), ((293528, 293565), 'pulumi.set', 'pulumi.set', (['self', '"""batch_size"""', 'value'], {}), "(self, 'batch_size', value)\n", (293538, 293565), False, 'import pulumi\n'), ((294039, 294089), 'pulumi.set', 'pulumi.set', (['__self__', '"""batch_cutoff"""', 'batch_cutoff'], {}), "(__self__, 'batch_cutoff', batch_cutoff)\n", (294049, 294089), False, 'import pulumi\n'), ((294098, 294140), 'pulumi.set', 'pulumi.set', (['__self__', '"""group_id"""', 'group_id'], {}), "(__self__, 'group_id', group_id)\n", (294108, 294140), False, 'import pulumi\n'), ((294149, 294187), 'pulumi.set', 'pulumi.set', (['__self__', '"""levels"""', 'levels'], {}), "(__self__, 'levels', levels)\n", (294159, 294187), False, 'import pulumi\n'), ((294196, 294246), 'pulumi.set', 'pulumi.set', (['__self__', '"""resource_ids"""', 'resource_ids'], {}), "(__self__, 'resource_ids', resource_ids)\n", (294206, 294246), False, 'import pulumi\n'), ((294255, 294309), 'pulumi.set', 'pulumi.set', (['__self__', '"""resource_types"""', 'resource_types'], {}), "(__self__, 'resource_types', resource_types)\n", (294265, 294309), False, 'import pulumi\n'), ((294522, 294554), 'pulumi.get', 'pulumi.get', (['self', '"""batch_cutoff"""'], {}), "(self, 'batch_cutoff')\n", (294532, 294554), False, 'import pulumi\n'), ((294643, 294682), 'pulumi.set', 'pulumi.set', (['self', '"""batch_cutoff"""', 'value'], {}), "(self, 'batch_cutoff', value)\n", (294653, 294682), False, 'import pulumi\n'), ((294793, 294821), 'pulumi.get', 'pulumi.get', (['self', '"""group_id"""'], {}), "(self, 'group_id')\n", (294803, 294821), False, 'import pulumi\n'), ((294902, 294937), 'pulumi.set', 'pulumi.set', (['self', '"""group_id"""', 'value'], {}), "(self, 'group_id', value)\n", (294912, 294937), False, 'import pulumi\n'), ((295054, 295080), 'pulumi.get', 'pulumi.get', (['self', '"""levels"""'], {}), "(self, 'levels')\n", (295064, 295080), False, 'import pulumi\n'), ((295181, 295214), 'pulumi.set', 'pulumi.set', (['self', '"""levels"""', 'value'], {}), "(self, 'levels', value)\n", (295191, 295214), False, 'import pulumi\n'), ((295357, 295389), 'pulumi.get', 'pulumi.get', (['self', '"""resource_ids"""'], {}), "(self, 'resource_ids')\n", (295367, 295389), False, 'import pulumi\n'), ((295502, 295541), 'pulumi.set', 'pulumi.set', (['self', '"""resource_ids"""', 'value'], {}), "(self, 'resource_ids', value)\n", (295512, 295541), False, 'import pulumi\n'), ((295688, 295722), 'pulumi.get', 'pulumi.get', (['self', '"""resource_types"""'], {}), "(self, 'resource_types')\n", (295698, 295722), False, 'import pulumi\n'), ((295839, 295880), 'pulumi.set', 'pulumi.set', (['self', '"""resource_types"""', 'value'], {}), "(self, 'resource_types', value)\n", (295849, 295880), False, 'import pulumi\n'), ((296005, 296035), 'pulumi.get', 'pulumi.get', (['self', '"""batch_size"""'], {}), "(self, 'batch_size')\n", (296015, 296035), False, 'import pulumi\n'), ((296130, 296167), 'pulumi.set', 'pulumi.set', (['self', '"""batch_size"""', 'value'], {}), "(self, 'batch_size', value)\n", (296140, 296167), False, 'import pulumi\n'), ((296557, 296607), 'pulumi.set', 'pulumi.set', (['__self__', '"""batch_cutoff"""', 'batch_cutoff'], {}), "(__self__, 'batch_cutoff', batch_cutoff)\n", (296567, 296607), False, 'import pulumi\n'), ((296616, 296658), 'pulumi.set', 'pulumi.set', (['__self__', '"""queue_id"""', 'queue_id'], {}), "(__self__, 'queue_id', queue_id)\n", (296626, 296658), False, 'import pulumi\n'), ((296667, 296729), 'pulumi.set', 'pulumi.set', (['__self__', '"""service_account_id"""', 'service_account_id'], {}), "(__self__, 'service_account_id', service_account_id)\n", (296677, 296729), False, 'import pulumi\n'), ((297060, 297092), 'pulumi.get', 'pulumi.get', (['self', '"""batch_cutoff"""'], {}), "(self, 'batch_cutoff')\n", (297070, 297092), False, 'import pulumi\n'), ((297181, 297220), 'pulumi.set', 'pulumi.set', (['self', '"""batch_cutoff"""', 'value'], {}), "(self, 'batch_cutoff', value)\n", (297191, 297220), False, 'import pulumi\n'), ((297331, 297359), 'pulumi.get', 'pulumi.get', (['self', '"""queue_id"""'], {}), "(self, 'queue_id')\n", (297341, 297359), False, 'import pulumi\n'), ((297440, 297475), 'pulumi.set', 'pulumi.set', (['self', '"""queue_id"""', 'value'], {}), "(self, 'queue_id', value)\n", (297450, 297475), False, 'import pulumi\n'), ((297605, 297643), 'pulumi.get', 'pulumi.get', (['self', '"""service_account_id"""'], {}), "(self, 'service_account_id')\n", (297615, 297643), False, 'import pulumi\n'), ((297744, 297789), 'pulumi.set', 'pulumi.set', (['self', '"""service_account_id"""', 'value'], {}), "(self, 'service_account_id', value)\n", (297754, 297789), False, 'import pulumi\n'), ((297914, 297944), 'pulumi.get', 'pulumi.get', (['self', '"""batch_size"""'], {}), "(self, 'batch_size')\n", (297924, 297944), False, 'import pulumi\n'), ((298039, 298076), 'pulumi.set', 'pulumi.set', (['self', '"""batch_size"""', 'value'], {}), "(self, 'batch_size', value)\n", (298049, 298076), False, 'import pulumi\n'), ((298217, 298255), 'pulumi.get', 'pulumi.get', (['self', '"""visibility_timeout"""'], {}), "(self, 'visibility_timeout')\n", (298227, 298255), False, 'import pulumi\n'), ((298366, 298411), 'pulumi.set', 'pulumi.set', (['self', '"""visibility_timeout"""', 'value'], {}), "(self, 'visibility_timeout', value)\n", (298376, 298411), False, 'import pulumi\n'), ((298867, 298911), 'pulumi.set', 'pulumi.set', (['__self__', '"""bucket_id"""', 'bucket_id'], {}), "(__self__, 'bucket_id', bucket_id)\n", (298877, 298911), False, 'import pulumi\n'), ((299434, 299463), 'pulumi.get', 'pulumi.get', (['self', '"""bucket_id"""'], {}), "(self, 'bucket_id')\n", (299444, 299463), False, 'import pulumi\n'), ((299546, 299582), 'pulumi.set', 'pulumi.set', (['self', '"""bucket_id"""', 'value'], {}), "(self, 'bucket_id', value)\n", (299556, 299582), False, 'import pulumi\n'), ((299686, 299712), 'pulumi.get', 'pulumi.get', (['self', '"""create"""'], {}), "(self, 'create')\n", (299696, 299712), False, 'import pulumi\n'), ((299800, 299833), 'pulumi.set', 'pulumi.set', (['self', '"""create"""', 'value'], {}), "(self, 'create', value)\n", (299810, 299833), False, 'import pulumi\n'), ((299937, 299963), 'pulumi.get', 'pulumi.get', (['self', '"""delete"""'], {}), "(self, 'delete')\n", (299947, 299963), False, 'import pulumi\n'), ((300051, 300084), 'pulumi.set', 'pulumi.set', (['self', '"""delete"""', 'value'], {}), "(self, 'delete', value)\n", (300061, 300084), False, 'import pulumi\n'), ((300187, 300213), 'pulumi.get', 'pulumi.get', (['self', '"""prefix"""'], {}), "(self, 'prefix')\n", (300197, 300213), False, 'import pulumi\n'), ((300300, 300333), 'pulumi.set', 'pulumi.set', (['self', '"""prefix"""', 'value'], {}), "(self, 'prefix', value)\n", (300310, 300333), False, 'import pulumi\n'), ((300436, 300462), 'pulumi.get', 'pulumi.get', (['self', '"""suffix"""'], {}), "(self, 'suffix')\n", (300446, 300462), False, 'import pulumi\n'), ((300549, 300582), 'pulumi.set', 'pulumi.set', (['self', '"""suffix"""', 'value'], {}), "(self, 'suffix', value)\n", (300559, 300582), False, 'import pulumi\n'), ((300686, 300712), 'pulumi.get', 'pulumi.get', (['self', '"""update"""'], {}), "(self, 'update')\n", (300696, 300712), False, 'import pulumi\n'), ((300800, 300833), 'pulumi.set', 'pulumi.set', (['self', '"""update"""', 'value'], {}), "(self, 'update', value)\n", (300810, 300833), False, 'import pulumi\n'), ((300979, 301035), 'pulumi.set', 'pulumi.set', (['__self__', '"""cron_expression"""', 'cron_expression'], {}), "(__self__, 'cron_expression', cron_expression)\n", (300989, 301035), False, 'import pulumi\n'), ((301160, 301195), 'pulumi.get', 'pulumi.get', (['self', '"""cron_expression"""'], {}), "(self, 'cron_expression')\n", (301170, 301195), False, 'import pulumi\n'), ((301290, 301332), 'pulumi.set', 'pulumi.set', (['self', '"""cron_expression"""', 'value'], {}), "(self, 'cron_expression', value)\n", (301300, 301332), False, 'import pulumi\n'), ((301805, 301831), 'pulumi.get', 'pulumi.get', (['self', '"""key_id"""'], {}), "(self, 'key_id')\n", (301815, 301831), False, 'import pulumi\n'), ((301918, 301951), 'pulumi.set', 'pulumi.set', (['self', '"""key_id"""', 'value'], {}), "(self, 'key_id', value)\n", (301928, 301951), False, 'import pulumi\n'), ((306838, 306880), 'pulumi.get', 'pulumi.get', (['self', '"""cluster_ca_certificate"""'], {}), "(self, 'cluster_ca_certificate')\n", (306848, 306880), False, 'import pulumi\n'), ((306999, 307048), 'pulumi.set', 'pulumi.set', (['self', '"""cluster_ca_certificate"""', 'value'], {}), "(self, 'cluster_ca_certificate', value)\n", (307009, 307048), False, 'import pulumi\n'), ((307298, 307337), 'pulumi.get', 'pulumi.get', (['self', '"""external_v4_address"""'], {}), "(self, 'external_v4_address')\n", (307308, 307337), False, 'import pulumi\n'), ((307450, 307496), 'pulumi.set', 'pulumi.set', (['self', '"""external_v4_address"""', 'value'], {}), "(self, 'external_v4_address', value)\n", (307460, 307496), False, 'import pulumi\n'), ((307793, 307833), 'pulumi.get', 'pulumi.get', (['self', '"""external_v4_endpoint"""'], {}), "(self, 'external_v4_endpoint')\n", (307803, 307833), False, 'import pulumi\n'), ((307948, 307995), 'pulumi.set', 'pulumi.set', (['self', '"""external_v4_endpoint"""', 'value'], {}), "(self, 'external_v4_endpoint', value)\n", (307958, 307995), False, 'import pulumi\n'), ((308245, 308284), 'pulumi.get', 'pulumi.get', (['self', '"""internal_v4_address"""'], {}), "(self, 'internal_v4_address')\n", (308255, 308284), False, 'import pulumi\n'), ((308397, 308443), 'pulumi.set', 'pulumi.set', (['self', '"""internal_v4_address"""', 'value'], {}), "(self, 'internal_v4_address', value)\n", (308407, 308443), False, 'import pulumi\n'), ((308711, 308751), 'pulumi.get', 'pulumi.get', (['self', '"""internal_v4_endpoint"""'], {}), "(self, 'internal_v4_endpoint')\n", (308721, 308751), False, 'import pulumi\n'), ((308866, 308913), 'pulumi.set', 'pulumi.set', (['self', '"""internal_v4_endpoint"""', 'value'], {}), "(self, 'internal_v4_endpoint', value)\n", (308876, 308913), False, 'import pulumi\n'), ((309519, 309557), 'pulumi.get', 'pulumi.get', (['self', '"""maintenance_policy"""'], {}), "(self, 'maintenance_policy')\n", (309529, 309557), False, 'import pulumi\n'), ((309711, 309756), 'pulumi.set', 'pulumi.set', (['self', '"""maintenance_policy"""', 'value'], {}), "(self, 'maintenance_policy', value)\n", (309721, 309756), False, 'import pulumi\n'), ((310011, 310040), 'pulumi.get', 'pulumi.get', (['self', '"""public_ip"""'], {}), "(self, 'public_ip')\n", (310021, 310040), False, 'import pulumi\n'), ((310134, 310170), 'pulumi.set', 'pulumi.set', (['self', '"""public_ip"""', 'value'], {}), "(self, 'public_ip', value)\n", (310144, 310170), False, 'import pulumi\n'), ((310456, 310484), 'pulumi.get', 'pulumi.get', (['self', '"""regional"""'], {}), "(self, 'regional')\n", (310466, 310484), False, 'import pulumi\n'), ((310609, 310644), 'pulumi.set', 'pulumi.set', (['self', '"""regional"""', 'value'], {}), "(self, 'regional', value)\n", (310619, 310644), False, 'import pulumi\n'), ((310919, 310957), 'pulumi.get', 'pulumi.get', (['self', '"""security_group_ids"""'], {}), "(self, 'security_group_ids')\n", (310929, 310957), False, 'import pulumi\n'), ((311092, 311137), 'pulumi.set', 'pulumi.set', (['self', '"""security_group_ids"""', 'value'], {}), "(self, 'security_group_ids', value)\n", (311102, 311137), False, 'import pulumi\n'), ((311347, 311374), 'pulumi.get', 'pulumi.get', (['self', '"""version"""'], {}), "(self, 'version')\n", (311357, 311374), False, 'import pulumi\n'), ((311463, 311497), 'pulumi.set', 'pulumi.set', (['self', '"""version"""', 'value'], {}), "(self, 'version', value)\n", (311473, 311497), False, 'import pulumi\n'), ((311776, 311808), 'pulumi.get', 'pulumi.get', (['self', '"""version_info"""'], {}), "(self, 'version_info')\n", (311786, 311808), False, 'import pulumi\n'), ((311944, 311983), 'pulumi.set', 'pulumi.set', (['self', '"""version_info"""', 'value'], {}), "(self, 'version_info', value)\n", (311954, 311983), False, 'import pulumi\n'), ((312255, 312280), 'pulumi.get', 'pulumi.get', (['self', '"""zonal"""'], {}), "(self, 'zonal')\n", (312265, 312280), False, 'import pulumi\n'), ((312396, 312428), 'pulumi.set', 'pulumi.set', (['self', '"""zonal"""', 'value'], {}), "(self, 'zonal', value)\n", (312406, 312428), False, 'import pulumi\n'), ((313434, 313484), 'pulumi.set', 'pulumi.set', (['__self__', '"""auto_upgrade"""', 'auto_upgrade'], {}), "(__self__, 'auto_upgrade', auto_upgrade)\n", (313444, 313484), False, 'import pulumi\n'), ((313874, 313906), 'pulumi.get', 'pulumi.get', (['self', '"""auto_upgrade"""'], {}), "(self, 'auto_upgrade')\n", (313884, 313906), False, 'import pulumi\n'), ((313996, 314035), 'pulumi.set', 'pulumi.set', (['self', '"""auto_upgrade"""', 'value'], {}), "(self, 'auto_upgrade', value)\n", (314006, 314035), False, 'import pulumi\n'), ((314637, 314676), 'pulumi.get', 'pulumi.get', (['self', '"""maintenance_windows"""'], {}), "(self, 'maintenance_windows')\n", (314647, 314676), False, 'import pulumi\n'), ((314873, 314919), 'pulumi.set', 'pulumi.set', (['self', '"""maintenance_windows"""', 'value'], {}), "(self, 'maintenance_windows', value)\n", (314883, 314919), False, 'import pulumi\n'), ((315201, 315243), 'pulumi.set', 'pulumi.set', (['__self__', '"""duration"""', 'duration'], {}), "(__self__, 'duration', duration)\n", (315211, 315243), False, 'import pulumi\n'), ((315252, 315298), 'pulumi.set', 'pulumi.set', (['__self__', '"""start_time"""', 'start_time'], {}), "(__self__, 'start_time', start_time)\n", (315262, 315298), False, 'import pulumi\n'), ((315466, 315494), 'pulumi.get', 'pulumi.get', (['self', '"""duration"""'], {}), "(self, 'duration')\n", (315476, 315494), False, 'import pulumi\n'), ((315575, 315610), 'pulumi.set', 'pulumi.set', (['self', '"""duration"""', 'value'], {}), "(self, 'duration', value)\n", (315585, 315610), False, 'import pulumi\n'), ((315725, 315755), 'pulumi.get', 'pulumi.get', (['self', '"""start_time"""'], {}), "(self, 'start_time')\n", (315735, 315755), False, 'import pulumi\n'), ((315840, 315877), 'pulumi.set', 'pulumi.set', (['self', '"""start_time"""', 'value'], {}), "(self, 'start_time', value)\n", (315850, 315877), False, 'import pulumi\n'), ((315977, 316000), 'pulumi.get', 'pulumi.get', (['self', '"""day"""'], {}), "(self, 'day')\n", (315987, 316000), False, 'import pulumi\n'), ((316081, 316111), 'pulumi.set', 'pulumi.set', (['self', '"""day"""', 'value'], {}), "(self, 'day', value)\n", (316091, 316111), False, 'import pulumi\n'), ((316764, 316802), 'pulumi.set', 'pulumi.set', (['__self__', '"""region"""', 'region'], {}), "(__self__, 'region', region)\n", (316774, 316802), False, 'import pulumi\n'), ((317121, 317147), 'pulumi.get', 'pulumi.get', (['self', '"""region"""'], {}), "(self, 'region')\n", (317131, 317147), False, 'import pulumi\n'), ((317224, 317257), 'pulumi.set', 'pulumi.set', (['self', '"""region"""', 'value'], {}), "(self, 'region', value)\n", (317234, 317257), False, 'import pulumi\n'), ((317558, 317587), 'pulumi.get', 'pulumi.get', (['self', '"""locations"""'], {}), "(self, 'locations')\n", (317568, 317587), False, 'import pulumi\n'), ((317746, 317782), 'pulumi.set', 'pulumi.set', (['self', '"""locations"""', 'value'], {}), "(self, 'locations', value)\n", (317756, 317782), False, 'import pulumi\n'), ((318535, 318564), 'pulumi.get', 'pulumi.get', (['self', '"""subnet_id"""'], {}), "(self, 'subnet_id')\n", (318545, 318564), False, 'import pulumi\n'), ((318657, 318693), 'pulumi.set', 'pulumi.set', (['self', '"""subnet_id"""', 'value'], {}), "(self, 'subnet_id', value)\n", (318667, 318693), False, 'import pulumi\n'), ((318866, 318890), 'pulumi.get', 'pulumi.get', (['self', '"""zone"""'], {}), "(self, 'zone')\n", (318876, 318890), False, 'import pulumi\n'), ((318973, 319004), 'pulumi.set', 'pulumi.set', (['self', '"""zone"""', 'value'], {}), "(self, 'zone', value)\n", (318983, 319004), False, 'import pulumi\n'), ((320939, 320974), 'pulumi.get', 'pulumi.get', (['self', '"""current_version"""'], {}), "(self, 'current_version')\n", (320949, 320974), False, 'import pulumi\n'), ((321079, 321121), 'pulumi.set', 'pulumi.set', (['self', '"""current_version"""', 'value'], {}), "(self, 'current_version', value)\n", (321089, 321121), False, 'import pulumi\n'), ((321541, 321583), 'pulumi.get', 'pulumi.get', (['self', '"""new_revision_available"""'], {}), "(self, 'new_revision_available')\n", (321551, 321583), False, 'import pulumi\n'), ((321703, 321752), 'pulumi.set', 'pulumi.set', (['self', '"""new_revision_available"""', 'value'], {}), "(self, 'new_revision_available', value)\n", (321713, 321752), False, 'import pulumi\n'), ((322072, 322112), 'pulumi.get', 'pulumi.get', (['self', '"""new_revision_summary"""'], {}), "(self, 'new_revision_summary')\n", (322082, 322112), False, 'import pulumi\n'), ((322227, 322274), 'pulumi.set', 'pulumi.set', (['self', '"""new_revision_summary"""', 'value'], {}), "(self, 'new_revision_summary', value)\n", (322237, 322274), False, 'import pulumi\n'), ((322575, 322613), 'pulumi.get', 'pulumi.get', (['self', '"""version_deprecated"""'], {}), "(self, 'version_deprecated')\n", (322585, 322613), False, 'import pulumi\n'), ((322725, 322770), 'pulumi.set', 'pulumi.set', (['self', '"""version_deprecated"""', 'value'], {}), "(self, 'version_deprecated', value)\n", (322735, 322770), False, 'import pulumi\n'), ((323512, 323541), 'pulumi.get', 'pulumi.get', (['self', '"""subnet_id"""'], {}), "(self, 'subnet_id')\n", (323522, 323541), False, 'import pulumi\n'), ((323634, 323670), 'pulumi.set', 'pulumi.set', (['self', '"""subnet_id"""', 'value'], {}), "(self, 'subnet_id', value)\n", (323644, 323670), False, 'import pulumi\n'), ((323843, 323867), 'pulumi.get', 'pulumi.get', (['self', '"""zone"""'], {}), "(self, 'zone')\n", (323853, 323867), False, 'import pulumi\n'), ((323950, 323981), 'pulumi.set', 'pulumi.set', (['self', '"""zone"""', 'value'], {}), "(self, 'zone', value)\n", (323960, 323981), False, 'import pulumi\n'), ((324715, 324741), 'pulumi.get', 'pulumi.get', (['self', '"""cilium"""'], {}), "(self, 'cilium')\n", (324725, 324741), False, 'import pulumi\n'), ((324875, 324908), 'pulumi.set', 'pulumi.set', (['self', '"""cilium"""', 'value'], {}), "(self, 'cilium', value)\n", (324885, 324908), False, 'import pulumi\n'), ((325960, 325989), 'pulumi.get', 'pulumi.get', (['self', '"""locations"""'], {}), "(self, 'locations')\n", (325970, 325989), False, 'import pulumi\n'), ((326152, 326188), 'pulumi.set', 'pulumi.set', (['self', '"""locations"""', 'value'], {}), "(self, 'locations', value)\n", (326162, 326188), False, 'import pulumi\n'), ((327418, 327447), 'pulumi.get', 'pulumi.get', (['self', '"""subnet_id"""'], {}), "(self, 'subnet_id')\n", (327428, 327447), False, 'import pulumi\n'), ((327540, 327576), 'pulumi.set', 'pulumi.set', (['self', '"""subnet_id"""', 'value'], {}), "(self, 'subnet_id', value)\n", (327550, 327576), False, 'import pulumi\n'), ((327783, 327807), 'pulumi.get', 'pulumi.get', (['self', '"""zone"""'], {}), "(self, 'zone')\n", (327793, 327807), False, 'import pulumi\n'), ((327890, 327921), 'pulumi.set', 'pulumi.set', (['self', '"""zone"""', 'value'], {}), "(self, 'zone', value)\n", (327900, 327921), False, 'import pulumi\n'), ((328447, 328499), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_expansion"""', 'max_expansion'], {}), "(__self__, 'max_expansion', max_expansion)\n", (328457, 328499), False, 'import pulumi\n'), ((328508, 328564), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_unavailable"""', 'max_unavailable'], {}), "(__self__, 'max_unavailable', max_unavailable)\n", (328518, 328564), False, 'import pulumi\n'), ((328832, 328865), 'pulumi.get', 'pulumi.get', (['self', '"""max_expansion"""'], {}), "(self, 'max_expansion')\n", (328842, 328865), False, 'import pulumi\n'), ((328956, 328996), 'pulumi.set', 'pulumi.set', (['self', '"""max_expansion"""', 'value'], {}), "(self, 'max_expansion', value)\n", (328966, 328996), False, 'import pulumi\n'), ((329234, 329269), 'pulumi.get', 'pulumi.get', (['self', '"""max_unavailable"""'], {}), "(self, 'max_unavailable')\n", (329244, 329269), False, 'import pulumi\n'), ((329364, 329406), 'pulumi.set', 'pulumi.set', (['self', '"""max_unavailable"""', 'value'], {}), "(self, 'max_unavailable', value)\n", (329374, 329406), False, 'import pulumi\n'), ((333707, 333736), 'pulumi.get', 'pulumi.get', (['self', '"""boot_disk"""'], {}), "(self, 'boot_disk')\n", (333717, 333736), False, 'import pulumi\n'), ((333875, 333911), 'pulumi.set', 'pulumi.set', (['self', '"""boot_disk"""', 'value'], {}), "(self, 'boot_disk', value)\n", (333885, 333911), False, 'import pulumi\n'), ((334514, 334542), 'pulumi.get', 'pulumi.get', (['self', '"""metadata"""'], {}), "(self, 'metadata')\n", (334524, 334542), False, 'import pulumi\n'), ((334661, 334696), 'pulumi.set', 'pulumi.set', (['self', '"""metadata"""', 'value'], {}), "(self, 'metadata', value)\n", (334671, 334696), False, 'import pulumi\n'), ((334896, 334919), 'pulumi.get', 'pulumi.get', (['self', '"""nat"""'], {}), "(self, 'nat')\n", (334906, 334919), False, 'import pulumi\n'), ((335001, 335031), 'pulumi.set', 'pulumi.set', (['self', '"""nat"""', 'value'], {}), "(self, 'nat', value)\n", (335011, 335031), False, 'import pulumi\n'), ((335291, 335336), 'pulumi.get', 'pulumi.get', (['self', '"""network_acceleration_type"""'], {}), "(self, 'network_acceleration_type')\n", (335301, 335336), False, 'import pulumi\n'), ((335461, 335513), 'pulumi.set', 'pulumi.set', (['self', '"""network_acceleration_type"""', 'value'], {}), "(self, 'network_acceleration_type', value)\n", (335471, 335513), False, 'import pulumi\n'), ((335875, 335913), 'pulumi.get', 'pulumi.get', (['self', '"""network_interfaces"""'], {}), "(self, 'network_interfaces')\n", (335885, 335913), False, 'import pulumi\n'), ((336102, 336147), 'pulumi.set', 'pulumi.set', (['self', '"""network_interfaces"""', 'value'], {}), "(self, 'network_interfaces', value)\n", (336112, 336147), False, 'import pulumi\n'), ((336440, 336476), 'pulumi.get', 'pulumi.get', (['self', '"""placement_policy"""'], {}), "(self, 'placement_policy')\n", (336450, 336476), False, 'import pulumi\n'), ((336636, 336679), 'pulumi.set', 'pulumi.set', (['self', '"""placement_policy"""', 'value'], {}), "(self, 'placement_policy', value)\n", (336646, 336679), False, 'import pulumi\n'), ((336922, 336953), 'pulumi.get', 'pulumi.get', (['self', '"""platform_id"""'], {}), "(self, 'platform_id')\n", (336932, 336953), False, 'import pulumi\n'), ((337050, 337088), 'pulumi.set', 'pulumi.set', (['self', '"""platform_id"""', 'value'], {}), "(self, 'platform_id', value)\n", (337060, 337088), False, 'import pulumi\n'), ((337241, 337270), 'pulumi.get', 'pulumi.get', (['self', '"""resources"""'], {}), "(self, 'resources')\n", (337251, 337270), False, 'import pulumi\n'), ((337410, 337446), 'pulumi.set', 'pulumi.set', (['self', '"""resources"""', 'value'], {}), "(self, 'resources', value)\n", (337420, 337446), False, 'import pulumi\n'), ((337761, 337798), 'pulumi.get', 'pulumi.get', (['self', '"""scheduling_policy"""'], {}), "(self, 'scheduling_policy')\n", (337771, 337798), False, 'import pulumi\n'), ((337961, 338005), 'pulumi.set', 'pulumi.set', (['self', '"""scheduling_policy"""', 'value'], {}), "(self, 'scheduling_policy', value)\n", (337971, 338005), False, 'import pulumi\n'), ((338718, 338742), 'pulumi.get', 'pulumi.get', (['self', '"""size"""'], {}), "(self, 'size')\n", (338728, 338742), False, 'import pulumi\n'), ((338825, 338856), 'pulumi.set', 'pulumi.set', (['self', '"""size"""', 'value'], {}), "(self, 'size', value)\n", (338835, 338856), False, 'import pulumi\n'), ((339004, 339028), 'pulumi.get', 'pulumi.get', (['self', '"""type"""'], {}), "(self, 'type')\n", (339014, 339028), False, 'import pulumi\n'), ((339111, 339142), 'pulumi.set', 'pulumi.set', (['self', '"""type"""', 'value'], {}), "(self, 'type', value)\n", (339121, 339142), False, 'import pulumi\n'), ((340232, 340278), 'pulumi.set', 'pulumi.set', (['__self__', '"""subnet_ids"""', 'subnet_ids'], {}), "(__self__, 'subnet_ids', subnet_ids)\n", (340242, 340278), False, 'import pulumi\n'), ((340816, 340846), 'pulumi.get', 'pulumi.get', (['self', '"""subnet_ids"""'], {}), "(self, 'subnet_ids')\n", (340826, 340846), False, 'import pulumi\n'), ((340955, 340992), 'pulumi.set', 'pulumi.set', (['self', '"""subnet_ids"""', 'value'], {}), "(self, 'subnet_ids', value)\n", (340965, 340992), False, 'import pulumi\n'), ((341199, 341223), 'pulumi.get', 'pulumi.get', (['self', '"""ipv4"""'], {}), "(self, 'ipv4')\n", (341209, 341223), False, 'import pulumi\n'), ((341307, 341338), 'pulumi.set', 'pulumi.set', (['self', '"""ipv4"""', 'value'], {}), "(self, 'ipv4', value)\n", (341317, 341338), False, 'import pulumi\n'), ((341595, 341619), 'pulumi.get', 'pulumi.get', (['self', '"""ipv6"""'], {}), "(self, 'ipv6')\n", (341605, 341619), False, 'import pulumi\n'), ((341703, 341734), 'pulumi.set', 'pulumi.set', (['self', '"""ipv6"""', 'value'], {}), "(self, 'ipv6', value)\n", (341713, 341734), False, 'import pulumi\n'), ((341934, 341957), 'pulumi.get', 'pulumi.get', (['self', '"""nat"""'], {}), "(self, 'nat')\n", (341944, 341957), False, 'import pulumi\n'), ((342039, 342069), 'pulumi.set', 'pulumi.set', (['self', '"""nat"""', 'value'], {}), "(self, 'nat', value)\n", (342049, 342069), False, 'import pulumi\n'), ((342307, 342345), 'pulumi.get', 'pulumi.get', (['self', '"""security_group_ids"""'], {}), "(self, 'security_group_ids')\n", (342317, 342345), False, 'import pulumi\n'), ((342480, 342525), 'pulumi.set', 'pulumi.set', (['self', '"""security_group_ids"""', 'value'], {}), "(self, 'security_group_ids', value)\n", (342490, 342525), False, 'import pulumi\n'), ((342849, 342911), 'pulumi.set', 'pulumi.set', (['__self__', '"""placement_group_id"""', 'placement_group_id'], {}), "(__self__, 'placement_group_id', placement_group_id)\n", (342859, 342911), False, 'import pulumi\n'), ((343141, 343179), 'pulumi.get', 'pulumi.get', (['self', '"""placement_group_id"""'], {}), "(self, 'placement_group_id')\n", (343151, 343179), False, 'import pulumi\n'), ((343280, 343325), 'pulumi.set', 'pulumi.set', (['self', '"""placement_group_id"""', 'value'], {}), "(self, 'placement_group_id', value)\n", (343290, 343325), False, 'import pulumi\n'), ((344154, 344187), 'pulumi.get', 'pulumi.get', (['self', '"""core_fraction"""'], {}), "(self, 'core_fraction')\n", (344164, 344187), False, 'import pulumi\n'), ((344288, 344328), 'pulumi.set', 'pulumi.set', (['self', '"""core_fraction"""', 'value'], {}), "(self, 'core_fraction', value)\n", (344298, 344328), False, 'import pulumi\n'), ((344430, 344455), 'pulumi.get', 'pulumi.get', (['self', '"""cores"""'], {}), "(self, 'cores')\n", (344440, 344455), False, 'import pulumi\n'), ((344540, 344572), 'pulumi.set', 'pulumi.set', (['self', '"""cores"""', 'value'], {}), "(self, 'cores', value)\n", (344550, 344572), False, 'import pulumi\n'), ((344673, 344697), 'pulumi.get', 'pulumi.get', (['self', '"""gpus"""'], {}), "(self, 'gpus')\n", (344683, 344697), False, 'import pulumi\n'), ((344780, 344811), 'pulumi.set', 'pulumi.set', (['self', '"""gpus"""', 'value'], {}), "(self, 'gpus', value)\n", (344790, 344811), False, 'import pulumi\n'), ((344916, 344942), 'pulumi.get', 'pulumi.get', (['self', '"""memory"""'], {}), "(self, 'memory')\n", (344926, 344942), False, 'import pulumi\n'), ((345031, 345064), 'pulumi.set', 'pulumi.set', (['self', '"""memory"""', 'value'], {}), "(self, 'memory', value)\n", (345041, 345064), False, 'import pulumi\n'), ((345708, 345739), 'pulumi.get', 'pulumi.get', (['self', '"""preemptible"""'], {}), "(self, 'preemptible')\n", (345718, 345739), False, 'import pulumi\n'), ((345837, 345875), 'pulumi.set', 'pulumi.set', (['self', '"""preemptible"""', 'value'], {}), "(self, 'preemptible', value)\n", (345847, 345875), False, 'import pulumi\n'), ((346833, 346881), 'pulumi.set', 'pulumi.set', (['__self__', '"""auto_repair"""', 'auto_repair'], {}), "(__self__, 'auto_repair', auto_repair)\n", (346843, 346881), False, 'import pulumi\n'), ((346890, 346940), 'pulumi.set', 'pulumi.set', (['__self__', '"""auto_upgrade"""', 'auto_upgrade'], {}), "(__self__, 'auto_upgrade', auto_upgrade)\n", (346900, 346940), False, 'import pulumi\n'), ((347321, 347352), 'pulumi.get', 'pulumi.get', (['self', '"""auto_repair"""'], {}), "(self, 'auto_repair')\n", (347331, 347352), False, 'import pulumi\n'), ((347440, 347478), 'pulumi.set', 'pulumi.set', (['self', '"""auto_repair"""', 'value'], {}), "(self, 'auto_repair', value)\n", (347450, 347478), False, 'import pulumi\n'), ((347740, 347772), 'pulumi.get', 'pulumi.get', (['self', '"""auto_upgrade"""'], {}), "(self, 'auto_upgrade')\n", (347750, 347772), False, 'import pulumi\n'), ((347862, 347901), 'pulumi.set', 'pulumi.set', (['self', '"""auto_upgrade"""', 'value'], {}), "(self, 'auto_upgrade', value)\n", (347872, 347901), False, 'import pulumi\n'), ((348277, 348316), 'pulumi.get', 'pulumi.get', (['self', '"""maintenance_windows"""'], {}), "(self, 'maintenance_windows')\n", (348287, 348316), False, 'import pulumi\n'), ((348509, 348555), 'pulumi.set', 'pulumi.set', (['self', '"""maintenance_windows"""', 'value'], {}), "(self, 'maintenance_windows', value)\n", (348519, 348555), False, 'import pulumi\n'), ((348833, 348875), 'pulumi.set', 'pulumi.set', (['__self__', '"""duration"""', 'duration'], {}), "(__self__, 'duration', duration)\n", (348843, 348875), False, 'import pulumi\n'), ((348884, 348930), 'pulumi.set', 'pulumi.set', (['__self__', '"""start_time"""', 'start_time'], {}), "(__self__, 'start_time', start_time)\n", (348894, 348930), False, 'import pulumi\n'), ((349098, 349126), 'pulumi.get', 'pulumi.get', (['self', '"""duration"""'], {}), "(self, 'duration')\n", (349108, 349126), False, 'import pulumi\n'), ((349207, 349242), 'pulumi.set', 'pulumi.set', (['self', '"""duration"""', 'value'], {}), "(self, 'duration', value)\n", (349217, 349242), False, 'import pulumi\n'), ((349357, 349387), 'pulumi.get', 'pulumi.get', (['self', '"""start_time"""'], {}), "(self, 'start_time')\n", (349367, 349387), False, 'import pulumi\n'), ((349472, 349509), 'pulumi.set', 'pulumi.set', (['self', '"""start_time"""', 'value'], {}), "(self, 'start_time', value)\n", (349482, 349509), False, 'import pulumi\n'), ((349609, 349632), 'pulumi.get', 'pulumi.get', (['self', '"""day"""'], {}), "(self, 'day')\n", (349619, 349632), False, 'import pulumi\n'), ((349713, 349743), 'pulumi.set', 'pulumi.set', (['self', '"""day"""', 'value'], {}), "(self, 'day', value)\n", (349723, 349743), False, 'import pulumi\n'), ((350877, 350907), 'pulumi.get', 'pulumi.get', (['self', '"""auto_scale"""'], {}), "(self, 'auto_scale')\n", (350887, 350907), False, 'import pulumi\n'), ((351044, 351081), 'pulumi.set', 'pulumi.set', (['self', '"""auto_scale"""', 'value'], {}), "(self, 'auto_scale', value)\n", (351054, 351081), False, 'import pulumi\n'), ((351361, 351392), 'pulumi.get', 'pulumi.get', (['self', '"""fixed_scale"""'], {}), "(self, 'fixed_scale')\n", (351371, 351392), False, 'import pulumi\n'), ((351532, 351570), 'pulumi.set', 'pulumi.set', (['self', '"""fixed_scale"""', 'value'], {}), "(self, 'fixed_scale', value)\n", (351542, 351570), False, 'import pulumi\n'), ((352092, 352132), 'pulumi.set', 'pulumi.set', (['__self__', '"""initial"""', 'initial'], {}), "(__self__, 'initial', initial)\n", (352102, 352132), False, 'import pulumi\n'), ((352141, 352173), 'pulumi.set', 'pulumi.set', (['__self__', '"""max"""', 'max'], {}), "(__self__, 'max', max)\n", (352151, 352173), False, 'import pulumi\n'), ((352182, 352214), 'pulumi.set', 'pulumi.set', (['__self__', '"""min"""', 'min'], {}), "(__self__, 'min', min)\n", (352192, 352214), False, 'import pulumi\n'), ((352387, 352414), 'pulumi.get', 'pulumi.get', (['self', '"""initial"""'], {}), "(self, 'initial')\n", (352397, 352414), False, 'import pulumi\n'), ((352493, 352527), 'pulumi.set', 'pulumi.set', (['self', '"""initial"""', 'value'], {}), "(self, 'initial', value)\n", (352503, 352527), False, 'import pulumi\n'), ((352696, 352719), 'pulumi.get', 'pulumi.get', (['self', '"""max"""'], {}), "(self, 'max')\n", (352706, 352719), False, 'import pulumi\n'), ((352790, 352820), 'pulumi.set', 'pulumi.set', (['self', '"""max"""', 'value'], {}), "(self, 'max', value)\n", (352800, 352820), False, 'import pulumi\n'), ((352989, 353012), 'pulumi.get', 'pulumi.get', (['self', '"""min"""'], {}), "(self, 'min')\n", (352999, 353012), False, 'import pulumi\n'), ((353083, 353113), 'pulumi.set', 'pulumi.set', (['self', '"""min"""', 'value'], {}), "(self, 'min', value)\n", (353093, 353113), False, 'import pulumi\n'), ((353634, 353658), 'pulumi.get', 'pulumi.get', (['self', '"""size"""'], {}), "(self, 'size')\n", (353644, 353658), False, 'import pulumi\n'), ((353741, 353772), 'pulumi.set', 'pulumi.set', (['self', '"""size"""', 'value'], {}), "(self, 'size', value)\n", (353751, 353772), False, 'import pulumi\n'), ((355709, 355744), 'pulumi.get', 'pulumi.get', (['self', '"""current_version"""'], {}), "(self, 'current_version')\n", (355719, 355744), False, 'import pulumi\n'), ((355849, 355891), 'pulumi.set', 'pulumi.set', (['self', '"""current_version"""', 'value'], {}), "(self, 'current_version', value)\n", (355859, 355891), False, 'import pulumi\n'), ((356314, 356356), 'pulumi.get', 'pulumi.get', (['self', '"""new_revision_available"""'], {}), "(self, 'new_revision_available')\n", (356324, 356356), False, 'import pulumi\n'), ((356476, 356525), 'pulumi.set', 'pulumi.set', (['self', '"""new_revision_available"""', 'value'], {}), "(self, 'new_revision_available', value)\n", (356486, 356525), False, 'import pulumi\n'), ((356845, 356885), 'pulumi.get', 'pulumi.get', (['self', '"""new_revision_summary"""'], {}), "(self, 'new_revision_summary')\n", (356855, 356885), False, 'import pulumi\n'), ((357000, 357047), 'pulumi.set', 'pulumi.set', (['self', '"""new_revision_summary"""', 'value'], {}), "(self, 'new_revision_summary', value)\n", (357010, 357047), False, 'import pulumi\n'), ((357351, 357389), 'pulumi.get', 'pulumi.get', (['self', '"""version_deprecated"""'], {}), "(self, 'version_deprecated')\n", (357361, 357389), False, 'import pulumi\n'), ((357501, 357546), 'pulumi.set', 'pulumi.set', (['self', '"""version_deprecated"""', 'value'], {}), "(self, 'version_deprecated', value)\n", (357511, 357546), False, 'import pulumi\n'), ((358122, 358172), 'pulumi.set', 'pulumi.set', (['__self__', '"""healthchecks"""', 'healthchecks'], {}), "(__self__, 'healthchecks', healthchecks)\n", (358132, 358172), False, 'import pulumi\n'), ((358181, 358237), 'pulumi.set', 'pulumi.set', (['__self__', '"""target_group_id"""', 'target_group_id'], {}), "(__self__, 'target_group_id', target_group_id)\n", (358191, 358237), False, 'import pulumi\n'), ((358505, 358537), 'pulumi.get', 'pulumi.get', (['self', '"""healthchecks"""'], {}), "(self, 'healthchecks')\n", (358515, 358537), False, 'import pulumi\n'), ((358704, 358743), 'pulumi.set', 'pulumi.set', (['self', '"""healthchecks"""', 'value'], {}), "(self, 'healthchecks', value)\n", (358714, 358743), False, 'import pulumi\n'), ((358923, 358958), 'pulumi.get', 'pulumi.get', (['self', '"""target_group_id"""'], {}), "(self, 'target_group_id')\n", (358933, 358958), False, 'import pulumi\n'), ((359053, 359095), 'pulumi.set', 'pulumi.set', (['self', '"""target_group_id"""', 'value'], {}), "(self, 'target_group_id', value)\n", (359063, 359095), False, 'import pulumi\n'), ((360832, 360866), 'pulumi.set', 'pulumi.set', (['__self__', '"""name"""', 'name'], {}), "(__self__, 'name', name)\n", (360842, 360866), False, 'import pulumi\n'), ((361686, 361710), 'pulumi.get', 'pulumi.get', (['self', '"""name"""'], {}), "(self, 'name')\n", (361696, 361710), False, 'import pulumi\n'), ((361783, 361814), 'pulumi.set', 'pulumi.set', (['self', '"""name"""', 'value'], {}), "(self, 'name', value)\n", (361793, 361814), False, 'import pulumi\n'), ((362082, 362119), 'pulumi.get', 'pulumi.get', (['self', '"""healthy_threshold"""'], {}), "(self, 'healthy_threshold')\n", (362092, 362119), False, 'import pulumi\n'), ((362228, 362272), 'pulumi.set', 'pulumi.set', (['self', '"""healthy_threshold"""', 'value'], {}), "(self, 'healthy_threshold', value)\n", (362238, 362272), False, 'import pulumi\n'), ((362564, 362596), 'pulumi.get', 'pulumi.get', (['self', '"""http_options"""'], {}), "(self, 'http_options')\n", (362574, 362596), False, 'import pulumi\n'), ((362760, 362799), 'pulumi.set', 'pulumi.set', (['self', '"""http_options"""', 'value'], {}), "(self, 'http_options', value)\n", (362770, 362799), False, 'import pulumi\n'), ((362998, 363026), 'pulumi.get', 'pulumi.get', (['self', '"""interval"""'], {}), "(self, 'interval')\n", (363008, 363026), False, 'import pulumi\n'), ((363117, 363152), 'pulumi.set', 'pulumi.set', (['self', '"""interval"""', 'value'], {}), "(self, 'interval', value)\n", (363127, 363152), False, 'import pulumi\n'), ((363440, 363471), 'pulumi.get', 'pulumi.get', (['self', '"""tcp_options"""'], {}), "(self, 'tcp_options')\n", (363450, 363471), False, 'import pulumi\n'), ((363632, 363670), 'pulumi.set', 'pulumi.set', (['self', '"""tcp_options"""', 'value'], {}), "(self, 'tcp_options', value)\n", (363642, 363670), False, 'import pulumi\n'), ((363895, 363922), 'pulumi.get', 'pulumi.get', (['self', '"""timeout"""'], {}), "(self, 'timeout')\n", (363905, 363922), False, 'import pulumi\n'), ((364011, 364045), 'pulumi.set', 'pulumi.set', (['self', '"""timeout"""', 'value'], {}), "(self, 'timeout', value)\n", (364021, 364045), False, 'import pulumi\n'), ((364312, 364351), 'pulumi.get', 'pulumi.get', (['self', '"""unhealthy_threshold"""'], {}), "(self, 'unhealthy_threshold')\n", (364322, 364351), False, 'import pulumi\n'), ((364464, 364510), 'pulumi.set', 'pulumi.set', (['self', '"""unhealthy_threshold"""', 'value'], {}), "(self, 'unhealthy_threshold', value)\n", (364474, 364510), False, 'import pulumi\n'), ((365004, 365038), 'pulumi.set', 'pulumi.set', (['__self__', '"""port"""', 'port'], {}), "(__self__, 'port', port)\n", (365014, 365038), False, 'import pulumi\n'), ((365264, 365288), 'pulumi.get', 'pulumi.get', (['self', '"""port"""'], {}), "(self, 'port')\n", (365274, 365288), False, 'import pulumi\n'), ((365361, 365392), 'pulumi.set', 'pulumi.set', (['self', '"""port"""', 'value'], {}), "(self, 'port', value)\n", (365371, 365392), False, 'import pulumi\n'), ((365654, 365678), 'pulumi.get', 'pulumi.get', (['self', '"""path"""'], {}), "(self, 'path')\n", (365664, 365678), False, 'import pulumi\n'), ((365761, 365792), 'pulumi.set', 'pulumi.set', (['self', '"""path"""', 'value'], {}), "(self, 'path', value)\n", (365771, 365792), False, 'import pulumi\n'), ((366058, 366092), 'pulumi.set', 'pulumi.set', (['__self__', '"""port"""', 'port'], {}), "(__self__, 'port', port)\n", (366068, 366092), False, 'import pulumi\n'), ((366242, 366266), 'pulumi.get', 'pulumi.get', (['self', '"""port"""'], {}), "(self, 'port')\n", (366252, 366266), False, 'import pulumi\n'), ((366339, 366370), 'pulumi.set', 'pulumi.set', (['self', '"""port"""', 'value'], {}), "(self, 'port', value)\n", (366349, 366370), False, 'import pulumi\n'), ((367730, 367764), 'pulumi.set', 'pulumi.set', (['__self__', '"""name"""', 'name'], {}), "(__self__, 'name', name)\n", (367740, 367764), False, 'import pulumi\n'), ((367773, 367807), 'pulumi.set', 'pulumi.set', (['__self__', '"""port"""', 'port'], {}), "(__self__, 'port', port)\n", (367783, 367807), False, 'import pulumi\n'), ((368460, 368484), 'pulumi.get', 'pulumi.get', (['self', '"""name"""'], {}), "(self, 'name')\n", (368470, 368484), False, 'import pulumi\n'), ((368557, 368588), 'pulumi.set', 'pulumi.set', (['self', '"""name"""', 'value'], {}), "(self, 'name', value)\n", (368567, 368588), False, 'import pulumi\n'), ((368738, 368762), 'pulumi.get', 'pulumi.get', (['self', '"""port"""'], {}), "(self, 'port')\n", (368748, 368762), False, 'import pulumi\n'), ((368835, 368866), 'pulumi.set', 'pulumi.set', (['self', '"""port"""', 'value'], {}), "(self, 'port', value)\n", (368845, 368866), False, 'import pulumi\n'), ((369165, 369206), 'pulumi.get', 'pulumi.get', (['self', '"""external_address_spec"""'], {}), "(self, 'external_address_spec')\n", (369175, 369206), False, 'import pulumi\n'), ((369374, 369422), 'pulumi.set', 'pulumi.set', (['self', '"""external_address_spec"""', 'value'], {}), "(self, 'external_address_spec', value)\n", (369384, 369422), False, 'import pulumi\n'), ((369721, 369762), 'pulumi.get', 'pulumi.get', (['self', '"""internal_address_spec"""'], {}), "(self, 'internal_address_spec')\n", (369731, 369762), False, 'import pulumi\n'), ((369930, 369978), 'pulumi.set', 'pulumi.set', (['self', '"""internal_address_spec"""', 'value'], {}), "(self, 'internal_address_spec', value)\n", (369940, 369978), False, 'import pulumi\n'), ((370181, 370209), 'pulumi.get', 'pulumi.get', (['self', '"""protocol"""'], {}), "(self, 'protocol')\n", (370191, 370209), False, 'import pulumi\n'), ((370300, 370335), 'pulumi.set', 'pulumi.set', (['self', '"""protocol"""', 'value'], {}), "(self, 'protocol', value)\n", (370310, 370335), False, 'import pulumi\n'), ((370556, 370587), 'pulumi.get', 'pulumi.get', (['self', '"""target_port"""'], {}), "(self, 'target_port')\n", (370566, 370587), False, 'import pulumi\n'), ((370684, 370722), 'pulumi.set', 'pulumi.set', (['self', '"""target_port"""', 'value'], {}), "(self, 'target_port', value)\n", (370694, 370722), False, 'import pulumi\n'), ((371803, 371830), 'pulumi.get', 'pulumi.get', (['self', '"""address"""'], {}), "(self, 'address')\n", (371813, 371830), False, 'import pulumi\n'), ((371919, 371953), 'pulumi.set', 'pulumi.set', (['self', '"""address"""', 'value'], {}), "(self, 'address', value)\n", (371929, 371953), False, 'import pulumi\n'), ((372232, 372262), 'pulumi.get', 'pulumi.get', (['self', '"""ip_version"""'], {}), "(self, 'ip_version')\n", (372242, 372262), False, 'import pulumi\n'), ((372357, 372394), 'pulumi.set', 'pulumi.set', (['self', '"""ip_version"""', 'value'], {}), "(self, 'ip_version', value)\n", (372367, 372394), False, 'import pulumi\n'), ((373172, 373216), 'pulumi.set', 'pulumi.set', (['__self__', '"""subnet_id"""', 'subnet_id'], {}), "(__self__, 'subnet_id', subnet_id)\n", (373182, 373216), False, 'import pulumi\n'), ((373599, 373628), 'pulumi.get', 'pulumi.get', (['self', '"""subnet_id"""'], {}), "(self, 'subnet_id')\n", (373609, 373628), False, 'import pulumi\n'), ((373711, 373747), 'pulumi.set', 'pulumi.set', (['self', '"""subnet_id"""', 'value'], {}), "(self, 'subnet_id', value)\n", (373721, 373747), False, 'import pulumi\n'), ((374030, 374057), 'pulumi.get', 'pulumi.get', (['self', '"""address"""'], {}), "(self, 'address')\n", (374040, 374057), False, 'import pulumi\n'), ((374146, 374180), 'pulumi.set', 'pulumi.set', (['self', '"""address"""', 'value'], {}), "(self, 'address', value)\n", (374156, 374180), False, 'import pulumi\n'), ((374459, 374489), 'pulumi.get', 'pulumi.get', (['self', '"""ip_version"""'], {}), "(self, 'ip_version')\n", (374469, 374489), False, 'import pulumi\n'), ((374584, 374621), 'pulumi.set', 'pulumi.set', (['self', '"""ip_version"""', 'value'], {}), "(self, 'ip_version', value)\n", (374594, 374621), False, 'import pulumi\n'), ((375109, 375149), 'pulumi.set', 'pulumi.set', (['__self__', '"""address"""', 'address'], {}), "(__self__, 'address', address)\n", (375119, 375149), False, 'import pulumi\n'), ((375158, 375202), 'pulumi.set', 'pulumi.set', (['__self__', '"""subnet_id"""', 'subnet_id'], {}), "(__self__, 'subnet_id', subnet_id)\n", (375168, 375202), False, 'import pulumi\n'), ((375354, 375381), 'pulumi.get', 'pulumi.get', (['self', '"""address"""'], {}), "(self, 'address')\n", (375364, 375381), False, 'import pulumi\n'), ((375460, 375494), 'pulumi.set', 'pulumi.set', (['self', '"""address"""', 'value'], {}), "(self, 'address', value)\n", (375470, 375494), False, 'import pulumi\n'), ((375800, 375829), 'pulumi.get', 'pulumi.get', (['self', '"""subnet_id"""'], {}), "(self, 'subnet_id')\n", (375810, 375829), False, 'import pulumi\n'), ((375912, 375948), 'pulumi.set', 'pulumi.set', (['self', '"""subnet_id"""', 'value'], {}), "(self, 'subnet_id', value)\n", (375922, 375948), False, 'import pulumi\n'), ((377314, 377343), 'pulumi.get', 'pulumi.get', (['self', '"""data_lens"""'], {}), "(self, 'data_lens')\n", (377324, 377343), False, 'import pulumi\n'), ((377437, 377473), 'pulumi.set', 'pulumi.set', (['self', '"""data_lens"""', 'value'], {}), "(self, 'data_lens', value)\n", (377447, 377473), False, 'import pulumi\n'), ((377676, 377703), 'pulumi.get', 'pulumi.get', (['self', '"""metrika"""'], {}), "(self, 'metrika')\n", (377686, 377703), False, 'import pulumi\n'), ((377793, 377827), 'pulumi.set', 'pulumi.set', (['self', '"""metrika"""', 'value'], {}), "(self, 'metrika', value)\n", (377803, 377827), False, 'import pulumi\n'), ((378029, 378059), 'pulumi.get', 'pulumi.get', (['self', '"""serverless"""'], {}), "(self, 'serverless')\n", (378039, 378059), False, 'import pulumi\n'), ((378155, 378192), 'pulumi.set', 'pulumi.set', (['self', '"""serverless"""', 'value'], {}), "(self, 'serverless', value)\n", (378165, 378192), False, 'import pulumi\n'), ((378403, 378430), 'pulumi.get', 'pulumi.get', (['self', '"""web_sql"""'], {}), "(self, 'web_sql')\n", (378413, 378430), False, 'import pulumi\n'), ((378520, 378554), 'pulumi.set', 'pulumi.set', (['self', '"""web_sql"""', 'value'], {}), "(self, 'web_sql', value)\n", (378530, 378554), False, 'import pulumi\n'), ((379309, 379334), 'pulumi.get', 'pulumi.get', (['self', '"""hours"""'], {}), "(self, 'hours')\n", (379319, 379334), False, 'import pulumi\n'), ((379419, 379451), 'pulumi.set', 'pulumi.set', (['self', '"""hours"""', 'value'], {}), "(self, 'hours', value)\n", (379429, 379451), False, 'import pulumi\n'), ((379631, 379658), 'pulumi.get', 'pulumi.get', (['self', '"""minutes"""'], {}), "(self, 'minutes')\n", (379641, 379658), False, 'import pulumi\n'), ((379747, 379781), 'pulumi.set', 'pulumi.set', (['self', '"""minutes"""', 'value'], {}), "(self, 'minutes', value)\n", (379757, 379781), False, 'import pulumi\n'), ((380395, 380439), 'pulumi.set', 'pulumi.set', (['__self__', '"""resources"""', 'resources'], {}), "(__self__, 'resources', resources)\n", (380405, 380439), False, 'import pulumi\n'), ((380784, 380813), 'pulumi.get', 'pulumi.get', (['self', '"""resources"""'], {}), "(self, 'resources')\n", (380794, 380813), False, 'import pulumi\n'), ((380938, 380974), 'pulumi.set', 'pulumi.set', (['self', '"""resources"""', 'value'], {}), "(self, 'resources', value)\n", (380948, 380974), False, 'import pulumi\n'), ((381187, 381213), 'pulumi.get', 'pulumi.get', (['self', '"""config"""'], {}), "(self, 'config')\n", (381197, 381213), False, 'import pulumi\n'), ((381339, 381372), 'pulumi.set', 'pulumi.set', (['self', '"""config"""', 'value'], {}), "(self, 'config', value)\n", (381349, 381372), False, 'import pulumi\n'), ((389898, 389938), 'pulumi.get', 'pulumi.get', (['self', '"""background_pool_size"""'], {}), "(self, 'background_pool_size')\n", (389908, 389938), False, 'import pulumi\n'), ((390053, 390100), 'pulumi.set', 'pulumi.set', (['self', '"""background_pool_size"""', 'value'], {}), "(self, 'background_pool_size', value)\n", (390063, 390100), False, 'import pulumi\n'), ((390261, 390310), 'pulumi.get', 'pulumi.get', (['self', '"""background_schedule_pool_size"""'], {}), "(self, 'background_schedule_pool_size')\n", (390271, 390310), False, 'import pulumi\n'), ((390443, 390499), 'pulumi.set', 'pulumi.set', (['self', '"""background_schedule_pool_size"""', 'value'], {}), "(self, 'background_schedule_pool_size', value)\n", (390453, 390499), False, 'import pulumi\n'), ((390781, 390813), 'pulumi.get', 'pulumi.get', (['self', '"""compressions"""'], {}), "(self, 'compressions')\n", (390791, 390813), False, 'import pulumi\n'), ((390986, 391025), 'pulumi.set', 'pulumi.set', (['self', '"""compressions"""', 'value'], {}), "(self, 'compressions', value)\n", (390996, 391025), False, 'import pulumi\n'), ((391152, 391183), 'pulumi.get', 'pulumi.get', (['self', '"""geobase_uri"""'], {}), "(self, 'geobase_uri')\n", (391162, 391183), False, 'import pulumi\n'), ((391280, 391318), 'pulumi.set', 'pulumi.set', (['self', '"""geobase_uri"""', 'value'], {}), "(self, 'geobase_uri', value)\n", (391290, 391318), False, 'import pulumi\n'), ((391630, 391666), 'pulumi.get', 'pulumi.get', (['self', '"""graphite_rollups"""'], {}), "(self, 'graphite_rollups')\n", (391640, 391666), False, 'import pulumi\n'), ((391850, 391893), 'pulumi.set', 'pulumi.set', (['self', '"""graphite_rollups"""', 'value'], {}), "(self, 'graphite_rollups', value)\n", (391860, 391893), False, 'import pulumi\n'), ((392138, 392163), 'pulumi.get', 'pulumi.get', (['self', '"""kafka"""'], {}), "(self, 'kafka')\n", (392148, 392163), False, 'import pulumi\n'), ((392292, 392324), 'pulumi.set', 'pulumi.set', (['self', '"""kafka"""', 'value'], {}), "(self, 'kafka', value)\n", (392302, 392324), False, 'import pulumi\n'), ((392631, 392663), 'pulumi.get', 'pulumi.get', (['self', '"""kafka_topics"""'], {}), "(self, 'kafka_topics')\n", (392641, 392663), False, 'import pulumi\n'), ((392835, 392874), 'pulumi.set', 'pulumi.set', (['self', '"""kafka_topics"""', 'value'], {}), "(self, 'kafka_topics', value)\n", (392845, 392874), False, 'import pulumi\n'), ((393014, 393052), 'pulumi.get', 'pulumi.get', (['self', '"""keep_alive_timeout"""'], {}), "(self, 'keep_alive_timeout')\n", (393024, 393052), False, 'import pulumi\n'), ((393163, 393208), 'pulumi.set', 'pulumi.set', (['self', '"""keep_alive_timeout"""', 'value'], {}), "(self, 'keep_alive_timeout', value)\n", (393173, 393208), False, 'import pulumi\n'), ((393331, 393360), 'pulumi.get', 'pulumi.get', (['self', '"""log_level"""'], {}), "(self, 'log_level')\n", (393341, 393360), False, 'import pulumi\n'), ((393453, 393489), 'pulumi.set', 'pulumi.set', (['self', '"""log_level"""', 'value'], {}), "(self, 'log_level', value)\n", (393463, 393489), False, 'import pulumi\n'), ((393623, 393658), 'pulumi.get', 'pulumi.get', (['self', '"""mark_cache_size"""'], {}), "(self, 'mark_cache_size')\n", (393633, 393658), False, 'import pulumi\n'), ((393763, 393805), 'pulumi.set', 'pulumi.set', (['self', '"""mark_cache_size"""', 'value'], {}), "(self, 'mark_cache_size', value)\n", (393773, 393805), False, 'import pulumi\n'), ((393953, 393995), 'pulumi.get', 'pulumi.get', (['self', '"""max_concurrent_queries"""'], {}), "(self, 'max_concurrent_queries')\n", (393963, 393995), False, 'import pulumi\n'), ((394114, 394163), 'pulumi.set', 'pulumi.set', (['self', '"""max_concurrent_queries"""', 'value'], {}), "(self, 'max_concurrent_queries', value)\n", (394124, 394163), False, 'import pulumi\n'), ((394298, 394333), 'pulumi.get', 'pulumi.get', (['self', '"""max_connections"""'], {}), "(self, 'max_connections')\n", (394308, 394333), False, 'import pulumi\n'), ((394438, 394480), 'pulumi.set', 'pulumi.set', (['self', '"""max_connections"""', 'value'], {}), "(self, 'max_connections', value)\n", (394448, 394480), False, 'import pulumi\n'), ((394634, 394680), 'pulumi.get', 'pulumi.get', (['self', '"""max_partition_size_to_drop"""'], {}), "(self, 'max_partition_size_to_drop')\n", (394644, 394680), False, 'import pulumi\n'), ((394807, 394860), 'pulumi.set', 'pulumi.set', (['self', '"""max_partition_size_to_drop"""', 'value'], {}), "(self, 'max_partition_size_to_drop', value)\n", (394817, 394860), False, 'import pulumi\n'), ((395006, 395048), 'pulumi.get', 'pulumi.get', (['self', '"""max_table_size_to_drop"""'], {}), "(self, 'max_table_size_to_drop')\n", (395016, 395048), False, 'import pulumi\n'), ((395167, 395216), 'pulumi.set', 'pulumi.set', (['self', '"""max_table_size_to_drop"""', 'value'], {}), "(self, 'max_table_size_to_drop', value)\n", (395177, 395216), False, 'import pulumi\n'), ((395488, 395518), 'pulumi.get', 'pulumi.get', (['self', '"""merge_tree"""'], {}), "(self, 'merge_tree')\n", (395498, 395518), False, 'import pulumi\n'), ((395661, 395698), 'pulumi.set', 'pulumi.set', (['self', '"""merge_tree"""', 'value'], {}), "(self, 'merge_tree', value)\n", (395671, 395698), False, 'import pulumi\n'), ((395839, 395877), 'pulumi.get', 'pulumi.get', (['self', '"""metric_log_enabled"""'], {}), "(self, 'metric_log_enabled')\n", (395849, 395877), False, 'import pulumi\n'), ((395989, 396034), 'pulumi.set', 'pulumi.set', (['self', '"""metric_log_enabled"""', 'value'], {}), "(self, 'metric_log_enabled', value)\n", (395999, 396034), False, 'import pulumi\n'), ((396187, 396232), 'pulumi.get', 'pulumi.get', (['self', '"""metric_log_retention_size"""'], {}), "(self, 'metric_log_retention_size')\n", (396197, 396232), False, 'import pulumi\n'), ((396357, 396409), 'pulumi.set', 'pulumi.set', (['self', '"""metric_log_retention_size"""', 'value'], {}), "(self, 'metric_log_retention_size', value)\n", (396367, 396409), False, 'import pulumi\n'), ((396562, 396607), 'pulumi.get', 'pulumi.get', (['self', '"""metric_log_retention_time"""'], {}), "(self, 'metric_log_retention_time')\n", (396572, 396607), False, 'import pulumi\n'), ((396732, 396784), 'pulumi.set', 'pulumi.set', (['self', '"""metric_log_retention_time"""', 'value'], {}), "(self, 'metric_log_retention_time', value)\n", (396742, 396784), False, 'import pulumi\n'), ((396933, 396976), 'pulumi.get', 'pulumi.get', (['self', '"""part_log_retention_size"""'], {}), "(self, 'part_log_retention_size')\n", (396943, 396976), False, 'import pulumi\n'), ((397097, 397147), 'pulumi.set', 'pulumi.set', (['self', '"""part_log_retention_size"""', 'value'], {}), "(self, 'part_log_retention_size', value)\n", (397107, 397147), False, 'import pulumi\n'), ((397296, 397339), 'pulumi.get', 'pulumi.get', (['self', '"""part_log_retention_time"""'], {}), "(self, 'part_log_retention_time')\n", (397306, 397339), False, 'import pulumi\n'), ((397460, 397510), 'pulumi.set', 'pulumi.set', (['self', '"""part_log_retention_time"""', 'value'], {}), "(self, 'part_log_retention_time', value)\n", (397470, 397510), False, 'import pulumi\n'), ((397661, 397705), 'pulumi.get', 'pulumi.get', (['self', '"""query_log_retention_size"""'], {}), "(self, 'query_log_retention_size')\n", (397671, 397705), False, 'import pulumi\n'), ((397828, 397879), 'pulumi.set', 'pulumi.set', (['self', '"""query_log_retention_size"""', 'value'], {}), "(self, 'query_log_retention_size', value)\n", (397838, 397879), False, 'import pulumi\n'), ((398030, 398074), 'pulumi.get', 'pulumi.get', (['self', '"""query_log_retention_time"""'], {}), "(self, 'query_log_retention_time')\n", (398040, 398074), False, 'import pulumi\n'), ((398197, 398248), 'pulumi.set', 'pulumi.set', (['self', '"""query_log_retention_time"""', 'value'], {}), "(self, 'query_log_retention_time', value)\n", (398207, 398248), False, 'import pulumi\n'), ((398400, 398444), 'pulumi.get', 'pulumi.get', (['self', '"""query_thread_log_enabled"""'], {}), "(self, 'query_thread_log_enabled')\n", (398410, 398444), False, 'import pulumi\n'), ((398568, 398619), 'pulumi.set', 'pulumi.set', (['self', '"""query_thread_log_enabled"""', 'value'], {}), "(self, 'query_thread_log_enabled', value)\n", (398578, 398619), False, 'import pulumi\n'), ((398783, 398834), 'pulumi.get', 'pulumi.get', (['self', '"""query_thread_log_retention_size"""'], {}), "(self, 'query_thread_log_retention_size')\n", (398793, 398834), False, 'import pulumi\n'), ((398971, 399029), 'pulumi.set', 'pulumi.set', (['self', '"""query_thread_log_retention_size"""', 'value'], {}), "(self, 'query_thread_log_retention_size', value)\n", (398981, 399029), False, 'import pulumi\n'), ((399193, 399244), 'pulumi.get', 'pulumi.get', (['self', '"""query_thread_log_retention_time"""'], {}), "(self, 'query_thread_log_retention_time')\n", (399203, 399244), False, 'import pulumi\n'), ((399381, 399439), 'pulumi.set', 'pulumi.set', (['self', '"""query_thread_log_retention_time"""', 'value'], {}), "(self, 'query_thread_log_retention_time', value)\n", (399391, 399439), False, 'import pulumi\n'), ((399693, 399721), 'pulumi.get', 'pulumi.get', (['self', '"""rabbitmq"""'], {}), "(self, 'rabbitmq')\n", (399703, 399721), False, 'import pulumi\n'), ((399859, 399894), 'pulumi.set', 'pulumi.set', (['self', '"""rabbitmq"""', 'value'], {}), "(self, 'rabbitmq', value)\n", (399869, 399894), False, 'import pulumi\n'), ((400031, 400067), 'pulumi.get', 'pulumi.get', (['self', '"""text_log_enabled"""'], {}), "(self, 'text_log_enabled')\n", (400041, 400067), False, 'import pulumi\n'), ((400175, 400218), 'pulumi.set', 'pulumi.set', (['self', '"""text_log_enabled"""', 'value'], {}), "(self, 'text_log_enabled', value)\n", (400185, 400218), False, 'import pulumi\n'), ((400350, 400384), 'pulumi.get', 'pulumi.get', (['self', '"""text_log_level"""'], {}), "(self, 'text_log_level')\n", (400360, 400384), False, 'import pulumi\n'), ((400487, 400528), 'pulumi.set', 'pulumi.set', (['self', '"""text_log_level"""', 'value'], {}), "(self, 'text_log_level', value)\n", (400497, 400528), False, 'import pulumi\n'), ((400677, 400720), 'pulumi.get', 'pulumi.get', (['self', '"""text_log_retention_size"""'], {}), "(self, 'text_log_retention_size')\n", (400687, 400720), False, 'import pulumi\n'), ((400841, 400891), 'pulumi.set', 'pulumi.set', (['self', '"""text_log_retention_size"""', 'value'], {}), "(self, 'text_log_retention_size', value)\n", (400851, 400891), False, 'import pulumi\n'), ((401040, 401083), 'pulumi.get', 'pulumi.get', (['self', '"""text_log_retention_time"""'], {}), "(self, 'text_log_retention_time')\n", (401050, 401083), False, 'import pulumi\n'), ((401204, 401254), 'pulumi.set', 'pulumi.set', (['self', '"""text_log_retention_time"""', 'value'], {}), "(self, 'text_log_retention_time', value)\n", (401214, 401254), False, 'import pulumi\n'), ((401359, 401387), 'pulumi.get', 'pulumi.get', (['self', '"""timezone"""'], {}), "(self, 'timezone')\n", (401369, 401387), False, 'import pulumi\n'), ((401478, 401513), 'pulumi.set', 'pulumi.set', (['self', '"""timezone"""', 'value'], {}), "(self, 'timezone', value)\n", (401488, 401513), False, 'import pulumi\n'), ((401652, 401689), 'pulumi.get', 'pulumi.get', (['self', '"""trace_log_enabled"""'], {}), "(self, 'trace_log_enabled')\n", (401662, 401689), False, 'import pulumi\n'), ((401799, 401843), 'pulumi.set', 'pulumi.set', (['self', '"""trace_log_enabled"""', 'value'], {}), "(self, 'trace_log_enabled', value)\n", (401809, 401843), False, 'import pulumi\n'), ((401994, 402038), 'pulumi.get', 'pulumi.get', (['self', '"""trace_log_retention_size"""'], {}), "(self, 'trace_log_retention_size')\n", (402004, 402038), False, 'import pulumi\n'), ((402161, 402212), 'pulumi.set', 'pulumi.set', (['self', '"""trace_log_retention_size"""', 'value'], {}), "(self, 'trace_log_retention_size', value)\n", (402171, 402212), False, 'import pulumi\n'), ((402363, 402407), 'pulumi.get', 'pulumi.get', (['self', '"""trace_log_retention_time"""'], {}), "(self, 'trace_log_retention_time')\n", (402373, 402407), False, 'import pulumi\n'), ((402530, 402581), 'pulumi.set', 'pulumi.set', (['self', '"""trace_log_retention_time"""', 'value'], {}), "(self, 'trace_log_retention_time', value)\n", (402540, 402581), False, 'import pulumi\n'), ((402731, 402774), 'pulumi.get', 'pulumi.get', (['self', '"""uncompressed_cache_size"""'], {}), "(self, 'uncompressed_cache_size')\n", (402741, 402774), False, 'import pulumi\n'), ((402895, 402945), 'pulumi.set', 'pulumi.set', (['self', '"""uncompressed_cache_size"""', 'value'], {}), "(self, 'uncompressed_cache_size', value)\n", (402905, 402945), False, 'import pulumi\n'), ((403826, 403864), 'pulumi.set', 'pulumi.set', (['__self__', '"""method"""', 'method'], {}), "(__self__, 'method', method)\n", (403836, 403864), False, 'import pulumi\n'), ((403873, 403925), 'pulumi.set', 'pulumi.set', (['__self__', '"""min_part_size"""', 'min_part_size'], {}), "(__self__, 'min_part_size', min_part_size)\n", (403883, 403925), False, 'import pulumi\n'), ((403934, 403998), 'pulumi.set', 'pulumi.set', (['__self__', '"""min_part_size_ratio"""', 'min_part_size_ratio'], {}), "(__self__, 'min_part_size_ratio', min_part_size_ratio)\n", (403944, 403998), False, 'import pulumi\n'), ((404192, 404218), 'pulumi.get', 'pulumi.get', (['self', '"""method"""'], {}), "(self, 'method')\n", (404202, 404218), False, 'import pulumi\n'), ((404295, 404328), 'pulumi.set', 'pulumi.set', (['self', '"""method"""', 'value'], {}), "(self, 'method', value)\n", (404305, 404328), False, 'import pulumi\n'), ((404655, 404688), 'pulumi.get', 'pulumi.get', (['self', '"""min_part_size"""'], {}), "(self, 'min_part_size')\n", (404665, 404688), False, 'import pulumi\n'), ((404779, 404819), 'pulumi.set', 'pulumi.set', (['self', '"""min_part_size"""', 'value'], {}), "(self, 'min_part_size', value)\n", (404789, 404819), False, 'import pulumi\n'), ((405178, 405217), 'pulumi.get', 'pulumi.get', (['self', '"""min_part_size_ratio"""'], {}), "(self, 'min_part_size_ratio')\n", (405188, 405217), False, 'import pulumi\n'), ((405322, 405368), 'pulumi.set', 'pulumi.set', (['self', '"""min_part_size_ratio"""', 'value'], {}), "(self, 'min_part_size_ratio', value)\n", (405332, 405368), False, 'import pulumi\n'), ((405929, 405963), 'pulumi.set', 'pulumi.set', (['__self__', '"""name"""', 'name'], {}), "(__self__, 'name', name)\n", (405939, 405963), False, 'import pulumi\n'), ((406210, 406234), 'pulumi.get', 'pulumi.get', (['self', '"""name"""'], {}), "(self, 'name')\n", (406220, 406234), False, 'import pulumi\n'), ((406307, 406338), 'pulumi.set', 'pulumi.set', (['self', '"""name"""', 'value'], {}), "(self, 'name', value)\n", (406317, 406338), False, 'import pulumi\n'), ((406582, 406610), 'pulumi.get', 'pulumi.get', (['self', '"""patterns"""'], {}), "(self, 'patterns')\n", (406592, 406610), False, 'import pulumi\n'), ((406785, 406820), 'pulumi.set', 'pulumi.set', (['self', '"""patterns"""', 'value'], {}), "(self, 'patterns', value)\n", (406795, 406820), False, 'import pulumi\n'), ((407559, 407601), 'pulumi.set', 'pulumi.set', (['__self__', '"""function"""', 'function'], {}), "(__self__, 'function', function)\n", (407569, 407601), False, 'import pulumi\n'), ((407931, 407959), 'pulumi.get', 'pulumi.get', (['self', '"""function"""'], {}), "(self, 'function')\n", (407941, 407959), False, 'import pulumi\n'), ((408040, 408075), 'pulumi.set', 'pulumi.set', (['self', '"""function"""', 'value'], {}), "(self, 'function', value)\n", (408050, 408075), False, 'import pulumi\n'), ((408262, 408288), 'pulumi.get', 'pulumi.get', (['self', '"""regexp"""'], {}), "(self, 'regexp')\n", (408272, 408288), False, 'import pulumi\n'), ((408375, 408408), 'pulumi.set', 'pulumi.set', (['self', '"""regexp"""', 'value'], {}), "(self, 'regexp', value)\n", (408385, 408408), False, 'import pulumi\n'), ((408659, 408689), 'pulumi.get', 'pulumi.get', (['self', '"""retentions"""'], {}), "(self, 'retentions')\n", (408669, 408689), False, 'import pulumi\n'), ((408877, 408914), 'pulumi.set', 'pulumi.set', (['self', '"""retentions"""', 'value'], {}), "(self, 'retentions', value)\n", (408887, 408914), False, 'import pulumi\n'), ((409332, 409364), 'pulumi.set', 'pulumi.set', (['__self__', '"""age"""', 'age'], {}), "(__self__, 'age', age)\n", (409342, 409364), False, 'import pulumi\n'), ((409373, 409417), 'pulumi.set', 'pulumi.set', (['__self__', '"""precision"""', 'precision'], {}), "(__self__, 'precision', precision)\n", (409383, 409417), False, 'import pulumi\n'), ((409568, 409591), 'pulumi.get', 'pulumi.get', (['self', '"""age"""'], {}), "(self, 'age')\n", (409578, 409591), False, 'import pulumi\n'), ((409662, 409692), 'pulumi.set', 'pulumi.set', (['self', '"""age"""', 'value'], {}), "(self, 'age', value)\n", (409672, 409692), False, 'import pulumi\n'), ((409876, 409905), 'pulumi.get', 'pulumi.get', (['self', '"""precision"""'], {}), "(self, 'precision')\n", (409886, 409905), False, 'import pulumi\n'), ((409988, 410024), 'pulumi.set', 'pulumi.set', (['self', '"""precision"""', 'value'], {}), "(self, 'precision', value)\n", (409998, 410024), False, 'import pulumi\n'), ((411414, 411448), 'pulumi.get', 'pulumi.get', (['self', '"""sasl_mechanism"""'], {}), "(self, 'sasl_mechanism')\n", (411424, 411448), False, 'import pulumi\n'), ((411551, 411592), 'pulumi.set', 'pulumi.set', (['self', '"""sasl_mechanism"""', 'value'], {}), "(self, 'sasl_mechanism', value)\n", (411561, 411592), False, 'import pulumi\n'), ((411786, 411819), 'pulumi.get', 'pulumi.get', (['self', '"""sasl_password"""'], {}), "(self, 'sasl_password')\n", (411796, 411819), False, 'import pulumi\n'), ((411920, 411960), 'pulumi.set', 'pulumi.set', (['self', '"""sasl_password"""', 'value'], {}), "(self, 'sasl_password', value)\n", (411930, 411960), False, 'import pulumi\n'), ((412149, 412182), 'pulumi.get', 'pulumi.get', (['self', '"""sasl_username"""'], {}), "(self, 'sasl_username')\n", (412159, 412182), False, 'import pulumi\n'), ((412283, 412323), 'pulumi.set', 'pulumi.set', (['self', '"""sasl_username"""', 'value'], {}), "(self, 'sasl_username', value)\n", (412293, 412323), False, 'import pulumi\n'), ((412545, 412582), 'pulumi.get', 'pulumi.get', (['self', '"""security_protocol"""'], {}), "(self, 'security_protocol')\n", (412555, 412582), False, 'import pulumi\n'), ((412691, 412735), 'pulumi.set', 'pulumi.set', (['self', '"""security_protocol"""', 'value'], {}), "(self, 'security_protocol', value)\n", (412701, 412735), False, 'import pulumi\n'), ((413264, 413298), 'pulumi.set', 'pulumi.set', (['__self__', '"""name"""', 'name'], {}), "(__self__, 'name', name)\n", (413274, 413298), False, 'import pulumi\n'), ((413545, 413569), 'pulumi.get', 'pulumi.get', (['self', '"""name"""'], {}), "(self, 'name')\n", (413555, 413569), False, 'import pulumi\n'), ((413642, 413673), 'pulumi.set', 'pulumi.set', (['self', '"""name"""', 'value'], {}), "(self, 'name', value)\n", (413652, 413673), False, 'import pulumi\n'), ((413916, 413944), 'pulumi.get', 'pulumi.get', (['self', '"""settings"""'], {}), "(self, 'settings')\n", (413926, 413944), False, 'import pulumi\n'), ((414092, 414127), 'pulumi.set', 'pulumi.set', (['self', '"""settings"""', 'value'], {}), "(self, 'settings', value)\n", (414102, 414127), False, 'import pulumi\n'), ((415530, 415564), 'pulumi.get', 'pulumi.get', (['self', '"""sasl_mechanism"""'], {}), "(self, 'sasl_mechanism')\n", (415540, 415564), False, 'import pulumi\n'), ((415667, 415708), 'pulumi.set', 'pulumi.set', (['self', '"""sasl_mechanism"""', 'value'], {}), "(self, 'sasl_mechanism', value)\n", (415677, 415708), False, 'import pulumi\n'), ((415902, 415935), 'pulumi.get', 'pulumi.get', (['self', '"""sasl_password"""'], {}), "(self, 'sasl_password')\n", (415912, 415935), False, 'import pulumi\n'), ((416036, 416076), 'pulumi.set', 'pulumi.set', (['self', '"""sasl_password"""', 'value'], {}), "(self, 'sasl_password', value)\n", (416046, 416076), False, 'import pulumi\n'), ((416265, 416298), 'pulumi.get', 'pulumi.get', (['self', '"""sasl_username"""'], {}), "(self, 'sasl_username')\n", (416275, 416298), False, 'import pulumi\n'), ((416399, 416439), 'pulumi.set', 'pulumi.set', (['self', '"""sasl_username"""', 'value'], {}), "(self, 'sasl_username', value)\n", (416409, 416439), False, 'import pulumi\n'), ((416661, 416698), 'pulumi.get', 'pulumi.get', (['self', '"""security_protocol"""'], {}), "(self, 'security_protocol')\n", (416671, 416698), False, 'import pulumi\n'), ((416807, 416851), 'pulumi.set', 'pulumi.set', (['self', '"""security_protocol"""', 'value'], {}), "(self, 'security_protocol', value)\n", (416817, 416851), False, 'import pulumi\n'), ((420753, 420812), 'pulumi.get', 'pulumi.get', (['self', '"""max_bytes_to_merge_at_min_space_in_pool"""'], {}), "(self, 'max_bytes_to_merge_at_min_space_in_pool')\n", (420763, 420812), False, 'import pulumi\n'), ((420965, 421031), 'pulumi.set', 'pulumi.set', (['self', '"""max_bytes_to_merge_at_min_space_in_pool"""', 'value'], {}), "(self, 'max_bytes_to_merge_at_min_space_in_pool', value)\n", (420975, 421031), False, 'import pulumi\n'), ((421350, 421400), 'pulumi.get', 'pulumi.get', (['self', '"""max_replicated_merges_in_queue"""'], {}), "(self, 'max_replicated_merges_in_queue')\n", (421360, 421400), False, 'import pulumi\n'), ((421535, 421592), 'pulumi.set', 'pulumi.set', (['self', '"""max_replicated_merges_in_queue"""', 'value'], {}), "(self, 'max_replicated_merges_in_queue', value)\n", (421545, 421592), False, 'import pulumi\n'), ((422148, 422225), 'pulumi.get', 'pulumi.get', (['self', '"""number_of_free_entries_in_pool_to_lower_max_size_of_merge"""'], {}), "(self, 'number_of_free_entries_in_pool_to_lower_max_size_of_merge')\n", (422158, 422225), False, 'import pulumi\n'), ((422414, 422502), 'pulumi.set', 'pulumi.set', (['self', '"""number_of_free_entries_in_pool_to_lower_max_size_of_merge"""', 'value'], {}), "(self,\n 'number_of_free_entries_in_pool_to_lower_max_size_of_merge', value)\n", (422424, 422502), False, 'import pulumi\n'), ((422838, 422879), 'pulumi.get', 'pulumi.get', (['self', '"""parts_to_delay_insert"""'], {}), "(self, 'parts_to_delay_insert')\n", (422848, 422879), False, 'import pulumi\n'), ((422996, 423044), 'pulumi.set', 'pulumi.set', (['self', '"""parts_to_delay_insert"""', 'value'], {}), "(self, 'parts_to_delay_insert', value)\n", (423006, 423044), False, 'import pulumi\n'), ((423366, 423407), 'pulumi.get', 'pulumi.get', (['self', '"""parts_to_throw_insert"""'], {}), "(self, 'parts_to_throw_insert')\n", (423376, 423407), False, 'import pulumi\n'), ((423524, 423572), 'pulumi.set', 'pulumi.set', (['self', '"""parts_to_throw_insert"""', 'value'], {}), "(self, 'parts_to_throw_insert', value)\n", (423534, 423572), False, 'import pulumi\n'), ((423890, 423941), 'pulumi.get', 'pulumi.get', (['self', '"""replicated_deduplication_window"""'], {}), "(self, 'replicated_deduplication_window')\n", (423900, 423941), False, 'import pulumi\n'), ((424078, 424136), 'pulumi.set', 'pulumi.set', (['self', '"""replicated_deduplication_window"""', 'value'], {}), "(self, 'replicated_deduplication_window', value)\n", (424088, 424136), False, 'import pulumi\n'), ((424472, 424531), 'pulumi.get', 'pulumi.get', (['self', '"""replicated_deduplication_window_seconds"""'], {}), "(self, 'replicated_deduplication_window_seconds')\n", (424482, 424531), False, 'import pulumi\n'), ((424684, 424750), 'pulumi.set', 'pulumi.set', (['self', '"""replicated_deduplication_window_seconds"""', 'value'], {}), "(self, 'replicated_deduplication_window_seconds', value)\n", (424694, 424750), False, 'import pulumi\n'), ((425462, 425490), 'pulumi.get', 'pulumi.get', (['self', '"""password"""'], {}), "(self, 'password')\n", (425472, 425490), False, 'import pulumi\n'), ((425581, 425616), 'pulumi.set', 'pulumi.set', (['self', '"""password"""', 'value'], {}), "(self, 'password', value)\n", (425591, 425616), False, 'import pulumi\n'), ((425772, 425800), 'pulumi.get', 'pulumi.get', (['self', '"""username"""'], {}), "(self, 'username')\n", (425782, 425800), False, 'import pulumi\n'), ((425891, 425926), 'pulumi.set', 'pulumi.set', (['self', '"""username"""', 'value'], {}), "(self, 'username', value)\n", (425901, 425926), False, 'import pulumi\n'), ((426550, 426594), 'pulumi.set', 'pulumi.set', (['__self__', '"""disk_size"""', 'disk_size'], {}), "(__self__, 'disk_size', disk_size)\n", (426560, 426594), False, 'import pulumi\n'), ((426603, 426653), 'pulumi.set', 'pulumi.set', (['__self__', '"""disk_type_id"""', 'disk_type_id'], {}), "(__self__, 'disk_type_id', disk_type_id)\n", (426613, 426653), False, 'import pulumi\n'), ((426662, 426724), 'pulumi.set', 'pulumi.set', (['__self__', '"""resource_preset_id"""', 'resource_preset_id'], {}), "(__self__, 'resource_preset_id', resource_preset_id)\n", (426672, 426724), False, 'import pulumi\n'), ((426936, 426965), 'pulumi.get', 'pulumi.get', (['self', '"""disk_size"""'], {}), "(self, 'disk_size')\n", (426946, 426965), False, 'import pulumi\n'), ((427048, 427084), 'pulumi.set', 'pulumi.set', (['self', '"""disk_size"""', 'value'], {}), "(self, 'disk_size', value)\n", (427058, 427084), False, 'import pulumi\n'), ((427404, 427436), 'pulumi.get', 'pulumi.get', (['self', '"""disk_type_id"""'], {}), "(self, 'disk_type_id')\n", (427414, 427436), False, 'import pulumi\n'), ((427525, 427564), 'pulumi.set', 'pulumi.set', (['self', '"""disk_type_id"""', 'value'], {}), "(self, 'disk_type_id', value)\n", (427535, 427564), False, 'import pulumi\n'), ((427694, 427732), 'pulumi.get', 'pulumi.get', (['self', '"""resource_preset_id"""'], {}), "(self, 'resource_preset_id')\n", (427704, 427732), False, 'import pulumi\n'), ((427833, 427878), 'pulumi.set', 'pulumi.set', (['self', '"""resource_preset_id"""', 'value'], {}), "(self, 'resource_preset_id', value)\n", (427843, 427878), False, 'import pulumi\n'), ((428195, 428235), 'pulumi.set', 'pulumi.set', (['__self__', '"""enabled"""', 'enabled'], {}), "(__self__, 'enabled', enabled)\n", (428205, 428235), False, 'import pulumi\n'), ((428461, 428488), 'pulumi.get', 'pulumi.get', (['self', '"""enabled"""'], {}), "(self, 'enabled')\n", (428471, 428488), False, 'import pulumi\n'), ((428568, 428602), 'pulumi.set', 'pulumi.set', (['self', '"""enabled"""', 'value'], {}), "(self, 'enabled', value)\n", (428578, 428602), False, 'import pulumi\n'), ((428844, 428878), 'pulumi.set', 'pulumi.set', (['__self__', '"""name"""', 'name'], {}), "(__self__, 'name', name)\n", (428854, 428878), False, 'import pulumi\n'), ((429037, 429061), 'pulumi.get', 'pulumi.get', (['self', '"""name"""'], {}), "(self, 'name')\n", (429047, 429061), False, 'import pulumi\n'), ((429134, 429165), 'pulumi.set', 'pulumi.set', (['self', '"""name"""', 'value'], {}), "(self, 'name', value)\n", (429144, 429165), False, 'import pulumi\n'), ((429775, 429809), 'pulumi.set', 'pulumi.set', (['__self__', '"""name"""', 'name'], {}), "(__self__, 'name', name)\n", (429785, 429809), False, 'import pulumi\n'), ((429818, 429852), 'pulumi.set', 'pulumi.set', (['__self__', '"""type"""', 'type'], {}), "(__self__, 'type', type)\n", (429828, 429852), False, 'import pulumi\n'), ((429861, 429893), 'pulumi.set', 'pulumi.set', (['__self__', '"""uri"""', 'uri'], {}), "(__self__, 'uri', uri)\n", (429871, 429893), False, 'import pulumi\n'), ((430052, 430076), 'pulumi.get', 'pulumi.get', (['self', '"""name"""'], {}), "(self, 'name')\n", (430062, 430076), False, 'import pulumi\n'), ((430149, 430180), 'pulumi.set', 'pulumi.set', (['self', '"""name"""', 'value'], {}), "(self, 'name', value)\n", (430159, 430180), False, 'import pulumi\n'), ((430434, 430458), 'pulumi.get', 'pulumi.get', (['self', '"""type"""'], {}), "(self, 'type')\n", (430444, 430458), False, 'import pulumi\n'), ((430531, 430562), 'pulumi.set', 'pulumi.set', (['self', '"""type"""', 'value'], {}), "(self, 'type', value)\n", (430541, 430562), False, 'import pulumi\n'), ((430757, 430780), 'pulumi.get', 'pulumi.get', (['self', '"""uri"""'], {}), "(self, 'uri')\n", (430767, 430780), False, 'import pulumi\n'), ((430851, 430881), 'pulumi.set', 'pulumi.set', (['self', '"""uri"""', 'value'], {}), "(self, 'uri', value)\n", (430861, 430881), False, 'import pulumi\n'), ((432242, 432276), 'pulumi.set', 'pulumi.set', (['__self__', '"""type"""', 'type'], {}), "(__self__, 'type', type)\n", (432252, 432276), False, 'import pulumi\n'), ((432285, 432319), 'pulumi.set', 'pulumi.set', (['__self__', '"""zone"""', 'zone'], {}), "(__self__, 'zone', zone)\n", (432295, 432319), False, 'import pulumi\n'), ((432946, 432970), 'pulumi.get', 'pulumi.get', (['self', '"""type"""'], {}), "(self, 'type')\n", (432956, 432970), False, 'import pulumi\n'), ((433043, 433074), 'pulumi.set', 'pulumi.set', (['self', '"""type"""', 'value'], {}), "(self, 'type', value)\n", (433053, 433074), False, 'import pulumi\n'), ((433384, 433408), 'pulumi.get', 'pulumi.get', (['self', '"""zone"""'], {}), "(self, 'zone')\n", (433394, 433408), False, 'import pulumi\n'), ((433481, 433512), 'pulumi.set', 'pulumi.set', (['self', '"""zone"""', 'value'], {}), "(self, 'zone', value)\n", (433491, 433512), False, 'import pulumi\n'), ((433780, 433816), 'pulumi.get', 'pulumi.get', (['self', '"""assign_public_ip"""'], {}), "(self, 'assign_public_ip')\n", (433790, 433816), False, 'import pulumi\n'), ((433924, 433967), 'pulumi.set', 'pulumi.set', (['self', '"""assign_public_ip"""', 'value'], {}), "(self, 'assign_public_ip', value)\n", (433934, 433967), False, 'import pulumi\n'), ((434145, 434169), 'pulumi.get', 'pulumi.get', (['self', '"""fqdn"""'], {}), "(self, 'fqdn')\n", (434155, 434169), False, 'import pulumi\n'), ((434252, 434283), 'pulumi.set', 'pulumi.set', (['self', '"""fqdn"""', 'value'], {}), "(self, 'fqdn', value)\n", (434262, 434283), False, 'import pulumi\n'), ((434489, 434519), 'pulumi.get', 'pulumi.get', (['self', '"""shard_name"""'], {}), "(self, 'shard_name')\n", (434499, 434519), False, 'import pulumi\n'), ((434614, 434651), 'pulumi.set', 'pulumi.set', (['self', '"""shard_name"""', 'value'], {}), "(self, 'shard_name', value)\n", (434624, 434651), False, 'import pulumi\n'), ((434926, 434955), 'pulumi.get', 'pulumi.get', (['self', '"""subnet_id"""'], {}), "(self, 'subnet_id')\n", (434936, 434955), False, 'import pulumi\n'), ((435048, 435084), 'pulumi.set', 'pulumi.set', (['self', '"""subnet_id"""', 'value'], {}), "(self, 'subnet_id', value)\n", (435058, 435084), False, 'import pulumi\n'), ((435836, 435870), 'pulumi.set', 'pulumi.set', (['__self__', '"""type"""', 'type'], {}), "(__self__, 'type', type)\n", (435846, 435870), False, 'import pulumi\n'), ((436273, 436297), 'pulumi.get', 'pulumi.get', (['self', '"""type"""'], {}), "(self, 'type')\n", (436283, 436297), False, 'import pulumi\n'), ((436370, 436401), 'pulumi.set', 'pulumi.set', (['self', '"""type"""', 'value'], {}), "(self, 'type', value)\n", (436380, 436401), False, 'import pulumi\n'), ((436660, 436683), 'pulumi.get', 'pulumi.get', (['self', '"""day"""'], {}), "(self, 'day')\n", (436670, 436683), False, 'import pulumi\n'), ((436764, 436794), 'pulumi.set', 'pulumi.set', (['self', '"""day"""', 'value'], {}), "(self, 'day', value)\n", (436774, 436794), False, 'import pulumi\n'), ((437012, 437036), 'pulumi.get', 'pulumi.get', (['self', '"""hour"""'], {}), "(self, 'hour')\n", (437022, 437036), False, 'import pulumi\n'), ((437119, 437150), 'pulumi.set', 'pulumi.set', (['self', '"""hour"""', 'value'], {}), "(self, 'hour', value)\n", (437129, 437150), False, 'import pulumi\n'), ((437755, 437789), 'pulumi.set', 'pulumi.set', (['__self__', '"""name"""', 'name'], {}), "(__self__, 'name', name)\n", (437765, 437789), False, 'import pulumi\n'), ((437798, 437832), 'pulumi.set', 'pulumi.set', (['__self__', '"""type"""', 'type'], {}), "(__self__, 'type', type)\n", (437808, 437832), False, 'import pulumi\n'), ((437841, 437873), 'pulumi.set', 'pulumi.set', (['__self__', '"""uri"""', 'uri'], {}), "(__self__, 'uri', uri)\n", (437851, 437873), False, 'import pulumi\n'), ((438032, 438056), 'pulumi.get', 'pulumi.get', (['self', '"""name"""'], {}), "(self, 'name')\n", (438042, 438056), False, 'import pulumi\n'), ((438129, 438160), 'pulumi.set', 'pulumi.set', (['self', '"""name"""', 'value'], {}), "(self, 'name', value)\n", (438139, 438160), False, 'import pulumi\n'), ((438414, 438438), 'pulumi.get', 'pulumi.get', (['self', '"""type"""'], {}), "(self, 'type')\n", (438424, 438438), False, 'import pulumi\n'), ((438511, 438542), 'pulumi.set', 'pulumi.set', (['self', '"""type"""', 'value'], {}), "(self, 'type', value)\n", (438521, 438542), False, 'import pulumi\n'), ((438737, 438760), 'pulumi.get', 'pulumi.get', (['self', '"""uri"""'], {}), "(self, 'uri')\n", (438747, 438760), False, 'import pulumi\n'), ((438831, 438861), 'pulumi.set', 'pulumi.set', (['self', '"""uri"""', 'value'], {}), "(self, 'uri', value)\n", (438841, 438861), False, 'import pulumi\n'), ((439445, 439479), 'pulumi.set', 'pulumi.set', (['__self__', '"""name"""', 'name'], {}), "(__self__, 'name', name)\n", (439455, 439479), False, 'import pulumi\n'), ((439488, 439536), 'pulumi.set', 'pulumi.set', (['__self__', '"""shard_names"""', 'shard_names'], {}), "(__self__, 'shard_names', shard_names)\n", (439498, 439536), False, 'import pulumi\n'), ((439792, 439816), 'pulumi.get', 'pulumi.get', (['self', '"""name"""'], {}), "(self, 'name')\n", (439802, 439816), False, 'import pulumi\n'), ((439889, 439920), 'pulumi.set', 'pulumi.set', (['self', '"""name"""', 'value'], {}), "(self, 'name', value)\n", (439899, 439920), False, 'import pulumi\n'), ((440146, 440177), 'pulumi.get', 'pulumi.get', (['self', '"""shard_names"""'], {}), "(self, 'shard_names')\n", (440156, 440177), False, 'import pulumi\n'), ((440288, 440326), 'pulumi.set', 'pulumi.set', (['self', '"""shard_names"""', 'value'], {}), "(self, 'shard_names', value)\n", (440298, 440326), False, 'import pulumi\n'), ((440498, 440529), 'pulumi.get', 'pulumi.get', (['self', '"""description"""'], {}), "(self, 'description')\n", (440508, 440529), False, 'import pulumi\n'), ((440626, 440664), 'pulumi.set', 'pulumi.set', (['self', '"""description"""', 'value'], {}), "(self, 'description', value)\n", (440636, 440664), False, 'import pulumi\n'), ((441806, 441840), 'pulumi.set', 'pulumi.set', (['__self__', '"""name"""', 'name'], {}), "(__self__, 'name', name)\n", (441816, 441840), False, 'import pulumi\n'), ((441849, 441891), 'pulumi.set', 'pulumi.set', (['__self__', '"""password"""', 'password'], {}), "(__self__, 'password', password)\n", (441859, 441891), False, 'import pulumi\n'), ((442317, 442341), 'pulumi.get', 'pulumi.get', (['self', '"""name"""'], {}), "(self, 'name')\n", (442327, 442341), False, 'import pulumi\n'), ((442414, 442445), 'pulumi.set', 'pulumi.set', (['self', '"""name"""', 'value'], {}), "(self, 'name', value)\n", (442424, 442445), False, 'import pulumi\n'), ((442596, 442624), 'pulumi.get', 'pulumi.get', (['self', '"""password"""'], {}), "(self, 'password')\n", (442606, 442624), False, 'import pulumi\n'), ((442705, 442740), 'pulumi.set', 'pulumi.set', (['self', '"""password"""', 'value'], {}), "(self, 'password', value)\n", (442715, 442740), False, 'import pulumi\n'), ((443016, 443047), 'pulumi.get', 'pulumi.get', (['self', '"""permissions"""'], {}), "(self, 'permissions')\n", (443026, 443047), False, 'import pulumi\n'), ((443205, 443243), 'pulumi.set', 'pulumi.set', (['self', '"""permissions"""', 'value'], {}), "(self, 'permissions', value)\n", (443215, 443243), False, 'import pulumi\n'), ((443489, 443515), 'pulumi.get', 'pulumi.get', (['self', '"""quotas"""'], {}), "(self, 'quotas')\n", (443499, 443515), False, 'import pulumi\n'), ((443658, 443691), 'pulumi.set', 'pulumi.set', (['self', '"""quotas"""', 'value'], {}), "(self, 'quotas', value)\n", (443668, 443691), False, 'import pulumi\n'), ((443912, 443940), 'pulumi.get', 'pulumi.get', (['self', '"""settings"""'], {}), "(self, 'settings')\n", (443922, 443940), False, 'import pulumi\n'), ((444066, 444101), 'pulumi.set', 'pulumi.set', (['self', '"""settings"""', 'value'], {}), "(self, 'settings', value)\n", (444076, 444101), False, 'import pulumi\n'), ((444394, 444446), 'pulumi.set', 'pulumi.set', (['__self__', '"""database_name"""', 'database_name'], {}), "(__self__, 'database_name', database_name)\n", (444404, 444446), False, 'import pulumi\n'), ((444662, 444695), 'pulumi.get', 'pulumi.get', (['self', '"""database_name"""'], {}), "(self, 'database_name')\n", (444672, 444695), False, 'import pulumi\n'), ((444786, 444826), 'pulumi.set', 'pulumi.set', (['self', '"""database_name"""', 'value'], {}), "(self, 'database_name', value)\n", (444796, 444826), False, 'import pulumi\n'), ((445926, 445986), 'pulumi.set', 'pulumi.set', (['__self__', '"""interval_duration"""', 'interval_duration'], {}), "(__self__, 'interval_duration', interval_duration)\n", (445936, 445986), False, 'import pulumi\n'), ((446656, 446693), 'pulumi.get', 'pulumi.get', (['self', '"""interval_duration"""'], {}), "(self, 'interval_duration')\n", (446666, 446693), False, 'import pulumi\n'), ((446792, 446836), 'pulumi.set', 'pulumi.set', (['self', '"""interval_duration"""', 'value'], {}), "(self, 'interval_duration', value)\n", (446802, 446836), False, 'import pulumi\n'), ((447015, 447041), 'pulumi.get', 'pulumi.get', (['self', '"""errors"""'], {}), "(self, 'errors')\n", (447025, 447041), False, 'import pulumi\n'), ((447128, 447161), 'pulumi.set', 'pulumi.set', (['self', '"""errors"""', 'value'], {}), "(self, 'errors', value)\n", (447138, 447161), False, 'import pulumi\n'), ((447387, 447421), 'pulumi.get', 'pulumi.get', (['self', '"""execution_time"""'], {}), "(self, 'execution_time')\n", (447397, 447421), False, 'import pulumi\n'), ((447524, 447565), 'pulumi.set', 'pulumi.set', (['self', '"""execution_time"""', 'value'], {}), "(self, 'execution_time', value)\n", (447534, 447565), False, 'import pulumi\n'), ((447730, 447757), 'pulumi.get', 'pulumi.get', (['self', '"""queries"""'], {}), "(self, 'queries')\n", (447740, 447757), False, 'import pulumi\n'), ((447846, 447880), 'pulumi.set', 'pulumi.set', (['self', '"""queries"""', 'value'], {}), "(self, 'queries', value)\n", (447856, 447880), False, 'import pulumi\n'), ((448130, 448159), 'pulumi.get', 'pulumi.get', (['self', '"""read_rows"""'], {}), "(self, 'read_rows')\n", (448140, 448159), False, 'import pulumi\n'), ((448252, 448288), 'pulumi.set', 'pulumi.set', (['self', '"""read_rows"""', 'value'], {}), "(self, 'read_rows', value)\n", (448262, 448288), False, 'import pulumi\n'), ((448493, 448524), 'pulumi.get', 'pulumi.get', (['self', '"""result_rows"""'], {}), "(self, 'result_rows')\n", (448503, 448524), False, 'import pulumi\n'), ((448621, 448659), 'pulumi.set', 'pulumi.set', (['self', '"""result_rows"""', 'value'], {}), "(self, 'result_rows', value)\n", (448631, 448659), False, 'import pulumi\n'), ((481275, 481315), 'pulumi.get', 'pulumi.get', (['self', '"""add_http_cors_header"""'], {}), "(self, 'add_http_cors_header')\n", (481285, 481315), False, 'import pulumi\n'), ((481431, 481478), 'pulumi.set', 'pulumi.set', (['self', '"""add_http_cors_header"""', 'value'], {}), "(self, 'add_http_cors_header', value)\n", (481441, 481478), False, 'import pulumi\n'), ((481664, 481693), 'pulumi.get', 'pulumi.get', (['self', '"""allow_ddl"""'], {}), "(self, 'allow_ddl')\n", (481674, 481693), False, 'import pulumi\n'), ((481787, 481823), 'pulumi.set', 'pulumi.set', (['self', '"""allow_ddl"""', 'value'], {}), "(self, 'allow_ddl', value)\n", (481797, 481823), False, 'import pulumi\n'), ((481991, 482018), 'pulumi.get', 'pulumi.get', (['self', '"""compile"""'], {}), "(self, 'compile')\n", (482001, 482018), False, 'import pulumi\n'), ((482108, 482142), 'pulumi.set', 'pulumi.set', (['self', '"""compile"""', 'value'], {}), "(self, 'compile', value)\n", (482118, 482142), False, 'import pulumi\n'), ((482350, 482389), 'pulumi.get', 'pulumi.get', (['self', '"""compile_expressions"""'], {}), "(self, 'compile_expressions')\n", (482360, 482389), False, 'import pulumi\n'), ((482503, 482549), 'pulumi.set', 'pulumi.set', (['self', '"""compile_expressions"""', 'value'], {}), "(self, 'compile_expressions', value)\n", (482513, 482549), False, 'import pulumi\n'), ((482802, 482837), 'pulumi.get', 'pulumi.get', (['self', '"""connect_timeout"""'], {}), "(self, 'connect_timeout')\n", (482812, 482837), False, 'import pulumi\n'), ((482942, 482984), 'pulumi.set', 'pulumi.set', (['self', '"""connect_timeout"""', 'value'], {}), "(self, 'connect_timeout', value)\n", (482952, 482984), False, 'import pulumi\n'), ((483279, 483328), 'pulumi.get', 'pulumi.get', (['self', '"""count_distinct_implementation"""'], {}), "(self, 'count_distinct_implementation')\n", (483289, 483328), False, 'import pulumi\n'), ((483461, 483517), 'pulumi.set', 'pulumi.set', (['self', '"""count_distinct_implementation"""', 'value'], {}), "(self, 'count_distinct_implementation', value)\n", (483471, 483517), False, 'import pulumi\n'), ((483762, 483804), 'pulumi.get', 'pulumi.get', (['self', '"""distinct_overflow_mode"""'], {}), "(self, 'distinct_overflow_mode')\n", (483772, 483804), False, 'import pulumi\n'), ((483923, 483972), 'pulumi.set', 'pulumi.set', (['self', '"""distinct_overflow_mode"""', 'value'], {}), "(self, 'distinct_overflow_mode', value)\n", (483933, 483972), False, 'import pulumi\n'), ((484238, 484298), 'pulumi.get', 'pulumi.get', (['self', '"""distributed_aggregation_memory_efficient"""'], {}), "(self, 'distributed_aggregation_memory_efficient')\n", (484248, 484298), False, 'import pulumi\n'), ((484454, 484521), 'pulumi.set', 'pulumi.set', (['self', '"""distributed_aggregation_memory_efficient"""', 'value'], {}), "(self, 'distributed_aggregation_memory_efficient', value)\n", (484464, 484521), False, 'import pulumi\n'), ((484754, 484802), 'pulumi.get', 'pulumi.get', (['self', '"""distributed_ddl_task_timeout"""'], {}), "(self, 'distributed_ddl_task_timeout')\n", (484764, 484802), False, 'import pulumi\n'), ((484933, 484988), 'pulumi.set', 'pulumi.set', (['self', '"""distributed_ddl_task_timeout"""', 'value'], {}), "(self, 'distributed_ddl_task_timeout', value)\n", (484943, 484988), False, 'import pulumi\n'), ((485221, 485265), 'pulumi.get', 'pulumi.get', (['self', '"""distributed_product_mode"""'], {}), "(self, 'distributed_product_mode')\n", (485231, 485265), False, 'import pulumi\n'), ((485388, 485439), 'pulumi.set', 'pulumi.set', (['self', '"""distributed_product_mode"""', 'value'], {}), "(self, 'distributed_product_mode', value)\n", (485398, 485439), False, 'import pulumi\n'), ((485685, 485746), 'pulumi.get', 'pulumi.get', (['self', '"""empty_result_for_aggregation_by_empty_set"""'], {}), "(self, 'empty_result_for_aggregation_by_empty_set')\n", (485695, 485746), False, 'import pulumi\n'), ((485904, 485972), 'pulumi.set', 'pulumi.set', (['self', '"""empty_result_for_aggregation_by_empty_set"""', 'value'], {}), "(self, 'empty_result_for_aggregation_by_empty_set', value)\n", (485914, 485972), False, 'import pulumi\n'), ((486228, 486271), 'pulumi.get', 'pulumi.get', (['self', '"""enable_http_compression"""'], {}), "(self, 'enable_http_compression')\n", (486238, 486271), False, 'import pulumi\n'), ((486393, 486443), 'pulumi.set', 'pulumi.set', (['self', '"""enable_http_compression"""', 'value'], {}), "(self, 'enable_http_compression', value)\n", (486403, 486443), False, 'import pulumi\n'), ((486751, 486821), 'pulumi.get', 'pulumi.get', (['self', '"""fallback_to_stale_replicas_for_distributed_queries"""'], {}), "(self, 'fallback_to_stale_replicas_for_distributed_queries')\n", (486761, 486821), False, 'import pulumi\n'), ((486997, 487074), 'pulumi.set', 'pulumi.set', (['self', '"""fallback_to_stale_replicas_for_distributed_queries"""', 'value'], {}), "(self, 'fallback_to_stale_replicas_for_distributed_queries', value)\n", (487007, 487074), False, 'import pulumi\n'), ((487309, 487348), 'pulumi.get', 'pulumi.get', (['self', '"""force_index_by_date"""'], {}), "(self, 'force_index_by_date')\n", (487319, 487348), False, 'import pulumi\n'), ((487462, 487508), 'pulumi.set', 'pulumi.set', (['self', '"""force_index_by_date"""', 'value'], {}), "(self, 'force_index_by_date', value)\n", (487472, 487508), False, 'import pulumi\n'), ((487752, 487789), 'pulumi.get', 'pulumi.get', (['self', '"""force_primary_key"""'], {}), "(self, 'force_primary_key')\n", (487762, 487789), False, 'import pulumi\n'), ((487899, 487943), 'pulumi.set', 'pulumi.set', (['self', '"""force_primary_key"""', 'value'], {}), "(self, 'force_primary_key', value)\n", (487909, 487943), False, 'import pulumi\n'), ((488192, 488234), 'pulumi.get', 'pulumi.get', (['self', '"""group_by_overflow_mode"""'], {}), "(self, 'group_by_overflow_mode')\n", (488202, 488234), False, 'import pulumi\n'), ((488353, 488402), 'pulumi.set', 'pulumi.set', (['self', '"""group_by_overflow_mode"""', 'value'], {}), "(self, 'group_by_overflow_mode', value)\n", (488363, 488402), False, 'import pulumi\n'), ((488687, 488735), 'pulumi.get', 'pulumi.get', (['self', '"""group_by_two_level_threshold"""'], {}), "(self, 'group_by_two_level_threshold')\n", (488697, 488735), False, 'import pulumi\n'), ((488866, 488921), 'pulumi.set', 'pulumi.set', (['self', '"""group_by_two_level_threshold"""', 'value'], {}), "(self, 'group_by_two_level_threshold', value)\n", (488876, 488921), False, 'import pulumi\n'), ((489218, 489272), 'pulumi.get', 'pulumi.get', (['self', '"""group_by_two_level_threshold_bytes"""'], {}), "(self, 'group_by_two_level_threshold_bytes')\n", (489228, 489272), False, 'import pulumi\n'), ((489415, 489476), 'pulumi.set', 'pulumi.set', (['self', '"""group_by_two_level_threshold_bytes"""', 'value'], {}), "(self, 'group_by_two_level_threshold_bytes', value)\n", (489425, 489476), False, 'import pulumi\n'), ((489703, 489746), 'pulumi.get', 'pulumi.get', (['self', '"""http_connection_timeout"""'], {}), "(self, 'http_connection_timeout')\n", (489713, 489746), False, 'import pulumi\n'), ((489867, 489917), 'pulumi.set', 'pulumi.set', (['self', '"""http_connection_timeout"""', 'value'], {}), "(self, 'http_connection_timeout', value)\n", (489877, 489917), False, 'import pulumi\n'), ((490216, 490266), 'pulumi.get', 'pulumi.get', (['self', '"""http_headers_progress_interval"""'], {}), "(self, 'http_headers_progress_interval')\n", (490226, 490266), False, 'import pulumi\n'), ((490401, 490458), 'pulumi.set', 'pulumi.set', (['self', '"""http_headers_progress_interval"""', 'value'], {}), "(self, 'http_headers_progress_interval', value)\n", (490411, 490458), False, 'import pulumi\n'), ((490679, 490719), 'pulumi.get', 'pulumi.get', (['self', '"""http_receive_timeout"""'], {}), "(self, 'http_receive_timeout')\n", (490689, 490719), False, 'import pulumi\n'), ((490834, 490881), 'pulumi.set', 'pulumi.set', (['self', '"""http_receive_timeout"""', 'value'], {}), "(self, 'http_receive_timeout', value)\n", (490844, 490881), False, 'import pulumi\n'), ((491096, 491133), 'pulumi.get', 'pulumi.get', (['self', '"""http_send_timeout"""'], {}), "(self, 'http_send_timeout')\n", (491106, 491133), False, 'import pulumi\n'), ((491242, 491286), 'pulumi.set', 'pulumi.set', (['self', '"""http_send_timeout"""', 'value'], {}), "(self, 'http_send_timeout', value)\n", (491252, 491286), False, 'import pulumi\n'), ((491615, 491675), 'pulumi.get', 'pulumi.get', (['self', '"""input_format_defaults_for_omitted_fields"""'], {}), "(self, 'input_format_defaults_for_omitted_fields')\n", (491625, 491675), False, 'import pulumi\n'), ((491831, 491898), 'pulumi.set', 'pulumi.set', (['self', '"""input_format_defaults_for_omitted_fields"""', 'value'], {}), "(self, 'input_format_defaults_for_omitted_fields', value)\n", (491841, 491898), False, 'import pulumi\n'), ((492203, 492264), 'pulumi.get', 'pulumi.get', (['self', '"""input_format_values_interpret_expressions"""'], {}), "(self, 'input_format_values_interpret_expressions')\n", (492213, 492264), False, 'import pulumi\n'), ((492422, 492490), 'pulumi.set', 'pulumi.set', (['self', '"""input_format_values_interpret_expressions"""', 'value'], {}), "(self, 'input_format_values_interpret_expressions', value)\n", (492432, 492490), False, 'import pulumi\n'), ((492680, 492713), 'pulumi.get', 'pulumi.get', (['self', '"""insert_quorum"""'], {}), "(self, 'insert_quorum')\n", (492690, 492713), False, 'import pulumi\n'), ((492814, 492854), 'pulumi.set', 'pulumi.set', (['self', '"""insert_quorum"""', 'value'], {}), "(self, 'insert_quorum', value)\n", (492824, 492854), False, 'import pulumi\n'), ((493075, 493116), 'pulumi.get', 'pulumi.get', (['self', '"""insert_quorum_timeout"""'], {}), "(self, 'insert_quorum_timeout')\n", (493085, 493116), False, 'import pulumi\n'), ((493233, 493281), 'pulumi.set', 'pulumi.set', (['self', '"""insert_quorum_timeout"""', 'value'], {}), "(self, 'insert_quorum_timeout', value)\n", (493243, 493281), False, 'import pulumi\n'), ((493506, 493544), 'pulumi.get', 'pulumi.get', (['self', '"""join_overflow_mode"""'], {}), "(self, 'join_overflow_mode')\n", (493516, 493544), False, 'import pulumi\n'), ((493655, 493700), 'pulumi.set', 'pulumi.set', (['self', '"""join_overflow_mode"""', 'value'], {}), "(self, 'join_overflow_mode', value)\n", (493665, 493700), False, 'import pulumi\n'), ((494000, 494034), 'pulumi.get', 'pulumi.get', (['self', '"""join_use_nulls"""'], {}), "(self, 'join_use_nulls')\n", (494010, 494034), False, 'import pulumi\n'), ((494138, 494179), 'pulumi.set', 'pulumi.set', (['self', '"""join_use_nulls"""', 'value'], {}), "(self, 'join_use_nulls', value)\n", (494148, 494179), False, 'import pulumi\n'), ((494471, 494521), 'pulumi.get', 'pulumi.get', (['self', '"""joined_subquery_requires_alias"""'], {}), "(self, 'joined_subquery_requires_alias')\n", (494481, 494521), False, 'import pulumi\n'), ((494657, 494714), 'pulumi.set', 'pulumi.set', (['self', '"""joined_subquery_requires_alias"""', 'value'], {}), "(self, 'joined_subquery_requires_alias', value)\n", (494667, 494714), False, 'import pulumi\n'), ((495003, 495061), 'pulumi.get', 'pulumi.get', (['self', '"""low_cardinality_allow_in_native_format"""'], {}), "(self, 'low_cardinality_allow_in_native_format')\n", (495013, 495061), False, 'import pulumi\n'), ((495213, 495278), 'pulumi.set', 'pulumi.set', (['self', '"""low_cardinality_allow_in_native_format"""', 'value'], {}), "(self, 'low_cardinality_allow_in_native_format', value)\n", (495223, 495278), False, 'import pulumi\n'), ((495476, 495509), 'pulumi.get', 'pulumi.get', (['self', '"""max_ast_depth"""'], {}), "(self, 'max_ast_depth')\n", (495486, 495509), False, 'import pulumi\n'), ((495610, 495650), 'pulumi.set', 'pulumi.set', (['self', '"""max_ast_depth"""', 'value'], {}), "(self, 'max_ast_depth', value)\n", (495620, 495650), False, 'import pulumi\n'), ((495857, 495893), 'pulumi.get', 'pulumi.get', (['self', '"""max_ast_elements"""'], {}), "(self, 'max_ast_elements')\n", (495867, 495893), False, 'import pulumi\n'), ((496000, 496043), 'pulumi.set', 'pulumi.set', (['self', '"""max_ast_elements"""', 'value'], {}), "(self, 'max_ast_elements', value)\n", (496010, 496043), False, 'import pulumi\n'), ((496293, 496327), 'pulumi.get', 'pulumi.get', (['self', '"""max_block_size"""'], {}), "(self, 'max_block_size')\n", (496303, 496327), False, 'import pulumi\n'), ((496430, 496471), 'pulumi.set', 'pulumi.set', (['self', '"""max_block_size"""', 'value'], {}), "(self, 'max_block_size', value)\n", (496440, 496471), False, 'import pulumi\n'), ((496744, 496798), 'pulumi.get', 'pulumi.get', (['self', '"""max_bytes_before_external_group_by"""'], {}), "(self, 'max_bytes_before_external_group_by')\n", (496754, 496798), False, 'import pulumi\n'), ((496941, 497002), 'pulumi.set', 'pulumi.set', (['self', '"""max_bytes_before_external_group_by"""', 'value'], {}), "(self, 'max_bytes_before_external_group_by', value)\n", (496951, 497002), False, 'import pulumi\n'), ((497339, 497389), 'pulumi.get', 'pulumi.get', (['self', '"""max_bytes_before_external_sort"""'], {}), "(self, 'max_bytes_before_external_sort')\n", (497349, 497389), False, 'import pulumi\n'), ((497524, 497581), 'pulumi.set', 'pulumi.set', (['self', '"""max_bytes_before_external_sort"""', 'value'], {}), "(self, 'max_bytes_before_external_sort', value)\n", (497534, 497581), False, 'import pulumi\n'), ((497848, 497889), 'pulumi.get', 'pulumi.get', (['self', '"""max_bytes_in_distinct"""'], {}), "(self, 'max_bytes_in_distinct')\n", (497858, 497889), False, 'import pulumi\n'), ((498006, 498054), 'pulumi.set', 'pulumi.set', (['self', '"""max_bytes_in_distinct"""', 'value'], {}), "(self, 'max_bytes_in_distinct', value)\n", (498016, 498054), False, 'import pulumi\n'), ((498283, 498320), 'pulumi.get', 'pulumi.get', (['self', '"""max_bytes_in_join"""'], {}), "(self, 'max_bytes_in_join')\n", (498293, 498320), False, 'import pulumi\n'), ((498429, 498473), 'pulumi.set', 'pulumi.set', (['self', '"""max_bytes_in_join"""', 'value'], {}), "(self, 'max_bytes_in_join', value)\n", (498439, 498473), False, 'import pulumi\n'), ((498728, 498764), 'pulumi.get', 'pulumi.get', (['self', '"""max_bytes_in_set"""'], {}), "(self, 'max_bytes_in_set')\n", (498738, 498764), False, 'import pulumi\n'), ((498871, 498914), 'pulumi.set', 'pulumi.set', (['self', '"""max_bytes_in_set"""', 'value'], {}), "(self, 'max_bytes_in_set', value)\n", (498881, 498914), False, 'import pulumi\n'), ((499190, 499227), 'pulumi.get', 'pulumi.get', (['self', '"""max_bytes_to_read"""'], {}), "(self, 'max_bytes_to_read')\n", (499200, 499227), False, 'import pulumi\n'), ((499336, 499380), 'pulumi.set', 'pulumi.set', (['self', '"""max_bytes_to_read"""', 'value'], {}), "(self, 'max_bytes_to_read', value)\n", (499346, 499380), False, 'import pulumi\n'), ((499647, 499684), 'pulumi.get', 'pulumi.get', (['self', '"""max_bytes_to_sort"""'], {}), "(self, 'max_bytes_to_sort')\n", (499657, 499684), False, 'import pulumi\n'), ((499793, 499837), 'pulumi.set', 'pulumi.set', (['self', '"""max_bytes_to_sort"""', 'value'], {}), "(self, 'max_bytes_to_sort', value)\n", (499803, 499837), False, 'import pulumi\n'), ((500159, 500200), 'pulumi.get', 'pulumi.get', (['self', '"""max_bytes_to_transfer"""'], {}), "(self, 'max_bytes_to_transfer')\n", (500169, 500200), False, 'import pulumi\n'), ((500317, 500365), 'pulumi.set', 'pulumi.set', (['self', '"""max_bytes_to_transfer"""', 'value'], {}), "(self, 'max_bytes_to_transfer', value)\n", (500327, 500365), False, 'import pulumi\n'), ((500624, 500663), 'pulumi.get', 'pulumi.get', (['self', '"""max_columns_to_read"""'], {}), "(self, 'max_columns_to_read')\n", (500634, 500663), False, 'import pulumi\n'), ((500776, 500822), 'pulumi.set', 'pulumi.set', (['self', '"""max_columns_to_read"""', 'value'], {}), "(self, 'max_columns_to_read', value)\n", (500786, 500822), False, 'import pulumi\n'), ((501051, 501089), 'pulumi.get', 'pulumi.get', (['self', '"""max_execution_time"""'], {}), "(self, 'max_execution_time')\n", (501061, 501089), False, 'import pulumi\n'), ((501200, 501245), 'pulumi.set', 'pulumi.set', (['self', '"""max_execution_time"""', 'value'], {}), "(self, 'max_execution_time', value)\n", (501210, 501245), False, 'import pulumi\n'), ((501499, 501544), 'pulumi.get', 'pulumi.get', (['self', '"""max_expanded_ast_elements"""'], {}), "(self, 'max_expanded_ast_elements')\n", (501509, 501544), False, 'import pulumi\n'), ((501669, 501721), 'pulumi.set', 'pulumi.set', (['self', '"""max_expanded_ast_elements"""', 'value'], {}), "(self, 'max_expanded_ast_elements', value)\n", (501679, 501721), False, 'import pulumi\n'), ((501974, 502015), 'pulumi.get', 'pulumi.get', (['self', '"""max_insert_block_size"""'], {}), "(self, 'max_insert_block_size')\n", (501984, 502015), False, 'import pulumi\n'), ((502132, 502180), 'pulumi.set', 'pulumi.set', (['self', '"""max_insert_block_size"""', 'value'], {}), "(self, 'max_insert_block_size', value)\n", (502142, 502180), False, 'import pulumi\n'), ((502434, 502470), 'pulumi.get', 'pulumi.get', (['self', '"""max_memory_usage"""'], {}), "(self, 'max_memory_usage')\n", (502444, 502470), False, 'import pulumi\n'), ((502577, 502620), 'pulumi.set', 'pulumi.set', (['self', '"""max_memory_usage"""', 'value'], {}), "(self, 'max_memory_usage', value)\n", (502587, 502620), False, 'import pulumi\n'), ((502900, 502945), 'pulumi.get', 'pulumi.get', (['self', '"""max_memory_usage_for_user"""'], {}), "(self, 'max_memory_usage_for_user')\n", (502910, 502945), False, 'import pulumi\n'), ((503070, 503122), 'pulumi.set', 'pulumi.set', (['self', '"""max_memory_usage_for_user"""', 'value'], {}), "(self, 'max_memory_usage_for_user', value)\n", (503080, 503122), False, 'import pulumi\n'), ((503376, 503417), 'pulumi.get', 'pulumi.get', (['self', '"""max_network_bandwidth"""'], {}), "(self, 'max_network_bandwidth')\n", (503386, 503417), False, 'import pulumi\n'), ((503534, 503582), 'pulumi.set', 'pulumi.set', (['self', '"""max_network_bandwidth"""', 'value'], {}), "(self, 'max_network_bandwidth', value)\n", (503544, 503582), False, 'import pulumi\n'), ((503852, 503902), 'pulumi.get', 'pulumi.get', (['self', '"""max_network_bandwidth_for_user"""'], {}), "(self, 'max_network_bandwidth_for_user')\n", (503862, 503902), False, 'import pulumi\n'), ((504037, 504094), 'pulumi.set', 'pulumi.set', (['self', '"""max_network_bandwidth_for_user"""', 'value'], {}), "(self, 'max_network_bandwidth_for_user', value)\n", (504047, 504094), False, 'import pulumi\n'), ((504344, 504378), 'pulumi.get', 'pulumi.get', (['self', '"""max_query_size"""'], {}), "(self, 'max_query_size')\n", (504354, 504378), False, 'import pulumi\n'), ((504481, 504522), 'pulumi.set', 'pulumi.set', (['self', '"""max_query_size"""', 'value'], {}), "(self, 'max_query_size', value)\n", (504491, 504522), False, 'import pulumi\n'), ((504788, 504849), 'pulumi.get', 'pulumi.get', (['self', '"""max_replica_delay_for_distributed_queries"""'], {}), "(self, 'max_replica_delay_for_distributed_queries')\n", (504798, 504849), False, 'import pulumi\n'), ((505006, 505074), 'pulumi.set', 'pulumi.set', (['self', '"""max_replica_delay_for_distributed_queries"""', 'value'], {}), "(self, 'max_replica_delay_for_distributed_queries', value)\n", (505016, 505074), False, 'import pulumi\n'), ((505284, 505320), 'pulumi.get', 'pulumi.get', (['self', '"""max_result_bytes"""'], {}), "(self, 'max_result_bytes')\n", (505294, 505320), False, 'import pulumi\n'), ((505427, 505470), 'pulumi.set', 'pulumi.set', (['self', '"""max_result_bytes"""', 'value'], {}), "(self, 'max_result_bytes', value)\n", (505437, 505470), False, 'import pulumi\n'), ((505677, 505712), 'pulumi.get', 'pulumi.get', (['self', '"""max_result_rows"""'], {}), "(self, 'max_result_rows')\n", (505687, 505712), False, 'import pulumi\n'), ((505817, 505859), 'pulumi.set', 'pulumi.set', (['self', '"""max_result_rows"""', 'value'], {}), "(self, 'max_result_rows', value)\n", (505827, 505859), False, 'import pulumi\n'), ((506099, 506139), 'pulumi.get', 'pulumi.get', (['self', '"""max_rows_in_distinct"""'], {}), "(self, 'max_rows_in_distinct')\n", (506109, 506139), False, 'import pulumi\n'), ((506254, 506301), 'pulumi.set', 'pulumi.set', (['self', '"""max_rows_in_distinct"""', 'value'], {}), "(self, 'max_rows_in_distinct', value)\n", (506264, 506301), False, 'import pulumi\n'), ((506527, 506563), 'pulumi.get', 'pulumi.get', (['self', '"""max_rows_in_join"""'], {}), "(self, 'max_rows_in_join')\n", (506537, 506563), False, 'import pulumi\n'), ((506670, 506713), 'pulumi.set', 'pulumi.set', (['self', '"""max_rows_in_join"""', 'value'], {}), "(self, 'max_rows_in_join', value)\n", (506680, 506713), False, 'import pulumi\n'), ((506965, 507000), 'pulumi.get', 'pulumi.get', (['self', '"""max_rows_in_set"""'], {}), "(self, 'max_rows_in_set')\n", (506975, 507000), False, 'import pulumi\n'), ((507105, 507147), 'pulumi.set', 'pulumi.set', (['self', '"""max_rows_in_set"""', 'value'], {}), "(self, 'max_rows_in_set', value)\n", (507115, 507147), False, 'import pulumi\n'), ((507398, 507438), 'pulumi.get', 'pulumi.get', (['self', '"""max_rows_to_group_by"""'], {}), "(self, 'max_rows_to_group_by')\n", (507408, 507438), False, 'import pulumi\n'), ((507553, 507600), 'pulumi.set', 'pulumi.set', (['self', '"""max_rows_to_group_by"""', 'value'], {}), "(self, 'max_rows_to_group_by', value)\n", (507563, 507600), False, 'import pulumi\n'), ((507853, 507889), 'pulumi.get', 'pulumi.get', (['self', '"""max_rows_to_read"""'], {}), "(self, 'max_rows_to_read')\n", (507863, 507889), False, 'import pulumi\n'), ((507996, 508039), 'pulumi.set', 'pulumi.set', (['self', '"""max_rows_to_read"""', 'value'], {}), "(self, 'max_rows_to_read', value)\n", (508006, 508039), False, 'import pulumi\n'), ((508283, 508319), 'pulumi.get', 'pulumi.get', (['self', '"""max_rows_to_sort"""'], {}), "(self, 'max_rows_to_sort')\n", (508293, 508319), False, 'import pulumi\n'), ((508426, 508469), 'pulumi.set', 'pulumi.set', (['self', '"""max_rows_to_sort"""', 'value'], {}), "(self, 'max_rows_to_sort', value)\n", (508436, 508469), False, 'import pulumi\n'), ((508768, 508808), 'pulumi.get', 'pulumi.get', (['self', '"""max_rows_to_transfer"""'], {}), "(self, 'max_rows_to_transfer')\n", (508778, 508808), False, 'import pulumi\n'), ((508923, 508970), 'pulumi.set', 'pulumi.set', (['self', '"""max_rows_to_transfer"""', 'value'], {}), "(self, 'max_rows_to_transfer', value)\n", (508933, 508970), False, 'import pulumi\n'), ((509287, 509328), 'pulumi.get', 'pulumi.get', (['self', '"""max_temporary_columns"""'], {}), "(self, 'max_temporary_columns')\n", (509297, 509328), False, 'import pulumi\n'), ((509445, 509493), 'pulumi.set', 'pulumi.set', (['self', '"""max_temporary_columns"""', 'value'], {}), "(self, 'max_temporary_columns', value)\n", (509455, 509493), False, 'import pulumi\n'), ((509828, 509879), 'pulumi.get', 'pulumi.get', (['self', '"""max_temporary_non_const_columns"""'], {}), "(self, 'max_temporary_non_const_columns')\n", (509838, 509879), False, 'import pulumi\n'), ((510016, 510074), 'pulumi.set', 'pulumi.set', (['self', '"""max_temporary_non_const_columns"""', 'value'], {}), "(self, 'max_temporary_non_const_columns', value)\n", (510026, 510074), False, 'import pulumi\n'), ((510340, 510371), 'pulumi.get', 'pulumi.get', (['self', '"""max_threads"""'], {}), "(self, 'max_threads')\n", (510350, 510371), False, 'import pulumi\n'), ((510468, 510506), 'pulumi.set', 'pulumi.set', (['self', '"""max_threads"""', 'value'], {}), "(self, 'max_threads', value)\n", (510478, 510506), False, 'import pulumi\n'), ((510843, 510896), 'pulumi.get', 'pulumi.get', (['self', '"""merge_tree_max_bytes_to_use_cache"""'], {}), "(self, 'merge_tree_max_bytes_to_use_cache')\n", (510853, 510896), False, 'import pulumi\n'), ((511037, 511097), 'pulumi.set', 'pulumi.set', (['self', '"""merge_tree_max_bytes_to_use_cache"""', 'value'], {}), "(self, 'merge_tree_max_bytes_to_use_cache', value)\n", (511047, 511097), False, 'import pulumi\n'), ((511430, 511482), 'pulumi.get', 'pulumi.get', (['self', '"""merge_tree_max_rows_to_use_cache"""'], {}), "(self, 'merge_tree_max_rows_to_use_cache')\n", (511440, 511482), False, 'import pulumi\n'), ((511621, 511680), 'pulumi.set', 'pulumi.set', (['self', '"""merge_tree_max_rows_to_use_cache"""', 'value'], {}), "(self, 'merge_tree_max_rows_to_use_cache', value)\n", (511631, 511680), False, 'import pulumi\n'), ((512093, 512153), 'pulumi.get', 'pulumi.get', (['self', '"""merge_tree_min_bytes_for_concurrent_read"""'], {}), "(self, 'merge_tree_min_bytes_for_concurrent_read')\n", (512103, 512153), False, 'import pulumi\n'), ((512308, 512375), 'pulumi.set', 'pulumi.set', (['self', '"""merge_tree_min_bytes_for_concurrent_read"""', 'value'], {}), "(self, 'merge_tree_min_bytes_for_concurrent_read', value)\n", (512318, 512375), False, 'import pulumi\n'), ((512788, 512847), 'pulumi.get', 'pulumi.get', (['self', '"""merge_tree_min_rows_for_concurrent_read"""'], {}), "(self, 'merge_tree_min_rows_for_concurrent_read')\n", (512798, 512847), False, 'import pulumi\n'), ((513000, 513066), 'pulumi.set', 'pulumi.set', (['self', '"""merge_tree_min_rows_for_concurrent_read"""', 'value'], {}), "(self, 'merge_tree_min_rows_for_concurrent_read', value)\n", (513010, 513066), False, 'import pulumi\n'), ((513333, 513379), 'pulumi.get', 'pulumi.get', (['self', '"""min_bytes_to_use_direct_io"""'], {}), "(self, 'min_bytes_to_use_direct_io')\n", (513343, 513379), False, 'import pulumi\n'), ((513506, 513559), 'pulumi.set', 'pulumi.set', (['self', '"""min_bytes_to_use_direct_io"""', 'value'], {}), "(self, 'min_bytes_to_use_direct_io', value)\n", (513516, 513559), False, 'import pulumi\n'), ((513821, 513861), 'pulumi.get', 'pulumi.get', (['self', '"""min_count_to_compile"""'], {}), "(self, 'min_count_to_compile')\n", (513831, 513861), False, 'import pulumi\n'), ((513976, 514023), 'pulumi.set', 'pulumi.set', (['self', '"""min_count_to_compile"""', 'value'], {}), "(self, 'min_count_to_compile', value)\n", (513986, 514023), False, 'import pulumi\n'), ((514311, 514362), 'pulumi.get', 'pulumi.get', (['self', '"""min_count_to_compile_expression"""'], {}), "(self, 'min_count_to_compile_expression')\n", (514321, 514362), False, 'import pulumi\n'), ((514499, 514557), 'pulumi.set', 'pulumi.set', (['self', '"""min_count_to_compile_expression"""', 'value'], {}), "(self, 'min_count_to_compile_expression', value)\n", (514509, 514557), False, 'import pulumi\n'), ((514775, 514814), 'pulumi.get', 'pulumi.get', (['self', '"""min_execution_speed"""'], {}), "(self, 'min_execution_speed')\n", (514785, 514814), False, 'import pulumi\n'), ((514927, 514973), 'pulumi.set', 'pulumi.set', (['self', '"""min_execution_speed"""', 'value'], {}), "(self, 'min_execution_speed', value)\n", (514937, 514973), False, 'import pulumi\n'), ((515203, 515248), 'pulumi.get', 'pulumi.get', (['self', '"""min_execution_speed_bytes"""'], {}), "(self, 'min_execution_speed_bytes')\n", (515213, 515248), False, 'import pulumi\n'), ((515373, 515425), 'pulumi.set', 'pulumi.set', (['self', '"""min_execution_speed_bytes"""', 'value'], {}), "(self, 'min_execution_speed_bytes', value)\n", (515383, 515425), False, 'import pulumi\n'), ((515714, 515761), 'pulumi.get', 'pulumi.get', (['self', '"""min_insert_block_size_bytes"""'], {}), "(self, 'min_insert_block_size_bytes')\n", (515724, 515761), False, 'import pulumi\n'), ((515890, 515944), 'pulumi.set', 'pulumi.set', (['self', '"""min_insert_block_size_bytes"""', 'value'], {}), "(self, 'min_insert_block_size_bytes', value)\n", (515900, 515944), False, 'import pulumi\n'), ((516230, 516276), 'pulumi.get', 'pulumi.get', (['self', '"""min_insert_block_size_rows"""'], {}), "(self, 'min_insert_block_size_rows')\n", (516240, 516276), False, 'import pulumi\n'), ((516403, 516456), 'pulumi.set', 'pulumi.set', (['self', '"""min_insert_block_size_rows"""', 'value'], {}), "(self, 'min_insert_block_size_rows', value)\n", (516413, 516456), False, 'import pulumi\n'), ((516866, 516924), 'pulumi.get', 'pulumi.get', (['self', '"""output_format_json_quote64bit_integers"""'], {}), "(self, 'output_format_json_quote64bit_integers')\n", (516876, 516924), False, 'import pulumi\n'), ((517076, 517141), 'pulumi.set', 'pulumi.set', (['self', '"""output_format_json_quote64bit_integers"""', 'value'], {}), "(self, 'output_format_json_quote64bit_integers', value)\n", (517086, 517141), False, 'import pulumi\n'), ((517406, 517460), 'pulumi.get', 'pulumi.get', (['self', '"""output_format_json_quote_denormals"""'], {}), "(self, 'output_format_json_quote_denormals')\n", (517416, 517460), False, 'import pulumi\n'), ((517604, 517665), 'pulumi.set', 'pulumi.set', (['self', '"""output_format_json_quote_denormals"""', 'value'], {}), "(self, 'output_format_json_quote_denormals', value)\n", (517614, 517665), False, 'import pulumi\n'), ((517818, 517846), 'pulumi.get', 'pulumi.get', (['self', '"""priority"""'], {}), "(self, 'priority')\n", (517828, 517846), False, 'import pulumi\n'), ((517937, 517972), 'pulumi.set', 'pulumi.set', (['self', '"""priority"""', 'value'], {}), "(self, 'priority', value)\n", (517947, 517972), False, 'import pulumi\n'), ((518152, 518182), 'pulumi.get', 'pulumi.get', (['self', '"""quota_mode"""'], {}), "(self, 'quota_mode')\n", (518162, 518182), False, 'import pulumi\n'), ((518277, 518314), 'pulumi.set', 'pulumi.set', (['self', '"""quota_mode"""', 'value'], {}), "(self, 'quota_mode', value)\n", (518287, 518314), False, 'import pulumi\n'), ((518542, 518580), 'pulumi.get', 'pulumi.get', (['self', '"""read_overflow_mode"""'], {}), "(self, 'read_overflow_mode')\n", (518552, 518580), False, 'import pulumi\n'), ((518691, 518736), 'pulumi.set', 'pulumi.set', (['self', '"""read_overflow_mode"""', 'value'], {}), "(self, 'read_overflow_mode', value)\n", (518701, 518736), False, 'import pulumi\n'), ((518953, 518981), 'pulumi.get', 'pulumi.get', (['self', '"""readonly"""'], {}), "(self, 'readonly')\n", (518963, 518981), False, 'import pulumi\n'), ((519072, 519107), 'pulumi.set', 'pulumi.set', (['self', '"""readonly"""', 'value'], {}), "(self, 'readonly', value)\n", (519082, 519107), False, 'import pulumi\n'), ((519360, 519395), 'pulumi.get', 'pulumi.get', (['self', '"""receive_timeout"""'], {}), "(self, 'receive_timeout')\n", (519370, 519395), False, 'import pulumi\n'), ((519500, 519542), 'pulumi.set', 'pulumi.set', (['self', '"""receive_timeout"""', 'value'], {}), "(self, 'receive_timeout', value)\n", (519510, 519542), False, 'import pulumi\n'), ((519862, 519915), 'pulumi.get', 'pulumi.get', (['self', '"""replication_alter_partitions_sync"""'], {}), "(self, 'replication_alter_partitions_sync')\n", (519872, 519915), False, 'import pulumi\n'), ((520056, 520116), 'pulumi.set', 'pulumi.set', (['self', '"""replication_alter_partitions_sync"""', 'value'], {}), "(self, 'replication_alter_partitions_sync', value)\n", (520066, 520116), False, 'import pulumi\n'), ((520347, 520387), 'pulumi.get', 'pulumi.get', (['self', '"""result_overflow_mode"""'], {}), "(self, 'result_overflow_mode')\n", (520357, 520387), False, 'import pulumi\n'), ((520502, 520549), 'pulumi.set', 'pulumi.set', (['self', '"""result_overflow_mode"""', 'value'], {}), "(self, 'result_overflow_mode', value)\n", (520512, 520549), False, 'import pulumi\n'), ((520807, 520856), 'pulumi.get', 'pulumi.get', (['self', '"""select_sequential_consistency"""'], {}), "(self, 'select_sequential_consistency')\n", (520817, 520856), False, 'import pulumi\n'), ((520990, 521046), 'pulumi.set', 'pulumi.set', (['self', '"""select_sequential_consistency"""', 'value'], {}), "(self, 'select_sequential_consistency', value)\n", (521000, 521046), False, 'import pulumi\n'), ((521335, 521384), 'pulumi.get', 'pulumi.get', (['self', '"""send_progress_in_http_headers"""'], {}), "(self, 'send_progress_in_http_headers')\n", (521345, 521384), False, 'import pulumi\n'), ((521518, 521574), 'pulumi.set', 'pulumi.set', (['self', '"""send_progress_in_http_headers"""', 'value'], {}), "(self, 'send_progress_in_http_headers', value)\n", (521528, 521574), False, 'import pulumi\n'), ((521818, 521850), 'pulumi.get', 'pulumi.get', (['self', '"""send_timeout"""'], {}), "(self, 'send_timeout')\n", (521828, 521850), False, 'import pulumi\n'), ((521949, 521988), 'pulumi.set', 'pulumi.set', (['self', '"""send_timeout"""', 'value'], {}), "(self, 'send_timeout', value)\n", (521959, 521988), False, 'import pulumi\n'), ((522224, 522261), 'pulumi.get', 'pulumi.get', (['self', '"""set_overflow_mode"""'], {}), "(self, 'set_overflow_mode')\n", (522234, 522261), False, 'import pulumi\n'), ((522370, 522414), 'pulumi.set', 'pulumi.set', (['self', '"""set_overflow_mode"""', 'value'], {}), "(self, 'set_overflow_mode', value)\n", (522380, 522414), False, 'import pulumi\n'), ((522658, 522701), 'pulumi.get', 'pulumi.get', (['self', '"""skip_unavailable_shards"""'], {}), "(self, 'skip_unavailable_shards')\n", (522668, 522701), False, 'import pulumi\n'), ((522823, 522873), 'pulumi.set', 'pulumi.set', (['self', '"""skip_unavailable_shards"""', 'value'], {}), "(self, 'skip_unavailable_shards', value)\n", (522833, 522873), False, 'import pulumi\n'), ((523101, 523139), 'pulumi.get', 'pulumi.get', (['self', '"""sort_overflow_mode"""'], {}), "(self, 'sort_overflow_mode')\n", (523111, 523139), False, 'import pulumi\n'), ((523250, 523295), 'pulumi.set', 'pulumi.set', (['self', '"""sort_overflow_mode"""', 'value'], {}), "(self, 'sort_overflow_mode', value)\n", (523260, 523295), False, 'import pulumi\n'), ((523518, 523559), 'pulumi.get', 'pulumi.get', (['self', '"""timeout_overflow_mode"""'], {}), "(self, 'timeout_overflow_mode')\n", (523528, 523559), False, 'import pulumi\n'), ((523676, 523724), 'pulumi.set', 'pulumi.set', (['self', '"""timeout_overflow_mode"""', 'value'], {}), "(self, 'timeout_overflow_mode', value)\n", (523686, 523724), False, 'import pulumi\n'), ((523949, 523991), 'pulumi.get', 'pulumi.get', (['self', '"""transfer_overflow_mode"""'], {}), "(self, 'transfer_overflow_mode')\n", (523959, 523991), False, 'import pulumi\n'), ((524110, 524159), 'pulumi.set', 'pulumi.set', (['self', '"""transfer_overflow_mode"""', 'value'], {}), "(self, 'transfer_overflow_mode', value)\n", (524120, 524159), False, 'import pulumi\n'), ((524379, 524416), 'pulumi.get', 'pulumi.get', (['self', '"""transform_null_in"""'], {}), "(self, 'transform_null_in')\n", (524389, 524416), False, 'import pulumi\n'), ((524526, 524570), 'pulumi.set', 'pulumi.set', (['self', '"""transform_null_in"""', 'value'], {}), "(self, 'transform_null_in', value)\n", (524536, 524570), False, 'import pulumi\n'), ((524798, 524840), 'pulumi.get', 'pulumi.get', (['self', '"""use_uncompressed_cache"""'], {}), "(self, 'use_uncompressed_cache')\n", (524808, 524840), False, 'import pulumi\n'), ((524960, 525009), 'pulumi.set', 'pulumi.set', (['self', '"""use_uncompressed_cache"""', 'value'], {}), "(self, 'use_uncompressed_cache', value)\n", (524970, 525009), False, 'import pulumi\n'), ((525772, 525801), 'pulumi.get', 'pulumi.get', (['self', '"""resources"""'], {}), "(self, 'resources')\n", (525782, 525801), False, 'import pulumi\n'), ((525935, 525971), 'pulumi.set', 'pulumi.set', (['self', '"""resources"""', 'value'], {}), "(self, 'resources', value)\n", (525945, 525971), False, 'import pulumi\n'), ((527167, 527196), 'pulumi.get', 'pulumi.get', (['self', '"""disk_size"""'], {}), "(self, 'disk_size')\n", (527177, 527196), False, 'import pulumi\n'), ((527289, 527325), 'pulumi.set', 'pulumi.set', (['self', '"""disk_size"""', 'value'], {}), "(self, 'disk_size', value)\n", (527299, 527325), False, 'import pulumi\n'), ((527655, 527687), 'pulumi.get', 'pulumi.get', (['self', '"""disk_type_id"""'], {}), "(self, 'disk_type_id')\n", (527665, 527687), False, 'import pulumi\n'), ((527786, 527825), 'pulumi.set', 'pulumi.set', (['self', '"""disk_type_id"""', 'value'], {}), "(self, 'disk_type_id', value)\n", (527796, 527825), False, 'import pulumi\n'), ((527965, 528003), 'pulumi.get', 'pulumi.get', (['self', '"""resource_preset_id"""'], {}), "(self, 'resource_preset_id')\n", (527975, 528003), False, 'import pulumi\n'), ((528114, 528159), 'pulumi.set', 'pulumi.set', (['self', '"""resource_preset_id"""', 'value'], {}), "(self, 'resource_preset_id', value)\n", (528124, 528159), False, 'import pulumi\n'), ((529568, 529622), 'pulumi.set', 'pulumi.set', (['__self__', '"""admin_password"""', 'admin_password'], {}), "(__self__, 'admin_password', admin_password)\n", (529578, 529622), False, 'import pulumi\n'), ((529631, 529675), 'pulumi.set', 'pulumi.set', (['__self__', '"""data_node"""', 'data_node'], {}), "(__self__, 'data_node', data_node)\n", (529641, 529675), False, 'import pulumi\n'), ((530224, 530258), 'pulumi.get', 'pulumi.get', (['self', '"""admin_password"""'], {}), "(self, 'admin_password')\n", (530234, 530258), False, 'import pulumi\n'), ((530351, 530392), 'pulumi.set', 'pulumi.set', (['self', '"""admin_password"""', 'value'], {}), "(self, 'admin_password', value)\n", (530361, 530392), False, 'import pulumi\n'), ((530667, 530696), 'pulumi.get', 'pulumi.get', (['self', '"""data_node"""'], {}), "(self, 'data_node')\n", (530677, 530696), False, 'import pulumi\n'), ((530819, 530855), 'pulumi.set', 'pulumi.set', (['self', '"""data_node"""', 'value'], {}), "(self, 'data_node', value)\n", (530829, 530855), False, 'import pulumi\n'), ((531153, 531180), 'pulumi.get', 'pulumi.get', (['self', '"""edition"""'], {}), "(self, 'edition')\n", (531163, 531180), False, 'import pulumi\n'), ((531269, 531303), 'pulumi.set', 'pulumi.set', (['self', '"""edition"""', 'value'], {}), "(self, 'edition', value)\n", (531279, 531303), False, 'import pulumi\n'), ((531596, 531627), 'pulumi.get', 'pulumi.get', (['self', '"""master_node"""'], {}), "(self, 'master_node')\n", (531606, 531627), False, 'import pulumi\n'), ((531766, 531804), 'pulumi.set', 'pulumi.set', (['self', '"""master_node"""', 'value'], {}), "(self, 'master_node', value)\n", (531776, 531804), False, 'import pulumi\n'), ((532007, 532034), 'pulumi.get', 'pulumi.get', (['self', '"""plugins"""'], {}), "(self, 'plugins')\n", (532017, 532034), False, 'import pulumi\n'), ((532147, 532181), 'pulumi.set', 'pulumi.set', (['self', '"""plugins"""', 'value'], {}), "(self, 'plugins', value)\n", (532157, 532181), False, 'import pulumi\n'), ((532343, 532370), 'pulumi.get', 'pulumi.get', (['self', '"""version"""'], {}), "(self, 'version')\n", (532353, 532370), False, 'import pulumi\n'), ((532459, 532493), 'pulumi.set', 'pulumi.set', (['self', '"""version"""', 'value'], {}), "(self, 'version', value)\n", (532469, 532493), False, 'import pulumi\n'), ((532926, 532970), 'pulumi.set', 'pulumi.set', (['__self__', '"""resources"""', 'resources'], {}), "(__self__, 'resources', resources)\n", (532936, 532970), False, 'import pulumi\n'), ((533257, 533286), 'pulumi.get', 'pulumi.get', (['self', '"""resources"""'], {}), "(self, 'resources')\n", (533267, 533286), False, 'import pulumi\n'), ((533418, 533454), 'pulumi.set', 'pulumi.set', (['self', '"""resources"""', 'value'], {}), "(self, 'resources', value)\n", (533428, 533454), False, 'import pulumi\n'), ((533942, 533986), 'pulumi.set', 'pulumi.set', (['__self__', '"""disk_size"""', 'disk_size'], {}), "(__self__, 'disk_size', disk_size)\n", (533952, 533986), False, 'import pulumi\n'), ((533995, 534045), 'pulumi.set', 'pulumi.set', (['__self__', '"""disk_type_id"""', 'disk_type_id'], {}), "(__self__, 'disk_type_id', disk_type_id)\n", (534005, 534045), False, 'import pulumi\n'), ((534054, 534116), 'pulumi.set', 'pulumi.set', (['__self__', '"""resource_preset_id"""', 'resource_preset_id'], {}), "(__self__, 'resource_preset_id', resource_preset_id)\n", (534064, 534116), False, 'import pulumi\n'), ((534318, 534347), 'pulumi.get', 'pulumi.get', (['self', '"""disk_size"""'], {}), "(self, 'disk_size')\n", (534328, 534347), False, 'import pulumi\n'), ((534430, 534466), 'pulumi.set', 'pulumi.set', (['self', '"""disk_size"""', 'value'], {}), "(self, 'disk_size', value)\n", (534440, 534466), False, 'import pulumi\n'), ((534660, 534692), 'pulumi.get', 'pulumi.get', (['self', '"""disk_type_id"""'], {}), "(self, 'disk_type_id')\n", (534670, 534692), False, 'import pulumi\n'), ((534781, 534820), 'pulumi.set', 'pulumi.set', (['self', '"""disk_type_id"""', 'value'], {}), "(self, 'disk_type_id', value)\n", (534791, 534820), False, 'import pulumi\n'), ((534950, 534988), 'pulumi.get', 'pulumi.get', (['self', '"""resource_preset_id"""'], {}), "(self, 'resource_preset_id')\n", (534960, 534988), False, 'import pulumi\n'), ((535089, 535134), 'pulumi.set', 'pulumi.set', (['self', '"""resource_preset_id"""', 'value'], {}), "(self, 'resource_preset_id', value)\n", (535099, 535134), False, 'import pulumi\n'), ((535573, 535617), 'pulumi.set', 'pulumi.set', (['__self__', '"""resources"""', 'resources'], {}), "(__self__, 'resources', resources)\n", (535583, 535617), False, 'import pulumi\n'), ((535906, 535935), 'pulumi.get', 'pulumi.get', (['self', '"""resources"""'], {}), "(self, 'resources')\n", (535916, 535935), False, 'import pulumi\n'), ((536069, 536105), 'pulumi.set', 'pulumi.set', (['self', '"""resources"""', 'value'], {}), "(self, 'resources', value)\n", (536079, 536105), False, 'import pulumi\n'), ((536595, 536639), 'pulumi.set', 'pulumi.set', (['__self__', '"""disk_size"""', 'disk_size'], {}), "(__self__, 'disk_size', disk_size)\n", (536605, 536639), False, 'import pulumi\n'), ((536648, 536698), 'pulumi.set', 'pulumi.set', (['__self__', '"""disk_type_id"""', 'disk_type_id'], {}), "(__self__, 'disk_type_id', disk_type_id)\n", (536658, 536698), False, 'import pulumi\n'), ((536707, 536769), 'pulumi.set', 'pulumi.set', (['__self__', '"""resource_preset_id"""', 'resource_preset_id'], {}), "(__self__, 'resource_preset_id', resource_preset_id)\n", (536717, 536769), False, 'import pulumi\n'), ((536971, 537000), 'pulumi.get', 'pulumi.get', (['self', '"""disk_size"""'], {}), "(self, 'disk_size')\n", (536981, 537000), False, 'import pulumi\n'), ((537083, 537119), 'pulumi.set', 'pulumi.set', (['self', '"""disk_size"""', 'value'], {}), "(self, 'disk_size', value)\n", (537093, 537119), False, 'import pulumi\n'), ((537313, 537345), 'pulumi.get', 'pulumi.get', (['self', '"""disk_type_id"""'], {}), "(self, 'disk_type_id')\n", (537323, 537345), False, 'import pulumi\n'), ((537434, 537473), 'pulumi.set', 'pulumi.set', (['self', '"""disk_type_id"""', 'value'], {}), "(self, 'disk_type_id', value)\n", (537444, 537473), False, 'import pulumi\n'), ((537603, 537641), 'pulumi.get', 'pulumi.get', (['self', '"""resource_preset_id"""'], {}), "(self, 'resource_preset_id')\n", (537613, 537641), False, 'import pulumi\n'), ((537742, 537787), 'pulumi.set', 'pulumi.set', (['self', '"""resource_preset_id"""', 'value'], {}), "(self, 'resource_preset_id', value)\n", (537752, 537787), False, 'import pulumi\n'), ((539065, 539099), 'pulumi.set', 'pulumi.set', (['__self__', '"""name"""', 'name'], {}), "(__self__, 'name', name)\n", (539075, 539099), False, 'import pulumi\n'), ((539108, 539142), 'pulumi.set', 'pulumi.set', (['__self__', '"""type"""', 'type'], {}), "(__self__, 'type', type)\n", (539118, 539142), False, 'import pulumi\n'), ((539151, 539185), 'pulumi.set', 'pulumi.set', (['__self__', '"""zone"""', 'zone'], {}), "(__self__, 'zone', zone)\n", (539161, 539185), False, 'import pulumi\n'), ((539611, 539635), 'pulumi.get', 'pulumi.get', (['self', '"""name"""'], {}), "(self, 'name')\n", (539621, 539635), False, 'import pulumi\n'), ((539708, 539739), 'pulumi.set', 'pulumi.set', (['self', '"""name"""', 'value'], {}), "(self, 'name', value)\n", (539718, 539739), False, 'import pulumi\n'), ((539943, 539967), 'pulumi.get', 'pulumi.get', (['self', '"""type"""'], {}), "(self, 'type')\n", (539953, 539967), False, 'import pulumi\n'), ((540040, 540071), 'pulumi.set', 'pulumi.set', (['self', '"""type"""', 'value'], {}), "(self, 'type', value)\n", (540050, 540071), False, 'import pulumi\n'), ((540384, 540408), 'pulumi.get', 'pulumi.get', (['self', '"""zone"""'], {}), "(self, 'zone')\n", (540394, 540408), False, 'import pulumi\n'), ((540481, 540512), 'pulumi.set', 'pulumi.set', (['self', '"""zone"""', 'value'], {}), "(self, 'zone', value)\n", (540491, 540512), False, 'import pulumi\n'), ((540780, 540816), 'pulumi.get', 'pulumi.get', (['self', '"""assign_public_ip"""'], {}), "(self, 'assign_public_ip')\n", (540790, 540816), False, 'import pulumi\n'), ((540924, 540967), 'pulumi.set', 'pulumi.set', (['self', '"""assign_public_ip"""', 'value'], {}), "(self, 'assign_public_ip', value)\n", (540934, 540967), False, 'import pulumi\n'), ((541145, 541169), 'pulumi.get', 'pulumi.get', (['self', '"""fqdn"""'], {}), "(self, 'fqdn')\n", (541155, 541169), False, 'import pulumi\n'), ((541252, 541283), 'pulumi.set', 'pulumi.set', (['self', '"""fqdn"""', 'value'], {}), "(self, 'fqdn', value)\n", (541262, 541283), False, 'import pulumi\n'), ((541566, 541595), 'pulumi.get', 'pulumi.get', (['self', '"""subnet_id"""'], {}), "(self, 'subnet_id')\n", (541576, 541595), False, 'import pulumi\n'), ((541688, 541724), 'pulumi.set', 'pulumi.set', (['self', '"""subnet_id"""', 'value'], {}), "(self, 'subnet_id', value)\n", (541698, 541724), False, 'import pulumi\n'), ((542601, 542630), 'pulumi.get', 'pulumi.get', (['self', '"""data_lens"""'], {}), "(self, 'data_lens')\n", (542611, 542630), False, 'import pulumi\n'), ((542724, 542760), 'pulumi.set', 'pulumi.set', (['self', '"""data_lens"""', 'value'], {}), "(self, 'data_lens', value)\n", (542734, 542760), False, 'import pulumi\n'), ((542968, 542995), 'pulumi.get', 'pulumi.get', (['self', '"""web_sql"""'], {}), "(self, 'web_sql')\n", (542978, 542995), False, 'import pulumi\n'), ((543085, 543119), 'pulumi.set', 'pulumi.set', (['self', '"""web_sql"""', 'value'], {}), "(self, 'web_sql', value)\n", (543095, 543119), False, 'import pulumi\n'), ((543891, 543916), 'pulumi.get', 'pulumi.get', (['self', '"""hours"""'], {}), "(self, 'hours')\n", (543901, 543916), False, 'import pulumi\n'), ((544001, 544033), 'pulumi.set', 'pulumi.set', (['self', '"""hours"""', 'value'], {}), "(self, 'hours', value)\n", (544011, 544033), False, 'import pulumi\n'), ((544219, 544246), 'pulumi.get', 'pulumi.get', (['self', '"""minutes"""'], {}), "(self, 'minutes')\n", (544229, 544246), False, 'import pulumi\n'), ((544335, 544369), 'pulumi.set', 'pulumi.set', (['self', '"""minutes"""', 'value'], {}), "(self, 'minutes', value)\n", (544345, 544369), False, 'import pulumi\n'), ((545423, 545459), 'pulumi.get', 'pulumi.get', (['self', '"""assign_public_ip"""'], {}), "(self, 'assign_public_ip')\n", (545433, 545459), False, 'import pulumi\n'), ((545567, 545610), 'pulumi.set', 'pulumi.set', (['self', '"""assign_public_ip"""', 'value'], {}), "(self, 'assign_public_ip', value)\n", (545577, 545610), False, 'import pulumi\n'), ((545799, 545823), 'pulumi.get', 'pulumi.get', (['self', '"""fqdn"""'], {}), "(self, 'fqdn')\n", (545809, 545823), False, 'import pulumi\n'), ((545906, 545937), 'pulumi.set', 'pulumi.set', (['self', '"""fqdn"""', 'value'], {}), "(self, 'fqdn', value)\n", (545916, 545937), False, 'import pulumi\n'), ((546367, 546411), 'pulumi.set', 'pulumi.set', (['__self__', '"""resources"""', 'resources'], {}), "(__self__, 'resources', resources)\n", (546377, 546411), False, 'import pulumi\n'), ((546699, 546728), 'pulumi.get', 'pulumi.get', (['self', '"""resources"""'], {}), "(self, 'resources')\n", (546709, 546728), False, 'import pulumi\n'), ((546858, 546894), 'pulumi.set', 'pulumi.set', (['self', '"""resources"""', 'value'], {}), "(self, 'resources', value)\n", (546868, 546894), False, 'import pulumi\n'), ((547164, 547208), 'pulumi.set', 'pulumi.set', (['__self__', '"""disk_size"""', 'disk_size'], {}), "(__self__, 'disk_size', disk_size)\n", (547174, 547208), False, 'import pulumi\n'), ((547217, 547267), 'pulumi.set', 'pulumi.set', (['__self__', '"""disk_type_id"""', 'disk_type_id'], {}), "(__self__, 'disk_type_id', disk_type_id)\n", (547227, 547267), False, 'import pulumi\n'), ((547276, 547338), 'pulumi.set', 'pulumi.set', (['__self__', '"""resource_preset_id"""', 'resource_preset_id'], {}), "(__self__, 'resource_preset_id', resource_preset_id)\n", (547286, 547338), False, 'import pulumi\n'), ((547451, 547480), 'pulumi.get', 'pulumi.get', (['self', '"""disk_size"""'], {}), "(self, 'disk_size')\n", (547461, 547480), False, 'import pulumi\n'), ((547563, 547599), 'pulumi.set', 'pulumi.set', (['self', '"""disk_size"""', 'value'], {}), "(self, 'disk_size', value)\n", (547573, 547599), False, 'import pulumi\n'), ((547717, 547749), 'pulumi.get', 'pulumi.get', (['self', '"""disk_type_id"""'], {}), "(self, 'disk_type_id')\n", (547727, 547749), False, 'import pulumi\n'), ((547838, 547877), 'pulumi.set', 'pulumi.set', (['self', '"""disk_type_id"""', 'value'], {}), "(self, 'disk_type_id', value)\n", (547848, 547877), False, 'import pulumi\n'), ((548007, 548045), 'pulumi.get', 'pulumi.get', (['self', '"""resource_preset_id"""'], {}), "(self, 'resource_preset_id')\n", (548017, 548045), False, 'import pulumi\n'), ((548146, 548191), 'pulumi.set', 'pulumi.set', (['self', '"""resource_preset_id"""', 'value'], {}), "(self, 'resource_preset_id', value)\n", (548156, 548191), False, 'import pulumi\n'), ((548728, 548752), 'pulumi.get', 'pulumi.get', (['self', '"""fqdn"""'], {}), "(self, 'fqdn')\n", (548738, 548752), False, 'import pulumi\n'), ((548835, 548866), 'pulumi.set', 'pulumi.set', (['self', '"""fqdn"""', 'value'], {}), "(self, 'fqdn', value)\n", (548845, 548866), False, 'import pulumi\n'), ((549299, 549343), 'pulumi.set', 'pulumi.set', (['__self__', '"""resources"""', 'resources'], {}), "(__self__, 'resources', resources)\n", (549309, 549343), False, 'import pulumi\n'), ((549632, 549661), 'pulumi.get', 'pulumi.get', (['self', '"""resources"""'], {}), "(self, 'resources')\n", (549642, 549661), False, 'import pulumi\n'), ((549792, 549828), 'pulumi.set', 'pulumi.set', (['self', '"""resources"""', 'value'], {}), "(self, 'resources', value)\n", (549802, 549828), False, 'import pulumi\n'), ((550099, 550143), 'pulumi.set', 'pulumi.set', (['__self__', '"""disk_size"""', 'disk_size'], {}), "(__self__, 'disk_size', disk_size)\n", (550109, 550143), False, 'import pulumi\n'), ((550152, 550202), 'pulumi.set', 'pulumi.set', (['__self__', '"""disk_type_id"""', 'disk_type_id'], {}), "(__self__, 'disk_type_id', disk_type_id)\n", (550162, 550202), False, 'import pulumi\n'), ((550211, 550273), 'pulumi.set', 'pulumi.set', (['__self__', '"""resource_preset_id"""', 'resource_preset_id'], {}), "(__self__, 'resource_preset_id', resource_preset_id)\n", (550221, 550273), False, 'import pulumi\n'), ((550386, 550415), 'pulumi.get', 'pulumi.get', (['self', '"""disk_size"""'], {}), "(self, 'disk_size')\n", (550396, 550415), False, 'import pulumi\n'), ((550498, 550534), 'pulumi.set', 'pulumi.set', (['self', '"""disk_size"""', 'value'], {}), "(self, 'disk_size', value)\n", (550508, 550534), False, 'import pulumi\n'), ((550652, 550684), 'pulumi.get', 'pulumi.get', (['self', '"""disk_type_id"""'], {}), "(self, 'disk_type_id')\n", (550662, 550684), False, 'import pulumi\n'), ((550773, 550812), 'pulumi.set', 'pulumi.set', (['self', '"""disk_type_id"""', 'value'], {}), "(self, 'disk_type_id', value)\n", (550783, 550812), False, 'import pulumi\n'), ((550942, 550980), 'pulumi.get', 'pulumi.get', (['self', '"""resource_preset_id"""'], {}), "(self, 'resource_preset_id')\n", (550952, 550980), False, 'import pulumi\n'), ((551081, 551126), 'pulumi.set', 'pulumi.set', (['self', '"""resource_preset_id"""', 'value'], {}), "(self, 'resource_preset_id', value)\n", (551091, 551126), False, 'import pulumi\n'), ((552773, 552809), 'pulumi.set', 'pulumi.set', (['__self__', '"""kafka"""', 'kafka'], {}), "(__self__, 'kafka', kafka)\n", (552783, 552809), False, 'import pulumi\n'), ((552818, 552858), 'pulumi.set', 'pulumi.set', (['__self__', '"""version"""', 'version'], {}), "(__self__, 'version', version)\n", (552828, 552858), False, 'import pulumi\n'), ((552867, 552903), 'pulumi.set', 'pulumi.set', (['__self__', '"""zones"""', 'zones'], {}), "(__self__, 'zones', zones)\n", (552877, 552903), False, 'import pulumi\n'), ((553657, 553682), 'pulumi.get', 'pulumi.get', (['self', '"""kafka"""'], {}), "(self, 'kafka')\n", (553667, 553682), False, 'import pulumi\n'), ((553786, 553818), 'pulumi.set', 'pulumi.set', (['self', '"""kafka"""', 'value'], {}), "(self, 'kafka', value)\n", (553796, 553818), False, 'import pulumi\n'), ((553982, 554009), 'pulumi.get', 'pulumi.get', (['self', '"""version"""'], {}), "(self, 'version')\n", (553992, 554009), False, 'import pulumi\n'), ((554088, 554122), 'pulumi.set', 'pulumi.set', (['self', '"""version"""', 'value'], {}), "(self, 'version', value)\n", (554098, 554122), False, 'import pulumi\n'), ((554298, 554323), 'pulumi.get', 'pulumi.get', (['self', '"""zones"""'], {}), "(self, 'zones')\n", (554308, 554323), False, 'import pulumi\n'), ((554422, 554454), 'pulumi.set', 'pulumi.set', (['self', '"""zones"""', 'value'], {}), "(self, 'zones', value)\n", (554432, 554454), False, 'import pulumi\n'), ((554716, 554752), 'pulumi.get', 'pulumi.get', (['self', '"""assign_public_ip"""'], {}), "(self, 'assign_public_ip')\n", (554726, 554752), False, 'import pulumi\n'), ((554860, 554903), 'pulumi.set', 'pulumi.set', (['self', '"""assign_public_ip"""', 'value'], {}), "(self, 'assign_public_ip', value)\n", (554870, 554903), False, 'import pulumi\n'), ((555126, 555159), 'pulumi.get', 'pulumi.get', (['self', '"""brokers_count"""'], {}), "(self, 'brokers_count')\n", (555136, 555159), False, 'import pulumi\n'), ((555260, 555300), 'pulumi.set', 'pulumi.set', (['self', '"""brokers_count"""', 'value'], {}), "(self, 'brokers_count', value)\n", (555270, 555300), False, 'import pulumi\n'), ((555536, 555571), 'pulumi.get', 'pulumi.get', (['self', '"""schema_registry"""'], {}), "(self, 'schema_registry')\n", (555546, 555571), False, 'import pulumi\n'), ((555677, 555719), 'pulumi.set', 'pulumi.set', (['self', '"""schema_registry"""', 'value'], {}), "(self, 'schema_registry', value)\n", (555687, 555719), False, 'import pulumi\n'), ((555960, 555996), 'pulumi.get', 'pulumi.get', (['self', '"""unmanaged_topics"""'], {}), "(self, 'unmanaged_topics')\n", (555970, 555996), False, 'import pulumi\n'), ((556104, 556147), 'pulumi.set', 'pulumi.set', (['self', '"""unmanaged_topics"""', 'value'], {}), "(self, 'unmanaged_topics', value)\n", (556114, 556147), False, 'import pulumi\n'), ((556396, 556425), 'pulumi.get', 'pulumi.get', (['self', '"""zookeeper"""'], {}), "(self, 'zookeeper')\n", (556406, 556425), False, 'import pulumi\n'), ((556551, 556587), 'pulumi.set', 'pulumi.set', (['self', '"""zookeeper"""', 'value'], {}), "(self, 'zookeeper', value)\n", (556561, 556587), False, 'import pulumi\n'), ((557244, 557288), 'pulumi.set', 'pulumi.set', (['__self__', '"""resources"""', 'resources'], {}), "(__self__, 'resources', resources)\n", (557254, 557288), False, 'import pulumi\n'), ((557647, 557676), 'pulumi.get', 'pulumi.get', (['self', '"""resources"""'], {}), "(self, 'resources')\n", (557657, 557676), False, 'import pulumi\n'), ((557797, 557833), 'pulumi.set', 'pulumi.set', (['self', '"""resources"""', 'value'], {}), "(self, 'resources', value)\n", (557807, 557833), False, 'import pulumi\n'), ((558114, 558146), 'pulumi.get', 'pulumi.get', (['self', '"""kafka_config"""'], {}), "(self, 'kafka_config')\n", (558124, 558146), False, 'import pulumi\n'), ((558285, 558324), 'pulumi.set', 'pulumi.set', (['self', '"""kafka_config"""', 'value'], {}), "(self, 'kafka_config', value)\n", (558295, 558324), False, 'import pulumi\n'), ((561638, 561683), 'pulumi.get', 'pulumi.get', (['self', '"""auto_create_topics_enable"""'], {}), "(self, 'auto_create_topics_enable')\n", (561648, 561683), False, 'import pulumi\n'), ((561809, 561861), 'pulumi.set', 'pulumi.set', (['self', '"""auto_create_topics_enable"""', 'value'], {}), "(self, 'auto_create_topics_enable', value)\n", (561819, 561861), False, 'import pulumi\n'), ((561998, 562034), 'pulumi.get', 'pulumi.get', (['self', '"""compression_type"""'], {}), "(self, 'compression_type')\n", (562008, 562034), False, 'import pulumi\n'), ((562141, 562184), 'pulumi.set', 'pulumi.set', (['self', '"""compression_type"""', 'value'], {}), "(self, 'compression_type', value)\n", (562151, 562184), False, 'import pulumi\n'), ((562340, 562386), 'pulumi.get', 'pulumi.get', (['self', '"""default_replication_factor"""'], {}), "(self, 'default_replication_factor')\n", (562350, 562386), False, 'import pulumi\n'), ((562513, 562566), 'pulumi.set', 'pulumi.set', (['self', '"""default_replication_factor"""', 'value'], {}), "(self, 'default_replication_factor', value)\n", (562523, 562566), False, 'import pulumi\n'), ((562723, 562770), 'pulumi.get', 'pulumi.get', (['self', '"""log_flush_interval_messages"""'], {}), "(self, 'log_flush_interval_messages')\n", (562733, 562770), False, 'import pulumi\n'), ((562899, 562953), 'pulumi.set', 'pulumi.set', (['self', '"""log_flush_interval_messages"""', 'value'], {}), "(self, 'log_flush_interval_messages', value)\n", (562909, 562953), False, 'import pulumi\n'), ((563098, 563139), 'pulumi.get', 'pulumi.get', (['self', '"""log_flush_interval_ms"""'], {}), "(self, 'log_flush_interval_ms')\n", (563108, 563139), False, 'import pulumi\n'), ((563256, 563304), 'pulumi.set', 'pulumi.set', (['self', '"""log_flush_interval_ms"""', 'value'], {}), "(self, 'log_flush_interval_ms', value)\n", (563266, 563304), False, 'import pulumi\n'), ((563468, 563519), 'pulumi.get', 'pulumi.get', (['self', '"""log_flush_scheduler_interval_ms"""'], {}), "(self, 'log_flush_scheduler_interval_ms')\n", (563478, 563519), False, 'import pulumi\n'), ((563656, 563714), 'pulumi.set', 'pulumi.set', (['self', '"""log_flush_scheduler_interval_ms"""', 'value'], {}), "(self, 'log_flush_scheduler_interval_ms', value)\n", (563666, 563714), False, 'import pulumi\n'), ((563850, 563885), 'pulumi.get', 'pulumi.get', (['self', '"""log_preallocate"""'], {}), "(self, 'log_preallocate')\n", (563860, 563885), False, 'import pulumi\n'), ((563991, 564033), 'pulumi.set', 'pulumi.set', (['self', '"""log_preallocate"""', 'value'], {}), "(self, 'log_preallocate', value)\n", (564001, 564033), False, 'import pulumi\n'), ((564175, 564214), 'pulumi.get', 'pulumi.get', (['self', '"""log_retention_bytes"""'], {}), "(self, 'log_retention_bytes')\n", (564185, 564214), False, 'import pulumi\n'), ((564327, 564373), 'pulumi.set', 'pulumi.set', (['self', '"""log_retention_bytes"""', 'value'], {}), "(self, 'log_retention_bytes', value)\n", (564337, 564373), False, 'import pulumi\n'), ((564515, 564554), 'pulumi.get', 'pulumi.get', (['self', '"""log_retention_hours"""'], {}), "(self, 'log_retention_hours')\n", (564525, 564554), False, 'import pulumi\n'), ((564667, 564713), 'pulumi.set', 'pulumi.set', (['self', '"""log_retention_hours"""', 'value'], {}), "(self, 'log_retention_hours', value)\n", (564677, 564713), False, 'import pulumi\n'), ((564859, 564900), 'pulumi.get', 'pulumi.get', (['self', '"""log_retention_minutes"""'], {}), "(self, 'log_retention_minutes')\n", (564869, 564900), False, 'import pulumi\n'), ((565017, 565065), 'pulumi.set', 'pulumi.set', (['self', '"""log_retention_minutes"""', 'value'], {}), "(self, 'log_retention_minutes', value)\n", (565027, 565065), False, 'import pulumi\n'), ((565201, 565237), 'pulumi.get', 'pulumi.get', (['self', '"""log_retention_ms"""'], {}), "(self, 'log_retention_ms')\n", (565211, 565237), False, 'import pulumi\n'), ((565344, 565387), 'pulumi.set', 'pulumi.set', (['self', '"""log_retention_ms"""', 'value'], {}), "(self, 'log_retention_ms', value)\n", (565354, 565387), False, 'import pulumi\n'), ((565525, 565562), 'pulumi.get', 'pulumi.get', (['self', '"""log_segment_bytes"""'], {}), "(self, 'log_segment_bytes')\n", (565535, 565562), False, 'import pulumi\n'), ((565671, 565715), 'pulumi.set', 'pulumi.set', (['self', '"""log_segment_bytes"""', 'value'], {}), "(self, 'log_segment_bytes', value)\n", (565681, 565715), False, 'import pulumi\n'), ((565848, 565882), 'pulumi.get', 'pulumi.get', (['self', '"""num_partitions"""'], {}), "(self, 'num_partitions')\n", (565858, 565882), False, 'import pulumi\n'), ((565985, 566026), 'pulumi.set', 'pulumi.set', (['self', '"""num_partitions"""', 'value'], {}), "(self, 'num_partitions', value)\n", (565995, 566026), False, 'import pulumi\n'), ((566183, 566230), 'pulumi.get', 'pulumi.get', (['self', '"""socket_receive_buffer_bytes"""'], {}), "(self, 'socket_receive_buffer_bytes')\n", (566193, 566230), False, 'import pulumi\n'), ((566359, 566413), 'pulumi.set', 'pulumi.set', (['self', '"""socket_receive_buffer_bytes"""', 'value'], {}), "(self, 'socket_receive_buffer_bytes', value)\n", (566369, 566413), False, 'import pulumi\n'), ((566564, 566608), 'pulumi.get', 'pulumi.get', (['self', '"""socket_send_buffer_bytes"""'], {}), "(self, 'socket_send_buffer_bytes')\n", (566574, 566608), False, 'import pulumi\n'), ((566731, 566782), 'pulumi.set', 'pulumi.set', (['self', '"""socket_send_buffer_bytes"""', 'value'], {}), "(self, 'socket_send_buffer_bytes', value)\n", (566741, 566782), False, 'import pulumi\n'), ((567397, 567441), 'pulumi.set', 'pulumi.set', (['__self__', '"""disk_size"""', 'disk_size'], {}), "(__self__, 'disk_size', disk_size)\n", (567407, 567441), False, 'import pulumi\n'), ((567450, 567500), 'pulumi.set', 'pulumi.set', (['__self__', '"""disk_type_id"""', 'disk_type_id'], {}), "(__self__, 'disk_type_id', disk_type_id)\n", (567460, 567500), False, 'import pulumi\n'), ((567509, 567571), 'pulumi.set', 'pulumi.set', (['__self__', '"""resource_preset_id"""', 'resource_preset_id'], {}), "(__self__, 'resource_preset_id', resource_preset_id)\n", (567519, 567571), False, 'import pulumi\n'), ((567783, 567812), 'pulumi.get', 'pulumi.get', (['self', '"""disk_size"""'], {}), "(self, 'disk_size')\n", (567793, 567812), False, 'import pulumi\n'), ((567895, 567931), 'pulumi.set', 'pulumi.set', (['self', '"""disk_size"""', 'value'], {}), "(self, 'disk_size', value)\n", (567905, 567931), False, 'import pulumi\n'), ((568246, 568278), 'pulumi.get', 'pulumi.get', (['self', '"""disk_type_id"""'], {}), "(self, 'disk_type_id')\n", (568256, 568278), False, 'import pulumi\n'), ((568367, 568406), 'pulumi.set', 'pulumi.set', (['self', '"""disk_type_id"""', 'value'], {}), "(self, 'disk_type_id', value)\n", (568377, 568406), False, 'import pulumi\n'), ((568536, 568574), 'pulumi.get', 'pulumi.get', (['self', '"""resource_preset_id"""'], {}), "(self, 'resource_preset_id')\n", (568546, 568574), False, 'import pulumi\n'), ((568675, 568720), 'pulumi.set', 'pulumi.set', (['self', '"""resource_preset_id"""', 'value'], {}), "(self, 'resource_preset_id', value)\n", (568685, 568720), False, 'import pulumi\n'), ((569487, 569516), 'pulumi.get', 'pulumi.get', (['self', '"""resources"""'], {}), "(self, 'resources')\n", (569497, 569516), False, 'import pulumi\n'), ((569651, 569687), 'pulumi.set', 'pulumi.set', (['self', '"""resources"""', 'value'], {}), "(self, 'resources', value)\n", (569661, 569687), False, 'import pulumi\n'), ((570879, 570908), 'pulumi.get', 'pulumi.get', (['self', '"""disk_size"""'], {}), "(self, 'disk_size')\n", (570889, 570908), False, 'import pulumi\n'), ((571001, 571037), 'pulumi.set', 'pulumi.set', (['self', '"""disk_size"""', 'value'], {}), "(self, 'disk_size', value)\n", (571011, 571037), False, 'import pulumi\n'), ((571362, 571394), 'pulumi.get', 'pulumi.get', (['self', '"""disk_type_id"""'], {}), "(self, 'disk_type_id')\n", (571372, 571394), False, 'import pulumi\n'), ((571493, 571532), 'pulumi.set', 'pulumi.set', (['self', '"""disk_type_id"""', 'value'], {}), "(self, 'disk_type_id', value)\n", (571503, 571532), False, 'import pulumi\n'), ((571672, 571710), 'pulumi.get', 'pulumi.get', (['self', '"""resource_preset_id"""'], {}), "(self, 'resource_preset_id')\n", (571682, 571710), False, 'import pulumi\n'), ((571821, 571866), 'pulumi.set', 'pulumi.set', (['self', '"""resource_preset_id"""', 'value'], {}), "(self, 'resource_preset_id', value)\n", (571831, 571866), False, 'import pulumi\n'), ((573669, 573705), 'pulumi.get', 'pulumi.get', (['self', '"""assign_public_ip"""'], {}), "(self, 'assign_public_ip')\n", (573679, 573705), False, 'import pulumi\n'), ((573813, 573856), 'pulumi.set', 'pulumi.set', (['self', '"""assign_public_ip"""', 'value'], {}), "(self, 'assign_public_ip', value)\n", (573823, 573856), False, 'import pulumi\n'), ((574011, 574037), 'pulumi.get', 'pulumi.get', (['self', '"""health"""'], {}), "(self, 'health')\n", (574021, 574037), False, 'import pulumi\n'), ((574124, 574157), 'pulumi.set', 'pulumi.set', (['self', '"""health"""', 'value'], {}), "(self, 'health', value)\n", (574134, 574157), False, 'import pulumi\n'), ((574313, 574337), 'pulumi.get', 'pulumi.get', (['self', '"""name"""'], {}), "(self, 'name')\n", (574323, 574337), False, 'import pulumi\n'), ((574420, 574451), 'pulumi.set', 'pulumi.set', (['self', '"""name"""', 'value'], {}), "(self, 'name', value)\n", (574430, 574451), False, 'import pulumi\n'), ((574621, 574645), 'pulumi.get', 'pulumi.get', (['self', '"""role"""'], {}), "(self, 'role')\n", (574631, 574645), False, 'import pulumi\n'), ((574728, 574759), 'pulumi.set', 'pulumi.set', (['self', '"""role"""', 'value'], {}), "(self, 'role', value)\n", (574738, 574759), False, 'import pulumi\n'), ((574963, 574992), 'pulumi.get', 'pulumi.get', (['self', '"""subnet_id"""'], {}), "(self, 'subnet_id')\n", (574973, 574992), False, 'import pulumi\n'), ((575085, 575121), 'pulumi.set', 'pulumi.set', (['self', '"""subnet_id"""', 'value'], {}), "(self, 'subnet_id', value)\n", (575095, 575121), False, 'import pulumi\n'), ((575328, 575355), 'pulumi.get', 'pulumi.get', (['self', '"""zone_id"""'], {}), "(self, 'zone_id')\n", (575338, 575355), False, 'import pulumi\n'), ((575444, 575478), 'pulumi.set', 'pulumi.set', (['self', '"""zone_id"""', 'value'], {}), "(self, 'zone_id', value)\n", (575454, 575478), False, 'import pulumi\n'), ((576256, 576290), 'pulumi.set', 'pulumi.set', (['__self__', '"""name"""', 'name'], {}), "(__self__, 'name', name)\n", (576266, 576290), False, 'import pulumi\n'), ((576299, 576345), 'pulumi.set', 'pulumi.set', (['__self__', '"""partitions"""', 'partitions'], {}), "(__self__, 'partitions', partitions)\n", (576309, 576345), False, 'import pulumi\n'), ((576354, 576416), 'pulumi.set', 'pulumi.set', (['__self__', '"""replication_factor"""', 'replication_factor'], {}), "(__self__, 'replication_factor', replication_factor)\n", (576364, 576416), False, 'import pulumi\n'), ((576662, 576686), 'pulumi.get', 'pulumi.get', (['self', '"""name"""'], {}), "(self, 'name')\n", (576672, 576686), False, 'import pulumi\n'), ((576759, 576790), 'pulumi.set', 'pulumi.set', (['self', '"""name"""', 'value'], {}), "(self, 'name', value)\n", (576769, 576790), False, 'import pulumi\n'), ((576957, 576987), 'pulumi.get', 'pulumi.get', (['self', '"""partitions"""'], {}), "(self, 'partitions')\n", (576967, 576987), False, 'import pulumi\n'), ((577072, 577109), 'pulumi.set', 'pulumi.set', (['self', '"""partitions"""', 'value'], {}), "(self, 'partitions', value)\n", (577082, 577109), False, 'import pulumi\n'), ((577335, 577373), 'pulumi.get', 'pulumi.get', (['self', '"""replication_factor"""'], {}), "(self, 'replication_factor')\n", (577345, 577373), False, 'import pulumi\n'), ((577474, 577519), 'pulumi.set', 'pulumi.set', (['self', '"""replication_factor"""', 'value'], {}), "(self, 'replication_factor', value)\n", (577484, 577519), False, 'import pulumi\n'), ((577786, 577818), 'pulumi.get', 'pulumi.get', (['self', '"""topic_config"""'], {}), "(self, 'topic_config')\n", (577796, 577818), False, 'import pulumi\n'), ((577951, 577990), 'pulumi.set', 'pulumi.set', (['self', '"""topic_config"""', 'value'], {}), "(self, 'topic_config', value)\n", (577961, 577990), False, 'import pulumi\n'), ((580562, 580596), 'pulumi.get', 'pulumi.get', (['self', '"""cleanup_policy"""'], {}), "(self, 'cleanup_policy')\n", (580572, 580596), False, 'import pulumi\n'), ((580699, 580740), 'pulumi.set', 'pulumi.set', (['self', '"""cleanup_policy"""', 'value'], {}), "(self, 'cleanup_policy', value)\n", (580709, 580740), False, 'import pulumi\n'), ((580877, 580913), 'pulumi.get', 'pulumi.get', (['self', '"""compression_type"""'], {}), "(self, 'compression_type')\n", (580887, 580913), False, 'import pulumi\n'), ((581020, 581063), 'pulumi.set', 'pulumi.set', (['self', '"""compression_type"""', 'value'], {}), "(self, 'compression_type', value)\n", (581030, 581063), False, 'import pulumi\n'), ((581205, 581244), 'pulumi.get', 'pulumi.get', (['self', '"""delete_retention_ms"""'], {}), "(self, 'delete_retention_ms')\n", (581215, 581244), False, 'import pulumi\n'), ((581357, 581403), 'pulumi.set', 'pulumi.set', (['self', '"""delete_retention_ms"""', 'value'], {}), "(self, 'delete_retention_ms', value)\n", (581367, 581403), False, 'import pulumi\n'), ((581546, 581586), 'pulumi.get', 'pulumi.get', (['self', '"""file_delete_delay_ms"""'], {}), "(self, 'file_delete_delay_ms')\n", (581556, 581586), False, 'import pulumi\n'), ((581701, 581748), 'pulumi.set', 'pulumi.set', (['self', '"""file_delete_delay_ms"""', 'value'], {}), "(self, 'file_delete_delay_ms', value)\n", (581711, 581748), False, 'import pulumi\n'), ((581881, 581915), 'pulumi.get', 'pulumi.get', (['self', '"""flush_messages"""'], {}), "(self, 'flush_messages')\n", (581891, 581915), False, 'import pulumi\n'), ((582018, 582059), 'pulumi.set', 'pulumi.set', (['self', '"""flush_messages"""', 'value'], {}), "(self, 'flush_messages', value)\n", (582028, 582059), False, 'import pulumi\n'), ((582180, 582208), 'pulumi.get', 'pulumi.get', (['self', '"""flush_ms"""'], {}), "(self, 'flush_ms')\n", (582190, 582208), False, 'import pulumi\n'), ((582299, 582334), 'pulumi.set', 'pulumi.set', (['self', '"""flush_ms"""', 'value'], {}), "(self, 'flush_ms', value)\n", (582309, 582334), False, 'import pulumi\n'), ((582472, 582509), 'pulumi.get', 'pulumi.get', (['self', '"""max_message_bytes"""'], {}), "(self, 'max_message_bytes')\n", (582482, 582509), False, 'import pulumi\n'), ((582618, 582662), 'pulumi.set', 'pulumi.set', (['self', '"""max_message_bytes"""', 'value'], {}), "(self, 'max_message_bytes', value)\n", (582628, 582662), False, 'import pulumi\n'), ((582807, 582848), 'pulumi.get', 'pulumi.get', (['self', '"""min_compaction_lag_ms"""'], {}), "(self, 'min_compaction_lag_ms')\n", (582817, 582848), False, 'import pulumi\n'), ((582965, 583013), 'pulumi.set', 'pulumi.set', (['self', '"""min_compaction_lag_ms"""', 'value'], {}), "(self, 'min_compaction_lag_ms', value)\n", (582975, 583013), False, 'import pulumi\n'), ((583155, 583194), 'pulumi.get', 'pulumi.get', (['self', '"""min_insync_replicas"""'], {}), "(self, 'min_insync_replicas')\n", (583165, 583194), False, 'import pulumi\n'), ((583307, 583353), 'pulumi.set', 'pulumi.set', (['self', '"""min_insync_replicas"""', 'value'], {}), "(self, 'min_insync_replicas', value)\n", (583317, 583353), False, 'import pulumi\n'), ((583462, 583493), 'pulumi.get', 'pulumi.get', (['self', '"""preallocate"""'], {}), "(self, 'preallocate')\n", (583472, 583493), False, 'import pulumi\n'), ((583591, 583629), 'pulumi.set', 'pulumi.set', (['self', '"""preallocate"""', 'value'], {}), "(self, 'preallocate', value)\n", (583601, 583629), False, 'import pulumi\n'), ((583764, 583799), 'pulumi.get', 'pulumi.get', (['self', '"""retention_bytes"""'], {}), "(self, 'retention_bytes')\n", (583774, 583799), False, 'import pulumi\n'), ((583904, 583946), 'pulumi.set', 'pulumi.set', (['self', '"""retention_bytes"""', 'value'], {}), "(self, 'retention_bytes', value)\n", (583914, 583946), False, 'import pulumi\n'), ((584075, 584107), 'pulumi.get', 'pulumi.get', (['self', '"""retention_ms"""'], {}), "(self, 'retention_ms')\n", (584085, 584107), False, 'import pulumi\n'), ((584206, 584245), 'pulumi.set', 'pulumi.set', (['self', '"""retention_ms"""', 'value'], {}), "(self, 'retention_ms', value)\n", (584216, 584245), False, 'import pulumi\n'), ((584376, 584409), 'pulumi.get', 'pulumi.get', (['self', '"""segment_bytes"""'], {}), "(self, 'segment_bytes')\n", (584386, 584409), False, 'import pulumi\n'), ((584510, 584550), 'pulumi.set', 'pulumi.set', (['self', '"""segment_bytes"""', 'value'], {}), "(self, 'segment_bytes', value)\n", (584520, 584550), False, 'import pulumi\n'), ((585189, 585223), 'pulumi.set', 'pulumi.set', (['__self__', '"""name"""', 'name'], {}), "(__self__, 'name', name)\n", (585199, 585223), False, 'import pulumi\n'), ((585232, 585274), 'pulumi.set', 'pulumi.set', (['__self__', '"""password"""', 'password'], {}), "(__self__, 'password', password)\n", (585242, 585274), False, 'import pulumi\n'), ((585517, 585541), 'pulumi.get', 'pulumi.get', (['self', '"""name"""'], {}), "(self, 'name')\n", (585527, 585541), False, 'import pulumi\n'), ((585614, 585645), 'pulumi.set', 'pulumi.set', (['self', '"""name"""', 'value'], {}), "(self, 'name', value)\n", (585624, 585645), False, 'import pulumi\n'), ((585798, 585826), 'pulumi.get', 'pulumi.get', (['self', '"""password"""'], {}), "(self, 'password')\n", (585808, 585826), False, 'import pulumi\n'), ((585907, 585942), 'pulumi.set', 'pulumi.set', (['self', '"""password"""', 'value'], {}), "(self, 'password', value)\n", (585917, 585942), False, 'import pulumi\n'), ((586213, 586244), 'pulumi.get', 'pulumi.get', (['self', '"""permissions"""'], {}), "(self, 'permissions')\n", (586223, 586244), False, 'import pulumi\n'), ((586397, 586435), 'pulumi.set', 'pulumi.set', (['self', '"""permissions"""', 'value'], {}), "(self, 'permissions', value)\n", (586407, 586435), False, 'import pulumi\n'), ((586832, 586866), 'pulumi.set', 'pulumi.set', (['__self__', '"""role"""', 'role'], {}), "(__self__, 'role', role)\n", (586842, 586866), False, 'import pulumi\n'), ((586875, 586921), 'pulumi.set', 'pulumi.set', (['__self__', '"""topic_name"""', 'topic_name'], {}), "(__self__, 'topic_name', topic_name)\n", (586885, 586921), False, 'import pulumi\n'), ((587081, 587105), 'pulumi.get', 'pulumi.get', (['self', '"""role"""'], {}), "(self, 'role')\n", (587091, 587105), False, 'import pulumi\n'), ((587178, 587209), 'pulumi.set', 'pulumi.set', (['self', '"""role"""', 'value'], {}), "(self, 'role', value)\n", (587188, 587209), False, 'import pulumi\n'), ((587416, 587446), 'pulumi.get', 'pulumi.get', (['self', '"""topic_name"""'], {}), "(self, 'topic_name')\n", (587426, 587446), False, 'import pulumi\n'), ((587531, 587568), 'pulumi.set', 'pulumi.set', (['self', '"""topic_name"""', 'value'], {}), "(self, 'topic_name', value)\n", (587541, 587568), False, 'import pulumi\n'), ((590133, 590167), 'pulumi.get', 'pulumi.get', (['self', '"""cleanup_policy"""'], {}), "(self, 'cleanup_policy')\n", (590143, 590167), False, 'import pulumi\n'), ((590270, 590311), 'pulumi.set', 'pulumi.set', (['self', '"""cleanup_policy"""', 'value'], {}), "(self, 'cleanup_policy', value)\n", (590280, 590311), False, 'import pulumi\n'), ((590448, 590484), 'pulumi.get', 'pulumi.get', (['self', '"""compression_type"""'], {}), "(self, 'compression_type')\n", (590458, 590484), False, 'import pulumi\n'), ((590591, 590634), 'pulumi.set', 'pulumi.set', (['self', '"""compression_type"""', 'value'], {}), "(self, 'compression_type', value)\n", (590601, 590634), False, 'import pulumi\n'), ((590776, 590815), 'pulumi.get', 'pulumi.get', (['self', '"""delete_retention_ms"""'], {}), "(self, 'delete_retention_ms')\n", (590786, 590815), False, 'import pulumi\n'), ((590928, 590974), 'pulumi.set', 'pulumi.set', (['self', '"""delete_retention_ms"""', 'value'], {}), "(self, 'delete_retention_ms', value)\n", (590938, 590974), False, 'import pulumi\n'), ((591117, 591157), 'pulumi.get', 'pulumi.get', (['self', '"""file_delete_delay_ms"""'], {}), "(self, 'file_delete_delay_ms')\n", (591127, 591157), False, 'import pulumi\n'), ((591272, 591319), 'pulumi.set', 'pulumi.set', (['self', '"""file_delete_delay_ms"""', 'value'], {}), "(self, 'file_delete_delay_ms', value)\n", (591282, 591319), False, 'import pulumi\n'), ((591452, 591486), 'pulumi.get', 'pulumi.get', (['self', '"""flush_messages"""'], {}), "(self, 'flush_messages')\n", (591462, 591486), False, 'import pulumi\n'), ((591589, 591630), 'pulumi.set', 'pulumi.set', (['self', '"""flush_messages"""', 'value'], {}), "(self, 'flush_messages', value)\n", (591599, 591630), False, 'import pulumi\n'), ((591751, 591779), 'pulumi.get', 'pulumi.get', (['self', '"""flush_ms"""'], {}), "(self, 'flush_ms')\n", (591761, 591779), False, 'import pulumi\n'), ((591870, 591905), 'pulumi.set', 'pulumi.set', (['self', '"""flush_ms"""', 'value'], {}), "(self, 'flush_ms', value)\n", (591880, 591905), False, 'import pulumi\n'), ((592043, 592080), 'pulumi.get', 'pulumi.get', (['self', '"""max_message_bytes"""'], {}), "(self, 'max_message_bytes')\n", (592053, 592080), False, 'import pulumi\n'), ((592189, 592233), 'pulumi.set', 'pulumi.set', (['self', '"""max_message_bytes"""', 'value'], {}), "(self, 'max_message_bytes', value)\n", (592199, 592233), False, 'import pulumi\n'), ((592378, 592419), 'pulumi.get', 'pulumi.get', (['self', '"""min_compaction_lag_ms"""'], {}), "(self, 'min_compaction_lag_ms')\n", (592388, 592419), False, 'import pulumi\n'), ((592536, 592584), 'pulumi.set', 'pulumi.set', (['self', '"""min_compaction_lag_ms"""', 'value'], {}), "(self, 'min_compaction_lag_ms', value)\n", (592546, 592584), False, 'import pulumi\n'), ((592726, 592765), 'pulumi.get', 'pulumi.get', (['self', '"""min_insync_replicas"""'], {}), "(self, 'min_insync_replicas')\n", (592736, 592765), False, 'import pulumi\n'), ((592878, 592924), 'pulumi.set', 'pulumi.set', (['self', '"""min_insync_replicas"""', 'value'], {}), "(self, 'min_insync_replicas', value)\n", (592888, 592924), False, 'import pulumi\n'), ((593033, 593064), 'pulumi.get', 'pulumi.get', (['self', '"""preallocate"""'], {}), "(self, 'preallocate')\n", (593043, 593064), False, 'import pulumi\n'), ((593162, 593200), 'pulumi.set', 'pulumi.set', (['self', '"""preallocate"""', 'value'], {}), "(self, 'preallocate', value)\n", (593172, 593200), False, 'import pulumi\n'), ((593335, 593370), 'pulumi.get', 'pulumi.get', (['self', '"""retention_bytes"""'], {}), "(self, 'retention_bytes')\n", (593345, 593370), False, 'import pulumi\n'), ((593475, 593517), 'pulumi.set', 'pulumi.set', (['self', '"""retention_bytes"""', 'value'], {}), "(self, 'retention_bytes', value)\n", (593485, 593517), False, 'import pulumi\n'), ((593646, 593678), 'pulumi.get', 'pulumi.get', (['self', '"""retention_ms"""'], {}), "(self, 'retention_ms')\n", (593656, 593678), False, 'import pulumi\n'), ((593777, 593816), 'pulumi.set', 'pulumi.set', (['self', '"""retention_ms"""', 'value'], {}), "(self, 'retention_ms', value)\n", (593787, 593816), False, 'import pulumi\n'), ((593947, 593980), 'pulumi.get', 'pulumi.get', (['self', '"""segment_bytes"""'], {}), "(self, 'segment_bytes')\n", (593957, 593980), False, 'import pulumi\n'), ((594081, 594121), 'pulumi.set', 'pulumi.set', (['self', '"""segment_bytes"""', 'value'], {}), "(self, 'segment_bytes', value)\n", (594091, 594121), False, 'import pulumi\n'), ((595232, 595272), 'pulumi.set', 'pulumi.set', (['__self__', '"""version"""', 'version'], {}), "(__self__, 'version', version)\n", (595242, 595272), False, 'import pulumi\n'), ((595802, 595829), 'pulumi.get', 'pulumi.get', (['self', '"""version"""'], {}), "(self, 'version')\n", (595812, 595829), False, 'import pulumi\n'), ((595908, 595942), 'pulumi.set', 'pulumi.set', (['self', '"""version"""', 'value'], {}), "(self, 'version', value)\n", (595918, 595942), False, 'import pulumi\n'), ((596198, 596224), 'pulumi.get', 'pulumi.get', (['self', '"""access"""'], {}), "(self, 'access')\n", (596208, 596224), False, 'import pulumi\n'), ((596350, 596383), 'pulumi.set', 'pulumi.set', (['self', '"""access"""', 'value'], {}), "(self, 'access', value)\n", (596360, 596383), False, 'import pulumi\n'), ((596695, 596734), 'pulumi.get', 'pulumi.get', (['self', '"""backup_window_start"""'], {}), "(self, 'backup_window_start')\n", (596705, 596734), False, 'import pulumi\n'), ((596897, 596943), 'pulumi.set', 'pulumi.set', (['self', '"""backup_window_start"""', 'value'], {}), "(self, 'backup_window_start', value)\n", (596907, 596943), False, 'import pulumi\n'), ((597258, 597307), 'pulumi.get', 'pulumi.get', (['self', '"""feature_compatibility_version"""'], {}), "(self, 'feature_compatibility_version')\n", (597268, 597307), False, 'import pulumi\n'), ((597440, 597496), 'pulumi.set', 'pulumi.set', (['self', '"""feature_compatibility_version"""', 'value'], {}), "(self, 'feature_compatibility_version', value)\n", (597450, 597496), False, 'import pulumi\n'), ((598031, 598060), 'pulumi.get', 'pulumi.get', (['self', '"""data_lens"""'], {}), "(self, 'data_lens')\n", (598041, 598060), False, 'import pulumi\n'), ((598154, 598190), 'pulumi.set', 'pulumi.set', (['self', '"""data_lens"""', 'value'], {}), "(self, 'data_lens', value)\n", (598164, 598190), False, 'import pulumi\n'), ((598955, 598980), 'pulumi.get', 'pulumi.get', (['self', '"""hours"""'], {}), "(self, 'hours')\n", (598965, 598980), False, 'import pulumi\n'), ((599065, 599097), 'pulumi.set', 'pulumi.set', (['self', '"""hours"""', 'value'], {}), "(self, 'hours', value)\n", (599075, 599097), False, 'import pulumi\n'), ((599277, 599304), 'pulumi.get', 'pulumi.get', (['self', '"""minutes"""'], {}), "(self, 'minutes')\n", (599287, 599304), False, 'import pulumi\n'), ((599393, 599427), 'pulumi.set', 'pulumi.set', (['self', '"""minutes"""', 'value'], {}), "(self, 'minutes', value)\n", (599403, 599427), False, 'import pulumi\n'), ((599700, 599734), 'pulumi.set', 'pulumi.set', (['__self__', '"""name"""', 'name'], {}), "(__self__, 'name', name)\n", (599710, 599734), False, 'import pulumi\n'), ((599927, 599951), 'pulumi.get', 'pulumi.get', (['self', '"""name"""'], {}), "(self, 'name')\n", (599937, 599951), False, 'import pulumi\n'), ((600024, 600055), 'pulumi.set', 'pulumi.set', (['self', '"""name"""', 'value'], {}), "(self, 'name', value)\n", (600034, 600055), False, 'import pulumi\n'), ((601733, 601777), 'pulumi.set', 'pulumi.set', (['__self__', '"""subnet_id"""', 'subnet_id'], {}), "(__self__, 'subnet_id', subnet_id)\n", (601743, 601777), False, 'import pulumi\n'), ((601786, 601826), 'pulumi.set', 'pulumi.set', (['__self__', '"""zone_id"""', 'zone_id'], {}), "(__self__, 'zone_id', zone_id)\n", (601796, 601826), False, 'import pulumi\n'), ((602615, 602644), 'pulumi.get', 'pulumi.get', (['self', '"""subnet_id"""'], {}), "(self, 'subnet_id')\n", (602625, 602644), False, 'import pulumi\n'), ((602727, 602763), 'pulumi.set', 'pulumi.set', (['self', '"""subnet_id"""', 'value'], {}), "(self, 'subnet_id', value)\n", (602737, 602763), False, 'import pulumi\n'), ((603088, 603115), 'pulumi.get', 'pulumi.get', (['self', '"""zone_id"""'], {}), "(self, 'zone_id')\n", (603098, 603115), False, 'import pulumi\n'), ((603194, 603228), 'pulumi.set', 'pulumi.set', (['self', '"""zone_id"""', 'value'], {}), "(self, 'zone_id', value)\n", (603204, 603228), False, 'import pulumi\n'), ((603494, 603530), 'pulumi.get', 'pulumi.get', (['self', '"""assign_public_ip"""'], {}), "(self, 'assign_public_ip')\n", (603504, 603530), False, 'import pulumi\n'), ((603638, 603681), 'pulumi.set', 'pulumi.set', (['self', '"""assign_public_ip"""', 'value'], {}), "(self, 'assign_public_ip', value)\n", (603648, 603681), False, 'import pulumi\n'), ((603840, 603866), 'pulumi.get', 'pulumi.get', (['self', '"""health"""'], {}), "(self, 'health')\n", (603850, 603866), False, 'import pulumi\n'), ((603953, 603986), 'pulumi.set', 'pulumi.set', (['self', '"""health"""', 'value'], {}), "(self, 'health', value)\n", (603963, 603986), False, 'import pulumi\n'), ((604189, 604213), 'pulumi.get', 'pulumi.get', (['self', '"""name"""'], {}), "(self, 'name')\n", (604199, 604213), False, 'import pulumi\n'), ((604296, 604327), 'pulumi.set', 'pulumi.set', (['self', '"""name"""', 'value'], {}), "(self, 'name', value)\n", (604306, 604327), False, 'import pulumi\n'), ((604515, 604539), 'pulumi.get', 'pulumi.get', (['self', '"""role"""'], {}), "(self, 'role')\n", (604525, 604539), False, 'import pulumi\n'), ((604622, 604653), 'pulumi.set', 'pulumi.set', (['self', '"""role"""', 'value'], {}), "(self, 'role', value)\n", (604632, 604653), False, 'import pulumi\n'), ((604859, 604889), 'pulumi.get', 'pulumi.get', (['self', '"""shard_name"""'], {}), "(self, 'shard_name')\n", (604869, 604889), False, 'import pulumi\n'), ((604984, 605021), 'pulumi.set', 'pulumi.set', (['self', '"""shard_name"""', 'value'], {}), "(self, 'shard_name', value)\n", (604994, 605021), False, 'import pulumi\n'), ((605285, 605309), 'pulumi.get', 'pulumi.get', (['self', '"""type"""'], {}), "(self, 'type')\n", (605295, 605309), False, 'import pulumi\n'), ((605392, 605423), 'pulumi.set', 'pulumi.set', (['self', '"""type"""', 'value'], {}), "(self, 'type', value)\n", (605402, 605423), False, 'import pulumi\n'), ((606172, 606206), 'pulumi.set', 'pulumi.set', (['__self__', '"""type"""', 'type'], {}), "(__self__, 'type', type)\n", (606182, 606206), False, 'import pulumi\n'), ((606609, 606633), 'pulumi.get', 'pulumi.get', (['self', '"""type"""'], {}), "(self, 'type')\n", (606619, 606633), False, 'import pulumi\n'), ((606706, 606737), 'pulumi.set', 'pulumi.set', (['self', '"""type"""', 'value'], {}), "(self, 'type', value)\n", (606716, 606737), False, 'import pulumi\n'), ((606996, 607019), 'pulumi.get', 'pulumi.get', (['self', '"""day"""'], {}), "(self, 'day')\n", (607006, 607019), False, 'import pulumi\n'), ((607100, 607130), 'pulumi.set', 'pulumi.set', (['self', '"""day"""', 'value'], {}), "(self, 'day', value)\n", (607110, 607130), False, 'import pulumi\n'), ((607348, 607372), 'pulumi.get', 'pulumi.get', (['self', '"""hour"""'], {}), "(self, 'hour')\n", (607358, 607372), False, 'import pulumi\n'), ((607455, 607486), 'pulumi.set', 'pulumi.set', (['self', '"""hour"""', 'value'], {}), "(self, 'hour', value)\n", (607465, 607486), False, 'import pulumi\n'), ((608093, 608137), 'pulumi.set', 'pulumi.set', (['__self__', '"""disk_size"""', 'disk_size'], {}), "(__self__, 'disk_size', disk_size)\n", (608103, 608137), False, 'import pulumi\n'), ((608146, 608196), 'pulumi.set', 'pulumi.set', (['__self__', '"""disk_type_id"""', 'disk_type_id'], {}), "(__self__, 'disk_type_id', disk_type_id)\n", (608156, 608196), False, 'import pulumi\n'), ((608205, 608267), 'pulumi.set', 'pulumi.set', (['__self__', '"""resource_preset_id"""', 'resource_preset_id'], {}), "(__self__, 'resource_preset_id', resource_preset_id)\n", (608215, 608267), False, 'import pulumi\n'), ((608477, 608506), 'pulumi.get', 'pulumi.get', (['self', '"""disk_size"""'], {}), "(self, 'disk_size')\n", (608487, 608506), False, 'import pulumi\n'), ((608589, 608625), 'pulumi.set', 'pulumi.set', (['self', '"""disk_size"""', 'value'], {}), "(self, 'disk_size', value)\n", (608599, 608625), False, 'import pulumi\n'), ((608943, 608975), 'pulumi.get', 'pulumi.get', (['self', '"""disk_type_id"""'], {}), "(self, 'disk_type_id')\n", (608953, 608975), False, 'import pulumi\n'), ((609064, 609103), 'pulumi.set', 'pulumi.set', (['self', '"""disk_type_id"""', 'value'], {}), "(self, 'disk_type_id', value)\n", (609074, 609103), False, 'import pulumi\n'), ((609233, 609271), 'pulumi.get', 'pulumi.get', (['self', '"""resource_preset_id"""'], {}), "(self, 'resource_preset_id')\n", (609243, 609271), False, 'import pulumi\n'), ((609372, 609417), 'pulumi.set', 'pulumi.set', (['self', '"""resource_preset_id"""', 'value'], {}), "(self, 'resource_preset_id', value)\n", (609382, 609417), False, 'import pulumi\n'), ((610090, 610124), 'pulumi.set', 'pulumi.set', (['__self__', '"""name"""', 'name'], {}), "(__self__, 'name', name)\n", (610100, 610124), False, 'import pulumi\n'), ((610133, 610175), 'pulumi.set', 'pulumi.set', (['__self__', '"""password"""', 'password'], {}), "(__self__, 'password', password)\n", (610143, 610175), False, 'import pulumi\n'), ((610465, 610489), 'pulumi.get', 'pulumi.get', (['self', '"""name"""'], {}), "(self, 'name')\n", (610475, 610489), False, 'import pulumi\n'), ((610562, 610593), 'pulumi.set', 'pulumi.set', (['self', '"""name"""', 'value'], {}), "(self, 'name', value)\n", (610572, 610593), False, 'import pulumi\n'), ((610746, 610774), 'pulumi.get', 'pulumi.get', (['self', '"""password"""'], {}), "(self, 'password')\n", (610756, 610774), False, 'import pulumi\n'), ((610855, 610890), 'pulumi.set', 'pulumi.set', (['self', '"""password"""', 'value'], {}), "(self, 'password', value)\n", (610865, 610890), False, 'import pulumi\n'), ((611163, 611194), 'pulumi.get', 'pulumi.get', (['self', '"""permissions"""'], {}), "(self, 'permissions')\n", (611173, 611194), False, 'import pulumi\n'), ((611349, 611387), 'pulumi.set', 'pulumi.set', (['self', '"""permissions"""', 'value'], {}), "(self, 'permissions', value)\n", (611359, 611387), False, 'import pulumi\n'), ((611761, 611813), 'pulumi.set', 'pulumi.set', (['__self__', '"""database_name"""', 'database_name'], {}), "(__self__, 'database_name', database_name)\n", (611771, 611813), False, 'import pulumi\n'), ((612108, 612141), 'pulumi.get', 'pulumi.get', (['self', '"""database_name"""'], {}), "(self, 'database_name')\n", (612118, 612141), False, 'import pulumi\n'), ((612232, 612272), 'pulumi.set', 'pulumi.set', (['self', '"""database_name"""', 'value'], {}), "(self, 'database_name', value)\n", (612242, 612272), False, 'import pulumi\n'), ((612398, 612423), 'pulumi.get', 'pulumi.get', (['self', '"""roles"""'], {}), "(self, 'roles')\n", (612408, 612423), False, 'import pulumi\n'), ((612532, 612564), 'pulumi.set', 'pulumi.set', (['self', '"""roles"""', 'value'], {}), "(self, 'roles', value)\n", (612542, 612564), False, 'import pulumi\n'), ((613510, 613539), 'pulumi.get', 'pulumi.get', (['self', '"""data_lens"""'], {}), "(self, 'data_lens')\n", (613520, 613539), False, 'import pulumi\n'), ((613633, 613669), 'pulumi.set', 'pulumi.set', (['self', '"""data_lens"""', 'value'], {}), "(self, 'data_lens', value)\n", (613643, 613669), False, 'import pulumi\n'), ((613950, 613977), 'pulumi.get', 'pulumi.get', (['self', '"""web_sql"""'], {}), "(self, 'web_sql')\n", (613960, 613977), False, 'import pulumi\n'), ((614067, 614101), 'pulumi.set', 'pulumi.set', (['self', '"""web_sql"""', 'value'], {}), "(self, 'web_sql', value)\n", (614077, 614101), False, 'import pulumi\n'), ((614851, 614876), 'pulumi.get', 'pulumi.get', (['self', '"""hours"""'], {}), "(self, 'hours')\n", (614861, 614876), False, 'import pulumi\n'), ((614961, 614993), 'pulumi.set', 'pulumi.set', (['self', '"""hours"""', 'value'], {}), "(self, 'hours', value)\n", (614971, 614993), False, 'import pulumi\n'), ((615173, 615200), 'pulumi.get', 'pulumi.get', (['self', '"""minutes"""'], {}), "(self, 'minutes')\n", (615183, 615200), False, 'import pulumi\n'), ((615289, 615323), 'pulumi.set', 'pulumi.set', (['self', '"""minutes"""', 'value'], {}), "(self, 'minutes', value)\n", (615299, 615323), False, 'import pulumi\n'), ((615742, 615776), 'pulumi.set', 'pulumi.set', (['__self__', '"""name"""', 'name'], {}), "(__self__, 'name', name)\n", (615752, 615776), False, 'import pulumi\n'), ((616117, 616141), 'pulumi.get', 'pulumi.get', (['self', '"""name"""'], {}), "(self, 'name')\n", (616127, 616141), False, 'import pulumi\n'), ((616214, 616245), 'pulumi.set', 'pulumi.set', (['self', '"""name"""', 'value'], {}), "(self, 'name', value)\n", (616224, 616245), False, 'import pulumi\n'), ((617936, 617970), 'pulumi.set', 'pulumi.set', (['__self__', '"""zone"""', 'zone'], {}), "(__self__, 'zone', zone)\n", (617946, 617970), False, 'import pulumi\n'), ((618759, 618783), 'pulumi.get', 'pulumi.get', (['self', '"""zone"""'], {}), "(self, 'zone')\n", (618769, 618783), False, 'import pulumi\n'), ((618856, 618887), 'pulumi.set', 'pulumi.set', (['self', '"""zone"""', 'value'], {}), "(self, 'zone', value)\n", (618866, 618887), False, 'import pulumi\n'), ((619164, 619200), 'pulumi.get', 'pulumi.get', (['self', '"""assign_public_ip"""'], {}), "(self, 'assign_public_ip')\n", (619174, 619200), False, 'import pulumi\n'), ((619308, 619351), 'pulumi.set', 'pulumi.set', (['self', '"""assign_public_ip"""', 'value'], {}), "(self, 'assign_public_ip', value)\n", (619318, 619351), False, 'import pulumi\n'), ((619529, 619553), 'pulumi.get', 'pulumi.get', (['self', '"""fqdn"""'], {}), "(self, 'fqdn')\n", (619539, 619553), False, 'import pulumi\n'), ((619636, 619667), 'pulumi.set', 'pulumi.set', (['self', '"""fqdn"""', 'value'], {}), "(self, 'fqdn', value)\n", (619646, 619667), False, 'import pulumi\n'), ((620018, 620042), 'pulumi.get', 'pulumi.get', (['self', '"""name"""'], {}), "(self, 'name')\n", (620028, 620042), False, 'import pulumi\n'), ((620125, 620156), 'pulumi.set', 'pulumi.set', (['self', '"""name"""', 'value'], {}), "(self, 'name', value)\n", (620135, 620156), False, 'import pulumi\n'), ((620420, 620458), 'pulumi.get', 'pulumi.get', (['self', '"""replication_source"""'], {}), "(self, 'replication_source')\n", (620430, 620458), False, 'import pulumi\n'), ((620569, 620614), 'pulumi.set', 'pulumi.set', (['self', '"""replication_source"""', 'value'], {}), "(self, 'replication_source', value)\n", (620579, 620614), False, 'import pulumi\n'), ((620958, 621001), 'pulumi.get', 'pulumi.get', (['self', '"""replication_source_name"""'], {}), "(self, 'replication_source_name')\n", (620968, 621001), False, 'import pulumi\n'), ((621122, 621172), 'pulumi.set', 'pulumi.set', (['self', '"""replication_source_name"""', 'value'], {}), "(self, 'replication_source_name', value)\n", (621132, 621172), False, 'import pulumi\n'), ((621447, 621476), 'pulumi.get', 'pulumi.get', (['self', '"""subnet_id"""'], {}), "(self, 'subnet_id')\n", (621457, 621476), False, 'import pulumi\n'), ((621569, 621605), 'pulumi.set', 'pulumi.set', (['self', '"""subnet_id"""', 'value'], {}), "(self, 'subnet_id', value)\n", (621579, 621605), False, 'import pulumi\n'), ((622315, 622349), 'pulumi.set', 'pulumi.set', (['__self__', '"""type"""', 'type'], {}), "(__self__, 'type', type)\n", (622325, 622349), False, 'import pulumi\n'), ((622752, 622776), 'pulumi.get', 'pulumi.get', (['self', '"""type"""'], {}), "(self, 'type')\n", (622762, 622776), False, 'import pulumi\n'), ((622849, 622880), 'pulumi.set', 'pulumi.set', (['self', '"""type"""', 'value'], {}), "(self, 'type', value)\n", (622859, 622880), False, 'import pulumi\n'), ((623111, 623134), 'pulumi.get', 'pulumi.get', (['self', '"""day"""'], {}), "(self, 'day')\n", (623121, 623134), False, 'import pulumi\n'), ((623215, 623245), 'pulumi.set', 'pulumi.set', (['self', '"""day"""', 'value'], {}), "(self, 'day', value)\n", (623225, 623245), False, 'import pulumi\n'), ((623454, 623478), 'pulumi.get', 'pulumi.get', (['self', '"""hour"""'], {}), "(self, 'hour')\n", (623464, 623478), False, 'import pulumi\n'), ((623561, 623592), 'pulumi.set', 'pulumi.set', (['self', '"""hour"""', 'value'], {}), "(self, 'hour', value)\n", (623571, 623592), False, 'import pulumi\n'), ((624056, 624100), 'pulumi.set', 'pulumi.set', (['__self__', '"""disk_size"""', 'disk_size'], {}), "(__self__, 'disk_size', disk_size)\n", (624066, 624100), False, 'import pulumi\n'), ((624109, 624159), 'pulumi.set', 'pulumi.set', (['__self__', '"""disk_type_id"""', 'disk_type_id'], {}), "(__self__, 'disk_type_id', disk_type_id)\n", (624119, 624159), False, 'import pulumi\n'), ((624168, 624230), 'pulumi.set', 'pulumi.set', (['__self__', '"""resource_preset_id"""', 'resource_preset_id'], {}), "(__self__, 'resource_preset_id', resource_preset_id)\n", (624178, 624230), False, 'import pulumi\n'), ((624438, 624467), 'pulumi.get', 'pulumi.get', (['self', '"""disk_size"""'], {}), "(self, 'disk_size')\n", (624448, 624467), False, 'import pulumi\n'), ((624550, 624586), 'pulumi.set', 'pulumi.set', (['self', '"""disk_size"""', 'value'], {}), "(self, 'disk_size', value)\n", (624560, 624586), False, 'import pulumi\n'), ((624772, 624804), 'pulumi.get', 'pulumi.get', (['self', '"""disk_type_id"""'], {}), "(self, 'disk_type_id')\n", (624782, 624804), False, 'import pulumi\n'), ((624893, 624932), 'pulumi.set', 'pulumi.set', (['self', '"""disk_type_id"""', 'value'], {}), "(self, 'disk_type_id', value)\n", (624903, 624932), False, 'import pulumi\n'), ((625062, 625100), 'pulumi.get', 'pulumi.get', (['self', '"""resource_preset_id"""'], {}), "(self, 'resource_preset_id')\n", (625072, 625100), False, 'import pulumi\n'), ((625201, 625246), 'pulumi.set', 'pulumi.set', (['self', '"""resource_preset_id"""', 'value'], {}), "(self, 'resource_preset_id', value)\n", (625211, 625246), False, 'import pulumi\n'), ((625876, 625920), 'pulumi.set', 'pulumi.set', (['__self__', '"""backup_id"""', 'backup_id'], {}), "(__self__, 'backup_id', backup_id)\n", (625886, 625920), False, 'import pulumi\n'), ((626317, 626346), 'pulumi.get', 'pulumi.get', (['self', '"""backup_id"""'], {}), "(self, 'backup_id')\n", (626327, 626346), False, 'import pulumi\n'), ((626429, 626465), 'pulumi.set', 'pulumi.set', (['self', '"""backup_id"""', 'value'], {}), "(self, 'backup_id', value)\n", (626439, 626465), False, 'import pulumi\n'), ((626744, 626768), 'pulumi.get', 'pulumi.get', (['self', '"""time"""'], {}), "(self, 'time')\n", (626754, 626768), False, 'import pulumi\n'), ((626851, 626882), 'pulumi.set', 'pulumi.set', (['self', '"""time"""', 'value'], {}), "(self, 'time', value)\n", (626861, 626882), False, 'import pulumi\n'), ((628729, 628763), 'pulumi.set', 'pulumi.set', (['__self__', '"""name"""', 'name'], {}), "(__self__, 'name', name)\n", (628739, 628763), False, 'import pulumi\n'), ((628772, 628814), 'pulumi.set', 'pulumi.set', (['__self__', '"""password"""', 'password'], {}), "(__self__, 'password', password)\n", (628782, 628814), False, 'import pulumi\n'), ((629612, 629636), 'pulumi.get', 'pulumi.get', (['self', '"""name"""'], {}), "(self, 'name')\n", (629622, 629636), False, 'import pulumi\n'), ((629709, 629740), 'pulumi.set', 'pulumi.set', (['self', '"""name"""', 'value'], {}), "(self, 'name', value)\n", (629719, 629740), False, 'import pulumi\n'), ((629893, 629921), 'pulumi.get', 'pulumi.get', (['self', '"""password"""'], {}), "(self, 'password')\n", (629903, 629921), False, 'import pulumi\n'), ((630002, 630037), 'pulumi.set', 'pulumi.set', (['self', '"""password"""', 'value'], {}), "(self, 'password', value)\n", (630012, 630037), False, 'import pulumi\n'), ((630384, 630425), 'pulumi.get', 'pulumi.get', (['self', '"""authentication_plugin"""'], {}), "(self, 'authentication_plugin')\n", (630394, 630425), False, 'import pulumi\n'), ((630542, 630590), 'pulumi.set', 'pulumi.set', (['self', '"""authentication_plugin"""', 'value'], {}), "(self, 'authentication_plugin', value)\n", (630552, 630590), False, 'import pulumi\n'), ((630928, 630965), 'pulumi.get', 'pulumi.get', (['self', '"""connection_limits"""'], {}), "(self, 'connection_limits')\n", (630938, 630965), False, 'import pulumi\n'), ((631112, 631156), 'pulumi.set', 'pulumi.set', (['self', '"""connection_limits"""', 'value'], {}), "(self, 'connection_limits', value)\n", (631122, 631156), False, 'import pulumi\n'), ((631571, 631609), 'pulumi.get', 'pulumi.get', (['self', '"""global_permissions"""'], {}), "(self, 'global_permissions')\n", (631581, 631609), False, 'import pulumi\n'), ((631744, 631789), 'pulumi.set', 'pulumi.set', (['self', '"""global_permissions"""', 'value'], {}), "(self, 'global_permissions', value)\n", (631754, 631789), False, 'import pulumi\n'), ((632060, 632091), 'pulumi.get', 'pulumi.get', (['self', '"""permissions"""'], {}), "(self, 'permissions')\n", (632070, 632091), False, 'import pulumi\n'), ((632244, 632282), 'pulumi.set', 'pulumi.set', (['self', '"""permissions"""', 'value'], {}), "(self, 'permissions', value)\n", (632254, 632282), False, 'import pulumi\n'), ((633754, 633798), 'pulumi.get', 'pulumi.get', (['self', '"""max_connections_per_hour"""'], {}), "(self, 'max_connections_per_hour')\n", (633764, 633798), False, 'import pulumi\n'), ((633921, 633972), 'pulumi.set', 'pulumi.set', (['self', '"""max_connections_per_hour"""', 'value'], {}), "(self, 'max_connections_per_hour', value)\n", (633931, 633972), False, 'import pulumi\n'), ((634175, 634217), 'pulumi.get', 'pulumi.get', (['self', '"""max_questions_per_hour"""'], {}), "(self, 'max_questions_per_hour')\n", (634185, 634217), False, 'import pulumi\n'), ((634336, 634385), 'pulumi.set', 'pulumi.set', (['self', '"""max_questions_per_hour"""', 'value'], {}), "(self, 'max_questions_per_hour', value)\n", (634346, 634385), False, 'import pulumi\n'), ((634582, 634622), 'pulumi.get', 'pulumi.get', (['self', '"""max_updates_per_hour"""'], {}), "(self, 'max_updates_per_hour')\n", (634592, 634622), False, 'import pulumi\n'), ((634737, 634784), 'pulumi.set', 'pulumi.set', (['self', '"""max_updates_per_hour"""', 'value'], {}), "(self, 'max_updates_per_hour', value)\n", (634747, 634784), False, 'import pulumi\n'), ((634982, 635022), 'pulumi.get', 'pulumi.get', (['self', '"""max_user_connections"""'], {}), "(self, 'max_user_connections')\n", (634992, 635022), False, 'import pulumi\n'), ((635137, 635184), 'pulumi.set', 'pulumi.set', (['self', '"""max_user_connections"""', 'value'], {}), "(self, 'max_user_connections', value)\n", (635147, 635184), False, 'import pulumi\n'), ((635903, 635955), 'pulumi.set', 'pulumi.set', (['__self__', '"""database_name"""', 'database_name'], {}), "(__self__, 'database_name', database_name)\n", (635913, 635955), False, 'import pulumi\n'), ((636250, 636283), 'pulumi.get', 'pulumi.get', (['self', '"""database_name"""'], {}), "(self, 'database_name')\n", (636260, 636283), False, 'import pulumi\n'), ((636374, 636414), 'pulumi.set', 'pulumi.set', (['self', '"""database_name"""', 'value'], {}), "(self, 'database_name', value)\n", (636384, 636414), False, 'import pulumi\n'), ((636841, 636866), 'pulumi.get', 'pulumi.get', (['self', '"""roles"""'], {}), "(self, 'roles')\n", (636851, 636866), False, 'import pulumi\n'), ((636975, 637007), 'pulumi.set', 'pulumi.set', (['self', '"""roles"""', 'value'], {}), "(self, 'roles', value)\n", (636985, 637007), False, 'import pulumi\n'), ((638595, 638637), 'pulumi.set', 'pulumi.set', (['__self__', '"""password"""', 'password'], {}), "(__self__, 'password', password)\n", (638605, 638637), False, 'import pulumi\n'), ((638646, 638686), 'pulumi.set', 'pulumi.set', (['__self__', '"""version"""', 'version'], {}), "(__self__, 'version', version)\n", (638656, 638686), False, 'import pulumi\n'), ((639505, 639533), 'pulumi.get', 'pulumi.get', (['self', '"""password"""'], {}), "(self, 'password')\n", (639515, 639533), False, 'import pulumi\n'), ((639614, 639649), 'pulumi.set', 'pulumi.set', (['self', '"""password"""', 'value'], {}), "(self, 'password', value)\n", (639624, 639649), False, 'import pulumi\n'), ((639811, 639838), 'pulumi.get', 'pulumi.get', (['self', '"""version"""'], {}), "(self, 'version')\n", (639821, 639838), False, 'import pulumi\n'), ((639917, 639951), 'pulumi.set', 'pulumi.set', (['self', '"""version"""', 'value'], {}), "(self, 'version', value)\n", (639927, 639951), False, 'import pulumi\n'), ((640151, 640180), 'pulumi.get', 'pulumi.get', (['self', '"""databases"""'], {}), "(self, 'databases')\n", (640161, 640180), False, 'import pulumi\n'), ((640273, 640309), 'pulumi.set', 'pulumi.set', (['self', '"""databases"""', 'value'], {}), "(self, 'databases', value)\n", (640283, 640309), False, 'import pulumi\n'), ((640710, 640746), 'pulumi.get', 'pulumi.get', (['self', '"""maxmemory_policy"""'], {}), "(self, 'maxmemory_policy')\n", (640720, 640746), False, 'import pulumi\n'), ((640853, 640896), 'pulumi.set', 'pulumi.set', (['self', '"""maxmemory_policy"""', 'value'], {}), "(self, 'maxmemory_policy', value)\n", (640863, 640896), False, 'import pulumi\n'), ((641141, 641183), 'pulumi.get', 'pulumi.get', (['self', '"""notify_keyspace_events"""'], {}), "(self, 'notify_keyspace_events')\n", (641151, 641183), False, 'import pulumi\n'), ((641302, 641351), 'pulumi.set', 'pulumi.set', (['self', '"""notify_keyspace_events"""', 'value'], {}), "(self, 'notify_keyspace_events', value)\n", (641312, 641351), False, 'import pulumi\n'), ((641584, 641627), 'pulumi.get', 'pulumi.get', (['self', '"""slowlog_log_slower_than"""'], {}), "(self, 'slowlog_log_slower_than')\n", (641594, 641627), False, 'import pulumi\n'), ((641748, 641798), 'pulumi.set', 'pulumi.set', (['self', '"""slowlog_log_slower_than"""', 'value'], {}), "(self, 'slowlog_log_slower_than', value)\n", (641758, 641798), False, 'import pulumi\n'), ((641989, 642024), 'pulumi.get', 'pulumi.get', (['self', '"""slowlog_max_len"""'], {}), "(self, 'slowlog_max_len')\n", (641999, 642024), False, 'import pulumi\n'), ((642129, 642171), 'pulumi.set', 'pulumi.set', (['self', '"""slowlog_max_len"""', 'value'], {}), "(self, 'slowlog_max_len', value)\n", (642139, 642171), False, 'import pulumi\n'), ((642366, 642393), 'pulumi.get', 'pulumi.get', (['self', '"""timeout"""'], {}), "(self, 'timeout')\n", (642376, 642393), False, 'import pulumi\n'), ((642482, 642516), 'pulumi.set', 'pulumi.set', (['self', '"""timeout"""', 'value'], {}), "(self, 'timeout', value)\n", (642492, 642516), False, 'import pulumi\n'), ((643447, 643481), 'pulumi.set', 'pulumi.set', (['__self__', '"""zone"""', 'zone'], {}), "(__self__, 'zone', zone)\n", (643457, 643481), False, 'import pulumi\n'), ((644047, 644071), 'pulumi.get', 'pulumi.get', (['self', '"""zone"""'], {}), "(self, 'zone')\n", (644057, 644071), False, 'import pulumi\n'), ((644144, 644175), 'pulumi.set', 'pulumi.set', (['self', '"""zone"""', 'value'], {}), "(self, 'zone', value)\n", (644154, 644175), False, 'import pulumi\n'), ((644353, 644377), 'pulumi.get', 'pulumi.get', (['self', '"""fqdn"""'], {}), "(self, 'fqdn')\n", (644363, 644377), False, 'import pulumi\n'), ((644460, 644491), 'pulumi.set', 'pulumi.set', (['self', '"""fqdn"""', 'value'], {}), "(self, 'fqdn', value)\n", (644470, 644491), False, 'import pulumi\n'), ((644697, 644727), 'pulumi.get', 'pulumi.get', (['self', '"""shard_name"""'], {}), "(self, 'shard_name')\n", (644707, 644727), False, 'import pulumi\n'), ((644822, 644859), 'pulumi.set', 'pulumi.set', (['self', '"""shard_name"""', 'value'], {}), "(self, 'shard_name', value)\n", (644832, 644859), False, 'import pulumi\n'), ((645142, 645171), 'pulumi.get', 'pulumi.get', (['self', '"""subnet_id"""'], {}), "(self, 'subnet_id')\n", (645152, 645171), False, 'import pulumi\n'), ((645264, 645300), 'pulumi.set', 'pulumi.set', (['self', '"""subnet_id"""', 'value'], {}), "(self, 'subnet_id', value)\n", (645274, 645300), False, 'import pulumi\n'), ((646047, 646081), 'pulumi.set', 'pulumi.set', (['__self__', '"""type"""', 'type'], {}), "(__self__, 'type', type)\n", (646057, 646081), False, 'import pulumi\n'), ((646484, 646508), 'pulumi.get', 'pulumi.get', (['self', '"""type"""'], {}), "(self, 'type')\n", (646494, 646508), False, 'import pulumi\n'), ((646581, 646612), 'pulumi.set', 'pulumi.set', (['self', '"""type"""', 'value'], {}), "(self, 'type', value)\n", (646591, 646612), False, 'import pulumi\n'), ((646871, 646894), 'pulumi.get', 'pulumi.get', (['self', '"""day"""'], {}), "(self, 'day')\n", (646881, 646894), False, 'import pulumi\n'), ((646975, 647005), 'pulumi.set', 'pulumi.set', (['self', '"""day"""', 'value'], {}), "(self, 'day', value)\n", (646985, 647005), False, 'import pulumi\n'), ((647223, 647247), 'pulumi.get', 'pulumi.get', (['self', '"""hour"""'], {}), "(self, 'hour')\n", (647233, 647247), False, 'import pulumi\n'), ((647330, 647361), 'pulumi.set', 'pulumi.set', (['self', '"""hour"""', 'value'], {}), "(self, 'hour', value)\n", (647340, 647361), False, 'import pulumi\n'), ((647877, 647921), 'pulumi.set', 'pulumi.set', (['__self__', '"""disk_size"""', 'disk_size'], {}), "(__self__, 'disk_size', disk_size)\n", (647887, 647921), False, 'import pulumi\n'), ((647930, 647992), 'pulumi.set', 'pulumi.set', (['__self__', '"""resource_preset_id"""', 'resource_preset_id'], {}), "(__self__, 'resource_preset_id', resource_preset_id)\n", (647940, 647992), False, 'import pulumi\n'), ((648294, 648323), 'pulumi.get', 'pulumi.get', (['self', '"""disk_size"""'], {}), "(self, 'disk_size')\n", (648304, 648323), False, 'import pulumi\n'), ((648406, 648442), 'pulumi.set', 'pulumi.set', (['self', '"""disk_size"""', 'value'], {}), "(self, 'disk_size', value)\n", (648416, 648442), False, 'import pulumi\n'), ((648572, 648610), 'pulumi.get', 'pulumi.get', (['self', '"""resource_preset_id"""'], {}), "(self, 'resource_preset_id')\n", (648582, 648610), False, 'import pulumi\n'), ((648711, 648756), 'pulumi.set', 'pulumi.set', (['self', '"""resource_preset_id"""', 'value'], {}), "(self, 'resource_preset_id', value)\n", (648721, 648756), False, 'import pulumi\n'), ((648993, 649025), 'pulumi.get', 'pulumi.get', (['self', '"""disk_type_id"""'], {}), "(self, 'disk_type_id')\n", (649003, 649025), False, 'import pulumi\n'), ((649124, 649163), 'pulumi.set', 'pulumi.set', (['self', '"""disk_type_id"""', 'value'], {}), "(self, 'disk_type_id', value)\n", (649134, 649163), False, 'import pulumi\n'), ((649917, 649942), 'pulumi.get', 'pulumi.get', (['self', '"""hours"""'], {}), "(self, 'hours')\n", (649927, 649942), False, 'import pulumi\n'), ((650027, 650059), 'pulumi.set', 'pulumi.set', (['self', '"""hours"""', 'value'], {}), "(self, 'hours', value)\n", (650037, 650059), False, 'import pulumi\n'), ((650239, 650266), 'pulumi.get', 'pulumi.get', (['self', '"""minutes"""'], {}), "(self, 'minutes')\n", (650249, 650266), False, 'import pulumi\n'), ((650355, 650389), 'pulumi.set', 'pulumi.set', (['self', '"""minutes"""', 'value'], {}), "(self, 'minutes', value)\n", (650365, 650389), False, 'import pulumi\n'), ((650620, 650654), 'pulumi.set', 'pulumi.set', (['__self__', '"""name"""', 'name'], {}), "(__self__, 'name', name)\n", (650630, 650654), False, 'import pulumi\n'), ((650803, 650827), 'pulumi.get', 'pulumi.get', (['self', '"""name"""'], {}), "(self, 'name')\n", (650813, 650827), False, 'import pulumi\n'), ((650900, 650931), 'pulumi.set', 'pulumi.set', (['self', '"""name"""', 'value'], {}), "(self, 'name', value)\n", (650910, 650931), False, 'import pulumi\n'), ((651833, 651867), 'pulumi.set', 'pulumi.set', (['__self__', '"""zone"""', 'zone'], {}), "(__self__, 'zone', zone)\n", (651843, 651867), False, 'import pulumi\n'), ((652333, 652357), 'pulumi.get', 'pulumi.get', (['self', '"""zone"""'], {}), "(self, 'zone')\n", (652343, 652357), False, 'import pulumi\n'), ((652430, 652461), 'pulumi.set', 'pulumi.set', (['self', '"""zone"""', 'value'], {}), "(self, 'zone', value)\n", (652440, 652461), False, 'import pulumi\n'), ((652772, 652808), 'pulumi.get', 'pulumi.get', (['self', '"""assign_public_ip"""'], {}), "(self, 'assign_public_ip')\n", (652782, 652808), False, 'import pulumi\n'), ((652916, 652959), 'pulumi.set', 'pulumi.set', (['self', '"""assign_public_ip"""', 'value'], {}), "(self, 'assign_public_ip', value)\n", (652926, 652959), False, 'import pulumi\n'), ((653137, 653161), 'pulumi.get', 'pulumi.get', (['self', '"""fqdn"""'], {}), "(self, 'fqdn')\n", (653147, 653161), False, 'import pulumi\n'), ((653244, 653275), 'pulumi.set', 'pulumi.set', (['self', '"""fqdn"""', 'value'], {}), "(self, 'fqdn', value)\n", (653254, 653275), False, 'import pulumi\n'), ((653550, 653579), 'pulumi.get', 'pulumi.get', (['self', '"""subnet_id"""'], {}), "(self, 'subnet_id')\n", (653560, 653579), False, 'import pulumi\n'), ((653672, 653708), 'pulumi.set', 'pulumi.set', (['self', '"""subnet_id"""', 'value'], {}), "(self, 'subnet_id', value)\n", (653682, 653708), False, 'import pulumi\n'), ((654184, 654228), 'pulumi.set', 'pulumi.set', (['__self__', '"""disk_size"""', 'disk_size'], {}), "(__self__, 'disk_size', disk_size)\n", (654194, 654228), False, 'import pulumi\n'), ((654237, 654287), 'pulumi.set', 'pulumi.set', (['__self__', '"""disk_type_id"""', 'disk_type_id'], {}), "(__self__, 'disk_type_id', disk_type_id)\n", (654247, 654287), False, 'import pulumi\n'), ((654296, 654358), 'pulumi.set', 'pulumi.set', (['__self__', '"""resource_preset_id"""', 'resource_preset_id'], {}), "(__self__, 'resource_preset_id', resource_preset_id)\n", (654306, 654358), False, 'import pulumi\n'), ((654570, 654599), 'pulumi.get', 'pulumi.get', (['self', '"""disk_size"""'], {}), "(self, 'disk_size')\n", (654580, 654599), False, 'import pulumi\n'), ((654682, 654718), 'pulumi.set', 'pulumi.set', (['self', '"""disk_size"""', 'value'], {}), "(self, 'disk_size', value)\n", (654692, 654718), False, 'import pulumi\n'), ((654908, 654940), 'pulumi.get', 'pulumi.get', (['self', '"""disk_type_id"""'], {}), "(self, 'disk_type_id')\n", (654918, 654940), False, 'import pulumi\n'), ((655029, 655068), 'pulumi.set', 'pulumi.set', (['self', '"""disk_type_id"""', 'value'], {}), "(self, 'disk_type_id', value)\n", (655039, 655068), False, 'import pulumi\n'), ((655198, 655236), 'pulumi.get', 'pulumi.get', (['self', '"""resource_preset_id"""'], {}), "(self, 'resource_preset_id')\n", (655208, 655236), False, 'import pulumi\n'), ((655337, 655382), 'pulumi.set', 'pulumi.set', (['self', '"""resource_preset_id"""', 'value'], {}), "(self, 'resource_preset_id', value)\n", (655347, 655382), False, 'import pulumi\n'), ((656017, 656051), 'pulumi.set', 'pulumi.set', (['__self__', '"""name"""', 'name'], {}), "(__self__, 'name', name)\n", (656027, 656051), False, 'import pulumi\n'), ((656060, 656102), 'pulumi.set', 'pulumi.set', (['__self__', '"""password"""', 'password'], {}), "(__self__, 'password', password)\n", (656070, 656102), False, 'import pulumi\n'), ((656348, 656372), 'pulumi.get', 'pulumi.get', (['self', '"""name"""'], {}), "(self, 'name')\n", (656358, 656372), False, 'import pulumi\n'), ((656445, 656476), 'pulumi.set', 'pulumi.set', (['self', '"""name"""', 'value'], {}), "(self, 'name', value)\n", (656455, 656476), False, 'import pulumi\n'), ((656629, 656657), 'pulumi.get', 'pulumi.get', (['self', '"""password"""'], {}), "(self, 'password')\n", (656639, 656657), False, 'import pulumi\n'), ((656738, 656773), 'pulumi.set', 'pulumi.set', (['self', '"""password"""', 'value'], {}), "(self, 'password', value)\n", (656748, 656773), False, 'import pulumi\n'), ((657048, 657079), 'pulumi.get', 'pulumi.get', (['self', '"""permissions"""'], {}), "(self, 'permissions')\n", (657058, 657079), False, 'import pulumi\n'), ((657236, 657274), 'pulumi.set', 'pulumi.set', (['self', '"""permissions"""', 'value'], {}), "(self, 'permissions', value)\n", (657246, 657274), False, 'import pulumi\n'), ((657914, 657966), 'pulumi.set', 'pulumi.set', (['__self__', '"""database_name"""', 'database_name'], {}), "(__self__, 'database_name', database_name)\n", (657924, 657966), False, 'import pulumi\n'), ((658261, 658294), 'pulumi.get', 'pulumi.get', (['self', '"""database_name"""'], {}), "(self, 'database_name')\n", (658271, 658294), False, 'import pulumi\n'), ((658385, 658425), 'pulumi.set', 'pulumi.set', (['self', '"""database_name"""', 'value'], {}), "(self, 'database_name', value)\n", (658395, 658425), False, 'import pulumi\n'), ((658776, 658801), 'pulumi.get', 'pulumi.get', (['self', '"""roles"""'], {}), "(self, 'roles')\n", (658786, 658801), False, 'import pulumi\n'), ((658910, 658942), 'pulumi.set', 'pulumi.set', (['self', '"""roles"""', 'value'], {}), "(self, 'roles', value)\n", (658920, 658942), False, 'import pulumi\n'), ((660104, 660160), 'pulumi.set', 'pulumi.set', (['__self__', '"""allowed_methods"""', 'allowed_methods'], {}), "(__self__, 'allowed_methods', allowed_methods)\n", (660114, 660160), False, 'import pulumi\n'), ((660169, 660225), 'pulumi.set', 'pulumi.set', (['__self__', '"""allowed_origins"""', 'allowed_origins'], {}), "(__self__, 'allowed_origins', allowed_origins)\n", (660179, 660225), False, 'import pulumi\n'), ((660816, 660851), 'pulumi.get', 'pulumi.get', (['self', '"""allowed_methods"""'], {}), "(self, 'allowed_methods')\n", (660826, 660851), False, 'import pulumi\n'), ((660970, 661012), 'pulumi.set', 'pulumi.set', (['self', '"""allowed_methods"""', 'value'], {}), "(self, 'allowed_methods', value)\n", (660980, 661012), False, 'import pulumi\n'), ((661230, 661265), 'pulumi.get', 'pulumi.get', (['self', '"""allowed_origins"""'], {}), "(self, 'allowed_origins')\n", (661240, 661265), False, 'import pulumi\n'), ((661384, 661426), 'pulumi.set', 'pulumi.set', (['self', '"""allowed_origins"""', 'value'], {}), "(self, 'allowed_origins', value)\n", (661394, 661426), False, 'import pulumi\n'), ((661654, 661689), 'pulumi.get', 'pulumi.get', (['self', '"""allowed_headers"""'], {}), "(self, 'allowed_headers')\n", (661664, 661689), False, 'import pulumi\n'), ((661818, 661860), 'pulumi.set', 'pulumi.set', (['self', '"""allowed_headers"""', 'value'], {}), "(self, 'allowed_headers', value)\n", (661828, 661860), False, 'import pulumi\n'), ((662090, 662124), 'pulumi.get', 'pulumi.get', (['self', '"""expose_headers"""'], {}), "(self, 'expose_headers')\n", (662100, 662124), False, 'import pulumi\n'), ((662251, 662292), 'pulumi.set', 'pulumi.set', (['self', '"""expose_headers"""', 'value'], {}), "(self, 'expose_headers', value)\n", (662261, 662292), False, 'import pulumi\n'), ((662545, 662580), 'pulumi.get', 'pulumi.get', (['self', '"""max_age_seconds"""'], {}), "(self, 'max_age_seconds')\n", (662555, 662580), False, 'import pulumi\n'), ((662685, 662727), 'pulumi.set', 'pulumi.set', (['self', '"""max_age_seconds"""', 'value'], {}), "(self, 'max_age_seconds', value)\n", (662695, 662727), False, 'import pulumi\n'), ((663197, 663245), 'pulumi.set', 'pulumi.set', (['__self__', '"""permissions"""', 'permissions'], {}), "(__self__, 'permissions', permissions)\n", (663207, 663245), False, 'import pulumi\n'), ((663254, 663288), 'pulumi.set', 'pulumi.set', (['__self__', '"""type"""', 'type'], {}), "(__self__, 'type', type)\n", (663264, 663288), False, 'import pulumi\n'), ((663553, 663584), 'pulumi.get', 'pulumi.get', (['self', '"""permissions"""'], {}), "(self, 'permissions')\n", (663563, 663584), False, 'import pulumi\n'), ((663695, 663733), 'pulumi.set', 'pulumi.set', (['self', '"""permissions"""', 'value'], {}), "(self, 'permissions', value)\n", (663705, 663733), False, 'import pulumi\n'), ((663824, 663848), 'pulumi.get', 'pulumi.get', (['self', '"""type"""'], {}), "(self, 'type')\n", (663834, 663848), False, 'import pulumi\n'), ((663921, 663952), 'pulumi.set', 'pulumi.set', (['self', '"""type"""', 'value'], {}), "(self, 'type', value)\n", (663931, 663952), False, 'import pulumi\n'), ((664171, 664193), 'pulumi.get', 'pulumi.get', (['self', '"""id"""'], {}), "(self, 'id')\n", (664181, 664193), False, 'import pulumi\n'), ((664272, 664301), 'pulumi.set', 'pulumi.set', (['self', '"""id"""', 'value'], {}), "(self, 'id', value)\n", (664282, 664301), False, 'import pulumi\n'), ((664401, 664424), 'pulumi.get', 'pulumi.get', (['self', '"""uri"""'], {}), "(self, 'uri')\n", (664411, 664424), False, 'import pulumi\n'), ((664505, 664535), 'pulumi.set', 'pulumi.set', (['self', '"""uri"""', 'value'], {}), "(self, 'uri', value)\n", (664515, 664535), False, 'import pulumi\n'), ((666685, 666725), 'pulumi.set', 'pulumi.set', (['__self__', '"""enabled"""', 'enabled'], {}), "(__self__, 'enabled', enabled)\n", (666695, 666725), False, 'import pulumi\n'), ((667711, 667738), 'pulumi.get', 'pulumi.get', (['self', '"""enabled"""'], {}), "(self, 'enabled')\n", (667721, 667738), False, 'import pulumi\n'), ((667818, 667852), 'pulumi.set', 'pulumi.set', (['self', '"""enabled"""', 'value'], {}), "(self, 'enabled', value)\n", (667828, 667852), False, 'import pulumi\n'), ((668172, 668230), 'pulumi.get', 'pulumi.get', (['self', '"""abort_incomplete_multipart_upload_days"""'], {}), "(self, 'abort_incomplete_multipart_upload_days')\n", (668182, 668230), False, 'import pulumi\n'), ((668381, 668446), 'pulumi.set', 'pulumi.set', (['self', '"""abort_incomplete_multipart_upload_days"""', 'value'], {}), "(self, 'abort_incomplete_multipart_upload_days', value)\n", (668391, 668446), False, 'import pulumi\n'), ((668686, 668716), 'pulumi.get', 'pulumi.get', (['self', '"""expiration"""'], {}), "(self, 'expiration')\n", (668696, 668716), False, 'import pulumi\n'), ((668850, 668887), 'pulumi.set', 'pulumi.set', (['self', '"""expiration"""', 'value'], {}), "(self, 'expiration', value)\n", (668860, 668887), False, 'import pulumi\n'), ((669106, 669128), 'pulumi.get', 'pulumi.get', (['self', '"""id"""'], {}), "(self, 'id')\n", (669116, 669128), False, 'import pulumi\n'), ((669207, 669236), 'pulumi.set', 'pulumi.set', (['self', '"""id"""', 'value'], {}), "(self, 'id', value)\n", (669217, 669236), False, 'import pulumi\n'), ((669555, 669604), 'pulumi.get', 'pulumi.get', (['self', '"""noncurrent_version_expiration"""'], {}), "(self, 'noncurrent_version_expiration')\n", (669565, 669604), False, 'import pulumi\n'), ((669793, 669849), 'pulumi.set', 'pulumi.set', (['self', '"""noncurrent_version_expiration"""', 'value'], {}), "(self, 'noncurrent_version_expiration', value)\n", (669803, 669849), False, 'import pulumi\n'), ((670199, 670249), 'pulumi.get', 'pulumi.get', (['self', '"""noncurrent_version_transitions"""'], {}), "(self, 'noncurrent_version_transitions')\n", (670209, 670249), False, 'import pulumi\n'), ((670464, 670521), 'pulumi.set', 'pulumi.set', (['self', '"""noncurrent_version_transitions"""', 'value'], {}), "(self, 'noncurrent_version_transitions', value)\n", (670474, 670521), False, 'import pulumi\n'), ((670733, 670759), 'pulumi.get', 'pulumi.get', (['self', '"""prefix"""'], {}), "(self, 'prefix')\n", (670743, 670759), False, 'import pulumi\n'), ((670846, 670879), 'pulumi.set', 'pulumi.set', (['self', '"""prefix"""', 'value'], {}), "(self, 'prefix', value)\n", (670856, 670879), False, 'import pulumi\n'), ((671149, 671180), 'pulumi.get', 'pulumi.get', (['self', '"""transitions"""'], {}), "(self, 'transitions')\n", (671159, 671180), False, 'import pulumi\n'), ((671340, 671378), 'pulumi.set', 'pulumi.set', (['self', '"""transitions"""', 'value'], {}), "(self, 'transitions', value)\n", (671350, 671378), False, 'import pulumi\n'), ((672733, 672757), 'pulumi.get', 'pulumi.get', (['self', '"""date"""'], {}), "(self, 'date')\n", (672743, 672757), False, 'import pulumi\n'), ((672840, 672871), 'pulumi.set', 'pulumi.set', (['self', '"""date"""', 'value'], {}), "(self, 'date', value)\n", (672850, 672871), False, 'import pulumi\n'), ((673099, 673123), 'pulumi.get', 'pulumi.get', (['self', '"""days"""'], {}), "(self, 'days')\n", (673109, 673123), False, 'import pulumi\n'), ((673206, 673237), 'pulumi.set', 'pulumi.set', (['self', '"""days"""', 'value'], {}), "(self, 'days', value)\n", (673216, 673237), False, 'import pulumi\n'), ((673626, 673674), 'pulumi.get', 'pulumi.get', (['self', '"""expired_object_delete_marker"""'], {}), "(self, 'expired_object_delete_marker')\n", (673636, 673674), False, 'import pulumi\n'), ((673806, 673861), 'pulumi.set', 'pulumi.set', (['self', '"""expired_object_delete_marker"""', 'value'], {}), "(self, 'expired_object_delete_marker', value)\n", (673816, 673861), False, 'import pulumi\n'), ((674437, 674461), 'pulumi.get', 'pulumi.get', (['self', '"""days"""'], {}), "(self, 'days')\n", (674447, 674461), False, 'import pulumi\n'), ((674544, 674575), 'pulumi.set', 'pulumi.set', (['self', '"""days"""', 'value'], {}), "(self, 'days', value)\n", (674554, 674575), False, 'import pulumi\n'), ((675119, 675171), 'pulumi.set', 'pulumi.set', (['__self__', '"""storage_class"""', 'storage_class'], {}), "(__self__, 'storage_class', storage_class)\n", (675129, 675171), False, 'import pulumi\n'), ((675529, 675562), 'pulumi.get', 'pulumi.get', (['self', '"""storage_class"""'], {}), "(self, 'storage_class')\n", (675539, 675562), False, 'import pulumi\n'), ((675653, 675693), 'pulumi.set', 'pulumi.set', (['self', '"""storage_class"""', 'value'], {}), "(self, 'storage_class', value)\n", (675663, 675693), False, 'import pulumi\n'), ((675894, 675918), 'pulumi.get', 'pulumi.get', (['self', '"""days"""'], {}), "(self, 'days')\n", (675904, 675918), False, 'import pulumi\n'), ((676001, 676032), 'pulumi.set', 'pulumi.set', (['self', '"""days"""', 'value'], {}), "(self, 'days', value)\n", (676011, 676032), False, 'import pulumi\n'), ((676745, 676797), 'pulumi.set', 'pulumi.set', (['__self__', '"""storage_class"""', 'storage_class'], {}), "(__self__, 'storage_class', storage_class)\n", (676755, 676797), False, 'import pulumi\n'), ((677211, 677244), 'pulumi.get', 'pulumi.get', (['self', '"""storage_class"""'], {}), "(self, 'storage_class')\n", (677221, 677244), False, 'import pulumi\n'), ((677335, 677375), 'pulumi.set', 'pulumi.set', (['self', '"""storage_class"""', 'value'], {}), "(self, 'storage_class', value)\n", (677345, 677375), False, 'import pulumi\n'), ((677589, 677613), 'pulumi.get', 'pulumi.get', (['self', '"""date"""'], {}), "(self, 'date')\n", (677599, 677613), False, 'import pulumi\n'), ((677696, 677727), 'pulumi.set', 'pulumi.set', (['self', '"""date"""', 'value'], {}), "(self, 'date', value)\n", (677706, 677727), False, 'import pulumi\n'), ((677955, 677979), 'pulumi.get', 'pulumi.get', (['self', '"""days"""'], {}), "(self, 'days')\n", (677965, 677979), False, 'import pulumi\n'), ((678062, 678093), 'pulumi.set', 'pulumi.set', (['self', '"""days"""', 'value'], {}), "(self, 'days', value)\n", (678072, 678093), False, 'import pulumi\n'), ((678524, 678576), 'pulumi.set', 'pulumi.set', (['__self__', '"""target_bucket"""', 'target_bucket'], {}), "(__self__, 'target_bucket', target_bucket)\n", (678534, 678576), False, 'import pulumi\n'), ((678890, 678923), 'pulumi.get', 'pulumi.get', (['self', '"""target_bucket"""'], {}), "(self, 'target_bucket')\n", (678900, 678923), False, 'import pulumi\n'), ((679014, 679054), 'pulumi.set', 'pulumi.set', (['self', '"""target_bucket"""', 'value'], {}), "(self, 'target_bucket', value)\n", (679024, 679054), False, 'import pulumi\n'), ((679258, 679291), 'pulumi.get', 'pulumi.get', (['self', '"""target_prefix"""'], {}), "(self, 'target_prefix')\n", (679268, 679291), False, 'import pulumi\n'), ((679392, 679432), 'pulumi.set', 'pulumi.set', (['self', '"""target_prefix"""', 'value'], {}), "(self, 'target_prefix', value)\n", (679402, 679432), False, 'import pulumi\n'), ((679850, 679884), 'pulumi.set', 'pulumi.set', (['__self__', '"""rule"""', 'rule'], {}), "(__self__, 'rule', rule)\n", (679860, 679884), False, 'import pulumi\n'), ((680148, 680172), 'pulumi.get', 'pulumi.get', (['self', '"""rule"""'], {}), "(self, 'rule')\n", (680158, 680172), False, 'import pulumi\n'), ((680298, 680329), 'pulumi.set', 'pulumi.set', (['self', '"""rule"""', 'value'], {}), "(self, 'rule', value)\n", (680308, 680329), False, 'import pulumi\n'), ((680883, 680991), 'pulumi.set', 'pulumi.set', (['__self__', '"""apply_server_side_encryption_by_default"""', 'apply_server_side_encryption_by_default'], {}), "(__self__, 'apply_server_side_encryption_by_default',\n apply_server_side_encryption_by_default)\n", (680893, 680991), False, 'import pulumi\n'), ((681357, 681416), 'pulumi.get', 'pulumi.get', (['self', '"""apply_server_side_encryption_by_default"""'], {}), "(self, 'apply_server_side_encryption_by_default')\n", (681367, 681416), False, 'import pulumi\n'), ((681646, 681712), 'pulumi.set', 'pulumi.set', (['self', '"""apply_server_side_encryption_by_default"""', 'value'], {}), "(self, 'apply_server_side_encryption_by_default', value)\n", (681656, 681712), False, 'import pulumi\n'), ((682231, 682291), 'pulumi.set', 'pulumi.set', (['__self__', '"""kms_master_key_id"""', 'kms_master_key_id'], {}), "(__self__, 'kms_master_key_id', kms_master_key_id)\n", (682241, 682291), False, 'import pulumi\n'), ((682300, 682352), 'pulumi.set', 'pulumi.set', (['__self__', '"""sse_algorithm"""', 'sse_algorithm'], {}), "(__self__, 'sse_algorithm', sse_algorithm)\n", (682310, 682352), False, 'import pulumi\n'), ((682566, 682603), 'pulumi.get', 'pulumi.get', (['self', '"""kms_master_key_id"""'], {}), "(self, 'kms_master_key_id')\n", (682576, 682603), False, 'import pulumi\n'), ((682702, 682746), 'pulumi.set', 'pulumi.set', (['self', '"""kms_master_key_id"""', 'value'], {}), "(self, 'kms_master_key_id', value)\n", (682712, 682746), False, 'import pulumi\n'), ((682976, 683009), 'pulumi.get', 'pulumi.get', (['self', '"""sse_algorithm"""'], {}), "(self, 'sse_algorithm')\n", (682986, 683009), False, 'import pulumi\n'), ((683100, 683140), 'pulumi.set', 'pulumi.set', (['self', '"""sse_algorithm"""', 'value'], {}), "(self, 'sse_algorithm', value)\n", (683110, 683140), False, 'import pulumi\n'), ((683881, 683908), 'pulumi.get', 'pulumi.get', (['self', '"""enabled"""'], {}), "(self, 'enabled')\n", (683891, 683908), False, 'import pulumi\n'), ((683998, 684032), 'pulumi.set', 'pulumi.set', (['self', '"""enabled"""', 'value'], {}), "(self, 'enabled', value)\n", (684008, 684032), False, 'import pulumi\n'), ((685907, 685941), 'pulumi.get', 'pulumi.get', (['self', '"""error_document"""'], {}), "(self, 'error_document')\n", (685917, 685941), False, 'import pulumi\n'), ((686044, 686085), 'pulumi.set', 'pulumi.set', (['self', '"""error_document"""', 'value'], {}), "(self, 'error_document', value)\n", (686054, 686085), False, 'import pulumi\n'), ((686354, 686388), 'pulumi.get', 'pulumi.get', (['self', '"""index_document"""'], {}), "(self, 'index_document')\n", (686364, 686388), False, 'import pulumi\n'), ((686491, 686532), 'pulumi.set', 'pulumi.set', (['self', '"""index_document"""', 'value'], {}), "(self, 'index_document', value)\n", (686501, 686532), False, 'import pulumi\n'), ((686957, 687001), 'pulumi.get', 'pulumi.get', (['self', '"""redirect_all_requests_to"""'], {}), "(self, 'redirect_all_requests_to')\n", (686967, 687001), False, 'import pulumi\n'), ((687124, 687175), 'pulumi.set', 'pulumi.set', (['self', '"""redirect_all_requests_to"""', 'value'], {}), "(self, 'redirect_all_requests_to', value)\n", (687134, 687175), False, 'import pulumi\n'), ((687519, 687552), 'pulumi.get', 'pulumi.get', (['self', '"""routing_rules"""'], {}), "(self, 'routing_rules')\n", (687529, 687552), False, 'import pulumi\n'), ((687653, 687693), 'pulumi.set', 'pulumi.set', (['self', '"""routing_rules"""', 'value'], {}), "(self, 'routing_rules', value)\n", (687663, 687693), False, 'import pulumi\n'), ((689018, 689045), 'pulumi.get', 'pulumi.get', (['self', '"""address"""'], {}), "(self, 'address')\n", (689028, 689045), False, 'import pulumi\n'), ((689134, 689168), 'pulumi.set', 'pulumi.set', (['self', '"""address"""', 'value'], {}), "(self, 'address', value)\n", (689144, 689168), False, 'import pulumi\n'), ((689406, 689450), 'pulumi.get', 'pulumi.get', (['self', '"""ddos_protection_provider"""'], {}), "(self, 'ddos_protection_provider')\n", (689416, 689450), False, 'import pulumi\n'), ((689573, 689624), 'pulumi.set', 'pulumi.set', (['self', '"""ddos_protection_provider"""', 'value'], {}), "(self, 'ddos_protection_provider', value)\n", (689583, 689624), False, 'import pulumi\n'), ((689841, 689885), 'pulumi.get', 'pulumi.get', (['self', '"""outgoing_smtp_capability"""'], {}), "(self, 'outgoing_smtp_capability')\n", (689851, 689885), False, 'import pulumi\n'), ((690008, 690059), 'pulumi.set', 'pulumi.set', (['self', '"""outgoing_smtp_capability"""', 'value'], {}), "(self, 'outgoing_smtp_capability', value)\n", (690018, 690059), False, 'import pulumi\n'), ((690241, 690268), 'pulumi.get', 'pulumi.get', (['self', '"""zone_id"""'], {}), "(self, 'zone_id')\n", (690251, 690268), False, 'import pulumi\n'), ((690357, 690391), 'pulumi.set', 'pulumi.set', (['self', '"""zone_id"""', 'value'], {}), "(self, 'zone_id', value)\n", (690367, 690391), False, 'import pulumi\n'), ((691544, 691586), 'pulumi.set', 'pulumi.set', (['__self__', '"""protocol"""', 'protocol'], {}), "(__self__, 'protocol', protocol)\n", (691554, 691586), False, 'import pulumi\n'), ((692624, 692652), 'pulumi.get', 'pulumi.get', (['self', '"""protocol"""'], {}), "(self, 'protocol')\n", (692634, 692652), False, 'import pulumi\n'), ((692733, 692768), 'pulumi.set', 'pulumi.set', (['self', '"""protocol"""', 'value'], {}), "(self, 'protocol', value)\n", (692743, 692768), False, 'import pulumi\n'), ((692943, 692974), 'pulumi.get', 'pulumi.get', (['self', '"""description"""'], {}), "(self, 'description')\n", (692953, 692974), False, 'import pulumi\n'), ((693071, 693109), 'pulumi.set', 'pulumi.set', (['self', '"""description"""', 'value'], {}), "(self, 'description', value)\n", (693081, 693109), False, 'import pulumi\n'), ((693232, 693261), 'pulumi.get', 'pulumi.get', (['self', '"""from_port"""'], {}), "(self, 'from_port')\n", (693242, 693261), False, 'import pulumi\n'), ((693354, 693390), 'pulumi.set', 'pulumi.set', (['self', '"""from_port"""', 'value'], {}), "(self, 'from_port', value)\n", (693364, 693390), False, 'import pulumi\n'), ((693547, 693569), 'pulumi.get', 'pulumi.get', (['self', '"""id"""'], {}), "(self, 'id')\n", (693557, 693569), False, 'import pulumi\n'), ((693648, 693677), 'pulumi.set', 'pulumi.set', (['self', '"""id"""', 'value'], {}), "(self, 'id', value)\n", (693658, 693677), False, 'import pulumi\n'), ((693881, 693907), 'pulumi.get', 'pulumi.get', (['self', '"""labels"""'], {}), "(self, 'labels')\n", (693891, 693907), False, 'import pulumi\n'), ((694022, 694055), 'pulumi.set', 'pulumi.set', (['self', '"""labels"""', 'value'], {}), "(self, 'labels', value)\n", (694032, 694055), False, 'import pulumi\n'), ((694156, 694180), 'pulumi.get', 'pulumi.get', (['self', '"""port"""'], {}), "(self, 'port')\n", (694166, 694180), False, 'import pulumi\n'), ((694263, 694294), 'pulumi.set', 'pulumi.set', (['self', '"""port"""', 'value'], {}), "(self, 'port', value)\n", (694273, 694294), False, 'import pulumi\n'), ((694433, 694470), 'pulumi.get', 'pulumi.get', (['self', '"""predefined_target"""'], {}), "(self, 'predefined_target')\n", (694443, 694470), False, 'import pulumi\n'), ((694579, 694623), 'pulumi.set', 'pulumi.set', (['self', '"""predefined_target"""', 'value'], {}), "(self, 'predefined_target', value)\n", (694589, 694623), False, 'import pulumi\n'), ((694761, 694798), 'pulumi.get', 'pulumi.get', (['self', '"""security_group_id"""'], {}), "(self, 'security_group_id')\n", (694771, 694798), False, 'import pulumi\n'), ((694907, 694951), 'pulumi.set', 'pulumi.set', (['self', '"""security_group_id"""', 'value'], {}), "(self, 'security_group_id', value)\n", (694917, 694951), False, 'import pulumi\n'), ((695070, 695097), 'pulumi.get', 'pulumi.get', (['self', '"""to_port"""'], {}), "(self, 'to_port')\n", (695080, 695097), False, 'import pulumi\n'), ((695186, 695220), 'pulumi.set', 'pulumi.set', (['self', '"""to_port"""', 'value'], {}), "(self, 'to_port', value)\n", (695196, 695220), False, 'import pulumi\n'), ((695376, 695410), 'pulumi.get', 'pulumi.get', (['self', '"""v4_cidr_blocks"""'], {}), "(self, 'v4_cidr_blocks')\n", (695386, 695410), False, 'import pulumi\n'), ((695537, 695578), 'pulumi.set', 'pulumi.set', (['self', '"""v4_cidr_blocks"""', 'value'], {}), "(self, 'v4_cidr_blocks', value)\n", (695547, 695578), False, 'import pulumi\n'), ((695734, 695768), 'pulumi.get', 'pulumi.get', (['self', '"""v6_cidr_blocks"""'], {}), "(self, 'v6_cidr_blocks')\n", (695744, 695768), False, 'import pulumi\n'), ((695895, 695936), 'pulumi.set', 'pulumi.set', (['self', '"""v6_cidr_blocks"""', 'value'], {}), "(self, 'v6_cidr_blocks', value)\n", (695905, 695936), False, 'import pulumi\n'), ((697090, 697132), 'pulumi.set', 'pulumi.set', (['__self__', '"""protocol"""', 'protocol'], {}), "(__self__, 'protocol', protocol)\n", (697100, 697132), False, 'import pulumi\n'), ((698170, 698198), 'pulumi.get', 'pulumi.get', (['self', '"""protocol"""'], {}), "(self, 'protocol')\n", (698180, 698198), False, 'import pulumi\n'), ((698279, 698314), 'pulumi.set', 'pulumi.set', (['self', '"""protocol"""', 'value'], {}), "(self, 'protocol', value)\n", (698289, 698314), False, 'import pulumi\n'), ((698489, 698520), 'pulumi.get', 'pulumi.get', (['self', '"""description"""'], {}), "(self, 'description')\n", (698499, 698520), False, 'import pulumi\n'), ((698617, 698655), 'pulumi.set', 'pulumi.set', (['self', '"""description"""', 'value'], {}), "(self, 'description', value)\n", (698627, 698655), False, 'import pulumi\n'), ((698778, 698807), 'pulumi.get', 'pulumi.get', (['self', '"""from_port"""'], {}), "(self, 'from_port')\n", (698788, 698807), False, 'import pulumi\n'), ((698900, 698936), 'pulumi.set', 'pulumi.set', (['self', '"""from_port"""', 'value'], {}), "(self, 'from_port', value)\n", (698910, 698936), False, 'import pulumi\n'), ((699093, 699115), 'pulumi.get', 'pulumi.get', (['self', '"""id"""'], {}), "(self, 'id')\n", (699103, 699115), False, 'import pulumi\n'), ((699194, 699223), 'pulumi.set', 'pulumi.set', (['self', '"""id"""', 'value'], {}), "(self, 'id', value)\n", (699204, 699223), False, 'import pulumi\n'), ((699427, 699453), 'pulumi.get', 'pulumi.get', (['self', '"""labels"""'], {}), "(self, 'labels')\n", (699437, 699453), False, 'import pulumi\n'), ((699568, 699601), 'pulumi.set', 'pulumi.set', (['self', '"""labels"""', 'value'], {}), "(self, 'labels', value)\n", (699578, 699601), False, 'import pulumi\n'), ((699702, 699726), 'pulumi.get', 'pulumi.get', (['self', '"""port"""'], {}), "(self, 'port')\n", (699712, 699726), False, 'import pulumi\n'), ((699809, 699840), 'pulumi.set', 'pulumi.set', (['self', '"""port"""', 'value'], {}), "(self, 'port', value)\n", (699819, 699840), False, 'import pulumi\n'), ((699979, 700016), 'pulumi.get', 'pulumi.get', (['self', '"""predefined_target"""'], {}), "(self, 'predefined_target')\n", (699989, 700016), False, 'import pulumi\n'), ((700125, 700169), 'pulumi.set', 'pulumi.set', (['self', '"""predefined_target"""', 'value'], {}), "(self, 'predefined_target', value)\n", (700135, 700169), False, 'import pulumi\n'), ((700307, 700344), 'pulumi.get', 'pulumi.get', (['self', '"""security_group_id"""'], {}), "(self, 'security_group_id')\n", (700317, 700344), False, 'import pulumi\n'), ((700453, 700497), 'pulumi.set', 'pulumi.set', (['self', '"""security_group_id"""', 'value'], {}), "(self, 'security_group_id', value)\n", (700463, 700497), False, 'import pulumi\n'), ((700616, 700643), 'pulumi.get', 'pulumi.get', (['self', '"""to_port"""'], {}), "(self, 'to_port')\n", (700626, 700643), False, 'import pulumi\n'), ((700732, 700766), 'pulumi.set', 'pulumi.set', (['self', '"""to_port"""', 'value'], {}), "(self, 'to_port', value)\n", (700742, 700766), False, 'import pulumi\n'), ((700922, 700956), 'pulumi.get', 'pulumi.get', (['self', '"""v4_cidr_blocks"""'], {}), "(self, 'v4_cidr_blocks')\n", (700932, 700956), False, 'import pulumi\n'), ((701083, 701124), 'pulumi.set', 'pulumi.set', (['self', '"""v4_cidr_blocks"""', 'value'], {}), "(self, 'v4_cidr_blocks', value)\n", (701093, 701124), False, 'import pulumi\n'), ((701280, 701314), 'pulumi.get', 'pulumi.get', (['self', '"""v6_cidr_blocks"""'], {}), "(self, 'v6_cidr_blocks')\n", (701290, 701314), False, 'import pulumi\n'), ((701441, 701482), 'pulumi.set', 'pulumi.set', (['self', '"""v6_cidr_blocks"""', 'value'], {}), "(self, 'v6_cidr_blocks', value)\n", (701451, 701482), False, 'import pulumi\n'), ((702332, 702370), 'pulumi.get', 'pulumi.get', (['self', '"""destination_prefix"""'], {}), "(self, 'destination_prefix')\n", (702342, 702370), False, 'import pulumi\n'), ((702481, 702526), 'pulumi.set', 'pulumi.set', (['self', '"""destination_prefix"""', 'value'], {}), "(self, 'destination_prefix', value)\n", (702491, 702526), False, 'import pulumi\n'), ((702719, 702755), 'pulumi.get', 'pulumi.get', (['self', '"""next_hop_address"""'], {}), "(self, 'next_hop_address')\n", (702729, 702755), False, 'import pulumi\n'), ((702862, 702905), 'pulumi.set', 'pulumi.set', (['self', '"""next_hop_address"""', 'value'], {}), "(self, 'next_hop_address', value)\n", (702872, 702905), False, 'import pulumi\n'), ((704041, 704083), 'pulumi.set', 'pulumi.set', (['__self__', '"""protocol"""', 'protocol'], {}), "(__self__, 'protocol', protocol)\n", (704051, 704083), False, 'import pulumi\n'), ((705121, 705149), 'pulumi.get', 'pulumi.get', (['self', '"""protocol"""'], {}), "(self, 'protocol')\n", (705131, 705149), False, 'import pulumi\n'), ((705230, 705265), 'pulumi.set', 'pulumi.set', (['self', '"""protocol"""', 'value'], {}), "(self, 'protocol', value)\n", (705240, 705265), False, 'import pulumi\n'), ((705440, 705471), 'pulumi.get', 'pulumi.get', (['self', '"""description"""'], {}), "(self, 'description')\n", (705450, 705471), False, 'import pulumi\n'), ((705568, 705606), 'pulumi.set', 'pulumi.set', (['self', '"""description"""', 'value'], {}), "(self, 'description', value)\n", (705578, 705606), False, 'import pulumi\n'), ((705729, 705758), 'pulumi.get', 'pulumi.get', (['self', '"""from_port"""'], {}), "(self, 'from_port')\n", (705739, 705758), False, 'import pulumi\n'), ((705851, 705887), 'pulumi.set', 'pulumi.set', (['self', '"""from_port"""', 'value'], {}), "(self, 'from_port', value)\n", (705861, 705887), False, 'import pulumi\n'), ((706034, 706056), 'pulumi.get', 'pulumi.get', (['self', '"""id"""'], {}), "(self, 'id')\n", (706044, 706056), False, 'import pulumi\n'), ((706135, 706164), 'pulumi.set', 'pulumi.set', (['self', '"""id"""', 'value'], {}), "(self, 'id', value)\n", (706145, 706164), False, 'import pulumi\n'), ((706368, 706394), 'pulumi.get', 'pulumi.get', (['self', '"""labels"""'], {}), "(self, 'labels')\n", (706378, 706394), False, 'import pulumi\n'), ((706509, 706542), 'pulumi.set', 'pulumi.set', (['self', '"""labels"""', 'value'], {}), "(self, 'labels', value)\n", (706519, 706542), False, 'import pulumi\n'), ((706643, 706667), 'pulumi.get', 'pulumi.get', (['self', '"""port"""'], {}), "(self, 'port')\n", (706653, 706667), False, 'import pulumi\n'), ((706750, 706781), 'pulumi.set', 'pulumi.set', (['self', '"""port"""', 'value'], {}), "(self, 'port', value)\n", (706760, 706781), False, 'import pulumi\n'), ((706920, 706957), 'pulumi.get', 'pulumi.get', (['self', '"""predefined_target"""'], {}), "(self, 'predefined_target')\n", (706930, 706957), False, 'import pulumi\n'), ((707066, 707110), 'pulumi.set', 'pulumi.set', (['self', '"""predefined_target"""', 'value'], {}), "(self, 'predefined_target', value)\n", (707076, 707110), False, 'import pulumi\n'), ((707248, 707285), 'pulumi.get', 'pulumi.get', (['self', '"""security_group_id"""'], {}), "(self, 'security_group_id')\n", (707258, 707285), False, 'import pulumi\n'), ((707394, 707438), 'pulumi.set', 'pulumi.set', (['self', '"""security_group_id"""', 'value'], {}), "(self, 'security_group_id', value)\n", (707404, 707438), False, 'import pulumi\n'), ((707557, 707584), 'pulumi.get', 'pulumi.get', (['self', '"""to_port"""'], {}), "(self, 'to_port')\n", (707567, 707584), False, 'import pulumi\n'), ((707673, 707707), 'pulumi.set', 'pulumi.set', (['self', '"""to_port"""', 'value'], {}), "(self, 'to_port', value)\n", (707683, 707707), False, 'import pulumi\n'), ((707863, 707897), 'pulumi.get', 'pulumi.get', (['self', '"""v4_cidr_blocks"""'], {}), "(self, 'v4_cidr_blocks')\n", (707873, 707897), False, 'import pulumi\n'), ((708024, 708065), 'pulumi.set', 'pulumi.set', (['self', '"""v4_cidr_blocks"""', 'value'], {}), "(self, 'v4_cidr_blocks', value)\n", (708034, 708065), False, 'import pulumi\n'), ((708221, 708255), 'pulumi.get', 'pulumi.get', (['self', '"""v6_cidr_blocks"""'], {}), "(self, 'v6_cidr_blocks')\n", (708231, 708255), False, 'import pulumi\n'), ((708382, 708423), 'pulumi.set', 'pulumi.set', (['self', '"""v6_cidr_blocks"""', 'value'], {}), "(self, 'v6_cidr_blocks', value)\n", (708392, 708423), False, 'import pulumi\n'), ((709560, 709602), 'pulumi.set', 'pulumi.set', (['__self__', '"""protocol"""', 'protocol'], {}), "(__self__, 'protocol', protocol)\n", (709570, 709602), False, 'import pulumi\n'), ((710640, 710668), 'pulumi.get', 'pulumi.get', (['self', '"""protocol"""'], {}), "(self, 'protocol')\n", (710650, 710668), False, 'import pulumi\n'), ((710749, 710784), 'pulumi.set', 'pulumi.set', (['self', '"""protocol"""', 'value'], {}), "(self, 'protocol', value)\n", (710759, 710784), False, 'import pulumi\n'), ((710959, 710990), 'pulumi.get', 'pulumi.get', (['self', '"""description"""'], {}), "(self, 'description')\n", (710969, 710990), False, 'import pulumi\n'), ((711087, 711125), 'pulumi.set', 'pulumi.set', (['self', '"""description"""', 'value'], {}), "(self, 'description', value)\n", (711097, 711125), False, 'import pulumi\n'), ((711248, 711277), 'pulumi.get', 'pulumi.get', (['self', '"""from_port"""'], {}), "(self, 'from_port')\n", (711258, 711277), False, 'import pulumi\n'), ((711370, 711406), 'pulumi.set', 'pulumi.set', (['self', '"""from_port"""', 'value'], {}), "(self, 'from_port', value)\n", (711380, 711406), False, 'import pulumi\n'), ((711553, 711575), 'pulumi.get', 'pulumi.get', (['self', '"""id"""'], {}), "(self, 'id')\n", (711563, 711575), False, 'import pulumi\n'), ((711654, 711683), 'pulumi.set', 'pulumi.set', (['self', '"""id"""', 'value'], {}), "(self, 'id', value)\n", (711664, 711683), False, 'import pulumi\n'), ((711887, 711913), 'pulumi.get', 'pulumi.get', (['self', '"""labels"""'], {}), "(self, 'labels')\n", (711897, 711913), False, 'import pulumi\n'), ((712028, 712061), 'pulumi.set', 'pulumi.set', (['self', '"""labels"""', 'value'], {}), "(self, 'labels', value)\n", (712038, 712061), False, 'import pulumi\n'), ((712162, 712186), 'pulumi.get', 'pulumi.get', (['self', '"""port"""'], {}), "(self, 'port')\n", (712172, 712186), False, 'import pulumi\n'), ((712269, 712300), 'pulumi.set', 'pulumi.set', (['self', '"""port"""', 'value'], {}), "(self, 'port', value)\n", (712279, 712300), False, 'import pulumi\n'), ((712439, 712476), 'pulumi.get', 'pulumi.get', (['self', '"""predefined_target"""'], {}), "(self, 'predefined_target')\n", (712449, 712476), False, 'import pulumi\n'), ((712585, 712629), 'pulumi.set', 'pulumi.set', (['self', '"""predefined_target"""', 'value'], {}), "(self, 'predefined_target', value)\n", (712595, 712629), False, 'import pulumi\n'), ((712767, 712804), 'pulumi.get', 'pulumi.get', (['self', '"""security_group_id"""'], {}), "(self, 'security_group_id')\n", (712777, 712804), False, 'import pulumi\n'), ((712913, 712957), 'pulumi.set', 'pulumi.set', (['self', '"""security_group_id"""', 'value'], {}), "(self, 'security_group_id', value)\n", (712923, 712957), False, 'import pulumi\n'), ((713076, 713103), 'pulumi.get', 'pulumi.get', (['self', '"""to_port"""'], {}), "(self, 'to_port')\n", (713086, 713103), False, 'import pulumi\n'), ((713192, 713226), 'pulumi.set', 'pulumi.set', (['self', '"""to_port"""', 'value'], {}), "(self, 'to_port', value)\n", (713202, 713226), False, 'import pulumi\n'), ((713382, 713416), 'pulumi.get', 'pulumi.get', (['self', '"""v4_cidr_blocks"""'], {}), "(self, 'v4_cidr_blocks')\n", (713392, 713416), False, 'import pulumi\n'), ((713543, 713584), 'pulumi.set', 'pulumi.set', (['self', '"""v4_cidr_blocks"""', 'value'], {}), "(self, 'v4_cidr_blocks', value)\n", (713553, 713584), False, 'import pulumi\n'), ((713740, 713774), 'pulumi.get', 'pulumi.get', (['self', '"""v6_cidr_blocks"""'], {}), "(self, 'v6_cidr_blocks')\n", (713750, 713774), False, 'import pulumi\n'), ((713901, 713942), 'pulumi.set', 'pulumi.set', (['self', '"""v6_cidr_blocks"""', 'value'], {}), "(self, 'v6_cidr_blocks', value)\n", (713911, 713942), False, 'import pulumi\n'), ((715056, 715087), 'pulumi.get', 'pulumi.get', (['self', '"""domain_name"""'], {}), "(self, 'domain_name')\n", (715066, 715087), False, 'import pulumi\n'), ((715184, 715222), 'pulumi.set', 'pulumi.set', (['self', '"""domain_name"""', 'value'], {}), "(self, 'domain_name', value)\n", (715194, 715222), False, 'import pulumi\n'), ((715453, 715492), 'pulumi.get', 'pulumi.get', (['self', '"""domain_name_servers"""'], {}), "(self, 'domain_name_servers')\n", (715463, 715492), False, 'import pulumi\n'), ((715629, 715675), 'pulumi.set', 'pulumi.set', (['self', '"""domain_name_servers"""', 'value'], {}), "(self, 'domain_name_servers', value)\n", (715639, 715675), False, 'import pulumi\n'), ((715883, 715914), 'pulumi.get', 'pulumi.get', (['self', '"""ntp_servers"""'], {}), "(self, 'ntp_servers')\n", (715893, 715914), False, 'import pulumi\n'), ((716035, 716073), 'pulumi.set', 'pulumi.set', (['self', '"""ntp_servers"""', 'value'], {}), "(self, 'ntp_servers', value)\n", (716045, 716073), False, 'import pulumi\n'), ((716792, 716818), 'pulumi.get', 'pulumi.get', (['self', '"""region"""'], {}), "(self, 'region')\n", (716802, 716818), False, 'import pulumi\n'), ((716942, 716975), 'pulumi.set', 'pulumi.set', (['self', '"""region"""', 'value'], {}), "(self, 'region', value)\n", (716952, 716975), False, 'import pulumi\n'), ((717226, 717256), 'pulumi.set', 'pulumi.set', (['__self__', '"""id"""', 'id'], {}), "(__self__, 'id', id)\n", (717236, 717256), False, 'import pulumi\n'), ((717420, 717442), 'pulumi.get', 'pulumi.get', (['self', '"""id"""'], {}), "(self, 'id')\n", (717430, 717442), False, 'import pulumi\n'), ((717511, 717540), 'pulumi.set', 'pulumi.set', (['self', '"""id"""', 'value'], {}), "(self, 'id', value)\n", (717521, 717540), False, 'import pulumi\n'), ((717955, 718003), 'pulumi.set', 'pulumi.set', (['__self__', '"""fixed_scale"""', 'fixed_scale'], {}), "(__self__, 'fixed_scale', fixed_scale)\n", (717965, 718003), False, 'import pulumi\n'), ((718293, 718324), 'pulumi.get', 'pulumi.get', (['self', '"""fixed_scale"""'], {}), "(self, 'fixed_scale')\n", (718303, 718324), False, 'import pulumi\n'), ((718455, 718493), 'pulumi.set', 'pulumi.set', (['self', '"""fixed_scale"""', 'value'], {}), "(self, 'fixed_scale', value)\n", (718465, 718493), False, 'import pulumi\n'), ((718765, 718799), 'pulumi.set', 'pulumi.set', (['__self__', '"""size"""', 'size'], {}), "(__self__, 'size', size)\n", (718775, 718799), False, 'import pulumi\n'), ((718975, 718999), 'pulumi.get', 'pulumi.get', (['self', '"""size"""'], {}), "(self, 'size')\n", (718985, 718999), False, 'import pulumi\n'), ((719072, 719103), 'pulumi.set', 'pulumi.set', (['self', '"""size"""', 'value'], {}), "(self, 'size', value)\n", (719082, 719103), False, 'import pulumi\n'), ((719644, 719692), 'pulumi.set', 'pulumi.set', (['__self__', '"""group_count"""', 'group_count'], {}), "(__self__, 'group_count', group_count)\n", (719654, 719692), False, 'import pulumi\n'), ((719701, 719757), 'pulumi.set', 'pulumi.set', (['__self__', '"""storage_type_id"""', 'storage_type_id'], {}), "(__self__, 'storage_type_id', storage_type_id)\n", (719711, 719757), False, 'import pulumi\n'), ((719981, 720012), 'pulumi.get', 'pulumi.get', (['self', '"""group_count"""'], {}), "(self, 'group_count')\n", (719991, 720012), False, 'import pulumi\n'), ((720099, 720137), 'pulumi.set', 'pulumi.set', (['self', '"""group_count"""', 'value'], {}), "(self, 'group_count', value)\n", (720109, 720137), False, 'import pulumi\n'), ((720424, 720459), 'pulumi.get', 'pulumi.get', (['self', '"""storage_type_id"""'], {}), "(self, 'storage_type_id')\n", (720434, 720459), False, 'import pulumi\n'), ((720554, 720596), 'pulumi.set', 'pulumi.set', (['self', '"""storage_type_id"""', 'value'], {}), "(self, 'storage_type_id', value)\n", (720564, 720596), False, 'import pulumi\n'), ((721936, 721984), 'pulumi.set', 'pulumi.set', (['__self__', '"""healthcheck"""', 'healthcheck'], {}), "(__self__, 'healthcheck', healthcheck)\n", (721946, 721984), False, 'import pulumi\n'), ((721993, 722061), 'pulumi.set', 'pulumi.set', (['__self__', '"""load_balancing_config"""', 'load_balancing_config'], {}), "(__self__, 'load_balancing_config', load_balancing_config)\n", (722003, 722061), False, 'import pulumi\n'), ((722070, 722104), 'pulumi.set', 'pulumi.set', (['__self__', '"""name"""', 'name'], {}), "(__self__, 'name', name)\n", (722080, 722104), False, 'import pulumi\n'), ((722113, 722147), 'pulumi.set', 'pulumi.set', (['__self__', '"""port"""', 'port'], {}), "(__self__, 'port', port)\n", (722123, 722147), False, 'import pulumi\n'), ((722156, 722214), 'pulumi.set', 'pulumi.set', (['__self__', '"""target_group_ids"""', 'target_group_ids'], {}), "(__self__, 'target_group_ids', target_group_ids)\n", (722166, 722214), False, 'import pulumi\n'), ((722223, 722255), 'pulumi.set', 'pulumi.set', (['__self__', '"""tls"""', 'tls'], {}), "(__self__, 'tls', tls)\n", (722233, 722255), False, 'import pulumi\n'), ((722264, 722302), 'pulumi.set', 'pulumi.set', (['__self__', '"""weight"""', 'weight'], {}), "(__self__, 'weight', weight)\n", (722274, 722302), False, 'import pulumi\n'), ((722553, 722584), 'pulumi.get', 'pulumi.get', (['self', '"""healthcheck"""'], {}), "(self, 'healthcheck')\n", (722563, 722584), False, 'import pulumi\n'), ((722700, 722738), 'pulumi.set', 'pulumi.set', (['self', '"""healthcheck"""', 'value'], {}), "(self, 'healthcheck', value)\n", (722710, 722738), False, 'import pulumi\n'), ((723045, 723086), 'pulumi.get', 'pulumi.get', (['self', '"""load_balancing_config"""'], {}), "(self, 'load_balancing_config')\n", (723055, 723086), False, 'import pulumi\n'), ((723230, 723278), 'pulumi.set', 'pulumi.set', (['self', '"""load_balancing_config"""', 'value'], {}), "(self, 'load_balancing_config', value)\n", (723240, 723278), False, 'import pulumi\n'), ((723416, 723440), 'pulumi.get', 'pulumi.get', (['self', '"""name"""'], {}), "(self, 'name')\n", (723426, 723440), False, 'import pulumi\n'), ((723499, 723530), 'pulumi.set', 'pulumi.set', (['self', '"""name"""', 'value'], {}), "(self, 'name', value)\n", (723509, 723530), False, 'import pulumi\n'), ((723666, 723690), 'pulumi.get', 'pulumi.get', (['self', '"""port"""'], {}), "(self, 'port')\n", (723676, 723690), False, 'import pulumi\n'), ((723749, 723780), 'pulumi.set', 'pulumi.set', (['self', '"""port"""', 'value'], {}), "(self, 'port', value)\n", (723759, 723780), False, 'import pulumi\n'), ((723976, 724012), 'pulumi.get', 'pulumi.get', (['self', '"""target_group_ids"""'], {}), "(self, 'target_group_ids')\n", (723986, 724012), False, 'import pulumi\n'), ((724105, 724148), 'pulumi.set', 'pulumi.set', (['self', '"""target_group_ids"""', 'value'], {}), "(self, 'target_group_ids', value)\n", (724115, 724148), False, 'import pulumi\n'), ((724375, 724398), 'pulumi.get', 'pulumi.get', (['self', '"""tls"""'], {}), "(self, 'tls')\n", (724385, 724398), False, 'import pulumi\n'), ((724490, 724520), 'pulumi.set', 'pulumi.set', (['self', '"""tls"""', 'value'], {}), "(self, 'tls', value)\n", (724500, 724520), False, 'import pulumi\n'), ((724746, 724772), 'pulumi.get', 'pulumi.get', (['self', '"""weight"""'], {}), "(self, 'weight')\n", (724756, 724772), False, 'import pulumi\n'), ((724835, 724868), 'pulumi.set', 'pulumi.set', (['self', '"""weight"""', 'value'], {}), "(self, 'weight', value)\n", (724845, 724868), False, 'import pulumi\n'), ((727054, 727112), 'pulumi.set', 'pulumi.set', (['__self__', '"""grpc_healthcheck"""', 'grpc_healthcheck'], {}), "(__self__, 'grpc_healthcheck', grpc_healthcheck)\n", (727064, 727112), False, 'import pulumi\n'), ((727121, 727179), 'pulumi.set', 'pulumi.set', (['__self__', '"""healthcheck_port"""', 'healthcheck_port'], {}), "(__self__, 'healthcheck_port', healthcheck_port)\n", (727131, 727179), False, 'import pulumi\n'), ((727188, 727248), 'pulumi.set', 'pulumi.set', (['__self__', '"""healthy_threshold"""', 'healthy_threshold'], {}), "(__self__, 'healthy_threshold', healthy_threshold)\n", (727198, 727248), False, 'import pulumi\n'), ((727257, 727315), 'pulumi.set', 'pulumi.set', (['__self__', '"""http_healthcheck"""', 'http_healthcheck'], {}), "(__self__, 'http_healthcheck', http_healthcheck)\n", (727267, 727315), False, 'import pulumi\n'), ((727324, 727366), 'pulumi.set', 'pulumi.set', (['__self__', '"""interval"""', 'interval'], {}), "(__self__, 'interval', interval)\n", (727334, 727366), False, 'import pulumi\n'), ((727375, 727447), 'pulumi.set', 'pulumi.set', (['__self__', '"""interval_jitter_percent"""', 'interval_jitter_percent'], {}), "(__self__, 'interval_jitter_percent', interval_jitter_percent)\n", (727385, 727447), False, 'import pulumi\n'), ((727456, 727518), 'pulumi.set', 'pulumi.set', (['__self__', '"""stream_healthcheck"""', 'stream_healthcheck'], {}), "(__self__, 'stream_healthcheck', stream_healthcheck)\n", (727466, 727518), False, 'import pulumi\n'), ((727527, 727567), 'pulumi.set', 'pulumi.set', (['__self__', '"""timeout"""', 'timeout'], {}), "(__self__, 'timeout', timeout)\n", (727537, 727567), False, 'import pulumi\n'), ((727576, 727640), 'pulumi.set', 'pulumi.set', (['__self__', '"""unhealthy_threshold"""', 'unhealthy_threshold'], {}), "(__self__, 'unhealthy_threshold', unhealthy_threshold)\n", (727586, 727640), False, 'import pulumi\n'), ((727944, 727980), 'pulumi.get', 'pulumi.get', (['self', '"""grpc_healthcheck"""'], {}), "(self, 'grpc_healthcheck')\n", (727954, 727980), False, 'import pulumi\n'), ((728121, 728164), 'pulumi.set', 'pulumi.set', (['self', '"""grpc_healthcheck"""', 'value'], {}), "(self, 'grpc_healthcheck', value)\n", (728131, 728164), False, 'import pulumi\n'), ((728356, 728392), 'pulumi.get', 'pulumi.get', (['self', '"""healthcheck_port"""'], {}), "(self, 'healthcheck_port')\n", (728366, 728392), False, 'import pulumi\n'), ((728475, 728518), 'pulumi.set', 'pulumi.set', (['self', '"""healthcheck_port"""', 'value'], {}), "(self, 'healthcheck_port', value)\n", (728485, 728518), False, 'import pulumi\n'), ((728876, 728913), 'pulumi.get', 'pulumi.get', (['self', '"""healthy_threshold"""'], {}), "(self, 'healthy_threshold')\n", (728886, 728913), False, 'import pulumi\n'), ((728998, 729042), 'pulumi.set', 'pulumi.set', (['self', '"""healthy_threshold"""', 'value'], {}), "(self, 'healthy_threshold', value)\n", (729008, 729042), False, 'import pulumi\n'), ((729346, 729382), 'pulumi.get', 'pulumi.get', (['self', '"""http_healthcheck"""'], {}), "(self, 'http_healthcheck')\n", (729356, 729382), False, 'import pulumi\n'), ((729523, 729566), 'pulumi.set', 'pulumi.set', (['self', '"""http_healthcheck"""', 'value'], {}), "(self, 'http_healthcheck', value)\n", (729533, 729566), False, 'import pulumi\n'), ((729711, 729739), 'pulumi.get', 'pulumi.get', (['self', '"""interval"""'], {}), "(self, 'interval')\n", (729721, 729739), False, 'import pulumi\n'), ((729806, 729841), 'pulumi.set', 'pulumi.set', (['self', '"""interval"""', 'value'], {}), "(self, 'interval', value)\n", (729816, 729841), False, 'import pulumi\n'), ((730179, 730222), 'pulumi.get', 'pulumi.get', (['self', '"""interval_jitter_percent"""'], {}), "(self, 'interval_jitter_percent')\n", (730189, 730222), False, 'import pulumi\n'), ((730321, 730371), 'pulumi.set', 'pulumi.set', (['self', '"""interval_jitter_percent"""', 'value'], {}), "(self, 'interval_jitter_percent', value)\n", (730331, 730371), False, 'import pulumi\n'), ((730683, 730721), 'pulumi.get', 'pulumi.get', (['self', '"""stream_healthcheck"""'], {}), "(self, 'stream_healthcheck')\n", (730693, 730721), False, 'import pulumi\n'), ((730868, 730913), 'pulumi.set', 'pulumi.set', (['self', '"""stream_healthcheck"""', 'value'], {}), "(self, 'stream_healthcheck', value)\n", (730878, 730913), False, 'import pulumi\n'), ((731067, 731094), 'pulumi.get', 'pulumi.get', (['self', '"""timeout"""'], {}), "(self, 'timeout')\n", (731077, 731094), False, 'import pulumi\n'), ((731159, 731193), 'pulumi.set', 'pulumi.set', (['self', '"""timeout"""', 'value'], {}), "(self, 'timeout', value)\n", (731169, 731193), False, 'import pulumi\n'), ((731537, 731576), 'pulumi.get', 'pulumi.get', (['self', '"""unhealthy_threshold"""'], {}), "(self, 'unhealthy_threshold')\n", (731547, 731576), False, 'import pulumi\n'), ((731665, 731711), 'pulumi.set', 'pulumi.set', (['self', '"""unhealthy_threshold"""', 'value'], {}), "(self, 'unhealthy_threshold', value)\n", (731675, 731711), False, 'import pulumi\n'), ((732001, 732051), 'pulumi.set', 'pulumi.set', (['__self__', '"""service_name"""', 'service_name'], {}), "(__self__, 'service_name', service_name)\n", (732011, 732051), False, 'import pulumi\n'), ((732257, 732289), 'pulumi.get', 'pulumi.get', (['self', '"""service_name"""'], {}), "(self, 'service_name')\n", (732267, 732289), False, 'import pulumi\n'), ((732364, 732403), 'pulumi.set', 'pulumi.set', (['self', '"""service_name"""', 'value'], {}), "(self, 'service_name', value)\n", (732374, 732403), False, 'import pulumi\n'), ((732802, 732836), 'pulumi.set', 'pulumi.set', (['__self__', '"""host"""', 'host'], {}), "(__self__, 'host', host)\n", (732812, 732836), False, 'import pulumi\n'), ((732845, 732881), 'pulumi.set', 'pulumi.set', (['__self__', '"""http2"""', 'http2'], {}), "(__self__, 'http2', http2)\n", (732855, 732881), False, 'import pulumi\n'), ((732890, 732924), 'pulumi.set', 'pulumi.set', (['__self__', '"""path"""', 'path'], {}), "(__self__, 'path', path)\n", (732900, 732924), False, 'import pulumi\n'), ((733068, 733092), 'pulumi.get', 'pulumi.get', (['self', '"""host"""'], {}), "(self, 'host')\n", (733078, 733092), False, 'import pulumi\n'), ((733151, 733182), 'pulumi.set', 'pulumi.set', (['self', '"""host"""', 'value'], {}), "(self, 'host', value)\n", (733161, 733182), False, 'import pulumi\n'), ((733331, 733356), 'pulumi.get', 'pulumi.get', (['self', '"""http2"""'], {}), "(self, 'http2')\n", (733341, 733356), False, 'import pulumi\n'), ((733418, 733450), 'pulumi.set', 'pulumi.set', (['self', '"""http2"""', 'value'], {}), "(self, 'http2', value)\n", (733428, 733450), False, 'import pulumi\n'), ((733570, 733594), 'pulumi.get', 'pulumi.get', (['self', '"""path"""'], {}), "(self, 'path')\n", (733580, 733594), False, 'import pulumi\n'), ((733653, 733684), 'pulumi.set', 'pulumi.set', (['self', '"""path"""', 'value'], {}), "(self, 'path', value)\n", (733663, 733684), False, 'import pulumi\n'), ((734054, 734094), 'pulumi.set', 'pulumi.set', (['__self__', '"""receive"""', 'receive'], {}), "(__self__, 'receive', receive)\n", (734064, 734094), False, 'import pulumi\n'), ((734103, 734137), 'pulumi.set', 'pulumi.set', (['__self__', '"""send"""', 'send'], {}), "(__self__, 'send', send)\n", (734113, 734137), False, 'import pulumi\n'), ((734283, 734310), 'pulumi.get', 'pulumi.get', (['self', '"""receive"""'], {}), "(self, 'receive')\n", (734293, 734310), False, 'import pulumi\n'), ((734375, 734409), 'pulumi.set', 'pulumi.set', (['self', '"""receive"""', 'value'], {}), "(self, 'receive', value)\n", (734385, 734409), False, 'import pulumi\n'), ((734588, 734612), 'pulumi.get', 'pulumi.get', (['self', '"""send"""'], {}), "(self, 'send')\n", (734598, 734612), False, 'import pulumi\n'), ((734671, 734702), 'pulumi.set', 'pulumi.set', (['self', '"""send"""', 'value'], {}), "(self, 'send', value)\n", (734681, 734702), False, 'import pulumi\n'), ((735593, 735683), 'pulumi.set', 'pulumi.set', (['__self__', '"""locality_aware_routing_percent"""', 'locality_aware_routing_percent'], {}), "(__self__, 'locality_aware_routing_percent',\n locality_aware_routing_percent)\n", (735603, 735683), False, 'import pulumi\n'), ((735688, 735744), 'pulumi.set', 'pulumi.set', (['__self__', '"""panic_threshold"""', 'panic_threshold'], {}), "(__self__, 'panic_threshold', panic_threshold)\n", (735698, 735744), False, 'import pulumi\n'), ((735753, 735809), 'pulumi.set', 'pulumi.set', (['__self__', '"""strict_locality"""', 'strict_locality'], {}), "(__self__, 'strict_locality', strict_locality)\n", (735763, 735809), False, 'import pulumi\n'), ((736095, 736145), 'pulumi.get', 'pulumi.get', (['self', '"""locality_aware_routing_percent"""'], {}), "(self, 'locality_aware_routing_percent')\n", (736105, 736145), False, 'import pulumi\n'), ((736256, 736313), 'pulumi.set', 'pulumi.set', (['self', '"""locality_aware_routing_percent"""', 'value'], {}), "(self, 'locality_aware_routing_percent', value)\n", (736266, 736313), False, 'import pulumi\n'), ((736714, 736749), 'pulumi.get', 'pulumi.get', (['self', '"""panic_threshold"""'], {}), "(self, 'panic_threshold')\n", (736724, 736749), False, 'import pulumi\n'), ((736830, 736872), 'pulumi.set', 'pulumi.set', (['self', '"""panic_threshold"""', 'value'], {}), "(self, 'panic_threshold', value)\n", (736840, 736872), False, 'import pulumi\n'), ((737132, 737167), 'pulumi.get', 'pulumi.get', (['self', '"""strict_locality"""'], {}), "(self, 'strict_locality')\n", (737142, 737167), False, 'import pulumi\n'), ((737249, 737291), 'pulumi.set', 'pulumi.set', (['self', '"""strict_locality"""', 'value'], {}), "(self, 'strict_locality', value)\n", (737259, 737291), False, 'import pulumi\n'), ((737864, 737896), 'pulumi.set', 'pulumi.set', (['__self__', '"""sni"""', 'sni'], {}), "(__self__, 'sni', sni)\n", (737874, 737896), False, 'import pulumi\n'), ((737905, 737967), 'pulumi.set', 'pulumi.set', (['__self__', '"""validation_context"""', 'validation_context'], {}), "(__self__, 'validation_context', validation_context)\n", (737915, 737967), False, 'import pulumi\n'), ((738360, 738383), 'pulumi.get', 'pulumi.get', (['self', '"""sni"""'], {}), "(self, 'sni')\n", (738370, 738383), False, 'import pulumi\n'), ((738440, 738470), 'pulumi.set', 'pulumi.set', (['self', '"""sni"""', 'value'], {}), "(self, 'sni', value)\n", (738450, 738470), False, 'import pulumi\n'), ((738639, 738677), 'pulumi.get', 'pulumi.get', (['self', '"""validation_context"""'], {}), "(self, 'validation_context')\n", (738649, 738677), False, 'import pulumi\n'), ((738816, 738861), 'pulumi.set', 'pulumi.set', (['self', '"""validation_context"""', 'value'], {}), "(self, 'validation_context', value)\n", (738826, 738861), False, 'import pulumi\n'), ((739060, 739118), 'pulumi.set', 'pulumi.set', (['__self__', '"""trusted_ca_bytes"""', 'trusted_ca_bytes'], {}), "(__self__, 'trusted_ca_bytes', trusted_ca_bytes)\n", (739070, 739118), False, 'import pulumi\n'), ((739127, 739179), 'pulumi.set', 'pulumi.set', (['__self__', '"""trusted_ca_id"""', 'trusted_ca_id'], {}), "(__self__, 'trusted_ca_id', trusted_ca_id)\n", (739137, 739179), False, 'import pulumi\n'), ((739291, 739327), 'pulumi.get', 'pulumi.get', (['self', '"""trusted_ca_bytes"""'], {}), "(self, 'trusted_ca_bytes')\n", (739301, 739327), False, 'import pulumi\n'), ((739410, 739453), 'pulumi.set', 'pulumi.set', (['self', '"""trusted_ca_bytes"""', 'value'], {}), "(self, 'trusted_ca_bytes', value)\n", (739420, 739453), False, 'import pulumi\n'), ((739559, 739592), 'pulumi.get', 'pulumi.get', (['self', '"""trusted_ca_id"""'], {}), "(self, 'trusted_ca_id')\n", (739569, 739592), False, 'import pulumi\n'), ((739669, 739709), 'pulumi.set', 'pulumi.set', (['self', '"""trusted_ca_id"""', 'value'], {}), "(self, 'trusted_ca_id', value)\n", (739679, 739709), False, 'import pulumi\n'), ((741144, 741192), 'pulumi.set', 'pulumi.set', (['__self__', '"""healthcheck"""', 'healthcheck'], {}), "(__self__, 'healthcheck', healthcheck)\n", (741154, 741192), False, 'import pulumi\n'), ((741201, 741237), 'pulumi.set', 'pulumi.set', (['__self__', '"""http2"""', 'http2'], {}), "(__self__, 'http2', http2)\n", (741211, 741237), False, 'import pulumi\n'), ((741246, 741314), 'pulumi.set', 'pulumi.set', (['__self__', '"""load_balancing_config"""', 'load_balancing_config'], {}), "(__self__, 'load_balancing_config', load_balancing_config)\n", (741256, 741314), False, 'import pulumi\n'), ((741323, 741357), 'pulumi.set', 'pulumi.set', (['__self__', '"""name"""', 'name'], {}), "(__self__, 'name', name)\n", (741333, 741357), False, 'import pulumi\n'), ((741366, 741400), 'pulumi.set', 'pulumi.set', (['__self__', '"""port"""', 'port'], {}), "(__self__, 'port', port)\n", (741376, 741400), False, 'import pulumi\n'), ((741409, 741467), 'pulumi.set', 'pulumi.set', (['__self__', '"""target_group_ids"""', 'target_group_ids'], {}), "(__self__, 'target_group_ids', target_group_ids)\n", (741419, 741467), False, 'import pulumi\n'), ((741476, 741508), 'pulumi.set', 'pulumi.set', (['__self__', '"""tls"""', 'tls'], {}), "(__self__, 'tls', tls)\n", (741486, 741508), False, 'import pulumi\n'), ((741517, 741555), 'pulumi.set', 'pulumi.set', (['__self__', '"""weight"""', 'weight'], {}), "(__self__, 'weight', weight)\n", (741527, 741555), False, 'import pulumi\n'), ((741806, 741837), 'pulumi.get', 'pulumi.get', (['self', '"""healthcheck"""'], {}), "(self, 'healthcheck')\n", (741816, 741837), False, 'import pulumi\n'), ((741953, 741991), 'pulumi.set', 'pulumi.set', (['self', '"""healthcheck"""', 'value'], {}), "(self, 'healthcheck', value)\n", (741963, 741991), False, 'import pulumi\n'), ((742140, 742165), 'pulumi.get', 'pulumi.get', (['self', '"""http2"""'], {}), "(self, 'http2')\n", (742150, 742165), False, 'import pulumi\n'), ((742227, 742259), 'pulumi.set', 'pulumi.set', (['self', '"""http2"""', 'value'], {}), "(self, 'http2', value)\n", (742237, 742259), False, 'import pulumi\n'), ((742566, 742607), 'pulumi.get', 'pulumi.get', (['self', '"""load_balancing_config"""'], {}), "(self, 'load_balancing_config')\n", (742576, 742607), False, 'import pulumi\n'), ((742751, 742799), 'pulumi.set', 'pulumi.set', (['self', '"""load_balancing_config"""', 'value'], {}), "(self, 'load_balancing_config', value)\n", (742761, 742799), False, 'import pulumi\n'), ((742937, 742961), 'pulumi.get', 'pulumi.get', (['self', '"""name"""'], {}), "(self, 'name')\n", (742947, 742961), False, 'import pulumi\n'), ((743020, 743051), 'pulumi.set', 'pulumi.set', (['self', '"""name"""', 'value'], {}), "(self, 'name', value)\n", (743030, 743051), False, 'import pulumi\n'), ((743187, 743211), 'pulumi.get', 'pulumi.get', (['self', '"""port"""'], {}), "(self, 'port')\n", (743197, 743211), False, 'import pulumi\n'), ((743270, 743301), 'pulumi.set', 'pulumi.set', (['self', '"""port"""', 'value'], {}), "(self, 'port', value)\n", (743280, 743301), False, 'import pulumi\n'), ((743497, 743533), 'pulumi.get', 'pulumi.get', (['self', '"""target_group_ids"""'], {}), "(self, 'target_group_ids')\n", (743507, 743533), False, 'import pulumi\n'), ((743626, 743669), 'pulumi.set', 'pulumi.set', (['self', '"""target_group_ids"""', 'value'], {}), "(self, 'target_group_ids', value)\n", (743636, 743669), False, 'import pulumi\n'), ((743896, 743919), 'pulumi.get', 'pulumi.get', (['self', '"""tls"""'], {}), "(self, 'tls')\n", (743906, 743919), False, 'import pulumi\n'), ((744011, 744041), 'pulumi.set', 'pulumi.set', (['self', '"""tls"""', 'value'], {}), "(self, 'tls', value)\n", (744021, 744041), False, 'import pulumi\n'), ((744267, 744293), 'pulumi.get', 'pulumi.get', (['self', '"""weight"""'], {}), "(self, 'weight')\n", (744277, 744293), False, 'import pulumi\n'), ((744356, 744389), 'pulumi.set', 'pulumi.set', (['self', '"""weight"""', 'value'], {}), "(self, 'weight', value)\n", (744366, 744389), False, 'import pulumi\n'), ((746575, 746633), 'pulumi.set', 'pulumi.set', (['__self__', '"""grpc_healthcheck"""', 'grpc_healthcheck'], {}), "(__self__, 'grpc_healthcheck', grpc_healthcheck)\n", (746585, 746633), False, 'import pulumi\n'), ((746642, 746700), 'pulumi.set', 'pulumi.set', (['__self__', '"""healthcheck_port"""', 'healthcheck_port'], {}), "(__self__, 'healthcheck_port', healthcheck_port)\n", (746652, 746700), False, 'import pulumi\n'), ((746709, 746769), 'pulumi.set', 'pulumi.set', (['__self__', '"""healthy_threshold"""', 'healthy_threshold'], {}), "(__self__, 'healthy_threshold', healthy_threshold)\n", (746719, 746769), False, 'import pulumi\n'), ((746778, 746836), 'pulumi.set', 'pulumi.set', (['__self__', '"""http_healthcheck"""', 'http_healthcheck'], {}), "(__self__, 'http_healthcheck', http_healthcheck)\n", (746788, 746836), False, 'import pulumi\n'), ((746845, 746887), 'pulumi.set', 'pulumi.set', (['__self__', '"""interval"""', 'interval'], {}), "(__self__, 'interval', interval)\n", (746855, 746887), False, 'import pulumi\n'), ((746896, 746968), 'pulumi.set', 'pulumi.set', (['__self__', '"""interval_jitter_percent"""', 'interval_jitter_percent'], {}), "(__self__, 'interval_jitter_percent', interval_jitter_percent)\n", (746906, 746968), False, 'import pulumi\n'), ((746977, 747039), 'pulumi.set', 'pulumi.set', (['__self__', '"""stream_healthcheck"""', 'stream_healthcheck'], {}), "(__self__, 'stream_healthcheck', stream_healthcheck)\n", (746987, 747039), False, 'import pulumi\n'), ((747048, 747088), 'pulumi.set', 'pulumi.set', (['__self__', '"""timeout"""', 'timeout'], {}), "(__self__, 'timeout', timeout)\n", (747058, 747088), False, 'import pulumi\n'), ((747097, 747161), 'pulumi.set', 'pulumi.set', (['__self__', '"""unhealthy_threshold"""', 'unhealthy_threshold'], {}), "(__self__, 'unhealthy_threshold', unhealthy_threshold)\n", (747107, 747161), False, 'import pulumi\n'), ((747465, 747501), 'pulumi.get', 'pulumi.get', (['self', '"""grpc_healthcheck"""'], {}), "(self, 'grpc_healthcheck')\n", (747475, 747501), False, 'import pulumi\n'), ((747642, 747685), 'pulumi.set', 'pulumi.set', (['self', '"""grpc_healthcheck"""', 'value'], {}), "(self, 'grpc_healthcheck', value)\n", (747652, 747685), False, 'import pulumi\n'), ((747877, 747913), 'pulumi.get', 'pulumi.get', (['self', '"""healthcheck_port"""'], {}), "(self, 'healthcheck_port')\n", (747887, 747913), False, 'import pulumi\n'), ((747996, 748039), 'pulumi.set', 'pulumi.set', (['self', '"""healthcheck_port"""', 'value'], {}), "(self, 'healthcheck_port', value)\n", (748006, 748039), False, 'import pulumi\n'), ((748397, 748434), 'pulumi.get', 'pulumi.get', (['self', '"""healthy_threshold"""'], {}), "(self, 'healthy_threshold')\n", (748407, 748434), False, 'import pulumi\n'), ((748519, 748563), 'pulumi.set', 'pulumi.set', (['self', '"""healthy_threshold"""', 'value'], {}), "(self, 'healthy_threshold', value)\n", (748529, 748563), False, 'import pulumi\n'), ((748867, 748903), 'pulumi.get', 'pulumi.get', (['self', '"""http_healthcheck"""'], {}), "(self, 'http_healthcheck')\n", (748877, 748903), False, 'import pulumi\n'), ((749044, 749087), 'pulumi.set', 'pulumi.set', (['self', '"""http_healthcheck"""', 'value'], {}), "(self, 'http_healthcheck', value)\n", (749054, 749087), False, 'import pulumi\n'), ((749232, 749260), 'pulumi.get', 'pulumi.get', (['self', '"""interval"""'], {}), "(self, 'interval')\n", (749242, 749260), False, 'import pulumi\n'), ((749327, 749362), 'pulumi.set', 'pulumi.set', (['self', '"""interval"""', 'value'], {}), "(self, 'interval', value)\n", (749337, 749362), False, 'import pulumi\n'), ((749700, 749743), 'pulumi.get', 'pulumi.get', (['self', '"""interval_jitter_percent"""'], {}), "(self, 'interval_jitter_percent')\n", (749710, 749743), False, 'import pulumi\n'), ((749842, 749892), 'pulumi.set', 'pulumi.set', (['self', '"""interval_jitter_percent"""', 'value'], {}), "(self, 'interval_jitter_percent', value)\n", (749852, 749892), False, 'import pulumi\n'), ((750204, 750242), 'pulumi.get', 'pulumi.get', (['self', '"""stream_healthcheck"""'], {}), "(self, 'stream_healthcheck')\n", (750214, 750242), False, 'import pulumi\n'), ((750389, 750434), 'pulumi.set', 'pulumi.set', (['self', '"""stream_healthcheck"""', 'value'], {}), "(self, 'stream_healthcheck', value)\n", (750399, 750434), False, 'import pulumi\n'), ((750588, 750615), 'pulumi.get', 'pulumi.get', (['self', '"""timeout"""'], {}), "(self, 'timeout')\n", (750598, 750615), False, 'import pulumi\n'), ((750680, 750714), 'pulumi.set', 'pulumi.set', (['self', '"""timeout"""', 'value'], {}), "(self, 'timeout', value)\n", (750690, 750714), False, 'import pulumi\n'), ((751058, 751097), 'pulumi.get', 'pulumi.get', (['self', '"""unhealthy_threshold"""'], {}), "(self, 'unhealthy_threshold')\n", (751068, 751097), False, 'import pulumi\n'), ((751186, 751232), 'pulumi.set', 'pulumi.set', (['self', '"""unhealthy_threshold"""', 'value'], {}), "(self, 'unhealthy_threshold', value)\n", (751196, 751232), False, 'import pulumi\n'), ((751522, 751572), 'pulumi.set', 'pulumi.set', (['__self__', '"""service_name"""', 'service_name'], {}), "(__self__, 'service_name', service_name)\n", (751532, 751572), False, 'import pulumi\n'), ((751778, 751810), 'pulumi.get', 'pulumi.get', (['self', '"""service_name"""'], {}), "(self, 'service_name')\n", (751788, 751810), False, 'import pulumi\n'), ((751885, 751924), 'pulumi.set', 'pulumi.set', (['self', '"""service_name"""', 'value'], {}), "(self, 'service_name', value)\n", (751895, 751924), False, 'import pulumi\n'), ((752323, 752357), 'pulumi.set', 'pulumi.set', (['__self__', '"""host"""', 'host'], {}), "(__self__, 'host', host)\n", (752333, 752357), False, 'import pulumi\n'), ((752366, 752402), 'pulumi.set', 'pulumi.set', (['__self__', '"""http2"""', 'http2'], {}), "(__self__, 'http2', http2)\n", (752376, 752402), False, 'import pulumi\n'), ((752411, 752445), 'pulumi.set', 'pulumi.set', (['__self__', '"""path"""', 'path'], {}), "(__self__, 'path', path)\n", (752421, 752445), False, 'import pulumi\n'), ((752589, 752613), 'pulumi.get', 'pulumi.get', (['self', '"""host"""'], {}), "(self, 'host')\n", (752599, 752613), False, 'import pulumi\n'), ((752672, 752703), 'pulumi.set', 'pulumi.set', (['self', '"""host"""', 'value'], {}), "(self, 'host', value)\n", (752682, 752703), False, 'import pulumi\n'), ((752852, 752877), 'pulumi.get', 'pulumi.get', (['self', '"""http2"""'], {}), "(self, 'http2')\n", (752862, 752877), False, 'import pulumi\n'), ((752939, 752971), 'pulumi.set', 'pulumi.set', (['self', '"""http2"""', 'value'], {}), "(self, 'http2', value)\n", (752949, 752971), False, 'import pulumi\n'), ((753091, 753115), 'pulumi.get', 'pulumi.get', (['self', '"""path"""'], {}), "(self, 'path')\n", (753101, 753115), False, 'import pulumi\n'), ((753174, 753205), 'pulumi.set', 'pulumi.set', (['self', '"""path"""', 'value'], {}), "(self, 'path', value)\n", (753184, 753205), False, 'import pulumi\n'), ((753575, 753615), 'pulumi.set', 'pulumi.set', (['__self__', '"""receive"""', 'receive'], {}), "(__self__, 'receive', receive)\n", (753585, 753615), False, 'import pulumi\n'), ((753624, 753658), 'pulumi.set', 'pulumi.set', (['__self__', '"""send"""', 'send'], {}), "(__self__, 'send', send)\n", (753634, 753658), False, 'import pulumi\n'), ((753804, 753831), 'pulumi.get', 'pulumi.get', (['self', '"""receive"""'], {}), "(self, 'receive')\n", (753814, 753831), False, 'import pulumi\n'), ((753896, 753930), 'pulumi.set', 'pulumi.set', (['self', '"""receive"""', 'value'], {}), "(self, 'receive', value)\n", (753906, 753930), False, 'import pulumi\n'), ((754109, 754133), 'pulumi.get', 'pulumi.get', (['self', '"""send"""'], {}), "(self, 'send')\n", (754119, 754133), False, 'import pulumi\n'), ((754192, 754223), 'pulumi.set', 'pulumi.set', (['self', '"""send"""', 'value'], {}), "(self, 'send', value)\n", (754202, 754223), False, 'import pulumi\n'), ((755114, 755204), 'pulumi.set', 'pulumi.set', (['__self__', '"""locality_aware_routing_percent"""', 'locality_aware_routing_percent'], {}), "(__self__, 'locality_aware_routing_percent',\n locality_aware_routing_percent)\n", (755124, 755204), False, 'import pulumi\n'), ((755209, 755265), 'pulumi.set', 'pulumi.set', (['__self__', '"""panic_threshold"""', 'panic_threshold'], {}), "(__self__, 'panic_threshold', panic_threshold)\n", (755219, 755265), False, 'import pulumi\n'), ((755274, 755330), 'pulumi.set', 'pulumi.set', (['__self__', '"""strict_locality"""', 'strict_locality'], {}), "(__self__, 'strict_locality', strict_locality)\n", (755284, 755330), False, 'import pulumi\n'), ((755616, 755666), 'pulumi.get', 'pulumi.get', (['self', '"""locality_aware_routing_percent"""'], {}), "(self, 'locality_aware_routing_percent')\n", (755626, 755666), False, 'import pulumi\n'), ((755777, 755834), 'pulumi.set', 'pulumi.set', (['self', '"""locality_aware_routing_percent"""', 'value'], {}), "(self, 'locality_aware_routing_percent', value)\n", (755787, 755834), False, 'import pulumi\n'), ((756235, 756270), 'pulumi.get', 'pulumi.get', (['self', '"""panic_threshold"""'], {}), "(self, 'panic_threshold')\n", (756245, 756270), False, 'import pulumi\n'), ((756351, 756393), 'pulumi.set', 'pulumi.set', (['self', '"""panic_threshold"""', 'value'], {}), "(self, 'panic_threshold', value)\n", (756361, 756393), False, 'import pulumi\n'), ((756653, 756688), 'pulumi.get', 'pulumi.get', (['self', '"""strict_locality"""'], {}), "(self, 'strict_locality')\n", (756663, 756688), False, 'import pulumi\n'), ((756770, 756812), 'pulumi.set', 'pulumi.set', (['self', '"""strict_locality"""', 'value'], {}), "(self, 'strict_locality', value)\n", (756780, 756812), False, 'import pulumi\n'), ((757385, 757417), 'pulumi.set', 'pulumi.set', (['__self__', '"""sni"""', 'sni'], {}), "(__self__, 'sni', sni)\n", (757395, 757417), False, 'import pulumi\n'), ((757426, 757488), 'pulumi.set', 'pulumi.set', (['__self__', '"""validation_context"""', 'validation_context'], {}), "(__self__, 'validation_context', validation_context)\n", (757436, 757488), False, 'import pulumi\n'), ((757881, 757904), 'pulumi.get', 'pulumi.get', (['self', '"""sni"""'], {}), "(self, 'sni')\n", (757891, 757904), False, 'import pulumi\n'), ((757961, 757991), 'pulumi.set', 'pulumi.set', (['self', '"""sni"""', 'value'], {}), "(self, 'sni', value)\n", (757971, 757991), False, 'import pulumi\n'), ((758160, 758198), 'pulumi.get', 'pulumi.get', (['self', '"""validation_context"""'], {}), "(self, 'validation_context')\n", (758170, 758198), False, 'import pulumi\n'), ((758337, 758382), 'pulumi.set', 'pulumi.set', (['self', '"""validation_context"""', 'value'], {}), "(self, 'validation_context', value)\n", (758347, 758382), False, 'import pulumi\n'), ((758581, 758639), 'pulumi.set', 'pulumi.set', (['__self__', '"""trusted_ca_bytes"""', 'trusted_ca_bytes'], {}), "(__self__, 'trusted_ca_bytes', trusted_ca_bytes)\n", (758591, 758639), False, 'import pulumi\n'), ((758648, 758700), 'pulumi.set', 'pulumi.set', (['__self__', '"""trusted_ca_id"""', 'trusted_ca_id'], {}), "(__self__, 'trusted_ca_id', trusted_ca_id)\n", (758658, 758700), False, 'import pulumi\n'), ((758812, 758848), 'pulumi.get', 'pulumi.get', (['self', '"""trusted_ca_bytes"""'], {}), "(self, 'trusted_ca_bytes')\n", (758822, 758848), False, 'import pulumi\n'), ((758931, 758974), 'pulumi.set', 'pulumi.set', (['self', '"""trusted_ca_bytes"""', 'value'], {}), "(self, 'trusted_ca_bytes', value)\n", (758941, 758974), False, 'import pulumi\n'), ((759080, 759113), 'pulumi.get', 'pulumi.get', (['self', '"""trusted_ca_id"""'], {}), "(self, 'trusted_ca_id')\n", (759090, 759113), False, 'import pulumi\n'), ((759190, 759230), 'pulumi.set', 'pulumi.set', (['self', '"""trusted_ca_id"""', 'value'], {}), "(self, 'trusted_ca_id', value)\n", (759200, 759230), False, 'import pulumi\n'), ((759383, 759455), 'pulumi.set', 'pulumi.set', (['__self__', '"""disk_placement_group_id"""', 'disk_placement_group_id'], {}), "(__self__, 'disk_placement_group_id', disk_placement_group_id)\n", (759393, 759455), False, 'import pulumi\n'), ((759580, 759623), 'pulumi.get', 'pulumi.get', (['self', '"""disk_placement_group_id"""'], {}), "(self, 'disk_placement_group_id')\n", (759590, 759623), False, 'import pulumi\n'), ((759720, 759770), 'pulumi.set', 'pulumi.set', (['self', '"""disk_placement_group_id"""', 'value'], {}), "(self, 'disk_placement_group_id', value)\n", (759730, 759770), False, 'import pulumi\n'), ((759918, 759980), 'pulumi.set', 'pulumi.set', (['__self__', '"""placement_group_id"""', 'placement_group_id'], {}), "(__self__, 'placement_group_id', placement_group_id)\n", (759928, 759980), False, 'import pulumi\n'), ((760096, 760134), 'pulumi.get', 'pulumi.get', (['self', '"""placement_group_id"""'], {}), "(self, 'placement_group_id')\n", (760106, 760134), False, 'import pulumi\n'), ((760221, 760266), 'pulumi.set', 'pulumi.set', (['self', '"""placement_group_id"""', 'value'], {}), "(self, 'placement_group_id', value)\n", (760231, 760266), False, 'import pulumi\n'), ((760842, 760873), 'pulumi.get', 'pulumi.get', (['self', '"""preemptible"""'], {}), "(self, 'preemptible')\n", (760852, 760873), False, 'import pulumi\n'), ((760957, 760995), 'pulumi.set', 'pulumi.set', (['self', '"""preemptible"""', 'value'], {}), "(self, 'preemptible', value)\n", (760967, 760995), False, 'import pulumi\n'), ((761246, 761278), 'pulumi.set', 'pulumi.set', (['__self__', '"""tag"""', 'tag'], {}), "(__self__, 'tag', tag)\n", (761256, 761278), False, 'import pulumi\n'), ((761599, 761622), 'pulumi.get', 'pulumi.get', (['self', '"""tag"""'], {}), "(self, 'tag')\n", (761609, 761622), False, 'import pulumi\n'), ((761679, 761709), 'pulumi.set', 'pulumi.set', (['self', '"""tag"""', 'value'], {}), "(self, 'tag', value)\n", (761689, 761709), False, 'import pulumi\n'), ((761839, 761879), 'pulumi.get', 'pulumi.get', (['self', '"""zone_instances_limit"""'], {}), "(self, 'zone_instances_limit')\n", (761849, 761879), False, 'import pulumi\n'), ((761980, 762027), 'pulumi.set', 'pulumi.set', (['self', '"""zone_instances_limit"""', 'value'], {}), "(self, 'zone_instances_limit', value)\n", (761990, 762027), False, 'import pulumi\n'), ((762155, 762194), 'pulumi.get', 'pulumi.get', (['self', '"""zone_requests_limit"""'], {}), "(self, 'zone_requests_limit')\n", (762165, 762194), False, 'import pulumi\n'), ((762293, 762339), 'pulumi.set', 'pulumi.set', (['self', '"""zone_requests_limit"""', 'value'], {}), "(self, 'zone_requests_limit', value)\n", (762303, 762339), False, 'import pulumi\n'), ((763051, 763091), 'pulumi.set', 'pulumi.set', (['__self__', '"""members"""', 'members'], {}), "(__self__, 'members', members)\n", (763061, 763091), False, 'import pulumi\n'), ((763100, 763134), 'pulumi.set', 'pulumi.set', (['__self__', '"""role"""', 'role'], {}), "(__self__, 'role', role)\n", (763110, 763134), False, 'import pulumi\n'), ((763563, 763590), 'pulumi.get', 'pulumi.get', (['self', '"""members"""'], {}), "(self, 'members')\n", (763573, 763590), False, 'import pulumi\n'), ((763665, 763699), 'pulumi.set', 'pulumi.set', (['self', '"""members"""', 'value'], {}), "(self, 'members', value)\n", (763675, 763699), False, 'import pulumi\n'), ((763937, 763961), 'pulumi.get', 'pulumi.get', (['self', '"""role"""'], {}), "(self, 'role')\n", (763947, 763961), False, 'import pulumi\n'), ((764020, 764051), 'pulumi.set', 'pulumi.set', (['self', '"""role"""', 'value'], {}), "(self, 'role', value)\n", (764030, 764051), False, 'import pulumi\n'), ((764354, 764394), 'pulumi.set', 'pulumi.set', (['__self__', '"""enabled"""', 'enabled'], {}), "(__self__, 'enabled', enabled)\n", (764364, 764394), False, 'import pulumi\n'), ((764617, 764644), 'pulumi.get', 'pulumi.get', (['self', '"""enabled"""'], {}), "(self, 'enabled')\n", (764627, 764644), False, 'import pulumi\n'), ((764710, 764744), 'pulumi.set', 'pulumi.set', (['self', '"""enabled"""', 'value'], {}), "(self, 'enabled', value)\n", (764720, 764744), False, 'import pulumi\n'), ((766194, 766230), 'pulumi.set', 'pulumi.set', (['__self__', '"""kafka"""', 'kafka'], {}), "(__self__, 'kafka', kafka)\n", (766204, 766230), False, 'import pulumi\n'), ((766239, 766279), 'pulumi.set', 'pulumi.set', (['__self__', '"""version"""', 'version'], {}), "(__self__, 'version', version)\n", (766249, 766279), False, 'import pulumi\n'), ((766288, 766324), 'pulumi.set', 'pulumi.set', (['__self__', '"""zones"""', 'zones'], {}), "(__self__, 'zones', zones)\n", (766298, 766324), False, 'import pulumi\n'), ((766333, 766377), 'pulumi.set', 'pulumi.set', (['__self__', '"""zookeeper"""', 'zookeeper'], {}), "(__self__, 'zookeeper', zookeeper)\n", (766343, 766377), False, 'import pulumi\n'), ((767040, 767065), 'pulumi.get', 'pulumi.get', (['self', '"""kafka"""'], {}), "(self, 'kafka')\n", (767050, 767065), False, 'import pulumi\n'), ((767158, 767190), 'pulumi.set', 'pulumi.set', (['self', '"""kafka"""', 'value'], {}), "(self, 'kafka', value)\n", (767168, 767190), False, 'import pulumi\n'), ((767351, 767378), 'pulumi.get', 'pulumi.get', (['self', '"""version"""'], {}), "(self, 'version')\n", (767361, 767378), False, 'import pulumi\n'), ((767443, 767477), 'pulumi.set', 'pulumi.set', (['self', '"""version"""', 'value'], {}), "(self, 'version', value)\n", (767453, 767477), False, 'import pulumi\n'), ((767636, 767661), 'pulumi.get', 'pulumi.get', (['self', '"""zones"""'], {}), "(self, 'zones')\n", (767646, 767661), False, 'import pulumi\n'), ((767732, 767764), 'pulumi.set', 'pulumi.set', (['self', '"""zones"""', 'value'], {}), "(self, 'zones', value)\n", (767742, 767764), False, 'import pulumi\n'), ((768003, 768032), 'pulumi.get', 'pulumi.get', (['self', '"""zookeeper"""'], {}), "(self, 'zookeeper')\n", (768013, 768032), False, 'import pulumi\n'), ((768137, 768173), 'pulumi.set', 'pulumi.set', (['self', '"""zookeeper"""', 'value'], {}), "(self, 'zookeeper', value)\n", (768147, 768173), False, 'import pulumi\n'), ((768403, 768439), 'pulumi.get', 'pulumi.get', (['self', '"""assign_public_ip"""'], {}), "(self, 'assign_public_ip')\n", (768413, 768439), False, 'import pulumi\n'), ((768533, 768576), 'pulumi.set', 'pulumi.set', (['self', '"""assign_public_ip"""', 'value'], {}), "(self, 'assign_public_ip', value)\n", (768543, 768576), False, 'import pulumi\n'), ((768776, 768809), 'pulumi.get', 'pulumi.get', (['self', '"""brokers_count"""'], {}), "(self, 'brokers_count')\n", (768786, 768809), False, 'import pulumi\n'), ((768896, 768936), 'pulumi.set', 'pulumi.set', (['self', '"""brokers_count"""', 'value'], {}), "(self, 'brokers_count', value)\n", (768906, 768936), False, 'import pulumi\n'), ((769178, 769213), 'pulumi.get', 'pulumi.get', (['self', '"""schema_registry"""'], {}), "(self, 'schema_registry')\n", (769188, 769213), False, 'import pulumi\n'), ((769305, 769347), 'pulumi.set', 'pulumi.set', (['self', '"""schema_registry"""', 'value'], {}), "(self, 'schema_registry', value)\n", (769315, 769347), False, 'import pulumi\n'), ((769594, 769630), 'pulumi.get', 'pulumi.get', (['self', '"""unmanaged_topics"""'], {}), "(self, 'unmanaged_topics')\n", (769604, 769630), False, 'import pulumi\n'), ((769724, 769767), 'pulumi.set', 'pulumi.set', (['self', '"""unmanaged_topics"""', 'value'], {}), "(self, 'unmanaged_topics', value)\n", (769734, 769767), False, 'import pulumi\n'), ((770405, 770449), 'pulumi.set', 'pulumi.set', (['__self__', '"""resources"""', 'resources'], {}), "(__self__, 'resources', resources)\n", (770415, 770449), False, 'import pulumi\n'), ((770808, 770837), 'pulumi.get', 'pulumi.get', (['self', '"""resources"""'], {}), "(self, 'resources')\n", (770818, 770837), False, 'import pulumi\n'), ((770947, 770983), 'pulumi.set', 'pulumi.set', (['self', '"""resources"""', 'value'], {}), "(self, 'resources', value)\n", (770957, 770983), False, 'import pulumi\n'), ((771264, 771296), 'pulumi.get', 'pulumi.get', (['self', '"""kafka_config"""'], {}), "(self, 'kafka_config')\n", (771274, 771296), False, 'import pulumi\n'), ((771424, 771463), 'pulumi.set', 'pulumi.set', (['self', '"""kafka_config"""', 'value'], {}), "(self, 'kafka_config', value)\n", (771434, 771463), False, 'import pulumi\n'), ((774556, 774601), 'pulumi.get', 'pulumi.get', (['self', '"""auto_create_topics_enable"""'], {}), "(self, 'auto_create_topics_enable')\n", (774566, 774601), False, 'import pulumi\n'), ((774713, 774765), 'pulumi.set', 'pulumi.set', (['self', '"""auto_create_topics_enable"""', 'value'], {}), "(self, 'auto_create_topics_enable', value)\n", (774723, 774765), False, 'import pulumi\n'), ((774888, 774924), 'pulumi.get', 'pulumi.get', (['self', '"""compression_type"""'], {}), "(self, 'compression_type')\n", (774898, 774924), False, 'import pulumi\n'), ((775017, 775060), 'pulumi.set', 'pulumi.set', (['self', '"""compression_type"""', 'value'], {}), "(self, 'compression_type', value)\n", (775027, 775060), False, 'import pulumi\n'), ((775202, 775248), 'pulumi.get', 'pulumi.get', (['self', '"""default_replication_factor"""'], {}), "(self, 'default_replication_factor')\n", (775212, 775248), False, 'import pulumi\n'), ((775361, 775414), 'pulumi.set', 'pulumi.set', (['self', '"""default_replication_factor"""', 'value'], {}), "(self, 'default_replication_factor', value)\n", (775371, 775414), False, 'import pulumi\n'), ((775557, 775604), 'pulumi.get', 'pulumi.get', (['self', '"""log_flush_interval_messages"""'], {}), "(self, 'log_flush_interval_messages')\n", (775567, 775604), False, 'import pulumi\n'), ((775719, 775773), 'pulumi.set', 'pulumi.set', (['self', '"""log_flush_interval_messages"""', 'value'], {}), "(self, 'log_flush_interval_messages', value)\n", (775729, 775773), False, 'import pulumi\n'), ((775904, 775945), 'pulumi.get', 'pulumi.get', (['self', '"""log_flush_interval_ms"""'], {}), "(self, 'log_flush_interval_ms')\n", (775914, 775945), False, 'import pulumi\n'), ((776048, 776096), 'pulumi.set', 'pulumi.set', (['self', '"""log_flush_interval_ms"""', 'value'], {}), "(self, 'log_flush_interval_ms', value)\n", (776058, 776096), False, 'import pulumi\n'), ((776246, 776297), 'pulumi.get', 'pulumi.get', (['self', '"""log_flush_scheduler_interval_ms"""'], {}), "(self, 'log_flush_scheduler_interval_ms')\n", (776256, 776297), False, 'import pulumi\n'), ((776420, 776478), 'pulumi.set', 'pulumi.set', (['self', '"""log_flush_scheduler_interval_ms"""', 'value'], {}), "(self, 'log_flush_scheduler_interval_ms', value)\n", (776430, 776478), False, 'import pulumi\n'), ((776600, 776635), 'pulumi.get', 'pulumi.get', (['self', '"""log_preallocate"""'], {}), "(self, 'log_preallocate')\n", (776610, 776635), False, 'import pulumi\n'), ((776727, 776769), 'pulumi.set', 'pulumi.set', (['self', '"""log_preallocate"""', 'value'], {}), "(self, 'log_preallocate', value)\n", (776737, 776769), False, 'import pulumi\n'), ((776897, 776936), 'pulumi.get', 'pulumi.get', (['self', '"""log_retention_bytes"""'], {}), "(self, 'log_retention_bytes')\n", (776907, 776936), False, 'import pulumi\n'), ((777035, 777081), 'pulumi.set', 'pulumi.set', (['self', '"""log_retention_bytes"""', 'value'], {}), "(self, 'log_retention_bytes', value)\n", (777045, 777081), False, 'import pulumi\n'), ((777209, 777248), 'pulumi.get', 'pulumi.get', (['self', '"""log_retention_hours"""'], {}), "(self, 'log_retention_hours')\n", (777219, 777248), False, 'import pulumi\n'), ((777347, 777393), 'pulumi.set', 'pulumi.set', (['self', '"""log_retention_hours"""', 'value'], {}), "(self, 'log_retention_hours', value)\n", (777357, 777393), False, 'import pulumi\n'), ((777525, 777566), 'pulumi.get', 'pulumi.get', (['self', '"""log_retention_minutes"""'], {}), "(self, 'log_retention_minutes')\n", (777535, 777566), False, 'import pulumi\n'), ((777669, 777717), 'pulumi.set', 'pulumi.set', (['self', '"""log_retention_minutes"""', 'value'], {}), "(self, 'log_retention_minutes', value)\n", (777679, 777717), False, 'import pulumi\n'), ((777839, 777875), 'pulumi.get', 'pulumi.get', (['self', '"""log_retention_ms"""'], {}), "(self, 'log_retention_ms')\n", (777849, 777875), False, 'import pulumi\n'), ((777968, 778011), 'pulumi.set', 'pulumi.set', (['self', '"""log_retention_ms"""', 'value'], {}), "(self, 'log_retention_ms', value)\n", (777978, 778011), False, 'import pulumi\n'), ((778135, 778172), 'pulumi.get', 'pulumi.get', (['self', '"""log_segment_bytes"""'], {}), "(self, 'log_segment_bytes')\n", (778145, 778172), False, 'import pulumi\n'), ((778267, 778311), 'pulumi.set', 'pulumi.set', (['self', '"""log_segment_bytes"""', 'value'], {}), "(self, 'log_segment_bytes', value)\n", (778277, 778311), False, 'import pulumi\n'), ((778430, 778464), 'pulumi.get', 'pulumi.get', (['self', '"""num_partitions"""'], {}), "(self, 'num_partitions')\n", (778440, 778464), False, 'import pulumi\n'), ((778553, 778594), 'pulumi.set', 'pulumi.set', (['self', '"""num_partitions"""', 'value'], {}), "(self, 'num_partitions', value)\n", (778563, 778594), False, 'import pulumi\n'), ((778737, 778784), 'pulumi.get', 'pulumi.get', (['self', '"""socket_receive_buffer_bytes"""'], {}), "(self, 'socket_receive_buffer_bytes')\n", (778747, 778784), False, 'import pulumi\n'), ((778899, 778953), 'pulumi.set', 'pulumi.set', (['self', '"""socket_receive_buffer_bytes"""', 'value'], {}), "(self, 'socket_receive_buffer_bytes', value)\n", (778909, 778953), False, 'import pulumi\n'), ((779090, 779134), 'pulumi.get', 'pulumi.get', (['self', '"""socket_send_buffer_bytes"""'], {}), "(self, 'socket_send_buffer_bytes')\n", (779100, 779134), False, 'import pulumi\n'), ((779243, 779294), 'pulumi.set', 'pulumi.set', (['self', '"""socket_send_buffer_bytes"""', 'value'], {}), "(self, 'socket_send_buffer_bytes', value)\n", (779253, 779294), False, 'import pulumi\n'), ((779864, 779908), 'pulumi.set', 'pulumi.set', (['__self__', '"""disk_size"""', 'disk_size'], {}), "(__self__, 'disk_size', disk_size)\n", (779874, 779908), False, 'import pulumi\n'), ((779917, 779967), 'pulumi.set', 'pulumi.set', (['__self__', '"""disk_type_id"""', 'disk_type_id'], {}), "(__self__, 'disk_type_id', disk_type_id)\n", (779927, 779967), False, 'import pulumi\n'), ((779976, 780038), 'pulumi.set', 'pulumi.set', (['__self__', '"""resource_preset_id"""', 'resource_preset_id'], {}), "(__self__, 'resource_preset_id', resource_preset_id)\n", (779986, 780038), False, 'import pulumi\n'), ((780247, 780276), 'pulumi.get', 'pulumi.get', (['self', '"""disk_size"""'], {}), "(self, 'disk_size')\n", (780257, 780276), False, 'import pulumi\n'), ((780345, 780381), 'pulumi.set', 'pulumi.set', (['self', '"""disk_size"""', 'value'], {}), "(self, 'disk_size', value)\n", (780355, 780381), False, 'import pulumi\n'), ((780693, 780725), 'pulumi.get', 'pulumi.get', (['self', '"""disk_type_id"""'], {}), "(self, 'disk_type_id')\n", (780703, 780725), False, 'import pulumi\n'), ((780800, 780839), 'pulumi.set', 'pulumi.set', (['self', '"""disk_type_id"""', 'value'], {}), "(self, 'disk_type_id', value)\n", (780810, 780839), False, 'import pulumi\n'), ((780955, 780993), 'pulumi.get', 'pulumi.get', (['self', '"""resource_preset_id"""'], {}), "(self, 'resource_preset_id')\n", (780965, 780993), False, 'import pulumi\n'), ((781080, 781125), 'pulumi.set', 'pulumi.set', (['self', '"""resource_preset_id"""', 'value'], {}), "(self, 'resource_preset_id', value)\n", (781090, 781125), False, 'import pulumi\n'), ((781512, 781556), 'pulumi.set', 'pulumi.set', (['__self__', '"""resources"""', 'resources'], {}), "(__self__, 'resources', resources)\n", (781522, 781556), False, 'import pulumi\n'), ((781819, 781848), 'pulumi.get', 'pulumi.get', (['self', '"""resources"""'], {}), "(self, 'resources')\n", (781829, 781848), False, 'import pulumi\n'), ((781962, 781998), 'pulumi.set', 'pulumi.set', (['self', '"""resources"""', 'value'], {}), "(self, 'resources', value)\n", (781972, 781998), False, 'import pulumi\n'), ((782572, 782616), 'pulumi.set', 'pulumi.set', (['__self__', '"""disk_size"""', 'disk_size'], {}), "(__self__, 'disk_size', disk_size)\n", (782582, 782616), False, 'import pulumi\n'), ((782625, 782675), 'pulumi.set', 'pulumi.set', (['__self__', '"""disk_type_id"""', 'disk_type_id'], {}), "(__self__, 'disk_type_id', disk_type_id)\n", (782635, 782675), False, 'import pulumi\n'), ((782684, 782746), 'pulumi.set', 'pulumi.set', (['__self__', '"""resource_preset_id"""', 'resource_preset_id'], {}), "(__self__, 'resource_preset_id', resource_preset_id)\n", (782694, 782746), False, 'import pulumi\n'), ((782955, 782984), 'pulumi.get', 'pulumi.get', (['self', '"""disk_size"""'], {}), "(self, 'disk_size')\n", (782965, 782984), False, 'import pulumi\n'), ((783053, 783089), 'pulumi.set', 'pulumi.set', (['self', '"""disk_size"""', 'value'], {}), "(self, 'disk_size', value)\n", (783063, 783089), False, 'import pulumi\n'), ((783401, 783433), 'pulumi.get', 'pulumi.get', (['self', '"""disk_type_id"""'], {}), "(self, 'disk_type_id')\n", (783411, 783433), False, 'import pulumi\n'), ((783508, 783547), 'pulumi.set', 'pulumi.set', (['self', '"""disk_type_id"""', 'value'], {}), "(self, 'disk_type_id', value)\n", (783518, 783547), False, 'import pulumi\n'), ((783663, 783701), 'pulumi.get', 'pulumi.get', (['self', '"""resource_preset_id"""'], {}), "(self, 'resource_preset_id')\n", (783673, 783701), False, 'import pulumi\n'), ((783788, 783833), 'pulumi.set', 'pulumi.set', (['self', '"""resource_preset_id"""', 'value'], {}), "(self, 'resource_preset_id', value)\n", (783798, 783833), False, 'import pulumi\n'), ((784643, 784689), 'pulumi.set', 'pulumi.set', (['__self__', '"""cluster_id"""', 'cluster_id'], {}), "(__self__, 'cluster_id', cluster_id)\n", (784653, 784689), False, 'import pulumi\n'), ((784698, 784732), 'pulumi.set', 'pulumi.set', (['__self__', '"""name"""', 'name'], {}), "(__self__, 'name', name)\n", (784708, 784732), False, 'import pulumi\n'), ((784741, 784787), 'pulumi.set', 'pulumi.set', (['__self__', '"""partitions"""', 'partitions'], {}), "(__self__, 'partitions', partitions)\n", (784751, 784787), False, 'import pulumi\n'), ((784796, 784858), 'pulumi.set', 'pulumi.set', (['__self__', '"""replication_factor"""', 'replication_factor'], {}), "(__self__, 'replication_factor', replication_factor)\n", (784806, 784858), False, 'import pulumi\n'), ((785120, 785150), 'pulumi.get', 'pulumi.get', (['self', '"""cluster_id"""'], {}), "(self, 'cluster_id')\n", (785130, 785150), False, 'import pulumi\n'), ((785221, 785258), 'pulumi.set', 'pulumi.set', (['self', '"""cluster_id"""', 'value'], {}), "(self, 'cluster_id', value)\n", (785231, 785258), False, 'import pulumi\n'), ((785398, 785422), 'pulumi.get', 'pulumi.get', (['self', '"""name"""'], {}), "(self, 'name')\n", (785408, 785422), False, 'import pulumi\n'), ((785481, 785512), 'pulumi.set', 'pulumi.set', (['self', '"""name"""', 'value'], {}), "(self, 'name', value)\n", (785491, 785512), False, 'import pulumi\n'), ((785676, 785706), 'pulumi.get', 'pulumi.get', (['self', '"""partitions"""'], {}), "(self, 'partitions')\n", (785686, 785706), False, 'import pulumi\n'), ((785777, 785814), 'pulumi.set', 'pulumi.set', (['self', '"""partitions"""', 'value'], {}), "(self, 'partitions', value)\n", (785787, 785814), False, 'import pulumi\n'), ((786037, 786075), 'pulumi.get', 'pulumi.get', (['self', '"""replication_factor"""'], {}), "(self, 'replication_factor')\n", (786047, 786075), False, 'import pulumi\n'), ((786162, 786207), 'pulumi.set', 'pulumi.set', (['self', '"""replication_factor"""', 'value'], {}), "(self, 'replication_factor', value)\n", (786172, 786207), False, 'import pulumi\n'), ((786474, 786506), 'pulumi.get', 'pulumi.get', (['self', '"""topic_config"""'], {}), "(self, 'topic_config')\n", (786484, 786506), False, 'import pulumi\n'), ((786628, 786667), 'pulumi.set', 'pulumi.set', (['self', '"""topic_config"""', 'value'], {}), "(self, 'topic_config', value)\n", (786638, 786667), False, 'import pulumi\n'), ((789046, 789080), 'pulumi.get', 'pulumi.get', (['self', '"""cleanup_policy"""'], {}), "(self, 'cleanup_policy')\n", (789056, 789080), False, 'import pulumi\n'), ((789169, 789210), 'pulumi.set', 'pulumi.set', (['self', '"""cleanup_policy"""', 'value'], {}), "(self, 'cleanup_policy', value)\n", (789179, 789210), False, 'import pulumi\n'), ((789333, 789369), 'pulumi.get', 'pulumi.get', (['self', '"""compression_type"""'], {}), "(self, 'compression_type')\n", (789343, 789369), False, 'import pulumi\n'), ((789462, 789505), 'pulumi.set', 'pulumi.set', (['self', '"""compression_type"""', 'value'], {}), "(self, 'compression_type', value)\n", (789472, 789505), False, 'import pulumi\n'), ((789633, 789672), 'pulumi.get', 'pulumi.get', (['self', '"""delete_retention_ms"""'], {}), "(self, 'delete_retention_ms')\n", (789643, 789672), False, 'import pulumi\n'), ((789771, 789817), 'pulumi.set', 'pulumi.set', (['self', '"""delete_retention_ms"""', 'value'], {}), "(self, 'delete_retention_ms', value)\n", (789781, 789817), False, 'import pulumi\n'), ((789946, 789986), 'pulumi.get', 'pulumi.get', (['self', '"""file_delete_delay_ms"""'], {}), "(self, 'file_delete_delay_ms')\n", (789956, 789986), False, 'import pulumi\n'), ((790087, 790134), 'pulumi.set', 'pulumi.set', (['self', '"""file_delete_delay_ms"""', 'value'], {}), "(self, 'file_delete_delay_ms', value)\n", (790097, 790134), False, 'import pulumi\n'), ((790253, 790287), 'pulumi.get', 'pulumi.get', (['self', '"""flush_messages"""'], {}), "(self, 'flush_messages')\n", (790263, 790287), False, 'import pulumi\n'), ((790376, 790417), 'pulumi.set', 'pulumi.set', (['self', '"""flush_messages"""', 'value'], {}), "(self, 'flush_messages', value)\n", (790386, 790417), False, 'import pulumi\n'), ((790524, 790552), 'pulumi.get', 'pulumi.get', (['self', '"""flush_ms"""'], {}), "(self, 'flush_ms')\n", (790534, 790552), False, 'import pulumi\n'), ((790629, 790664), 'pulumi.set', 'pulumi.set', (['self', '"""flush_ms"""', 'value'], {}), "(self, 'flush_ms', value)\n", (790639, 790664), False, 'import pulumi\n'), ((790788, 790825), 'pulumi.get', 'pulumi.get', (['self', '"""max_message_bytes"""'], {}), "(self, 'max_message_bytes')\n", (790798, 790825), False, 'import pulumi\n'), ((790920, 790964), 'pulumi.set', 'pulumi.set', (['self', '"""max_message_bytes"""', 'value'], {}), "(self, 'max_message_bytes', value)\n", (790930, 790964), False, 'import pulumi\n'), ((791095, 791136), 'pulumi.get', 'pulumi.get', (['self', '"""min_compaction_lag_ms"""'], {}), "(self, 'min_compaction_lag_ms')\n", (791105, 791136), False, 'import pulumi\n'), ((791239, 791287), 'pulumi.set', 'pulumi.set', (['self', '"""min_compaction_lag_ms"""', 'value'], {}), "(self, 'min_compaction_lag_ms', value)\n", (791249, 791287), False, 'import pulumi\n'), ((791415, 791454), 'pulumi.get', 'pulumi.get', (['self', '"""min_insync_replicas"""'], {}), "(self, 'min_insync_replicas')\n", (791425, 791454), False, 'import pulumi\n'), ((791553, 791599), 'pulumi.set', 'pulumi.set', (['self', '"""min_insync_replicas"""', 'value'], {}), "(self, 'min_insync_replicas', value)\n", (791563, 791599), False, 'import pulumi\n'), ((791694, 791725), 'pulumi.get', 'pulumi.get', (['self', '"""preallocate"""'], {}), "(self, 'preallocate')\n", (791704, 791725), False, 'import pulumi\n'), ((791809, 791847), 'pulumi.set', 'pulumi.set', (['self', '"""preallocate"""', 'value'], {}), "(self, 'preallocate', value)\n", (791819, 791847), False, 'import pulumi\n'), ((791968, 792003), 'pulumi.get', 'pulumi.get', (['self', '"""retention_bytes"""'], {}), "(self, 'retention_bytes')\n", (791978, 792003), False, 'import pulumi\n'), ((792094, 792136), 'pulumi.set', 'pulumi.set', (['self', '"""retention_bytes"""', 'value'], {}), "(self, 'retention_bytes', value)\n", (792104, 792136), False, 'import pulumi\n'), ((792251, 792283), 'pulumi.get', 'pulumi.get', (['self', '"""retention_ms"""'], {}), "(self, 'retention_ms')\n", (792261, 792283), False, 'import pulumi\n'), ((792368, 792407), 'pulumi.set', 'pulumi.set', (['self', '"""retention_ms"""', 'value'], {}), "(self, 'retention_ms', value)\n", (792378, 792407), False, 'import pulumi\n'), ((792524, 792557), 'pulumi.get', 'pulumi.get', (['self', '"""segment_bytes"""'], {}), "(self, 'segment_bytes')\n", (792534, 792557), False, 'import pulumi\n'), ((792644, 792684), 'pulumi.set', 'pulumi.set', (['self', '"""segment_bytes"""', 'value'], {}), "(self, 'segment_bytes', value)\n", (792654, 792684), False, 'import pulumi\n'), ((793245, 793279), 'pulumi.set', 'pulumi.set', (['__self__', '"""name"""', 'name'], {}), "(__self__, 'name', name)\n", (793255, 793279), False, 'import pulumi\n'), ((793288, 793330), 'pulumi.set', 'pulumi.set', (['__self__', '"""password"""', 'password'], {}), "(__self__, 'password', password)\n", (793298, 793330), False, 'import pulumi\n'), ((793567, 793591), 'pulumi.get', 'pulumi.get', (['self', '"""name"""'], {}), "(self, 'name')\n", (793577, 793591), False, 'import pulumi\n'), ((793650, 793681), 'pulumi.set', 'pulumi.set', (['self', '"""name"""', 'value'], {}), "(self, 'name', value)\n", (793660, 793681), False, 'import pulumi\n'), ((793831, 793859), 'pulumi.get', 'pulumi.get', (['self', '"""password"""'], {}), "(self, 'password')\n", (793841, 793859), False, 'import pulumi\n'), ((793926, 793961), 'pulumi.set', 'pulumi.set', (['self', '"""password"""', 'value'], {}), "(self, 'password', value)\n", (793936, 793961), False, 'import pulumi\n'), ((794218, 794249), 'pulumi.get', 'pulumi.get', (['self', '"""permissions"""'], {}), "(self, 'permissions')\n", (794228, 794249), False, 'import pulumi\n'), ((794377, 794415), 'pulumi.set', 'pulumi.set', (['self', '"""permissions"""', 'value'], {}), "(self, 'permissions', value)\n", (794387, 794415), False, 'import pulumi\n'), ((794766, 794800), 'pulumi.set', 'pulumi.set', (['__self__', '"""role"""', 'role'], {}), "(__self__, 'role', role)\n", (794776, 794800), False, 'import pulumi\n'), ((794809, 794855), 'pulumi.set', 'pulumi.set', (['__self__', '"""topic_name"""', 'topic_name'], {}), "(__self__, 'topic_name', topic_name)\n", (794819, 794855), False, 'import pulumi\n'), ((794997, 795021), 'pulumi.get', 'pulumi.get', (['self', '"""role"""'], {}), "(self, 'role')\n", (795007, 795021), False, 'import pulumi\n'), ((795080, 795111), 'pulumi.set', 'pulumi.set', (['self', '"""role"""', 'value'], {}), "(self, 'role', value)\n", (795090, 795111), False, 'import pulumi\n'), ((795315, 795345), 'pulumi.get', 'pulumi.get', (['self', '"""topic_name"""'], {}), "(self, 'topic_name')\n", (795325, 795345), False, 'import pulumi\n'), ((795416, 795453), 'pulumi.set', 'pulumi.set', (['self', '"""topic_name"""', 'value'], {}), "(self, 'topic_name', value)\n", (795426, 795453), False, 'import pulumi\n'), ((795909, 795953), 'pulumi.set', 'pulumi.set', (['__self__', '"""data_lens"""', 'data_lens'], {}), "(__self__, 'data_lens', data_lens)\n", (795919, 795953), False, 'import pulumi\n'), ((795962, 796002), 'pulumi.set', 'pulumi.set', (['__self__', '"""web_sql"""', 'web_sql'], {}), "(__self__, 'web_sql', web_sql)\n", (795972, 796002), False, 'import pulumi\n'), ((796214, 796243), 'pulumi.get', 'pulumi.get', (['self', '"""data_lens"""'], {}), "(self, 'data_lens')\n", (796224, 796243), False, 'import pulumi\n'), ((796313, 796349), 'pulumi.set', 'pulumi.set', (['self', '"""data_lens"""', 'value'], {}), "(self, 'data_lens', value)\n", (796323, 796349), False, 'import pulumi\n'), ((796606, 796633), 'pulumi.get', 'pulumi.get', (['self', '"""web_sql"""'], {}), "(self, 'web_sql')\n", (796616, 796633), False, 'import pulumi\n'), ((796699, 796733), 'pulumi.set', 'pulumi.set', (['self', '"""web_sql"""', 'value'], {}), "(self, 'web_sql', value)\n", (796709, 796733), False, 'import pulumi\n'), ((15698, 15746), 'pulumi.set', 'pulumi.set', (['__self__', '"""healthcheck"""', 'healthcheck'], {}), "(__self__, 'healthcheck', healthcheck)\n", (15708, 15746), False, 'import pulumi\n'), ((15805, 15873), 'pulumi.set', 'pulumi.set', (['__self__', '"""load_balancing_config"""', 'load_balancing_config'], {}), "(__self__, 'load_balancing_config', load_balancing_config)\n", (15815, 15873), False, 'import pulumi\n'), ((15915, 15949), 'pulumi.set', 'pulumi.set', (['__self__', '"""port"""', 'port'], {}), "(__self__, 'port', port)\n", (15925, 15949), False, 'import pulumi\n'), ((15990, 16022), 'pulumi.set', 'pulumi.set', (['__self__', '"""tls"""', 'tls'], {}), "(__self__, 'tls', tls)\n", (16000, 16022), False, 'import pulumi\n'), ((16066, 16104), 'pulumi.set', 'pulumi.set', (['__self__', '"""weight"""', 'weight'], {}), "(__self__, 'weight', weight)\n", (16076, 16104), False, 'import pulumi\n'), ((21649, 21707), 'pulumi.set', 'pulumi.set', (['__self__', '"""grpc_healthcheck"""', 'grpc_healthcheck'], {}), "(__self__, 'grpc_healthcheck', grpc_healthcheck)\n", (21659, 21707), False, 'import pulumi\n'), ((21761, 21819), 'pulumi.set', 'pulumi.set', (['__self__', '"""healthcheck_port"""', 'healthcheck_port'], {}), "(__self__, 'healthcheck_port', healthcheck_port)\n", (21771, 21819), False, 'import pulumi\n'), ((21874, 21934), 'pulumi.set', 'pulumi.set', (['__self__', '"""healthy_threshold"""', 'healthy_threshold'], {}), "(__self__, 'healthy_threshold', healthy_threshold)\n", (21884, 21934), False, 'import pulumi\n'), ((21988, 22046), 'pulumi.set', 'pulumi.set', (['__self__', '"""http_healthcheck"""', 'http_healthcheck'], {}), "(__self__, 'http_healthcheck', http_healthcheck)\n", (21998, 22046), False, 'import pulumi\n'), ((22107, 22179), 'pulumi.set', 'pulumi.set', (['__self__', '"""interval_jitter_percent"""', 'interval_jitter_percent'], {}), "(__self__, 'interval_jitter_percent', interval_jitter_percent)\n", (22117, 22179), False, 'import pulumi\n'), ((22235, 22297), 'pulumi.set', 'pulumi.set', (['__self__', '"""stream_healthcheck"""', 'stream_healthcheck'], {}), "(__self__, 'stream_healthcheck', stream_healthcheck)\n", (22245, 22297), False, 'import pulumi\n'), ((22354, 22418), 'pulumi.set', 'pulumi.set', (['__self__', '"""unhealthy_threshold"""', 'unhealthy_threshold'], {}), "(__self__, 'unhealthy_threshold', unhealthy_threshold)\n", (22364, 22418), False, 'import pulumi\n'), ((27227, 27277), 'pulumi.set', 'pulumi.set', (['__self__', '"""service_name"""', 'service_name'], {}), "(__self__, 'service_name', service_name)\n", (27237, 27277), False, 'import pulumi\n'), ((28249, 28283), 'pulumi.set', 'pulumi.set', (['__self__', '"""host"""', 'host'], {}), "(__self__, 'host', host)\n", (28259, 28283), False, 'import pulumi\n'), ((28326, 28362), 'pulumi.set', 'pulumi.set', (['__self__', '"""http2"""', 'http2'], {}), "(__self__, 'http2', http2)\n", (28336, 28362), False, 'import pulumi\n'), ((29712, 29752), 'pulumi.set', 'pulumi.set', (['__self__', '"""receive"""', 'receive'], {}), "(__self__, 'receive', receive)\n", (29722, 29752), False, 'import pulumi\n'), ((29794, 29828), 'pulumi.set', 'pulumi.set', (['__self__', '"""send"""', 'send'], {}), "(__self__, 'send', send)\n", (29804, 29828), False, 'import pulumi\n'), ((31553, 31643), 'pulumi.set', 'pulumi.set', (['__self__', '"""locality_aware_routing_percent"""', 'locality_aware_routing_percent'], {}), "(__self__, 'locality_aware_routing_percent',\n locality_aware_routing_percent)\n", (31563, 31643), False, 'import pulumi\n'), ((31692, 31748), 'pulumi.set', 'pulumi.set', (['__self__', '"""panic_threshold"""', 'panic_threshold'], {}), "(__self__, 'panic_threshold', panic_threshold)\n", (31702, 31748), False, 'import pulumi\n'), ((31801, 31857), 'pulumi.set', 'pulumi.set', (['__self__', '"""strict_locality"""', 'strict_locality'], {}), "(__self__, 'strict_locality', strict_locality)\n", (31811, 31857), False, 'import pulumi\n'), ((34180, 34212), 'pulumi.set', 'pulumi.set', (['__self__', '"""sni"""', 'sni'], {}), "(__self__, 'sni', sni)\n", (34190, 34212), False, 'import pulumi\n'), ((34268, 34330), 'pulumi.set', 'pulumi.set', (['__self__', '"""validation_context"""', 'validation_context'], {}), "(__self__, 'validation_context', validation_context)\n", (34278, 34330), False, 'import pulumi\n'), ((35639, 35697), 'pulumi.set', 'pulumi.set', (['__self__', '"""trusted_ca_bytes"""', 'trusted_ca_bytes'], {}), "(__self__, 'trusted_ca_bytes', trusted_ca_bytes)\n", (35649, 35697), False, 'import pulumi\n'), ((35748, 35800), 'pulumi.set', 'pulumi.set', (['__self__', '"""trusted_ca_id"""', 'trusted_ca_id'], {}), "(__self__, 'trusted_ca_id', trusted_ca_id)\n", (35758, 35800), False, 'import pulumi\n'), ((38336, 38384), 'pulumi.set', 'pulumi.set', (['__self__', '"""healthcheck"""', 'healthcheck'], {}), "(__self__, 'healthcheck', healthcheck)\n", (38346, 38384), False, 'import pulumi\n'), ((38427, 38463), 'pulumi.set', 'pulumi.set', (['__self__', '"""http2"""', 'http2'], {}), "(__self__, 'http2', http2)\n", (38437, 38463), False, 'import pulumi\n'), ((38522, 38590), 'pulumi.set', 'pulumi.set', (['__self__', '"""load_balancing_config"""', 'load_balancing_config'], {}), "(__self__, 'load_balancing_config', load_balancing_config)\n", (38532, 38590), False, 'import pulumi\n'), ((38632, 38666), 'pulumi.set', 'pulumi.set', (['__self__', '"""port"""', 'port'], {}), "(__self__, 'port', port)\n", (38642, 38666), False, 'import pulumi\n'), ((38707, 38739), 'pulumi.set', 'pulumi.set', (['__self__', '"""tls"""', 'tls'], {}), "(__self__, 'tls', tls)\n", (38717, 38739), False, 'import pulumi\n'), ((38783, 38821), 'pulumi.set', 'pulumi.set', (['__self__', '"""weight"""', 'weight'], {}), "(__self__, 'weight', weight)\n", (38793, 38821), False, 'import pulumi\n'), ((44682, 44740), 'pulumi.set', 'pulumi.set', (['__self__', '"""grpc_healthcheck"""', 'grpc_healthcheck'], {}), "(__self__, 'grpc_healthcheck', grpc_healthcheck)\n", (44692, 44740), False, 'import pulumi\n'), ((44794, 44852), 'pulumi.set', 'pulumi.set', (['__self__', '"""healthcheck_port"""', 'healthcheck_port'], {}), "(__self__, 'healthcheck_port', healthcheck_port)\n", (44804, 44852), False, 'import pulumi\n'), ((44907, 44967), 'pulumi.set', 'pulumi.set', (['__self__', '"""healthy_threshold"""', 'healthy_threshold'], {}), "(__self__, 'healthy_threshold', healthy_threshold)\n", (44917, 44967), False, 'import pulumi\n'), ((45021, 45079), 'pulumi.set', 'pulumi.set', (['__self__', '"""http_healthcheck"""', 'http_healthcheck'], {}), "(__self__, 'http_healthcheck', http_healthcheck)\n", (45031, 45079), False, 'import pulumi\n'), ((45140, 45212), 'pulumi.set', 'pulumi.set', (['__self__', '"""interval_jitter_percent"""', 'interval_jitter_percent'], {}), "(__self__, 'interval_jitter_percent', interval_jitter_percent)\n", (45150, 45212), False, 'import pulumi\n'), ((45268, 45330), 'pulumi.set', 'pulumi.set', (['__self__', '"""stream_healthcheck"""', 'stream_healthcheck'], {}), "(__self__, 'stream_healthcheck', stream_healthcheck)\n", (45278, 45330), False, 'import pulumi\n'), ((45387, 45451), 'pulumi.set', 'pulumi.set', (['__self__', '"""unhealthy_threshold"""', 'unhealthy_threshold'], {}), "(__self__, 'unhealthy_threshold', unhealthy_threshold)\n", (45397, 45451), False, 'import pulumi\n'), ((50260, 50310), 'pulumi.set', 'pulumi.set', (['__self__', '"""service_name"""', 'service_name'], {}), "(__self__, 'service_name', service_name)\n", (50270, 50310), False, 'import pulumi\n'), ((51282, 51316), 'pulumi.set', 'pulumi.set', (['__self__', '"""host"""', 'host'], {}), "(__self__, 'host', host)\n", (51292, 51316), False, 'import pulumi\n'), ((51359, 51395), 'pulumi.set', 'pulumi.set', (['__self__', '"""http2"""', 'http2'], {}), "(__self__, 'http2', http2)\n", (51369, 51395), False, 'import pulumi\n'), ((52745, 52785), 'pulumi.set', 'pulumi.set', (['__self__', '"""receive"""', 'receive'], {}), "(__self__, 'receive', receive)\n", (52755, 52785), False, 'import pulumi\n'), ((52827, 52861), 'pulumi.set', 'pulumi.set', (['__self__', '"""send"""', 'send'], {}), "(__self__, 'send', send)\n", (52837, 52861), False, 'import pulumi\n'), ((54586, 54676), 'pulumi.set', 'pulumi.set', (['__self__', '"""locality_aware_routing_percent"""', 'locality_aware_routing_percent'], {}), "(__self__, 'locality_aware_routing_percent',\n locality_aware_routing_percent)\n", (54596, 54676), False, 'import pulumi\n'), ((54725, 54781), 'pulumi.set', 'pulumi.set', (['__self__', '"""panic_threshold"""', 'panic_threshold'], {}), "(__self__, 'panic_threshold', panic_threshold)\n", (54735, 54781), False, 'import pulumi\n'), ((54834, 54890), 'pulumi.set', 'pulumi.set', (['__self__', '"""strict_locality"""', 'strict_locality'], {}), "(__self__, 'strict_locality', strict_locality)\n", (54844, 54890), False, 'import pulumi\n'), ((57213, 57245), 'pulumi.set', 'pulumi.set', (['__self__', '"""sni"""', 'sni'], {}), "(__self__, 'sni', sni)\n", (57223, 57245), False, 'import pulumi\n'), ((57301, 57363), 'pulumi.set', 'pulumi.set', (['__self__', '"""validation_context"""', 'validation_context'], {}), "(__self__, 'validation_context', validation_context)\n", (57311, 57363), False, 'import pulumi\n'), ((58672, 58730), 'pulumi.set', 'pulumi.set', (['__self__', '"""trusted_ca_bytes"""', 'trusted_ca_bytes'], {}), "(__self__, 'trusted_ca_bytes', trusted_ca_bytes)\n", (58682, 58730), False, 'import pulumi\n'), ((58781, 58833), 'pulumi.set', 'pulumi.set', (['__self__', '"""trusted_ca_id"""', 'trusted_ca_id'], {}), "(__self__, 'trusted_ca_id', trusted_ca_id)\n", (58791, 58833), False, 'import pulumi\n'), ((61130, 61186), 'pulumi.set', 'pulumi.set', (['__self__', '"""disable_traffic"""', 'disable_traffic'], {}), "(__self__, 'disable_traffic', disable_traffic)\n", (61140, 61186), False, 'import pulumi\n'), ((63323, 63367), 'pulumi.set', 'pulumi.set', (['__self__', '"""endpoints"""', 'endpoints'], {}), "(__self__, 'endpoints', endpoints)\n", (63333, 63367), False, 'import pulumi\n'), ((63409, 63443), 'pulumi.set', 'pulumi.set', (['__self__', '"""http"""', 'http'], {}), "(__self__, 'http', http)\n", (63419, 63443), False, 'import pulumi\n'), ((63484, 63516), 'pulumi.set', 'pulumi.set', (['__self__', '"""tls"""', 'tls'], {}), "(__self__, 'tls', tls)\n", (63494, 63516), False, 'import pulumi\n'), ((67620, 67688), 'pulumi.set', 'pulumi.set', (['__self__', '"""external_ipv4_address"""', 'external_ipv4_address'], {}), "(__self__, 'external_ipv4_address', external_ipv4_address)\n", (67630, 67688), False, 'import pulumi\n'), ((67747, 67815), 'pulumi.set', 'pulumi.set', (['__self__', '"""external_ipv6_address"""', 'external_ipv6_address'], {}), "(__self__, 'external_ipv6_address', external_ipv6_address)\n", (67757, 67815), False, 'import pulumi\n'), ((67874, 67942), 'pulumi.set', 'pulumi.set', (['__self__', '"""internal_ipv4_address"""', 'internal_ipv4_address'], {}), "(__self__, 'internal_ipv4_address', internal_ipv4_address)\n", (67884, 67942), False, 'import pulumi\n'), ((69972, 70012), 'pulumi.set', 'pulumi.set', (['__self__', '"""address"""', 'address'], {}), "(__self__, 'address', address)\n", (69982, 70012), False, 'import pulumi\n'), ((70692, 70732), 'pulumi.set', 'pulumi.set', (['__self__', '"""address"""', 'address'], {}), "(__self__, 'address', address)\n", (70702, 70732), False, 'import pulumi\n'), ((71570, 71610), 'pulumi.set', 'pulumi.set', (['__self__', '"""address"""', 'address'], {}), "(__self__, 'address', address)\n", (71580, 71610), False, 'import pulumi\n'), ((71657, 71701), 'pulumi.set', 'pulumi.set', (['__self__', '"""subnet_id"""', 'subnet_id'], {}), "(__self__, 'subnet_id', subnet_id)\n", (71667, 71701), False, 'import pulumi\n'), ((73084, 73124), 'pulumi.set', 'pulumi.set', (['__self__', '"""handler"""', 'handler'], {}), "(__self__, 'handler', handler)\n", (73094, 73124), False, 'import pulumi\n'), ((73171, 73215), 'pulumi.set', 'pulumi.set', (['__self__', '"""redirects"""', 'redirects'], {}), "(__self__, 'redirects', redirects)\n", (73181, 73215), False, 'import pulumi\n'), ((74894, 74944), 'pulumi.set', 'pulumi.set', (['__self__', '"""allow_http10"""', 'allow_http10'], {}), "(__self__, 'allow_http10', allow_http10)\n", (74904, 74944), False, 'import pulumi\n'), ((74995, 75047), 'pulumi.set', 'pulumi.set', (['__self__', '"""http2_options"""', 'http2_options'], {}), "(__self__, 'http2_options', http2_options)\n", (75005, 75047), False, 'import pulumi\n'), ((75099, 75153), 'pulumi.set', 'pulumi.set', (['__self__', '"""http_router_id"""', 'http_router_id'], {}), "(__self__, 'http_router_id', http_router_id)\n", (75109, 75153), False, 'import pulumi\n'), ((76794, 76864), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_concurrent_streams"""', 'max_concurrent_streams'], {}), "(__self__, 'max_concurrent_streams', max_concurrent_streams)\n", (76804, 76864), False, 'import pulumi\n'), ((77512, 77564), 'pulumi.set', 'pulumi.set', (['__self__', '"""http_to_https"""', 'http_to_https'], {}), "(__self__, 'http_to_https', http_to_https)\n", (77522, 77564), False, 'import pulumi\n'), ((78640, 78690), 'pulumi.set', 'pulumi.set', (['__self__', '"""sni_handlers"""', 'sni_handlers'], {}), "(__self__, 'sni_handlers', sni_handlers)\n", (78650, 78690), False, 'import pulumi\n'), ((80569, 80619), 'pulumi.set', 'pulumi.set', (['__self__', '"""http_handler"""', 'http_handler'], {}), "(__self__, 'http_handler', http_handler)\n", (80579, 80619), False, 'import pulumi\n'), ((82543, 82593), 'pulumi.set', 'pulumi.set', (['__self__', '"""allow_http10"""', 'allow_http10'], {}), "(__self__, 'allow_http10', allow_http10)\n", (82553, 82593), False, 'import pulumi\n'), ((82644, 82696), 'pulumi.set', 'pulumi.set', (['__self__', '"""http2_options"""', 'http2_options'], {}), "(__self__, 'http2_options', http2_options)\n", (82654, 82696), False, 'import pulumi\n'), ((82748, 82802), 'pulumi.set', 'pulumi.set', (['__self__', '"""http_router_id"""', 'http_router_id'], {}), "(__self__, 'http_router_id', http_router_id)\n", (82758, 82802), False, 'import pulumi\n'), ((84494, 84564), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_concurrent_streams"""', 'max_concurrent_streams'], {}), "(__self__, 'max_concurrent_streams', max_concurrent_streams)\n", (84504, 84564), False, 'import pulumi\n'), ((87790, 87840), 'pulumi.set', 'pulumi.set', (['__self__', '"""http_handler"""', 'http_handler'], {}), "(__self__, 'http_handler', http_handler)\n", (87800, 87840), False, 'import pulumi\n'), ((89779, 89829), 'pulumi.set', 'pulumi.set', (['__self__', '"""allow_http10"""', 'allow_http10'], {}), "(__self__, 'allow_http10', allow_http10)\n", (89789, 89829), False, 'import pulumi\n'), ((89880, 89932), 'pulumi.set', 'pulumi.set', (['__self__', '"""http2_options"""', 'http2_options'], {}), "(__self__, 'http2_options', http2_options)\n", (89890, 89932), False, 'import pulumi\n'), ((89984, 90038), 'pulumi.set', 'pulumi.set', (['__self__', '"""http_router_id"""', 'http_router_id'], {}), "(__self__, 'http_router_id', http_router_id)\n", (89994, 90038), False, 'import pulumi\n'), ((91739, 91809), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_concurrent_streams"""', 'max_concurrent_streams'], {}), "(__self__, 'max_concurrent_streams', max_concurrent_streams)\n", (91749, 91809), False, 'import pulumi\n'), ((94470, 94508), 'pulumi.set', 'pulumi.set', (['__self__', '"""append"""', 'append'], {}), "(__self__, 'append', append)\n", (94480, 94508), False, 'import pulumi\n'), ((94550, 94584), 'pulumi.set', 'pulumi.set', (['__self__', '"""name"""', 'name'], {}), "(__self__, 'name', name)\n", (94560, 94584), False, 'import pulumi\n'), ((94628, 94666), 'pulumi.set', 'pulumi.set', (['__self__', '"""remove"""', 'remove'], {}), "(__self__, 'remove', remove)\n", (94638, 94666), False, 'import pulumi\n'), ((94711, 94751), 'pulumi.set', 'pulumi.set', (['__self__', '"""replace"""', 'replace'], {}), "(__self__, 'replace', replace)\n", (94721, 94751), False, 'import pulumi\n'), ((97017, 97055), 'pulumi.set', 'pulumi.set', (['__self__', '"""append"""', 'append'], {}), "(__self__, 'append', append)\n", (97027, 97055), False, 'import pulumi\n'), ((97097, 97131), 'pulumi.set', 'pulumi.set', (['__self__', '"""name"""', 'name'], {}), "(__self__, 'name', name)\n", (97107, 97131), False, 'import pulumi\n'), ((97175, 97213), 'pulumi.set', 'pulumi.set', (['__self__', '"""remove"""', 'remove'], {}), "(__self__, 'remove', remove)\n", (97185, 97213), False, 'import pulumi\n'), ((97258, 97298), 'pulumi.set', 'pulumi.set', (['__self__', '"""replace"""', 'replace'], {}), "(__self__, 'replace', replace)\n", (97268, 97298), False, 'import pulumi\n'), ((99429, 99475), 'pulumi.set', 'pulumi.set', (['__self__', '"""grpc_route"""', 'grpc_route'], {}), "(__self__, 'grpc_route', grpc_route)\n", (99439, 99475), False, 'import pulumi\n'), ((99523, 99569), 'pulumi.set', 'pulumi.set', (['__self__', '"""http_route"""', 'http_route'], {}), "(__self__, 'http_route', http_route)\n", (99533, 99569), False, 'import pulumi\n'), ((99611, 99645), 'pulumi.set', 'pulumi.set', (['__self__', '"""name"""', 'name'], {}), "(__self__, 'name', name)\n", (99621, 99645), False, 'import pulumi\n'), ((101888, 101938), 'pulumi.set', 'pulumi.set', (['__self__', '"""grpc_matches"""', 'grpc_matches'], {}), "(__self__, 'grpc_matches', grpc_matches)\n", (101898, 101938), False, 'import pulumi\n'), ((101993, 102053), 'pulumi.set', 'pulumi.set', (['__self__', '"""grpc_route_action"""', 'grpc_route_action'], {}), "(__self__, 'grpc_route_action', grpc_route_action)\n", (102003, 102053), False, 'import pulumi\n'), ((102118, 102203), 'pulumi.set', 'pulumi.set', (['__self__', '"""grpc_status_response_action"""', 'grpc_status_response_action'], {}), "(__self__, 'grpc_status_response_action', grpc_status_response_action\n )\n", (102128, 102203), False, 'import pulumi\n'), ((104273, 104307), 'pulumi.set', 'pulumi.set', (['__self__', '"""fqmn"""', 'fqmn'], {}), "(__self__, 'fqmn', fqmn)\n", (104283, 104307), False, 'import pulumi\n'), ((105016, 105052), 'pulumi.set', 'pulumi.set', (['__self__', '"""exact"""', 'exact'], {}), "(__self__, 'exact', exact)\n", (105026, 105052), False, 'import pulumi\n'), ((105096, 105134), 'pulumi.set', 'pulumi.set', (['__self__', '"""prefix"""', 'prefix'], {}), "(__self__, 'prefix', prefix)\n", (105106, 105134), False, 'import pulumi\n'), ((107033, 107093), 'pulumi.set', 'pulumi.set', (['__self__', '"""auto_host_rewrite"""', 'auto_host_rewrite'], {}), "(__self__, 'auto_host_rewrite', auto_host_rewrite)\n", (107043, 107093), False, 'import pulumi\n'), ((107143, 107193), 'pulumi.set', 'pulumi.set', (['__self__', '"""host_rewrite"""', 'host_rewrite'], {}), "(__self__, 'host_rewrite', host_rewrite)\n", (107153, 107193), False, 'import pulumi\n'), ((107243, 107293), 'pulumi.set', 'pulumi.set', (['__self__', '"""idle_timeout"""', 'idle_timeout'], {}), "(__self__, 'idle_timeout', idle_timeout)\n", (107253, 107293), False, 'import pulumi\n'), ((107342, 107390), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_timeout"""', 'max_timeout'], {}), "(__self__, 'max_timeout', max_timeout)\n", (107352, 107390), False, 'import pulumi\n'), ((110085, 110123), 'pulumi.set', 'pulumi.set', (['__self__', '"""status"""', 'status'], {}), "(__self__, 'status', status)\n", (110095, 110123), False, 'import pulumi\n'), ((111906, 111976), 'pulumi.set', 'pulumi.set', (['__self__', '"""direct_response_action"""', 'direct_response_action'], {}), "(__self__, 'direct_response_action', direct_response_action)\n", (111916, 111976), False, 'import pulumi\n'), ((112026, 112076), 'pulumi.set', 'pulumi.set', (['__self__', '"""http_matches"""', 'http_matches'], {}), "(__self__, 'http_matches', http_matches)\n", (112036, 112076), False, 'import pulumi\n'), ((112131, 112191), 'pulumi.set', 'pulumi.set', (['__self__', '"""http_route_action"""', 'http_route_action'], {}), "(__self__, 'http_route_action', http_route_action)\n", (112141, 112191), False, 'import pulumi\n'), ((112244, 112300), 'pulumi.set', 'pulumi.set', (['__self__', '"""redirect_action"""', 'redirect_action'], {}), "(__self__, 'redirect_action', redirect_action)\n", (112254, 112300), False, 'import pulumi\n'), ((114969, 115003), 'pulumi.set', 'pulumi.set', (['__self__', '"""body"""', 'body'], {}), "(__self__, 'body', body)\n", (114979, 115003), False, 'import pulumi\n'), ((115047, 115085), 'pulumi.set', 'pulumi.set', (['__self__', '"""status"""', 'status'], {}), "(__self__, 'status', status)\n", (115057, 115085), False, 'import pulumi\n'), ((116406, 116456), 'pulumi.set', 'pulumi.set', (['__self__', '"""http_methods"""', 'http_methods'], {}), "(__self__, 'http_methods', http_methods)\n", (116416, 116456), False, 'import pulumi\n'), ((116498, 116532), 'pulumi.set', 'pulumi.set', (['__self__', '"""path"""', 'path'], {}), "(__self__, 'path', path)\n", (116508, 116532), False, 'import pulumi\n'), ((117600, 117636), 'pulumi.set', 'pulumi.set', (['__self__', '"""exact"""', 'exact'], {}), "(__self__, 'exact', exact)\n", (117610, 117636), False, 'import pulumi\n'), ((117680, 117718), 'pulumi.set', 'pulumi.set', (['__self__', '"""prefix"""', 'prefix'], {}), "(__self__, 'prefix', prefix)\n", (117690, 117718), False, 'import pulumi\n'), ((120092, 120152), 'pulumi.set', 'pulumi.set', (['__self__', '"""auto_host_rewrite"""', 'auto_host_rewrite'], {}), "(__self__, 'auto_host_rewrite', auto_host_rewrite)\n", (120102, 120152), False, 'import pulumi\n'), ((120202, 120252), 'pulumi.set', 'pulumi.set', (['__self__', '"""host_rewrite"""', 'host_rewrite'], {}), "(__self__, 'host_rewrite', host_rewrite)\n", (120212, 120252), False, 'import pulumi\n'), ((120302, 120352), 'pulumi.set', 'pulumi.set', (['__self__', '"""idle_timeout"""', 'idle_timeout'], {}), "(__self__, 'idle_timeout', idle_timeout)\n", (120312, 120352), False, 'import pulumi\n'), ((120404, 120458), 'pulumi.set', 'pulumi.set', (['__self__', '"""prefix_rewrite"""', 'prefix_rewrite'], {}), "(__self__, 'prefix_rewrite', prefix_rewrite)\n", (120414, 120458), False, 'import pulumi\n'), ((120503, 120543), 'pulumi.set', 'pulumi.set', (['__self__', '"""timeout"""', 'timeout'], {}), "(__self__, 'timeout', timeout)\n", (120513, 120543), False, 'import pulumi\n'), ((120594, 120646), 'pulumi.set', 'pulumi.set', (['__self__', '"""upgrade_types"""', 'upgrade_types'], {}), "(__self__, 'upgrade_types', upgrade_types)\n", (120604, 120646), False, 'import pulumi\n'), ((125239, 125289), 'pulumi.set', 'pulumi.set', (['__self__', '"""remove_query"""', 'remove_query'], {}), "(__self__, 'remove_query', remove_query)\n", (125249, 125289), False, 'import pulumi\n'), ((125339, 125389), 'pulumi.set', 'pulumi.set', (['__self__', '"""replace_host"""', 'replace_host'], {}), "(__self__, 'replace_host', replace_host)\n", (125349, 125389), False, 'import pulumi\n'), ((125439, 125489), 'pulumi.set', 'pulumi.set', (['__self__', '"""replace_path"""', 'replace_path'], {}), "(__self__, 'replace_path', replace_path)\n", (125449, 125489), False, 'import pulumi\n'), ((125539, 125589), 'pulumi.set', 'pulumi.set', (['__self__', '"""replace_port"""', 'replace_port'], {}), "(__self__, 'replace_port', replace_port)\n", (125549, 125589), False, 'import pulumi\n'), ((125641, 125695), 'pulumi.set', 'pulumi.set', (['__self__', '"""replace_prefix"""', 'replace_prefix'], {}), "(__self__, 'replace_prefix', replace_prefix)\n", (125651, 125695), False, 'import pulumi\n'), ((125747, 125801), 'pulumi.set', 'pulumi.set', (['__self__', '"""replace_scheme"""', 'replace_scheme'], {}), "(__self__, 'replace_scheme', replace_scheme)\n", (125757, 125801), False, 'import pulumi\n'), ((125852, 125904), 'pulumi.set', 'pulumi.set', (['__self__', '"""response_code"""', 'response_code'], {}), "(__self__, 'response_code', response_code)\n", (125862, 125904), False, 'import pulumi\n'), ((130803, 130851), 'pulumi.set', 'pulumi.set', (['__self__', '"""auto_delete"""', 'auto_delete'], {}), "(__self__, 'auto_delete', auto_delete)\n", (130813, 130851), False, 'import pulumi\n'), ((130900, 130948), 'pulumi.set', 'pulumi.set', (['__self__', '"""device_name"""', 'device_name'], {}), "(__self__, 'device_name', device_name)\n", (130910, 130948), False, 'import pulumi\n'), ((130993, 131033), 'pulumi.set', 'pulumi.set', (['__self__', '"""disk_id"""', 'disk_id'], {}), "(__self__, 'disk_id', disk_id)\n", (131003, 131033), False, 'import pulumi\n'), ((131088, 131148), 'pulumi.set', 'pulumi.set', (['__self__', '"""initialize_params"""', 'initialize_params'], {}), "(__self__, 'initialize_params', initialize_params)\n", (131098, 131148), False, 'import pulumi\n'), ((131190, 131224), 'pulumi.set', 'pulumi.set', (['__self__', '"""mode"""', 'mode'], {}), "(__self__, 'mode', mode)\n", (131200, 131224), False, 'import pulumi\n'), ((134357, 134405), 'pulumi.set', 'pulumi.set', (['__self__', '"""description"""', 'description'], {}), "(__self__, 'description', description)\n", (134367, 134405), False, 'import pulumi\n'), ((134451, 134493), 'pulumi.set', 'pulumi.set', (['__self__', '"""image_id"""', 'image_id'], {}), "(__self__, 'image_id', image_id)\n", (134461, 134493), False, 'import pulumi\n'), ((134535, 134569), 'pulumi.set', 'pulumi.set', (['__self__', '"""name"""', 'name'], {}), "(__self__, 'name', name)\n", (134545, 134569), False, 'import pulumi\n'), ((134611, 134645), 'pulumi.set', 'pulumi.set', (['__self__', '"""size"""', 'size'], {}), "(__self__, 'size', size)\n", (134621, 134645), False, 'import pulumi\n'), ((134694, 134742), 'pulumi.set', 'pulumi.set', (['__self__', '"""snapshot_id"""', 'snapshot_id'], {}), "(__self__, 'snapshot_id', snapshot_id)\n", (134704, 134742), False, 'import pulumi\n'), ((134784, 134818), 'pulumi.set', 'pulumi.set', (['__self__', '"""type"""', 'type'], {}), "(__self__, 'type', type)\n", (134794, 134818), False, 'import pulumi\n'), ((138690, 138776), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_opening_traffic_duration"""', 'max_opening_traffic_duration'], {}), "(__self__, 'max_opening_traffic_duration',\n max_opening_traffic_duration)\n", (138700, 138776), False, 'import pulumi\n'), ((138824, 138878), 'pulumi.set', 'pulumi.set', (['__self__', '"""status_message"""', 'status_message'], {}), "(__self__, 'status_message', status_message)\n", (138834, 138878), False, 'import pulumi\n'), ((138940, 139014), 'pulumi.set', 'pulumi.set', (['__self__', '"""target_group_description"""', 'target_group_description'], {}), "(__self__, 'target_group_description', target_group_description)\n", (138950, 139014), False, 'import pulumi\n'), ((139067, 139123), 'pulumi.set', 'pulumi.set', (['__self__', '"""target_group_id"""', 'target_group_id'], {}), "(__self__, 'target_group_id', target_group_id)\n", (139077, 139123), False, 'import pulumi\n'), ((139180, 139244), 'pulumi.set', 'pulumi.set', (['__self__', '"""target_group_labels"""', 'target_group_labels'], {}), "(__self__, 'target_group_labels', target_group_labels)\n", (139190, 139244), False, 'import pulumi\n'), ((139299, 139359), 'pulumi.set', 'pulumi.set', (['__self__', '"""target_group_name"""', 'target_group_name'], {}), "(__self__, 'target_group_name', target_group_name)\n", (139309, 139359), False, 'import pulumi\n'), ((143887, 143937), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_creating"""', 'max_creating'], {}), "(__self__, 'max_creating', max_creating)\n", (143897, 143937), False, 'import pulumi\n'), ((143987, 144037), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_deleting"""', 'max_deleting'], {}), "(__self__, 'max_deleting', max_deleting)\n", (143997, 144037), False, 'import pulumi\n'), ((144091, 144149), 'pulumi.set', 'pulumi.set', (['__self__', '"""startup_duration"""', 'startup_duration'], {}), "(__self__, 'startup_duration', startup_duration)\n", (144101, 144149), False, 'import pulumi\n'), ((144195, 144237), 'pulumi.set', 'pulumi.set', (['__self__', '"""strategy"""', 'strategy'], {}), "(__self__, 'strategy', strategy)\n", (144205, 144237), False, 'import pulumi\n'), ((148699, 148759), 'pulumi.set', 'pulumi.set', (['__self__', '"""healthy_threshold"""', 'healthy_threshold'], {}), "(__self__, 'healthy_threshold', healthy_threshold)\n", (148709, 148759), False, 'import pulumi\n'), ((148809, 148859), 'pulumi.set', 'pulumi.set', (['__self__', '"""http_options"""', 'http_options'], {}), "(__self__, 'http_options', http_options)\n", (148819, 148859), False, 'import pulumi\n'), ((148905, 148947), 'pulumi.set', 'pulumi.set', (['__self__', '"""interval"""', 'interval'], {}), "(__self__, 'interval', interval)\n", (148915, 148947), False, 'import pulumi\n'), ((148996, 149044), 'pulumi.set', 'pulumi.set', (['__self__', '"""tcp_options"""', 'tcp_options'], {}), "(__self__, 'tcp_options', tcp_options)\n", (149006, 149044), False, 'import pulumi\n'), ((149089, 149129), 'pulumi.set', 'pulumi.set', (['__self__', '"""timeout"""', 'timeout'], {}), "(__self__, 'timeout', timeout)\n", (149099, 149129), False, 'import pulumi\n'), ((149186, 149250), 'pulumi.set', 'pulumi.set', (['__self__', '"""unhealthy_threshold"""', 'unhealthy_threshold'], {}), "(__self__, 'unhealthy_threshold', unhealthy_threshold)\n", (149196, 149250), False, 'import pulumi\n'), ((155498, 155532), 'pulumi.set', 'pulumi.set', (['__self__', '"""fqdn"""', 'fqdn'], {}), "(__self__, 'fqdn', fqdn)\n", (155508, 155532), False, 'import pulumi\n'), ((155581, 155629), 'pulumi.set', 'pulumi.set', (['__self__', '"""instance_id"""', 'instance_id'], {}), "(__self__, 'instance_id', instance_id)\n", (155591, 155629), False, 'import pulumi\n'), ((155671, 155705), 'pulumi.set', 'pulumi.set', (['__self__', '"""name"""', 'name'], {}), "(__self__, 'name', name)\n", (155681, 155705), False, 'import pulumi\n'), ((155761, 155823), 'pulumi.set', 'pulumi.set', (['__self__', '"""network_interfaces"""', 'network_interfaces'], {}), "(__self__, 'network_interfaces', network_interfaces)\n", (155771, 155823), False, 'import pulumi\n'), ((155867, 155905), 'pulumi.set', 'pulumi.set', (['__self__', '"""status"""', 'status'], {}), "(__self__, 'status', status)\n", (155877, 155905), False, 'import pulumi\n'), ((155960, 156020), 'pulumi.set', 'pulumi.set', (['__self__', '"""status_changed_at"""', 'status_changed_at'], {}), "(__self__, 'status_changed_at', status_changed_at)\n", (155970, 156020), False, 'import pulumi\n'), ((156072, 156126), 'pulumi.set', 'pulumi.set', (['__self__', '"""status_message"""', 'status_message'], {}), "(__self__, 'status_message', status_message)\n", (156082, 156126), False, 'import pulumi\n'), ((156171, 156211), 'pulumi.set', 'pulumi.set', (['__self__', '"""zone_id"""', 'zone_id'], {}), "(__self__, 'zone_id', zone_id)\n", (156181, 156211), False, 'import pulumi\n'), ((161346, 161382), 'pulumi.set', 'pulumi.set', (['__self__', '"""index"""', 'index'], {}), "(__self__, 'index', index)\n", (161356, 161382), False, 'import pulumi\n'), ((161430, 161476), 'pulumi.set', 'pulumi.set', (['__self__', '"""ip_address"""', 'ip_address'], {}), "(__self__, 'ip_address', ip_address)\n", (161440, 161476), False, 'import pulumi\n'), ((161518, 161552), 'pulumi.set', 'pulumi.set', (['__self__', '"""ipv4"""', 'ipv4'], {}), "(__self__, 'ipv4', ipv4)\n", (161528, 161552), False, 'import pulumi\n'), ((161594, 161628), 'pulumi.set', 'pulumi.set', (['__self__', '"""ipv6"""', 'ipv6'], {}), "(__self__, 'ipv6', ipv6)\n", (161604, 161628), False, 'import pulumi\n'), ((161678, 161728), 'pulumi.set', 'pulumi.set', (['__self__', '"""ipv6_address"""', 'ipv6_address'], {}), "(__self__, 'ipv6_address', ipv6_address)\n", (161688, 161728), False, 'import pulumi\n'), ((161777, 161825), 'pulumi.set', 'pulumi.set', (['__self__', '"""mac_address"""', 'mac_address'], {}), "(__self__, 'mac_address', mac_address)\n", (161787, 161825), False, 'import pulumi\n'), ((161866, 161898), 'pulumi.set', 'pulumi.set', (['__self__', '"""nat"""', 'nat'], {}), "(__self__, 'nat', nat)\n", (161876, 161898), False, 'import pulumi\n'), ((161950, 162004), 'pulumi.set', 'pulumi.set', (['__self__', '"""nat_ip_address"""', 'nat_ip_address'], {}), "(__self__, 'nat_ip_address', nat_ip_address)\n", (161960, 162004), False, 'import pulumi\n'), ((162056, 162110), 'pulumi.set', 'pulumi.set', (['__self__', '"""nat_ip_version"""', 'nat_ip_version'], {}), "(__self__, 'nat_ip_version', nat_ip_version)\n", (162066, 162110), False, 'import pulumi\n'), ((162157, 162201), 'pulumi.set', 'pulumi.set', (['__self__', '"""subnet_id"""', 'subnet_id'], {}), "(__self__, 'subnet_id', subnet_id)\n", (162167, 162201), False, 'import pulumi\n'), ((170825, 170873), 'pulumi.set', 'pulumi.set', (['__self__', '"""description"""', 'description'], {}), "(__self__, 'description', description)\n", (170835, 170873), False, 'import pulumi\n'), ((170919, 170961), 'pulumi.set', 'pulumi.set', (['__self__', '"""hostname"""', 'hostname'], {}), "(__self__, 'hostname', hostname)\n", (170929, 170961), False, 'import pulumi\n'), ((171005, 171043), 'pulumi.set', 'pulumi.set', (['__self__', '"""labels"""', 'labels'], {}), "(__self__, 'labels', labels)\n", (171015, 171043), False, 'import pulumi\n'), ((171089, 171131), 'pulumi.set', 'pulumi.set', (['__self__', '"""metadata"""', 'metadata'], {}), "(__self__, 'metadata', metadata)\n", (171099, 171131), False, 'import pulumi\n'), ((171173, 171207), 'pulumi.set', 'pulumi.set', (['__self__', '"""name"""', 'name'], {}), "(__self__, 'name', name)\n", (171183, 171207), False, 'import pulumi\n'), ((171261, 171319), 'pulumi.set', 'pulumi.set', (['__self__', '"""network_settings"""', 'network_settings'], {}), "(__self__, 'network_settings', network_settings)\n", (171271, 171319), False, 'import pulumi\n'), ((171373, 171431), 'pulumi.set', 'pulumi.set', (['__self__', '"""placement_policy"""', 'placement_policy'], {}), "(__self__, 'placement_policy', placement_policy)\n", (171383, 171431), False, 'import pulumi\n'), ((171480, 171528), 'pulumi.set', 'pulumi.set', (['__self__', '"""platform_id"""', 'platform_id'], {}), "(__self__, 'platform_id', platform_id)\n", (171490, 171528), False, 'import pulumi\n'), ((171583, 171643), 'pulumi.set', 'pulumi.set', (['__self__', '"""scheduling_policy"""', 'scheduling_policy'], {}), "(__self__, 'scheduling_policy', scheduling_policy)\n", (171593, 171643), False, 'import pulumi\n'), ((171696, 171752), 'pulumi.set', 'pulumi.set', (['__self__', '"""secondary_disks"""', 'secondary_disks'], {}), "(__self__, 'secondary_disks', secondary_disks)\n", (171706, 171752), False, 'import pulumi\n'), ((171808, 171870), 'pulumi.set', 'pulumi.set', (['__self__', '"""service_account_id"""', 'service_account_id'], {}), "(__self__, 'service_account_id', service_account_id)\n", (171818, 171870), False, 'import pulumi\n'), ((180452, 180500), 'pulumi.set', 'pulumi.set', (['__self__', '"""device_name"""', 'device_name'], {}), "(__self__, 'device_name', device_name)\n", (180462, 180500), False, 'import pulumi\n'), ((180545, 180585), 'pulumi.set', 'pulumi.set', (['__self__', '"""disk_id"""', 'disk_id'], {}), "(__self__, 'disk_id', disk_id)\n", (180555, 180585), False, 'import pulumi\n'), ((180640, 180700), 'pulumi.set', 'pulumi.set', (['__self__', '"""initialize_params"""', 'initialize_params'], {}), "(__self__, 'initialize_params', initialize_params)\n", (180650, 180700), False, 'import pulumi\n'), ((180742, 180776), 'pulumi.set', 'pulumi.set', (['__self__', '"""mode"""', 'mode'], {}), "(__self__, 'mode', mode)\n", (180752, 180776), False, 'import pulumi\n'), ((183421, 183469), 'pulumi.set', 'pulumi.set', (['__self__', '"""description"""', 'description'], {}), "(__self__, 'description', description)\n", (183431, 183469), False, 'import pulumi\n'), ((183515, 183557), 'pulumi.set', 'pulumi.set', (['__self__', '"""image_id"""', 'image_id'], {}), "(__self__, 'image_id', image_id)\n", (183525, 183557), False, 'import pulumi\n'), ((183599, 183633), 'pulumi.set', 'pulumi.set', (['__self__', '"""size"""', 'size'], {}), "(__self__, 'size', size)\n", (183609, 183633), False, 'import pulumi\n'), ((183682, 183730), 'pulumi.set', 'pulumi.set', (['__self__', '"""snapshot_id"""', 'snapshot_id'], {}), "(__self__, 'snapshot_id', snapshot_id)\n", (183692, 183730), False, 'import pulumi\n'), ((183772, 183806), 'pulumi.set', 'pulumi.set', (['__self__', '"""type"""', 'type'], {}), "(__self__, 'type', type)\n", (183782, 183806), False, 'import pulumi\n'), ((188178, 188226), 'pulumi.set', 'pulumi.set', (['__self__', '"""dns_records"""', 'dns_records'], {}), "(__self__, 'dns_records', dns_records)\n", (188188, 188226), False, 'import pulumi\n'), ((188274, 188320), 'pulumi.set', 'pulumi.set', (['__self__', '"""ip_address"""', 'ip_address'], {}), "(__self__, 'ip_address', ip_address)\n", (188284, 188320), False, 'import pulumi\n'), ((188362, 188396), 'pulumi.set', 'pulumi.set', (['__self__', '"""ipv4"""', 'ipv4'], {}), "(__self__, 'ipv4', ipv4)\n", (188372, 188396), False, 'import pulumi\n'), ((188438, 188472), 'pulumi.set', 'pulumi.set', (['__self__', '"""ipv6"""', 'ipv6'], {}), "(__self__, 'ipv6', ipv6)\n", (188448, 188472), False, 'import pulumi\n'), ((188522, 188572), 'pulumi.set', 'pulumi.set', (['__self__', '"""ipv6_address"""', 'ipv6_address'], {}), "(__self__, 'ipv6_address', ipv6_address)\n", (188532, 188572), False, 'import pulumi\n'), ((188626, 188684), 'pulumi.set', 'pulumi.set', (['__self__', '"""ipv6_dns_records"""', 'ipv6_dns_records'], {}), "(__self__, 'ipv6_dns_records', ipv6_dns_records)\n", (188636, 188684), False, 'import pulumi\n'), ((188725, 188757), 'pulumi.set', 'pulumi.set', (['__self__', '"""nat"""', 'nat'], {}), "(__self__, 'nat', nat)\n", (188735, 188757), False, 'import pulumi\n'), ((188810, 188866), 'pulumi.set', 'pulumi.set', (['__self__', '"""nat_dns_records"""', 'nat_dns_records'], {}), "(__self__, 'nat_dns_records', nat_dns_records)\n", (188820, 188866), False, 'import pulumi\n'), ((188918, 188972), 'pulumi.set', 'pulumi.set', (['__self__', '"""nat_ip_address"""', 'nat_ip_address'], {}), "(__self__, 'nat_ip_address', nat_ip_address)\n", (188928, 188972), False, 'import pulumi\n'), ((189020, 189066), 'pulumi.set', 'pulumi.set', (['__self__', '"""network_id"""', 'network_id'], {}), "(__self__, 'network_id', network_id)\n", (189030, 189066), False, 'import pulumi\n'), ((189122, 189184), 'pulumi.set', 'pulumi.set', (['__self__', '"""security_group_ids"""', 'security_group_ids'], {}), "(__self__, 'security_group_ids', security_group_ids)\n", (189132, 189184), False, 'import pulumi\n'), ((189232, 189278), 'pulumi.set', 'pulumi.set', (['__self__', '"""subnet_ids"""', 'subnet_ids'], {}), "(__self__, 'subnet_ids', subnet_ids)\n", (189242, 189278), False, 'import pulumi\n'), ((195027, 195075), 'pulumi.set', 'pulumi.set', (['__self__', '"""dns_zone_id"""', 'dns_zone_id'], {}), "(__self__, 'dns_zone_id', dns_zone_id)\n", (195037, 195075), False, 'import pulumi\n'), ((195116, 195148), 'pulumi.set', 'pulumi.set', (['__self__', '"""ptr"""', 'ptr'], {}), "(__self__, 'ptr', ptr)\n", (195126, 195148), False, 'import pulumi\n'), ((195189, 195221), 'pulumi.set', 'pulumi.set', (['__self__', '"""ttl"""', 'ttl'], {}), "(__self__, 'ttl', ttl)\n", (195199, 195221), False, 'import pulumi\n'), ((197264, 197312), 'pulumi.set', 'pulumi.set', (['__self__', '"""dns_zone_id"""', 'dns_zone_id'], {}), "(__self__, 'dns_zone_id', dns_zone_id)\n", (197274, 197312), False, 'import pulumi\n'), ((197353, 197385), 'pulumi.set', 'pulumi.set', (['__self__', '"""ptr"""', 'ptr'], {}), "(__self__, 'ptr', ptr)\n", (197363, 197385), False, 'import pulumi\n'), ((197426, 197458), 'pulumi.set', 'pulumi.set', (['__self__', '"""ttl"""', 'ttl'], {}), "(__self__, 'ttl', ttl)\n", (197436, 197458), False, 'import pulumi\n'), ((199500, 199548), 'pulumi.set', 'pulumi.set', (['__self__', '"""dns_zone_id"""', 'dns_zone_id'], {}), "(__self__, 'dns_zone_id', dns_zone_id)\n", (199510, 199548), False, 'import pulumi\n'), ((199589, 199621), 'pulumi.set', 'pulumi.set', (['__self__', '"""ptr"""', 'ptr'], {}), "(__self__, 'ptr', ptr)\n", (199599, 199621), False, 'import pulumi\n'), ((199662, 199694), 'pulumi.set', 'pulumi.set', (['__self__', '"""ttl"""', 'ttl'], {}), "(__self__, 'ttl', ttl)\n", (199672, 199694), False, 'import pulumi\n'), ((201303, 201337), 'pulumi.set', 'pulumi.set', (['__self__', '"""type"""', 'type'], {}), "(__self__, 'type', type)\n", (201313, 201337), False, 'import pulumi\n'), ((203231, 203283), 'pulumi.set', 'pulumi.set', (['__self__', '"""core_fraction"""', 'core_fraction'], {}), "(__self__, 'core_fraction', core_fraction)\n", (203241, 203283), False, 'import pulumi\n'), ((203325, 203359), 'pulumi.set', 'pulumi.set', (['__self__', '"""gpus"""', 'gpus'], {}), "(__self__, 'gpus', gpus)\n", (203335, 203359), False, 'import pulumi\n'), ((204948, 204996), 'pulumi.set', 'pulumi.set', (['__self__', '"""preemptible"""', 'preemptible'], {}), "(__self__, 'preemptible', preemptible)\n", (204958, 204996), False, 'import pulumi\n'), ((206425, 206473), 'pulumi.set', 'pulumi.set', (['__self__', '"""device_name"""', 'device_name'], {}), "(__self__, 'device_name', device_name)\n", (206435, 206473), False, 'import pulumi\n'), ((206518, 206558), 'pulumi.set', 'pulumi.set', (['__self__', '"""disk_id"""', 'disk_id'], {}), "(__self__, 'disk_id', disk_id)\n", (206528, 206558), False, 'import pulumi\n'), ((206613, 206673), 'pulumi.set', 'pulumi.set', (['__self__', '"""initialize_params"""', 'initialize_params'], {}), "(__self__, 'initialize_params', initialize_params)\n", (206623, 206673), False, 'import pulumi\n'), ((206715, 206749), 'pulumi.set', 'pulumi.set', (['__self__', '"""mode"""', 'mode'], {}), "(__self__, 'mode', mode)\n", (206725, 206749), False, 'import pulumi\n'), ((209409, 209457), 'pulumi.set', 'pulumi.set', (['__self__', '"""description"""', 'description'], {}), "(__self__, 'description', description)\n", (209419, 209457), False, 'import pulumi\n'), ((209503, 209545), 'pulumi.set', 'pulumi.set', (['__self__', '"""image_id"""', 'image_id'], {}), "(__self__, 'image_id', image_id)\n", (209513, 209545), False, 'import pulumi\n'), ((209587, 209621), 'pulumi.set', 'pulumi.set', (['__self__', '"""size"""', 'size'], {}), "(__self__, 'size', size)\n", (209597, 209621), False, 'import pulumi\n'), ((209670, 209718), 'pulumi.set', 'pulumi.set', (['__self__', '"""snapshot_id"""', 'snapshot_id'], {}), "(__self__, 'snapshot_id', snapshot_id)\n", (209680, 209718), False, 'import pulumi\n'), ((209760, 209794), 'pulumi.set', 'pulumi.set', (['__self__', '"""type"""', 'type'], {}), "(__self__, 'type', type)\n", (209770, 209794), False, 'import pulumi\n'), ((212786, 212872), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_opening_traffic_duration"""', 'max_opening_traffic_duration'], {}), "(__self__, 'max_opening_traffic_duration',\n max_opening_traffic_duration)\n", (212796, 212872), False, 'import pulumi\n'), ((212920, 212974), 'pulumi.set', 'pulumi.set', (['__self__', '"""status_message"""', 'status_message'], {}), "(__self__, 'status_message', status_message)\n", (212930, 212974), False, 'import pulumi\n'), ((213036, 213110), 'pulumi.set', 'pulumi.set', (['__self__', '"""target_group_description"""', 'target_group_description'], {}), "(__self__, 'target_group_description', target_group_description)\n", (213046, 213110), False, 'import pulumi\n'), ((213163, 213219), 'pulumi.set', 'pulumi.set', (['__self__', '"""target_group_id"""', 'target_group_id'], {}), "(__self__, 'target_group_id', target_group_id)\n", (213173, 213219), False, 'import pulumi\n'), ((213276, 213340), 'pulumi.set', 'pulumi.set', (['__self__', '"""target_group_labels"""', 'target_group_labels'], {}), "(__self__, 'target_group_labels', target_group_labels)\n", (213286, 213340), False, 'import pulumi\n'), ((213395, 213455), 'pulumi.set', 'pulumi.set', (['__self__', '"""target_group_name"""', 'target_group_name'], {}), "(__self__, 'target_group_name', target_group_name)\n", (213405, 213455), False, 'import pulumi\n'), ((217107, 217153), 'pulumi.set', 'pulumi.set', (['__self__', '"""auto_scale"""', 'auto_scale'], {}), "(__self__, 'auto_scale', auto_scale)\n", (217117, 217153), False, 'import pulumi\n'), ((217202, 217250), 'pulumi.set', 'pulumi.set', (['__self__', '"""fixed_scale"""', 'fixed_scale'], {}), "(__self__, 'fixed_scale', fixed_scale)\n", (217212, 217250), False, 'import pulumi\n'), ((217303, 217359), 'pulumi.set', 'pulumi.set', (['__self__', '"""test_auto_scale"""', 'test_auto_scale'], {}), "(__self__, 'test_auto_scale', test_auto_scale)\n", (217313, 217359), False, 'import pulumi\n'), ((221312, 221382), 'pulumi.set', 'pulumi.set', (['__self__', '"""cpu_utilization_target"""', 'cpu_utilization_target'], {}), "(__self__, 'cpu_utilization_target', cpu_utilization_target)\n", (221322, 221382), False, 'import pulumi\n'), ((221432, 221482), 'pulumi.set', 'pulumi.set', (['__self__', '"""custom_rules"""', 'custom_rules'], {}), "(__self__, 'custom_rules', custom_rules)\n", (221442, 221482), False, 'import pulumi\n'), ((221528, 221570), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_size"""', 'max_size'], {}), "(__self__, 'max_size', max_size)\n", (221538, 221570), False, 'import pulumi\n'), ((221621, 221673), 'pulumi.set', 'pulumi.set', (['__self__', '"""min_zone_size"""', 'min_zone_size'], {}), "(__self__, 'min_zone_size', min_zone_size)\n", (221631, 221673), False, 'import pulumi\n'), ((221733, 221803), 'pulumi.set', 'pulumi.set', (['__self__', '"""stabilization_duration"""', 'stabilization_duration'], {}), "(__self__, 'stabilization_duration', stabilization_duration)\n", (221743, 221803), False, 'import pulumi\n'), ((221856, 221912), 'pulumi.set', 'pulumi.set', (['__self__', '"""warmup_duration"""', 'warmup_duration'], {}), "(__self__, 'warmup_duration', warmup_duration)\n", (221866, 221912), False, 'import pulumi\n'), ((227688, 227732), 'pulumi.set', 'pulumi.set', (['__self__', '"""folder_id"""', 'folder_id'], {}), "(__self__, 'folder_id', folder_id)\n", (227698, 227732), False, 'import pulumi\n'), ((227776, 227814), 'pulumi.set', 'pulumi.set', (['__self__', '"""labels"""', 'labels'], {}), "(__self__, 'labels', labels)\n", (227786, 227814), False, 'import pulumi\n'), ((227859, 227899), 'pulumi.set', 'pulumi.set', (['__self__', '"""service"""', 'service'], {}), "(__self__, 'service', service)\n", (227869, 227899), False, 'import pulumi\n'), ((233739, 233809), 'pulumi.set', 'pulumi.set', (['__self__', '"""cpu_utilization_target"""', 'cpu_utilization_target'], {}), "(__self__, 'cpu_utilization_target', cpu_utilization_target)\n", (233749, 233809), False, 'import pulumi\n'), ((233859, 233909), 'pulumi.set', 'pulumi.set', (['__self__', '"""custom_rules"""', 'custom_rules'], {}), "(__self__, 'custom_rules', custom_rules)\n", (233869, 233909), False, 'import pulumi\n'), ((233955, 233997), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_size"""', 'max_size'], {}), "(__self__, 'max_size', max_size)\n", (233965, 233997), False, 'import pulumi\n'), ((234048, 234100), 'pulumi.set', 'pulumi.set', (['__self__', '"""min_zone_size"""', 'min_zone_size'], {}), "(__self__, 'min_zone_size', min_zone_size)\n", (234058, 234100), False, 'import pulumi\n'), ((234160, 234230), 'pulumi.set', 'pulumi.set', (['__self__', '"""stabilization_duration"""', 'stabilization_duration'], {}), "(__self__, 'stabilization_duration', stabilization_duration)\n", (234170, 234230), False, 'import pulumi\n'), ((234283, 234339), 'pulumi.set', 'pulumi.set', (['__self__', '"""warmup_duration"""', 'warmup_duration'], {}), "(__self__, 'warmup_duration', warmup_duration)\n", (234293, 234339), False, 'import pulumi\n'), ((240127, 240171), 'pulumi.set', 'pulumi.set', (['__self__', '"""folder_id"""', 'folder_id'], {}), "(__self__, 'folder_id', folder_id)\n", (240137, 240171), False, 'import pulumi\n'), ((240215, 240253), 'pulumi.set', 'pulumi.set', (['__self__', '"""labels"""', 'labels'], {}), "(__self__, 'labels', labels)\n", (240225, 240253), False, 'import pulumi\n'), ((240298, 240338), 'pulumi.set', 'pulumi.set', (['__self__', '"""service"""', 'service'], {}), "(__self__, 'service', service)\n", (240308, 240338), False, 'import pulumi\n'), ((246324, 246372), 'pulumi.set', 'pulumi.set', (['__self__', '"""dns_records"""', 'dns_records'], {}), "(__self__, 'dns_records', dns_records)\n", (246334, 246372), False, 'import pulumi\n'), ((246415, 246451), 'pulumi.set', 'pulumi.set', (['__self__', '"""index"""', 'index'], {}), "(__self__, 'index', index)\n", (246425, 246451), False, 'import pulumi\n'), ((246499, 246545), 'pulumi.set', 'pulumi.set', (['__self__', '"""ip_address"""', 'ip_address'], {}), "(__self__, 'ip_address', ip_address)\n", (246509, 246545), False, 'import pulumi\n'), ((246587, 246621), 'pulumi.set', 'pulumi.set', (['__self__', '"""ipv4"""', 'ipv4'], {}), "(__self__, 'ipv4', ipv4)\n", (246597, 246621), False, 'import pulumi\n'), ((246663, 246697), 'pulumi.set', 'pulumi.set', (['__self__', '"""ipv6"""', 'ipv6'], {}), "(__self__, 'ipv6', ipv6)\n", (246673, 246697), False, 'import pulumi\n'), ((246747, 246797), 'pulumi.set', 'pulumi.set', (['__self__', '"""ipv6_address"""', 'ipv6_address'], {}), "(__self__, 'ipv6_address', ipv6_address)\n", (246757, 246797), False, 'import pulumi\n'), ((246851, 246909), 'pulumi.set', 'pulumi.set', (['__self__', '"""ipv6_dns_records"""', 'ipv6_dns_records'], {}), "(__self__, 'ipv6_dns_records', ipv6_dns_records)\n", (246861, 246909), False, 'import pulumi\n'), ((246958, 247006), 'pulumi.set', 'pulumi.set', (['__self__', '"""mac_address"""', 'mac_address'], {}), "(__self__, 'mac_address', mac_address)\n", (246968, 247006), False, 'import pulumi\n'), ((247047, 247079), 'pulumi.set', 'pulumi.set', (['__self__', '"""nat"""', 'nat'], {}), "(__self__, 'nat', nat)\n", (247057, 247079), False, 'import pulumi\n'), ((247132, 247188), 'pulumi.set', 'pulumi.set', (['__self__', '"""nat_dns_records"""', 'nat_dns_records'], {}), "(__self__, 'nat_dns_records', nat_dns_records)\n", (247142, 247188), False, 'import pulumi\n'), ((247240, 247294), 'pulumi.set', 'pulumi.set', (['__self__', '"""nat_ip_address"""', 'nat_ip_address'], {}), "(__self__, 'nat_ip_address', nat_ip_address)\n", (247250, 247294), False, 'import pulumi\n'), ((247346, 247400), 'pulumi.set', 'pulumi.set', (['__self__', '"""nat_ip_version"""', 'nat_ip_version'], {}), "(__self__, 'nat_ip_version', nat_ip_version)\n", (247356, 247400), False, 'import pulumi\n'), ((247456, 247518), 'pulumi.set', 'pulumi.set', (['__self__', '"""security_group_ids"""', 'security_group_ids'], {}), "(__self__, 'security_group_ids', security_group_ids)\n", (247466, 247518), False, 'import pulumi\n'), ((254131, 254179), 'pulumi.set', 'pulumi.set', (['__self__', '"""dns_zone_id"""', 'dns_zone_id'], {}), "(__self__, 'dns_zone_id', dns_zone_id)\n", (254141, 254179), False, 'import pulumi\n'), ((254220, 254252), 'pulumi.set', 'pulumi.set', (['__self__', '"""ptr"""', 'ptr'], {}), "(__self__, 'ptr', ptr)\n", (254230, 254252), False, 'import pulumi\n'), ((254293, 254325), 'pulumi.set', 'pulumi.set', (['__self__', '"""ttl"""', 'ttl'], {}), "(__self__, 'ttl', ttl)\n", (254303, 254325), False, 'import pulumi\n'), ((256377, 256425), 'pulumi.set', 'pulumi.set', (['__self__', '"""dns_zone_id"""', 'dns_zone_id'], {}), "(__self__, 'dns_zone_id', dns_zone_id)\n", (256387, 256425), False, 'import pulumi\n'), ((256466, 256498), 'pulumi.set', 'pulumi.set', (['__self__', '"""ptr"""', 'ptr'], {}), "(__self__, 'ptr', ptr)\n", (256476, 256498), False, 'import pulumi\n'), ((256539, 256571), 'pulumi.set', 'pulumi.set', (['__self__', '"""ttl"""', 'ttl'], {}), "(__self__, 'ttl', ttl)\n", (256549, 256571), False, 'import pulumi\n'), ((258622, 258670), 'pulumi.set', 'pulumi.set', (['__self__', '"""dns_zone_id"""', 'dns_zone_id'], {}), "(__self__, 'dns_zone_id', dns_zone_id)\n", (258632, 258670), False, 'import pulumi\n'), ((258711, 258743), 'pulumi.set', 'pulumi.set', (['__self__', '"""ptr"""', 'ptr'], {}), "(__self__, 'ptr', ptr)\n", (258721, 258743), False, 'import pulumi\n'), ((258784, 258816), 'pulumi.set', 'pulumi.set', (['__self__', '"""ttl"""', 'ttl'], {}), "(__self__, 'ttl', ttl)\n", (258794, 258816), False, 'import pulumi\n'), ((261587, 261639), 'pulumi.set', 'pulumi.set', (['__self__', '"""core_fraction"""', 'core_fraction'], {}), "(__self__, 'core_fraction', core_fraction)\n", (261597, 261639), False, 'import pulumi\n'), ((261681, 261715), 'pulumi.set', 'pulumi.set', (['__self__', '"""gpus"""', 'gpus'], {}), "(__self__, 'gpus', gpus)\n", (261691, 261715), False, 'import pulumi\n'), ((263271, 263319), 'pulumi.set', 'pulumi.set', (['__self__', '"""preemptible"""', 'preemptible'], {}), "(__self__, 'preemptible', preemptible)\n", (263281, 263319), False, 'import pulumi\n'), ((264648, 264696), 'pulumi.set', 'pulumi.set', (['__self__', '"""auto_delete"""', 'auto_delete'], {}), "(__self__, 'auto_delete', auto_delete)\n", (264658, 264696), False, 'import pulumi\n'), ((264745, 264793), 'pulumi.set', 'pulumi.set', (['__self__', '"""device_name"""', 'device_name'], {}), "(__self__, 'device_name', device_name)\n", (264755, 264793), False, 'import pulumi\n'), ((264835, 264869), 'pulumi.set', 'pulumi.set', (['__self__', '"""mode"""', 'mode'], {}), "(__self__, 'mode', mode)\n", (264845, 264869), False, 'import pulumi\n'), ((267319, 267357), 'pulumi.set', 'pulumi.set', (['__self__', '"""hadoop"""', 'hadoop'], {}), "(__self__, 'hadoop', hadoop)\n", (267329, 267357), False, 'import pulumi\n'), ((267405, 267451), 'pulumi.set', 'pulumi.set', (['__self__', '"""version_id"""', 'version_id'], {}), "(__self__, 'version_id', version_id)\n", (267415, 267451), False, 'import pulumi\n'), ((269745, 269791), 'pulumi.set', 'pulumi.set', (['__self__', '"""properties"""', 'properties'], {}), "(__self__, 'properties', properties)\n", (269755, 269791), False, 'import pulumi\n'), ((269837, 269879), 'pulumi.set', 'pulumi.set', (['__self__', '"""services"""', 'services'], {}), "(__self__, 'services', services)\n", (269847, 269879), False, 'import pulumi\n'), ((269932, 269988), 'pulumi.set', 'pulumi.set', (['__self__', '"""ssh_public_keys"""', 'ssh_public_keys'], {}), "(__self__, 'ssh_public_keys', ssh_public_keys)\n", (269942, 269988), False, 'import pulumi\n'), ((273175, 273237), 'pulumi.set', 'pulumi.set', (['__self__', '"""autoscaling_config"""', 'autoscaling_config'], {}), "(__self__, 'autoscaling_config', autoscaling_config)\n", (273185, 273237), False, 'import pulumi\n'), ((273277, 273307), 'pulumi.set', 'pulumi.set', (['__self__', '"""id"""', 'id'], {}), "(__self__, 'id', id)\n", (273287, 273307), False, 'import pulumi\n'), ((278134, 278204), 'pulumi.set', 'pulumi.set', (['__self__', '"""cpu_utilization_target"""', 'cpu_utilization_target'], {}), "(__self__, 'cpu_utilization_target', cpu_utilization_target)\n", (278144, 278204), False, 'import pulumi\n'), ((278262, 278328), 'pulumi.set', 'pulumi.set', (['__self__', '"""decommission_timeout"""', 'decommission_timeout'], {}), "(__self__, 'decommission_timeout', decommission_timeout)\n", (278272, 278328), False, 'import pulumi\n'), ((278386, 278452), 'pulumi.set', 'pulumi.set', (['__self__', '"""measurement_duration"""', 'measurement_duration'], {}), "(__self__, 'measurement_duration', measurement_duration)\n", (278396, 278452), False, 'import pulumi\n'), ((278501, 278549), 'pulumi.set', 'pulumi.set', (['__self__', '"""preemptible"""', 'preemptible'], {}), "(__self__, 'preemptible', preemptible)\n", (278511, 278549), False, 'import pulumi\n'), ((278609, 278679), 'pulumi.set', 'pulumi.set', (['__self__', '"""stabilization_duration"""', 'stabilization_duration'], {}), "(__self__, 'stabilization_duration', stabilization_duration)\n", (278619, 278679), False, 'import pulumi\n'), ((278732, 278788), 'pulumi.set', 'pulumi.set', (['__self__', '"""warmup_duration"""', 'warmup_duration'], {}), "(__self__, 'warmup_duration', warmup_duration)\n", (278742, 278788), False, 'import pulumi\n'), ((283273, 283323), 'pulumi.set', 'pulumi.set', (['__self__', '"""disk_type_id"""', 'disk_type_id'], {}), "(__self__, 'disk_type_id', disk_type_id)\n", (283283, 283323), False, 'import pulumi\n'), ((285484, 285522), 'pulumi.set', 'pulumi.set', (['__self__', '"""sha256"""', 'sha256'], {}), "(__self__, 'sha256', sha256)\n", (285494, 285522), False, 'import pulumi\n'), ((286697, 286763), 'pulumi.set', 'pulumi.set', (['__self__', '"""zone_instances_limit"""', 'zone_instances_limit'], {}), "(__self__, 'zone_instances_limit', zone_instances_limit)\n", (286707, 286763), False, 'import pulumi\n'), ((286820, 286884), 'pulumi.set', 'pulumi.set', (['__self__', '"""zone_requests_limit"""', 'zone_requests_limit'], {}), "(__self__, 'zone_requests_limit', zone_requests_limit)\n", (286830, 286884), False, 'import pulumi\n'), ((289146, 289200), 'pulumi.set', 'pulumi.set', (['__self__', '"""retry_attempts"""', 'retry_attempts'], {}), "(__self__, 'retry_attempts', retry_attempts)\n", (289156, 289200), False, 'import pulumi\n'), ((289252, 289306), 'pulumi.set', 'pulumi.set', (['__self__', '"""retry_interval"""', 'retry_interval'], {}), "(__self__, 'retry_interval', retry_interval)\n", (289262, 289306), False, 'import pulumi\n'), ((289362, 289424), 'pulumi.set', 'pulumi.set', (['__self__', '"""service_account_id"""', 'service_account_id'], {}), "(__self__, 'service_account_id', service_account_id)\n", (289372, 289424), False, 'import pulumi\n'), ((289465, 289497), 'pulumi.set', 'pulumi.set', (['__self__', '"""tag"""', 'tag'], {}), "(__self__, 'tag', tag)\n", (289475, 289497), False, 'import pulumi\n'), ((291255, 291299), 'pulumi.set', 'pulumi.set', (['__self__', '"""device_id"""', 'device_id'], {}), "(__self__, 'device_id', device_id)\n", (291265, 291299), False, 'import pulumi\n'), ((291342, 291378), 'pulumi.set', 'pulumi.set', (['__self__', '"""topic"""', 'topic'], {}), "(__self__, 'topic', topic)\n", (291352, 291378), False, 'import pulumi\n'), ((292621, 292667), 'pulumi.set', 'pulumi.set', (['__self__', '"""batch_size"""', 'batch_size'], {}), "(__self__, 'batch_size', batch_size)\n", (292631, 292667), False, 'import pulumi\n'), ((294357, 294403), 'pulumi.set', 'pulumi.set', (['__self__', '"""batch_size"""', 'batch_size'], {}), "(__self__, 'batch_size', batch_size)\n", (294367, 294403), False, 'import pulumi\n'), ((296777, 296823), 'pulumi.set', 'pulumi.set', (['__self__', '"""batch_size"""', 'batch_size'], {}), "(__self__, 'batch_size', batch_size)\n", (296787, 296823), False, 'import pulumi\n'), ((296879, 296941), 'pulumi.set', 'pulumi.set', (['__self__', '"""visibility_timeout"""', 'visibility_timeout'], {}), "(__self__, 'visibility_timeout', visibility_timeout)\n", (296889, 296941), False, 'import pulumi\n'), ((298955, 298993), 'pulumi.set', 'pulumi.set', (['__self__', '"""create"""', 'create'], {}), "(__self__, 'create', create)\n", (298965, 298993), False, 'import pulumi\n'), ((299037, 299075), 'pulumi.set', 'pulumi.set', (['__self__', '"""delete"""', 'delete'], {}), "(__self__, 'delete', delete)\n", (299047, 299075), False, 'import pulumi\n'), ((299119, 299157), 'pulumi.set', 'pulumi.set', (['__self__', '"""prefix"""', 'prefix'], {}), "(__self__, 'prefix', prefix)\n", (299129, 299157), False, 'import pulumi\n'), ((299201, 299239), 'pulumi.set', 'pulumi.set', (['__self__', '"""suffix"""', 'suffix'], {}), "(__self__, 'suffix', suffix)\n", (299211, 299239), False, 'import pulumi\n'), ((299283, 299321), 'pulumi.set', 'pulumi.set', (['__self__', '"""update"""', 'update'], {}), "(__self__, 'update', update)\n", (299293, 299321), False, 'import pulumi\n'), ((301606, 301644), 'pulumi.set', 'pulumi.set', (['__self__', '"""key_id"""', 'key_id'], {}), "(__self__, 'key_id', key_id)\n", (301616, 301644), False, 'import pulumi\n'), ((305323, 305393), 'pulumi.set', 'pulumi.set', (['__self__', '"""cluster_ca_certificate"""', 'cluster_ca_certificate'], {}), "(__self__, 'cluster_ca_certificate', cluster_ca_certificate)\n", (305333, 305393), False, 'import pulumi\n'), ((305450, 305514), 'pulumi.set', 'pulumi.set', (['__self__', '"""external_v4_address"""', 'external_v4_address'], {}), "(__self__, 'external_v4_address', external_v4_address)\n", (305460, 305514), False, 'import pulumi\n'), ((305572, 305638), 'pulumi.set', 'pulumi.set', (['__self__', '"""external_v4_endpoint"""', 'external_v4_endpoint'], {}), "(__self__, 'external_v4_endpoint', external_v4_endpoint)\n", (305582, 305638), False, 'import pulumi\n'), ((305695, 305759), 'pulumi.set', 'pulumi.set', (['__self__', '"""internal_v4_address"""', 'internal_v4_address'], {}), "(__self__, 'internal_v4_address', internal_v4_address)\n", (305705, 305759), False, 'import pulumi\n'), ((305817, 305883), 'pulumi.set', 'pulumi.set', (['__self__', '"""internal_v4_endpoint"""', 'internal_v4_endpoint'], {}), "(__self__, 'internal_v4_endpoint', internal_v4_endpoint)\n", (305827, 305883), False, 'import pulumi\n'), ((305939, 306001), 'pulumi.set', 'pulumi.set', (['__self__', '"""maintenance_policy"""', 'maintenance_policy'], {}), "(__self__, 'maintenance_policy', maintenance_policy)\n", (305949, 306001), False, 'import pulumi\n'), ((306048, 306092), 'pulumi.set', 'pulumi.set', (['__self__', '"""public_ip"""', 'public_ip'], {}), "(__self__, 'public_ip', public_ip)\n", (306058, 306092), False, 'import pulumi\n'), ((306138, 306180), 'pulumi.set', 'pulumi.set', (['__self__', '"""regional"""', 'regional'], {}), "(__self__, 'regional', regional)\n", (306148, 306180), False, 'import pulumi\n'), ((306236, 306298), 'pulumi.set', 'pulumi.set', (['__self__', '"""security_group_ids"""', 'security_group_ids'], {}), "(__self__, 'security_group_ids', security_group_ids)\n", (306246, 306298), False, 'import pulumi\n'), ((306343, 306383), 'pulumi.set', 'pulumi.set', (['__self__', '"""version"""', 'version'], {}), "(__self__, 'version', version)\n", (306353, 306383), False, 'import pulumi\n'), ((306433, 306483), 'pulumi.set', 'pulumi.set', (['__self__', '"""version_info"""', 'version_info'], {}), "(__self__, 'version_info', version_info)\n", (306443, 306483), False, 'import pulumi\n'), ((306526, 306562), 'pulumi.set', 'pulumi.set', (['__self__', '"""zonal"""', 'zonal'], {}), "(__self__, 'zonal', zonal)\n", (306536, 306562), False, 'import pulumi\n'), ((313541, 313605), 'pulumi.set', 'pulumi.set', (['__self__', '"""maintenance_windows"""', 'maintenance_windows'], {}), "(__self__, 'maintenance_windows', maintenance_windows)\n", (313551, 313605), False, 'import pulumi\n'), ((315339, 315371), 'pulumi.set', 'pulumi.set', (['__self__', '"""day"""', 'day'], {}), "(__self__, 'day', day)\n", (315349, 315371), False, 'import pulumi\n'), ((316849, 316893), 'pulumi.set', 'pulumi.set', (['__self__', '"""locations"""', 'locations'], {}), "(__self__, 'locations', locations)\n", (316859, 316893), False, 'import pulumi\n'), ((318231, 318275), 'pulumi.set', 'pulumi.set', (['__self__', '"""subnet_id"""', 'subnet_id'], {}), "(__self__, 'subnet_id', subnet_id)\n", (318241, 318275), False, 'import pulumi\n'), ((318317, 318351), 'pulumi.set', 'pulumi.set', (['__self__', '"""zone"""', 'zone'], {}), "(__self__, 'zone', zone)\n", (318327, 318351), False, 'import pulumi\n'), ((320291, 320347), 'pulumi.set', 'pulumi.set', (['__self__', '"""current_version"""', 'current_version'], {}), "(__self__, 'current_version', current_version)\n", (320301, 320347), False, 'import pulumi\n'), ((320407, 320477), 'pulumi.set', 'pulumi.set', (['__self__', '"""new_revision_available"""', 'new_revision_available'], {}), "(__self__, 'new_revision_available', new_revision_available)\n", (320417, 320477), False, 'import pulumi\n'), ((320535, 320601), 'pulumi.set', 'pulumi.set', (['__self__', '"""new_revision_summary"""', 'new_revision_summary'], {}), "(__self__, 'new_revision_summary', new_revision_summary)\n", (320545, 320601), False, 'import pulumi\n'), ((320657, 320719), 'pulumi.set', 'pulumi.set', (['__self__', '"""version_deprecated"""', 'version_deprecated'], {}), "(__self__, 'version_deprecated', version_deprecated)\n", (320667, 320719), False, 'import pulumi\n'), ((323208, 323252), 'pulumi.set', 'pulumi.set', (['__self__', '"""subnet_id"""', 'subnet_id'], {}), "(__self__, 'subnet_id', subnet_id)\n", (323218, 323252), False, 'import pulumi\n'), ((323294, 323328), 'pulumi.set', 'pulumi.set', (['__self__', '"""zone"""', 'zone'], {}), "(__self__, 'zone', zone)\n", (323304, 323328), False, 'import pulumi\n'), ((324421, 324459), 'pulumi.set', 'pulumi.set', (['__self__', '"""cilium"""', 'cilium'], {}), "(__self__, 'cilium', cilium)\n", (324431, 324459), False, 'import pulumi\n'), ((325576, 325620), 'pulumi.set', 'pulumi.set', (['__self__', '"""locations"""', 'locations'], {}), "(__self__, 'locations', locations)\n", (325586, 325620), False, 'import pulumi\n'), ((326721, 326867), 'warnings.warn', 'warnings.warn', (['"""The \'subnet_id\' field has been deprecated. Please use \'subnet_ids under network_interface\' instead."""', 'DeprecationWarning'], {}), '(\n "The \'subnet_id\' field has been deprecated. Please use \'subnet_ids under network_interface\' instead."\n , DeprecationWarning)\n', (326734, 326867), False, 'import warnings\n'), ((326874, 327027), 'pulumi.log.warn', 'pulumi.log.warn', (['"""subnet_id is deprecated: The \'subnet_id\' field has been deprecated. Please use \'subnet_ids under network_interface\' instead."""'], {}), '(\n "subnet_id is deprecated: The \'subnet_id\' field has been deprecated. Please use \'subnet_ids under network_interface\' instead."\n )\n', (326889, 327027), False, 'import pulumi\n'), ((327068, 327112), 'pulumi.set', 'pulumi.set', (['__self__', '"""subnet_id"""', 'subnet_id'], {}), "(__self__, 'subnet_id', subnet_id)\n", (327078, 327112), False, 'import pulumi\n'), ((327154, 327188), 'pulumi.set', 'pulumi.set', (['__self__', '"""zone"""', 'zone'], {}), "(__self__, 'zone', zone)\n", (327164, 327188), False, 'import pulumi\n'), ((332212, 332256), 'pulumi.set', 'pulumi.set', (['__self__', '"""boot_disk"""', 'boot_disk'], {}), "(__self__, 'boot_disk', boot_disk)\n", (332222, 332256), False, 'import pulumi\n'), ((332302, 332344), 'pulumi.set', 'pulumi.set', (['__self__', '"""metadata"""', 'metadata'], {}), "(__self__, 'metadata', metadata)\n", (332312, 332344), False, 'import pulumi\n'), ((332385, 332518), 'warnings.warn', 'warnings.warn', (['"""The \'nat\' field has been deprecated. Please use \'nat under network_interface\' instead."""', 'DeprecationWarning'], {}), '(\n "The \'nat\' field has been deprecated. Please use \'nat under network_interface\' instead."\n , DeprecationWarning)\n', (332398, 332518), False, 'import warnings\n'), ((332525, 332659), 'pulumi.log.warn', 'pulumi.log.warn', (['"""nat is deprecated: The \'nat\' field has been deprecated. Please use \'nat under network_interface\' instead."""'], {}), '(\n "nat is deprecated: The \'nat\' field has been deprecated. Please use \'nat under network_interface\' instead."\n )\n', (332540, 332659), False, 'import pulumi\n'), ((332694, 332726), 'pulumi.set', 'pulumi.set', (['__self__', '"""nat"""', 'nat'], {}), "(__self__, 'nat', nat)\n", (332704, 332726), False, 'import pulumi\n'), ((332789, 332865), 'pulumi.set', 'pulumi.set', (['__self__', '"""network_acceleration_type"""', 'network_acceleration_type'], {}), "(__self__, 'network_acceleration_type', network_acceleration_type)\n", (332799, 332865), False, 'import pulumi\n'), ((332921, 332983), 'pulumi.set', 'pulumi.set', (['__self__', '"""network_interfaces"""', 'network_interfaces'], {}), "(__self__, 'network_interfaces', network_interfaces)\n", (332931, 332983), False, 'import pulumi\n'), ((333037, 333095), 'pulumi.set', 'pulumi.set', (['__self__', '"""placement_policy"""', 'placement_policy'], {}), "(__self__, 'placement_policy', placement_policy)\n", (333047, 333095), False, 'import pulumi\n'), ((333144, 333192), 'pulumi.set', 'pulumi.set', (['__self__', '"""platform_id"""', 'platform_id'], {}), "(__self__, 'platform_id', platform_id)\n", (333154, 333192), False, 'import pulumi\n'), ((333239, 333283), 'pulumi.set', 'pulumi.set', (['__self__', '"""resources"""', 'resources'], {}), "(__self__, 'resources', resources)\n", (333249, 333283), False, 'import pulumi\n'), ((333338, 333398), 'pulumi.set', 'pulumi.set', (['__self__', '"""scheduling_policy"""', 'scheduling_policy'], {}), "(__self__, 'scheduling_policy', scheduling_policy)\n", (333348, 333398), False, 'import pulumi\n'), ((338432, 338466), 'pulumi.set', 'pulumi.set', (['__self__', '"""size"""', 'size'], {}), "(__self__, 'size', size)\n", (338442, 338466), False, 'import pulumi\n'), ((338508, 338542), 'pulumi.set', 'pulumi.set', (['__self__', '"""type"""', 'type'], {}), "(__self__, 'type', type)\n", (338518, 338542), False, 'import pulumi\n'), ((340320, 340354), 'pulumi.set', 'pulumi.set', (['__self__', '"""ipv4"""', 'ipv4'], {}), "(__self__, 'ipv4', ipv4)\n", (340330, 340354), False, 'import pulumi\n'), ((340396, 340430), 'pulumi.set', 'pulumi.set', (['__self__', '"""ipv6"""', 'ipv6'], {}), "(__self__, 'ipv6', ipv6)\n", (340406, 340430), False, 'import pulumi\n'), ((340471, 340503), 'pulumi.set', 'pulumi.set', (['__self__', '"""nat"""', 'nat'], {}), "(__self__, 'nat', nat)\n", (340481, 340503), False, 'import pulumi\n'), ((340559, 340621), 'pulumi.set', 'pulumi.set', (['__self__', '"""security_group_ids"""', 'security_group_ids'], {}), "(__self__, 'security_group_ids', security_group_ids)\n", (340569, 340621), False, 'import pulumi\n'), ((343734, 343786), 'pulumi.set', 'pulumi.set', (['__self__', '"""core_fraction"""', 'core_fraction'], {}), "(__self__, 'core_fraction', core_fraction)\n", (343744, 343786), False, 'import pulumi\n'), ((343829, 343865), 'pulumi.set', 'pulumi.set', (['__self__', '"""cores"""', 'cores'], {}), "(__self__, 'cores', cores)\n", (343839, 343865), False, 'import pulumi\n'), ((343907, 343941), 'pulumi.set', 'pulumi.set', (['__self__', '"""gpus"""', 'gpus'], {}), "(__self__, 'gpus', gpus)\n", (343917, 343941), False, 'import pulumi\n'), ((343985, 344023), 'pulumi.set', 'pulumi.set', (['__self__', '"""memory"""', 'memory'], {}), "(__self__, 'memory', memory)\n", (343995, 344023), False, 'import pulumi\n'), ((345446, 345494), 'pulumi.set', 'pulumi.set', (['__self__', '"""preemptible"""', 'preemptible'], {}), "(__self__, 'preemptible', preemptible)\n", (345456, 345494), False, 'import pulumi\n'), ((346997, 347061), 'pulumi.set', 'pulumi.set', (['__self__', '"""maintenance_windows"""', 'maintenance_windows'], {}), "(__self__, 'maintenance_windows', maintenance_windows)\n", (347007, 347061), False, 'import pulumi\n'), ((348971, 349003), 'pulumi.set', 'pulumi.set', (['__self__', '"""day"""', 'day'], {}), "(__self__, 'day', day)\n", (348981, 349003), False, 'import pulumi\n'), ((350457, 350503), 'pulumi.set', 'pulumi.set', (['__self__', '"""auto_scale"""', 'auto_scale'], {}), "(__self__, 'auto_scale', auto_scale)\n", (350467, 350503), False, 'import pulumi\n'), ((350552, 350600), 'pulumi.set', 'pulumi.set', (['__self__', '"""fixed_scale"""', 'fixed_scale'], {}), "(__self__, 'fixed_scale', fixed_scale)\n", (350562, 350600), False, 'import pulumi\n'), ((353424, 353458), 'pulumi.set', 'pulumi.set', (['__self__', '"""size"""', 'size'], {}), "(__self__, 'size', size)\n", (353434, 353458), False, 'import pulumi\n'), ((355061, 355117), 'pulumi.set', 'pulumi.set', (['__self__', '"""current_version"""', 'current_version'], {}), "(__self__, 'current_version', current_version)\n", (355071, 355117), False, 'import pulumi\n'), ((355177, 355247), 'pulumi.set', 'pulumi.set', (['__self__', '"""new_revision_available"""', 'new_revision_available'], {}), "(__self__, 'new_revision_available', new_revision_available)\n", (355187, 355247), False, 'import pulumi\n'), ((355305, 355371), 'pulumi.set', 'pulumi.set', (['__self__', '"""new_revision_summary"""', 'new_revision_summary'], {}), "(__self__, 'new_revision_summary', new_revision_summary)\n", (355315, 355371), False, 'import pulumi\n'), ((355427, 355489), 'pulumi.set', 'pulumi.set', (['__self__', '"""version_deprecated"""', 'version_deprecated'], {}), "(__self__, 'version_deprecated', version_deprecated)\n", (355437, 355489), False, 'import pulumi\n'), ((360921, 360981), 'pulumi.set', 'pulumi.set', (['__self__', '"""healthy_threshold"""', 'healthy_threshold'], {}), "(__self__, 'healthy_threshold', healthy_threshold)\n", (360931, 360981), False, 'import pulumi\n'), ((361031, 361081), 'pulumi.set', 'pulumi.set', (['__self__', '"""http_options"""', 'http_options'], {}), "(__self__, 'http_options', http_options)\n", (361041, 361081), False, 'import pulumi\n'), ((361127, 361169), 'pulumi.set', 'pulumi.set', (['__self__', '"""interval"""', 'interval'], {}), "(__self__, 'interval', interval)\n", (361137, 361169), False, 'import pulumi\n'), ((361218, 361266), 'pulumi.set', 'pulumi.set', (['__self__', '"""tcp_options"""', 'tcp_options'], {}), "(__self__, 'tcp_options', tcp_options)\n", (361228, 361266), False, 'import pulumi\n'), ((361311, 361351), 'pulumi.set', 'pulumi.set', (['__self__', '"""timeout"""', 'timeout'], {}), "(__self__, 'timeout', timeout)\n", (361321, 361351), False, 'import pulumi\n'), ((361408, 361472), 'pulumi.set', 'pulumi.set', (['__self__', '"""unhealthy_threshold"""', 'unhealthy_threshold'], {}), "(__self__, 'unhealthy_threshold', unhealthy_threshold)\n", (361418, 361472), False, 'import pulumi\n'), ((365080, 365114), 'pulumi.set', 'pulumi.set', (['__self__', '"""path"""', 'path'], {}), "(__self__, 'path', path)\n", (365090, 365114), False, 'import pulumi\n'), ((367866, 367934), 'pulumi.set', 'pulumi.set', (['__self__', '"""external_address_spec"""', 'external_address_spec'], {}), "(__self__, 'external_address_spec', external_address_spec)\n", (367876, 367934), False, 'import pulumi\n'), ((367993, 368061), 'pulumi.set', 'pulumi.set', (['__self__', '"""internal_address_spec"""', 'internal_address_spec'], {}), "(__self__, 'internal_address_spec', internal_address_spec)\n", (368003, 368061), False, 'import pulumi\n'), ((368107, 368149), 'pulumi.set', 'pulumi.set', (['__self__', '"""protocol"""', 'protocol'], {}), "(__self__, 'protocol', protocol)\n", (368117, 368149), False, 'import pulumi\n'), ((368198, 368246), 'pulumi.set', 'pulumi.set', (['__self__', '"""target_port"""', 'target_port'], {}), "(__self__, 'target_port', target_port)\n", (368208, 368246), False, 'import pulumi\n'), ((371386, 371426), 'pulumi.set', 'pulumi.set', (['__self__', '"""address"""', 'address'], {}), "(__self__, 'address', address)\n", (371396, 371426), False, 'import pulumi\n'), ((371474, 371520), 'pulumi.set', 'pulumi.set', (['__self__', '"""ip_version"""', 'ip_version'], {}), "(__self__, 'ip_version', ip_version)\n", (371484, 371520), False, 'import pulumi\n'), ((373261, 373301), 'pulumi.set', 'pulumi.set', (['__self__', '"""address"""', 'address'], {}), "(__self__, 'address', address)\n", (373271, 373301), False, 'import pulumi\n'), ((373349, 373395), 'pulumi.set', 'pulumi.set', (['__self__', '"""ip_version"""', 'ip_version'], {}), "(__self__, 'ip_version', ip_version)\n", (373359, 373395), False, 'import pulumi\n'), ((376790, 376834), 'pulumi.set', 'pulumi.set', (['__self__', '"""data_lens"""', 'data_lens'], {}), "(__self__, 'data_lens', data_lens)\n", (376800, 376834), False, 'import pulumi\n'), ((376879, 376919), 'pulumi.set', 'pulumi.set', (['__self__', '"""metrika"""', 'metrika'], {}), "(__self__, 'metrika', metrika)\n", (376889, 376919), False, 'import pulumi\n'), ((376967, 377013), 'pulumi.set', 'pulumi.set', (['__self__', '"""serverless"""', 'serverless'], {}), "(__self__, 'serverless', serverless)\n", (376977, 377013), False, 'import pulumi\n'), ((377058, 377098), 'pulumi.set', 'pulumi.set', (['__self__', '"""web_sql"""', 'web_sql'], {}), "(__self__, 'web_sql', web_sql)\n", (377068, 377098), False, 'import pulumi\n'), ((379012, 379048), 'pulumi.set', 'pulumi.set', (['__self__', '"""hours"""', 'hours'], {}), "(__self__, 'hours', hours)\n", (379022, 379048), False, 'import pulumi\n'), ((379093, 379133), 'pulumi.set', 'pulumi.set', (['__self__', '"""minutes"""', 'minutes'], {}), "(__self__, 'minutes', minutes)\n", (379103, 379133), False, 'import pulumi\n'), ((380483, 380521), 'pulumi.set', 'pulumi.set', (['__self__', '"""config"""', 'config'], {}), "(__self__, 'config', config)\n", (380493, 380521), False, 'import pulumi\n'), ((385562, 385628), 'pulumi.set', 'pulumi.set', (['__self__', '"""background_pool_size"""', 'background_pool_size'], {}), "(__self__, 'background_pool_size', background_pool_size)\n", (385572, 385628), False, 'import pulumi\n'), ((385695, 385783), 'pulumi.set', 'pulumi.set', (['__self__', '"""background_schedule_pool_size"""', 'background_schedule_pool_size'], {}), "(__self__, 'background_schedule_pool_size',\n background_schedule_pool_size)\n", (385705, 385783), False, 'import pulumi\n'), ((385829, 385879), 'pulumi.set', 'pulumi.set', (['__self__', '"""compressions"""', 'compressions'], {}), "(__self__, 'compressions', compressions)\n", (385839, 385879), False, 'import pulumi\n'), ((385928, 385976), 'pulumi.set', 'pulumi.set', (['__self__', '"""geobase_uri"""', 'geobase_uri'], {}), "(__self__, 'geobase_uri', geobase_uri)\n", (385938, 385976), False, 'import pulumi\n'), ((386030, 386088), 'pulumi.set', 'pulumi.set', (['__self__', '"""graphite_rollups"""', 'graphite_rollups'], {}), "(__self__, 'graphite_rollups', graphite_rollups)\n", (386040, 386088), False, 'import pulumi\n'), ((386131, 386167), 'pulumi.set', 'pulumi.set', (['__self__', '"""kafka"""', 'kafka'], {}), "(__self__, 'kafka', kafka)\n", (386141, 386167), False, 'import pulumi\n'), ((386217, 386267), 'pulumi.set', 'pulumi.set', (['__self__', '"""kafka_topics"""', 'kafka_topics'], {}), "(__self__, 'kafka_topics', kafka_topics)\n", (386227, 386267), False, 'import pulumi\n'), ((386323, 386385), 'pulumi.set', 'pulumi.set', (['__self__', '"""keep_alive_timeout"""', 'keep_alive_timeout'], {}), "(__self__, 'keep_alive_timeout', keep_alive_timeout)\n", (386333, 386385), False, 'import pulumi\n'), ((386432, 386476), 'pulumi.set', 'pulumi.set', (['__self__', '"""log_level"""', 'log_level'], {}), "(__self__, 'log_level', log_level)\n", (386442, 386476), False, 'import pulumi\n'), ((386529, 386585), 'pulumi.set', 'pulumi.set', (['__self__', '"""mark_cache_size"""', 'mark_cache_size'], {}), "(__self__, 'mark_cache_size', mark_cache_size)\n", (386539, 386585), False, 'import pulumi\n'), ((386645, 386715), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_concurrent_queries"""', 'max_concurrent_queries'], {}), "(__self__, 'max_concurrent_queries', max_concurrent_queries)\n", (386655, 386715), False, 'import pulumi\n'), ((386768, 386824), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_connections"""', 'max_connections'], {}), "(__self__, 'max_connections', max_connections)\n", (386778, 386824), False, 'import pulumi\n'), ((386888, 386966), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_partition_size_to_drop"""', 'max_partition_size_to_drop'], {}), "(__self__, 'max_partition_size_to_drop', max_partition_size_to_drop)\n", (386898, 386966), False, 'import pulumi\n'), ((387026, 387096), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_table_size_to_drop"""', 'max_table_size_to_drop'], {}), "(__self__, 'max_table_size_to_drop', max_table_size_to_drop)\n", (387036, 387096), False, 'import pulumi\n'), ((387144, 387190), 'pulumi.set', 'pulumi.set', (['__self__', '"""merge_tree"""', 'merge_tree'], {}), "(__self__, 'merge_tree', merge_tree)\n", (387154, 387190), False, 'import pulumi\n'), ((387246, 387308), 'pulumi.set', 'pulumi.set', (['__self__', '"""metric_log_enabled"""', 'metric_log_enabled'], {}), "(__self__, 'metric_log_enabled', metric_log_enabled)\n", (387256, 387308), False, 'import pulumi\n'), ((387371, 387447), 'pulumi.set', 'pulumi.set', (['__self__', '"""metric_log_retention_size"""', 'metric_log_retention_size'], {}), "(__self__, 'metric_log_retention_size', metric_log_retention_size)\n", (387381, 387447), False, 'import pulumi\n'), ((387510, 387586), 'pulumi.set', 'pulumi.set', (['__self__', '"""metric_log_retention_time"""', 'metric_log_retention_time'], {}), "(__self__, 'metric_log_retention_time', metric_log_retention_time)\n", (387520, 387586), False, 'import pulumi\n'), ((387647, 387719), 'pulumi.set', 'pulumi.set', (['__self__', '"""part_log_retention_size"""', 'part_log_retention_size'], {}), "(__self__, 'part_log_retention_size', part_log_retention_size)\n", (387657, 387719), False, 'import pulumi\n'), ((387780, 387852), 'pulumi.set', 'pulumi.set', (['__self__', '"""part_log_retention_time"""', 'part_log_retention_time'], {}), "(__self__, 'part_log_retention_time', part_log_retention_time)\n", (387790, 387852), False, 'import pulumi\n'), ((387914, 387988), 'pulumi.set', 'pulumi.set', (['__self__', '"""query_log_retention_size"""', 'query_log_retention_size'], {}), "(__self__, 'query_log_retention_size', query_log_retention_size)\n", (387924, 387988), False, 'import pulumi\n'), ((388050, 388124), 'pulumi.set', 'pulumi.set', (['__self__', '"""query_log_retention_time"""', 'query_log_retention_time'], {}), "(__self__, 'query_log_retention_time', query_log_retention_time)\n", (388060, 388124), False, 'import pulumi\n'), ((388186, 388260), 'pulumi.set', 'pulumi.set', (['__self__', '"""query_thread_log_enabled"""', 'query_thread_log_enabled'], {}), "(__self__, 'query_thread_log_enabled', query_thread_log_enabled)\n", (388196, 388260), False, 'import pulumi\n'), ((388329, 388421), 'pulumi.set', 'pulumi.set', (['__self__', '"""query_thread_log_retention_size"""', 'query_thread_log_retention_size'], {}), "(__self__, 'query_thread_log_retention_size',\n query_thread_log_retention_size)\n", (388339, 388421), False, 'import pulumi\n'), ((388486, 388578), 'pulumi.set', 'pulumi.set', (['__self__', '"""query_thread_log_retention_time"""', 'query_thread_log_retention_time'], {}), "(__self__, 'query_thread_log_retention_time',\n query_thread_log_retention_time)\n", (388496, 388578), False, 'import pulumi\n'), ((388620, 388662), 'pulumi.set', 'pulumi.set', (['__self__', '"""rabbitmq"""', 'rabbitmq'], {}), "(__self__, 'rabbitmq', rabbitmq)\n", (388630, 388662), False, 'import pulumi\n'), ((388716, 388774), 'pulumi.set', 'pulumi.set', (['__self__', '"""text_log_enabled"""', 'text_log_enabled'], {}), "(__self__, 'text_log_enabled', text_log_enabled)\n", (388726, 388774), False, 'import pulumi\n'), ((388826, 388880), 'pulumi.set', 'pulumi.set', (['__self__', '"""text_log_level"""', 'text_log_level'], {}), "(__self__, 'text_log_level', text_log_level)\n", (388836, 388880), False, 'import pulumi\n'), ((388941, 389013), 'pulumi.set', 'pulumi.set', (['__self__', '"""text_log_retention_size"""', 'text_log_retention_size'], {}), "(__self__, 'text_log_retention_size', text_log_retention_size)\n", (388951, 389013), False, 'import pulumi\n'), ((389074, 389146), 'pulumi.set', 'pulumi.set', (['__self__', '"""text_log_retention_time"""', 'text_log_retention_time'], {}), "(__self__, 'text_log_retention_time', text_log_retention_time)\n", (389084, 389146), False, 'import pulumi\n'), ((389192, 389234), 'pulumi.set', 'pulumi.set', (['__self__', '"""timezone"""', 'timezone'], {}), "(__self__, 'timezone', timezone)\n", (389202, 389234), False, 'import pulumi\n'), ((389289, 389349), 'pulumi.set', 'pulumi.set', (['__self__', '"""trace_log_enabled"""', 'trace_log_enabled'], {}), "(__self__, 'trace_log_enabled', trace_log_enabled)\n", (389299, 389349), False, 'import pulumi\n'), ((389411, 389485), 'pulumi.set', 'pulumi.set', (['__self__', '"""trace_log_retention_size"""', 'trace_log_retention_size'], {}), "(__self__, 'trace_log_retention_size', trace_log_retention_size)\n", (389421, 389485), False, 'import pulumi\n'), ((389547, 389621), 'pulumi.set', 'pulumi.set', (['__self__', '"""trace_log_retention_time"""', 'trace_log_retention_time'], {}), "(__self__, 'trace_log_retention_time', trace_log_retention_time)\n", (389557, 389621), False, 'import pulumi\n'), ((389682, 389754), 'pulumi.set', 'pulumi.set', (['__self__', '"""uncompressed_cache_size"""', 'uncompressed_cache_size'], {}), "(__self__, 'uncompressed_cache_size', uncompressed_cache_size)\n", (389692, 389754), False, 'import pulumi\n'), ((406009, 406051), 'pulumi.set', 'pulumi.set', (['__self__', '"""patterns"""', 'patterns'], {}), "(__self__, 'patterns', patterns)\n", (406019, 406051), False, 'import pulumi\n'), ((407645, 407683), 'pulumi.set', 'pulumi.set', (['__self__', '"""regexp"""', 'regexp'], {}), "(__self__, 'regexp', regexp)\n", (407655, 407683), False, 'import pulumi\n'), ((407731, 407777), 'pulumi.set', 'pulumi.set', (['__self__', '"""retentions"""', 'retentions'], {}), "(__self__, 'retentions', retentions)\n", (407741, 407777), False, 'import pulumi\n'), ((410829, 410883), 'pulumi.set', 'pulumi.set', (['__self__', '"""sasl_mechanism"""', 'sasl_mechanism'], {}), "(__self__, 'sasl_mechanism', sasl_mechanism)\n", (410839, 410883), False, 'import pulumi\n'), ((410934, 410986), 'pulumi.set', 'pulumi.set', (['__self__', '"""sasl_password"""', 'sasl_password'], {}), "(__self__, 'sasl_password', sasl_password)\n", (410944, 410986), False, 'import pulumi\n'), ((411037, 411089), 'pulumi.set', 'pulumi.set', (['__self__', '"""sasl_username"""', 'sasl_username'], {}), "(__self__, 'sasl_username', sasl_username)\n", (411047, 411089), False, 'import pulumi\n'), ((411144, 411204), 'pulumi.set', 'pulumi.set', (['__self__', '"""security_protocol"""', 'security_protocol'], {}), "(__self__, 'security_protocol', security_protocol)\n", (411154, 411204), False, 'import pulumi\n'), ((413344, 413386), 'pulumi.set', 'pulumi.set', (['__self__', '"""settings"""', 'settings'], {}), "(__self__, 'settings', settings)\n", (413354, 413386), False, 'import pulumi\n'), ((414945, 414999), 'pulumi.set', 'pulumi.set', (['__self__', '"""sasl_mechanism"""', 'sasl_mechanism'], {}), "(__self__, 'sasl_mechanism', sasl_mechanism)\n", (414955, 414999), False, 'import pulumi\n'), ((415050, 415102), 'pulumi.set', 'pulumi.set', (['__self__', '"""sasl_password"""', 'sasl_password'], {}), "(__self__, 'sasl_password', sasl_password)\n", (415060, 415102), False, 'import pulumi\n'), ((415153, 415205), 'pulumi.set', 'pulumi.set', (['__self__', '"""sasl_username"""', 'sasl_username'], {}), "(__self__, 'sasl_username', sasl_username)\n", (415163, 415205), False, 'import pulumi\n'), ((415260, 415320), 'pulumi.set', 'pulumi.set', (['__self__', '"""security_protocol"""', 'security_protocol'], {}), "(__self__, 'security_protocol', security_protocol)\n", (415270, 415320), False, 'import pulumi\n'), ((419309, 419417), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_bytes_to_merge_at_min_space_in_pool"""', 'max_bytes_to_merge_at_min_space_in_pool'], {}), "(__self__, 'max_bytes_to_merge_at_min_space_in_pool',\n max_bytes_to_merge_at_min_space_in_pool)\n", (419319, 419417), False, 'import pulumi\n'), ((419481, 419571), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_replicated_merges_in_queue"""', 'max_replicated_merges_in_queue'], {}), "(__self__, 'max_replicated_merges_in_queue',\n max_replicated_merges_in_queue)\n", (419491, 419571), False, 'import pulumi\n'), ((419662, 419810), 'pulumi.set', 'pulumi.set', (['__self__', '"""number_of_free_entries_in_pool_to_lower_max_size_of_merge"""', 'number_of_free_entries_in_pool_to_lower_max_size_of_merge'], {}), "(__self__,\n 'number_of_free_entries_in_pool_to_lower_max_size_of_merge',\n number_of_free_entries_in_pool_to_lower_max_size_of_merge)\n", (419672, 419810), False, 'import pulumi\n'), ((419861, 419929), 'pulumi.set', 'pulumi.set', (['__self__', '"""parts_to_delay_insert"""', 'parts_to_delay_insert'], {}), "(__self__, 'parts_to_delay_insert', parts_to_delay_insert)\n", (419871, 419929), False, 'import pulumi\n'), ((419988, 420056), 'pulumi.set', 'pulumi.set', (['__self__', '"""parts_to_throw_insert"""', 'parts_to_throw_insert'], {}), "(__self__, 'parts_to_throw_insert', parts_to_throw_insert)\n", (419998, 420056), False, 'import pulumi\n'), ((420125, 420217), 'pulumi.set', 'pulumi.set', (['__self__', '"""replicated_deduplication_window"""', 'replicated_deduplication_window'], {}), "(__self__, 'replicated_deduplication_window',\n replicated_deduplication_window)\n", (420135, 420217), False, 'import pulumi\n'), ((420290, 420398), 'pulumi.set', 'pulumi.set', (['__self__', '"""replicated_deduplication_window_seconds"""', 'replicated_deduplication_window_seconds'], {}), "(__self__, 'replicated_deduplication_window_seconds',\n replicated_deduplication_window_seconds)\n", (420300, 420398), False, 'import pulumi\n'), ((425171, 425213), 'pulumi.set', 'pulumi.set', (['__self__', '"""password"""', 'password'], {}), "(__self__, 'password', password)\n", (425181, 425213), False, 'import pulumi\n'), ((425259, 425301), 'pulumi.set', 'pulumi.set', (['__self__', '"""username"""', 'username'], {}), "(__self__, 'username', username)\n", (425269, 425301), False, 'import pulumi\n'), ((432373, 432431), 'pulumi.set', 'pulumi.set', (['__self__', '"""assign_public_ip"""', 'assign_public_ip'], {}), "(__self__, 'assign_public_ip', assign_public_ip)\n", (432383, 432431), False, 'import pulumi\n'), ((432473, 432507), 'pulumi.set', 'pulumi.set', (['__self__', '"""fqdn"""', 'fqdn'], {}), "(__self__, 'fqdn', fqdn)\n", (432483, 432507), False, 'import pulumi\n'), ((432555, 432601), 'pulumi.set', 'pulumi.set', (['__self__', '"""shard_name"""', 'shard_name'], {}), "(__self__, 'shard_name', shard_name)\n", (432565, 432601), False, 'import pulumi\n'), ((432648, 432692), 'pulumi.set', 'pulumi.set', (['__self__', '"""subnet_id"""', 'subnet_id'], {}), "(__self__, 'subnet_id', subnet_id)\n", (432658, 432692), False, 'import pulumi\n'), ((435911, 435943), 'pulumi.set', 'pulumi.set', (['__self__', '"""day"""', 'day'], {}), "(__self__, 'day', day)\n", (435921, 435943), False, 'import pulumi\n'), ((435985, 436019), 'pulumi.set', 'pulumi.set', (['__self__', '"""hour"""', 'hour'], {}), "(__self__, 'hour', hour)\n", (435995, 436019), False, 'import pulumi\n'), ((439585, 439633), 'pulumi.set', 'pulumi.set', (['__self__', '"""description"""', 'description'], {}), "(__self__, 'description', description)\n", (439595, 439633), False, 'import pulumi\n'), ((441940, 441988), 'pulumi.set', 'pulumi.set', (['__self__', '"""permissions"""', 'permissions'], {}), "(__self__, 'permissions', permissions)\n", (441950, 441988), False, 'import pulumi\n'), ((442032, 442070), 'pulumi.set', 'pulumi.set', (['__self__', '"""quotas"""', 'quotas'], {}), "(__self__, 'quotas', quotas)\n", (442042, 442070), False, 'import pulumi\n'), ((442116, 442158), 'pulumi.set', 'pulumi.set', (['__self__', '"""settings"""', 'settings'], {}), "(__self__, 'settings', settings)\n", (442126, 442158), False, 'import pulumi\n'), ((446030, 446068), 'pulumi.set', 'pulumi.set', (['__self__', '"""errors"""', 'errors'], {}), "(__self__, 'errors', errors)\n", (446040, 446068), False, 'import pulumi\n'), ((446120, 446174), 'pulumi.set', 'pulumi.set', (['__self__', '"""execution_time"""', 'execution_time'], {}), "(__self__, 'execution_time', execution_time)\n", (446130, 446174), False, 'import pulumi\n'), ((446219, 446259), 'pulumi.set', 'pulumi.set', (['__self__', '"""queries"""', 'queries'], {}), "(__self__, 'queries', queries)\n", (446229, 446259), False, 'import pulumi\n'), ((446306, 446350), 'pulumi.set', 'pulumi.set', (['__self__', '"""read_rows"""', 'read_rows'], {}), "(__self__, 'read_rows', read_rows)\n", (446316, 446350), False, 'import pulumi\n'), ((446399, 446447), 'pulumi.set', 'pulumi.set', (['__self__', '"""result_rows"""', 'result_rows'], {}), "(__self__, 'result_rows', result_rows)\n", (446409, 446447), False, 'import pulumi\n'), ((468784, 468850), 'pulumi.set', 'pulumi.set', (['__self__', '"""add_http_cors_header"""', 'add_http_cors_header'], {}), "(__self__, 'add_http_cors_header', add_http_cors_header)\n", (468794, 468850), False, 'import pulumi\n'), ((468897, 468941), 'pulumi.set', 'pulumi.set', (['__self__', '"""allow_ddl"""', 'allow_ddl'], {}), "(__self__, 'allow_ddl', allow_ddl)\n", (468907, 468941), False, 'import pulumi\n'), ((468986, 469026), 'pulumi.set', 'pulumi.set', (['__self__', '"""compile"""', 'compile'], {}), "(__self__, 'compile', compile)\n", (468996, 469026), False, 'import pulumi\n'), ((469083, 469147), 'pulumi.set', 'pulumi.set', (['__self__', '"""compile_expressions"""', 'compile_expressions'], {}), "(__self__, 'compile_expressions', compile_expressions)\n", (469093, 469147), False, 'import pulumi\n'), ((469200, 469256), 'pulumi.set', 'pulumi.set', (['__self__', '"""connect_timeout"""', 'connect_timeout'], {}), "(__self__, 'connect_timeout', connect_timeout)\n", (469210, 469256), False, 'import pulumi\n'), ((469323, 469411), 'pulumi.set', 'pulumi.set', (['__self__', '"""count_distinct_implementation"""', 'count_distinct_implementation'], {}), "(__self__, 'count_distinct_implementation',\n count_distinct_implementation)\n", (469333, 469411), False, 'import pulumi\n'), ((469467, 469537), 'pulumi.set', 'pulumi.set', (['__self__', '"""distinct_overflow_mode"""', 'distinct_overflow_mode'], {}), "(__self__, 'distinct_overflow_mode', distinct_overflow_mode)\n", (469477, 469537), False, 'import pulumi\n'), ((469615, 469725), 'pulumi.set', 'pulumi.set', (['__self__', '"""distributed_aggregation_memory_efficient"""', 'distributed_aggregation_memory_efficient'], {}), "(__self__, 'distributed_aggregation_memory_efficient',\n distributed_aggregation_memory_efficient)\n", (469625, 469725), False, 'import pulumi\n'), ((469787, 469873), 'pulumi.set', 'pulumi.set', (['__self__', '"""distributed_ddl_task_timeout"""', 'distributed_ddl_task_timeout'], {}), "(__self__, 'distributed_ddl_task_timeout',\n distributed_ddl_task_timeout)\n", (469797, 469873), False, 'import pulumi\n'), ((469931, 470005), 'pulumi.set', 'pulumi.set', (['__self__', '"""distributed_product_mode"""', 'distributed_product_mode'], {}), "(__self__, 'distributed_product_mode', distributed_product_mode)\n", (469941, 470005), False, 'import pulumi\n'), ((470084, 470196), 'pulumi.set', 'pulumi.set', (['__self__', '"""empty_result_for_aggregation_by_empty_set"""', 'empty_result_for_aggregation_by_empty_set'], {}), "(__self__, 'empty_result_for_aggregation_by_empty_set',\n empty_result_for_aggregation_by_empty_set)\n", (470094, 470196), False, 'import pulumi\n'), ((470253, 470325), 'pulumi.set', 'pulumi.set', (['__self__', '"""enable_http_compression"""', 'enable_http_compression'], {}), "(__self__, 'enable_http_compression', enable_http_compression)\n", (470263, 470325), False, 'import pulumi\n'), ((470413, 470543), 'pulumi.set', 'pulumi.set', (['__self__', '"""fallback_to_stale_replicas_for_distributed_queries"""', 'fallback_to_stale_replicas_for_distributed_queries'], {}), "(__self__, 'fallback_to_stale_replicas_for_distributed_queries',\n fallback_to_stale_replicas_for_distributed_queries)\n", (470423, 470543), False, 'import pulumi\n'), ((470596, 470660), 'pulumi.set', 'pulumi.set', (['__self__', '"""force_index_by_date"""', 'force_index_by_date'], {}), "(__self__, 'force_index_by_date', force_index_by_date)\n", (470606, 470660), False, 'import pulumi\n'), ((470715, 470775), 'pulumi.set', 'pulumi.set', (['__self__', '"""force_primary_key"""', 'force_primary_key'], {}), "(__self__, 'force_primary_key', force_primary_key)\n", (470725, 470775), False, 'import pulumi\n'), ((470835, 470905), 'pulumi.set', 'pulumi.set', (['__self__', '"""group_by_overflow_mode"""', 'group_by_overflow_mode'], {}), "(__self__, 'group_by_overflow_mode', group_by_overflow_mode)\n", (470845, 470905), False, 'import pulumi\n'), ((470971, 471057), 'pulumi.set', 'pulumi.set', (['__self__', '"""group_by_two_level_threshold"""', 'group_by_two_level_threshold'], {}), "(__self__, 'group_by_two_level_threshold',\n group_by_two_level_threshold)\n", (470981, 471057), False, 'import pulumi\n'), ((471125, 471223), 'pulumi.set', 'pulumi.set', (['__self__', '"""group_by_two_level_threshold_bytes"""', 'group_by_two_level_threshold_bytes'], {}), "(__self__, 'group_by_two_level_threshold_bytes',\n group_by_two_level_threshold_bytes)\n", (471135, 471223), False, 'import pulumi\n'), ((471280, 471352), 'pulumi.set', 'pulumi.set', (['__self__', '"""http_connection_timeout"""', 'http_connection_timeout'], {}), "(__self__, 'http_connection_timeout', http_connection_timeout)\n", (471290, 471352), False, 'import pulumi\n'), ((471420, 471510), 'pulumi.set', 'pulumi.set', (['__self__', '"""http_headers_progress_interval"""', 'http_headers_progress_interval'], {}), "(__self__, 'http_headers_progress_interval',\n http_headers_progress_interval)\n", (471430, 471510), False, 'import pulumi\n'), ((471564, 471630), 'pulumi.set', 'pulumi.set', (['__self__', '"""http_receive_timeout"""', 'http_receive_timeout'], {}), "(__self__, 'http_receive_timeout', http_receive_timeout)\n", (471574, 471630), False, 'import pulumi\n'), ((471685, 471745), 'pulumi.set', 'pulumi.set', (['__self__', '"""http_send_timeout"""', 'http_send_timeout'], {}), "(__self__, 'http_send_timeout', http_send_timeout)\n", (471695, 471745), False, 'import pulumi\n'), ((471823, 471933), 'pulumi.set', 'pulumi.set', (['__self__', '"""input_format_defaults_for_omitted_fields"""', 'input_format_defaults_for_omitted_fields'], {}), "(__self__, 'input_format_defaults_for_omitted_fields',\n input_format_defaults_for_omitted_fields)\n", (471833, 471933), False, 'import pulumi\n'), ((472008, 472120), 'pulumi.set', 'pulumi.set', (['__self__', '"""input_format_values_interpret_expressions"""', 'input_format_values_interpret_expressions'], {}), "(__self__, 'input_format_values_interpret_expressions',\n input_format_values_interpret_expressions)\n", (472018, 472120), False, 'import pulumi\n'), ((472167, 472219), 'pulumi.set', 'pulumi.set', (['__self__', '"""insert_quorum"""', 'insert_quorum'], {}), "(__self__, 'insert_quorum', insert_quorum)\n", (472177, 472219), False, 'import pulumi\n'), ((472278, 472346), 'pulumi.set', 'pulumi.set', (['__self__', '"""insert_quorum_timeout"""', 'insert_quorum_timeout'], {}), "(__self__, 'insert_quorum_timeout', insert_quorum_timeout)\n", (472288, 472346), False, 'import pulumi\n'), ((472402, 472464), 'pulumi.set', 'pulumi.set', (['__self__', '"""join_overflow_mode"""', 'join_overflow_mode'], {}), "(__self__, 'join_overflow_mode', join_overflow_mode)\n", (472412, 472464), False, 'import pulumi\n'), ((472516, 472570), 'pulumi.set', 'pulumi.set', (['__self__', '"""join_use_nulls"""', 'join_use_nulls'], {}), "(__self__, 'join_use_nulls', join_use_nulls)\n", (472526, 472570), False, 'import pulumi\n'), ((472638, 472728), 'pulumi.set', 'pulumi.set', (['__self__', '"""joined_subquery_requires_alias"""', 'joined_subquery_requires_alias'], {}), "(__self__, 'joined_subquery_requires_alias',\n joined_subquery_requires_alias)\n", (472648, 472728), False, 'import pulumi\n'), ((472800, 472906), 'pulumi.set', 'pulumi.set', (['__self__', '"""low_cardinality_allow_in_native_format"""', 'low_cardinality_allow_in_native_format'], {}), "(__self__, 'low_cardinality_allow_in_native_format',\n low_cardinality_allow_in_native_format)\n", (472810, 472906), False, 'import pulumi\n'), ((472953, 473005), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_ast_depth"""', 'max_ast_depth'], {}), "(__self__, 'max_ast_depth', max_ast_depth)\n", (472963, 473005), False, 'import pulumi\n'), ((473059, 473117), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_ast_elements"""', 'max_ast_elements'], {}), "(__self__, 'max_ast_elements', max_ast_elements)\n", (473069, 473117), False, 'import pulumi\n'), ((473169, 473223), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_block_size"""', 'max_block_size'], {}), "(__self__, 'max_block_size', max_block_size)\n", (473179, 473223), False, 'import pulumi\n'), ((473295, 473393), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_bytes_before_external_group_by"""', 'max_bytes_before_external_group_by'], {}), "(__self__, 'max_bytes_before_external_group_by',\n max_bytes_before_external_group_by)\n", (473305, 473393), False, 'import pulumi\n'), ((473457, 473547), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_bytes_before_external_sort"""', 'max_bytes_before_external_sort'], {}), "(__self__, 'max_bytes_before_external_sort',\n max_bytes_before_external_sort)\n", (473467, 473547), False, 'import pulumi\n'), ((473602, 473670), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_bytes_in_distinct"""', 'max_bytes_in_distinct'], {}), "(__self__, 'max_bytes_in_distinct', max_bytes_in_distinct)\n", (473612, 473670), False, 'import pulumi\n'), ((473725, 473785), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_bytes_in_join"""', 'max_bytes_in_join'], {}), "(__self__, 'max_bytes_in_join', max_bytes_in_join)\n", (473735, 473785), False, 'import pulumi\n'), ((473839, 473897), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_bytes_in_set"""', 'max_bytes_in_set'], {}), "(__self__, 'max_bytes_in_set', max_bytes_in_set)\n", (473849, 473897), False, 'import pulumi\n'), ((473952, 474012), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_bytes_to_read"""', 'max_bytes_to_read'], {}), "(__self__, 'max_bytes_to_read', max_bytes_to_read)\n", (473962, 474012), False, 'import pulumi\n'), ((474067, 474127), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_bytes_to_sort"""', 'max_bytes_to_sort'], {}), "(__self__, 'max_bytes_to_sort', max_bytes_to_sort)\n", (474077, 474127), False, 'import pulumi\n'), ((474186, 474254), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_bytes_to_transfer"""', 'max_bytes_to_transfer'], {}), "(__self__, 'max_bytes_to_transfer', max_bytes_to_transfer)\n", (474196, 474254), False, 'import pulumi\n'), ((474311, 474375), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_columns_to_read"""', 'max_columns_to_read'], {}), "(__self__, 'max_columns_to_read', max_columns_to_read)\n", (474321, 474375), False, 'import pulumi\n'), ((474431, 474493), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_execution_time"""', 'max_execution_time'], {}), "(__self__, 'max_execution_time', max_execution_time)\n", (474441, 474493), False, 'import pulumi\n'), ((474556, 474632), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_expanded_ast_elements"""', 'max_expanded_ast_elements'], {}), "(__self__, 'max_expanded_ast_elements', max_expanded_ast_elements)\n", (474566, 474632), False, 'import pulumi\n'), ((474691, 474759), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_insert_block_size"""', 'max_insert_block_size'], {}), "(__self__, 'max_insert_block_size', max_insert_block_size)\n", (474701, 474759), False, 'import pulumi\n'), ((474813, 474871), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_memory_usage"""', 'max_memory_usage'], {}), "(__self__, 'max_memory_usage', max_memory_usage)\n", (474823, 474871), False, 'import pulumi\n'), ((474934, 475010), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_memory_usage_for_user"""', 'max_memory_usage_for_user'], {}), "(__self__, 'max_memory_usage_for_user', max_memory_usage_for_user)\n", (474944, 475010), False, 'import pulumi\n'), ((475069, 475137), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_network_bandwidth"""', 'max_network_bandwidth'], {}), "(__self__, 'max_network_bandwidth', max_network_bandwidth)\n", (475079, 475137), False, 'import pulumi\n'), ((475205, 475295), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_network_bandwidth_for_user"""', 'max_network_bandwidth_for_user'], {}), "(__self__, 'max_network_bandwidth_for_user',\n max_network_bandwidth_for_user)\n", (475215, 475295), False, 'import pulumi\n'), ((475343, 475397), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_query_size"""', 'max_query_size'], {}), "(__self__, 'max_query_size', max_query_size)\n", (475353, 475397), False, 'import pulumi\n'), ((475476, 475588), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_replica_delay_for_distributed_queries"""', 'max_replica_delay_for_distributed_queries'], {}), "(__self__, 'max_replica_delay_for_distributed_queries',\n max_replica_delay_for_distributed_queries)\n", (475486, 475588), False, 'import pulumi\n'), ((475638, 475696), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_result_bytes"""', 'max_result_bytes'], {}), "(__self__, 'max_result_bytes', max_result_bytes)\n", (475648, 475696), False, 'import pulumi\n'), ((475749, 475805), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_result_rows"""', 'max_result_rows'], {}), "(__self__, 'max_result_rows', max_result_rows)\n", (475759, 475805), False, 'import pulumi\n'), ((475863, 475929), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_rows_in_distinct"""', 'max_rows_in_distinct'], {}), "(__self__, 'max_rows_in_distinct', max_rows_in_distinct)\n", (475873, 475929), False, 'import pulumi\n'), ((475983, 476041), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_rows_in_join"""', 'max_rows_in_join'], {}), "(__self__, 'max_rows_in_join', max_rows_in_join)\n", (475993, 476041), False, 'import pulumi\n'), ((476094, 476150), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_rows_in_set"""', 'max_rows_in_set'], {}), "(__self__, 'max_rows_in_set', max_rows_in_set)\n", (476104, 476150), False, 'import pulumi\n'), ((476208, 476274), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_rows_to_group_by"""', 'max_rows_to_group_by'], {}), "(__self__, 'max_rows_to_group_by', max_rows_to_group_by)\n", (476218, 476274), False, 'import pulumi\n'), ((476328, 476386), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_rows_to_read"""', 'max_rows_to_read'], {}), "(__self__, 'max_rows_to_read', max_rows_to_read)\n", (476338, 476386), False, 'import pulumi\n'), ((476440, 476498), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_rows_to_sort"""', 'max_rows_to_sort'], {}), "(__self__, 'max_rows_to_sort', max_rows_to_sort)\n", (476450, 476498), False, 'import pulumi\n'), ((476556, 476622), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_rows_to_transfer"""', 'max_rows_to_transfer'], {}), "(__self__, 'max_rows_to_transfer', max_rows_to_transfer)\n", (476566, 476622), False, 'import pulumi\n'), ((476681, 476749), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_temporary_columns"""', 'max_temporary_columns'], {}), "(__self__, 'max_temporary_columns', max_temporary_columns)\n", (476691, 476749), False, 'import pulumi\n'), ((476818, 476910), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_temporary_non_const_columns"""', 'max_temporary_non_const_columns'], {}), "(__self__, 'max_temporary_non_const_columns',\n max_temporary_non_const_columns)\n", (476828, 476910), False, 'import pulumi\n'), ((476955, 477003), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_threads"""', 'max_threads'], {}), "(__self__, 'max_threads', max_threads)\n", (476965, 477003), False, 'import pulumi\n'), ((477074, 477170), 'pulumi.set', 'pulumi.set', (['__self__', '"""merge_tree_max_bytes_to_use_cache"""', 'merge_tree_max_bytes_to_use_cache'], {}), "(__self__, 'merge_tree_max_bytes_to_use_cache',\n merge_tree_max_bytes_to_use_cache)\n", (477084, 477170), False, 'import pulumi\n'), ((477236, 477330), 'pulumi.set', 'pulumi.set', (['__self__', '"""merge_tree_max_rows_to_use_cache"""', 'merge_tree_max_rows_to_use_cache'], {}), "(__self__, 'merge_tree_max_rows_to_use_cache',\n merge_tree_max_rows_to_use_cache)\n", (477246, 477330), False, 'import pulumi\n'), ((477404, 477514), 'pulumi.set', 'pulumi.set', (['__self__', '"""merge_tree_min_bytes_for_concurrent_read"""', 'merge_tree_min_bytes_for_concurrent_read'], {}), "(__self__, 'merge_tree_min_bytes_for_concurrent_read',\n merge_tree_min_bytes_for_concurrent_read)\n", (477414, 477514), False, 'import pulumi\n'), ((477587, 477695), 'pulumi.set', 'pulumi.set', (['__self__', '"""merge_tree_min_rows_for_concurrent_read"""', 'merge_tree_min_rows_for_concurrent_read'], {}), "(__self__, 'merge_tree_min_rows_for_concurrent_read',\n merge_tree_min_rows_for_concurrent_read)\n", (477597, 477695), False, 'import pulumi\n'), ((477755, 477833), 'pulumi.set', 'pulumi.set', (['__self__', '"""min_bytes_to_use_direct_io"""', 'min_bytes_to_use_direct_io'], {}), "(__self__, 'min_bytes_to_use_direct_io', min_bytes_to_use_direct_io)\n", (477765, 477833), False, 'import pulumi\n'), ((477891, 477957), 'pulumi.set', 'pulumi.set', (['__self__', '"""min_count_to_compile"""', 'min_count_to_compile'], {}), "(__self__, 'min_count_to_compile', min_count_to_compile)\n", (477901, 477957), False, 'import pulumi\n'), ((478026, 478118), 'pulumi.set', 'pulumi.set', (['__self__', '"""min_count_to_compile_expression"""', 'min_count_to_compile_expression'], {}), "(__self__, 'min_count_to_compile_expression',\n min_count_to_compile_expression)\n", (478036, 478118), False, 'import pulumi\n'), ((478171, 478235), 'pulumi.set', 'pulumi.set', (['__self__', '"""min_execution_speed"""', 'min_execution_speed'], {}), "(__self__, 'min_execution_speed', min_execution_speed)\n", (478181, 478235), False, 'import pulumi\n'), ((478298, 478374), 'pulumi.set', 'pulumi.set', (['__self__', '"""min_execution_speed_bytes"""', 'min_execution_speed_bytes'], {}), "(__self__, 'min_execution_speed_bytes', min_execution_speed_bytes)\n", (478308, 478374), False, 'import pulumi\n'), ((478439, 478524), 'pulumi.set', 'pulumi.set', (['__self__', '"""min_insert_block_size_bytes"""', 'min_insert_block_size_bytes'], {}), "(__self__, 'min_insert_block_size_bytes', min_insert_block_size_bytes\n )\n", (478449, 478524), False, 'import pulumi\n'), ((478583, 478661), 'pulumi.set', 'pulumi.set', (['__self__', '"""min_insert_block_size_rows"""', 'min_insert_block_size_rows'], {}), "(__self__, 'min_insert_block_size_rows', min_insert_block_size_rows)\n", (478593, 478661), False, 'import pulumi\n'), ((478737, 478843), 'pulumi.set', 'pulumi.set', (['__self__', '"""output_format_json_quote64bit_integers"""', 'output_format_json_quote64bit_integers'], {}), "(__self__, 'output_format_json_quote64bit_integers',\n output_format_json_quote64bit_integers)\n", (478747, 478843), False, 'import pulumi\n'), ((478911, 479009), 'pulumi.set', 'pulumi.set', (['__self__', '"""output_format_json_quote_denormals"""', 'output_format_json_quote_denormals'], {}), "(__self__, 'output_format_json_quote_denormals',\n output_format_json_quote_denormals)\n", (478921, 479009), False, 'import pulumi\n'), ((479051, 479093), 'pulumi.set', 'pulumi.set', (['__self__', '"""priority"""', 'priority'], {}), "(__self__, 'priority', priority)\n", (479061, 479093), False, 'import pulumi\n'), ((479141, 479187), 'pulumi.set', 'pulumi.set', (['__self__', '"""quota_mode"""', 'quota_mode'], {}), "(__self__, 'quota_mode', quota_mode)\n", (479151, 479187), False, 'import pulumi\n'), ((479243, 479305), 'pulumi.set', 'pulumi.set', (['__self__', '"""read_overflow_mode"""', 'read_overflow_mode'], {}), "(__self__, 'read_overflow_mode', read_overflow_mode)\n", (479253, 479305), False, 'import pulumi\n'), ((479351, 479393), 'pulumi.set', 'pulumi.set', (['__self__', '"""readonly"""', 'readonly'], {}), "(__self__, 'readonly', readonly)\n", (479361, 479393), False, 'import pulumi\n'), ((479446, 479502), 'pulumi.set', 'pulumi.set', (['__self__', '"""receive_timeout"""', 'receive_timeout'], {}), "(__self__, 'receive_timeout', receive_timeout)\n", (479456, 479502), False, 'import pulumi\n'), ((479573, 479669), 'pulumi.set', 'pulumi.set', (['__self__', '"""replication_alter_partitions_sync"""', 'replication_alter_partitions_sync'], {}), "(__self__, 'replication_alter_partitions_sync',\n replication_alter_partitions_sync)\n", (479583, 479669), False, 'import pulumi\n'), ((479723, 479789), 'pulumi.set', 'pulumi.set', (['__self__', '"""result_overflow_mode"""', 'result_overflow_mode'], {}), "(__self__, 'result_overflow_mode', result_overflow_mode)\n", (479733, 479789), False, 'import pulumi\n'), ((479856, 479944), 'pulumi.set', 'pulumi.set', (['__self__', '"""select_sequential_consistency"""', 'select_sequential_consistency'], {}), "(__self__, 'select_sequential_consistency',\n select_sequential_consistency)\n", (479866, 479944), False, 'import pulumi\n'), ((480007, 480095), 'pulumi.set', 'pulumi.set', (['__self__', '"""send_progress_in_http_headers"""', 'send_progress_in_http_headers'], {}), "(__self__, 'send_progress_in_http_headers',\n send_progress_in_http_headers)\n", (480017, 480095), False, 'import pulumi\n'), ((480141, 480191), 'pulumi.set', 'pulumi.set', (['__self__', '"""send_timeout"""', 'send_timeout'], {}), "(__self__, 'send_timeout', send_timeout)\n", (480151, 480191), False, 'import pulumi\n'), ((480246, 480306), 'pulumi.set', 'pulumi.set', (['__self__', '"""set_overflow_mode"""', 'set_overflow_mode'], {}), "(__self__, 'set_overflow_mode', set_overflow_mode)\n", (480256, 480306), False, 'import pulumi\n'), ((480367, 480439), 'pulumi.set', 'pulumi.set', (['__self__', '"""skip_unavailable_shards"""', 'skip_unavailable_shards'], {}), "(__self__, 'skip_unavailable_shards', skip_unavailable_shards)\n", (480377, 480439), False, 'import pulumi\n'), ((480495, 480557), 'pulumi.set', 'pulumi.set', (['__self__', '"""sort_overflow_mode"""', 'sort_overflow_mode'], {}), "(__self__, 'sort_overflow_mode', sort_overflow_mode)\n", (480505, 480557), False, 'import pulumi\n'), ((480616, 480684), 'pulumi.set', 'pulumi.set', (['__self__', '"""timeout_overflow_mode"""', 'timeout_overflow_mode'], {}), "(__self__, 'timeout_overflow_mode', timeout_overflow_mode)\n", (480626, 480684), False, 'import pulumi\n'), ((480744, 480814), 'pulumi.set', 'pulumi.set', (['__self__', '"""transfer_overflow_mode"""', 'transfer_overflow_mode'], {}), "(__self__, 'transfer_overflow_mode', transfer_overflow_mode)\n", (480754, 480814), False, 'import pulumi\n'), ((480869, 480929), 'pulumi.set', 'pulumi.set', (['__self__', '"""transform_null_in"""', 'transform_null_in'], {}), "(__self__, 'transform_null_in', transform_null_in)\n", (480879, 480929), False, 'import pulumi\n'), ((480989, 481059), 'pulumi.set', 'pulumi.set', (['__self__', '"""use_uncompressed_cache"""', 'use_uncompressed_cache'], {}), "(__self__, 'use_uncompressed_cache', use_uncompressed_cache)\n", (480999, 481059), False, 'import pulumi\n'), ((525456, 525500), 'pulumi.set', 'pulumi.set', (['__self__', '"""resources"""', 'resources'], {}), "(__self__, 'resources', resources)\n", (525466, 525500), False, 'import pulumi\n'), ((526683, 526727), 'pulumi.set', 'pulumi.set', (['__self__', '"""disk_size"""', 'disk_size'], {}), "(__self__, 'disk_size', disk_size)\n", (526693, 526727), False, 'import pulumi\n'), ((526777, 526827), 'pulumi.set', 'pulumi.set', (['__self__', '"""disk_type_id"""', 'disk_type_id'], {}), "(__self__, 'disk_type_id', disk_type_id)\n", (526787, 526827), False, 'import pulumi\n'), ((526883, 526945), 'pulumi.set', 'pulumi.set', (['__self__', '"""resource_preset_id"""', 'resource_preset_id'], {}), "(__self__, 'resource_preset_id', resource_preset_id)\n", (526893, 526945), False, 'import pulumi\n'), ((529720, 529760), 'pulumi.set', 'pulumi.set', (['__self__', '"""edition"""', 'edition'], {}), "(__self__, 'edition', edition)\n", (529730, 529760), False, 'import pulumi\n'), ((529809, 529857), 'pulumi.set', 'pulumi.set', (['__self__', '"""master_node"""', 'master_node'], {}), "(__self__, 'master_node', master_node)\n", (529819, 529857), False, 'import pulumi\n'), ((529902, 529942), 'pulumi.set', 'pulumi.set', (['__self__', '"""plugins"""', 'plugins'], {}), "(__self__, 'plugins', plugins)\n", (529912, 529942), False, 'import pulumi\n'), ((529987, 530027), 'pulumi.set', 'pulumi.set', (['__self__', '"""version"""', 'version'], {}), "(__self__, 'version', version)\n", (529997, 530027), False, 'import pulumi\n'), ((539239, 539297), 'pulumi.set', 'pulumi.set', (['__self__', '"""assign_public_ip"""', 'assign_public_ip'], {}), "(__self__, 'assign_public_ip', assign_public_ip)\n", (539249, 539297), False, 'import pulumi\n'), ((539339, 539373), 'pulumi.set', 'pulumi.set', (['__self__', '"""fqdn"""', 'fqdn'], {}), "(__self__, 'fqdn', fqdn)\n", (539349, 539373), False, 'import pulumi\n'), ((539420, 539464), 'pulumi.set', 'pulumi.set', (['__self__', '"""subnet_id"""', 'subnet_id'], {}), "(__self__, 'subnet_id', subnet_id)\n", (539430, 539464), False, 'import pulumi\n'), ((542236, 542280), 'pulumi.set', 'pulumi.set', (['__self__', '"""data_lens"""', 'data_lens'], {}), "(__self__, 'data_lens', data_lens)\n", (542246, 542280), False, 'import pulumi\n'), ((542325, 542365), 'pulumi.set', 'pulumi.set', (['__self__', '"""web_sql"""', 'web_sql'], {}), "(__self__, 'web_sql', web_sql)\n", (542335, 542365), False, 'import pulumi\n'), ((543588, 543624), 'pulumi.set', 'pulumi.set', (['__self__', '"""hours"""', 'hours'], {}), "(__self__, 'hours', hours)\n", (543598, 543624), False, 'import pulumi\n'), ((543669, 543709), 'pulumi.set', 'pulumi.set', (['__self__', '"""minutes"""', 'minutes'], {}), "(__self__, 'minutes', minutes)\n", (543679, 543709), False, 'import pulumi\n'), ((544969, 545027), 'pulumi.set', 'pulumi.set', (['__self__', '"""assign_public_ip"""', 'assign_public_ip'], {}), "(__self__, 'assign_public_ip', assign_public_ip)\n", (544979, 545027), False, 'import pulumi\n'), ((545069, 545103), 'pulumi.set', 'pulumi.set', (['__self__', '"""fqdn"""', 'fqdn'], {}), "(__self__, 'fqdn', fqdn)\n", (545079, 545103), False, 'import pulumi\n'), ((548505, 548539), 'pulumi.set', 'pulumi.set', (['__self__', '"""fqdn"""', 'fqdn'], {}), "(__self__, 'fqdn', fqdn)\n", (548515, 548539), False, 'import pulumi\n'), ((552957, 553015), 'pulumi.set', 'pulumi.set', (['__self__', '"""assign_public_ip"""', 'assign_public_ip'], {}), "(__self__, 'assign_public_ip', assign_public_ip)\n", (552967, 553015), False, 'import pulumi\n'), ((553066, 553118), 'pulumi.set', 'pulumi.set', (['__self__', '"""brokers_count"""', 'brokers_count'], {}), "(__self__, 'brokers_count', brokers_count)\n", (553076, 553118), False, 'import pulumi\n'), ((553171, 553227), 'pulumi.set', 'pulumi.set', (['__self__', '"""schema_registry"""', 'schema_registry'], {}), "(__self__, 'schema_registry', schema_registry)\n", (553181, 553227), False, 'import pulumi\n'), ((553281, 553339), 'pulumi.set', 'pulumi.set', (['__self__', '"""unmanaged_topics"""', 'unmanaged_topics'], {}), "(__self__, 'unmanaged_topics', unmanaged_topics)\n", (553291, 553339), False, 'import pulumi\n'), ((553386, 553430), 'pulumi.set', 'pulumi.set', (['__self__', '"""zookeeper"""', 'zookeeper'], {}), "(__self__, 'zookeeper', zookeeper)\n", (553396, 553430), False, 'import pulumi\n'), ((557338, 557388), 'pulumi.set', 'pulumi.set', (['__self__', '"""kafka_config"""', 'kafka_config'], {}), "(__self__, 'kafka_config', kafka_config)\n", (557348, 557388), False, 'import pulumi\n'), ((559633, 559709), 'pulumi.set', 'pulumi.set', (['__self__', '"""auto_create_topics_enable"""', 'auto_create_topics_enable'], {}), "(__self__, 'auto_create_topics_enable', auto_create_topics_enable)\n", (559643, 559709), False, 'import pulumi\n'), ((559763, 559821), 'pulumi.set', 'pulumi.set', (['__self__', '"""compression_type"""', 'compression_type'], {}), "(__self__, 'compression_type', compression_type)\n", (559773, 559821), False, 'import pulumi\n'), ((559885, 559963), 'pulumi.set', 'pulumi.set', (['__self__', '"""default_replication_factor"""', 'default_replication_factor'], {}), "(__self__, 'default_replication_factor', default_replication_factor)\n", (559895, 559963), False, 'import pulumi\n'), ((560028, 560113), 'pulumi.set', 'pulumi.set', (['__self__', '"""log_flush_interval_messages"""', 'log_flush_interval_messages'], {}), "(__self__, 'log_flush_interval_messages', log_flush_interval_messages\n )\n", (560038, 560113), False, 'import pulumi\n'), ((560167, 560235), 'pulumi.set', 'pulumi.set', (['__self__', '"""log_flush_interval_ms"""', 'log_flush_interval_ms'], {}), "(__self__, 'log_flush_interval_ms', log_flush_interval_ms)\n", (560177, 560235), False, 'import pulumi\n'), ((560304, 560396), 'pulumi.set', 'pulumi.set', (['__self__', '"""log_flush_scheduler_interval_ms"""', 'log_flush_scheduler_interval_ms'], {}), "(__self__, 'log_flush_scheduler_interval_ms',\n log_flush_scheduler_interval_ms)\n", (560314, 560396), False, 'import pulumi\n'), ((560445, 560501), 'pulumi.set', 'pulumi.set', (['__self__', '"""log_preallocate"""', 'log_preallocate'], {}), "(__self__, 'log_preallocate', log_preallocate)\n", (560455, 560501), False, 'import pulumi\n'), ((560558, 560622), 'pulumi.set', 'pulumi.set', (['__self__', '"""log_retention_bytes"""', 'log_retention_bytes'], {}), "(__self__, 'log_retention_bytes', log_retention_bytes)\n", (560568, 560622), False, 'import pulumi\n'), ((560679, 560743), 'pulumi.set', 'pulumi.set', (['__self__', '"""log_retention_hours"""', 'log_retention_hours'], {}), "(__self__, 'log_retention_hours', log_retention_hours)\n", (560689, 560743), False, 'import pulumi\n'), ((560802, 560870), 'pulumi.set', 'pulumi.set', (['__self__', '"""log_retention_minutes"""', 'log_retention_minutes'], {}), "(__self__, 'log_retention_minutes', log_retention_minutes)\n", (560812, 560870), False, 'import pulumi\n'), ((560924, 560982), 'pulumi.set', 'pulumi.set', (['__self__', '"""log_retention_ms"""', 'log_retention_ms'], {}), "(__self__, 'log_retention_ms', log_retention_ms)\n", (560934, 560982), False, 'import pulumi\n'), ((561037, 561097), 'pulumi.set', 'pulumi.set', (['__self__', '"""log_segment_bytes"""', 'log_segment_bytes'], {}), "(__self__, 'log_segment_bytes', log_segment_bytes)\n", (561047, 561097), False, 'import pulumi\n'), ((561149, 561203), 'pulumi.set', 'pulumi.set', (['__self__', '"""num_partitions"""', 'num_partitions'], {}), "(__self__, 'num_partitions', num_partitions)\n", (561159, 561203), False, 'import pulumi\n'), ((561268, 561353), 'pulumi.set', 'pulumi.set', (['__self__', '"""socket_receive_buffer_bytes"""', 'socket_receive_buffer_bytes'], {}), "(__self__, 'socket_receive_buffer_bytes', socket_receive_buffer_bytes\n )\n", (561278, 561353), False, 'import pulumi\n'), ((561410, 561484), 'pulumi.set', 'pulumi.set', (['__self__', '"""socket_send_buffer_bytes"""', 'socket_send_buffer_bytes'], {}), "(__self__, 'socket_send_buffer_bytes', socket_send_buffer_bytes)\n", (561420, 561484), False, 'import pulumi\n'), ((569170, 569214), 'pulumi.set', 'pulumi.set', (['__self__', '"""resources"""', 'resources'], {}), "(__self__, 'resources', resources)\n", (569180, 569214), False, 'import pulumi\n'), ((570395, 570439), 'pulumi.set', 'pulumi.set', (['__self__', '"""disk_size"""', 'disk_size'], {}), "(__self__, 'disk_size', disk_size)\n", (570405, 570439), False, 'import pulumi\n'), ((570489, 570539), 'pulumi.set', 'pulumi.set', (['__self__', '"""disk_type_id"""', 'disk_type_id'], {}), "(__self__, 'disk_type_id', disk_type_id)\n", (570499, 570539), False, 'import pulumi\n'), ((570595, 570657), 'pulumi.set', 'pulumi.set', (['__self__', '"""resource_preset_id"""', 'resource_preset_id'], {}), "(__self__, 'resource_preset_id', resource_preset_id)\n", (570605, 570657), False, 'import pulumi\n'), ((572939, 572997), 'pulumi.set', 'pulumi.set', (['__self__', '"""assign_public_ip"""', 'assign_public_ip'], {}), "(__self__, 'assign_public_ip', assign_public_ip)\n", (572949, 572997), False, 'import pulumi\n'), ((573041, 573079), 'pulumi.set', 'pulumi.set', (['__self__', '"""health"""', 'health'], {}), "(__self__, 'health', health)\n", (573051, 573079), False, 'import pulumi\n'), ((573121, 573155), 'pulumi.set', 'pulumi.set', (['__self__', '"""name"""', 'name'], {}), "(__self__, 'name', name)\n", (573131, 573155), False, 'import pulumi\n'), ((573197, 573231), 'pulumi.set', 'pulumi.set', (['__self__', '"""role"""', 'role'], {}), "(__self__, 'role', role)\n", (573207, 573231), False, 'import pulumi\n'), ((573278, 573322), 'pulumi.set', 'pulumi.set', (['__self__', '"""subnet_id"""', 'subnet_id'], {}), "(__self__, 'subnet_id', subnet_id)\n", (573288, 573322), False, 'import pulumi\n'), ((573367, 573407), 'pulumi.set', 'pulumi.set', (['__self__', '"""zone_id"""', 'zone_id'], {}), "(__self__, 'zone_id', zone_id)\n", (573377, 573407), False, 'import pulumi\n'), ((576466, 576516), 'pulumi.set', 'pulumi.set', (['__self__', '"""topic_config"""', 'topic_config'], {}), "(__self__, 'topic_config', topic_config)\n", (576476, 576516), False, 'import pulumi\n'), ((579052, 579106), 'pulumi.set', 'pulumi.set', (['__self__', '"""cleanup_policy"""', 'cleanup_policy'], {}), "(__self__, 'cleanup_policy', cleanup_policy)\n", (579062, 579106), False, 'import pulumi\n'), ((579160, 579218), 'pulumi.set', 'pulumi.set', (['__self__', '"""compression_type"""', 'compression_type'], {}), "(__self__, 'compression_type', compression_type)\n", (579170, 579218), False, 'import pulumi\n'), ((579275, 579339), 'pulumi.set', 'pulumi.set', (['__self__', '"""delete_retention_ms"""', 'delete_retention_ms'], {}), "(__self__, 'delete_retention_ms', delete_retention_ms)\n", (579285, 579339), False, 'import pulumi\n'), ((579397, 579463), 'pulumi.set', 'pulumi.set', (['__self__', '"""file_delete_delay_ms"""', 'file_delete_delay_ms'], {}), "(__self__, 'file_delete_delay_ms', file_delete_delay_ms)\n", (579407, 579463), False, 'import pulumi\n'), ((579515, 579569), 'pulumi.set', 'pulumi.set', (['__self__', '"""flush_messages"""', 'flush_messages'], {}), "(__self__, 'flush_messages', flush_messages)\n", (579525, 579569), False, 'import pulumi\n'), ((579615, 579657), 'pulumi.set', 'pulumi.set', (['__self__', '"""flush_ms"""', 'flush_ms'], {}), "(__self__, 'flush_ms', flush_ms)\n", (579625, 579657), False, 'import pulumi\n'), ((579712, 579772), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_message_bytes"""', 'max_message_bytes'], {}), "(__self__, 'max_message_bytes', max_message_bytes)\n", (579722, 579772), False, 'import pulumi\n'), ((579831, 579899), 'pulumi.set', 'pulumi.set', (['__self__', '"""min_compaction_lag_ms"""', 'min_compaction_lag_ms'], {}), "(__self__, 'min_compaction_lag_ms', min_compaction_lag_ms)\n", (579841, 579899), False, 'import pulumi\n'), ((579956, 580020), 'pulumi.set', 'pulumi.set', (['__self__', '"""min_insync_replicas"""', 'min_insync_replicas'], {}), "(__self__, 'min_insync_replicas', min_insync_replicas)\n", (579966, 580020), False, 'import pulumi\n'), ((580069, 580117), 'pulumi.set', 'pulumi.set', (['__self__', '"""preallocate"""', 'preallocate'], {}), "(__self__, 'preallocate', preallocate)\n", (580079, 580117), False, 'import pulumi\n'), ((580170, 580226), 'pulumi.set', 'pulumi.set', (['__self__', '"""retention_bytes"""', 'retention_bytes'], {}), "(__self__, 'retention_bytes', retention_bytes)\n", (580180, 580226), False, 'import pulumi\n'), ((580276, 580326), 'pulumi.set', 'pulumi.set', (['__self__', '"""retention_ms"""', 'retention_ms'], {}), "(__self__, 'retention_ms', retention_ms)\n", (580286, 580326), False, 'import pulumi\n'), ((580377, 580429), 'pulumi.set', 'pulumi.set', (['__self__', '"""segment_bytes"""', 'segment_bytes'], {}), "(__self__, 'segment_bytes', segment_bytes)\n", (580387, 580429), False, 'import pulumi\n'), ((585323, 585371), 'pulumi.set', 'pulumi.set', (['__self__', '"""permissions"""', 'permissions'], {}), "(__self__, 'permissions', permissions)\n", (585333, 585371), False, 'import pulumi\n'), ((588623, 588677), 'pulumi.set', 'pulumi.set', (['__self__', '"""cleanup_policy"""', 'cleanup_policy'], {}), "(__self__, 'cleanup_policy', cleanup_policy)\n", (588633, 588677), False, 'import pulumi\n'), ((588731, 588789), 'pulumi.set', 'pulumi.set', (['__self__', '"""compression_type"""', 'compression_type'], {}), "(__self__, 'compression_type', compression_type)\n", (588741, 588789), False, 'import pulumi\n'), ((588846, 588910), 'pulumi.set', 'pulumi.set', (['__self__', '"""delete_retention_ms"""', 'delete_retention_ms'], {}), "(__self__, 'delete_retention_ms', delete_retention_ms)\n", (588856, 588910), False, 'import pulumi\n'), ((588968, 589034), 'pulumi.set', 'pulumi.set', (['__self__', '"""file_delete_delay_ms"""', 'file_delete_delay_ms'], {}), "(__self__, 'file_delete_delay_ms', file_delete_delay_ms)\n", (588978, 589034), False, 'import pulumi\n'), ((589086, 589140), 'pulumi.set', 'pulumi.set', (['__self__', '"""flush_messages"""', 'flush_messages'], {}), "(__self__, 'flush_messages', flush_messages)\n", (589096, 589140), False, 'import pulumi\n'), ((589186, 589228), 'pulumi.set', 'pulumi.set', (['__self__', '"""flush_ms"""', 'flush_ms'], {}), "(__self__, 'flush_ms', flush_ms)\n", (589196, 589228), False, 'import pulumi\n'), ((589283, 589343), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_message_bytes"""', 'max_message_bytes'], {}), "(__self__, 'max_message_bytes', max_message_bytes)\n", (589293, 589343), False, 'import pulumi\n'), ((589402, 589470), 'pulumi.set', 'pulumi.set', (['__self__', '"""min_compaction_lag_ms"""', 'min_compaction_lag_ms'], {}), "(__self__, 'min_compaction_lag_ms', min_compaction_lag_ms)\n", (589412, 589470), False, 'import pulumi\n'), ((589527, 589591), 'pulumi.set', 'pulumi.set', (['__self__', '"""min_insync_replicas"""', 'min_insync_replicas'], {}), "(__self__, 'min_insync_replicas', min_insync_replicas)\n", (589537, 589591), False, 'import pulumi\n'), ((589640, 589688), 'pulumi.set', 'pulumi.set', (['__self__', '"""preallocate"""', 'preallocate'], {}), "(__self__, 'preallocate', preallocate)\n", (589650, 589688), False, 'import pulumi\n'), ((589741, 589797), 'pulumi.set', 'pulumi.set', (['__self__', '"""retention_bytes"""', 'retention_bytes'], {}), "(__self__, 'retention_bytes', retention_bytes)\n", (589751, 589797), False, 'import pulumi\n'), ((589847, 589897), 'pulumi.set', 'pulumi.set', (['__self__', '"""retention_ms"""', 'retention_ms'], {}), "(__self__, 'retention_ms', retention_ms)\n", (589857, 589897), False, 'import pulumi\n'), ((589948, 590000), 'pulumi.set', 'pulumi.set', (['__self__', '"""segment_bytes"""', 'segment_bytes'], {}), "(__self__, 'segment_bytes', segment_bytes)\n", (589958, 590000), False, 'import pulumi\n'), ((595316, 595354), 'pulumi.set', 'pulumi.set', (['__self__', '"""access"""', 'access'], {}), "(__self__, 'access', access)\n", (595326, 595354), False, 'import pulumi\n'), ((595411, 595475), 'pulumi.set', 'pulumi.set', (['__self__', '"""backup_window_start"""', 'backup_window_start'], {}), "(__self__, 'backup_window_start', backup_window_start)\n", (595421, 595475), False, 'import pulumi\n'), ((595542, 595630), 'pulumi.set', 'pulumi.set', (['__self__', '"""feature_compatibility_version"""', 'feature_compatibility_version'], {}), "(__self__, 'feature_compatibility_version',\n feature_compatibility_version)\n", (595552, 595630), False, 'import pulumi\n'), ((597804, 597848), 'pulumi.set', 'pulumi.set', (['__self__', '"""data_lens"""', 'data_lens'], {}), "(__self__, 'data_lens', data_lens)\n", (597814, 597848), False, 'import pulumi\n'), ((598658, 598694), 'pulumi.set', 'pulumi.set', (['__self__', '"""hours"""', 'hours'], {}), "(__self__, 'hours', hours)\n", (598668, 598694), False, 'import pulumi\n'), ((598739, 598779), 'pulumi.set', 'pulumi.set', (['__self__', '"""minutes"""', 'minutes'], {}), "(__self__, 'minutes', minutes)\n", (598749, 598779), False, 'import pulumi\n'), ((601880, 601938), 'pulumi.set', 'pulumi.set', (['__self__', '"""assign_public_ip"""', 'assign_public_ip'], {}), "(__self__, 'assign_public_ip', assign_public_ip)\n", (601890, 601938), False, 'import pulumi\n'), ((601982, 602020), 'pulumi.set', 'pulumi.set', (['__self__', '"""health"""', 'health'], {}), "(__self__, 'health', health)\n", (601992, 602020), False, 'import pulumi\n'), ((602062, 602096), 'pulumi.set', 'pulumi.set', (['__self__', '"""name"""', 'name'], {}), "(__self__, 'name', name)\n", (602072, 602096), False, 'import pulumi\n'), ((602138, 602172), 'pulumi.set', 'pulumi.set', (['__self__', '"""role"""', 'role'], {}), "(__self__, 'role', role)\n", (602148, 602172), False, 'import pulumi\n'), ((602220, 602266), 'pulumi.set', 'pulumi.set', (['__self__', '"""shard_name"""', 'shard_name'], {}), "(__self__, 'shard_name', shard_name)\n", (602230, 602266), False, 'import pulumi\n'), ((602308, 602342), 'pulumi.set', 'pulumi.set', (['__self__', '"""type"""', 'type'], {}), "(__self__, 'type', type)\n", (602318, 602342), False, 'import pulumi\n'), ((606247, 606279), 'pulumi.set', 'pulumi.set', (['__self__', '"""day"""', 'day'], {}), "(__self__, 'day', day)\n", (606257, 606279), False, 'import pulumi\n'), ((606321, 606355), 'pulumi.set', 'pulumi.set', (['__self__', '"""hour"""', 'hour'], {}), "(__self__, 'hour', hour)\n", (606331, 606355), False, 'import pulumi\n'), ((610224, 610272), 'pulumi.set', 'pulumi.set', (['__self__', '"""permissions"""', 'permissions'], {}), "(__self__, 'permissions', permissions)\n", (610234, 610272), False, 'import pulumi\n'), ((611856, 611892), 'pulumi.set', 'pulumi.set', (['__self__', '"""roles"""', 'roles'], {}), "(__self__, 'roles', roles)\n", (611866, 611892), False, 'import pulumi\n'), ((613145, 613189), 'pulumi.set', 'pulumi.set', (['__self__', '"""data_lens"""', 'data_lens'], {}), "(__self__, 'data_lens', data_lens)\n", (613155, 613189), False, 'import pulumi\n'), ((613234, 613274), 'pulumi.set', 'pulumi.set', (['__self__', '"""web_sql"""', 'web_sql'], {}), "(__self__, 'web_sql', web_sql)\n", (613244, 613274), False, 'import pulumi\n'), ((614554, 614590), 'pulumi.set', 'pulumi.set', (['__self__', '"""hours"""', 'hours'], {}), "(__self__, 'hours', hours)\n", (614564, 614590), False, 'import pulumi\n'), ((614635, 614675), 'pulumi.set', 'pulumi.set', (['__self__', '"""minutes"""', 'minutes'], {}), "(__self__, 'minutes', minutes)\n", (614645, 614675), False, 'import pulumi\n'), ((618024, 618082), 'pulumi.set', 'pulumi.set', (['__self__', '"""assign_public_ip"""', 'assign_public_ip'], {}), "(__self__, 'assign_public_ip', assign_public_ip)\n", (618034, 618082), False, 'import pulumi\n'), ((618124, 618158), 'pulumi.set', 'pulumi.set', (['__self__', '"""fqdn"""', 'fqdn'], {}), "(__self__, 'fqdn', fqdn)\n", (618134, 618158), False, 'import pulumi\n'), ((618200, 618234), 'pulumi.set', 'pulumi.set', (['__self__', '"""name"""', 'name'], {}), "(__self__, 'name', name)\n", (618210, 618234), False, 'import pulumi\n'), ((618290, 618352), 'pulumi.set', 'pulumi.set', (['__self__', '"""replication_source"""', 'replication_source'], {}), "(__self__, 'replication_source', replication_source)\n", (618300, 618352), False, 'import pulumi\n'), ((618413, 618485), 'pulumi.set', 'pulumi.set', (['__self__', '"""replication_source_name"""', 'replication_source_name'], {}), "(__self__, 'replication_source_name', replication_source_name)\n", (618423, 618485), False, 'import pulumi\n'), ((618532, 618576), 'pulumi.set', 'pulumi.set', (['__self__', '"""subnet_id"""', 'subnet_id'], {}), "(__self__, 'subnet_id', subnet_id)\n", (618542, 618576), False, 'import pulumi\n'), ((622390, 622422), 'pulumi.set', 'pulumi.set', (['__self__', '"""day"""', 'day'], {}), "(__self__, 'day', day)\n", (622400, 622422), False, 'import pulumi\n'), ((622464, 622498), 'pulumi.set', 'pulumi.set', (['__self__', '"""hour"""', 'hour'], {}), "(__self__, 'hour', hour)\n", (622474, 622498), False, 'import pulumi\n'), ((625962, 625996), 'pulumi.set', 'pulumi.set', (['__self__', '"""time"""', 'time'], {}), "(__self__, 'time', time)\n", (625972, 625996), False, 'import pulumi\n'), ((628873, 628941), 'pulumi.set', 'pulumi.set', (['__self__', '"""authentication_plugin"""', 'authentication_plugin'], {}), "(__self__, 'authentication_plugin', authentication_plugin)\n", (628883, 628941), False, 'import pulumi\n'), ((628996, 629056), 'pulumi.set', 'pulumi.set', (['__self__', '"""connection_limits"""', 'connection_limits'], {}), "(__self__, 'connection_limits', connection_limits)\n", (629006, 629056), False, 'import pulumi\n'), ((629112, 629174), 'pulumi.set', 'pulumi.set', (['__self__', '"""global_permissions"""', 'global_permissions'], {}), "(__self__, 'global_permissions', global_permissions)\n", (629122, 629174), False, 'import pulumi\n'), ((629223, 629271), 'pulumi.set', 'pulumi.set', (['__self__', '"""permissions"""', 'permissions'], {}), "(__self__, 'permissions', permissions)\n", (629233, 629271), False, 'import pulumi\n'), ((633093, 633167), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_connections_per_hour"""', 'max_connections_per_hour'], {}), "(__self__, 'max_connections_per_hour', max_connections_per_hour)\n", (633103, 633167), False, 'import pulumi\n'), ((633227, 633297), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_questions_per_hour"""', 'max_questions_per_hour'], {}), "(__self__, 'max_questions_per_hour', max_questions_per_hour)\n", (633237, 633297), False, 'import pulumi\n'), ((633355, 633421), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_updates_per_hour"""', 'max_updates_per_hour'], {}), "(__self__, 'max_updates_per_hour', max_updates_per_hour)\n", (633365, 633421), False, 'import pulumi\n'), ((633479, 633545), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_user_connections"""', 'max_user_connections'], {}), "(__self__, 'max_user_connections', max_user_connections)\n", (633489, 633545), False, 'import pulumi\n'), ((635998, 636034), 'pulumi.set', 'pulumi.set', (['__self__', '"""roles"""', 'roles'], {}), "(__self__, 'roles', roles)\n", (636008, 636034), False, 'import pulumi\n'), ((638733, 638777), 'pulumi.set', 'pulumi.set', (['__self__', '"""databases"""', 'databases'], {}), "(__self__, 'databases', databases)\n", (638743, 638777), False, 'import pulumi\n'), ((638831, 638889), 'pulumi.set', 'pulumi.set', (['__self__', '"""maxmemory_policy"""', 'maxmemory_policy'], {}), "(__self__, 'maxmemory_policy', maxmemory_policy)\n", (638841, 638889), False, 'import pulumi\n'), ((638949, 639019), 'pulumi.set', 'pulumi.set', (['__self__', '"""notify_keyspace_events"""', 'notify_keyspace_events'], {}), "(__self__, 'notify_keyspace_events', notify_keyspace_events)\n", (638959, 639019), False, 'import pulumi\n'), ((639080, 639152), 'pulumi.set', 'pulumi.set', (['__self__', '"""slowlog_log_slower_than"""', 'slowlog_log_slower_than'], {}), "(__self__, 'slowlog_log_slower_than', slowlog_log_slower_than)\n", (639090, 639152), False, 'import pulumi\n'), ((639205, 639261), 'pulumi.set', 'pulumi.set', (['__self__', '"""slowlog_max_len"""', 'slowlog_max_len'], {}), "(__self__, 'slowlog_max_len', slowlog_max_len)\n", (639215, 639261), False, 'import pulumi\n'), ((639306, 639346), 'pulumi.set', 'pulumi.set', (['__self__', '"""timeout"""', 'timeout'], {}), "(__self__, 'timeout', timeout)\n", (639316, 639346), False, 'import pulumi\n'), ((643523, 643557), 'pulumi.set', 'pulumi.set', (['__self__', '"""fqdn"""', 'fqdn'], {}), "(__self__, 'fqdn', fqdn)\n", (643533, 643557), False, 'import pulumi\n'), ((643605, 643651), 'pulumi.set', 'pulumi.set', (['__self__', '"""shard_name"""', 'shard_name'], {}), "(__self__, 'shard_name', shard_name)\n", (643615, 643651), False, 'import pulumi\n'), ((643698, 643742), 'pulumi.set', 'pulumi.set', (['__self__', '"""subnet_id"""', 'subnet_id'], {}), "(__self__, 'subnet_id', subnet_id)\n", (643708, 643742), False, 'import pulumi\n'), ((646122, 646154), 'pulumi.set', 'pulumi.set', (['__self__', '"""day"""', 'day'], {}), "(__self__, 'day', day)\n", (646132, 646154), False, 'import pulumi\n'), ((646196, 646230), 'pulumi.set', 'pulumi.set', (['__self__', '"""hour"""', 'hour'], {}), "(__self__, 'hour', hour)\n", (646206, 646230), False, 'import pulumi\n'), ((648042, 648092), 'pulumi.set', 'pulumi.set', (['__self__', '"""disk_type_id"""', 'disk_type_id'], {}), "(__self__, 'disk_type_id', disk_type_id)\n", (648052, 648092), False, 'import pulumi\n'), ((649620, 649656), 'pulumi.set', 'pulumi.set', (['__self__', '"""hours"""', 'hours'], {}), "(__self__, 'hours', hours)\n", (649630, 649656), False, 'import pulumi\n'), ((649701, 649741), 'pulumi.set', 'pulumi.set', (['__self__', '"""minutes"""', 'minutes'], {}), "(__self__, 'minutes', minutes)\n", (649711, 649741), False, 'import pulumi\n'), ((651921, 651979), 'pulumi.set', 'pulumi.set', (['__self__', '"""assign_public_ip"""', 'assign_public_ip'], {}), "(__self__, 'assign_public_ip', assign_public_ip)\n", (651931, 651979), False, 'import pulumi\n'), ((652021, 652055), 'pulumi.set', 'pulumi.set', (['__self__', '"""fqdn"""', 'fqdn'], {}), "(__self__, 'fqdn', fqdn)\n", (652031, 652055), False, 'import pulumi\n'), ((652102, 652146), 'pulumi.set', 'pulumi.set', (['__self__', '"""subnet_id"""', 'subnet_id'], {}), "(__self__, 'subnet_id', subnet_id)\n", (652112, 652146), False, 'import pulumi\n'), ((656151, 656199), 'pulumi.set', 'pulumi.set', (['__self__', '"""permissions"""', 'permissions'], {}), "(__self__, 'permissions', permissions)\n", (656161, 656199), False, 'import pulumi\n'), ((658009, 658045), 'pulumi.set', 'pulumi.set', (['__self__', '"""roles"""', 'roles'], {}), "(__self__, 'roles', roles)\n", (658019, 658045), False, 'import pulumi\n'), ((660278, 660334), 'pulumi.set', 'pulumi.set', (['__self__', '"""allowed_headers"""', 'allowed_headers'], {}), "(__self__, 'allowed_headers', allowed_headers)\n", (660288, 660334), False, 'import pulumi\n'), ((660386, 660440), 'pulumi.set', 'pulumi.set', (['__self__', '"""expose_headers"""', 'expose_headers'], {}), "(__self__, 'expose_headers', expose_headers)\n", (660396, 660440), False, 'import pulumi\n'), ((660493, 660549), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_age_seconds"""', 'max_age_seconds'], {}), "(__self__, 'max_age_seconds', max_age_seconds)\n", (660503, 660549), False, 'import pulumi\n'), ((663328, 663358), 'pulumi.set', 'pulumi.set', (['__self__', '"""id"""', 'id'], {}), "(__self__, 'id', id)\n", (663338, 663358), False, 'import pulumi\n'), ((663399, 663431), 'pulumi.set', 'pulumi.set', (['__self__', '"""uri"""', 'uri'], {}), "(__self__, 'uri', uri)\n", (663409, 663431), False, 'import pulumi\n'), ((666801, 666907), 'pulumi.set', 'pulumi.set', (['__self__', '"""abort_incomplete_multipart_upload_days"""', 'abort_incomplete_multipart_upload_days'], {}), "(__self__, 'abort_incomplete_multipart_upload_days',\n abort_incomplete_multipart_upload_days)\n", (666811, 666907), False, 'import pulumi\n'), ((666951, 666997), 'pulumi.set', 'pulumi.set', (['__self__', '"""expiration"""', 'expiration'], {}), "(__self__, 'expiration', expiration)\n", (666961, 666997), False, 'import pulumi\n'), ((667037, 667067), 'pulumi.set', 'pulumi.set', (['__self__', '"""id"""', 'id'], {}), "(__self__, 'id', id)\n", (667047, 667067), False, 'import pulumi\n'), ((667134, 667222), 'pulumi.set', 'pulumi.set', (['__self__', '"""noncurrent_version_expiration"""', 'noncurrent_version_expiration'], {}), "(__self__, 'noncurrent_version_expiration',\n noncurrent_version_expiration)\n", (667144, 667222), False, 'import pulumi\n'), ((667286, 667376), 'pulumi.set', 'pulumi.set', (['__self__', '"""noncurrent_version_transitions"""', 'noncurrent_version_transitions'], {}), "(__self__, 'noncurrent_version_transitions',\n noncurrent_version_transitions)\n", (667296, 667376), False, 'import pulumi\n'), ((667416, 667454), 'pulumi.set', 'pulumi.set', (['__self__', '"""prefix"""', 'prefix'], {}), "(__self__, 'prefix', prefix)\n", (667426, 667454), False, 'import pulumi\n'), ((667503, 667551), 'pulumi.set', 'pulumi.set', (['__self__', '"""transitions"""', 'transitions'], {}), "(__self__, 'transitions', transitions)\n", (667513, 667551), False, 'import pulumi\n'), ((672261, 672295), 'pulumi.set', 'pulumi.set', (['__self__', '"""date"""', 'date'], {}), "(__self__, 'date', date)\n", (672271, 672295), False, 'import pulumi\n'), ((672337, 672371), 'pulumi.set', 'pulumi.set', (['__self__', '"""days"""', 'days'], {}), "(__self__, 'days', days)\n", (672347, 672371), False, 'import pulumi\n'), ((672437, 672523), 'pulumi.set', 'pulumi.set', (['__self__', '"""expired_object_delete_marker"""', 'expired_object_delete_marker'], {}), "(__self__, 'expired_object_delete_marker',\n expired_object_delete_marker)\n", (672447, 672523), False, 'import pulumi\n'), ((674206, 674240), 'pulumi.set', 'pulumi.set', (['__self__', '"""days"""', 'days'], {}), "(__self__, 'days', days)\n", (674216, 674240), False, 'import pulumi\n'), ((675213, 675247), 'pulumi.set', 'pulumi.set', (['__self__', '"""days"""', 'days'], {}), "(__self__, 'days', days)\n", (675223, 675247), False, 'import pulumi\n'), ((676839, 676873), 'pulumi.set', 'pulumi.set', (['__self__', '"""date"""', 'date'], {}), "(__self__, 'date', date)\n", (676849, 676873), False, 'import pulumi\n'), ((676915, 676949), 'pulumi.set', 'pulumi.set', (['__self__', '"""days"""', 'days'], {}), "(__self__, 'days', days)\n", (676925, 676949), False, 'import pulumi\n'), ((678627, 678679), 'pulumi.set', 'pulumi.set', (['__self__', '"""target_prefix"""', 'target_prefix'], {}), "(__self__, 'target_prefix', target_prefix)\n", (678637, 678679), False, 'import pulumi\n'), ((683553, 683593), 'pulumi.set', 'pulumi.set', (['__self__', '"""enabled"""', 'enabled'], {}), "(__self__, 'enabled', enabled)\n", (683563, 683593), False, 'import pulumi\n'), ((685276, 685330), 'pulumi.set', 'pulumi.set', (['__self__', '"""error_document"""', 'error_document'], {}), "(__self__, 'error_document', error_document)\n", (685286, 685330), False, 'import pulumi\n'), ((685382, 685436), 'pulumi.set', 'pulumi.set', (['__self__', '"""index_document"""', 'index_document'], {}), "(__self__, 'index_document', index_document)\n", (685392, 685436), False, 'import pulumi\n'), ((685498, 685572), 'pulumi.set', 'pulumi.set', (['__self__', '"""redirect_all_requests_to"""', 'redirect_all_requests_to'], {}), "(__self__, 'redirect_all_requests_to', redirect_all_requests_to)\n", (685508, 685572), False, 'import pulumi\n'), ((685623, 685675), 'pulumi.set', 'pulumi.set', (['__self__', '"""routing_rules"""', 'routing_rules'], {}), "(__self__, 'routing_rules', routing_rules)\n", (685633, 685675), False, 'import pulumi\n'), ((688471, 688511), 'pulumi.set', 'pulumi.set', (['__self__', '"""address"""', 'address'], {}), "(__self__, 'address', address)\n", (688481, 688511), False, 'import pulumi\n'), ((688573, 688647), 'pulumi.set', 'pulumi.set', (['__self__', '"""ddos_protection_provider"""', 'ddos_protection_provider'], {}), "(__self__, 'ddos_protection_provider', ddos_protection_provider)\n", (688583, 688647), False, 'import pulumi\n'), ((688709, 688783), 'pulumi.set', 'pulumi.set', (['__self__', '"""outgoing_smtp_capability"""', 'outgoing_smtp_capability'], {}), "(__self__, 'outgoing_smtp_capability', outgoing_smtp_capability)\n", (688719, 688783), False, 'import pulumi\n'), ((688828, 688868), 'pulumi.set', 'pulumi.set', (['__self__', '"""zone_id"""', 'zone_id'], {}), "(__self__, 'zone_id', zone_id)\n", (688838, 688868), False, 'import pulumi\n'), ((691635, 691683), 'pulumi.set', 'pulumi.set', (['__self__', '"""description"""', 'description'], {}), "(__self__, 'description', description)\n", (691645, 691683), False, 'import pulumi\n'), ((691730, 691774), 'pulumi.set', 'pulumi.set', (['__self__', '"""from_port"""', 'from_port'], {}), "(__self__, 'from_port', from_port)\n", (691740, 691774), False, 'import pulumi\n'), ((691814, 691844), 'pulumi.set', 'pulumi.set', (['__self__', '"""id"""', 'id'], {}), "(__self__, 'id', id)\n", (691824, 691844), False, 'import pulumi\n'), ((691888, 691926), 'pulumi.set', 'pulumi.set', (['__self__', '"""labels"""', 'labels'], {}), "(__self__, 'labels', labels)\n", (691898, 691926), False, 'import pulumi\n'), ((691968, 692002), 'pulumi.set', 'pulumi.set', (['__self__', '"""port"""', 'port'], {}), "(__self__, 'port', port)\n", (691978, 692002), False, 'import pulumi\n'), ((692057, 692117), 'pulumi.set', 'pulumi.set', (['__self__', '"""predefined_target"""', 'predefined_target'], {}), "(__self__, 'predefined_target', predefined_target)\n", (692067, 692117), False, 'import pulumi\n'), ((692172, 692232), 'pulumi.set', 'pulumi.set', (['__self__', '"""security_group_id"""', 'security_group_id'], {}), "(__self__, 'security_group_id', security_group_id)\n", (692182, 692232), False, 'import pulumi\n'), ((692277, 692317), 'pulumi.set', 'pulumi.set', (['__self__', '"""to_port"""', 'to_port'], {}), "(__self__, 'to_port', to_port)\n", (692287, 692317), False, 'import pulumi\n'), ((692369, 692423), 'pulumi.set', 'pulumi.set', (['__self__', '"""v4_cidr_blocks"""', 'v4_cidr_blocks'], {}), "(__self__, 'v4_cidr_blocks', v4_cidr_blocks)\n", (692379, 692423), False, 'import pulumi\n'), ((692475, 692529), 'pulumi.set', 'pulumi.set', (['__self__', '"""v6_cidr_blocks"""', 'v6_cidr_blocks'], {}), "(__self__, 'v6_cidr_blocks', v6_cidr_blocks)\n", (692485, 692529), False, 'import pulumi\n'), ((697181, 697229), 'pulumi.set', 'pulumi.set', (['__self__', '"""description"""', 'description'], {}), "(__self__, 'description', description)\n", (697191, 697229), False, 'import pulumi\n'), ((697276, 697320), 'pulumi.set', 'pulumi.set', (['__self__', '"""from_port"""', 'from_port'], {}), "(__self__, 'from_port', from_port)\n", (697286, 697320), False, 'import pulumi\n'), ((697360, 697390), 'pulumi.set', 'pulumi.set', (['__self__', '"""id"""', 'id'], {}), "(__self__, 'id', id)\n", (697370, 697390), False, 'import pulumi\n'), ((697434, 697472), 'pulumi.set', 'pulumi.set', (['__self__', '"""labels"""', 'labels'], {}), "(__self__, 'labels', labels)\n", (697444, 697472), False, 'import pulumi\n'), ((697514, 697548), 'pulumi.set', 'pulumi.set', (['__self__', '"""port"""', 'port'], {}), "(__self__, 'port', port)\n", (697524, 697548), False, 'import pulumi\n'), ((697603, 697663), 'pulumi.set', 'pulumi.set', (['__self__', '"""predefined_target"""', 'predefined_target'], {}), "(__self__, 'predefined_target', predefined_target)\n", (697613, 697663), False, 'import pulumi\n'), ((697718, 697778), 'pulumi.set', 'pulumi.set', (['__self__', '"""security_group_id"""', 'security_group_id'], {}), "(__self__, 'security_group_id', security_group_id)\n", (697728, 697778), False, 'import pulumi\n'), ((697823, 697863), 'pulumi.set', 'pulumi.set', (['__self__', '"""to_port"""', 'to_port'], {}), "(__self__, 'to_port', to_port)\n", (697833, 697863), False, 'import pulumi\n'), ((697915, 697969), 'pulumi.set', 'pulumi.set', (['__self__', '"""v4_cidr_blocks"""', 'v4_cidr_blocks'], {}), "(__self__, 'v4_cidr_blocks', v4_cidr_blocks)\n", (697925, 697969), False, 'import pulumi\n'), ((698021, 698075), 'pulumi.set', 'pulumi.set', (['__self__', '"""v6_cidr_blocks"""', 'v6_cidr_blocks'], {}), "(__self__, 'v6_cidr_blocks', v6_cidr_blocks)\n", (698031, 698075), False, 'import pulumi\n'), ((701954, 702016), 'pulumi.set', 'pulumi.set', (['__self__', '"""destination_prefix"""', 'destination_prefix'], {}), "(__self__, 'destination_prefix', destination_prefix)\n", (701964, 702016), False, 'import pulumi\n'), ((702070, 702128), 'pulumi.set', 'pulumi.set', (['__self__', '"""next_hop_address"""', 'next_hop_address'], {}), "(__self__, 'next_hop_address', next_hop_address)\n", (702080, 702128), False, 'import pulumi\n'), ((704132, 704180), 'pulumi.set', 'pulumi.set', (['__self__', '"""description"""', 'description'], {}), "(__self__, 'description', description)\n", (704142, 704180), False, 'import pulumi\n'), ((704227, 704271), 'pulumi.set', 'pulumi.set', (['__self__', '"""from_port"""', 'from_port'], {}), "(__self__, 'from_port', from_port)\n", (704237, 704271), False, 'import pulumi\n'), ((704311, 704341), 'pulumi.set', 'pulumi.set', (['__self__', '"""id"""', 'id'], {}), "(__self__, 'id', id)\n", (704321, 704341), False, 'import pulumi\n'), ((704385, 704423), 'pulumi.set', 'pulumi.set', (['__self__', '"""labels"""', 'labels'], {}), "(__self__, 'labels', labels)\n", (704395, 704423), False, 'import pulumi\n'), ((704465, 704499), 'pulumi.set', 'pulumi.set', (['__self__', '"""port"""', 'port'], {}), "(__self__, 'port', port)\n", (704475, 704499), False, 'import pulumi\n'), ((704554, 704614), 'pulumi.set', 'pulumi.set', (['__self__', '"""predefined_target"""', 'predefined_target'], {}), "(__self__, 'predefined_target', predefined_target)\n", (704564, 704614), False, 'import pulumi\n'), ((704669, 704729), 'pulumi.set', 'pulumi.set', (['__self__', '"""security_group_id"""', 'security_group_id'], {}), "(__self__, 'security_group_id', security_group_id)\n", (704679, 704729), False, 'import pulumi\n'), ((704774, 704814), 'pulumi.set', 'pulumi.set', (['__self__', '"""to_port"""', 'to_port'], {}), "(__self__, 'to_port', to_port)\n", (704784, 704814), False, 'import pulumi\n'), ((704866, 704920), 'pulumi.set', 'pulumi.set', (['__self__', '"""v4_cidr_blocks"""', 'v4_cidr_blocks'], {}), "(__self__, 'v4_cidr_blocks', v4_cidr_blocks)\n", (704876, 704920), False, 'import pulumi\n'), ((704972, 705026), 'pulumi.set', 'pulumi.set', (['__self__', '"""v6_cidr_blocks"""', 'v6_cidr_blocks'], {}), "(__self__, 'v6_cidr_blocks', v6_cidr_blocks)\n", (704982, 705026), False, 'import pulumi\n'), ((709651, 709699), 'pulumi.set', 'pulumi.set', (['__self__', '"""description"""', 'description'], {}), "(__self__, 'description', description)\n", (709661, 709699), False, 'import pulumi\n'), ((709746, 709790), 'pulumi.set', 'pulumi.set', (['__self__', '"""from_port"""', 'from_port'], {}), "(__self__, 'from_port', from_port)\n", (709756, 709790), False, 'import pulumi\n'), ((709830, 709860), 'pulumi.set', 'pulumi.set', (['__self__', '"""id"""', 'id'], {}), "(__self__, 'id', id)\n", (709840, 709860), False, 'import pulumi\n'), ((709904, 709942), 'pulumi.set', 'pulumi.set', (['__self__', '"""labels"""', 'labels'], {}), "(__self__, 'labels', labels)\n", (709914, 709942), False, 'import pulumi\n'), ((709984, 710018), 'pulumi.set', 'pulumi.set', (['__self__', '"""port"""', 'port'], {}), "(__self__, 'port', port)\n", (709994, 710018), False, 'import pulumi\n'), ((710073, 710133), 'pulumi.set', 'pulumi.set', (['__self__', '"""predefined_target"""', 'predefined_target'], {}), "(__self__, 'predefined_target', predefined_target)\n", (710083, 710133), False, 'import pulumi\n'), ((710188, 710248), 'pulumi.set', 'pulumi.set', (['__self__', '"""security_group_id"""', 'security_group_id'], {}), "(__self__, 'security_group_id', security_group_id)\n", (710198, 710248), False, 'import pulumi\n'), ((710293, 710333), 'pulumi.set', 'pulumi.set', (['__self__', '"""to_port"""', 'to_port'], {}), "(__self__, 'to_port', to_port)\n", (710303, 710333), False, 'import pulumi\n'), ((710385, 710439), 'pulumi.set', 'pulumi.set', (['__self__', '"""v4_cidr_blocks"""', 'v4_cidr_blocks'], {}), "(__self__, 'v4_cidr_blocks', v4_cidr_blocks)\n", (710395, 710439), False, 'import pulumi\n'), ((710491, 710545), 'pulumi.set', 'pulumi.set', (['__self__', '"""v6_cidr_blocks"""', 'v6_cidr_blocks'], {}), "(__self__, 'v6_cidr_blocks', v6_cidr_blocks)\n", (710501, 710545), False, 'import pulumi\n'), ((714618, 714666), 'pulumi.set', 'pulumi.set', (['__self__', '"""domain_name"""', 'domain_name'], {}), "(__self__, 'domain_name', domain_name)\n", (714628, 714666), False, 'import pulumi\n'), ((714723, 714787), 'pulumi.set', 'pulumi.set', (['__self__', '"""domain_name_servers"""', 'domain_name_servers'], {}), "(__self__, 'domain_name_servers', domain_name_servers)\n", (714733, 714787), False, 'import pulumi\n'), ((714836, 714884), 'pulumi.set', 'pulumi.set', (['__self__', '"""ntp_servers"""', 'ntp_servers'], {}), "(__self__, 'ntp_servers', ntp_servers)\n", (714846, 714884), False, 'import pulumi\n'), ((716499, 716537), 'pulumi.set', 'pulumi.set', (['__self__', '"""region"""', 'region'], {}), "(__self__, 'region', region)\n", (716509, 716537), False, 'import pulumi\n'), ((760595, 760643), 'pulumi.set', 'pulumi.set', (['__self__', '"""preemptible"""', 'preemptible'], {}), "(__self__, 'preemptible', preemptible)\n", (760605, 760643), False, 'import pulumi\n'), ((761336, 761402), 'pulumi.set', 'pulumi.set', (['__self__', '"""zone_instances_limit"""', 'zone_instances_limit'], {}), "(__self__, 'zone_instances_limit', zone_instances_limit)\n", (761346, 761402), False, 'import pulumi\n'), ((761459, 761523), 'pulumi.set', 'pulumi.set', (['__self__', '"""zone_requests_limit"""', 'zone_requests_limit'], {}), "(__self__, 'zone_requests_limit', zone_requests_limit)\n", (761469, 761523), False, 'import pulumi\n'), ((766431, 766489), 'pulumi.set', 'pulumi.set', (['__self__', '"""assign_public_ip"""', 'assign_public_ip'], {}), "(__self__, 'assign_public_ip', assign_public_ip)\n", (766441, 766489), False, 'import pulumi\n'), ((766540, 766592), 'pulumi.set', 'pulumi.set', (['__self__', '"""brokers_count"""', 'brokers_count'], {}), "(__self__, 'brokers_count', brokers_count)\n", (766550, 766592), False, 'import pulumi\n'), ((766645, 766701), 'pulumi.set', 'pulumi.set', (['__self__', '"""schema_registry"""', 'schema_registry'], {}), "(__self__, 'schema_registry', schema_registry)\n", (766655, 766701), False, 'import pulumi\n'), ((766755, 766813), 'pulumi.set', 'pulumi.set', (['__self__', '"""unmanaged_topics"""', 'unmanaged_topics'], {}), "(__self__, 'unmanaged_topics', unmanaged_topics)\n", (766765, 766813), False, 'import pulumi\n'), ((770499, 770549), 'pulumi.set', 'pulumi.set', (['__self__', '"""kafka_config"""', 'kafka_config'], {}), "(__self__, 'kafka_config', kafka_config)\n", (770509, 770549), False, 'import pulumi\n'), ((772565, 772641), 'pulumi.set', 'pulumi.set', (['__self__', '"""auto_create_topics_enable"""', 'auto_create_topics_enable'], {}), "(__self__, 'auto_create_topics_enable', auto_create_topics_enable)\n", (772575, 772641), False, 'import pulumi\n'), ((772695, 772753), 'pulumi.set', 'pulumi.set', (['__self__', '"""compression_type"""', 'compression_type'], {}), "(__self__, 'compression_type', compression_type)\n", (772705, 772753), False, 'import pulumi\n'), ((772817, 772895), 'pulumi.set', 'pulumi.set', (['__self__', '"""default_replication_factor"""', 'default_replication_factor'], {}), "(__self__, 'default_replication_factor', default_replication_factor)\n", (772827, 772895), False, 'import pulumi\n'), ((772960, 773045), 'pulumi.set', 'pulumi.set', (['__self__', '"""log_flush_interval_messages"""', 'log_flush_interval_messages'], {}), "(__self__, 'log_flush_interval_messages', log_flush_interval_messages\n )\n", (772970, 773045), False, 'import pulumi\n'), ((773099, 773167), 'pulumi.set', 'pulumi.set', (['__self__', '"""log_flush_interval_ms"""', 'log_flush_interval_ms'], {}), "(__self__, 'log_flush_interval_ms', log_flush_interval_ms)\n", (773109, 773167), False, 'import pulumi\n'), ((773236, 773328), 'pulumi.set', 'pulumi.set', (['__self__', '"""log_flush_scheduler_interval_ms"""', 'log_flush_scheduler_interval_ms'], {}), "(__self__, 'log_flush_scheduler_interval_ms',\n log_flush_scheduler_interval_ms)\n", (773246, 773328), False, 'import pulumi\n'), ((773377, 773433), 'pulumi.set', 'pulumi.set', (['__self__', '"""log_preallocate"""', 'log_preallocate'], {}), "(__self__, 'log_preallocate', log_preallocate)\n", (773387, 773433), False, 'import pulumi\n'), ((773490, 773554), 'pulumi.set', 'pulumi.set', (['__self__', '"""log_retention_bytes"""', 'log_retention_bytes'], {}), "(__self__, 'log_retention_bytes', log_retention_bytes)\n", (773500, 773554), False, 'import pulumi\n'), ((773611, 773675), 'pulumi.set', 'pulumi.set', (['__self__', '"""log_retention_hours"""', 'log_retention_hours'], {}), "(__self__, 'log_retention_hours', log_retention_hours)\n", (773621, 773675), False, 'import pulumi\n'), ((773734, 773802), 'pulumi.set', 'pulumi.set', (['__self__', '"""log_retention_minutes"""', 'log_retention_minutes'], {}), "(__self__, 'log_retention_minutes', log_retention_minutes)\n", (773744, 773802), False, 'import pulumi\n'), ((773856, 773914), 'pulumi.set', 'pulumi.set', (['__self__', '"""log_retention_ms"""', 'log_retention_ms'], {}), "(__self__, 'log_retention_ms', log_retention_ms)\n", (773866, 773914), False, 'import pulumi\n'), ((773969, 774029), 'pulumi.set', 'pulumi.set', (['__self__', '"""log_segment_bytes"""', 'log_segment_bytes'], {}), "(__self__, 'log_segment_bytes', log_segment_bytes)\n", (773979, 774029), False, 'import pulumi\n'), ((774081, 774135), 'pulumi.set', 'pulumi.set', (['__self__', '"""num_partitions"""', 'num_partitions'], {}), "(__self__, 'num_partitions', num_partitions)\n", (774091, 774135), False, 'import pulumi\n'), ((774200, 774285), 'pulumi.set', 'pulumi.set', (['__self__', '"""socket_receive_buffer_bytes"""', 'socket_receive_buffer_bytes'], {}), "(__self__, 'socket_receive_buffer_bytes', socket_receive_buffer_bytes\n )\n", (774210, 774285), False, 'import pulumi\n'), ((774342, 774416), 'pulumi.set', 'pulumi.set', (['__self__', '"""socket_send_buffer_bytes"""', 'socket_send_buffer_bytes'], {}), "(__self__, 'socket_send_buffer_bytes', socket_send_buffer_bytes)\n", (774352, 774416), False, 'import pulumi\n'), ((784908, 784958), 'pulumi.set', 'pulumi.set', (['__self__', '"""topic_config"""', 'topic_config'], {}), "(__self__, 'topic_config', topic_config)\n", (784918, 784958), False, 'import pulumi\n'), ((787550, 787604), 'pulumi.set', 'pulumi.set', (['__self__', '"""cleanup_policy"""', 'cleanup_policy'], {}), "(__self__, 'cleanup_policy', cleanup_policy)\n", (787560, 787604), False, 'import pulumi\n'), ((787658, 787716), 'pulumi.set', 'pulumi.set', (['__self__', '"""compression_type"""', 'compression_type'], {}), "(__self__, 'compression_type', compression_type)\n", (787668, 787716), False, 'import pulumi\n'), ((787773, 787837), 'pulumi.set', 'pulumi.set', (['__self__', '"""delete_retention_ms"""', 'delete_retention_ms'], {}), "(__self__, 'delete_retention_ms', delete_retention_ms)\n", (787783, 787837), False, 'import pulumi\n'), ((787895, 787961), 'pulumi.set', 'pulumi.set', (['__self__', '"""file_delete_delay_ms"""', 'file_delete_delay_ms'], {}), "(__self__, 'file_delete_delay_ms', file_delete_delay_ms)\n", (787905, 787961), False, 'import pulumi\n'), ((788013, 788067), 'pulumi.set', 'pulumi.set', (['__self__', '"""flush_messages"""', 'flush_messages'], {}), "(__self__, 'flush_messages', flush_messages)\n", (788023, 788067), False, 'import pulumi\n'), ((788113, 788155), 'pulumi.set', 'pulumi.set', (['__self__', '"""flush_ms"""', 'flush_ms'], {}), "(__self__, 'flush_ms', flush_ms)\n", (788123, 788155), False, 'import pulumi\n'), ((788210, 788270), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_message_bytes"""', 'max_message_bytes'], {}), "(__self__, 'max_message_bytes', max_message_bytes)\n", (788220, 788270), False, 'import pulumi\n'), ((788329, 788397), 'pulumi.set', 'pulumi.set', (['__self__', '"""min_compaction_lag_ms"""', 'min_compaction_lag_ms'], {}), "(__self__, 'min_compaction_lag_ms', min_compaction_lag_ms)\n", (788339, 788397), False, 'import pulumi\n'), ((788454, 788518), 'pulumi.set', 'pulumi.set', (['__self__', '"""min_insync_replicas"""', 'min_insync_replicas'], {}), "(__self__, 'min_insync_replicas', min_insync_replicas)\n", (788464, 788518), False, 'import pulumi\n'), ((788567, 788615), 'pulumi.set', 'pulumi.set', (['__self__', '"""preallocate"""', 'preallocate'], {}), "(__self__, 'preallocate', preallocate)\n", (788577, 788615), False, 'import pulumi\n'), ((788668, 788724), 'pulumi.set', 'pulumi.set', (['__self__', '"""retention_bytes"""', 'retention_bytes'], {}), "(__self__, 'retention_bytes', retention_bytes)\n", (788678, 788724), False, 'import pulumi\n'), ((788774, 788824), 'pulumi.set', 'pulumi.set', (['__self__', '"""retention_ms"""', 'retention_ms'], {}), "(__self__, 'retention_ms', retention_ms)\n", (788784, 788824), False, 'import pulumi\n'), ((788875, 788927), 'pulumi.set', 'pulumi.set', (['__self__', '"""segment_bytes"""', 'segment_bytes'], {}), "(__self__, 'segment_bytes', segment_bytes)\n", (788885, 788927), False, 'import pulumi\n'), ((793379, 793427), 'pulumi.set', 'pulumi.set', (['__self__', '"""permissions"""', 'permissions'], {}), "(__self__, 'permissions', permissions)\n", (793389, 793427), False, 'import pulumi\n')] |
"""
Project resources
Many configuration and scripting resources are extracted here.
"""
from shared.tools.snapshot.utils import encode, hashmapToDict
def extract_project_props(client_context):
global_props = client_context.getGlobalProps()
configuration = {
'permissions': hashmapToDict(global_props.getPermissionEnabledMap()),
'roles': {
'client': dict((category, [role.strip()
for role
in role_string.split(',')
if role
])
for category, role_string
in hashmapToDict(
global_props.getRequiredClientRolesMap()
).items()),
'delete' : [role.strip() for role in global_props.getRequiredDeleteRoles()],
'publish' : [role.strip() for role in global_props.getRequiredPublishRoles()],
'resource': [role.strip() for role in global_props.getRequiredResourceRoles()],
'required': [role.strip() for role in global_props.getRequiredRoles()],
'save' : [role.strip() for role in global_props.getRequiredSaveRoles()],
'view' : [role.strip() for role in global_props.getRequiredViewRoles()],
},
'auditing': global_props.isAuditingEnabled(),
'legacy': global_props.isLegacyProject(),
'commitMessageMode': global_props.getCommitMessageMode().toString(), # enum
'defaultSQLTagsProviderRate': global_props.getSqltagsClientPollRate(),
}
defaultable_attributes = set([
'auditProfileName',
'authProfileName',
'defaultDatasourceName',
'defaultSQLTagsProviderName',
'publishMode',
])
for attribute in defaultable_attributes:
try: # to get the Java getter first
# it's slightly more reliable than the Jython auto-attribute, in general
getter_name = 'get' + attribute[0].upper() + attribute[1:]
value = getattr(global_props, getter_name)()
except AttributeError:
try: # the Jython attribute
value = getattr(global_props, attribute)
except AttributeError:
value = None
if value is None:
continue
configuration[attribute] = value
return dict([
encode(configuration),
])
def extract_gatewayevents(resource_objects):
assert len(resource_objects) == 1, 'Resource is expected to be contained in one root object'
client_script_config = resource_objects[0]
scripts = {}
script = client_script_config.getStartupScript()
if script:
scripts['startup.py'] = script
script = client_script_config.getShutdownScript()
if script:
scripts['shutdown.py'] = script
timer_scripts = client_script_config.getTimerScripts()
for timer_script in timer_scripts:
suffix, serialized = encode({
'enabled': timer_script.isEnabled(),
'timing': 'delay' if timer_script.isFixedDelay() else 'rate',
'period': timer_script.getDelay(),
'threading': 'shared' if timer_script.isSharedThread() else 'dedicated',
})
scripts['timer/%s%s' % (timer_script.getName(), suffix)] = serialized
scripts['timer/%s.py' % timer_script.getName()] = timer_scripts[timer_script]
for tag_script in client_script_config.getTagChangeScripts():
suffix, serialized = encode({
'name': tag_script.getName(),
'tags': [tag_path for tag_path in tag_script.getPaths()],
'triggers': [t.toString() for t in tag_script.getChangeTypes()],
'enabled': tag_script.isEnabled(),
})
scripts['tag-change/%s%s' % (tag_script.getName(), suffix)] = serialized
scripts['tag-change/%s.py' % tag_script.getName()] = tag_script.getScript()
message_scripts = client_script_config.getMessageHandlerScripts()
for message_script in message_scripts:
suffix, serialized = encode({
'name': message_script.getName(),
'threading': str(message_script.getThreadType()),
'enabled': message_script.isEnabled(),
})
scripts['message/%s%s' % (message_script.getName(),suffix)] = serialized
scripts['message/%s.py' % message_script.getName()] = message_scripts[message_script]
return scripts
def extract_clientevents(resource_objects):
assert len(resource_objects) == 1, 'Resource is expected to be contained in one root object'
client_script_config = resource_objects[0]
scripts = {}
script = client_script_config.getStartupScript()
if script:
scripts['startup.py'] = script
script = client_script_config.getShutdownScript()
if script:
scripts['shutdown.py'] = script
script = client_script_config.getShutdownAllowedScript()
if script:
scripts['shutdown-intercept.py'] = script
key_schema_pattern = re.compile("(\[(?P<modifiers>.*)\] )?(?P<key>.*) \((?P<action>.*)\)")
key_modifier_pattern = re.compile("(Button \d|\w+)")
key_scripts = client_script_config.getKeyScripts()
for kix, key_script in enumerate(key_scripts):
key_config = key_schema_pattern.match(key_script.getDisplay()).groupdict()
suffix, serialized = encode({
'action': key_config['action'],
'key': key_config['key'].replace("'", ''),
'modifiers': key_modifier_pattern.findall(key_config['modifiers']) if key_config['modifiers'] else []
})
scripts['key/%s%s' % (key_script.getDisplay(), suffix)] = serialized
scripts['key/%s.py' % key_script.getDisplay()] = key_scripts[key_script]
timer_scripts = client_script_config.getTimerScripts()
for timer_script in timer_scripts:
suffix, serialized = encode({
'enabled': timer_script.isEnabled(),
'timing': 'delay' if timer_script.isFixedDelay() else 'rate',
'period': timer_script.getDelay(),
'threading': 'shared' if timer_script.isSharedThread() else 'dedicated',
})
scripts['timer/%s%s' % (timer_script.getName(), suffix)] = serialized
scripts['timer/%s.py' % timer_script.getName()] = timer_scripts[timer_script]
for tag_script in client_script_config.getTagChangeScripts():
suffix, serialized = encode({
'name': tag_script.getName(),
'tags': [tag_path for tag_path in tag_script.getPaths()],
'triggers': [t.toString() for t in tag_script.getChangeTypes()],
'enabled': tag_script.isEnabled(),
})
scripts['tag-change/%s%s' % (tag_script.getName(), suffix)] = serialized
scripts['tag-change/%s.py' % tag_script.getName()] = tag_script.getScript()
def traverse_menu(parent_path, menu_node, mutable_dict):
for mix, child in enumerate(menu_node.getChildren() or []):
suffix, serialized = encode({
'name': child.getName(),
'icon': child.getIconPath(),
'mnemonic': child.getMnemonic(),
'description': child.getDescription(),
'accelerator': child.getAccelerator(),
})
mutable_dict['%s/entry-%02d%s' % ('/'.join(parent_path), mix, suffix)] = serialized
mutable_dict['%s/entry-%02d.py' % ('/'.join(parent_path), mix)] = child.getScript()
traverse_menu(parent_path + [child.getName() or ('Submenu-%02d' % mix)], child, mutable_dict)
menu_root = client_script_config.getMenuRoot()
traverse_menu(['menu'], menu_root, scripts)
message_scripts = client_script_config.getMessageHandlerScripts()
for message_script in message_scripts:
suffix, serialized = encode({
'name': message_script.getName(),
'threading': str(message_script.getThreadType()),
'enabled': message_script.isEnabled(),
})
scripts['message/%s%s' % (message_script.getName(), suffix)] = serialized
scripts['message/%s.py' % message_script.getName()] = message_scripts[message_script]
return scripts
def extract_namedquery(resource_objects):
assert len(resource_objects) == 1, 'Resource is expected to be contained in one root object'
named_query = resource_objects[0]
info = {
'query': named_query.getQuery(),
'database': named_query.getDatabase() or '-default-',
'parameters': dict(
(param.getIdentifier(), {
'sql_type' : str(param.getSqlType()),
'type' : str(param.getType()),
'identifier': str(param.getIdentifier()),
})
for param
in named_query.getParameters()
),
'type': named_query.getType(),
}
return dict([
('.sql', format_sql(info['query'])),
encode(info),
])
def extract_project_script(resource_objects):
assert len(resource_objects) == 1, 'Resource is expected to be contained in one root object'
script = resource_objects[0]
return {
'.py': script,
}
# Ready for the dispatcher
EXTRACTORS = {
'sr.script.project': extract_project_script,
'named-query': extract_namedquery,
'client.event.scripts': extract_clientevents,
'event.scripts': extract_gatewayevents,
'project/properties': extract_project_props,
}
| [
"shared.tools.snapshot.utils.encode"
] | [((2021, 2042), 'shared.tools.snapshot.utils.encode', 'encode', (['configuration'], {}), '(configuration)\n', (2027, 2042), False, 'from shared.tools.snapshot.utils import encode, hashmapToDict\n'), ((7840, 7852), 'shared.tools.snapshot.utils.encode', 'encode', (['info'], {}), '(info)\n', (7846, 7852), False, 'from shared.tools.snapshot.utils import encode, hashmapToDict\n')] |
# -*- coding: utf-8 -*-
"""Polynomial techniques for fitting baselines to experimental data.
Created on Feb. 27, 2021
@author: <NAME>
The function penalized_poly was adapted from MATLAB code from
https://www.mathworks.com/matlabcentral/fileexchange/27429-background-correction
(accessed March 18, 2021), which was licensed under the BSD-2-clause below.
License: 2-clause BSD
Copyright (c) 2012, <NAME>
All rights reserved.
Redistribution and use in source and binary forms, with or without
modification, are permitted provided that the following conditions are
met:
* Redistributions of source code must retain the above copyright
notice, this list of conditions and the following disclaimer.
* Redistributions in binary form must reproduce the above copyright
notice, this list of conditions and the following disclaimer in
the documentation and/or other materials provided with the distribution
THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE
LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
POSSIBILITY OF SUCH DAMAGE.
The function loess was adapted from code from https://gist.github.com/agramfort/850437
(accessed March 25, 2021), which was licensed under the BSD-3-clause below.
# Authors: <NAME> <<EMAIL>>
#
# License: BSD (3-clause)
Copyright (c) 2015, <NAME>
All rights reserved.
Redistribution and use in source and binary forms, with or without
modification, are permitted provided that the following conditions are met:
1. Redistributions of source code must retain the above copyright notice, this
list of conditions and the following disclaimer.
2. Redistributions in binary form must reproduce the above copyright notice,
this list of conditions and the following disclaimer in the documentation
and/or other materials provided with the distribution.
3. Neither the name of the copyright holder nor the names of its
contributors may be used to endorse or promote products derived from
this software without specific prior written permission.
THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE
FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR
SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER
CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY,
OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
"""
from math import ceil
import warnings
import numpy as np
from . import _weighting
from ._algorithm_setup import _get_vander, _setup_polynomial
from ._compat import jit, prange
from .utils import (
_MIN_FLOAT, ParameterWarning, _convert_coef, _interp_inplace, _inverted_sort,
relative_difference
)
def poly(data, x_data=None, poly_order=2, weights=None, return_coef=False):
"""
Computes a polynomial that fits the baseline of the data.
Parameters
----------
data : array-like, shape (N,)
The y-values of the measured data, with N data points.
x_data : array-like, shape (N,), optional
The x-values of the measured data. Default is None, which will create an
array from -1 to 1 with N points.
poly_order : int, optional
The polynomial order for fitting the baseline. Default is 2.
weights : array-like, shape (N,), optional
The weighting array. If None (default), then will be an array with
size equal to N and all values set to 1.
return_coef : bool, optional
If True, will convert the polynomial coefficients for the fit baseline to
a form that fits the input x_data and return them in the params dictionary.
Default is False, since the conversion takes time.
Returns
-------
baseline : numpy.ndarray, shape (N,)
The calculated baseline.
params : dict
A dictionary with the following items:
* 'weights': numpy.ndarray, shape (N,)
The weight array used for fitting the data.
* 'coef': numpy.ndarray, shape (poly_order,)
Only if `return_coef` is True. The array of polynomial parameters
for the baseline, in increasing order. Can be used to create a
polynomial using numpy.polynomial.polynomial.Polynomial().
Notes
-----
To only fit regions without peaks, supply a weight array with zero values
at the indices where peaks are located.
"""
y, x, weight_array, original_domain = _setup_polynomial(data, x_data, weights)
fit_polynomial = np.polynomial.Polynomial.fit(x, y, poly_order, w=np.sqrt(weight_array))
baseline = fit_polynomial(x)
params = {'weights': weight_array}
if return_coef:
params['coef'] = fit_polynomial.convert(window=original_domain).coef
return baseline, params
def modpoly(data, x_data=None, poly_order=2, tol=1e-3, max_iter=250, weights=None,
use_original=False, mask_initial_peaks=False, return_coef=False):
"""
The modified polynomial (ModPoly) baseline algorithm.
Parameters
----------
data : array-like, shape (N,)
The y-values of the measured data, with N data points.
x_data : array-like, shape (N,), optional
The x-values of the measured data. Default is None, which will create an
array from -1 to 1 with N points.
poly_order : int, optional
The polynomial order for fitting the baseline. Default is 2.
tol : float, optional
The exit criteria. Default is 1e-3.
max_iter : int, optional
The maximum number of iterations. Default is 250.
weights : array-like, shape (N,), optional
The weighting array. If None (default), then will be an array with
size equal to N and all values set to 1.
use_original : bool, optional
If False (default), will compare the baseline of each iteration with
the y-values of that iteration [1]_ when choosing minimum values. If True,
will compare the baseline with the original y-values given by `data` [2]_.
mask_initial_peaks : bool, optional
If True, will mask any data where the initial baseline fit + the standard
deviation of the residual is less than measured data [3]_. Default is False.
return_coef : bool, optional
If True, will convert the polynomial coefficients for the fit baseline to
a form that fits the input x_data and return them in the params dictionary.
Default is False, since the conversion takes time.
Returns
-------
baseline : numpy.ndarray, shape (N,)
The calculated baseline.
params : dict
A dictionary with the following items:
* 'weights': numpy.ndarray, shape (N,)
The weight array used for fitting the data.
* 'tol_history': numpy.ndarray
An array containing the calculated tolerance values for
each iteration. The length of the array is the number of iterations
completed. If the last value in the array is greater than the input
`tol` value, then the function did not converge.
* 'coef': numpy.ndarray, shape (poly_order + 1,)
Only if `return_coef` is True. The array of polynomial parameters
for the baseline, in increasing order. Can be used to create a
polynomial using numpy.polynomial.polynomial.Polynomial().
Notes
-----
Algorithm originally developed in [2]_ and then slightly modified in [1]_.
References
----------
.. [1] <NAME>., et al. Baseline correction by improved iterative polynomial
fitting with automatic threshold. Chemometrics and Intelligent
Laboratory Systems, 2006, 82, 59-65.
.. [2] <NAME>., et al. Automated method for subtraction of fluorescence
from biological raman spectra. Applied Spectroscopy, 2003, 57(11),
1363-1367.
.. [3] <NAME>., et al. Automated Autofluorescence Background Subtraction
Algorithm for Biomedical Raman Spectroscopy, Applied Spectroscopy,
2007, 61(11), 1225-1232.
"""
y, x, weight_array, original_domain, vander, pseudo_inverse = _setup_polynomial(
data, x_data, weights, poly_order, True, True, True
)
sqrt_w = np.sqrt(weight_array)
if use_original:
y0 = y
coef = np.dot(pseudo_inverse, sqrt_w * y)
baseline = np.dot(vander, coef)
if mask_initial_peaks:
# use baseline + deviation since without deviation, half of y should be above baseline
weight_array[baseline + np.std(y - baseline) < y] = 0
sqrt_w = np.sqrt(weight_array)
vander, pseudo_inverse = _get_vander(x, poly_order, sqrt_w)
tol_history = np.empty(max_iter)
for i in range(max_iter):
baseline_old = baseline
y = np.minimum(y0 if use_original else y, baseline)
coef = np.dot(pseudo_inverse, sqrt_w * y)
baseline = np.dot(vander, coef)
calc_difference = relative_difference(baseline_old, baseline)
tol_history[i] = calc_difference
if calc_difference < tol:
break
params = {'weights': weight_array, 'tol_history': tol_history[:i + 1]}
if return_coef:
params['coef'] = _convert_coef(coef, original_domain)
return baseline, params
def imodpoly(data, x_data=None, poly_order=2, tol=1e-3, max_iter=250, weights=None,
use_original=False, mask_initial_peaks=True, return_coef=False, num_std=1):
"""
The improved modofied polynomial (IModPoly) baseline algorithm.
Parameters
----------
data : array-like, shape (N,)
The y-values of the measured data, with N data points.
x_data : array-like, shape (N,), optional
The x-values of the measured data. Default is None, which will create an
array from -1 to 1 with N points.
poly_order : int, optional
The polynomial order for fitting the baseline. Default is 2.
tol : float, optional
The exit criteria. Default is 1e-3.
max_iter : int, optional
The maximum number of iterations. Default is 250.
weights : array-like, shape (N,), optional
The weighting array. If None (default), then will be an array with
size equal to N and all values set to 1.
use_original : bool, optional
If False (default), will compare the baseline of each iteration with
the y-values of that iteration [4]_ when choosing minimum values. If True,
will compare the baseline with the original y-values given by `data` [5]_.
mask_initial_peaks : bool, optional
If True (default), will mask any data where the initial baseline fit +
the standard deviation of the residual is less than measured data [6]_.
return_coef : bool, optional
If True, will convert the polynomial coefficients for the fit baseline to
a form that fits the input x_data and return them in the params dictionary.
Default is False, since the conversion takes time.
num_std : float, optional
The number of standard deviations to include when thresholding. Default
is 1.
Returns
-------
baseline : numpy.ndarray, shape (N,)
The calculated baseline.
params : dict
A dictionary with the following items:
* 'weights': numpy.ndarray, shape (N,)
The weight array used for fitting the data.
* 'tol_history': numpy.ndarray
An array containing the calculated tolerance values for
each iteration. The length of the array is the number of iterations
completed. If the last value in the array is greater than the input
`tol` value, then the function did not converge.
* 'coef': numpy.ndarray, shape (poly_order + 1,)
Only if `return_coef` is True. The array of polynomial parameters
for the baseline, in increasing order. Can be used to create a
polynomial using numpy.polynomial.polynomial.Polynomial().
Notes
-----
Algorithm originally developed in [6]_.
References
----------
.. [4] <NAME>., et al. Baseline correction by improved iterative polynomial
fitting with automatic threshold. Chemometrics and Intelligent
Laboratory Systems, 2006, 82, 59-65.
.. [5] <NAME>., et al. Automated method for subtraction of fluorescence
from biological raman spectra. Applied Spectroscopy, 2003, 57(11),
1363-1367.
.. [6] <NAME>., et al. Automated Autofluorescence Background Subtraction
Algorithm for Biomedical Raman Spectroscopy, Applied Spectroscopy,
2007, 61(11), 1225-1232.
"""
y, x, weight_array, original_domain, vander, pseudo_inverse = _setup_polynomial(
data, x_data, weights, poly_order, True, True, True
)
sqrt_w = np.sqrt(weight_array)
if use_original:
y0 = y
coef = np.dot(pseudo_inverse, sqrt_w * y)
baseline = np.dot(vander, coef)
deviation = np.std(y - baseline)
if mask_initial_peaks:
weight_array[baseline + deviation < y] = 0
sqrt_w = np.sqrt(weight_array)
vander, pseudo_inverse = _get_vander(x, poly_order, sqrt_w)
tol_history = np.empty(max_iter)
for i in range(max_iter):
y = np.minimum(y0 if use_original else y, baseline + num_std * deviation)
coef = np.dot(pseudo_inverse, sqrt_w * y)
baseline = np.dot(vander, coef)
new_deviation = np.std(y - baseline)
# use new_deviation as dividing term in relative difference
calc_difference = relative_difference(new_deviation, deviation)
tol_history[i] = calc_difference
if calc_difference < tol:
break
deviation = new_deviation
params = {'weights': weight_array, 'tol_history': tol_history[:i + 1]}
if return_coef:
params['coef'] = _convert_coef(coef, original_domain)
return baseline, params
# adapted from (https://www.mathworks.com/matlabcentral/fileexchange/27429-background-correction);
# see license above
def _huber_loss(residual, threshold=1.0, alpha_factor=0.99, symmetric=True):
"""
The Huber non-quadratic cost function.
Parameters
----------
residual : numpy.ndarray, shape (N,)
The residual array.
threshold : float, optional
Any residual values below the threshold are given quadratic loss.
Default is 1.0.
alpha_factor : float, optional
The scale between 0 and 1 to multiply the cost function's alpha_max
value (see Notes below). Default is 0.99.
symmetric : bool, optional
If True (default), the cost function is symmetric and applies the same
weighting for positive and negative values. If False, will apply weights
asymmetrically so that only positive weights are given the non-quadratic
weigting and negative weights have normal, quadratic weighting.
Returns
-------
weights : numpy.ndarray, shape (N,)
The weight array.
Notes
-----
The returned result is
-residual + alpha_factor * alpha_max * phi'(residual)
where phi'(x) is the derivative of the huber loss function, phi(x).
References
----------
<NAME>., et al. Background removal from spectra by designing and
minimising a non-quadratic cost function. Chemometrics and Intelligent
Laboratory Systems, 2005, 76(2), 121–133.
"""
alpha = alpha_factor * 0.5 # alpha_max for huber is 0.5
if symmetric:
mask = (np.abs(residual) < threshold)
weights = (
mask * residual * (2 * alpha - 1)
+ (~mask) * 2 * alpha * threshold * np.sign(residual)
)
else:
mask = (residual < threshold)
weights = (
mask * residual * (2 * alpha - 1)
+ (~mask) * (2 * alpha * threshold - residual)
)
return weights
# adapted from (https://www.mathworks.com/matlabcentral/fileexchange/27429-background-correction);
# see license above
def _truncated_quadratic_loss(residual, threshold=1.0, alpha_factor=0.99, symmetric=True):
"""
The Truncated-Quadratic non-quadratic cost function.
Parameters
----------
residual : numpy.ndarray, shape (N,)
The residual array.
threshold : float, optional
Any residual values below the threshold are given quadratic loss.
Default is 1.0.
alpha_factor : float, optional
The scale between 0 and 1 to multiply the cost function's alpha_max
value (see Notes below). Default is 0.99.
symmetric : bool, optional
If True (default), the cost function is symmetric and applies the same
weighting for positive and negative values. If False, will apply weights
asymmetrically so that only positive weights are given the non-quadratic
weigting and negative weights have normal, quadratic weighting.
Returns
-------
weights : numpy.ndarray, shape (N,)
The weight array.
Notes
-----
The returned result is
-residual + alpha_factor * alpha_max * phi'(residual)
where phi'(x) is the derivative of the truncated quadratic function, phi(x).
References
----------
Mazet, V., et al. Background removal from spectra by designing and
minimising a non-quadratic cost function. Chemometrics and Intelligent
Laboratory Systems, 2005, 76(2), 121–133.
"""
alpha = alpha_factor * 0.5 # alpha_max for truncated quadratic is 0.5
if symmetric:
mask = (np.abs(residual) < threshold)
else:
mask = (residual < threshold)
return mask * residual * (2 * alpha - 1) - (~mask) * residual
def _indec_loss(residual, threshold=1.0, alpha_factor=0.99, symmetric=True):
"""
The Indec non-quadratic cost function.
Parameters
----------
residual : numpy.ndarray, shape (N,)
The residual array.
threshold : float, optional
Any residual values below the threshold are given quadratic loss.
Default is 1.0.
alpha_factor : float, optional
The scale between 0 and 1 to multiply the cost function's alpha_max
value (see Notes below). Default is 0.99.
symmetric : bool, optional
If True (default), the cost function is symmetric and applies the same
weighting for positive and negative values. If False, will apply weights
asymmetrically so that only positive weights are given the non-quadratic
weigting and negative weights have normal, quadratic weighting.
Returns
-------
weights : numpy.ndarray, shape (N,)
The weight array.
Notes
-----
The returned result is
-residual + alpha_factor * alpha_max * phi'(residual)
where phi'(x) is the derivative of the Indec function, phi(x).
References
----------
<NAME>., et al. Goldindec: A Novel Algorithm for Raman Spectrum Baseline
Correction. Applied Spectroscopy, 2015, 69(7), 834-842.
<NAME>., et al. Background removal from spectra by designing and
minimising a non-quadratic cost function. Chemometrics and Intelligent
Laboratory Systems, 2005, 76(2), 121–133.
"""
alpha = alpha_factor * 0.5 # alpha_max for indec is 0.5
if symmetric:
mask = (np.abs(residual) < threshold)
multiple = np.sign(residual)
else:
mask = (residual < threshold)
# multiple=1 is same as sign(residual) since residual is always > 0
# for asymmetric case, but this allows not doing the sign calculation
multiple = 1
weights = (
mask * residual * (2 * alpha - 1)
- (~mask) * (
residual + alpha * multiple * threshold**3 / np.maximum(2 * residual**2, _MIN_FLOAT)
)
)
return weights
def _identify_loss_method(loss_method):
"""
Identifies the symmetry for the given loss method.
Parameters
----------
loss_method : str
The loss method to use. Should have the symmetry identifier as
the prefix.
Returns
-------
symmetric : bool
True if `loss_method` had 's_' or 'symmetric_' as the prefix, else False.
str
The input `loss_method` value without the first section that indicated
the symmetry.
Raises
------
ValueError
Raised if the loss method does not have the correct form.
"""
prefix, *split_method = loss_method.lower().split('_')
if prefix not in ('a', 's', 'asymmetric', 'symmetric') or not split_method:
raise ValueError('must specify loss function symmetry by appending "a_" or "s_"')
if prefix in ('a', 'asymmetric'):
symmetric = False
else:
symmetric = True
return symmetric, '_'.join(split_method)
# adapted from (https://www.mathworks.com/matlabcentral/fileexchange/27429-background-correction);
# see license above
def penalized_poly(data, x_data=None, poly_order=2, tol=1e-3, max_iter=250,
weights=None, cost_function='asymmetric_truncated_quadratic',
threshold=None, alpha_factor=0.99, return_coef=False):
"""
Fits a polynomial baseline using a non-quadratic cost function.
The non-quadratic cost functions penalize residuals with larger values,
giving a more robust fit compared to normal least-squares.
Parameters
----------
data : array-like, shape (N,)
The y-values of the measured data, with N data points.
x_data : array-like, shape (N,), optional
The x-values of the measured data. Default is None, which will create an
array from -1 to 1 with N points.
poly_order : int, optional
The polynomial order for fitting the baseline. Default is 2.
tol : float, optional
The exit criteria. Default is 1e-3.
max_iter : int, optional
The maximum number of iterations. Default is 250.
weights : array-like, shape (N,), optional
The weighting array. If None (default), then will be an array with
size equal to N and all values set to 1.
cost_function : str, optional
The non-quadratic cost function to minimize. Must indicate symmetry of the
method by appending 'a' or 'asymmetric' for asymmetric loss, and 's' or
'symmetric' for symmetric loss. Default is 'asymmetric_truncated_quadratic'.
Available methods, and their associated reference, are:
* 'asymmetric_truncated_quadratic'[7]_
* 'symmetric_truncated_quadratic'[7]_
* 'asymmetric_huber'[7]_
* 'symmetric_huber'[7]_
* 'asymmetric_indec'[8]_
* 'symmetric_indec'[8]_
threshold : float, optional
The threshold value for the loss method, where the function goes from
quadratic loss (such as used for least squares) to non-quadratic. For
symmetric loss methods, residual values with absolute value less than
threshold will have quadratic loss. For asymmetric loss methods, residual
values less than the threshold will have quadratic loss. Default is None,
which sets `threshold` to one-tenth of the standard deviation of the input
data.
alpha_factor : float, optional
A value between 0 and 1 that controls the value of the penalty. Default is
0.99. Typically should not need to change this value.
return_coef : bool, optional
If True, will convert the polynomial coefficients for the fit baseline to
a form that fits the input x_data and return them in the params dictionary.
Default is False, since the conversion takes time.
Returns
-------
baseline : numpy.ndarray, shape (N,)
The calculated baseline.
params : dict
A dictionary with the following items:
* 'weights': numpy.ndarray, shape (N,)
The weight array used for fitting the data.
* 'tol_history': numpy.ndarray
An array containing the calculated tolerance values for
each iteration. The length of the array is the number of iterations
completed. If the last value in the array is greater than the input
`tol` value, then the function did not converge.
* 'coef': numpy.ndarray, shape (poly_order + 1,)
Only if `return_coef` is True. The array of polynomial parameters
for the baseline, in increasing order. Can be used to create a
polynomial using numpy.polynomial.polynomial.Polynomial().
Raises
------
ValueError
Raised if `alpha_factor` is not between 0 and 1.
Notes
-----
In baseline literature, this procedure is sometimes called "backcor".
References
----------
.. [7] <NAME>., et al. Background removal from spectra by designing and
minimising a non-quadratic cost function. Chemometrics and Intelligent
Laboratory Systems, 2005, 76(2), 121–133.
.. [8] <NAME>., et al. Goldindec: A Novel Algorithm for Raman Spectrum Baseline
Correction. Applied Spectroscopy, 2015, 69(7), 834-842.
"""
if not 0 < alpha_factor <= 1:
raise ValueError('alpha_factor must be between 0 and 1')
symmetric_loss, method = _identify_loss_method(cost_function)
loss_function = {
'huber': _huber_loss,
'truncated_quadratic': _truncated_quadratic_loss,
'indec': _indec_loss
}[method]
y, x, weight_array, original_domain, vander, pseudo_inverse = _setup_polynomial(
data, x_data, weights, poly_order, return_vander=True, return_pinv=True
)
if threshold is None:
threshold = np.std(y) / 10
loss_kwargs = {
'threshold': threshold, 'alpha_factor': alpha_factor, 'symmetric': symmetric_loss
}
sqrt_w = np.sqrt(weight_array)
y = sqrt_w * y
coef = np.dot(pseudo_inverse, y)
baseline = np.dot(vander, coef)
tol_history = np.empty(max_iter)
for i in range(max_iter):
baseline_old = baseline
coef = np.dot(pseudo_inverse, y + loss_function(y - sqrt_w * baseline, **loss_kwargs))
baseline = np.dot(vander, coef)
calc_difference = relative_difference(baseline_old, baseline)
tol_history[i] = calc_difference
if calc_difference < tol:
break
params = {'weights': weight_array, 'tol_history': tol_history[:i + 1]}
if return_coef:
params['coef'] = _convert_coef(coef, original_domain)
return baseline, params
def _tukey_square(residual, scale=3, symmetric=False):
"""
The square root of Tukey's bisquare function.
Parameters
----------
residual : numpy.ndarray, shape (N,)
The residual array of the fit.
scale : float, optional
A scale factor applied to the weighted residuals to control the
robustness of the fit. Default is 3.0.
symmetric : bool, optional
If False (default), will apply weighting asymmetrically, with residuals
< 0 having full weight. If True, will apply weighting the same for both
positive and negative residuals, which is regular LOESS.
Returns
-------
weights : numpy.ndarray, shape (N,)
The weighting array.
Notes
-----
The function is technically sqrt(Tukey's bisquare) since the outer
power of 2 is not performed. This is intentional, so that the square
root for weighting in least squares does not need to be done, speeding
up the calculation.
References
----------
<NAME>., et al., Baseline subtraction using robust local regression
estimation. J. Quantitative Spectroscopy and Radiative Transfer, 2001, 68,
179-193.
"""
if symmetric:
inner = residual / scale
weights = np.maximum(0, 1 - inner * inner)
else:
weights = np.ones_like(residual)
mask = residual > 0
inner = residual[mask] / scale
weights[mask] = np.maximum(0, 1 - inner * inner)
return weights
def _median_absolute_value(values):
"""
Computes the median absolute value (MAV) of an array.
Parameters
----------
values : array-like
The array of values to use for the calculation.
Returns
-------
float
The scaled median absolute value for the input array.
Notes
-----
The 1/0.6744897501960817 scale factor is to make the result comparable to the
standard deviation of a Gaussian distribution. The divisor is obtained by
calculating the value at which the cumulative distribution function of a Gaussian
distribution is 0.75 (see https://en.wikipedia.org/wiki/Median_absolute_deviation),
which can be obtained by::
from scipy.special import ndtri
ndtri(0.75) # equals 0.6744897501960817
To calculate the median absolute difference (MAD) using this function, simply do::
_median_absolute_value(values - np.median(values))
References
----------
<NAME>., et al., Baseline subtraction using robust local regression
estimation. J. Quantitative Spectroscopy and Radiative Transfer, 2001, 68,
179-193.
https://en.wikipedia.org/wiki/Median_absolute_deviation.
"""
return np.median(np.abs(values)) / 0.6744897501960817
@jit(nopython=True, cache=True)
def _loess_solver(AT, b):
"""
Solves the equation `A x = b` given `A.T` and `b`.
Parameters
----------
AT : numpy.ndarray, shape (M, N)
The transposed `A` matrix.
b : numpy.ndarray, shape (N,)
The `b` array.
Returns
-------
numpy.ndarray, shape (N,)
The solution to the normal equation.
Notes
-----
Uses np.linalg.solve (which uses LU decomposition) rather than np.linalg.lstsq
(which uses SVD) since solve is ~30-60% faster. np.linalg.solve requires ``A.T * A``,
which squares the condition number of ``A``, but on tested datasets the relative
difference when using solve vs lstsq (using np.allclose) is ~1e-10 to 1e-13 for
poly_orders of 1 or 2, which seems fine; the relative differences increase to
~1e-6 to 1e-9 for a poly_order of 3, and ~1e-4 to 1e-6 for a poly_order of 4, but
loess should use a poly_order <= 2, so that should not be a problem.
"""
return np.linalg.solve(AT.dot(AT.T), AT.dot(b))
@jit(nopython=True, cache=True, parallel=True)
def _fill_skips(x, baseline, skips):
"""
Fills in the skipped baseline points using linear interpolation.
Parameters
----------
x : numpy.ndarray
The array of x-values.
baseline : numpy.ndarray
The array of baseline values with all fit points allocated. All skipped points
will be filled in using interpolation.
skips : numpy.ndarray, shape (G, 2)
The array of left and right indices that define the windows for interpolation,
with length G being the number of interpolation segments. Indices are set such
that `baseline[skips[i][0]:skips[i][1]]` will have fitted values at the first
and last indices and all other values (the slice [1:-1]) will be calculated by
interpolation.
Notes
-----
All changes to `baseline` are done inplace.
"""
for i in prange(skips.shape[0]):
window = skips[i]
left = window[0]
right = window[1]
_interp_inplace(x[left:right], baseline[left:right], baseline[left], baseline[right - 1])
# adapted from (https://gist.github.com/agramfort/850437); see license above
@jit(nopython=True, cache=True, parallel=True)
def _loess_low_memory(x, y, weights, coefs, vander, num_x, windows, fits):
"""
A version of loess that uses near constant memory.
The distance-weighted kernel for each x-value is computed each loop, rather
than cached, so memory usage is low but the calculation is slightly slower.
Parameters
----------
x : numpy.ndarray, shape (N,)
The x-values of the measured data, with N data points.
y : numpy.ndarray, shape (N,)
The y-values of the measured data, with N points.
weights : numpy.ndarray, shape (N,)
The array of weights.
coefs : numpy.ndarray, shape (N, poly_order + 1)
The array of polynomial coefficients (with polynomial order poly_order),
for each value in `x`.
vander : numpy.ndarray, shape (N, poly_order + 1)
The Vandermonde matrix for the `x` array.
num_x : int
The number of data points in `x`, also known as N.
windows : numpy.ndarray, shape (F, 2)
An array of left and right indices that define the fitting window for each fit
x-value. The length is F, which is the total number of fit points. If `fit_dx`
is <= 0, F is equal to N, the total number of x-values.
fits : numpy.ndarray, shape (F,)
The array of indices indicating which x-values to fit.
Notes
-----
The coefficient array, `coefs`, is modified inplace.
"""
baseline = np.empty(num_x)
y_fit = y * weights
vander_fit = vander.T * weights
for idx in prange(fits.shape[0]):
i = fits[idx]
window = windows[idx]
left = window[0]
right = window[1]
difference = np.abs(x[left:right] - x[i])
difference = difference / max(difference[0], difference[-1])
difference = difference * difference * difference
difference = 1 - difference
kernel = np.sqrt(difference * difference * difference)
coef = _loess_solver(
kernel * vander_fit[:, left:right], kernel * y_fit[left:right]
)
baseline[i] = vander[i].dot(coef)
coefs[i] = coef
return baseline
# adapted from (https://gist.github.com/agramfort/850437); see license above
@jit(nopython=True, cache=True, parallel=True)
def _loess_first_loop(x, y, weights, coefs, vander, total_points, num_x, windows, fits):
"""
The initial fit for loess that also caches the window values for each x-value.
Parameters
----------
x : numpy.ndarray, shape (N,)
The x-values of the measured data, with N data points.
y : numpy.ndarray, shape (N,)
The y-values of the measured data, with N points.
weights : numpy.ndarray, shape (N,)
The array of weights.
coefs : numpy.ndarray, shape (N, poly_order + 1)
The array of polynomial coefficients (with polynomial order poly_order),
for each value in `x`.
vander : numpy.ndarray, shape (N, poly_order + 1)
The Vandermonde matrix for the `x` array.
total_points : int
The number of points to include when fitting each x-value.
num_x : int
The number of data points in `x`, also known as N.
windows : numpy.ndarray, shape (F, 2)
An array of left and right indices that define the fitting window for each fit
x-value. The length is F, which is the total number of fit points. If `fit_dx`
is <= 0, F is equal to N, the total number of x-values.
fits : numpy.ndarray, shape (F,)
The array of indices indicating which x-values to fit.
Returns
-------
kernels : numpy.ndarray, shape (num_x, total_points)
The array containing the distance-weighted kernel for each x-value.
Notes
-----
The coefficient array, `coefs`, is modified inplace.
"""
kernels = np.empty((num_x, total_points))
baseline = np.empty(num_x)
y_fit = y * weights
vander_fit = vander.T * weights
for idx in prange(fits.shape[0]):
i = fits[idx]
window = windows[idx]
left = window[0]
right = window[1]
difference = np.abs(x[left:right] - x[i])
difference = difference / max(difference[0], difference[-1])
difference = difference * difference * difference
difference = 1 - difference
kernel = np.sqrt(difference * difference * difference)
kernels[i] = kernel
coef = _loess_solver(
kernel * vander_fit[:, left:right], kernel * y_fit[left:right]
)
baseline[i] = vander[i].dot(coef)
coefs[i] = coef
return kernels, baseline
@jit(nopython=True, cache=True, parallel=True)
def _loess_nonfirst_loops(y, weights, coefs, vander, kernels, windows, num_x, fits):
"""
The loess fit to use after the first loop that uses the cached window values.
Parameters
----------
y : numpy.ndarray, shape (N,)
The y-values of the measured data, with N points.
weights : numpy.ndarray, shape (N,)
The array of weights.
coefs : numpy.ndarray, shape (N, poly_order + 1)
The array of polynomial coefficients (with polynomial order poly_order),
for each value in `x`.
vander : numpy.ndarray, shape (N, poly_order + 1)
The Vandermonde matrix for the `x` array.
kernels : numpy.ndarray, shape (N, total_points)
The array containing the distance-weighted kernel for each x-value. Each
kernel has a length of total_points.
windows : numpy.ndarray, shape (F, 2)
An array of left and right indices that define the fitting window for each fit
x-value. The length is F, which is the total number of fit points. If `fit_dx`
is <= 0, F is equal to N, the total number of x-values.
num_x : int
The total number of values, N.
fits : numpy.ndarray, shape (F,)
The array of indices indicating which x-values to fit.
Notes
-----
The coefficient array, `coefs`, is modified inplace.
"""
baseline = np.empty(num_x)
y_fit = y * weights
vander_fit = vander.T * weights
for idx in prange(fits.shape[0]):
i = fits[idx]
window = windows[idx]
left = window[0]
right = window[1]
kernel = kernels[i]
coef = _loess_solver(
kernel * vander_fit[:, left:right], kernel * y_fit[left:right]
)
baseline[i] = vander[i].dot(coef)
coefs[i] = coef
return baseline
@jit(nopython=True, cache=True)
def _determine_fits(x, num_x, total_points, delta):
"""
Determines the x-values to fit and the left and right indices for each fit x-value.
The windows are set before fitting so that fitting can be done in parallel
when numba is installed, since the left and right indices would otherwise
need to be determined in order. Similarly, determining which x-values to fit would
not be able to be done in parallel since it requires knowledge of the last x-value
fit.
Parameters
----------
x : numpy.ndarray, shape (N,)
The array of x-values.
num_x : int
The total number of x-values, N.
total_points : int
The number of values to include in each fitting window.
delta : float
If `delta` is > 0, will skip all but the last x-value in the range x_last + `delta`,
where x_last is the last x-value to be fit. Fits all x-values if `delta` is <= 0.
Returns
-------
windows : numpy.ndarray, shape (F, 2)
An array of left and right indices that define the fitting window for each fit
x-value. The length is F, which is the total number of fit points. If `fit_dx`
is <= 0, F is equal to N, the total number of x-values. Indices are set such
that the number of values in `x[windows[i][0]:windows[i][1]] is equal to
`total_points`.
fits : numpy.ndarray, shape (F,)
The array of indices indicating which x-values to fit.
skips : numpy.ndarray, shape (G, 2)
The array of left and right indices that define the windows for interpolation,
with length G being the number of interpolation segments. G is 0 if `fit_dx` is
<= 0. Indices are set such that `baseline[skips[i][0]:skips[i][1]]` will have
fitted values at the first and last indices and all other values (the slice [1:-1])
will be calculated by interpolation.
Notes
-----
The dtype `np.intp` is used for `fits`, `skips`, and `windows` to be consistent with
numpy since numpy internally uses that type when referring to indices.
"""
# faster to allocate array and return only filled in sections
# rather than constanly appending to a list
if delta > 0:
check_fits = True
fits = np.empty(num_x, dtype=np.intp)
fits[0] = 0 # always fit first item
skips = np.empty((num_x, 2), dtype=np.intp)
else:
# TODO maybe use another function when fitting all points in order
# to skip the if check_fits check for every x-value; does it affect
# calculation time that much?
check_fits = False
# TODO once numba minimum version is >= 0.47, can use dtype kwarg in np.arange
fits = np.arange(num_x).astype(np.intp)
# numba cannot compile in nopython mode when directly creating
# np.array([], dtype=np.intp), so work-around by creating np.array([[0, 0]])
# and then index with [:total_skips], which becomes np.array([])
# since total_skips is 0 when delta is <= 0.
skips = np.array([[0, 0]], dtype=np.intp)
windows = np.empty((num_x, 2), dtype=np.intp)
windows[0] = (0, total_points)
total_fits = 1
total_skips = 0
skip_start = 0
skip_range = x[0] + delta
left = 0
right = total_points
for i in range(1, num_x - 1):
x_val = x[i]
if check_fits:
# use x[i+1] rather than x[i] since it ensures that the last value within
# the range x_last_fit + delta is used; x[i+1] is also guranteed to be >= x[i]
if x[i + 1] < skip_range:
if not skip_start:
skip_start = i
continue
else:
skip_range = x_val + delta
fits[total_fits] = i
if skip_start:
skips[total_skips] = (skip_start - 1, i + 1)
total_skips += 1
skip_start = 0
while right < num_x and x_val - x[left] > x[right] - x_val:
left += 1
right += 1
window = windows[total_fits]
window[0] = left
window[1] = right
total_fits += 1
if skip_start: # fit second to last x-value
fits[total_fits] = num_x - 2
if x[-1] - x[-2] < x[-2] - x[num_x - total_points]:
windows[total_fits] = (num_x - total_points, num_x)
else:
windows[total_fits] = (num_x - total_points - 1, num_x - 1)
total_fits += 1
skips[total_skips] = (skip_start - 1, num_x - 1)
total_skips += 1
# always fit last item
fits[total_fits] = num_x - 1
windows[total_fits] = (num_x - total_points, num_x)
total_fits += 1
return windows[:total_fits], fits[:total_fits], skips[:total_skips]
def loess(data, x_data=None, fraction=0.2, total_points=None, poly_order=1, scale=3.0,
tol=1e-3, max_iter=10, symmetric_weights=False, use_threshold=False, num_std=1,
use_original=False, weights=None, return_coef=False, conserve_memory=True, delta=0.0):
"""
Locally estimated scatterplot smoothing (LOESS).
Performs polynomial regression at each data point using the nearest points.
Parameters
----------
data : array-like, shape (N,)
The y-values of the measured data, with N data points.
x_data : array-like, shape (N,), optional
The x-values of the measured data. Default is None, which will create an
array from -1 to 1 with N points.
fraction : float, optional
The fraction of N data points to include for the fitting on each point.
Default is 0.2. Not used if `total_points` is not None.
total_points : int, optional
The total number of points to include for the fitting on each point. Default
is None, which will use `fraction` * N to determine the number of points.
scale : float, optional
A scale factor applied to the weighted residuals to control the robustness
of the fit. Default is 3.0, as used in [9]_. Note that the original loess
procedure in [10]_ used a `scale` of ~4.05.
poly_order : int, optional
The polynomial order for fitting the baseline. Default is 1.
tol : float, optional
The exit criteria. Default is 1e-3.
max_iter : int, optional
The maximum number of iterations. Default is 10.
symmetric_weights : bool, optional
If False (default), will apply weighting asymmetrically, with residuals
< 0 having a weight of 1, according to [9]_. If True, will apply weighting
the same for both positive and negative residuals, which is regular LOESS.
If `use_threshold` is True, this parameter is ignored.
use_threshold : bool, optional
If False (default), will compute weights each iteration to perform the
robust fitting, which is regular LOESS. If True, will apply a threshold
on the data being fit each iteration, based on the maximum values of the
data and the fit baseline, as proposed by [11]_, similar to the modpoly
and imodpoly techniques.
num_std : float, optional
The number of standard deviations to include when thresholding. Default
is 1, which is the value used for the imodpoly technique. Only used if
`use_threshold` is True.
use_original : bool, optional
If False (default), will compare the baseline of each iteration with
the y-values of that iteration [12]_ when choosing minimum values for
thresholding. If True, will compare the baseline with the original
y-values given by `data` [13]_. Only used if `use_threshold` is True.
weights : array-like, shape (N,), optional
The weighting array. If None (default), then will be an array with
size equal to N and all values set to 1.
return_coef : bool, optional
If True, will convert the polynomial coefficients for the fit baseline to
a form that fits the input x_data and return them in the params dictionary.
Default is False, since the conversion takes time.
conserve_memory : bool, optional
If False, will cache the distance-weighted kernels for each value
in `x_data` on the first iteration and reuse them on subsequent iterations to
save time. The shape of the array of kernels is (len(`x_data`), `total_points`).
If True (default), will recalculate the kernels each iteration, which uses very
little memory, but is slower. Can usually set to False unless `x_data` and`total_points`
are quite large and the function causes memory issues when cacheing the kernels. If
numba is installed, there is no significant time difference since the calculations are
sped up.
delta : float, optional
If `delta` is > 0, will skip all but the last x-value in the range x_last + `delta`,
where x_last is the last x-value to be fit using weighted least squares, and instead
use linear interpolation to calculate the fit for those x-values (same behavior as in
statsmodels [14]_ and Cleveland's original Fortran lowess implementation [15]_).
Fits all x-values if `delta` is <= 0. Default is 0.0. Note that `x_data` is scaled to
fit in the range [-1, 1], so `delta` should likewise be scaled. For example, if the
desired `delta` value was ``0.01 * (max(x_data) - min(x_data))``, then the
correctly scaled `delta` would be 0.02 (ie. ``0.01 * (1 - (-1))``).
Returns
-------
baseline : numpy.ndarray, shape (N,)
The calculated baseline.
params : dict
A dictionary with the following items:
* 'weights': numpy.ndarray, shape (N,)
The weight array used for fitting the data. Does NOT contain the
individual distance-weighted kernels for each x-value.
* 'tol_history': numpy.ndarray
An array containing the calculated tolerance values for
each iteration. The length of the array is the number of iterations
completed. If the last value in the array is greater than the input
`tol` value, then the function did not converge.
* 'coef': numpy.ndarray, shape (N, poly_order + 1)
Only if `return_coef` is True. The array of polynomial parameters
for the baseline, in increasing order. Can be used to create a polynomial
using numpy.polynomial.polynomial.Polynomial(). If `delta` is > 0, the
coefficients for any skipped x-value will all be 0.
Raises
------
ValueError
Raised if the number of points per window for the fitting is less than
`poly_order` + 1 or greater than the total number of points.
Notes
-----
The iterative, robust, aspect of the fitting can be achieved either through
reweighting based on the residuals (the typical usage), or thresholding the
fit data based on the residuals, as proposed by [11]_, similar to the modpoly
and imodpoly techniques.
In baseline literature, this procedure is sometimes called "rbe", meaning
"robust baseline estimate".
References
----------
.. [9] <NAME>., et al. Baseline subtraction using robust local
regression estimation. J. Quantitative Spectroscopy and Radiative
Transfer, 2001, 68, 179-193.
.. [10] <NAME>. Robust locally weighted regression and smoothing
scatterplots. Journal of the American Statistical Association,
1979, 74(368), 829-836.
.. [11] <NAME>. Comparison of Several Methods of Chromatographic
Baseline Removal with a New Approach Based on Quantile Regression.
Chromatographia, 2011, 73, 721-731.
.. [12] <NAME>., et al. Baseline correction by improved iterative polynomial
fitting with automatic threshold. Chemometrics and Intelligent
Laboratory Systems, 2006, 82, 59-65.
.. [13] <NAME>., et al. Automated method for subtraction of fluorescence
from biological raman spectra. Applied Spectroscopy, 2003, 57(11),
1363-1367.
.. [14] https://github.com/statsmodels/statsmodels.
.. [15] https://www.netlib.org/go (lowess.f is the file).
"""
y, x, weight_array, original_domain = _setup_polynomial(data, x_data, weights, poly_order)
num_x = x.shape[0]
if total_points is None:
total_points = ceil(fraction * num_x)
if total_points < poly_order + 1:
raise ValueError('total points must be greater than polynomial order + 1')
elif total_points > num_x:
raise ValueError((
'points per window is higher than total number of points; lower either '
'"fraction" or "total_points"'
))
elif poly_order > 2:
warnings.warn(
('polynomial orders greater than 2 can have numerical issues;'
' consider using a polynomial order of 1 or 2 instead'), ParameterWarning
)
sort_x = x_data is not None
if sort_x:
sort_order = np.argsort(x, kind='mergesort') # to ensure x is increasing
x = x[sort_order]
y = y[sort_order]
weight_array = weight_array[sort_order]
if use_original:
y0 = y
# find the indices for fitting beforehand so that the fitting can be done
# in parallel; cast delta as float so numba does not have to compile for
# both int and float
windows, fits, skips = _determine_fits(x, num_x, total_points, float(delta))
# np.polynomial.polynomial.polyvander returns a Fortran-ordered array, which
# when matrix multiplied with the C-ordered coefficient array gives a warning
# when using numba, so convert Vandermonde matrix to C-ordering.
vander = np.ascontiguousarray(_get_vander(x, poly_order, calc_pinv=False))
baseline = y
coefs = np.zeros((num_x, poly_order + 1))
tol_history = np.empty(max_iter + 1)
sqrt_w = np.sqrt(weight_array)
# do max_iter + 1 since a max_iter of 0 would return y as baseline otherwise
for i in range(max_iter + 1):
baseline_old = baseline
if conserve_memory:
baseline = _loess_low_memory(
x, y, sqrt_w, coefs, vander, num_x, windows, fits
)
elif i == 0:
kernels, baseline = _loess_first_loop(
x, y, sqrt_w, coefs, vander, total_points, num_x, windows, fits
)
else:
baseline = _loess_nonfirst_loops(
y, sqrt_w, coefs, vander, kernels, windows, num_x, fits
)
_fill_skips(x, baseline, skips)
calc_difference = relative_difference(baseline_old, baseline)
tol_history[i] = calc_difference
if calc_difference < tol:
break
if use_threshold:
y = np.minimum(
y0 if use_original else y, baseline + num_std * np.std(y - baseline)
)
else:
residual = y - baseline
# TODO median_absolute_value can be 0 if more than half of residuals are
# 0 (perfect fit); can that ever really happen? if so, should prevent dividing by 0
sqrt_w = _tukey_square(
residual / _median_absolute_value(residual), scale, symmetric_weights
)
params = {'weights': sqrt_w**2, 'tol_history': tol_history[:i + 1]}
if return_coef:
# TODO maybe leave out the coefficients from the rest of the calculations
# since they are otherwise unused, and just fit x vs baseline here; would
# save a little memory; is providing coefficients for loess even useful?
params['coef'] = np.array([_convert_coef(coef, original_domain) for coef in coefs])
if sort_x:
inverted_order = _inverted_sort(sort_order)
baseline = baseline[inverted_order]
params['weights'] = params['weights'][inverted_order]
if return_coef:
params['coef'] = params['coef'][inverted_order]
return baseline, params
def quant_reg(data, x_data=None, poly_order=2, quantile=0.05, tol=1e-6, max_iter=250,
weights=None, eps=None, return_coef=False):
"""
Approximates the baseline of the data using quantile regression.
Parameters
----------
data : array-like, shape (N,)
The y-values of the measured data, with N data points.
x_data : array-like, shape (N,), optional
The x-values of the measured data. Default is None, which will create an
array from -1 to 1 with N points.
poly_order : int, optional
The polynomial order for fitting the baseline. Default is 2.
quantile : float, optional
The quantile at which to fit the baseline. Default is 0.05.
tol : float, optional
The exit criteria. Default is 1e-6. For extreme quantiles (`quantile` < 0.01
or `quantile` > 0.99), may need to use a lower value to get a good fit.
max_iter : int, optional
The maximum number of iterations. Default is 250. For extreme quantiles
(`quantile` < 0.01 or `quantile` > 0.99), may need to use a higher value to
ensure convergence.
weights : array-like, shape (N,), optional
The weighting array. If None (default), then will be an array with
size equal to N and all values set to 1.
eps : float, optional
A small value added to the square of the residual to prevent dividing by 0.
Default is None, which uses the square of the maximum-absolute-value of the
fit each iteration multiplied by 1e-6.
return_coef : bool, optional
If True, will convert the polynomial coefficients for the fit baseline to
a form that fits the input `x_data` and return them in the params dictionary.
Default is False, since the conversion takes time.
Returns
-------
baseline : numpy.ndarray, shape (N,)
The calculated baseline.
params : dict
A dictionary with the following items:
* 'weights': numpy.ndarray, shape (N,)
The weight array used for fitting the data.
* 'tol_history': numpy.ndarray
An array containing the calculated tolerance values for
each iteration. The length of the array is the number of iterations
completed. If the last value in the array is greater than the input
`tol` value, then the function did not converge.
* 'coef': numpy.ndarray, shape (poly_order + 1,)
Only if `return_coef` is True. The array of polynomial parameters
for the baseline, in increasing order. Can be used to create a
polynomial using numpy.polynomial.polynomial.Polynomial().
Raises
------
ValueError
Raised if `quantile` is not between 0 and 1.
Notes
-----
Application of quantile regression for baseline fitting as described in [16]_.
Performs quantile regression using iteratively reweighted least squares (IRLS)
as described in [17]_.
References
----------
.. [16] Komsta, Ł. Comparison of Several Methods of Chromatographic
Baseline Removal with a New Approach Based on Quantile Regression.
Chromatographia, 2011, 73, 721-731.
.. [17] <NAME>., et al. Simultaneous estimation of quantile curves using
quantile sheets. AStA Advances in Statistical Analysis, 2013, 97, 77-87.
"""
# TODO provide a way to estimate best poly_order based on AIC like in Komsta? could be
# useful for all polynomial methods; maybe could be an optimizer function
if not 0 < quantile < 1:
raise ValueError('quantile must be between 0 and 1.')
y, x, weight_array, original_domain, vander = _setup_polynomial(
data, x_data, weights, poly_order, return_vander=True
)
# estimate first iteration using least squares
coef = np.linalg.lstsq(vander * weight_array[:, None], y * weight_array, None)[0]
baseline = vander @ coef
tol_history = np.empty(max_iter)
for i in range(max_iter):
baseline_old = baseline
weight_array = np.sqrt(_weighting._quantile(y, baseline, quantile, eps))
coef = np.linalg.lstsq(vander * weight_array[:, None], y * weight_array, None)[0]
baseline = vander @ coef
# relative_difference(baseline_old, baseline, 1) gives nearly same result and
# the l2 norm is faster to calculate, so use that instead of l1 norm
calc_difference = relative_difference(baseline_old, baseline)
tol_history[i] = calc_difference
if calc_difference < tol:
break
params = {'weights': weight_array**2, 'tol_history': tol_history[:i + 1]}
if return_coef:
params['coef'] = _convert_coef(coef, original_domain)
return baseline, params
def goldindec(data, x_data=None, poly_order=2, tol=1e-3, max_iter=250, weights=None,
cost_function='asymmetric_indec', peak_ratio=0.5, alpha_factor=0.99,
tol_2=1e-3, tol_3=1e-6, max_iter_2=100, return_coef=False):
"""
Fits a polynomial baseline using a non-quadratic cost function.
The non-quadratic cost functions penalize residuals with larger values,
giving a more robust fit compared to normal least-squares.
Parameters
----------
data : array-like, shape (N,)
The y-values of the measured data, with N data points.
x_data : array-like, shape (N,), optional
The x-values of the measured data. Default is None, which will create an
array from -1 to 1 with N points.
poly_order : int, optional
The polynomial order for fitting the baseline. Default is 2.
tol : float, optional
The exit criteria for the fitting with a given threshold value. Default is 1e-3.
max_iter : int, optional
The maximum number of iterations for fitting a threshold value. Default is 250.
weights : array-like, shape (N,), optional
The weighting array. If None (default), then will be an array with
size equal to N and all values set to 1.
cost_function : str, optional
The non-quadratic cost function to minimize. Unlike :func:`.penalized_poly`,
this function only works with asymmetric cost functions, so the symmetry prefix
('a' or 'asymmetric') is optional (eg. 'indec' and 'a_indec' are the same). Default
is 'asymmetric_indec'. Available methods, and their associated reference, are:
* 'asymmetric_indec'[18]_
* 'asymmetric_truncated_quadratic'[19]_
* 'asymmetric_huber'[19]_
peak_ratio : float, optional
A value between 0 and 1 that designates how many points in the data belong
to peaks. Values are valid within ~10% of the actual peak ratio. Default is 0.5.
alpha_factor : float, optional
A value between 0 and 1 that controls the value of the penalty. Default is
0.99. Typically should not need to change this value.
tol_2 : float, optional
The exit criteria for the difference between the optimal up-down ratio (number of
points above 0 in the residual compared to number of points below 0) and the up-down
ratio for a given threshold value. Default is 1e-3.
tol_3 : float, optional
The exit criteria for the relative change in the threshold value. Default is 1e-6.
max_iter_2 : float, optional
The number of iterations for iterating between different threshold values.
Default is 100.
return_coef : bool, optional
If True, will convert the polynomial coefficients for the fit baseline to
a form that fits the input x_data and return them in the params dictionary.
Default is False, since the conversion takes time.
Returns
-------
baseline : numpy.ndarray, shape (N,)
The calculated baseline.
params : dict
A dictionary with the following items:
* 'weights': numpy.ndarray, shape (N,)
The weight array used for fitting the data.
* 'tol_history': numpy.ndarray, shape (J, K)
An array containing the calculated tolerance values for each iteration
of both threshold values and fit values. Index 0 are the tolerence values
for the difference in up-down ratios, index 1 are the tolerance values for
the relative change in the threshold, and indices >= 2 are the tolerance values
for each fit. All values that were not used in fitting have values of 0. Shape J
is 2 plus the number of iterations for the threshold to converge (related to
`max_iter_2`, `tol_2`, `tol_3`), and shape K is the maximum of the number of
iterations for the threshold and the maximum number of iterations for all of
the fits of the various threshold values (related to `max_iter` and `tol`).
* 'threshold' : float
The optimal threshold value. Could be used in :func:`.penalized_poly`
for fitting other similar data.
* 'coef': numpy.ndarray, shape (poly_order + 1,)
Only if `return_coef` is True. The array of polynomial parameters
for the baseline, in increasing order. Can be used to create a
polynomial using numpy.polynomial.polynomial.Polynomial().
Raises
------
ValueError
Raised if `alpha_factor` or `peak_ratio` are not between 0 and 1, or if the
specified cost function is symmetric.
References
----------
.. [18] <NAME>., et al. Goldindec: A Novel Algorithm for Raman Spectrum Baseline
Correction. Applied Spectroscopy, 2015, 69(7), 834-842.
.. [19] <NAME>., et al. Background removal from spectra by designing and
minimising a non-quadratic cost function. Chemometrics and Intelligent
Laboratory Systems, 2005, 76(2), 121–133.
"""
if not 0 < alpha_factor <= 1:
raise ValueError('alpha_factor must be between 0 and 1')
elif not 0 < peak_ratio < 1:
raise ValueError('peak_ratio must be between 0 and 1')
try:
symmetric_loss, method = _identify_loss_method(cost_function)
except ValueError: # do not require a prefix since cost must be asymmetric
symmetric_loss, method = _identify_loss_method('a_' + cost_function)
if symmetric_loss:
# symmetric cost functions don't work due to how the up-down ratio vs
# peak_ratio function was created in the reference; in theory, could simulate
# spectra with both positive and negative peaks following the reference
# and build another empirical function, but would likely need to also
# add other parameters detailing the percent of positive vs negative peaks,
# etc., so it's not worth the effort
raise ValueError('goldindec only works for asymmetric cost functions')
loss_function = {
'huber': _huber_loss,
'truncated_quadratic': _truncated_quadratic_loss,
'indec': _indec_loss
}[method]
y, x, weight_array, original_domain, vander, pseudo_inverse = _setup_polynomial(
data, x_data, weights, poly_order, return_vander=True, return_pinv=True
)
num_y = y.shape[0]
up_down_ratio_goal = (
0.7679 + 11.2358 * peak_ratio - 39.7064 * peak_ratio**2 + 92.3583 * peak_ratio**3
)
# TODO reference states threshold must be <= 2 for half-quadratic minimization to
# be valid for indec cost function, and normalized y so that threshold is always <= 2;
# however, it seems to work fine without normalization; just be aware in case errors
# occur, may have to normalize y in both this function and penalized_poly
sqrt_w = np.sqrt(weight_array)
y_fit = sqrt_w * y
coef = np.dot(pseudo_inverse, y_fit)
initial_baseline = np.dot(vander, coef)
a = 0
# reference used b=1, but normalized y before fitting; instead, set b as max of
# initial residual
b = abs((y - initial_baseline).max())
threshold = a + 0.618 * (b - a)
loss_kwargs = {
'threshold': threshold, 'alpha_factor': alpha_factor,
'symmetric': symmetric_loss
}
# have to use zeros rather than empty for tol_history since each inner fit may
# have a different number of iterations
tol_history = np.zeros((max_iter_2 + 2, max(max_iter, max_iter_2)))
j_max = 0
for i in range(max_iter_2):
baseline = initial_baseline
for j in range(max_iter):
baseline_old = baseline
coef = np.dot(
pseudo_inverse, y_fit + loss_function(y_fit - sqrt_w * baseline, **loss_kwargs)
)
baseline = np.dot(vander, coef)
calc_difference = relative_difference(baseline_old, baseline)
tol_history[i + 2, j] = calc_difference
if calc_difference < tol:
break
if j > j_max:
j_max = j
up_count = (y > baseline).sum()
up_down_ratio = up_count / max(1, num_y - up_count)
calc_difference = up_down_ratio - up_down_ratio_goal
tol_history[0, i] = calc_difference
if calc_difference > tol_2:
a = threshold
elif calc_difference < -tol_2:
b = threshold
else:
break
threshold = a + 0.618 * (b - a)
# this exit criteria was not stated in the reference, but the change in threshold
# becomes zero fairly quickly, so need to also exit rather than needlessly
# continuing to calculate with the same threshold value
calc_difference = relative_difference(loss_kwargs['threshold'], threshold)
tol_history[1, i] = calc_difference
if calc_difference < tol_3:
break
loss_kwargs['threshold'] = threshold
params = {
'weights': weight_array, 'tol_history': tol_history[:i + 3, :max(i, j_max) + 1],
'threshold': loss_kwargs['threshold']
}
if return_coef:
params['coef'] = _convert_coef(coef, original_domain)
return baseline, params
| [
"numpy.ones_like",
"numpy.abs",
"numpy.sqrt",
"numpy.minimum",
"math.ceil",
"numpy.argsort",
"numpy.array",
"numpy.dot",
"numpy.zeros",
"numpy.empty",
"numpy.sign",
"numpy.linalg.lstsq",
"numpy.std",
"warnings.warn",
"numpy.maximum",
"numpy.arange"
] | [((9204, 9225), 'numpy.sqrt', 'np.sqrt', (['weight_array'], {}), '(weight_array)\n', (9211, 9225), True, 'import numpy as np\n'), ((9274, 9308), 'numpy.dot', 'np.dot', (['pseudo_inverse', '(sqrt_w * y)'], {}), '(pseudo_inverse, sqrt_w * y)\n', (9280, 9308), True, 'import numpy as np\n'), ((9324, 9344), 'numpy.dot', 'np.dot', (['vander', 'coef'], {}), '(vander, coef)\n', (9330, 9344), True, 'import numpy as np\n'), ((9655, 9673), 'numpy.empty', 'np.empty', (['max_iter'], {}), '(max_iter)\n', (9663, 9673), True, 'import numpy as np\n'), ((13785, 13806), 'numpy.sqrt', 'np.sqrt', (['weight_array'], {}), '(weight_array)\n', (13792, 13806), True, 'import numpy as np\n'), ((13855, 13889), 'numpy.dot', 'np.dot', (['pseudo_inverse', '(sqrt_w * y)'], {}), '(pseudo_inverse, sqrt_w * y)\n', (13861, 13889), True, 'import numpy as np\n'), ((13905, 13925), 'numpy.dot', 'np.dot', (['vander', 'coef'], {}), '(vander, coef)\n', (13911, 13925), True, 'import numpy as np\n'), ((13942, 13962), 'numpy.std', 'np.std', (['(y - baseline)'], {}), '(y - baseline)\n', (13948, 13962), True, 'import numpy as np\n'), ((14167, 14185), 'numpy.empty', 'np.empty', (['max_iter'], {}), '(max_iter)\n', (14175, 14185), True, 'import numpy as np\n'), ((26700, 26721), 'numpy.sqrt', 'np.sqrt', (['weight_array'], {}), '(weight_array)\n', (26707, 26721), True, 'import numpy as np\n'), ((26753, 26778), 'numpy.dot', 'np.dot', (['pseudo_inverse', 'y'], {}), '(pseudo_inverse, y)\n', (26759, 26778), True, 'import numpy as np\n'), ((26794, 26814), 'numpy.dot', 'np.dot', (['vander', 'coef'], {}), '(vander, coef)\n', (26800, 26814), True, 'import numpy as np\n'), ((26833, 26851), 'numpy.empty', 'np.empty', (['max_iter'], {}), '(max_iter)\n', (26841, 26851), True, 'import numpy as np\n'), ((33845, 33860), 'numpy.empty', 'np.empty', (['num_x'], {}), '(num_x)\n', (33853, 33860), True, 'import numpy as np\n'), ((36211, 36242), 'numpy.empty', 'np.empty', (['(num_x, total_points)'], {}), '((num_x, total_points))\n', (36219, 36242), True, 'import numpy as np\n'), ((36258, 36273), 'numpy.empty', 'np.empty', (['num_x'], {}), '(num_x)\n', (36266, 36273), True, 'import numpy as np\n'), ((38394, 38409), 'numpy.empty', 'np.empty', (['num_x'], {}), '(num_x)\n', (38402, 38409), True, 'import numpy as np\n'), ((41977, 42012), 'numpy.empty', 'np.empty', (['(num_x, 2)'], {'dtype': 'np.intp'}), '((num_x, 2), dtype=np.intp)\n', (41985, 42012), True, 'import numpy as np\n'), ((52749, 52782), 'numpy.zeros', 'np.zeros', (['(num_x, poly_order + 1)'], {}), '((num_x, poly_order + 1))\n', (52757, 52782), True, 'import numpy as np\n'), ((52801, 52823), 'numpy.empty', 'np.empty', (['(max_iter + 1)'], {}), '(max_iter + 1)\n', (52809, 52823), True, 'import numpy as np\n'), ((52837, 52858), 'numpy.sqrt', 'np.sqrt', (['weight_array'], {}), '(weight_array)\n', (52844, 52858), True, 'import numpy as np\n'), ((58879, 58897), 'numpy.empty', 'np.empty', (['max_iter'], {}), '(max_iter)\n', (58887, 58897), True, 'import numpy as np\n'), ((66570, 66591), 'numpy.sqrt', 'np.sqrt', (['weight_array'], {}), '(weight_array)\n', (66577, 66591), True, 'import numpy as np\n'), ((66627, 66656), 'numpy.dot', 'np.dot', (['pseudo_inverse', 'y_fit'], {}), '(pseudo_inverse, y_fit)\n', (66633, 66656), True, 'import numpy as np\n'), ((66680, 66700), 'numpy.dot', 'np.dot', (['vander', 'coef'], {}), '(vander, coef)\n', (66686, 66700), True, 'import numpy as np\n'), ((9546, 9567), 'numpy.sqrt', 'np.sqrt', (['weight_array'], {}), '(weight_array)\n', (9553, 9567), True, 'import numpy as np\n'), ((9748, 9795), 'numpy.minimum', 'np.minimum', (['(y0 if use_original else y)', 'baseline'], {}), '(y0 if use_original else y, baseline)\n', (9758, 9795), True, 'import numpy as np\n'), ((9811, 9845), 'numpy.dot', 'np.dot', (['pseudo_inverse', '(sqrt_w * y)'], {}), '(pseudo_inverse, sqrt_w * y)\n', (9817, 9845), True, 'import numpy as np\n'), ((9865, 9885), 'numpy.dot', 'np.dot', (['vander', 'coef'], {}), '(vander, coef)\n', (9871, 9885), True, 'import numpy as np\n'), ((14058, 14079), 'numpy.sqrt', 'np.sqrt', (['weight_array'], {}), '(weight_array)\n', (14065, 14079), True, 'import numpy as np\n'), ((14228, 14297), 'numpy.minimum', 'np.minimum', (['(y0 if use_original else y)', '(baseline + num_std * deviation)'], {}), '(y0 if use_original else y, baseline + num_std * deviation)\n', (14238, 14297), True, 'import numpy as np\n'), ((14313, 14347), 'numpy.dot', 'np.dot', (['pseudo_inverse', '(sqrt_w * y)'], {}), '(pseudo_inverse, sqrt_w * y)\n', (14319, 14347), True, 'import numpy as np\n'), ((14367, 14387), 'numpy.dot', 'np.dot', (['vander', 'coef'], {}), '(vander, coef)\n', (14373, 14387), True, 'import numpy as np\n'), ((14412, 14432), 'numpy.std', 'np.std', (['(y - baseline)'], {}), '(y - baseline)\n', (14418, 14432), True, 'import numpy as np\n'), ((20273, 20290), 'numpy.sign', 'np.sign', (['residual'], {}), '(residual)\n', (20280, 20290), True, 'import numpy as np\n'), ((27028, 27048), 'numpy.dot', 'np.dot', (['vander', 'coef'], {}), '(vander, coef)\n', (27034, 27048), True, 'import numpy as np\n'), ((28659, 28691), 'numpy.maximum', 'np.maximum', (['(0)', '(1 - inner * inner)'], {}), '(0, 1 - inner * inner)\n', (28669, 28691), True, 'import numpy as np\n'), ((28720, 28742), 'numpy.ones_like', 'np.ones_like', (['residual'], {}), '(residual)\n', (28732, 28742), True, 'import numpy as np\n'), ((28834, 28866), 'numpy.maximum', 'np.maximum', (['(0)', '(1 - inner * inner)'], {}), '(0, 1 - inner * inner)\n', (28844, 28866), True, 'import numpy as np\n'), ((34084, 34112), 'numpy.abs', 'np.abs', (['(x[left:right] - x[i])'], {}), '(x[left:right] - x[i])\n', (34090, 34112), True, 'import numpy as np\n'), ((34293, 34338), 'numpy.sqrt', 'np.sqrt', (['(difference * difference * difference)'], {}), '(difference * difference * difference)\n', (34300, 34338), True, 'import numpy as np\n'), ((36497, 36525), 'numpy.abs', 'np.abs', (['(x[left:right] - x[i])'], {}), '(x[left:right] - x[i])\n', (36503, 36525), True, 'import numpy as np\n'), ((36706, 36751), 'numpy.sqrt', 'np.sqrt', (['(difference * difference * difference)'], {}), '(difference * difference * difference)\n', (36713, 36751), True, 'import numpy as np\n'), ((41141, 41171), 'numpy.empty', 'np.empty', (['num_x'], {'dtype': 'np.intp'}), '(num_x, dtype=np.intp)\n', (41149, 41171), True, 'import numpy as np\n'), ((41233, 41268), 'numpy.empty', 'np.empty', (['(num_x, 2)'], {'dtype': 'np.intp'}), '((num_x, 2), dtype=np.intp)\n', (41241, 41268), True, 'import numpy as np\n'), ((41928, 41961), 'numpy.array', 'np.array', (['[[0, 0]]'], {'dtype': 'np.intp'}), '([[0, 0]], dtype=np.intp)\n', (41936, 41961), True, 'import numpy as np\n'), ((51319, 51341), 'math.ceil', 'ceil', (['(fraction * num_x)'], {}), '(fraction * num_x)\n', (51323, 51341), False, 'from math import ceil\n'), ((51948, 51979), 'numpy.argsort', 'np.argsort', (['x'], {'kind': '"""mergesort"""'}), "(x, kind='mergesort')\n", (51958, 51979), True, 'import numpy as np\n'), ((58757, 58828), 'numpy.linalg.lstsq', 'np.linalg.lstsq', (['(vander * weight_array[:, None])', '(y * weight_array)', 'None'], {}), '(vander * weight_array[:, None], y * weight_array, None)\n', (58772, 58828), True, 'import numpy as np\n'), ((5537, 5558), 'numpy.sqrt', 'np.sqrt', (['weight_array'], {}), '(weight_array)\n', (5544, 5558), True, 'import numpy as np\n'), ((16471, 16487), 'numpy.abs', 'np.abs', (['residual'], {}), '(residual)\n', (16477, 16487), True, 'import numpy as np\n'), ((18482, 18498), 'numpy.abs', 'np.abs', (['residual'], {}), '(residual)\n', (18488, 18498), True, 'import numpy as np\n'), ((20224, 20240), 'numpy.abs', 'np.abs', (['residual'], {}), '(residual)\n', (20230, 20240), True, 'import numpy as np\n'), ((26555, 26564), 'numpy.std', 'np.std', (['y'], {}), '(y)\n', (26561, 26564), True, 'import numpy as np\n'), ((30109, 30123), 'numpy.abs', 'np.abs', (['values'], {}), '(values)\n', (30115, 30123), True, 'import numpy as np\n'), ((59056, 59127), 'numpy.linalg.lstsq', 'np.linalg.lstsq', (['(vander * weight_array[:, None])', '(y * weight_array)', 'None'], {}), '(vander * weight_array[:, None], y * weight_array, None)\n', (59071, 59127), True, 'import numpy as np\n'), ((67532, 67552), 'numpy.dot', 'np.dot', (['vander', 'coef'], {}), '(vander, coef)\n', (67538, 67552), True, 'import numpy as np\n'), ((16615, 16632), 'numpy.sign', 'np.sign', (['residual'], {}), '(residual)\n', (16622, 16632), True, 'import numpy as np\n'), ((41597, 41613), 'numpy.arange', 'np.arange', (['num_x'], {}), '(num_x)\n', (41606, 41613), True, 'import numpy as np\n'), ((51693, 51849), 'warnings.warn', 'warnings.warn', (['"""polynomial orders greater than 2 can have numerical issues; consider using a polynomial order of 1 or 2 instead"""', 'ParameterWarning'], {}), "(\n 'polynomial orders greater than 2 can have numerical issues; consider using a polynomial order of 1 or 2 instead'\n , ParameterWarning)\n", (51706, 51849), False, 'import warnings\n'), ((9499, 9519), 'numpy.std', 'np.std', (['(y - baseline)'], {}), '(y - baseline)\n', (9505, 9519), True, 'import numpy as np\n'), ((20651, 20692), 'numpy.maximum', 'np.maximum', (['(2 * residual ** 2)', '_MIN_FLOAT'], {}), '(2 * residual ** 2, _MIN_FLOAT)\n', (20661, 20692), True, 'import numpy as np\n'), ((53792, 53812), 'numpy.std', 'np.std', (['(y - baseline)'], {}), '(y - baseline)\n', (53798, 53812), True, 'import numpy as np\n')] |
from django.urls import path
from . import views
urlpatterns = [
path('<int:sproperty_id>/',views.contact,name="contact"),
path('',views.contact_submit,name="contact-submit"),
] | [
"django.urls.path"
] | [((75, 133), 'django.urls.path', 'path', (['"""<int:sproperty_id>/"""', 'views.contact'], {'name': '"""contact"""'}), "('<int:sproperty_id>/', views.contact, name='contact')\n", (79, 133), False, 'from django.urls import path\n'), ((137, 190), 'django.urls.path', 'path', (['""""""', 'views.contact_submit'], {'name': '"""contact-submit"""'}), "('', views.contact_submit, name='contact-submit')\n", (141, 190), False, 'from django.urls import path\n')] |
#!/usr/bin/env python
#
# Copyright 2010 Google Inc.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
#
from google.appengine.ext import db
from google.appengine.ext import webapp
from google.appengine.ext.webapp import util
class IndexHost(db.Model):
record_type = db.StringProperty()
record_name = db.StringProperty()
listed = db.BooleanProperty()
class NameServer(db.Model):
ip = db.StringProperty()
hostname = db.StringProperty()
name = db.StringProperty()
listed = db.BooleanProperty()
city = db.StringProperty()
region = db.StringProperty()
country = db.StringProperty()
country_code = db.StringProperty()
coordinates = db.GeoPtProperty()
is_global = db.BooleanProperty()
is_regional = db.BooleanProperty()
is_custom = db.BooleanProperty()
url = db.LinkProperty()
timestamp = db.DateTimeProperty(auto_now_add=True)
class Submission(db.Model):
client_id = db.IntegerProperty()
submit_id = db.IntegerProperty()
class_c = db.StringProperty()
timestamp = db.DateTimeProperty(auto_now_add=True)
listed = db.BooleanProperty()
hidden = db.BooleanProperty()
city = db.StringProperty()
region = db.StringProperty()
country = db.StringProperty()
country_code = db.StringProperty()
coordinates = db.GeoPtProperty()
# de-normalized data, also duplicated in RunResults (though much slower)
best_nameserver = db.ReferenceProperty(NameServer, collection_name='best_submissions')
best_improvement = db.FloatProperty()
primary_nameserver = db.ReferenceProperty(NameServer, collection_name="primary_submissions")
class SubmissionConfig(db.Model):
submission = db.ReferenceProperty(Submission, collection_name='config')
input_source = db.StringProperty()
benchmark_thread_count = db.IntegerProperty()
health_thread_count = db.IntegerProperty()
health_timeout = db.FloatProperty()
timeout = db.FloatProperty()
query_count = db.IntegerProperty()
run_count = db.IntegerProperty()
platform = db.StringProperty()
version = db.StringProperty()
class SubmissionNameServer(db.Model):
nameserver = db.ReferenceProperty(NameServer, collection_name='submissions')
submission = db.ReferenceProperty(Submission, collection_name='nameservers')
is_error_prone = db.BooleanProperty()
is_disabled = db.BooleanProperty()
is_reference = db.BooleanProperty()
overall_average = db.FloatProperty()
check_average = db.FloatProperty()
averages = db.ListProperty(float)
duration_min = db.FloatProperty()
duration_max = db.FloatProperty()
error_count = db.IntegerProperty()
timeout_count = db.IntegerProperty()
nx_count = db.IntegerProperty()
position = db.IntegerProperty()
sys_position = db.IntegerProperty()
version = db.StringProperty()
node_ids = db.ListProperty(str)
# TODO(tstromberg): Remove obsoleted improvement variable
improvement = db.FloatProperty()
diff = db.FloatProperty()
notes = db.ListProperty(str)
port_behavior = db.StringProperty()
# Store one row per run for run_results, since we do not need to do much with them.
class RunResult(db.Model):
submission_nameserver = db.ReferenceProperty(SubmissionNameServer, collection_name='results')
run_number = db.IntegerProperty()
durations = db.ListProperty(float)
answer_counts = db.ListProperty(int)
# We may want to compare index results, so we will store one row per record
class IndexResult(db.Model):
submission_nameserver = db.ReferenceProperty(SubmissionNameServer, collection_name='index_results')
index_host = db.ReferenceProperty(IndexHost, collection_name='results')
duration = db.FloatProperty()
answer_count = db.IntegerProperty()
ttl = db.IntegerProperty()
response = db.StringProperty()
| [
"google.appengine.ext.db.FloatProperty",
"google.appengine.ext.db.BooleanProperty",
"google.appengine.ext.db.GeoPtProperty",
"google.appengine.ext.db.IntegerProperty",
"google.appengine.ext.db.ReferenceProperty",
"google.appengine.ext.db.ListProperty",
"google.appengine.ext.db.DateTimeProperty",
"goog... | [((766, 785), 'google.appengine.ext.db.StringProperty', 'db.StringProperty', ([], {}), '()\n', (783, 785), False, 'from google.appengine.ext import db\n'), ((802, 821), 'google.appengine.ext.db.StringProperty', 'db.StringProperty', ([], {}), '()\n', (819, 821), False, 'from google.appengine.ext import db\n'), ((833, 853), 'google.appengine.ext.db.BooleanProperty', 'db.BooleanProperty', ([], {}), '()\n', (851, 853), False, 'from google.appengine.ext import db\n'), ((890, 909), 'google.appengine.ext.db.StringProperty', 'db.StringProperty', ([], {}), '()\n', (907, 909), False, 'from google.appengine.ext import db\n'), ((923, 942), 'google.appengine.ext.db.StringProperty', 'db.StringProperty', ([], {}), '()\n', (940, 942), False, 'from google.appengine.ext import db\n'), ((952, 971), 'google.appengine.ext.db.StringProperty', 'db.StringProperty', ([], {}), '()\n', (969, 971), False, 'from google.appengine.ext import db\n'), ((983, 1003), 'google.appengine.ext.db.BooleanProperty', 'db.BooleanProperty', ([], {}), '()\n', (1001, 1003), False, 'from google.appengine.ext import db\n'), ((1013, 1032), 'google.appengine.ext.db.StringProperty', 'db.StringProperty', ([], {}), '()\n', (1030, 1032), False, 'from google.appengine.ext import db\n'), ((1044, 1063), 'google.appengine.ext.db.StringProperty', 'db.StringProperty', ([], {}), '()\n', (1061, 1063), False, 'from google.appengine.ext import db\n'), ((1076, 1095), 'google.appengine.ext.db.StringProperty', 'db.StringProperty', ([], {}), '()\n', (1093, 1095), False, 'from google.appengine.ext import db\n'), ((1113, 1132), 'google.appengine.ext.db.StringProperty', 'db.StringProperty', ([], {}), '()\n', (1130, 1132), False, 'from google.appengine.ext import db\n'), ((1149, 1167), 'google.appengine.ext.db.GeoPtProperty', 'db.GeoPtProperty', ([], {}), '()\n', (1165, 1167), False, 'from google.appengine.ext import db\n'), ((1182, 1202), 'google.appengine.ext.db.BooleanProperty', 'db.BooleanProperty', ([], {}), '()\n', (1200, 1202), False, 'from google.appengine.ext import db\n'), ((1219, 1239), 'google.appengine.ext.db.BooleanProperty', 'db.BooleanProperty', ([], {}), '()\n', (1237, 1239), False, 'from google.appengine.ext import db\n'), ((1254, 1274), 'google.appengine.ext.db.BooleanProperty', 'db.BooleanProperty', ([], {}), '()\n', (1272, 1274), False, 'from google.appengine.ext import db\n'), ((1283, 1300), 'google.appengine.ext.db.LinkProperty', 'db.LinkProperty', ([], {}), '()\n', (1298, 1300), False, 'from google.appengine.ext import db\n'), ((1315, 1353), 'google.appengine.ext.db.DateTimeProperty', 'db.DateTimeProperty', ([], {'auto_now_add': '(True)'}), '(auto_now_add=True)\n', (1334, 1353), False, 'from google.appengine.ext import db\n'), ((1400, 1420), 'google.appengine.ext.db.IntegerProperty', 'db.IntegerProperty', ([], {}), '()\n', (1418, 1420), False, 'from google.appengine.ext import db\n'), ((1435, 1455), 'google.appengine.ext.db.IntegerProperty', 'db.IntegerProperty', ([], {}), '()\n', (1453, 1455), False, 'from google.appengine.ext import db\n'), ((1468, 1487), 'google.appengine.ext.db.StringProperty', 'db.StringProperty', ([], {}), '()\n', (1485, 1487), False, 'from google.appengine.ext import db\n'), ((1502, 1540), 'google.appengine.ext.db.DateTimeProperty', 'db.DateTimeProperty', ([], {'auto_now_add': '(True)'}), '(auto_now_add=True)\n', (1521, 1540), False, 'from google.appengine.ext import db\n'), ((1552, 1572), 'google.appengine.ext.db.BooleanProperty', 'db.BooleanProperty', ([], {}), '()\n', (1570, 1572), False, 'from google.appengine.ext import db\n'), ((1584, 1604), 'google.appengine.ext.db.BooleanProperty', 'db.BooleanProperty', ([], {}), '()\n', (1602, 1604), False, 'from google.appengine.ext import db\n'), ((1614, 1633), 'google.appengine.ext.db.StringProperty', 'db.StringProperty', ([], {}), '()\n', (1631, 1633), False, 'from google.appengine.ext import db\n'), ((1645, 1664), 'google.appengine.ext.db.StringProperty', 'db.StringProperty', ([], {}), '()\n', (1662, 1664), False, 'from google.appengine.ext import db\n'), ((1677, 1696), 'google.appengine.ext.db.StringProperty', 'db.StringProperty', ([], {}), '()\n', (1694, 1696), False, 'from google.appengine.ext import db\n'), ((1714, 1733), 'google.appengine.ext.db.StringProperty', 'db.StringProperty', ([], {}), '()\n', (1731, 1733), False, 'from google.appengine.ext import db\n'), ((1750, 1768), 'google.appengine.ext.db.GeoPtProperty', 'db.GeoPtProperty', ([], {}), '()\n', (1766, 1768), False, 'from google.appengine.ext import db\n'), ((1867, 1935), 'google.appengine.ext.db.ReferenceProperty', 'db.ReferenceProperty', (['NameServer'], {'collection_name': '"""best_submissions"""'}), "(NameServer, collection_name='best_submissions')\n", (1887, 1935), False, 'from google.appengine.ext import db\n'), ((1957, 1975), 'google.appengine.ext.db.FloatProperty', 'db.FloatProperty', ([], {}), '()\n', (1973, 1975), False, 'from google.appengine.ext import db\n'), ((1999, 2070), 'google.appengine.ext.db.ReferenceProperty', 'db.ReferenceProperty', (['NameServer'], {'collection_name': '"""primary_submissions"""'}), "(NameServer, collection_name='primary_submissions')\n", (2019, 2070), False, 'from google.appengine.ext import db\n'), ((2121, 2179), 'google.appengine.ext.db.ReferenceProperty', 'db.ReferenceProperty', (['Submission'], {'collection_name': '"""config"""'}), "(Submission, collection_name='config')\n", (2141, 2179), False, 'from google.appengine.ext import db\n'), ((2199, 2218), 'google.appengine.ext.db.StringProperty', 'db.StringProperty', ([], {}), '()\n', (2216, 2218), False, 'from google.appengine.ext import db\n'), ((2246, 2266), 'google.appengine.ext.db.IntegerProperty', 'db.IntegerProperty', ([], {}), '()\n', (2264, 2266), False, 'from google.appengine.ext import db\n'), ((2291, 2311), 'google.appengine.ext.db.IntegerProperty', 'db.IntegerProperty', ([], {}), '()\n', (2309, 2311), False, 'from google.appengine.ext import db\n'), ((2331, 2349), 'google.appengine.ext.db.FloatProperty', 'db.FloatProperty', ([], {}), '()\n', (2347, 2349), False, 'from google.appengine.ext import db\n'), ((2362, 2380), 'google.appengine.ext.db.FloatProperty', 'db.FloatProperty', ([], {}), '()\n', (2378, 2380), False, 'from google.appengine.ext import db\n'), ((2397, 2417), 'google.appengine.ext.db.IntegerProperty', 'db.IntegerProperty', ([], {}), '()\n', (2415, 2417), False, 'from google.appengine.ext import db\n'), ((2432, 2452), 'google.appengine.ext.db.IntegerProperty', 'db.IntegerProperty', ([], {}), '()\n', (2450, 2452), False, 'from google.appengine.ext import db\n'), ((2466, 2485), 'google.appengine.ext.db.StringProperty', 'db.StringProperty', ([], {}), '()\n', (2483, 2485), False, 'from google.appengine.ext import db\n'), ((2498, 2517), 'google.appengine.ext.db.StringProperty', 'db.StringProperty', ([], {}), '()\n', (2515, 2517), False, 'from google.appengine.ext import db\n'), ((2574, 2637), 'google.appengine.ext.db.ReferenceProperty', 'db.ReferenceProperty', (['NameServer'], {'collection_name': '"""submissions"""'}), "(NameServer, collection_name='submissions')\n", (2594, 2637), False, 'from google.appengine.ext import db\n'), ((2653, 2716), 'google.appengine.ext.db.ReferenceProperty', 'db.ReferenceProperty', (['Submission'], {'collection_name': '"""nameservers"""'}), "(Submission, collection_name='nameservers')\n", (2673, 2716), False, 'from google.appengine.ext import db\n'), ((2738, 2758), 'google.appengine.ext.db.BooleanProperty', 'db.BooleanProperty', ([], {}), '()\n', (2756, 2758), False, 'from google.appengine.ext import db\n'), ((2775, 2795), 'google.appengine.ext.db.BooleanProperty', 'db.BooleanProperty', ([], {}), '()\n', (2793, 2795), False, 'from google.appengine.ext import db\n'), ((2813, 2833), 'google.appengine.ext.db.BooleanProperty', 'db.BooleanProperty', ([], {}), '()\n', (2831, 2833), False, 'from google.appengine.ext import db\n'), ((2854, 2872), 'google.appengine.ext.db.FloatProperty', 'db.FloatProperty', ([], {}), '()\n', (2870, 2872), False, 'from google.appengine.ext import db\n'), ((2891, 2909), 'google.appengine.ext.db.FloatProperty', 'db.FloatProperty', ([], {}), '()\n', (2907, 2909), False, 'from google.appengine.ext import db\n'), ((2923, 2945), 'google.appengine.ext.db.ListProperty', 'db.ListProperty', (['float'], {}), '(float)\n', (2938, 2945), False, 'from google.appengine.ext import db\n'), ((2963, 2981), 'google.appengine.ext.db.FloatProperty', 'db.FloatProperty', ([], {}), '()\n', (2979, 2981), False, 'from google.appengine.ext import db\n'), ((2999, 3017), 'google.appengine.ext.db.FloatProperty', 'db.FloatProperty', ([], {}), '()\n', (3015, 3017), False, 'from google.appengine.ext import db\n'), ((3034, 3054), 'google.appengine.ext.db.IntegerProperty', 'db.IntegerProperty', ([], {}), '()\n', (3052, 3054), False, 'from google.appengine.ext import db\n'), ((3073, 3093), 'google.appengine.ext.db.IntegerProperty', 'db.IntegerProperty', ([], {}), '()\n', (3091, 3093), False, 'from google.appengine.ext import db\n'), ((3107, 3127), 'google.appengine.ext.db.IntegerProperty', 'db.IntegerProperty', ([], {}), '()\n', (3125, 3127), False, 'from google.appengine.ext import db\n'), ((3141, 3161), 'google.appengine.ext.db.IntegerProperty', 'db.IntegerProperty', ([], {}), '()\n', (3159, 3161), False, 'from google.appengine.ext import db\n'), ((3179, 3199), 'google.appengine.ext.db.IntegerProperty', 'db.IntegerProperty', ([], {}), '()\n', (3197, 3199), False, 'from google.appengine.ext import db\n'), ((3212, 3231), 'google.appengine.ext.db.StringProperty', 'db.StringProperty', ([], {}), '()\n', (3229, 3231), False, 'from google.appengine.ext import db\n'), ((3245, 3265), 'google.appengine.ext.db.ListProperty', 'db.ListProperty', (['str'], {}), '(str)\n', (3260, 3265), False, 'from google.appengine.ext import db\n'), ((3342, 3360), 'google.appengine.ext.db.FloatProperty', 'db.FloatProperty', ([], {}), '()\n', (3358, 3360), False, 'from google.appengine.ext import db\n'), ((3370, 3388), 'google.appengine.ext.db.FloatProperty', 'db.FloatProperty', ([], {}), '()\n', (3386, 3388), False, 'from google.appengine.ext import db\n'), ((3399, 3419), 'google.appengine.ext.db.ListProperty', 'db.ListProperty', (['str'], {}), '(str)\n', (3414, 3419), False, 'from google.appengine.ext import db\n'), ((3438, 3457), 'google.appengine.ext.db.StringProperty', 'db.StringProperty', ([], {}), '()\n', (3455, 3457), False, 'from google.appengine.ext import db\n'), ((3596, 3665), 'google.appengine.ext.db.ReferenceProperty', 'db.ReferenceProperty', (['SubmissionNameServer'], {'collection_name': '"""results"""'}), "(SubmissionNameServer, collection_name='results')\n", (3616, 3665), False, 'from google.appengine.ext import db\n'), ((3681, 3701), 'google.appengine.ext.db.IntegerProperty', 'db.IntegerProperty', ([], {}), '()\n', (3699, 3701), False, 'from google.appengine.ext import db\n'), ((3716, 3738), 'google.appengine.ext.db.ListProperty', 'db.ListProperty', (['float'], {}), '(float)\n', (3731, 3738), False, 'from google.appengine.ext import db\n'), ((3757, 3777), 'google.appengine.ext.db.ListProperty', 'db.ListProperty', (['int'], {}), '(int)\n', (3772, 3777), False, 'from google.appengine.ext import db\n'), ((3910, 3985), 'google.appengine.ext.db.ReferenceProperty', 'db.ReferenceProperty', (['SubmissionNameServer'], {'collection_name': '"""index_results"""'}), "(SubmissionNameServer, collection_name='index_results')\n", (3930, 3985), False, 'from google.appengine.ext import db\n'), ((4001, 4059), 'google.appengine.ext.db.ReferenceProperty', 'db.ReferenceProperty', (['IndexHost'], {'collection_name': '"""results"""'}), "(IndexHost, collection_name='results')\n", (4021, 4059), False, 'from google.appengine.ext import db\n'), ((4073, 4091), 'google.appengine.ext.db.FloatProperty', 'db.FloatProperty', ([], {}), '()\n', (4089, 4091), False, 'from google.appengine.ext import db\n'), ((4109, 4129), 'google.appengine.ext.db.IntegerProperty', 'db.IntegerProperty', ([], {}), '()\n', (4127, 4129), False, 'from google.appengine.ext import db\n'), ((4138, 4158), 'google.appengine.ext.db.IntegerProperty', 'db.IntegerProperty', ([], {}), '()\n', (4156, 4158), False, 'from google.appengine.ext import db\n'), ((4172, 4191), 'google.appengine.ext.db.StringProperty', 'db.StringProperty', ([], {}), '()\n', (4189, 4191), False, 'from google.appengine.ext import db\n')] |
"""
Running operational space control with the PyGame display, using an exponential
additive signal when to push away from joints.
The target location can be moved by clicking on the background.
"""
import numpy as np
from abr_control.arms import threejoint as arm
# from abr_control.arms import twojoint as arm
from abr_control.interfaces import PyGame
from abr_control.controllers import OSC, AvoidJointLimits, Damping
print('\nClick to move the target.\n')
# initialize our robot config
robot_config = arm.Config(use_cython=True)
# create our arm simulation
arm_sim = arm.ArmSim(robot_config)
avoid = AvoidJointLimits(
robot_config,
min_joint_angles=[np.pi/5.0]*robot_config.N_JOINTS,
max_joint_angles=[np.pi/2.0]*robot_config.N_JOINTS,
max_torque=[100.0]*robot_config.N_JOINTS)
# damp the movements of the arm
damping = Damping(robot_config, kv=10)
# create an operational space controller
ctrlr = OSC(robot_config, kp=100, null_controllers=[avoid, damping],
# control (x, y) out of [x, y, z, alpha, beta, gamma]
ctrlr_dof=[True, True, False, False, False, False])
def on_click(self, mouse_x, mouse_y):
self.target[0] = self.mouse_x
self.target[1] = self.mouse_y
# create our interface
interface = PyGame(robot_config, arm_sim,
dt=.001, on_click=on_click,
q_init=[np.pi/4, np.pi/2, np.pi/2])
interface.connect()
# create a target [x, y, z]]
target_xyz = [0, 2, 0]
# create a target orientation [alpha, beta, gamma]
target_angles = [0, 0, 0]
interface.set_target(target_xyz)
try:
print('\nSimulation starting...\n')
count = 0
while 1:
# get arm feedback
feedback = interface.get_feedback()
hand_xyz = robot_config.Tx('EE', feedback['q'])
target = np.hstack([target_xyz, target_angles])
# generate an operational space control signal
u = ctrlr.generate(
q=feedback['q'],
dq=feedback['dq'],
target=target,
)
new_target_xy = interface.get_mousexy()
if new_target_xy is not None:
target_xyz[:2] = new_target_xy
interface.set_target(target_xyz)
# apply the control signal, step the sim forward
interface.send_forces(
u, update_display=True if count % 20 == 0 else False)
count += 1
finally:
# stop and reset the simulation
interface.disconnect()
print('Simulation terminated...')
| [
"numpy.hstack",
"abr_control.controllers.OSC",
"abr_control.interfaces.PyGame",
"abr_control.controllers.AvoidJointLimits",
"abr_control.arms.threejoint.Config",
"abr_control.controllers.Damping",
"abr_control.arms.threejoint.ArmSim"
] | [((509, 536), 'abr_control.arms.threejoint.Config', 'arm.Config', ([], {'use_cython': '(True)'}), '(use_cython=True)\n', (519, 536), True, 'from abr_control.arms import threejoint as arm\n'), ((575, 599), 'abr_control.arms.threejoint.ArmSim', 'arm.ArmSim', (['robot_config'], {}), '(robot_config)\n', (585, 599), True, 'from abr_control.arms import threejoint as arm\n'), ((609, 804), 'abr_control.controllers.AvoidJointLimits', 'AvoidJointLimits', (['robot_config'], {'min_joint_angles': '([np.pi / 5.0] * robot_config.N_JOINTS)', 'max_joint_angles': '([np.pi / 2.0] * robot_config.N_JOINTS)', 'max_torque': '([100.0] * robot_config.N_JOINTS)'}), '(robot_config, min_joint_angles=[np.pi / 5.0] *\n robot_config.N_JOINTS, max_joint_angles=[np.pi / 2.0] * robot_config.\n N_JOINTS, max_torque=[100.0] * robot_config.N_JOINTS)\n', (625, 804), False, 'from abr_control.controllers import OSC, AvoidJointLimits, Damping\n'), ((845, 873), 'abr_control.controllers.Damping', 'Damping', (['robot_config'], {'kv': '(10)'}), '(robot_config, kv=10)\n', (852, 873), False, 'from abr_control.controllers import OSC, AvoidJointLimits, Damping\n'), ((923, 1040), 'abr_control.controllers.OSC', 'OSC', (['robot_config'], {'kp': '(100)', 'null_controllers': '[avoid, damping]', 'ctrlr_dof': '[True, True, False, False, False, False]'}), '(robot_config, kp=100, null_controllers=[avoid, damping], ctrlr_dof=[\n True, True, False, False, False, False])\n', (926, 1040), False, 'from abr_control.controllers import OSC, AvoidJointLimits, Damping\n'), ((1258, 1363), 'abr_control.interfaces.PyGame', 'PyGame', (['robot_config', 'arm_sim'], {'dt': '(0.001)', 'on_click': 'on_click', 'q_init': '[np.pi / 4, np.pi / 2, np.pi / 2]'}), '(robot_config, arm_sim, dt=0.001, on_click=on_click, q_init=[np.pi / \n 4, np.pi / 2, np.pi / 2])\n', (1264, 1363), False, 'from abr_control.interfaces import PyGame\n'), ((1793, 1831), 'numpy.hstack', 'np.hstack', (['[target_xyz, target_angles]'], {}), '([target_xyz, target_angles])\n', (1802, 1831), True, 'import numpy as np\n')] |
import itertools
from unittest import TestCase
import numpy as np
from utils.data import ArrayInfo, image_array_to_rgb
from utils.data.mappers import *
class ImageUtilsTestCase(TestCase):
def test_image_array_to_rgb(self):
np.random.seed(1234)
def f(batch_size, n_channels, channel_last, the_channel_last,
use_info, bit_depth, dequantize, scale_to):
shape = [31, 32]
if n_channels is not None:
if the_channel_last:
shape = shape + [n_channels]
else:
shape = [n_channels] + shape
the_info = ArrayInfo(
shape=shape, min_val=0, max_val=255, is_discrete=True,
n_discrete_vals=256, bit_depth=8)
x = np.random.randint(0, 256, size=batch_size + shape)
mappers = []
ans_mappers = None
if bit_depth not in (None, 8):
mappers.append(ReduceToBitDepth(bit_depth))
ans_mappers = ReduceToBitDepth(bit_depth)
if dequantize:
mappers.append(Dequantize(epsilon=1e-5))
if scale_to:
mappers.append(ScaleToRange(*scale_to))
if mappers:
m = ArrayMapperList(mappers)
y_the_info = m.fit(the_info)
y = m.transform(x)
else:
y_the_info = the_info
y = x
ans = x
if ans_mappers is not None:
ans_mappers.fit(the_info)
ans = ans_mappers.transform(ans)
if n_channels is None:
ans = np.reshape(ans, ans.shape + (1,))
elif not the_channel_last:
ans = np.transpose(
ans,
list(range(len(ans.shape) - 3)) + [-2, -1, -3]
)
info = y_the_info if use_info else None
out = image_array_to_rgb(y, info, channel_last)
self.assertEqual(out.dtype, np.uint8)
np.testing.assert_equal(out, ans)
for (batch_size, n_channels, channel_last, the_channel_last,
use_info, bit_depth, dequantize, scale_to) in itertools.product(
([], [7], [3, 4]),
(None, 1, 3),
(None, True, False),
(True, False),
(True, False),
(8, 5),
(True, False),
(None, (0, 1), (-1, 1),),
):
# skip inconsistent parameter combination
if (n_channels is None and channel_last is not None) or \
(channel_last is not None and channel_last != the_channel_last):
continue
if n_channels is None and batch_size:
continue
# use_info = False is not supported along with dequantize or bit-depth
if not use_info and (dequantize or bit_depth != 8):
continue
f(batch_size, n_channels, channel_last, the_channel_last,
use_info, bit_depth, dequantize, scale_to)
| [
"numpy.reshape",
"numpy.testing.assert_equal",
"itertools.product",
"utils.data.ArrayInfo",
"utils.data.image_array_to_rgb",
"numpy.random.randint",
"numpy.random.seed"
] | [((240, 260), 'numpy.random.seed', 'np.random.seed', (['(1234)'], {}), '(1234)\n', (254, 260), True, 'import numpy as np\n'), ((2216, 2375), 'itertools.product', 'itertools.product', (['([], [7], [3, 4])', '(None, 1, 3)', '(None, True, False)', '(True, False)', '(True, False)', '(8, 5)', '(True, False)', '(None, (0, 1), (-1, 1))'], {}), '(([], [7], [3, 4]), (None, 1, 3), (None, True, False), (\n True, False), (True, False), (8, 5), (True, False), (None, (0, 1), (-1, 1))\n )\n', (2233, 2375), False, 'import itertools\n'), ((639, 741), 'utils.data.ArrayInfo', 'ArrayInfo', ([], {'shape': 'shape', 'min_val': '(0)', 'max_val': '(255)', 'is_discrete': '(True)', 'n_discrete_vals': '(256)', 'bit_depth': '(8)'}), '(shape=shape, min_val=0, max_val=255, is_discrete=True,\n n_discrete_vals=256, bit_depth=8)\n', (648, 741), False, 'from utils.data import ArrayInfo, image_array_to_rgb\n'), ((787, 837), 'numpy.random.randint', 'np.random.randint', (['(0)', '(256)'], {'size': '(batch_size + shape)'}), '(0, 256, size=batch_size + shape)\n', (804, 837), True, 'import numpy as np\n'), ((1949, 1990), 'utils.data.image_array_to_rgb', 'image_array_to_rgb', (['y', 'info', 'channel_last'], {}), '(y, info, channel_last)\n', (1967, 1990), False, 'from utils.data import ArrayInfo, image_array_to_rgb\n'), ((2053, 2086), 'numpy.testing.assert_equal', 'np.testing.assert_equal', (['out', 'ans'], {}), '(out, ans)\n', (2076, 2086), True, 'import numpy as np\n'), ((1659, 1692), 'numpy.reshape', 'np.reshape', (['ans', '(ans.shape + (1,))'], {}), '(ans, ans.shape + (1,))\n', (1669, 1692), True, 'import numpy as np\n')] |
# -*- coding: utf-8 -*-
# closeness-server (c) <NAME>
from closeness_server import create_app
app = create_app()
| [
"closeness_server.create_app"
] | [((101, 113), 'closeness_server.create_app', 'create_app', ([], {}), '()\n', (111, 113), False, 'from closeness_server import create_app\n')] |
from sqlalchemy import Table, Column, Integer, String
from sqlalchemy.orm import mapper
from .database import metadata, db_session
class User(object):
query = db_session.query_property()
def __init__(self, name=None, email=None):
self.name = name
self.email = email
def __repr__(self):
return '<User %r>' % (self.name)
users = Table(
'users', metadata,
Column('id', Integer, primary_key=True),
Column('name', String(50), unique=True),
Column('email', String(120), unique=True)
)
mapper(User, users)
| [
"sqlalchemy.orm.mapper",
"sqlalchemy.String",
"sqlalchemy.Column"
] | [((538, 557), 'sqlalchemy.orm.mapper', 'mapper', (['User', 'users'], {}), '(User, users)\n', (544, 557), False, 'from sqlalchemy.orm import mapper\n'), ((404, 443), 'sqlalchemy.Column', 'Column', (['"""id"""', 'Integer'], {'primary_key': '(True)'}), "('id', Integer, primary_key=True)\n", (410, 443), False, 'from sqlalchemy import Table, Column, Integer, String\n'), ((464, 474), 'sqlalchemy.String', 'String', (['(50)'], {}), '(50)\n', (470, 474), False, 'from sqlalchemy import Table, Column, Integer, String\n'), ((510, 521), 'sqlalchemy.String', 'String', (['(120)'], {}), '(120)\n', (516, 521), False, 'from sqlalchemy import Table, Column, Integer, String\n')] |
import dash_bootstrap_components as dbc
import dash_html_components as html
"""
Creation of jumbotrons for a better Homepage display
"""
left_jumbotron = dbc.Col(
html.Div(
[
html.H2("Your Favorite Finance Hub", className="display-3"),
html.Hr(className="my-2"),
html.P(
"Get the latest news from the news agencies "
"And track the live Stock & Crypto data. Free."
),
],
className="h-100 p-5 text-white bg-dark rounded-1",
),
md=12,
)
right_jumbotron = dbc.Col(
html.Div(
[
html.Hr(className="my-2"),
],
className="h-100 p-5 bg-light border rounded-3",
),
md=16,
)
jumbotron = dbc.Row(
[left_jumbotron],
className="align-items-md-stretch",
)
second_jumbotron = dbc.Row(children=[right_jumbotron],
className="align-items-md-stretch"
)
| [
"dash_bootstrap_components.Row",
"dash_html_components.P",
"dash_html_components.H2",
"dash_html_components.Hr"
] | [((744, 805), 'dash_bootstrap_components.Row', 'dbc.Row', (['[left_jumbotron]'], {'className': '"""align-items-md-stretch"""'}), "([left_jumbotron], className='align-items-md-stretch')\n", (751, 805), True, 'import dash_bootstrap_components as dbc\n'), ((836, 907), 'dash_bootstrap_components.Row', 'dbc.Row', ([], {'children': '[right_jumbotron]', 'className': '"""align-items-md-stretch"""'}), "(children=[right_jumbotron], className='align-items-md-stretch')\n", (843, 907), True, 'import dash_bootstrap_components as dbc\n'), ((200, 259), 'dash_html_components.H2', 'html.H2', (['"""Your Favorite Finance Hub"""'], {'className': '"""display-3"""'}), "('Your Favorite Finance Hub', className='display-3')\n", (207, 259), True, 'import dash_html_components as html\n'), ((273, 298), 'dash_html_components.Hr', 'html.Hr', ([], {'className': '"""my-2"""'}), "(className='my-2')\n", (280, 298), True, 'import dash_html_components as html\n'), ((312, 420), 'dash_html_components.P', 'html.P', (['"""Get the latest news from the news agencies And track the live Stock & Crypto data. Free."""'], {}), "(\n 'Get the latest news from the news agencies And track the live Stock & Crypto data. Free.'\n )\n", (318, 420), True, 'import dash_html_components as html\n'), ((616, 641), 'dash_html_components.Hr', 'html.Hr', ([], {'className': '"""my-2"""'}), "(className='my-2')\n", (623, 641), True, 'import dash_html_components as html\n')] |
#!/usr/bin/python
# -*- coding: UTF-8 -*-
from numpy import *
import heapq
class solutions:
def childs_with_sugers(self,childs,sugers):
"""
分糖给小朋友,一个小朋友只能拿一块糖,糖不能分割
:param childs:
:param sugers:
:return:
"""
childs=sorted(childs)
sugers=sorted(sugers)
res=[]
# if len(childs) > len(sugers) :
j=0
for i,child in enumerate(childs):
while j < len(sugers):
if sugers[j] >= child:
res.append([child,sugers[j]])
j+=1
break
j=j+1
else: # 正常结束 while 循环 则运行下面代码
print('suger is not enough') # 糖已经分配完了,有的小朋友没有糖吃
break # 跳出for 循环
return res
def regions_overlap(self,regions,L):
"""
区间覆盖问题
给定一个长度为 m的区间,再给出 n条线段的起点和终点(注意这里是闭区间),
求最少使用多少条线段可以将整个区间完全覆盖。
ref: https://www.cnblogs.com/acgoto/p/9824723.html
:param regions: [ [2,6],[1,4],[3,6],[3,7],[6,8],[2,4],[3,5] ]
:param L: 8
:return: [ [1,4] ,[3,7],[6,8] ]
"""
regions=sorted(regions, key=lambda d: d[0]) # 按照区间的 左端点进行排序
# print(regions)
right_most=1
res=[]
while right_most<L:
left_small=list(filter(lambda x: x[0] <= right_most, regions)) # 过滤出左端点 小于 right_most 的区间
right_max=max(left_small,key=lambda x:x[1]) # 选这些区间 中 右端点最大的一个
res.append(right_max)
right_most=right_max[1] #更新 已覆盖线段的 右端点
return res
def max_regions_not_intersect(self, regions, L):
"""
最多 不相交区间(活动选择问题)
假设我们有 n 个区间,区间的起始端点和结束端点分别是[l1, r1],[l2, r2],[l3, r3],……,[ln, rn]。我们从这 n 个区间中选出一部分区间,
这部分区间满足两两不相交(端点相交的情况不算相交),最多能选出多少个区间呢?
ref: https://time.geekbang.org/column/article/73188
:param regions: [[6,8],[2,4],[3,5],[1,5],[5,9],[8,10]]
:param L: 10
:return: [[2,4],[6,8],[8,10]]
"""
regions = sorted(regions, key=lambda d: d[0]) # 按照区间的 左端点进行排序
# print(regions)
right_most=0
res=[]
while right_most<L:
left_small=list(filter(lambda x: x[0] >= right_most, regions)) # 过滤出左端点 大于 right_most 的区间 ,这样能避免重合
right_max=min(left_small,key=lambda x:x[1]) # 选这些区间 中 右端点最小的一个,这样 能留出更多的剩余空间
res.append(right_max)
right_most=right_max[1] #更新 已覆盖线段的 右端点
return res
def activity_selection(self, regions):
"""
活动选择问题
假设我们有 n 个活动,活动的 开始端点和结束时间 分别是[l1, r1],[l2, r2],[l3, r3],……,[ln, rn]
我们从这 n 个 活动中选出一部分活动,
这部分活动的时间不冲突,即 满足 两两不相交(端点相交的情况不算相交),最多能选出多少个活动呢?
贪心策略: 结束时间 早的 活动 优先
:param regions: [[6,8],[2,4],[3,5],[1,5],[5,9],[8,10]]
:return: [[2,4],[6,8],[8,10]]
"""
regions = sorted(regions, key=lambda d: d[1]) # 按照 结束时间 对活动 进行排序
selected=[] #被选中的活动
next_earliest_start=0 # 保证与上一个活动不冲突的 情况下,下一个活动的最早开始时间
for region in regions:
start=region[0]
end=region[1]
if start >= next_earliest_start: # 活动的开始时间 满足 要求
selected.append(region)
next_earliest_start=end
return selected
def minimum_delay_scheduling(self, duration, deadline):
"""
最小延迟 调度 问题
n 项任务,每一项任务 消耗的时间为 duration,每一项任务的 截止时间为 deadline
任务超过截止时间才完成会产生延迟
求 所有任务中 发生最大延迟的任务 所产生的延迟 达到最小的 调度策略,在这一策略下的 最大延迟
贪心策略: 截止时间 早的任务优先
:param duration: [5,8,4,10,3]
:param deadline: [10,12,15,11,20]
:return:
"""
N=len(duration) # 任务个数
deadline=array(deadline)
jobs_deadline= [ (i,deadline[i]) for i in range(N)] # (任务标号, 任务截止时间)
jobs_deadline = sorted(jobs_deadline, key=lambda ele: ele[1])
jobs_delay = zeros(N, dtype=int) # 记录 各个任务的 延迟时间
current_time=0 # 记录 当前的时间点
for job in jobs_deadline:
job_NO=job[0]
job_deadline=job[1]
current_time+=duration[job_NO]
if current_time > job_deadline: # 时间超过了 deadline
jobs_delay[job_NO]= current_time-job_deadline
print('jobs_delay:',jobs_delay)
max_delay=max(jobs_delay)
return max_delay
def bulk_bag_problem(self, weights,values, capacity):
"""
散装背包问题 (背包问题变形)
与普通背包问题 的关键区别:每件物品均可分解
一个可以容纳 100kg 物品的背包, 有 5 种豆子,每种豆子的 总量 和 总价值都各不相同。
为了让背包中所装物品的总价值最大,我们如何选择在背包中装哪些豆子?每种豆子又该装多少呢
采用贪心策略:尽可能多放 单位重量价值 最大的 物品
:param weights: [100,30,60,20,50]
:param values: [100,90,120,80,75]
:param capacity: 100
:return:
"""
N=len(weights) # 物品种类
weights=array(weights)
values=array(values)
unit_weight_value= values/weights
unit_weight_value= [ (i,unit_weight_value[i]) for i in range(N)] # (标号,物品的 单位重量的价值)
unit_weight_value= sorted(unit_weight_value,key=lambda ele:ele[1],reverse=True)
#按照 物品单位重量的价值 逆序排序 [(3, 4.0), (1, 3.0), (2, 2.0), (4, 1.5), (0, 1.0)]
bag_weight=capacity
bag_value=0
bag_items= zeros(N, dtype=float) # 背包中 每一样物品的 重量
for item in unit_weight_value:
if bag_weight <=0:
break
item_NO=item[0]
if bag_weight >= weights[item_NO]: # 背包 容量 足以装下全部的 item_NO 物品
bag_weight -= weights[item_NO] #
bag_value += values[item_NO]
bag_items[item_NO]=weights[item_NO]
else: # 背包 容量 不足
bag_items[item_NO]=bag_weight # 剩下的背包容量 全部装 item_NO 物品
bag_weight = 0 #
bag_value += bag_weight*item[1]
return bag_value,bag_items
def optimal_loading(self, weights, capacity):
"""
最优装载问题 (背包问题变形)
即 物品的价值都为 1 的 01 背包问题
n 个集装箱 重量为 weights, 船的载重能力限制为 capacity, 每一个集装箱 都小于 capacity,
如何 选择 使得 可以装载更多 的集装箱
采用贪心策略:轻的物品优先装入
:param weights: [100,30,60,20,50]
:param capacity: 100
:return:
"""
N=len(weights) # 物品种类
weights=array(weights)
items_weight= [ (i,weights[i]) for i in range(N)] # (标号,物品的重量)
items_weight = sorted(items_weight, key=lambda ele: ele[1])
bag_weight=capacity
bag_items= [] # 记录 放入背包的物品
for item in items_weight:
if bag_weight <=0:
break
item_NO=item[0]
if bag_weight >= item[1] : # 背包 容量 能装下 item_NO 物品
bag_weight -= weights[item_NO]
bag_items.append(item_NO)
return bag_items
class ComapreHeap(object):
def __init__(self, initial=None, key=lambda x: x):
self.key = key
if initial:
self._data = [(key(item), item) for item in initial]
heapq.heapify(self._data)
else:
self._data = []
def push(self, item):
heapq.heappush(self._data, (self.key(item), item))
def pop(self):
return heapq.heappop(self._data)[1]
class TreeNode(object):
def __init__(self,key=None,value=None):
self.key=key
self.value=value
self.left=None
self.right=None
class huffman_tree:
"""
霍夫曼 前缀编码
使用 {0,1} 按照字符出现的 频率,并根据贪心策略 生成二叉编码树对字符集进行前缀编码
ref:
(1)《算法导论》
(2)https://time.geekbang.org/column/article/73188
"""
def __init__(self, char_list):
self.huffman_encode_tree=self.__encode(char_list)
self.__decode_all()
def encode(self, string):
"""
将字符串 编码为 以 {0,1} 表示的 字节流
:param string: 'acbf'
:return: '01001011100'
"""
bytes=[]
for char in string:
bytes.append(self.char_dict[char])
return ''.join(bytes)
def decode(self,bytes):
"""
解码一段 以 {0,1} 表示的 字节流
:param bytes: '01001011100'
:return: 'acbf'
"""
root = self.huffman_encode_tree
p=root
res_string=[]
for byte in bytes:
if byte=='0': # 走左子树
p=p.left
else: # 走右子树
p=p.right
#TODO : 解码失败的处理
if p.left==None and p.right==None: # 走到叶子节点了
res_string.append(p.key)
p=root
return ''.join(res_string)
def __decode_all(self):
"""
返回 霍夫曼 编码树 上 char_list 中所有字符 和其 对应的 编码
:return: char_dict= {'a': '0', 'c': '100', 'b': '101', 'f': '1100', 'e': '1101', 'd': '111'}
"""
root=self.huffman_encode_tree
self.char_dict={}
self.__tree_pre_order(root,[])
for char,char_bytes in self.char_dict.items():
self.char_dict[char]=''.join(str(e) for e in char_bytes)
return self.char_dict
def __tree_pre_order(self,root,pre_list):
if root.left== None and root.right == None: # 说明到达叶子节点
self.char_dict[root.key]=pre_list
else:
if root.left !=None:
self.__tree_pre_order(root.left,pre_list+[0])
if root.right!=None:
self.__tree_pre_order(root.right, pre_list + [1])
def __encode(self,char_list):
"""
将 char_list 中的字符,生成一颗 huffman 编码树
:param char_list: [('a',45),('b',13),('c',12),('d',16),('e',9),('f',5)]
:return:
"""
leaf_nodes= [TreeNode(ele[0] ,ele[1]) for ele in char_list ]
heap_nodes=ComapreHeap(leaf_nodes,key=lambda x:x.value)
N=len(char_list)
root_node=None
for i in range(N-1): # N 为叶节点个数,要执行N-1次的 叶节点的合并操作
root_node=TreeNode()
left_node=heap_nodes.pop()
right_node=heap_nodes.pop()
root_node.key='s'+str(i) # 非叶子节点的 Key
root_node.value=left_node.value+ right_node.value
root_node.left=left_node
root_node.right=right_node
# print('root:',root_node.value)
# print('root.left:', root_node.left.value)
# print('root.right:', root_node.right.value)
heap_nodes.push(root_node)
return root_node
if __name__ == '__main__':
sol = solutions()
childs=[3,4,5,6,7,8] # 小孩0-小孩5 想要的糖果的 重量
sugers=[1,2,3,4,5] # 现有的各个糖果的 重量
# print(sol.childs_with_sugers(childs,sugers))
regions=[ [2,6],[1,4],[3,6],[3,7],[6,8],[2,4],[3,5] ]
L=8
# print(sol.regions_overlap(regions,L))
regions=[[6,8],[2,4],[3,5],[1,5],[5,9],[8,10]]
L=10
# print(sol.max_regions_not_intersect(regions, L))
# print(sol.activity_selection(regions))
duration= [5, 8, 4, 10, 3]
deadline= [10, 12, 15, 11, 20]
# print(sol.minimum_delay_scheduling(duration,deadline))
weights= [100, 30, 60, 20, 50]
values= [100, 90, 120, 80, 75]
capacity= 100
# print(sol.bulk_bag_problem(weights,values,capacity))
# print(sol.optimal_loading(weights,capacity))
char_list=[('a',45),('b',13),('c',12),('d',16),('e',9),('f',5)]
huffman_tree=huffman_tree(char_list)
print('char_dict:',huffman_tree.char_dict)
bytes= '01001011100'
print(huffman_tree.decode(bytes))
print(huffman_tree.encode('acbf'))
| [
"heapq.heappop",
"heapq.heapify"
] | [((7105, 7130), 'heapq.heapify', 'heapq.heapify', (['self._data'], {}), '(self._data)\n', (7118, 7130), False, 'import heapq\n'), ((7294, 7319), 'heapq.heappop', 'heapq.heappop', (['self._data'], {}), '(self._data)\n', (7307, 7319), False, 'import heapq\n')] |
# -*- coding: utf-8 -*-
# flake8: noqa
# noreorder
"""
Pytube: a very serious Python library for downloading YouTube Videos.
"""
__title__ = 'my_pytube'
__version__ = '9.5.2'
__author__ = '<NAME>'
__license__ = 'MIT License'
__copyright__ = 'Copyright 2019 <NAME>'
#import logging
#import query
#import streams
#import captions
#import contrib
#import __main__
from my_pytube.logging import create_logger
from my_pytube.query import CaptionQuery
from my_pytube.query import StreamQuery
from my_pytube.streams import Stream
from my_pytube.captions import Caption
from my_pytube.contrib.playlist import Playlist
from my_pytube.__main__ import YouTube
logger = create_logger()
logger.info('%s v%s', __title__, __version__)
| [
"my_pytube.logging.create_logger"
] | [((663, 678), 'my_pytube.logging.create_logger', 'create_logger', ([], {}), '()\n', (676, 678), False, 'from my_pytube.logging import create_logger\n')] |
## @ingroup Methods-Aerodynamics-Airfoil_Panel_Method
# panel_geometry.py
# Created: Mar 2021, <NAME>
# ---------------------------------------
#-------------------------------
# Imports
# ----------------------------------------------------------------------
import SUAVE
from SUAVE.Core import Units
import numpy as np
# ----------------------------------------------------------------------
# panel_geometry.py
# ----------------------------------------------------------------------
## @ingroup Methods-Aerodynamics-Airfoil_Panel_Method
def panel_geometry(x,y,npanel,nalpha,nRe):
"""Computes airfoil surface panelization parameters for later use in
the computation of the matrix of influence coefficients.
Assumptions:
None
Source:
None
Inputs:
x - Vector of x coordinates of the surface nodes [unitless]
y - Vector of y coordinates of the surface nodes [unitless]
npanel - Number of panels on the airfoil [unitless]
Outputs:
l - Panel lengths [unitless]
st - np.sin(theta) for each panel [radians]
ct - np.cos(theta) for each panel [radians]
xbar - x-coordinate of the midpoint of each panel [unitless]
ybar - y-coordinate of the midpoint of each panel [unitless]
Properties Used:
N/A
"""
# compute various geometrical quantities
l = np.sqrt((x[1:] -x[:-1])**2 +(y[1:] -y[:-1])**2)
st = (y[1:] -y[:-1])/l
ct = (x[1:] -x[:-1])/l
xbar = (x[1:] +x[:-1])/2
ybar = (y[1:] +y[:-1])/2
norm = np.zeros((npanel,2,nalpha,nRe))
norm[:,0,:,:] = -st
norm[:,1,:,:] = ct
return l,st,ct,xbar,ybar,norm
| [
"numpy.zeros",
"numpy.sqrt"
] | [((1778, 1832), 'numpy.sqrt', 'np.sqrt', (['((x[1:] - x[:-1]) ** 2 + (y[1:] - y[:-1]) ** 2)'], {}), '((x[1:] - x[:-1]) ** 2 + (y[1:] - y[:-1]) ** 2)\n', (1785, 1832), True, 'import numpy as np\n'), ((1962, 1996), 'numpy.zeros', 'np.zeros', (['(npanel, 2, nalpha, nRe)'], {}), '((npanel, 2, nalpha, nRe))\n', (1970, 1996), True, 'import numpy as np\n')] |
#!/usr/bin/python
"""
(dummy-)Mutates multiple residues on a PDB-formatted structure.
HADDOCK will then reconstruct the residue according to its topology.
Usage: python pdb_multimutate.py pdbFL <mutation list file>
The format of mutation list:
chain resi resn_wt resn_mut
Example: python pdb_multimutate.py 3mrb.pdb list_mutations
for mutating residue 7 Serine of chain A to Alanine:
A 7 SER ALA
Author: {0}
Email: {1}
"""
from __future__ import print_function
import os
import sys
import re
__author__ = "<NAME>; <NAME>; <NAME>"
__email__ = "<EMAIL>; <EMAIL>; <EMAIL>"
USAGE = __doc__.format(__author__, __email__)
def check_input(args):
"""Checks whether to read from stdin/file and validates user input/options."""
if len(args) == 2:
if not os.path.isfile(args[0]):
sys.stderr.write('File not found: ' + args[0] + '\n')
sys.stderr.write(USAGE)
sys.exit(1)
else:
sys.stderr.write(USAGE)
sys.exit(1)
def mutate(structure_fhandle, chain, resi, resn_wt, resn_mut):
mutated_structure = []
flag = 0 # 0: residue-to-be-mutated NOT found in structure; 1 found.
atom_set = set(["CA", "C", "O", "N", "CB"]) # keep main chain atoms and CB atom of side chain
for line in structure_fhandle:
if line[0:4] == 'ATOM' or line[0:6] == 'HETATM' or line[0:6] == 'ANISOU':
s_chain = line[21].strip()
s_resi = line[22:26].strip()
s_resn = line[17:20].strip()
s_atom = line[12:16].strip()
if s_chain == chain and s_resi == resi and s_resn == resn_wt:
flag = 1
if s_atom in atom_set:
line = line[0:17]+resn_mut+line[20:]
else:
continue
mutated_structure.append(line)
if flag ==0:
sys.stderr.write('WARNING: ' + chain + ":" + resi +":" + resn_wt + ' does NOT exist in the strcuture\n')
return mutated_structure
def _print_mutants(pdbFL, mutationFL):
# read pdb file into memory
f_pdb = open(pdbFL, 'r')
structure = [l for l in f_pdb]
f_pdb.close()
new_pdbFL = os.path.splitext(pdbFL)[0] + '_mutated.pdb'
# mutate pdb file
print("Generated mutant files:")
f_mut = open(mutationFL, 'r')
for line in f_mut:
line = re.sub('[\n\r]','', line)
i = line.split()
if len(i) == 4:
chain, resi, resn_wt, resn_mut = i
print (i)
else:
sys.stderr.write('WARNING: Unrecognized mutation format in line "{0}"\n'.format(" ".join(i)))
continue
structure = mutate(structure, chain, resi, resn_wt, resn_mut)
if structure:
m_file = open(new_pdbFL, 'w')
print(new_pdbFL + ' generated')
m_file.write(''.join(structure))
m_file.close()
if __name__ == "__main__":
check_input(sys.argv[1:])
pdbFL = sys.argv[1]
mutationFL = sys.argv[2]
_print_mutants(pdbFL, mutationFL)
| [
"os.path.splitext",
"os.path.isfile",
"sys.stderr.write",
"sys.exit",
"re.sub"
] | [((970, 993), 'sys.stderr.write', 'sys.stderr.write', (['USAGE'], {}), '(USAGE)\n', (986, 993), False, 'import sys\n'), ((1002, 1013), 'sys.exit', 'sys.exit', (['(1)'], {}), '(1)\n', (1010, 1013), False, 'import sys\n'), ((1869, 1981), 'sys.stderr.write', 'sys.stderr.write', (["('WARNING: ' + chain + ':' + resi + ':' + resn_wt +\n ' does NOT exist in the strcuture\\n')"], {}), '(\'WARNING: \' + chain + \':\' + resi + \':\' + resn_wt +\n """ does NOT exist in the strcuture\n""")\n', (1885, 1981), False, 'import sys\n'), ((2352, 2378), 're.sub', 're.sub', (["'[\\n\\r]'", '""""""', 'line'], {}), "('[\\n\\r]', '', line)\n", (2358, 2378), False, 'import re\n'), ((801, 824), 'os.path.isfile', 'os.path.isfile', (['args[0]'], {}), '(args[0])\n', (815, 824), False, 'import os\n'), ((838, 891), 'sys.stderr.write', 'sys.stderr.write', (["('File not found: ' + args[0] + '\\n')"], {}), "('File not found: ' + args[0] + '\\n')\n", (854, 891), False, 'import sys\n'), ((904, 927), 'sys.stderr.write', 'sys.stderr.write', (['USAGE'], {}), '(USAGE)\n', (920, 927), False, 'import sys\n'), ((940, 951), 'sys.exit', 'sys.exit', (['(1)'], {}), '(1)\n', (948, 951), False, 'import sys\n'), ((2176, 2199), 'os.path.splitext', 'os.path.splitext', (['pdbFL'], {}), '(pdbFL)\n', (2192, 2199), False, 'import os\n')] |
import marshal
from multiprocessing import Condition, Process, Queue, Pipe
import os
from threading import Timer
from types import FunctionType
import pickle
from celery import Celery
from flask import Flask, url_for
from grams.grams import Histogram
from grams.markov import MC
import time
def make_app():
def make_model():
def _make_model(corpus, n_sentences=10):
# global cv
def _generate():
markovchain = MC(corpus)
return markovchain.generate
generate = _generate()
child_conn.send(generate(n_sentences))
while True:
if parent_conn.poll():
## previously sent message got consumed
# send another
child_conn.send(generate(n_sentences))
parent_conn, child_conn = Pipe(duplex=True)
with open("res/the_adventures_of_sherlock_holmes.txt", "r") as f:
f_out = f.read()
make_process = Process(target=_make_model, args=(f_out,))
make_process.start()
return parent_conn, make_process
# init app
flask_app = Flask(__name__)
parent_conn, make_process = make_model()
@flask_app.route("/")
def home():
if parent_conn.poll():
return parent_conn.recv()
return "loading..."
return flask_app
flask_app = make_app()
if __name__ == "__main__":
flask_app.run(debug=True, port=8080)
| [
"multiprocessing.Process",
"multiprocessing.Pipe",
"grams.markov.MC",
"flask.Flask"
] | [((1155, 1170), 'flask.Flask', 'Flask', (['__name__'], {}), '(__name__)\n', (1160, 1170), False, 'from flask import Flask, url_for\n'), ((864, 881), 'multiprocessing.Pipe', 'Pipe', ([], {'duplex': '(True)'}), '(duplex=True)\n', (868, 881), False, 'from multiprocessing import Condition, Process, Queue, Pipe\n'), ((1009, 1051), 'multiprocessing.Process', 'Process', ([], {'target': '_make_model', 'args': '(f_out,)'}), '(target=_make_model, args=(f_out,))\n', (1016, 1051), False, 'from multiprocessing import Condition, Process, Queue, Pipe\n'), ((468, 478), 'grams.markov.MC', 'MC', (['corpus'], {}), '(corpus)\n', (470, 478), False, 'from grams.markov import MC\n')] |
import abc
import numbers
from typing import Union
import numpy as np
from river import base, optim, utils
VectorLike = Union[utils.VectorDict, np.ndarray]
__all__ = ["Initializer", "Scheduler", "Optimizer", "Loss"]
class Initializer(base.Base, abc.ABC):
"""An initializer is used to set initial weights in a model."""
@abc.abstractmethod
def __call__(self, shape=1):
"""Returns a fresh set of weights.
Parameters
----------
shape
Indicates how many weights to return. If `1`, then a single scalar value will be
returned.
"""
class Scheduler(base.Base, abc.ABC):
"""Can be used to program the learning rate schedule of an `optim.base.Optimizer`."""
@abc.abstractmethod
def get(self, t: int) -> float:
"""Returns the learning rate at a given iteration.
Parameters
----------
t
The iteration number.
"""
def __repr__(self):
return f"{self.__class__.__name__}({vars(self)})"
class Optimizer(base.Base):
"""Optimizer interface.
Every optimizer inherits from this base interface.
Parameters
----------
lr
Attributes
----------
learning_rate : float
Returns the current learning rate value.
"""
def __init__(self, lr: Union[Scheduler, float]):
if isinstance(lr, numbers.Number):
lr = optim.schedulers.Constant(lr)
self.lr = lr
self.n_iterations = 0
@property
def learning_rate(self) -> float:
return self.lr.get(self.n_iterations)
def look_ahead(self, w: dict) -> dict:
"""Updates a weight vector before a prediction is made.
Parameters:
w (dict): A dictionary of weight parameters. The weights are modified in-place.
Returns:
The updated weights.
"""
return w
def _step_with_dict(self, w: dict, g: dict) -> dict:
raise NotImplementedError
def _step_with_vector(self, w: VectorLike, g: VectorLike) -> VectorLike:
raise NotImplementedError
def step(
self, w: Union[dict, VectorLike], g: Union[dict, VectorLike]
) -> Union[dict, VectorLike]:
"""Updates a weight vector given a gradient.
Parameters
----------
w
A vector-like object containing weights. The weights are modified in-place.
g
A vector-like object of gradients.
Returns
-------
The updated weights.
"""
if isinstance(w, VectorLike.__args__) and isinstance(g, VectorLike.__args__):
try:
w = self._step_with_vector(w, g)
self.n_iterations += 1
return w
except NotImplementedError:
pass
w = self._step_with_dict(w, g)
self.n_iterations += 1
return w
def __repr__(self):
return f"{self.__class__.__name__}({vars(self)})"
class Loss(base.Base, abc.ABC):
"""Base class for all loss functions."""
def __repr__(self):
return f"{self.__class__.__name__}({vars(self)})"
@abc.abstractmethod
def __call__(self, y_true, y_pred):
"""Returns the loss.
Parameters
----------
y_true
Ground truth(s).
y_pred
Prediction(s).
Returns
-------
The loss(es).
"""
@abc.abstractmethod
def gradient(self, y_true, y_pred):
"""Return the gradient with respect to y_pred.
Parameters
----------
y_true
Ground truth(s).
y_pred
Prediction(s).
Returns
-------
The gradient(s).
"""
@abc.abstractmethod
def mean_func(self, y_pred):
"""Mean function.
This is the inverse of the link function. Typically, a loss function takes as input the raw
output of a model. In the case of classification, the raw output would be logits. The mean
function can be used to convert the raw output into a value that makes sense to the user,
such as a probability.
Parameters
----------
y_pred
Raw prediction(s).
Returns
-------
The adjusted prediction(s).
References
----------
[^1]: [Wikipedia section on link and mean function](https://www.wikiwand.com/en/Generalized_linear_model#/Link_function)
"""
| [
"river.optim.schedulers.Constant"
] | [((1421, 1450), 'river.optim.schedulers.Constant', 'optim.schedulers.Constant', (['lr'], {}), '(lr)\n', (1446, 1450), False, 'from river import base, optim, utils\n')] |
#!/usr/bin/env python3
"""Quantize and convert a keras model to tensorflowjs model."""
import argparse
import pathlib
import tempfile
import numpy as np
import tensorflow as tf
import tensorflowjs as tfjs
def main() -> None:
"""Execute the main routine."""
##
# Parse command-line arguments
##
parser = argparse.ArgumentParser(description=__doc__)
parser.add_argument(
"--keras_path",
help="Path to the keras model",
required=True)
parser.add_argument(
"--tfjs_dir",
help="Path to the tensorflowjs model",
required=True)
args = parser.parse_args()
keras_path = pathlib.Path(args.keras_path)
tfjs_dir = pathlib.Path(args.tfjs_dir)
##
# Check command-line arguments
##
if not keras_path.exists():
raise FileNotFoundError("Keras model does not exist: {}".format(
keras_path))
if not keras_path.is_file():
raise RuntimeError("Expected keras model to be a file: {}".format(
keras_path))
if tfjs_dir.exists() and not tfjs_dir.is_dir():
raise NotADirectoryError(
"Expected tfjs model path to be a directory: {}".format(
tfjs_dir))
##
# Prepare
##
tfjs_dir.mkdir(exist_ok=True, parents=True)
##
# Quantize and convert
#
# (paraphrased from tensorflowjs/converters/converter.py
##
model = tf.keras.models.load_model(str(keras_path))
with tempfile.TemporaryDirectory(
prefix='savedmodel', dir=str(tfjs_dir)) as temp_savedmodel_dir:
tf.keras.experimental.export_saved_model(
model, str(temp_savedmodel_dir), serving_only=True)
quantization_dtype = np.uint8 # quantize to 1-byte
skip_op_check = False # default
strip_debug_ops = False # default
tfjs.converters.tf_saved_model_conversion_v2.convert_tf_saved_model(
str(temp_savedmodel_dir),
str(tfjs_dir),
signature_def='serving_default',
saved_model_tags='serve',
quantization_dtype=quantization_dtype,
skip_op_check=skip_op_check,
strip_debug_ops=strip_debug_ops)
if __name__ == "__main__":
main()
| [
"argparse.ArgumentParser",
"pathlib.Path"
] | [((329, 373), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {'description': '__doc__'}), '(description=__doc__)\n', (352, 373), False, 'import argparse\n'), ((654, 683), 'pathlib.Path', 'pathlib.Path', (['args.keras_path'], {}), '(args.keras_path)\n', (666, 683), False, 'import pathlib\n'), ((699, 726), 'pathlib.Path', 'pathlib.Path', (['args.tfjs_dir'], {}), '(args.tfjs_dir)\n', (711, 726), False, 'import pathlib\n')] |
from flask import render_template_string
from werkzeug.test import EnvironBuilder, run_wsgi_app
class TestTemplateGlobals:
def test_static_mode(self, single_use_app):
@single_use_app.route("/test-globals")
def test_globals():
return render_template_string("""{{ static_mode }}""")
client = single_use_app.test_client()
single_use_app.config["STATIC_MODE"] = False
assert client.get("/test-globals").get_data(as_text=True) == "False"
single_use_app.config["STATIC_MODE"] = True
assert client.get("/test-globals").get_data(as_text=True) == "True"
assert client.get("/test-globals?static_mode=no").get_data(as_text=True) == "False"
assert client.get("/test-globals?static_mode=yes").get_data(as_text=True) == "True"
class TestHeaders:
def test_http_headers(self, single_use_app):
builder = EnvironBuilder(path="/", method="GET")
env = builder.get_environ()
(app_iter, status, headers) = run_wsgi_app(single_use_app, env)
assert headers.get("X-Frame-Options") == "deny"
assert headers.get("X-Content-Type-Options") == "nosniff"
assert headers.get("X-XSS-Protection") == "1; mode=block"
assert headers.get("X-Permitted-Cross-Domain-Policies") == "none"
assert headers.get("Content-Security-Policy")
| [
"flask.render_template_string",
"werkzeug.test.EnvironBuilder",
"werkzeug.test.run_wsgi_app"
] | [((895, 933), 'werkzeug.test.EnvironBuilder', 'EnvironBuilder', ([], {'path': '"""/"""', 'method': '"""GET"""'}), "(path='/', method='GET')\n", (909, 933), False, 'from werkzeug.test import EnvironBuilder, run_wsgi_app\n'), ((1009, 1042), 'werkzeug.test.run_wsgi_app', 'run_wsgi_app', (['single_use_app', 'env'], {}), '(single_use_app, env)\n', (1021, 1042), False, 'from werkzeug.test import EnvironBuilder, run_wsgi_app\n'), ((267, 310), 'flask.render_template_string', 'render_template_string', (['"""{{ static_mode }}"""'], {}), "('{{ static_mode }}')\n", (289, 310), False, 'from flask import render_template_string\n')] |
#!/usr/bin/env python
########################################################################
# RSA2ELK, by <NAME>
# Converts Netwitness log parser configuration to Logstash configuration
# see https://github.com/blookot/rsa2elk
########################################################################
import config
import re
import csv
import sys
# carriage return for logstash conf, just \n for linux
CR = "\n"
# insert n times a tab
def t(n):
t=""
for i in range(0,n):
t=t+"\t"
return t
# Replace all RSA date syntax with LS date filter one
def convertDate(s):
dateList = set()
pattern = re.compile(",'([^']+)'")
for dStr in pattern.finditer(s):
c = dStr.group(1)
# first, replace all non coding (not having a %) chars, like T, cf https://www.elastic.co/guide/en/logstash/current/plugins-filters-date.html#plugins-filters-date-match
c = re.sub("([^%])([a-zA-Z]+)", r"\1'\2'", c)
# replace the specific chars by their logstash date filter equivalent
c = c.replace("%C", "M/d/yy H:m:s")
c = c.replace("%R", "MMMM")
c = c.replace("%B", "MMM")
c = c.replace("%M", "MM")
c = c.replace("%G", "M")
c = c.replace("%D", "dd")
c = c.replace("%F", "d")
c = c.replace("%H", "HH")
c = c.replace("%I", "HH")
c = c.replace("%N", "H")
c = c.replace("%T", "mm")
c = c.replace("%U", "m")
c = c.replace("%J", "D")
c = c.replace("%P", "a")
c = c.replace("%S", "ss")
c = c.replace("%O", "s")
c = c.replace("%Y", "yy")
c = c.replace("%W", "yyyy")
c = c.replace("%Z", "H:m:s")
c = c.replace("%A", "D")
c = c.replace("%Q", "a") # AM/PM
c = c.replace("%K", "") # undocumented... seen in cef.xml
c = c.replace("%L", "") # undocumented... seen in cef.xml
c = c.replace("%E", "") # undocumented... seen in v20_trendmicromsg.xml
c = c.replace("%X", "UNIX")
if '%' in c: print("Missing a condition in date conversion: " + c)
dateList.add(c)
return "\"" + "\", \"".join(dateList) + "\""
# converting STRCAT
def convertStrcat(s):
c = ""
regex = re.compile("^[a-zA-Z]+.*") # just saying the string starts with a letter
# grab first (
iFirstPar = s.find("(")
iEndPar = s.find(")", iFirstPar)
if iFirstPar > 0 and iEndPar > 0:
idParts = str.split(s[iFirstPar+1:iEndPar], ",")
# check for static strings vs fields
for idPart in idParts:
if "'" in idPart:
# catenate a string
c = c + str.strip(idPart.replace("'", ""))
elif "\"" in idPart:
# catenate a string
c = c + str.strip(idPart.r("\"", ""))
elif regex.match(idPart.strip()):
# first character is [a-z] ie a field
c = c + "%{" + idPart.strip() + "}"
elif idPart == "\t":
# just a tab
c = c + "\\t"
else:
# any other char, just append
c = c + idPart.strip()
return c
else:
if config.DEBUG: print("Couldn't parse STRCAT string")
return ""
# escaping " in grok content, and adding anchors if passed as param
def escapeGrok(s):
if s == "": return s
if config.NO_GROK_ANCHORS:
return "\"" + str.strip(s.replace("\"","\\\"")) + "\""
else:
return "\"^" + str.strip(s.replace("\"","\\\"")) + "$\""
# escaping " in dissect content
def escapeDissect(s):
if s == "": return s
return "\"" + str.strip(s.replace("\"","\\\"")) + "\""
# escapes a string in logstash (only escaping ")
def escapeString(s):
return str.strip(s.replace("\"","\\\""))
# escape special characters in grok : \ . ^ $ * + - ? ( ) [ ] { } |
def escapeRegex(s):
# s = str(s,'utf-8')
s = s.replace("\\", "\\\\"); s = s.replace(".", "\\."); s = s.replace("^", "\\^"); s = s.replace("$", "\\$")
s = s.replace("*", "\\*"); s = s.replace("+", "\\+"); s = s.replace("-", "\\-"); s = s.replace("?", "\\?")
s = s.replace("(", "\\("); s = s.replace(")", "\\)"); s = s.replace("[", "\\["); s = s.replace("]", "\\]")
s = s.replace("{", "\\{"); s = s.replace("}", "\\}"); s = s.replace("|", "\\|")
s = s.replace(chr(9), "\\t"); s = s.replace(chr(10), "\\n"); s = s.replace(chr(13), "\\r"); s = s.replace(" ", "\\s")
return s
# remove dots in field names
def removeDots(fldName):
if "." in fldName:
return "[" + fldName.replace(".","][") + "]"
else:
return fldName
# transform a nested field into a nested es mapping
def generateFieldMapping(fldName, fldType):
# either key.subkey as RSA format
if "." in fldName:
# split on .
flds = fldName.split(".")
if len(flds) == 2: config.esMap["mappings"]["properties"][flds[0]]["properties"][flds[1]] = {"type": fldType}
elif len(flds) == 3: config.esMap["mappings"]["properties"][flds[0]]["properties"][flds[1]]["properties"][flds[2]] = {"type": fldType}
elif len(flds) == 4: config.esMap["mappings"]["properties"][flds[0]]["properties"][flds[1]]["properties"][flds[2]]["properties"][flds[3]] = {"type": fldType}
elif len(flds) == 5: config.esMap["mappings"]["properties"][flds[0]]["properties"][flds[1]]["properties"][flds[2]]["properties"][flds[3]]["properties"][flds[4]] = {"type": fldType}
else:
print("Error: are you really using more than 5 levels of nested field?")
sys.exit(-1)
# or [key][subkey] after mutate
elif "][" in fldName:
# split on ][
flds = fldName[1:-1].split("][")
if len(flds) == 2: config.esMap["mappings"]["properties"][flds[0]]["properties"][flds[1]] = {"type": fldType}
elif len(flds) == 3: config.esMap["mappings"]["properties"][flds[0]]["properties"][flds[1]]["properties"][flds[2]] = {"type": fldType}
elif len(flds) == 4: config.esMap["mappings"]["properties"][flds[0]]["properties"][flds[1]]["properties"][flds[2]]["properties"][flds[3]] = {"type": fldType}
elif len(flds) == 5: config.esMap["mappings"]["properties"][flds[0]]["properties"][flds[1]]["properties"][flds[2]]["properties"][flds[3]]["properties"][flds[4]] = {"type": fldType}
else:
print("Error: are you really using more than 5 levels of nested field?")
sys.exit(-1)
# plain key
else:
config.esMap["mappings"]["properties"][fldName] = {"type": fldType}
# read table-map and populate dicts
def initMapping():
noComma = False; noSemiColon = False
# trying to open table map file with , delimiter first (french way!)
with open(config.MAPPING_FILE,'r') as csvFile:
try:
reader = csv.DictReader(csvFile, delimiter=',')
for row in reader:
# if field is flagged as "Transient", we don't take it into consideration
if row['ecsName'] != "" and row['flags'] != "Transient":
config.ecsField[row['envisionName']] = row['ecsName']
config.ecsType[row['envisionName']] = row['ecsType']
except:
noComma = True
if noComma:
with open(config.MAPPING_FILE,'r') as csvFile:
try:
reader = csv.DictReader(csvFile, delimiter=';')
for row in reader:
# TODO: not sure what I should do with the fields marked "Transient"
if row['ecsName'] != "": # and row['flags'] != "Transient":
config.ecsField[row['envisionName']] = row['ecsName']
config.ecsType[row['envisionName']] = row['ecsType']
except:
noSemiColon = True
if noComma and noSemiColon:
print("CSV mapping file expecting ; or , delimiter. Exiting...")
sys.exit(-1)
# get a possible valuemap func and store the fields
def getValueMap(fld,vmFunc):
# form is *getEventCategoryActivity(action)
pattern = re.compile("\*([^\(]+)\(([^\)]+)\)")
m = pattern.match(vmFunc)
if m is not None:
k,v = m.group(1),m.group(2)
# let's look for the func in the valuemap funcs we've stored
if k in config.valueMap:
# record the func parameter as the key
config.valueMap[k]["fld"] = v
# record the destination field as well
config.valueMap[k]["newFld"] = fld
config.allFields.add(fld)
| [
"csv.DictReader",
"re.compile",
"config.allFields.add",
"sys.exit",
"re.sub"
] | [((625, 649), 're.compile', 're.compile', (['""",\'([^\']+)\'"""'], {}), '(",\'([^\']+)\'")\n', (635, 649), False, 'import re\n'), ((2228, 2254), 're.compile', 're.compile', (['"""^[a-zA-Z]+.*"""'], {}), "('^[a-zA-Z]+.*')\n", (2238, 2254), False, 'import re\n'), ((8064, 8105), 're.compile', 're.compile', (['"""\\\\*([^\\\\(]+)\\\\(([^\\\\)]+)\\\\)"""'], {}), "('\\\\*([^\\\\(]+)\\\\(([^\\\\)]+)\\\\)')\n", (8074, 8105), False, 'import re\n'), ((902, 944), 're.sub', 're.sub', (['"""([^%])([a-zA-Z]+)"""', '"""\\\\1\'\\\\2\'"""', 'c'], {}), '(\'([^%])([a-zA-Z]+)\', "\\\\1\'\\\\2\'", c)\n', (908, 944), False, 'import re\n'), ((7907, 7919), 'sys.exit', 'sys.exit', (['(-1)'], {}), '(-1)\n', (7915, 7919), False, 'import sys\n'), ((6805, 6843), 'csv.DictReader', 'csv.DictReader', (['csvFile'], {'delimiter': '""","""'}), "(csvFile, delimiter=',')\n", (6819, 6843), False, 'import csv\n'), ((8494, 8519), 'config.allFields.add', 'config.allFields.add', (['fld'], {}), '(fld)\n', (8514, 8519), False, 'import config\n'), ((7341, 7379), 'csv.DictReader', 'csv.DictReader', (['csvFile'], {'delimiter': '""";"""'}), "(csvFile, delimiter=';')\n", (7355, 7379), False, 'import csv\n'), ((5570, 5582), 'sys.exit', 'sys.exit', (['(-1)'], {}), '(-1)\n', (5578, 5582), False, 'import sys\n'), ((6435, 6447), 'sys.exit', 'sys.exit', (['(-1)'], {}), '(-1)\n', (6443, 6447), False, 'import sys\n')] |
# -*- coding: utf-8 -*-
import os
import shutil
import io
import zipfile
from commons.Utils import path_utils
class ZipHelper(object):
@classmethod
def _get_arcname(cls, old_arcname, new_arcname):
if old_arcname is not None:
return os.path.join(old_arcname, new_arcname)
else:
return new_arcname
@classmethod
def zip(cls, src, dest, arcname=None):
if os.path.isdir(src):
for root, filepath_list, filename_list in os.walk(src):
for filepath in filepath_list:
ZipHelper.zip(
src=os.path.join(root, filepath), dest=dest, arcname=ZipHelper._get_arcname(arcname, filepath)
)
for filename in filename_list:
with zipfile.ZipFile(dest, 'a', zipfile.ZIP_DEFLATED) as zf:
zf.write(os.path.join(root, filename), arcname=ZipHelper._get_arcname(arcname, filename))
break
elif os.path.isfile(src):
filepath, filename = os.path.split(src)
with zipfile.ZipFile(dest, 'a', zipfile.ZIP_DEFLATED) as zf:
zf.write(src, arcname=ZipHelper._get_arcname(arcname, filename))
@classmethod
def zip_file(cls, file, zipname=None, exclude_parent=False):
if not os.path.isfile(file):
return False
folder, filename = os.path.split(file)
if exclude_parent:
base_path, pathname = path_utils.get_path_split(folder)
if zipname is None:
zipname = os.path.join(base_path, "{0}.zip".format(pathname))
ZipHelper.zip(file, zipname, arcname=None)
else:
filename_base, filename_ext = os.path.splitext(filename)
if zipname is None:
zipname = os.path.join(folder, "{0}.zip".format(filename_base))
ZipHelper.zip(file, zipname, arcname=None)
return True
@classmethod
def zip_folder(cls, folder, zipname=None, exclude_parent=False):
if not os.path.isdir(folder):
return False
base_path, pathname = path_utils.get_path_split(folder)
if zipname is None:
zipname = os.path.join(base_path, "{0}.zip".format(pathname))
if exclude_parent:
ZipHelper.zip(folder, zipname, arcname=pathname)
else:
ZipHelper.zip(folder, zipname, arcname=None)
return True
@classmethod
def unzip(cls, zipname, dest=None):
"""服务器端解压文件生成r_list.dat文件"""
zf = zipfile.ZipFile(zipname, 'r')
dest = dest.replace('\\', '/')
if dest.endswith('/'):
dest = dest[:-1]
for filename in zf.namelist():
fname = filename
new_file = '%s/%s' % (dest, fname)
if new_file.endswith('/') or new_file.endswith('\\'):
if not os.path.exists(new_file):
os.makedirs(new_file)
else:
filepath = os.path.split(new_file)[0]
if not os.path.exists(filepath):
os.makedirs(filepath)
try:
f = open(new_file, 'wb')
except UnicodeEncodeError:
f = open(new_file.encode('utf8'), 'wb')
f.write(zf.read(filename))
f.close()
class InMemoryZip(object):
"""用法:
imz = InMemoryZip()
imz.append("test.txt", "Another test").append("test2.txt", "Still another")
imz.writetofile("test.zip")
"""
def __init__(self):
# Create the in-memory file-like object
self.in_memory_zip = io.StringIO()
def append(self, filename_in_zip, file_contents):
"""Appends a file with name filename_in_zip and contents of
file_contents to the in-memory zip."""
# Get a handle to the in-memory zip in append mode
zf = zipfile.ZipFile(self.in_memory_zip, "a", zipfile.ZIP_DEFLATED, False)
# Write the file to the in-memory zip
zf.writestr(filename_in_zip, file_contents)
# Mark the files as having been created on Windows so that
# Unix permissions are not inferred as 0000
for zfile in zf.filelist:
zfile.create_system = 0
return self
def read(self):
'''Returns a string with the contents of the in-memory zip.'''
self.in_memory_zip.seek(0)
return self.in_memory_zip.read()
def writetofile(self, filename):
'''Writes the in-memory zip to a file.'''
with open(filename, "wb") as f:
f.write(self.read())
| [
"os.path.exists",
"zipfile.ZipFile",
"os.makedirs",
"os.path.join",
"os.path.splitext",
"os.path.split",
"os.path.isfile",
"os.path.isdir",
"commons.Utils.path_utils.get_path_split",
"io.StringIO",
"os.walk"
] | [((419, 437), 'os.path.isdir', 'os.path.isdir', (['src'], {}), '(src)\n', (432, 437), False, 'import os\n'), ((1407, 1426), 'os.path.split', 'os.path.split', (['file'], {}), '(file)\n', (1420, 1426), False, 'import os\n'), ((2139, 2172), 'commons.Utils.path_utils.get_path_split', 'path_utils.get_path_split', (['folder'], {}), '(folder)\n', (2164, 2172), False, 'from commons.Utils import path_utils\n'), ((2564, 2593), 'zipfile.ZipFile', 'zipfile.ZipFile', (['zipname', '"""r"""'], {}), "(zipname, 'r')\n", (2579, 2593), False, 'import zipfile\n'), ((3652, 3665), 'io.StringIO', 'io.StringIO', ([], {}), '()\n', (3663, 3665), False, 'import io\n'), ((3908, 3977), 'zipfile.ZipFile', 'zipfile.ZipFile', (['self.in_memory_zip', '"""a"""', 'zipfile.ZIP_DEFLATED', '(False)'], {}), "(self.in_memory_zip, 'a', zipfile.ZIP_DEFLATED, False)\n", (3923, 3977), False, 'import zipfile\n'), ((263, 301), 'os.path.join', 'os.path.join', (['old_arcname', 'new_arcname'], {}), '(old_arcname, new_arcname)\n', (275, 301), False, 'import os\n'), ((493, 505), 'os.walk', 'os.walk', (['src'], {}), '(src)\n', (500, 505), False, 'import os\n'), ((1007, 1026), 'os.path.isfile', 'os.path.isfile', (['src'], {}), '(src)\n', (1021, 1026), False, 'import os\n'), ((1332, 1352), 'os.path.isfile', 'os.path.isfile', (['file'], {}), '(file)\n', (1346, 1352), False, 'import os\n'), ((1489, 1522), 'commons.Utils.path_utils.get_path_split', 'path_utils.get_path_split', (['folder'], {}), '(folder)\n', (1514, 1522), False, 'from commons.Utils import path_utils\n'), ((1744, 1770), 'os.path.splitext', 'os.path.splitext', (['filename'], {}), '(filename)\n', (1760, 1770), False, 'import os\n'), ((2060, 2081), 'os.path.isdir', 'os.path.isdir', (['folder'], {}), '(folder)\n', (2073, 2081), False, 'import os\n'), ((1061, 1079), 'os.path.split', 'os.path.split', (['src'], {}), '(src)\n', (1074, 1079), False, 'import os\n'), ((1097, 1145), 'zipfile.ZipFile', 'zipfile.ZipFile', (['dest', '"""a"""', 'zipfile.ZIP_DEFLATED'], {}), "(dest, 'a', zipfile.ZIP_DEFLATED)\n", (1112, 1145), False, 'import zipfile\n'), ((2897, 2921), 'os.path.exists', 'os.path.exists', (['new_file'], {}), '(new_file)\n', (2911, 2921), False, 'import os\n'), ((2943, 2964), 'os.makedirs', 'os.makedirs', (['new_file'], {}), '(new_file)\n', (2954, 2964), False, 'import os\n'), ((3010, 3033), 'os.path.split', 'os.path.split', (['new_file'], {}), '(new_file)\n', (3023, 3033), False, 'import os\n'), ((3060, 3084), 'os.path.exists', 'os.path.exists', (['filepath'], {}), '(filepath)\n', (3074, 3084), False, 'import os\n'), ((3106, 3127), 'os.makedirs', 'os.makedirs', (['filepath'], {}), '(filepath)\n', (3117, 3127), False, 'import os\n'), ((802, 850), 'zipfile.ZipFile', 'zipfile.ZipFile', (['dest', '"""a"""', 'zipfile.ZIP_DEFLATED'], {}), "(dest, 'a', zipfile.ZIP_DEFLATED)\n", (817, 850), False, 'import zipfile\n'), ((617, 645), 'os.path.join', 'os.path.join', (['root', 'filepath'], {}), '(root, filepath)\n', (629, 645), False, 'import os\n'), ((891, 919), 'os.path.join', 'os.path.join', (['root', 'filename'], {}), '(root, filename)\n', (903, 919), False, 'import os\n')] |