code
stringlengths
22
1.05M
apis
listlengths
1
3.31k
extract_api
stringlengths
75
3.25M
import typer import deciphon_cli.data as data __all__ = ["app"] app = typer.Typer() @app.command() def default(): typer.echo(data.env_example_content(), nl=False)
[ "deciphon_cli.data.env_example_content", "typer.Typer" ]
[((73, 86), 'typer.Typer', 'typer.Typer', ([], {}), '()\n', (84, 86), False, 'import typer\n'), ((134, 160), 'deciphon_cli.data.env_example_content', 'data.env_example_content', ([], {}), '()\n', (158, 160), True, 'import deciphon_cli.data as data\n')]
import olll import numpy as np test1 = [[1,0,0,1,1,0,1],[0,1,0,5,0,0,0],[0,0,1,0,5,0,5]] test2 = [[1,0,0,2,-1,1],[0,1,0,3,-4,-2],[0,0,1,5,-10,-8]] test3 = [[1,0,0,1,1,0,1], [0,1,0,4,-1,0,-1], [0,0,1,1,1,0,1]] test4 = [[1,0,0,2,5,3],[0,1,0,1,1,1,],[0,0,1,4,-2,0]] test5 = [[1,0,0,0,0,0,2,1,1,2],[0,1,0,0,0,0,1,1,-1,-1],[0,0,1,0,0,0,-1,0,-2,-3],[0,0,0,1,0,0,1,-1,1,-1],[0,0,0,0,1,0,-1,2,-4,-3],[0,0,0,0,0,1,1,0,0,1]] test6 = [[1, 0, 0, 5, 0, 0, 0],[0, 1, 0, 0, 5, 0, 5],[0, 0, 1, 1, 1, 0, 1]] test7 = [[1, 0, 0, 20, 0, 0, 0],[0, 1, 0, 0, 20, 0, 20],[0, 0, 1, 4, 4, 0, 4]] test8 = [[1, 0, 0, 10, 0, 0, 0],[0, 1, 0, 0, 10, 0, 10],[0, 0, 1, 2, 2, 0, 2]] n = input("Please enter n: \n") n = int(n) k = input("Please enter k: \n") k = int(k) p = input("Please enter p: \n") p = int(p) id = np.identity(k) A = [[]] * k print("Please enter the generating set:\n") for i in range(k): print("\nEnter the generator a[",i,"]: ") a = list(map(int,input().strip().split()))[:n] #print(i, a) a = [x * (2**p) for x in a] y = list(id[i]) print(y[i],type(y[i])) A[i] = y+a print(A[i], type(A[i])) print(A, type(A)) print(test7, type(test7)) rb = olll.reduction(test7,0.75) print("Basis: ", rb)
[ "numpy.identity", "olll.reduction" ]
[((787, 801), 'numpy.identity', 'np.identity', (['k'], {}), '(k)\n', (798, 801), True, 'import numpy as np\n'), ((1169, 1196), 'olll.reduction', 'olll.reduction', (['test7', '(0.75)'], {}), '(test7, 0.75)\n', (1183, 1196), False, 'import olll\n')]
from __future__ import annotations import logging from typing import Any from homeassistant.components.integration.sensor import ( TRAPEZOIDAL_METHOD, IntegrationSensor, ) from homeassistant.components.sensor import DOMAIN as SENSOR_DOMAIN from homeassistant.const import CONF_NAME, TIME_HOURS from homeassistant.helpers.entity import async_generate_entity_id from homeassistant.helpers.typing import HomeAssistantType from custom_components.powercalc.common import SourceEntity from custom_components.powercalc.const import ( ATTR_SOURCE_DOMAIN, ATTR_SOURCE_ENTITY, CONF_ENERGY_SENSOR_NAMING, ) from custom_components.powercalc.migrate import async_migrate_entity_id from .power import VirtualPowerSensor ENERGY_ICON = "mdi:lightning-bolt" ENTITY_ID_FORMAT = SENSOR_DOMAIN + ".{}" _LOGGER = logging.getLogger(__name__) async def create_energy_sensor( hass: HomeAssistantType, sensor_config: dict, power_sensor: VirtualPowerSensor, source_entity: SourceEntity, ) -> VirtualEnergySensor: """Create the energy sensor entity""" name_pattern = sensor_config.get(CONF_ENERGY_SENSOR_NAMING) name = sensor_config.get(CONF_NAME) or source_entity.name name = name_pattern.format(name) object_id = sensor_config.get(CONF_NAME) or source_entity.object_id entity_id = async_generate_entity_id( ENTITY_ID_FORMAT, name_pattern.format(object_id), hass=hass ) unique_id = None if source_entity.unique_id: unique_id = f"{source_entity.unique_id}_energy" async_migrate_entity_id(hass, "sensor", unique_id, entity_id) _LOGGER.debug("Creating energy sensor: %s", name) return VirtualEnergySensor( source_entity=power_sensor.entity_id, unique_id=unique_id, entity_id=entity_id, name=name, round_digits=4, unit_prefix="k", unit_of_measurement=None, unit_time=TIME_HOURS, integration_method=TRAPEZOIDAL_METHOD, powercalc_source_entity=source_entity.entity_id, powercalc_source_domain=source_entity.domain, ) class VirtualEnergySensor(IntegrationSensor): """Virtual energy sensor, totalling kWh""" def __init__( self, source_entity, unique_id, entity_id, name, round_digits, unit_prefix, unit_time, unit_of_measurement, integration_method, powercalc_source_entity: str, powercalc_source_domain: str, ): super().__init__( source_entity, name, round_digits, unit_prefix, unit_time, unit_of_measurement, integration_method, ) self._powercalc_source_entity = powercalc_source_entity self._powercalc_source_domain = powercalc_source_domain self.entity_id = entity_id if unique_id: self._attr_unique_id = unique_id @property def extra_state_attributes(self) -> dict[str, Any]: """Return the state attributes of the acceleration sensor.""" state_attr = super().extra_state_attributes state_attr[ATTR_SOURCE_ENTITY] = self._powercalc_source_entity state_attr[ATTR_SOURCE_DOMAIN] = self._powercalc_source_domain return state_attr @property def icon(self): return ENERGY_ICON
[ "logging.getLogger", "custom_components.powercalc.migrate.async_migrate_entity_id" ]
[((818, 845), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (835, 845), False, 'import logging\n'), ((1542, 1603), 'custom_components.powercalc.migrate.async_migrate_entity_id', 'async_migrate_entity_id', (['hass', '"""sensor"""', 'unique_id', 'entity_id'], {}), "(hass, 'sensor', unique_id, entity_id)\n", (1565, 1603), False, 'from custom_components.powercalc.migrate import async_migrate_entity_id\n')]
import numpy as np import matplotlib.pyplot as plt import ipywidgets from mesostat.utils.opencv_helper import cvWriter from mesostat.utils.arrays import numpy_merge_dimensions from sklearn.decomposition import PCA def distance_matrix(data): nDim, nTime = data.shape dataExtr = np.repeat(data[..., None], nTime, axis=2) delta = dataExtr - dataExtr.transpose((0,2,1)) return np.linalg.norm(delta, axis=0) def decorrelate(data): # Leading dimension must be channels if data.ndim > 2: dataEff = numpy_merge_dimensions(data, 1, data.ndim+1) else: dataEff = data pca = PCA(n_components=48) rez = pca.fit_transform(dataEff.T) print(rez.shape) rez /= np.std(rez, axis=0) return rez.T.reshape(data.shape) class RecurrencePlot: def __init__(self, data, w0): self.nPoint, self.nTime = data.shape # Compute the recurrence plots with no threshold self.dist = distance_matrix(data) # Plot first picdata = self.get_plotdata(w0) self.fig, self.ax = plt.subplots(figsize=(5, 5)) self.pic = self.ax.imshow(picdata, cmap='binary', origin='lower') self.ax.set_title('Recurrence Plot', fontsize=16) def get_plotdata(self, w): thr = np.percentile(self.dist.flatten(), w) return self.dist <= thr def update(self, w): picnew = self.get_plotdata(w) self.pic.set_data(picnew) self.fig.canvas.draw_idle() def interact(self): ipywidgets.interact(self.update, w=(0, 100, 1)) def write_video(self, fname, frate=10.0, codec='XVID'): frameDim = (self.nTime, self.nTime) with cvWriter(fname, frameDim, frate=frate, codec=codec) as writer: for w in range(101): data = self.get_plotdata(w) writer.write(data) class RecurrencePlotMultitrial: def __init__(self, data3D, w0): self.nTrial, self.nPoint, self.nTime = data3D.shape binarize = lambda dist, w: dist <= np.percentile(dist, w) # Compute the recurrence plots with no threshold self.rezMat = np.zeros((101, self.nTime, self.nTime), dtype=float) for data in data3D: dist = distance_matrix(data) for w in range(101): self.rezMat[w] += binarize(dist, w).astype(float) self.rezMat /= self.nTrial # Plot first self.fig, self.ax = plt.subplots(figsize=(5, 5)) self.pic = self.ax.imshow(self.rezMat[w0], cmap='viridis', origin='lower', vmin=0, vmax=1) self.ax.set_title('Recurrence Plot', fontsize=16) def update(self, w): self.pic.set_data(self.rezMat[w]) self.fig.canvas.draw_idle() def interact(self): ipywidgets.interact(self.update, w=(0, 100, 1)) def write_video(self, fname, frate=10.0, codec='XVID'): frameDim = (self.nTime, self.nTime) with cvWriter(fname, frameDim, frate=frate, codec=codec) as writer: for w in range(101): writer.write(self.rezMat[w])
[ "numpy.repeat", "sklearn.decomposition.PCA", "numpy.linalg.norm", "numpy.zeros", "mesostat.utils.opencv_helper.cvWriter", "ipywidgets.interact", "numpy.std", "numpy.percentile", "mesostat.utils.arrays.numpy_merge_dimensions", "matplotlib.pyplot.subplots" ]
[((288, 329), 'numpy.repeat', 'np.repeat', (['data[..., None]', 'nTime'], {'axis': '(2)'}), '(data[..., None], nTime, axis=2)\n', (297, 329), True, 'import numpy as np\n'), ((392, 421), 'numpy.linalg.norm', 'np.linalg.norm', (['delta'], {'axis': '(0)'}), '(delta, axis=0)\n', (406, 421), True, 'import numpy as np\n'), ((617, 637), 'sklearn.decomposition.PCA', 'PCA', ([], {'n_components': '(48)'}), '(n_components=48)\n', (620, 637), False, 'from sklearn.decomposition import PCA\n'), ((711, 730), 'numpy.std', 'np.std', (['rez'], {'axis': '(0)'}), '(rez, axis=0)\n', (717, 730), True, 'import numpy as np\n'), ((528, 574), 'mesostat.utils.arrays.numpy_merge_dimensions', 'numpy_merge_dimensions', (['data', '(1)', '(data.ndim + 1)'], {}), '(data, 1, data.ndim + 1)\n', (550, 574), False, 'from mesostat.utils.arrays import numpy_merge_dimensions\n'), ((1061, 1089), 'matplotlib.pyplot.subplots', 'plt.subplots', ([], {'figsize': '(5, 5)'}), '(figsize=(5, 5))\n', (1073, 1089), True, 'import matplotlib.pyplot as plt\n'), ((1505, 1552), 'ipywidgets.interact', 'ipywidgets.interact', (['self.update'], {'w': '(0, 100, 1)'}), '(self.update, w=(0, 100, 1))\n', (1524, 1552), False, 'import ipywidgets\n'), ((2123, 2175), 'numpy.zeros', 'np.zeros', (['(101, self.nTime, self.nTime)'], {'dtype': 'float'}), '((101, self.nTime, self.nTime), dtype=float)\n', (2131, 2175), True, 'import numpy as np\n'), ((2430, 2458), 'matplotlib.pyplot.subplots', 'plt.subplots', ([], {'figsize': '(5, 5)'}), '(figsize=(5, 5))\n', (2442, 2458), True, 'import matplotlib.pyplot as plt\n'), ((2753, 2800), 'ipywidgets.interact', 'ipywidgets.interact', (['self.update'], {'w': '(0, 100, 1)'}), '(self.update, w=(0, 100, 1))\n', (2772, 2800), False, 'import ipywidgets\n'), ((1672, 1723), 'mesostat.utils.opencv_helper.cvWriter', 'cvWriter', (['fname', 'frameDim'], {'frate': 'frate', 'codec': 'codec'}), '(fname, frameDim, frate=frate, codec=codec)\n', (1680, 1723), False, 'from mesostat.utils.opencv_helper import cvWriter\n'), ((2920, 2971), 'mesostat.utils.opencv_helper.cvWriter', 'cvWriter', (['fname', 'frameDim'], {'frate': 'frate', 'codec': 'codec'}), '(fname, frameDim, frate=frate, codec=codec)\n', (2928, 2971), False, 'from mesostat.utils.opencv_helper import cvWriter\n'), ((2020, 2042), 'numpy.percentile', 'np.percentile', (['dist', 'w'], {}), '(dist, w)\n', (2033, 2042), True, 'import numpy as np\n')]
import sys import os import numpy as np import scipy.io as sio import random from decimal import Decimal import argparse import csv from keras.models import load_model import f_model from f_preprocess import fill_length # Usage: python rematch_challenge.py test_file_path def arg_parse(): """ Parse arguements """ parser = argparse.ArgumentParser(description='Rematch test of ECG Contest') parser.add_argument("--test_path", dest='test_path', help= "the file path of Test Data", default="your test_path", type=str) #You need to write your test data path with the argparse parameter. #For your convenience when testing with local data, you can write your local test set path to default return parser.parse_args() def main(): args = arg_parse() test_path = args.test_path print(test_path) ## Add your codes to classify normal and diseases. model01_path = './model/model_01.h5' #model02_path = '/media/uuser/data/final_run/model/model_02.h5' #modelxg_path = './model/model.pkl' #feature_path = './data/feature.csv' keysname = ('I','II','III','aVR','aVL','aVF', \ 'V1','V2','V3','V4','V5','V6','age','sex') t_len = 25000 len_target=t_len model_01 = f_model.build_model_01(num_classes=10,len_target=len_target) model_01.load_weights(model01_path) ## Classify the samples of the test set and write the results into answers.txt, ## and each row representing a prediction of one sample. ## Here we use random numbers as prediction labels as an example and ## you should replace it with your own results. Data_list = os.listdir(test_path) classes = np.asarray([1,1,2,3,4,5,6,7,8,9]) with open('answers.csv', 'w') as csvfile: writer = csv.writer(csvfile) writer.writerow(['File_name', 'label1', 'label2', 'label3', 'label4', 'label5', 'label6', 'label7', 'label8', 'label9', 'label10']) for file_name in Data_list: if file_name.endswith('.mat'): answer = [] record_name = file_name.strip('.mat') answer.append(record_name) # model 01 ecg = np.empty([t_len,12]) mypath=test_path+file_name data = sio.loadmat(mypath) # read 12 leads for lead in range(12): temp=data[keysname[lead]] ecg[:,lead] = fill_length(temp,t_len) data_x = ecg.reshape((1,t_len,12)) pred_1 = model_01.predict(data_x) # model 02 # model xgboost preds = pred_1 preds[preds>=0.5] = 1 preds[preds<0.5] = 0 pred_out = preds * classes y_out =[] for i in range(10): if pred_out[0][i]==classes[i]: y_out.append(i) for x in range(10-len(y_out)): y_out.append('') writer.writerow(answer+y_out) csvfile.close() if __name__ == "__main__": main()
[ "f_preprocess.fill_length", "os.listdir", "argparse.ArgumentParser", "csv.writer", "numpy.asarray", "scipy.io.loadmat", "numpy.empty", "f_model.build_model_01" ]
[((343, 409), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {'description': '"""Rematch test of ECG Contest"""'}), "(description='Rematch test of ECG Contest')\n", (366, 409), False, 'import argparse\n'), ((1293, 1354), 'f_model.build_model_01', 'f_model.build_model_01', ([], {'num_classes': '(10)', 'len_target': 'len_target'}), '(num_classes=10, len_target=len_target)\n', (1315, 1354), False, 'import f_model\n'), ((1685, 1706), 'os.listdir', 'os.listdir', (['test_path'], {}), '(test_path)\n', (1695, 1706), False, 'import os\n'), ((1721, 1763), 'numpy.asarray', 'np.asarray', (['[1, 1, 2, 3, 4, 5, 6, 7, 8, 9]'], {}), '([1, 1, 2, 3, 4, 5, 6, 7, 8, 9])\n', (1731, 1763), True, 'import numpy as np\n'), ((1818, 1837), 'csv.writer', 'csv.writer', (['csvfile'], {}), '(csvfile)\n', (1828, 1837), False, 'import csv\n'), ((2232, 2253), 'numpy.empty', 'np.empty', (['[t_len, 12]'], {}), '([t_len, 12])\n', (2240, 2253), True, 'import numpy as np\n'), ((2319, 2338), 'scipy.io.loadmat', 'sio.loadmat', (['mypath'], {}), '(mypath)\n', (2330, 2338), True, 'import scipy.io as sio\n'), ((2490, 2514), 'f_preprocess.fill_length', 'fill_length', (['temp', 't_len'], {}), '(temp, t_len)\n', (2501, 2514), False, 'from f_preprocess import fill_length\n')]
#!/usr/bin/env python # coding=utf-8 # This is free and unencumbered software released into the public domain. # # Anyone is free to copy, modify, publish, use, compile, sell, or # distribute this software, either in source code form or as a compiled # binary, for any purpose, commercial or non-commercial, and by any # means. # # In jurisdictions that recognize copyright laws, the author or authors # of this software dedicate any and all copyright interest in the # software to the public domain. We make this dedication for the benefit # of the public at large and to the detriment of our heirs and # successors. We intend this dedication to be an overt act of # relinquishment in perpetuity of all present and future rights to this # software under copyright law. # # THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, # EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF # MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. # IN NO EVENT SHALL THE AUTHORS BE LIABLE FOR ANY CLAIM, DAMAGES OR # OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, # ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR # OTHER DEALINGS IN THE SOFTWARE. # # For more information, please refer to <http://unlicense.org/> """ This script demonstrates how to use the bmreports.UnitData class to get information about Constraint Payments. It does not attempt to cater for long or short days and simply assumes that there will be 48 settlement periods. """ import argparse from datetime import datetime, timedelta, date from pywind.bmreports.unit import UnitData def mkdate(datestr): return datetime.strptime(datestr, '%Y-%m-%d').date() def main(): parser = argparse.ArgumentParser(description='Get Constraint Payment information for yesterday') parser.add_argument('--date', action='store', type=mkdate, help='Date to get results for') parser.add_argument('--period', action='store', help='Period to get data for') args = parser.parse_args() data = {} ud = UnitData({'date': args.date or date.today() - timedelta(days=2)}) pr = [args.period] or range(1,49) for period in pr: ud.period = period if ud.get_data(): data[period] = ud.data else: print ("Unable to get data for %s, period %d" % (ud.date.strftime("%d %b %Y"), period)) for period, units in sorted(data.iteritems()): print ("Period: ", period) for unit in sorted(units, key=lambda x: x['ngc']): print (" ", unit['ngc'], unit['lead']) if unit['bid'].has_key('volume'): print (" BID: ", unit['bid']['volume']+'MWh ', unit['bid']['cashflow']) if unit['offer'].has_key('volume'): print (" OFFER: ", unit['offer']['volume']+'MWh ', unit['offer']['cashflow']) if __name__ == '__main__': main()
[ "datetime.datetime.strptime", "datetime.date.today", "argparse.ArgumentParser", "datetime.timedelta" ]
[((1746, 1838), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {'description': '"""Get Constraint Payment information for yesterday"""'}), "(description=\n 'Get Constraint Payment information for yesterday')\n", (1769, 1838), False, 'import argparse\n'), ((1673, 1711), 'datetime.datetime.strptime', 'datetime.strptime', (['datestr', '"""%Y-%m-%d"""'], {}), "(datestr, '%Y-%m-%d')\n", (1690, 1711), False, 'from datetime import datetime, timedelta, date\n'), ((2098, 2110), 'datetime.date.today', 'date.today', ([], {}), '()\n', (2108, 2110), False, 'from datetime import datetime, timedelta, date\n'), ((2113, 2130), 'datetime.timedelta', 'timedelta', ([], {'days': '(2)'}), '(days=2)\n', (2122, 2130), False, 'from datetime import datetime, timedelta, date\n')]
"""Run a minimal bot.""" import nonebot from nonebot.adapters.cqhttp import Bot nonebot.init() app = nonebot.get_asgi() driver = nonebot.get_driver() driver.register_adapter("cqhttp", Bot) nonebot.load_builtin_plugins() nonebot.load_plugins("nonebot_plugin_guess") if __name__ == "__main__": # nonebot.run() # nonebot.run(app="bot:app") nonebot.run(app="minimal_bot:app")
[ "nonebot.run", "nonebot.load_builtin_plugins", "nonebot.get_driver", "nonebot.load_plugins", "nonebot.get_asgi", "nonebot.init" ]
[((81, 95), 'nonebot.init', 'nonebot.init', ([], {}), '()\n', (93, 95), False, 'import nonebot\n'), ((102, 120), 'nonebot.get_asgi', 'nonebot.get_asgi', ([], {}), '()\n', (118, 120), False, 'import nonebot\n'), ((131, 151), 'nonebot.get_driver', 'nonebot.get_driver', ([], {}), '()\n', (149, 151), False, 'import nonebot\n'), ((192, 222), 'nonebot.load_builtin_plugins', 'nonebot.load_builtin_plugins', ([], {}), '()\n', (220, 222), False, 'import nonebot\n'), ((223, 267), 'nonebot.load_plugins', 'nonebot.load_plugins', (['"""nonebot_plugin_guess"""'], {}), "('nonebot_plugin_guess')\n", (243, 267), False, 'import nonebot\n'), ((353, 387), 'nonebot.run', 'nonebot.run', ([], {'app': '"""minimal_bot:app"""'}), "(app='minimal_bot:app')\n", (364, 387), False, 'import nonebot\n')]
from lib.appController import driver_queue from lib.pyapp import Pyapp import threading # driver 多线程运行是进行线程之间的数据隔离 local = threading.local() # 配置在实例化时,去mq中获取创建好的driver,如果调试page则需要传递driver class BasePage(object): def __init__(self, driver=None): if driver is None: local.driver = driver_queue.get() local.pyapp = Pyapp(local.driver) else: local.driver = driver local.pyapp = Pyapp(driver) def quit(self): local.app.quit() def reset_package(self): local.pyapp.reset() class QQ_Login_Page(BasePage): def login(self): local.pyapp.click('android=>new UiSelector().text("登 录")') def username(self): local.pyapp.type('content=>请输入QQ号码或手机或邮箱', 3408467505) def passwd(self): local.pyapp.type('content=>密码 安全', '<PASSWORD>') def login_check(self, name): return local.pyapp.wait_and_save_exception('android=>new UiSelector().text("登 录")', name) class Page(QQ_Login_Page): pass
[ "threading.local", "lib.pyapp.Pyapp", "lib.appController.driver_queue.get" ]
[((124, 141), 'threading.local', 'threading.local', ([], {}), '()\n', (139, 141), False, 'import threading\n'), ((306, 324), 'lib.appController.driver_queue.get', 'driver_queue.get', ([], {}), '()\n', (322, 324), False, 'from lib.appController import driver_queue\n'), ((351, 370), 'lib.pyapp.Pyapp', 'Pyapp', (['local.driver'], {}), '(local.driver)\n', (356, 370), False, 'from lib.pyapp import Pyapp\n'), ((445, 458), 'lib.pyapp.Pyapp', 'Pyapp', (['driver'], {}), '(driver)\n', (450, 458), False, 'from lib.pyapp import Pyapp\n')]
from NoviCypher import FileCypher from os import path import sys Version = '1.0.0b' def encode(p, r, k, c): print(f"encoding {p}\nrows={r}\nkey={k}\nchunk={c}") FileCypher(p, rows=r, chunk=c, key=k) input() def decode(p, k): print(f"Decoding {p}\nkey={k}") if not FileCypher.decrypt_file(p, k): input() arg = sys.argv len_argv = len(arg) if len_argv > 1: if arg[1] in ['-h', '--help']: print(f"folder/File rows Cypher_size") quit() a = arg[1] else: a = input('File/Folder >') if not path.exists(a): print(f"Path doesn't exist: {a}") quit() key = input('Key >') if path.isfile(a): a, b = path.splitext(a) if b != '.ncy': print(f'File must be .ncy but it is {b}') decode(a, key) elif path.isdir(a): if len_argv > 2: row = arg[2] else: row = input('Number of rows >') if len_argv > 3: chunk = arg[3] else: chunk = input('Cypher size >') encode(a, int(row), key, int(chunk)) else: print("Unknown path")
[ "os.path.exists", "NoviCypher.FileCypher.decrypt_file", "os.path.splitext", "os.path.isfile", "os.path.isdir", "NoviCypher.FileCypher" ]
[((588, 602), 'os.path.isfile', 'path.isfile', (['a'], {}), '(a)\n', (599, 602), False, 'from os import path\n'), ((167, 204), 'NoviCypher.FileCypher', 'FileCypher', (['p'], {'rows': 'r', 'chunk': 'c', 'key': 'k'}), '(p, rows=r, chunk=c, key=k)\n', (177, 204), False, 'from NoviCypher import FileCypher\n'), ((505, 519), 'os.path.exists', 'path.exists', (['a'], {}), '(a)\n', (516, 519), False, 'from os import path\n'), ((612, 628), 'os.path.splitext', 'path.splitext', (['a'], {}), '(a)\n', (625, 628), False, 'from os import path\n'), ((711, 724), 'os.path.isdir', 'path.isdir', (['a'], {}), '(a)\n', (721, 724), False, 'from os import path\n'), ((275, 304), 'NoviCypher.FileCypher.decrypt_file', 'FileCypher.decrypt_file', (['p', 'k'], {}), '(p, k)\n', (298, 304), False, 'from NoviCypher import FileCypher\n')]
from typing import List import cv2 from fellbeast.bounding_box import BoundingBox from fellbeast.configurations import CHECK_FOR_NEW_FACE_FREQUENCY from fellbeast.utils import get_closest_coordinate OPENCV_OBJECT_TRACKERS = { "csrt": cv2.TrackerCSRT_create, "kcf": cv2.TrackerKCF_create, "boosting": cv2.TrackerBoosting_create, "mil": cv2.TrackerMIL_create, "tld": cv2.TrackerTLD_create, "medianflow": cv2.TrackerMedianFlow_create, "mosse": cv2.TrackerMOSSE_create } class BaseObjectTracker(object): bounding_box = None def init(self, _, bounding_box): raise NotImplementedError def update_tracker(self, _): raise NotImplementedError class ObjectTracker(BaseObjectTracker): def __init__(self): self.tracker = cv2.TrackerCSRT_create() def init(self, frame, bounding_box): self.tracker.init(frame, bounding_box.tracker_format) def update_tracker(self, frame): success, bounding_box = self.tracker.update(frame) return success, BoundingBox.from_tracker(*bounding_box.astype(int)) class MultipleObjectTracker(BaseObjectTracker): def __init__(self, logger): self.logger = logger self.tracker = None self.lost_tracking = True self.bounding_boxes = list() self.objects_data = dict() def init(self, frame, bounding_boxes: List[BoundingBox], tracker_type='medianflow'): self.logger.info('Object Tracker initialized') self.tracker = cv2.MultiTracker_create() self.bounding_boxes = bounding_boxes for bounding_box in bounding_boxes: self.tracker.add(OPENCV_OBJECT_TRACKERS[tracker_type](), frame, bounding_box.tracker_format) def update_tracker(self, frame): success, tracker_bounding_boxes = self.tracker.update(frame) bounding_boxes = [BoundingBox.from_tracker(*bounding_box.astype(int)) for bounding_box in tracker_bounding_boxes] self.bounding_boxes = bounding_boxes return success, bounding_boxes def track_faces(self, frame, frame_number, camera): scan_for_new_faces = frame_number % CHECK_FOR_NEW_FACE_FREQUENCY == 0 # Initial face detection if self.lost_tracking: self.logger.info('Object Tracker - Detecting Faces') # Getting all faces and trying to recognise them self.bounding_boxes = camera.face_detector.detect(frame, method='deepface') # If faces where detected they are recognized if len(self.bounding_boxes) > 0: self.objects_data = {face_bounding_box.bounding_box_center: { 'name': camera.face_recognition.find_face_in_encodings(image=frame, face_bounding_box=face_bounding_box), 'bounding_box': face_bounding_box} for face_bounding_box in self.bounding_boxes} self.init(frame, bounding_boxes=self.bounding_boxes) self.lost_tracking = False # Periodic scanning for new faces elif scan_for_new_faces: new_faces_bounding_box = camera.face_detector.detect(frame, method='deepface') # If there are new faces setting the lost_tracking indicator to True if len(new_faces_bounding_box) > len(self.bounding_boxes): self.lost_tracking = True # Updating tracker with new frame else: (success, faces_bounding_boxes) = self.update_tracker(frame) old_coordinates = list(self.objects_data.keys()) updated_objects_data = {face_bounding_box.bounding_box_center: { 'name': self.objects_data[get_closest_coordinate(face_bounding_box.bounding_box_center, old_coordinates)]['name'], 'bounding_box': face_bounding_box} for face_bounding_box in faces_bounding_boxes} self.objects_data = updated_objects_data self.lost_tracking = not success return self.objects_data
[ "fellbeast.utils.get_closest_coordinate", "cv2.MultiTracker_create", "cv2.TrackerCSRT_create" ]
[((786, 810), 'cv2.TrackerCSRT_create', 'cv2.TrackerCSRT_create', ([], {}), '()\n', (808, 810), False, 'import cv2\n'), ((1501, 1526), 'cv2.MultiTracker_create', 'cv2.MultiTracker_create', ([], {}), '()\n', (1524, 1526), False, 'import cv2\n'), ((3766, 3844), 'fellbeast.utils.get_closest_coordinate', 'get_closest_coordinate', (['face_bounding_box.bounding_box_center', 'old_coordinates'], {}), '(face_bounding_box.bounding_box_center, old_coordinates)\n', (3788, 3844), False, 'from fellbeast.utils import get_closest_coordinate\n')]
"""Views for Zinnia channels""" from django.views.generic.list import ListView from zinnia.models.entry import Entry from zinnia.settings import PAGINATION class EntryChannel(ListView): """View for displaying a custom selection of entries based on a search pattern, useful for SEO/SMO pages""" query = '' paginate_by = PAGINATION def get_queryset(self): """Override the get_queryset method to do the search""" return Entry.published.search(self.query)
[ "zinnia.models.entry.Entry.published.search" ]
[((457, 491), 'zinnia.models.entry.Entry.published.search', 'Entry.published.search', (['self.query'], {}), '(self.query)\n', (479, 491), False, 'from zinnia.models.entry import Entry\n')]
from bge import logic, events, render from mathutils import Vector class mouseScroll: def __init__ (self, cont): #get Dependencies self.cont = cont self.camera = cont.owner self.mouse = logic.mouse x = render.getWindowWidth()//2 y = render.getWindowHeight()//2 self.screen_center = (x, y) render.setMousePosition(x + 1, y + 1) #show Mouse render.showMouse(1) def main (self): #check for individual Mouse Position on Screen if self.mouse.position[1] <= 0.0: #print("Scroll Screen Foward") self._scrollY(0.5) if self.mouse.position[1] >= 1: #print("Scroll Screen Down") self._scrollY(-0.5) if self.mouse.position[0] <= 0.0: #print("Scroll Screen Left") self._scrollX(-0.5) if self.mouse.position[0] >= 1: #print("Scroll Screen Right") self._scrollX(0.5) def _scrollY (self, dir): #move Camera in Y axis self.camera.position.y += dir def _scrollX (self, dir): #move Camera in X axis self.camera.position.x += dir class keyboardScroll: def __init__(self, cont): #get Dependencies self.cont = cont self.camera = cont.owner self.keyboard = logic.keyboard def main (self): #check for individual Keyboard Events if self.keyboard.events[events.UPARROWKEY] == logic.KX_INPUT_ACTIVE: #print("Activate Forward!") self._scrollY(0.1) if self.keyboard.events[events.DOWNARROWKEY] == logic.KX_INPUT_ACTIVE: #print("Activate Backward!") self._scrollY(-0.1) if self.keyboard.events[events.LEFTARROWKEY] == logic.KX_INPUT_ACTIVE: #print("Activate Left!") self._scrollX(-0.1) if self.keyboard.events[events.RIGHTARROWKEY] == logic.KX_INPUT_ACTIVE: #print("Activate Right!") self._scrollX(0.1) def _scrollY (self, dir): #move Camera in Y axis self.camera.position.y += dir def _scrollX (self, dir): #move Camera in X axis self.camera.position.x += dir
[ "bge.render.showMouse", "bge.render.getWindowHeight", "bge.render.setMousePosition", "bge.render.getWindowWidth" ]
[((382, 419), 'bge.render.setMousePosition', 'render.setMousePosition', (['(x + 1)', '(y + 1)'], {}), '(x + 1, y + 1)\n', (405, 419), False, 'from bge import logic, events, render\n'), ((448, 467), 'bge.render.showMouse', 'render.showMouse', (['(1)'], {}), '(1)\n', (464, 467), False, 'from bge import logic, events, render\n'), ((271, 294), 'bge.render.getWindowWidth', 'render.getWindowWidth', ([], {}), '()\n', (292, 294), False, 'from bge import logic, events, render\n'), ((310, 334), 'bge.render.getWindowHeight', 'render.getWindowHeight', ([], {}), '()\n', (332, 334), False, 'from bge import logic, events, render\n')]
# Copyright 2020 The Magenta Authors. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Defines the SVGDecoder model.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function import copy from magenta.models.image_stylization import ops # from magenta.models.svg_vae import image_vae_joint from magenta.models.svg_vae import svg_decoder_loss_joint from tensor2tensor.layers import common_hparams from tensor2tensor.layers import common_layers from tensor2tensor.utils import registry from tensor2tensor.utils import t2t_model from tensor2tensor.utils import trainer_lib import tensorflow.compat.v1 as tf from tensorflow.contrib import rnn as contrib_rnn import numpy as np import tensorflow_probability as tfp tfd = tfp.distributions @registry.register_model class SVGDecoder(t2t_model.T2TModel): """Defines the SVGDecoder model.""" def body(self, features): if self._hparams.initializer == 'orthogonal': raise ValueError('LSTM models fail with orthogonal initializer.') train = self._hparams.mode == tf.estimator.ModeKeys.TRAIN return self.render2cmd_v3_internal(features, self._hparams, train) def loss(self, logits, features): # logits should be dict with 'outputs', which is image. targets = tf.reshape(features['targets'], [-1, 64, 64, 1]) weights = common_layers.weights_all(targets) loss_num = tf.pow(logits - targets, 2) return tf.reduce_sum(loss_num * weights), tf.reduce_sum(weights) def vae_internal(self, features, hparams, train): # inputs and targets should both be images with dims [batch, 64, 64, 1] inputs, targets = features['inputs'], features['targets'] inputs = tf.reshape(inputs, [-1, 64, 64, 1]) targets = tf.reshape(targets, [-1, 64, 64, 1]) clss = features['cls'] with tf.variable_scope('vae_internal', reuse=tf.AUTO_REUSE): # encoder enc_out = self.visual_encoder(inputs, clss, hparams, train) enc_out = tf.reshape(enc_out, [-1, 2 * hparams.bottleneck_bits]) # bottleneck sampled_bottleneck, b_loss = self.bottleneck(enc_out) losses = {'bottleneck_kl': tf.reduce_mean(b_loss)} if 'bottleneck' in features: if common_layers.shape_list(features['bottleneck'])[0] == 0: # return bottleneck for interpolation # set losses['training'] = 0 so top() isn't called on it # potential todo: use losses dict so we have kl_loss here for non stop # gradient models return sampled_bottleneck, {'training': 0.0} else: # we want to use the given bottleneck sampled_bottleneck = features['bottleneck'] # finalize bottleneck unbottleneck = sampled_bottleneck # decoder. dec_out = self.visual_decoder(unbottleneck, clss, hparams) # calculate training loss here lol rec_loss = -dec_out.log_prob(inputs) elbo = tf.reduce_mean(-(b_loss + rec_loss)) losses['rec_loss'] = tf.reduce_mean(rec_loss) losses['training'] = -elbo if (not hasattr(self, 'summarized_imgs') and self._hparams.mode != tf.estimator.ModeKeys.PREDICT): self.summarized_imgs = True with tf.name_scope(None), tf.name_scope('train' if train else 'test'): tf.summary.image('rendered_out', dec_out.mean()) tf.summary.image('rendered_og', inputs) # print(common_layers.shape_list(dec_out.mean())) # print("sample bottle", common_layers.shape_list(sampled_bottleneck)) # print("b_loss", common_layers.shape_list(b_loss)) return sampled_bottleneck, dec_out.mean(), losses def bottleneck(self, x): z_size = self.hparams.bottleneck_bits x_shape = common_layers.shape_list(x) with tf.variable_scope('bottleneck', reuse=tf.AUTO_REUSE): mu = x[..., :self.hparams.bottleneck_bits] if self.hparams.mode != tf.estimator.ModeKeys.TRAIN: return mu, 0.0 # No sampling or kl loss on eval. log_sigma = x[..., self.hparams.bottleneck_bits:] epsilon = tf.random_normal(x_shape[:-1] + [z_size]) z = mu + tf.exp(log_sigma / 2) * epsilon kl = 0.5 * tf.reduce_mean(tf.exp(log_sigma) + tf.square(mu) - 1. - log_sigma, axis=-1) # This is the 'free bits' trick mentioned in Kingma et al. (2016) free_bits = self.hparams.free_bits kl_loss = tf.reduce_mean(tf.maximum(kl - free_bits, 0.0)) return z, kl_loss * self.hparams.kl_beta def visual_encoder(self, inputs, clss, hparams, train): # goes from [batch, 64, 64, 1] to [batch, hidden_size] with tf.variable_scope('visual_encoder', reuse=tf.AUTO_REUSE): ret = inputs clss = tf.reshape(clss, [-1]) # conv layer, followed by instance norm + FiLM ret = tf.layers.Conv2D(hparams.base_depth, 5, 1, padding='SAME', activation=None)(ret) ret = ops.conditional_instance_norm(ret, clss, hparams.num_categories) ret = tf.nn.relu(ret) ret = tf.layers.Conv2D(hparams.base_depth, 5, 2, padding='SAME', activation=None)(ret) ret = ops.conditional_instance_norm(ret, clss, hparams.num_categories) ret = tf.nn.relu(ret) ret = tf.layers.Conv2D(2 * hparams.base_depth, 5, 1, padding='SAME', activation=None)(ret) ret = ops.conditional_instance_norm(ret, clss, hparams.num_categories) ret = tf.nn.relu(ret) ret = tf.layers.Conv2D(2 * hparams.base_depth, 5, 2, padding='SAME', activation=None)(ret) ret = ops.conditional_instance_norm(ret, clss, hparams.num_categories) ret = tf.nn.relu(ret) # new conv layer, to bring shape down ret = tf.layers.Conv2D(2 * hparams.bottleneck_bits, 4, 2, padding='SAME', activation=None)(ret) ret = ops.conditional_instance_norm(ret, clss, hparams.num_categories) ret = tf.nn.relu(ret) # new conv layer, to bring shape down ret = tf.layers.Conv2D(2 * hparams.bottleneck_bits, 4, 2, padding='SAME', activation=None)(ret) ret = ops.conditional_instance_norm(ret, clss, hparams.num_categories) ret = tf.nn.relu(ret) # ret has 1024 ret = tf.layers.flatten(ret) ret = tf.layers.dense(ret, 2 * hparams.bottleneck_bits, activation=None) return ret def visual_decoder(self, bottleneck, clss, hparams): # goes from [batch, bottleneck_bits] to [batch, 64, 64, 1] with tf.variable_scope('visual_decoder', reuse=tf.AUTO_REUSE): # unbottleneck ret = tf.layers.dense(bottleneck, 1024, activation=None) ret = tf.reshape(ret, [-1, 4, 4, 64]) clss = tf.reshape(clss, [-1]) # new deconv to bring shape up ret = tf.layers.Conv2DTranspose(2 * hparams.base_depth, 4, 2, padding='SAME', activation=None)(ret) ret = ops.conditional_instance_norm(ret, clss, hparams.num_categories) ret = tf.nn.relu(ret) # new deconv to bring shape up ret = tf.layers.Conv2DTranspose(2 * hparams.base_depth, 4, 2, padding='SAME', activation=None)(ret) ret = ops.conditional_instance_norm(ret, clss, hparams.num_categories) ret = tf.nn.relu(ret) ret = tf.layers.Conv2DTranspose(2 * hparams.base_depth, 5, padding='SAME', activation=None)(ret) ret = ops.conditional_instance_norm(ret, clss, hparams.num_categories) ret = tf.nn.relu(ret) ret = tf.layers.Conv2DTranspose(2 * hparams.base_depth, 5, 2, padding='SAME', activation=None)(ret) ret = ops.conditional_instance_norm(ret, clss, hparams.num_categories) ret = tf.nn.relu(ret) ret = tf.layers.Conv2DTranspose(hparams.base_depth, 5, padding='SAME', activation=None)(ret) ret = ops.conditional_instance_norm(ret, clss, hparams.num_categories) ret = tf.nn.relu(ret) ret = tf.layers.Conv2DTranspose(hparams.base_depth, 5, 2, padding='SAME', activation=None)(ret) ret = ops.conditional_instance_norm(ret, clss, hparams.num_categories) ret = tf.nn.relu(ret) ret = tf.layers.Conv2DTranspose(hparams.base_depth, 5, padding='SAME', activation=None)(ret) ret = ops.conditional_instance_norm(ret, clss, hparams.num_categories) ret = tf.nn.relu(ret) ret = tf.layers.Conv2D(1, 5, padding='SAME', activation=None)(ret) ret = tfd.Independent(tfd.Bernoulli(logits=ret), reinterpreted_batch_ndims=3, name='image') return ret def pretrained_visual_encoder(self, features, hparams, train): # we want the exact hparams used for training this vv # vae_hparams = trainer_lib.create_hparams( # hparams.vae_hparam_set, hparams.vae_hparams, # data_dir=hparams.vae_data_dir, problem_name=hparams.vae_problem) # go back to root variable scope with tf.variable_scope(tf.VariableScope(tf.AUTO_REUSE, ''), reuse=tf.AUTO_REUSE, auxiliary_name_scope=False): # vae = image_vae_joint.ImageVAE(vae_hparams, mode=self._hparams.mode, # problem_hparams=vae_hparams.problem_hparams) # source image feat vae_features_source = copy.copy(features) vae_features_source['inputs'] = tf.reshape(vae_features_source['source_psr'][:, -1, :], [-1, 64, 64, 1]) vae_features_source['targets'] = vae_features_source['inputs'] vae_features_source['cls'] = vae_features_source['targets_cls'] # we want vae to return bottleneck # vae_features_source['bottleneck'] = tf.zeros((0, 128)) # we want vae return all sampled_bottleneck_source, dec_out_source, losses_source = self.vae_internal(vae_features_source, hparams, train) # vae.initialize_from_ckpt(hparams.vae_ckpt_dir) if tf.executing_eagerly(): sampled_bottleneck_source, dec_out_source, losses_source = self.vae_internal(vae_features_source, hparams, train) # the real input to vae will be features['rendered_targets'] vae_features_target = copy.copy(features) #print('checking shape') # print(vae_features_target['targets_psr']) # input() vae_features_target['inputs'] = tf.reshape(vae_features_target['targets_psr'][:, -1, :], [-1, 64, 64, 1]) vae_features_target['targets'] = vae_features_target['inputs'] vae_features_target['cls'] = vae_features_target['targets_cls'] # we want vae to return bottleneck # vae_features_target['bottleneck'] = tf.zeros((0, 128)) sampled_bottleneck_target, dec_out_target, losses_target = self.vae_internal(vae_features_target, hparams, train) if tf.executing_eagerly(): sampled_bottleneck_target, dec_out_target, losses_target = self.vae_internal(vae_features_target, hparams, train) vae_losses = {} for k in losses_source.keys(): vae_losses[k] = losses_source[k] + losses_target[k] return sampled_bottleneck_target - sampled_bottleneck_source, dec_out_source, dec_out_target, vae_losses def render2cmd_v3_internal(self, features, hparams, train): # inputs and targets are both sequences with # shape = [batch, seq_len, 1, hparams.problem.feature_dim] targets = features['targets'] source = features['source'] losses = {} sampled_bottleneck, dec_out_source, dec_out_target, vae_losses = self.pretrained_visual_encoder(features, hparams, train) losses.update(vae_losses) # if hparams.sg_bottleneck: # sampled_bottleneck = tf.stop_gradient(sampled_bottleneck) with tf.variable_scope('render2cmd_v3_internal'): # override bottleneck, or return it, if requested if 'bottleneck' in features: if common_layers.shape_list(features['bottleneck'])[0] == 0: # return sampled_bottleneck, # set losses['training'] = 0 so self.top() doesn't get called on it return sampled_bottleneck, {'training': 0.0} else: # we want to use the given bottleneck sampled_bottleneck = features['bottleneck'] # finalize bottleneck unbottleneck_dim = hparams.hidden_size * 2 # twice because using LSTM if hparams.twice_decoder: unbottleneck_dim = unbottleneck_dim * 2 dec_initial_state = [] # LSTM encoder _, encoder_output_states = self.lstm_encoder( common_layers.flatten4d3d(source), hparams) # print(features['targets'].shape) # print('run stacking...') # print(sampled_bottleneck.shape) # print(source.shape) # input() for hi in range(hparams.num_hidden_layers): unbottleneck = self.unbottleneck(sampled_bottleneck, unbottleneck_dim, name_append='_{}'.format(hi)) c, h = encoder_output_states[hi] # print(unbottleneck.shape) # print(c.shape, h.shape) # first_dim = common_layers.shape_list(unbottleneck)[0] # print(first_dim) #c = tf.tile(c,[first_dim,1]) #h = tf.tile(h,[first_dim,1]) # input() dec_initial_state.append( tf.nn.rnn_cell.LSTMStateTuple( c=tf.concat([unbottleneck[:, :unbottleneck_dim // 2], c], 1), h=tf.concat([unbottleneck[:, unbottleneck_dim // 2:], h], 1))) dec_initial_state = tuple(dec_initial_state) #print('checkshape dec_initial_state') # print(dec_initial_state) # input() shifted_targets = common_layers.shift_right(targets) # Add 1 to account for the padding added to the left from shift_right targets_length = common_layers.length_from_embedding( shifted_targets) + 1 # LSTM decoder hparams_decoder = copy.copy(hparams) if hparams.twice_decoder: hparams_decoder.hidden_size = 2 * hparams.hidden_size if hparams.mode == tf.estimator.ModeKeys.PREDICT: decoder_outputs, _ = self.lstm_decoder_infer( common_layers.flatten4d3d(shifted_targets), targets_length, hparams_decoder, features['targets_cls'], train, initial_state=dec_initial_state, bottleneck=sampled_bottleneck) else: decoder_outputs, _ = self.lstm_decoder( common_layers.flatten4d3d(shifted_targets), targets_length, hparams_decoder, features['targets_cls'], train, initial_state=dec_initial_state, bottleneck=sampled_bottleneck) ret = tf.expand_dims(decoder_outputs, axis=2) return ret, losses def lstm_decoder_infer(self, inputs, sequence_length, hparams, clss, train, initial_state=None, bottleneck=None): # IN PREDICT MODE, RUN tf.while RNN max_decode_length = 51 batch_size = common_layers.shape_list(inputs)[0] zero_pad, logits_so_far = self.create_initial_input_for_decode(batch_size) layers = contrib_rnn.MultiRNNCell([ self.lstm_cell(hparams, train) for _ in range(hparams.num_hidden_layers) ]) if initial_state is None: raise Exception('initial state should be init from bottleneck!') # append one-hot class to bottleneck, which will be given per step clss = tf.reshape(clss, [-1]) if not hparams.use_cls: clss = tf.zeros_like(clss) if hparams.condition_on_sln: sln = tf.reshape(sequence_length, [-1]) bottleneck = tf.concat((bottleneck, tf.one_hot(clss, hparams.num_categories), tf.one_hot(sln, max_decode_length)), -1) else: bottleneck = tf.concat((bottleneck, tf.one_hot(clss, hparams.num_categories)), -1) def infer_step(logits_so_far, current_hidden): """Inference step of LSTM while loop.""" # unflatten hidden: current_hidden = tuple(tf.nn.rnn_cell.LSTMStateTuple(c=s[0], h=s[1]) for s in current_hidden) # put logits_so_far through top tm = self._problem_hparams.modality['targets'] # need to reuse top params reset_scope = tf.variable_scope(tf.VariableScope(tf.AUTO_REUSE, ''), reuse=tf.AUTO_REUSE, auxiliary_name_scope=False) top_scope = tf.variable_scope('svg_decoder/{}_modality'.format(tm), reuse=tf.AUTO_REUSE) with reset_scope, top_scope: samples_so_far = self.hparams.top['targets']( logits_so_far, None, self.hparams, self.problem_hparams.vocab_size) # append a zero pad to the samples. this effectively shifts the samples # right, but, unlike shift_right, by not removing the last element, we # allow an empty samples_so_far to not be empty after padding samples_so_far = tf.concat([zero_pad, samples_so_far], axis=1) shifted_targets = common_layers.flatten4d3d(samples_so_far) # now take the very last one here, will be the actual input to the rnn shifted_targets = shifted_targets[:, -1:, :] # tile and append the bottleneck to inputs sln_offset = 0 if hparams.condition_on_sln: sln_offset = 51 pre_tile_y = tf.reshape(bottleneck, [common_layers.shape_list(bottleneck)[0], 1, hparams.bottleneck_bits + hparams.num_categories + sln_offset]) overlay_x = tf.tile(pre_tile_y, [1, common_layers.shape_list(shifted_targets)[1], 1]) inputs = tf.concat([shifted_targets, overlay_x], -1) seq_len_batch = tf.ones([common_layers.shape_list(inputs)[0]]) # RUN PRE-LSTM LAYER with tf.variable_scope('pre_decoder', reuse=tf.AUTO_REUSE): inputs = tf.layers.dense(inputs, hparams.hidden_size, name='bottom') inputs = tf.nn.tanh(inputs) # RUN LSTM with tf.variable_scope('lstm_decoder', reuse=tf.AUTO_REUSE): next_step, next_state = tf.nn.dynamic_rnn( layers, inputs, seq_len_batch, initial_state=current_hidden, dtype=tf.float32, time_major=False) next_step = tf.expand_dims(next_step, [1]) logits_so_far = tf.concat([logits_so_far, next_step], 1) #print('concat success') # input() # flatten state next_state = tuple((s.c, s.h) for s in next_state) return logits_so_far, next_state def while_exit_cond(logits_so_far, unused_current_hidden): length = common_layers.shape_list(logits_so_far)[1] return length < max_decode_length # passing state must be flattened: initial_state = tuple([(s.c, s.h) for s in initial_state]) # actually run tf.while: logits, final_state = tf.while_loop( while_exit_cond, infer_step, [logits_so_far, initial_state], shape_invariants=[ tf.TensorShape([None, None, 1, hparams.hidden_size]), tuple([(s[0].get_shape(), s[1].get_shape()) for s in initial_state]), ], back_prop=False, parallel_iterations=1 ) # logits should be returned in 3d mode: logits = common_layers.flatten4d3d(logits) return logits, final_state def lstm_decoder(self, inputs, sequence_length, hparams, clss, train, initial_state=None, bottleneck=None): # NOT IN PREDICT MODE. JUST RUN TEACHER-FORCED RNN: layers = contrib_rnn.MultiRNNCell([ self.lstm_cell(hparams, train) for _ in range(hparams.num_hidden_layers) ]) # append one-hot class to bottleneck, which will be given per step clss = tf.reshape(clss, [-1]) if not hparams.use_cls: clss = tf.zeros_like(clss) if hparams.condition_on_sln: sln = tf.reshape(sequence_length, [-1]) bottleneck = tf.concat((bottleneck, tf.one_hot(clss, hparams.num_categories), tf.one_hot(sln, 51)), -1) else: bottleneck = tf.concat((bottleneck, tf.one_hot(clss, hparams.num_categories)), -1) # tile and append the bottleneck to inputs sln_offset = 0 if hparams.condition_on_sln: sln_offset = 51 pre_tile_y = tf.reshape( bottleneck, [common_layers.shape_list(bottleneck)[0], 1, hparams.bottleneck_bits + hparams.num_categories + sln_offset]) overlay_x = tf.tile( pre_tile_y, [1, common_layers.shape_list(inputs)[1], 1]) inputs = tf.concat([inputs, overlay_x], -1) with tf.variable_scope('pre_decoder', reuse=tf.AUTO_REUSE): inputs = tf.layers.dense( inputs, hparams.hidden_size, name='bottom') inputs = tf.nn.tanh(inputs) # print(inputs) # print(initial_state) # input() with tf.variable_scope('lstm_decoder', reuse=tf.AUTO_REUSE): return tf.nn.dynamic_rnn( layers, inputs, sequence_length, initial_state=initial_state, dtype=tf.float32, time_major=False) def lstm_encoder(self, inputs, hparams): batch_size = common_layers.shape_list(inputs)[0] length = 51 # a 4-layer LSTM cell = tf.nn.rnn_cell.LSTMCell(256, state_is_tuple=True) if hparams.mode != tf.estimator.ModeKeys.PREDICT: cell = tf.nn.rnn_cell.DropoutWrapper( cell=cell, output_keep_prob=0.5) cell1 = tf.nn.rnn_cell.LSTMCell(256, state_is_tuple=True) if hparams.mode != tf.estimator.ModeKeys.PREDICT: cell1 = tf.nn.rnn_cell.DropoutWrapper( cell=cell1, output_keep_prob=0.5) cell2 = tf.nn.rnn_cell.LSTMCell(256, state_is_tuple=True) if hparams.mode != tf.estimator.ModeKeys.PREDICT: cell2 = tf.nn.rnn_cell.DropoutWrapper( cell=cell2, output_keep_prob=0.5) cell3 = tf.nn.rnn_cell.LSTMCell(256, state_is_tuple=True) if hparams.mode != tf.estimator.ModeKeys.PREDICT: cell3 = tf.nn.rnn_cell.DropoutWrapper( cell=cell3, output_keep_prob=0.5) stack = tf.nn.rnn_cell.MultiRNNCell( [cell, cell1, cell2, cell3], state_is_tuple=True) initial_state = stack.zero_state(batch_size, dtype=tf.float32) # a two layer LSTM le_output, le_output_states = tf.nn.dynamic_rnn( cell=stack, inputs=inputs, sequence_length=tf.fill([batch_size], length), initial_state=initial_state, dtype=tf.float32, time_major=False ) #print('check lstm encoder') # print(le_output_states) # input() return le_output, le_output_states def lstm_cell(self, hparams, train): keep_prob = 1.0 - hparams.rec_dropout * tf.to_float(train) recurrent_dropout_cell = contrib_rnn.LayerNormBasicLSTMCell( hparams.hidden_size + 256, layer_norm=hparams.layer_norm, dropout_keep_prob=keep_prob) if hparams.ff_dropout: return contrib_rnn.DropoutWrapper( recurrent_dropout_cell, input_keep_prob=keep_prob) return recurrent_dropout_cell def unbottleneck(self, x, res_size, reuse=tf.AUTO_REUSE, name_append=''): with tf.variable_scope('unbottleneck{}'.format(name_append), reuse=reuse): x = tf.layers.dense(x, res_size, name='dense', activation='tanh') return x def create_initial_input_for_decode(self, batch_size): # Create an initial output tensor. This will be passed # to the infer_step, which adds one timestep at every iteration. dim = self._problem_hparams.vocab_size['targets'] hdim = self._hparams.hidden_size + 256 initial_output = tf.zeros((batch_size, 0, 1, hdim), dtype=tf.float32) zero_pad = tf.zeros((batch_size, 1, 1, dim), dtype=tf.float32) # Hack: foldl complains when the output shape is less specified than the # input shape, so we confuse it about the input shape. initial_output = tf.slice(initial_output, [0, 0, 0, 0], common_layers.shape_list(initial_output)) zero_pad = tf.slice(zero_pad, [0, 0, 0, 0], common_layers.shape_list(zero_pad)) return zero_pad, initial_output def _greedy_infer(self, features, extra_decode_length, use_tpu=False): # extra_decode_length is set to 0, unused. infer_features = copy.copy(features) if 'targets' not in infer_features: infer_features['targets'] = infer_features['infer_targets'] logits, losses = self(infer_features) # pylint: disable=not-callable return { 'outputs': logits, 'scores': None, 'logits': logits, 'losses': losses, } @registry.register_hparams def svg_decoder(): """Basic hparams for SVG decoder.""" hparams = common_hparams.basic_params1() hparams.daisy_chain_variables = False hparams.batch_size = 128 hparams.hidden_size = 1024 hparams.num_hidden_layers = 2 hparams.initializer = 'uniform_unit_scaling' hparams.initializer_gain = 1.0 hparams.weight_decay = 0.0 hparams.num_hidden_layers = 4 hparams.force_full_predict = True hparams.dropout = 0.5 hparams.learning_rate_warmup_steps = 100000 # LSTM-specific hparams hparams.add_hparam('vocab_size', None) # VAE params hparams.add_hparam('base_depth', 32) hparams.add_hparam('bottleneck_bits', 32) # loss hparams hparams.add_hparam('kl_beta', 300) hparams.add_hparam('free_bits_div', 4) hparams.add_hparam('free_bits', 0.15) # loss params hparams.add_hparam('soft_k', 10) hparams.add_hparam('mdn_k', 1) # params required by LayerNormLSTMCell, for us to just use recurrent dropout hparams.add_hparam('layer_norm', False) hparams.add_hparam('ff_dropout', True) hparams.add_hparam('rec_dropout', 0.3) # Decode architecture hparams hparams.add_hparam('twice_decoder', False) hparams.add_hparam('sg_bottleneck', False) hparams.add_hparam('condition_on_sln', False) hparams.add_hparam('use_cls', True) # MDN loss hparams hparams.add_hparam('num_mixture', 50) hparams.add_hparam('mix_temperature', 0.0001) hparams.add_hparam('gauss_temperature', 0.0001) hparams.add_hparam('dont_reduce_loss', False) # VAE meta hparams (to load image encoder) # hparams.add_hparam('vae_ckpt_dir', '') # hparams.add_hparam('vae_hparams', 'base_depth=32, hidden_size=32') # hparams.add_hparam('vae_data_dir', '') # hparams.add_hparam('vae_hparam_set', 'image_vae') # hparams.add_hparam('vae_problem', 'glyph_azzn_problem') # data format hparams hparams.add_hparam('num_categories', 62) # problem hparams (required, don't modify) hparams.add_hparam('absolute', False) hparams.add_hparam('just_render', False) hparams.add_hparam('plus_render', False) # modality hparams hparams.bottom = { 'inputs': svg_decoder_loss_joint.real_svg_bottom, 'targets': svg_decoder_loss_joint.real_svg_bottom, } hparams.top = {'targets': svg_decoder_loss_joint.real_svg_top} hparams.loss = {'targets': svg_decoder_loss_joint.real_svg_loss} return hparams
[ "tensorflow.compat.v1.one_hot", "tensor2tensor.layers.common_layers.length_from_embedding", "tensorflow.compat.v1.exp", "tensor2tensor.layers.common_layers.shape_list", "tensorflow.compat.v1.maximum", "tensorflow.compat.v1.concat", "tensorflow.compat.v1.zeros", "tensorflow.compat.v1.nn.rnn_cell.MultiR...
[((27888, 27918), 'tensor2tensor.layers.common_hparams.basic_params1', 'common_hparams.basic_params1', ([], {}), '()\n', (27916, 27918), False, 'from tensor2tensor.layers import common_hparams\n'), ((1819, 1867), 'tensorflow.compat.v1.reshape', 'tf.reshape', (["features['targets']", '[-1, 64, 64, 1]'], {}), "(features['targets'], [-1, 64, 64, 1])\n", (1829, 1867), True, 'import tensorflow.compat.v1 as tf\n'), ((1886, 1920), 'tensor2tensor.layers.common_layers.weights_all', 'common_layers.weights_all', (['targets'], {}), '(targets)\n', (1911, 1920), False, 'from tensor2tensor.layers import common_layers\n'), ((1940, 1967), 'tensorflow.compat.v1.pow', 'tf.pow', (['(logits - targets)', '(2)'], {}), '(logits - targets, 2)\n', (1946, 1967), True, 'import tensorflow.compat.v1 as tf\n'), ((2259, 2294), 'tensorflow.compat.v1.reshape', 'tf.reshape', (['inputs', '[-1, 64, 64, 1]'], {}), '(inputs, [-1, 64, 64, 1])\n', (2269, 2294), True, 'import tensorflow.compat.v1 as tf\n'), ((2313, 2349), 'tensorflow.compat.v1.reshape', 'tf.reshape', (['targets', '[-1, 64, 64, 1]'], {}), '(targets, [-1, 64, 64, 1])\n', (2323, 2349), True, 'import tensorflow.compat.v1 as tf\n'), ((4537, 4564), 'tensor2tensor.layers.common_layers.shape_list', 'common_layers.shape_list', (['x'], {}), '(x)\n', (4561, 4564), False, 'from tensor2tensor.layers import common_layers\n'), ((17605, 17627), 'tensorflow.compat.v1.reshape', 'tf.reshape', (['clss', '[-1]'], {}), '(clss, [-1])\n', (17615, 17627), True, 'import tensorflow.compat.v1 as tf\n'), ((21964, 21997), 'tensor2tensor.layers.common_layers.flatten4d3d', 'common_layers.flatten4d3d', (['logits'], {}), '(logits)\n', (21989, 21997), False, 'from tensor2tensor.layers import common_layers\n'), ((22459, 22481), 'tensorflow.compat.v1.reshape', 'tf.reshape', (['clss', '[-1]'], {}), '(clss, [-1])\n', (22469, 22481), True, 'import tensorflow.compat.v1 as tf\n'), ((23421, 23455), 'tensorflow.compat.v1.concat', 'tf.concat', (['[inputs, overlay_x]', '(-1)'], {}), '([inputs, overlay_x], -1)\n', (23430, 23455), True, 'import tensorflow.compat.v1 as tf\n'), ((24136, 24185), 'tensorflow.compat.v1.nn.rnn_cell.LSTMCell', 'tf.nn.rnn_cell.LSTMCell', (['(256)'], {'state_is_tuple': '(True)'}), '(256, state_is_tuple=True)\n', (24159, 24185), True, 'import tensorflow.compat.v1 as tf\n'), ((24360, 24409), 'tensorflow.compat.v1.nn.rnn_cell.LSTMCell', 'tf.nn.rnn_cell.LSTMCell', (['(256)'], {'state_is_tuple': '(True)'}), '(256, state_is_tuple=True)\n', (24383, 24409), True, 'import tensorflow.compat.v1 as tf\n'), ((24586, 24635), 'tensorflow.compat.v1.nn.rnn_cell.LSTMCell', 'tf.nn.rnn_cell.LSTMCell', (['(256)'], {'state_is_tuple': '(True)'}), '(256, state_is_tuple=True)\n', (24609, 24635), True, 'import tensorflow.compat.v1 as tf\n'), ((24812, 24861), 'tensorflow.compat.v1.nn.rnn_cell.LSTMCell', 'tf.nn.rnn_cell.LSTMCell', (['(256)'], {'state_is_tuple': '(True)'}), '(256, state_is_tuple=True)\n', (24835, 24861), True, 'import tensorflow.compat.v1 as tf\n'), ((25038, 25115), 'tensorflow.compat.v1.nn.rnn_cell.MultiRNNCell', 'tf.nn.rnn_cell.MultiRNNCell', (['[cell, cell1, cell2, cell3]'], {'state_is_tuple': '(True)'}), '([cell, cell1, cell2, cell3], state_is_tuple=True)\n', (25065, 25115), True, 'import tensorflow.compat.v1 as tf\n'), ((25779, 25905), 'tensorflow.contrib.rnn.LayerNormBasicLSTMCell', 'contrib_rnn.LayerNormBasicLSTMCell', (['(hparams.hidden_size + 256)'], {'layer_norm': 'hparams.layer_norm', 'dropout_keep_prob': 'keep_prob'}), '(hparams.hidden_size + 256, layer_norm=\n hparams.layer_norm, dropout_keep_prob=keep_prob)\n', (25813, 25905), True, 'from tensorflow.contrib import rnn as contrib_rnn\n'), ((26709, 26761), 'tensorflow.compat.v1.zeros', 'tf.zeros', (['(batch_size, 0, 1, hdim)'], {'dtype': 'tf.float32'}), '((batch_size, 0, 1, hdim), dtype=tf.float32)\n', (26717, 26761), True, 'import tensorflow.compat.v1 as tf\n'), ((26781, 26832), 'tensorflow.compat.v1.zeros', 'tf.zeros', (['(batch_size, 1, 1, dim)'], {'dtype': 'tf.float32'}), '((batch_size, 1, 1, dim), dtype=tf.float32)\n', (26789, 26832), True, 'import tensorflow.compat.v1 as tf\n'), ((27425, 27444), 'copy.copy', 'copy.copy', (['features'], {}), '(features)\n', (27434, 27444), False, 'import copy\n'), ((1983, 2016), 'tensorflow.compat.v1.reduce_sum', 'tf.reduce_sum', (['(loss_num * weights)'], {}), '(loss_num * weights)\n', (1996, 2016), True, 'import tensorflow.compat.v1 as tf\n'), ((2018, 2040), 'tensorflow.compat.v1.reduce_sum', 'tf.reduce_sum', (['weights'], {}), '(weights)\n', (2031, 2040), True, 'import tensorflow.compat.v1 as tf\n'), ((2396, 2450), 'tensorflow.compat.v1.variable_scope', 'tf.variable_scope', (['"""vae_internal"""'], {'reuse': 'tf.AUTO_REUSE'}), "('vae_internal', reuse=tf.AUTO_REUSE)\n", (2413, 2450), True, 'import tensorflow.compat.v1 as tf\n'), ((2568, 2622), 'tensorflow.compat.v1.reshape', 'tf.reshape', (['enc_out', '[-1, 2 * hparams.bottleneck_bits]'], {}), '(enc_out, [-1, 2 * hparams.bottleneck_bits])\n', (2578, 2622), True, 'import tensorflow.compat.v1 as tf\n'), ((3662, 3698), 'tensorflow.compat.v1.reduce_mean', 'tf.reduce_mean', (['(-(b_loss + rec_loss))'], {}), '(-(b_loss + rec_loss))\n', (3676, 3698), True, 'import tensorflow.compat.v1 as tf\n'), ((3732, 3756), 'tensorflow.compat.v1.reduce_mean', 'tf.reduce_mean', (['rec_loss'], {}), '(rec_loss)\n', (3746, 3756), True, 'import tensorflow.compat.v1 as tf\n'), ((4578, 4630), 'tensorflow.compat.v1.variable_scope', 'tf.variable_scope', (['"""bottleneck"""'], {'reuse': 'tf.AUTO_REUSE'}), "('bottleneck', reuse=tf.AUTO_REUSE)\n", (4595, 4630), True, 'import tensorflow.compat.v1 as tf\n'), ((4902, 4943), 'tensorflow.compat.v1.random_normal', 'tf.random_normal', (['(x_shape[:-1] + [z_size])'], {}), '(x_shape[:-1] + [z_size])\n', (4918, 4943), True, 'import tensorflow.compat.v1 as tf\n'), ((5477, 5533), 'tensorflow.compat.v1.variable_scope', 'tf.variable_scope', (['"""visual_encoder"""'], {'reuse': 'tf.AUTO_REUSE'}), "('visual_encoder', reuse=tf.AUTO_REUSE)\n", (5494, 5533), True, 'import tensorflow.compat.v1 as tf\n'), ((5579, 5601), 'tensorflow.compat.v1.reshape', 'tf.reshape', (['clss', '[-1]'], {}), '(clss, [-1])\n', (5589, 5601), True, 'import tensorflow.compat.v1 as tf\n'), ((5814, 5878), 'magenta.models.image_stylization.ops.conditional_instance_norm', 'ops.conditional_instance_norm', (['ret', 'clss', 'hparams.num_categories'], {}), '(ret, clss, hparams.num_categories)\n', (5843, 5878), False, 'from magenta.models.image_stylization import ops\n'), ((5897, 5912), 'tensorflow.compat.v1.nn.relu', 'tf.nn.relu', (['ret'], {}), '(ret)\n', (5907, 5912), True, 'import tensorflow.compat.v1 as tf\n'), ((6066, 6130), 'magenta.models.image_stylization.ops.conditional_instance_norm', 'ops.conditional_instance_norm', (['ret', 'clss', 'hparams.num_categories'], {}), '(ret, clss, hparams.num_categories)\n', (6095, 6130), False, 'from magenta.models.image_stylization import ops\n'), ((6149, 6164), 'tensorflow.compat.v1.nn.relu', 'tf.nn.relu', (['ret'], {}), '(ret)\n', (6159, 6164), True, 'import tensorflow.compat.v1 as tf\n'), ((6322, 6386), 'magenta.models.image_stylization.ops.conditional_instance_norm', 'ops.conditional_instance_norm', (['ret', 'clss', 'hparams.num_categories'], {}), '(ret, clss, hparams.num_categories)\n', (6351, 6386), False, 'from magenta.models.image_stylization import ops\n'), ((6405, 6420), 'tensorflow.compat.v1.nn.relu', 'tf.nn.relu', (['ret'], {}), '(ret)\n', (6415, 6420), True, 'import tensorflow.compat.v1 as tf\n'), ((6578, 6642), 'magenta.models.image_stylization.ops.conditional_instance_norm', 'ops.conditional_instance_norm', (['ret', 'clss', 'hparams.num_categories'], {}), '(ret, clss, hparams.num_categories)\n', (6607, 6642), False, 'from magenta.models.image_stylization import ops\n'), ((6661, 6676), 'tensorflow.compat.v1.nn.relu', 'tf.nn.relu', (['ret'], {}), '(ret)\n', (6671, 6676), True, 'import tensorflow.compat.v1 as tf\n'), ((6889, 6953), 'magenta.models.image_stylization.ops.conditional_instance_norm', 'ops.conditional_instance_norm', (['ret', 'clss', 'hparams.num_categories'], {}), '(ret, clss, hparams.num_categories)\n', (6918, 6953), False, 'from magenta.models.image_stylization import ops\n'), ((6972, 6987), 'tensorflow.compat.v1.nn.relu', 'tf.nn.relu', (['ret'], {}), '(ret)\n', (6982, 6987), True, 'import tensorflow.compat.v1 as tf\n'), ((7200, 7264), 'magenta.models.image_stylization.ops.conditional_instance_norm', 'ops.conditional_instance_norm', (['ret', 'clss', 'hparams.num_categories'], {}), '(ret, clss, hparams.num_categories)\n', (7229, 7264), False, 'from magenta.models.image_stylization import ops\n'), ((7283, 7298), 'tensorflow.compat.v1.nn.relu', 'tf.nn.relu', (['ret'], {}), '(ret)\n', (7293, 7298), True, 'import tensorflow.compat.v1 as tf\n'), ((7345, 7367), 'tensorflow.compat.v1.layers.flatten', 'tf.layers.flatten', (['ret'], {}), '(ret)\n', (7362, 7367), True, 'import tensorflow.compat.v1 as tf\n'), ((7386, 7452), 'tensorflow.compat.v1.layers.dense', 'tf.layers.dense', (['ret', '(2 * hparams.bottleneck_bits)'], {'activation': 'None'}), '(ret, 2 * hparams.bottleneck_bits, activation=None)\n', (7401, 7452), True, 'import tensorflow.compat.v1 as tf\n'), ((7611, 7667), 'tensorflow.compat.v1.variable_scope', 'tf.variable_scope', (['"""visual_decoder"""'], {'reuse': 'tf.AUTO_REUSE'}), "('visual_decoder', reuse=tf.AUTO_REUSE)\n", (7628, 7667), True, 'import tensorflow.compat.v1 as tf\n'), ((7714, 7764), 'tensorflow.compat.v1.layers.dense', 'tf.layers.dense', (['bottleneck', '(1024)'], {'activation': 'None'}), '(bottleneck, 1024, activation=None)\n', (7729, 7764), True, 'import tensorflow.compat.v1 as tf\n'), ((7783, 7814), 'tensorflow.compat.v1.reshape', 'tf.reshape', (['ret', '[-1, 4, 4, 64]'], {}), '(ret, [-1, 4, 4, 64])\n', (7793, 7814), True, 'import tensorflow.compat.v1 as tf\n'), ((7834, 7856), 'tensorflow.compat.v1.reshape', 'tf.reshape', (['clss', '[-1]'], {}), '(clss, [-1])\n', (7844, 7856), True, 'import tensorflow.compat.v1 as tf\n'), ((8075, 8139), 'magenta.models.image_stylization.ops.conditional_instance_norm', 'ops.conditional_instance_norm', (['ret', 'clss', 'hparams.num_categories'], {}), '(ret, clss, hparams.num_categories)\n', (8104, 8139), False, 'from magenta.models.image_stylization import ops\n'), ((8158, 8173), 'tensorflow.compat.v1.nn.relu', 'tf.nn.relu', (['ret'], {}), '(ret)\n', (8168, 8173), True, 'import tensorflow.compat.v1 as tf\n'), ((8392, 8456), 'magenta.models.image_stylization.ops.conditional_instance_norm', 'ops.conditional_instance_norm', (['ret', 'clss', 'hparams.num_categories'], {}), '(ret, clss, hparams.num_categories)\n', (8421, 8456), False, 'from magenta.models.image_stylization import ops\n'), ((8475, 8490), 'tensorflow.compat.v1.nn.relu', 'tf.nn.relu', (['ret'], {}), '(ret)\n', (8485, 8490), True, 'import tensorflow.compat.v1 as tf\n'), ((8663, 8727), 'magenta.models.image_stylization.ops.conditional_instance_norm', 'ops.conditional_instance_norm', (['ret', 'clss', 'hparams.num_categories'], {}), '(ret, clss, hparams.num_categories)\n', (8692, 8727), False, 'from magenta.models.image_stylization import ops\n'), ((8746, 8761), 'tensorflow.compat.v1.nn.relu', 'tf.nn.relu', (['ret'], {}), '(ret)\n', (8756, 8761), True, 'import tensorflow.compat.v1 as tf\n'), ((8937, 9001), 'magenta.models.image_stylization.ops.conditional_instance_norm', 'ops.conditional_instance_norm', (['ret', 'clss', 'hparams.num_categories'], {}), '(ret, clss, hparams.num_categories)\n', (8966, 9001), False, 'from magenta.models.image_stylization import ops\n'), ((9020, 9035), 'tensorflow.compat.v1.nn.relu', 'tf.nn.relu', (['ret'], {}), '(ret)\n', (9030, 9035), True, 'import tensorflow.compat.v1 as tf\n'), ((9204, 9268), 'magenta.models.image_stylization.ops.conditional_instance_norm', 'ops.conditional_instance_norm', (['ret', 'clss', 'hparams.num_categories'], {}), '(ret, clss, hparams.num_categories)\n', (9233, 9268), False, 'from magenta.models.image_stylization import ops\n'), ((9287, 9302), 'tensorflow.compat.v1.nn.relu', 'tf.nn.relu', (['ret'], {}), '(ret)\n', (9297, 9302), True, 'import tensorflow.compat.v1 as tf\n'), ((9474, 9538), 'magenta.models.image_stylization.ops.conditional_instance_norm', 'ops.conditional_instance_norm', (['ret', 'clss', 'hparams.num_categories'], {}), '(ret, clss, hparams.num_categories)\n', (9503, 9538), False, 'from magenta.models.image_stylization import ops\n'), ((9557, 9572), 'tensorflow.compat.v1.nn.relu', 'tf.nn.relu', (['ret'], {}), '(ret)\n', (9567, 9572), True, 'import tensorflow.compat.v1 as tf\n'), ((9741, 9805), 'magenta.models.image_stylization.ops.conditional_instance_norm', 'ops.conditional_instance_norm', (['ret', 'clss', 'hparams.num_categories'], {}), '(ret, clss, hparams.num_categories)\n', (9770, 9805), False, 'from magenta.models.image_stylization import ops\n'), ((9824, 9839), 'tensorflow.compat.v1.nn.relu', 'tf.nn.relu', (['ret'], {}), '(ret)\n', (9834, 9839), True, 'import tensorflow.compat.v1 as tf\n'), ((10862, 10881), 'copy.copy', 'copy.copy', (['features'], {}), '(features)\n', (10871, 10881), False, 'import copy\n'), ((10926, 10998), 'tensorflow.compat.v1.reshape', 'tf.reshape', (["vae_features_source['source_psr'][:, -1, :]", '[-1, 64, 64, 1]'], {}), "(vae_features_source['source_psr'][:, -1, :], [-1, 64, 64, 1])\n", (10936, 10998), True, 'import tensorflow.compat.v1 as tf\n'), ((11556, 11578), 'tensorflow.compat.v1.executing_eagerly', 'tf.executing_eagerly', ([], {}), '()\n', (11576, 11578), True, 'import tensorflow.compat.v1 as tf\n'), ((11818, 11837), 'copy.copy', 'copy.copy', (['features'], {}), '(features)\n', (11827, 11837), False, 'import copy\n'), ((11997, 12070), 'tensorflow.compat.v1.reshape', 'tf.reshape', (["vae_features_target['targets_psr'][:, -1, :]", '[-1, 64, 64, 1]'], {}), "(vae_features_target['targets_psr'][:, -1, :], [-1, 64, 64, 1])\n", (12007, 12070), True, 'import tensorflow.compat.v1 as tf\n'), ((12529, 12551), 'tensorflow.compat.v1.executing_eagerly', 'tf.executing_eagerly', ([], {}), '()\n', (12549, 12551), True, 'import tensorflow.compat.v1 as tf\n'), ((13503, 13546), 'tensorflow.compat.v1.variable_scope', 'tf.variable_scope', (['"""render2cmd_v3_internal"""'], {}), "('render2cmd_v3_internal')\n", (13520, 13546), True, 'import tensorflow.compat.v1 as tf\n'), ((15704, 15738), 'tensor2tensor.layers.common_layers.shift_right', 'common_layers.shift_right', (['targets'], {}), '(targets)\n', (15729, 15738), False, 'from tensor2tensor.layers import common_layers\n'), ((15982, 16000), 'copy.copy', 'copy.copy', (['hparams'], {}), '(hparams)\n', (15991, 16000), False, 'import copy\n'), ((16833, 16872), 'tensorflow.compat.v1.expand_dims', 'tf.expand_dims', (['decoder_outputs'], {'axis': '(2)'}), '(decoder_outputs, axis=2)\n', (16847, 16872), True, 'import tensorflow.compat.v1 as tf\n'), ((17142, 17174), 'tensor2tensor.layers.common_layers.shape_list', 'common_layers.shape_list', (['inputs'], {}), '(inputs)\n', (17166, 17174), False, 'from tensor2tensor.layers import common_layers\n'), ((17679, 17698), 'tensorflow.compat.v1.zeros_like', 'tf.zeros_like', (['clss'], {}), '(clss)\n', (17692, 17698), True, 'import tensorflow.compat.v1 as tf\n'), ((17754, 17787), 'tensorflow.compat.v1.reshape', 'tf.reshape', (['sequence_length', '[-1]'], {}), '(sequence_length, [-1])\n', (17764, 17787), True, 'import tensorflow.compat.v1 as tf\n'), ((19383, 19428), 'tensorflow.compat.v1.concat', 'tf.concat', (['[zero_pad, samples_so_far]'], {'axis': '(1)'}), '([zero_pad, samples_so_far], axis=1)\n', (19392, 19428), True, 'import tensorflow.compat.v1 as tf\n'), ((19459, 19500), 'tensor2tensor.layers.common_layers.flatten4d3d', 'common_layers.flatten4d3d', (['samples_so_far'], {}), '(samples_so_far)\n', (19484, 19500), False, 'from tensor2tensor.layers import common_layers\n'), ((20177, 20220), 'tensorflow.compat.v1.concat', 'tf.concat', (['[shifted_targets, overlay_x]', '(-1)'], {}), '([shifted_targets, overlay_x], -1)\n', (20186, 20220), True, 'import tensorflow.compat.v1 as tf\n'), ((20850, 20880), 'tensorflow.compat.v1.expand_dims', 'tf.expand_dims', (['next_step', '[1]'], {}), '(next_step, [1])\n', (20864, 20880), True, 'import tensorflow.compat.v1 as tf\n'), ((20910, 20950), 'tensorflow.compat.v1.concat', 'tf.concat', (['[logits_so_far, next_step]', '(1)'], {}), '([logits_so_far, next_step], 1)\n', (20919, 20950), True, 'import tensorflow.compat.v1 as tf\n'), ((22533, 22552), 'tensorflow.compat.v1.zeros_like', 'tf.zeros_like', (['clss'], {}), '(clss)\n', (22546, 22552), True, 'import tensorflow.compat.v1 as tf\n'), ((22608, 22641), 'tensorflow.compat.v1.reshape', 'tf.reshape', (['sequence_length', '[-1]'], {}), '(sequence_length, [-1])\n', (22618, 22641), True, 'import tensorflow.compat.v1 as tf\n'), ((23470, 23523), 'tensorflow.compat.v1.variable_scope', 'tf.variable_scope', (['"""pre_decoder"""'], {'reuse': 'tf.AUTO_REUSE'}), "('pre_decoder', reuse=tf.AUTO_REUSE)\n", (23487, 23523), True, 'import tensorflow.compat.v1 as tf\n'), ((23546, 23605), 'tensorflow.compat.v1.layers.dense', 'tf.layers.dense', (['inputs', 'hparams.hidden_size'], {'name': '"""bottom"""'}), "(inputs, hparams.hidden_size, name='bottom')\n", (23561, 23605), True, 'import tensorflow.compat.v1 as tf\n'), ((23644, 23662), 'tensorflow.compat.v1.nn.tanh', 'tf.nn.tanh', (['inputs'], {}), '(inputs)\n', (23654, 23662), True, 'import tensorflow.compat.v1 as tf\n'), ((23749, 23803), 'tensorflow.compat.v1.variable_scope', 'tf.variable_scope', (['"""lstm_decoder"""'], {'reuse': 'tf.AUTO_REUSE'}), "('lstm_decoder', reuse=tf.AUTO_REUSE)\n", (23766, 23803), True, 'import tensorflow.compat.v1 as tf\n'), ((23824, 23944), 'tensorflow.compat.v1.nn.dynamic_rnn', 'tf.nn.dynamic_rnn', (['layers', 'inputs', 'sequence_length'], {'initial_state': 'initial_state', 'dtype': 'tf.float32', 'time_major': '(False)'}), '(layers, inputs, sequence_length, initial_state=\n initial_state, dtype=tf.float32, time_major=False)\n', (23841, 23944), True, 'import tensorflow.compat.v1 as tf\n'), ((24040, 24072), 'tensor2tensor.layers.common_layers.shape_list', 'common_layers.shape_list', (['inputs'], {}), '(inputs)\n', (24064, 24072), False, 'from tensor2tensor.layers import common_layers\n'), ((24263, 24325), 'tensorflow.compat.v1.nn.rnn_cell.DropoutWrapper', 'tf.nn.rnn_cell.DropoutWrapper', ([], {'cell': 'cell', 'output_keep_prob': '(0.5)'}), '(cell=cell, output_keep_prob=0.5)\n', (24292, 24325), True, 'import tensorflow.compat.v1 as tf\n'), ((24488, 24551), 'tensorflow.compat.v1.nn.rnn_cell.DropoutWrapper', 'tf.nn.rnn_cell.DropoutWrapper', ([], {'cell': 'cell1', 'output_keep_prob': '(0.5)'}), '(cell=cell1, output_keep_prob=0.5)\n', (24517, 24551), True, 'import tensorflow.compat.v1 as tf\n'), ((24714, 24777), 'tensorflow.compat.v1.nn.rnn_cell.DropoutWrapper', 'tf.nn.rnn_cell.DropoutWrapper', ([], {'cell': 'cell2', 'output_keep_prob': '(0.5)'}), '(cell=cell2, output_keep_prob=0.5)\n', (24743, 24777), True, 'import tensorflow.compat.v1 as tf\n'), ((24940, 25003), 'tensorflow.compat.v1.nn.rnn_cell.DropoutWrapper', 'tf.nn.rnn_cell.DropoutWrapper', ([], {'cell': 'cell3', 'output_keep_prob': '(0.5)'}), '(cell=cell3, output_keep_prob=0.5)\n', (24969, 25003), True, 'import tensorflow.compat.v1 as tf\n'), ((25989, 26066), 'tensorflow.contrib.rnn.DropoutWrapper', 'contrib_rnn.DropoutWrapper', (['recurrent_dropout_cell'], {'input_keep_prob': 'keep_prob'}), '(recurrent_dropout_cell, input_keep_prob=keep_prob)\n', (26015, 26066), True, 'from tensorflow.contrib import rnn as contrib_rnn\n'), ((26300, 26361), 'tensorflow.compat.v1.layers.dense', 'tf.layers.dense', (['x', 'res_size'], {'name': '"""dense"""', 'activation': '"""tanh"""'}), "(x, res_size, name='dense', activation='tanh')\n", (26315, 26361), True, 'import tensorflow.compat.v1 as tf\n'), ((27075, 27115), 'tensor2tensor.layers.common_layers.shape_list', 'common_layers.shape_list', (['initial_output'], {}), '(initial_output)\n', (27099, 27115), False, 'from tensor2tensor.layers import common_layers\n'), ((27197, 27231), 'tensor2tensor.layers.common_layers.shape_list', 'common_layers.shape_list', (['zero_pad'], {}), '(zero_pad)\n', (27221, 27231), False, 'from tensor2tensor.layers import common_layers\n'), ((2754, 2776), 'tensorflow.compat.v1.reduce_mean', 'tf.reduce_mean', (['b_loss'], {}), '(b_loss)\n', (2768, 2776), True, 'import tensorflow.compat.v1 as tf\n'), ((5258, 5289), 'tensorflow.compat.v1.maximum', 'tf.maximum', (['(kl - free_bits)', '(0.0)'], {}), '(kl - free_bits, 0.0)\n', (5268, 5289), True, 'import tensorflow.compat.v1 as tf\n'), ((5680, 5755), 'tensorflow.compat.v1.layers.Conv2D', 'tf.layers.Conv2D', (['hparams.base_depth', '(5)', '(1)'], {'padding': '"""SAME"""', 'activation': 'None'}), "(hparams.base_depth, 5, 1, padding='SAME', activation=None)\n", (5696, 5755), True, 'import tensorflow.compat.v1 as tf\n'), ((5932, 6007), 'tensorflow.compat.v1.layers.Conv2D', 'tf.layers.Conv2D', (['hparams.base_depth', '(5)', '(2)'], {'padding': '"""SAME"""', 'activation': 'None'}), "(hparams.base_depth, 5, 2, padding='SAME', activation=None)\n", (5948, 6007), True, 'import tensorflow.compat.v1 as tf\n'), ((6184, 6263), 'tensorflow.compat.v1.layers.Conv2D', 'tf.layers.Conv2D', (['(2 * hparams.base_depth)', '(5)', '(1)'], {'padding': '"""SAME"""', 'activation': 'None'}), "(2 * hparams.base_depth, 5, 1, padding='SAME', activation=None)\n", (6200, 6263), True, 'import tensorflow.compat.v1 as tf\n'), ((6440, 6519), 'tensorflow.compat.v1.layers.Conv2D', 'tf.layers.Conv2D', (['(2 * hparams.base_depth)', '(5)', '(2)'], {'padding': '"""SAME"""', 'activation': 'None'}), "(2 * hparams.base_depth, 5, 2, padding='SAME', activation=None)\n", (6456, 6519), True, 'import tensorflow.compat.v1 as tf\n'), ((6746, 6834), 'tensorflow.compat.v1.layers.Conv2D', 'tf.layers.Conv2D', (['(2 * hparams.bottleneck_bits)', '(4)', '(2)'], {'padding': '"""SAME"""', 'activation': 'None'}), "(2 * hparams.bottleneck_bits, 4, 2, padding='SAME',\n activation=None)\n", (6762, 6834), True, 'import tensorflow.compat.v1 as tf\n'), ((7057, 7145), 'tensorflow.compat.v1.layers.Conv2D', 'tf.layers.Conv2D', (['(2 * hparams.bottleneck_bits)', '(4)', '(2)'], {'padding': '"""SAME"""', 'activation': 'None'}), "(2 * hparams.bottleneck_bits, 4, 2, padding='SAME',\n activation=None)\n", (7073, 7145), True, 'import tensorflow.compat.v1 as tf\n'), ((7919, 8011), 'tensorflow.compat.v1.layers.Conv2DTranspose', 'tf.layers.Conv2DTranspose', (['(2 * hparams.base_depth)', '(4)', '(2)'], {'padding': '"""SAME"""', 'activation': 'None'}), "(2 * hparams.base_depth, 4, 2, padding='SAME',\n activation=None)\n", (7944, 8011), True, 'import tensorflow.compat.v1 as tf\n'), ((8236, 8328), 'tensorflow.compat.v1.layers.Conv2DTranspose', 'tf.layers.Conv2DTranspose', (['(2 * hparams.base_depth)', '(4)', '(2)'], {'padding': '"""SAME"""', 'activation': 'None'}), "(2 * hparams.base_depth, 4, 2, padding='SAME',\n activation=None)\n", (8261, 8328), True, 'import tensorflow.compat.v1 as tf\n'), ((8510, 8599), 'tensorflow.compat.v1.layers.Conv2DTranspose', 'tf.layers.Conv2DTranspose', (['(2 * hparams.base_depth)', '(5)'], {'padding': '"""SAME"""', 'activation': 'None'}), "(2 * hparams.base_depth, 5, padding='SAME',\n activation=None)\n", (8535, 8599), True, 'import tensorflow.compat.v1 as tf\n'), ((8781, 8873), 'tensorflow.compat.v1.layers.Conv2DTranspose', 'tf.layers.Conv2DTranspose', (['(2 * hparams.base_depth)', '(5)', '(2)'], {'padding': '"""SAME"""', 'activation': 'None'}), "(2 * hparams.base_depth, 5, 2, padding='SAME',\n activation=None)\n", (8806, 8873), True, 'import tensorflow.compat.v1 as tf\n'), ((9055, 9141), 'tensorflow.compat.v1.layers.Conv2DTranspose', 'tf.layers.Conv2DTranspose', (['hparams.base_depth', '(5)'], {'padding': '"""SAME"""', 'activation': 'None'}), "(hparams.base_depth, 5, padding='SAME', activation\n =None)\n", (9080, 9141), True, 'import tensorflow.compat.v1 as tf\n'), ((9322, 9410), 'tensorflow.compat.v1.layers.Conv2DTranspose', 'tf.layers.Conv2DTranspose', (['hparams.base_depth', '(5)', '(2)'], {'padding': '"""SAME"""', 'activation': 'None'}), "(hparams.base_depth, 5, 2, padding='SAME',\n activation=None)\n", (9347, 9410), True, 'import tensorflow.compat.v1 as tf\n'), ((9592, 9678), 'tensorflow.compat.v1.layers.Conv2DTranspose', 'tf.layers.Conv2DTranspose', (['hparams.base_depth', '(5)'], {'padding': '"""SAME"""', 'activation': 'None'}), "(hparams.base_depth, 5, padding='SAME', activation\n =None)\n", (9617, 9678), True, 'import tensorflow.compat.v1 as tf\n'), ((9859, 9914), 'tensorflow.compat.v1.layers.Conv2D', 'tf.layers.Conv2D', (['(1)', '(5)'], {'padding': '"""SAME"""', 'activation': 'None'}), "(1, 5, padding='SAME', activation=None)\n", (9875, 9914), True, 'import tensorflow.compat.v1 as tf\n'), ((10504, 10539), 'tensorflow.compat.v1.VariableScope', 'tf.VariableScope', (['tf.AUTO_REUSE', '""""""'], {}), "(tf.AUTO_REUSE, '')\n", (10520, 10539), True, 'import tensorflow.compat.v1 as tf\n'), ((14424, 14457), 'tensor2tensor.layers.common_layers.flatten4d3d', 'common_layers.flatten4d3d', (['source'], {}), '(source)\n', (14449, 14457), False, 'from tensor2tensor.layers import common_layers\n'), ((15850, 15902), 'tensor2tensor.layers.common_layers.length_from_embedding', 'common_layers.length_from_embedding', (['shifted_targets'], {}), '(shifted_targets)\n', (15885, 15902), False, 'from tensor2tensor.layers import common_layers\n'), ((18605, 18640), 'tensorflow.compat.v1.VariableScope', 'tf.VariableScope', (['tf.AUTO_REUSE', '""""""'], {}), "(tf.AUTO_REUSE, '')\n", (18621, 18640), True, 'import tensorflow.compat.v1 as tf\n'), ((20348, 20401), 'tensorflow.compat.v1.variable_scope', 'tf.variable_scope', (['"""pre_decoder"""'], {'reuse': 'tf.AUTO_REUSE'}), "('pre_decoder', reuse=tf.AUTO_REUSE)\n", (20365, 20401), True, 'import tensorflow.compat.v1 as tf\n'), ((20428, 20487), 'tensorflow.compat.v1.layers.dense', 'tf.layers.dense', (['inputs', 'hparams.hidden_size'], {'name': '"""bottom"""'}), "(inputs, hparams.hidden_size, name='bottom')\n", (20443, 20487), True, 'import tensorflow.compat.v1 as tf\n'), ((20513, 20531), 'tensorflow.compat.v1.nn.tanh', 'tf.nn.tanh', (['inputs'], {}), '(inputs)\n', (20523, 20531), True, 'import tensorflow.compat.v1 as tf\n'), ((20573, 20627), 'tensorflow.compat.v1.variable_scope', 'tf.variable_scope', (['"""lstm_decoder"""'], {'reuse': 'tf.AUTO_REUSE'}), "('lstm_decoder', reuse=tf.AUTO_REUSE)\n", (20590, 20627), True, 'import tensorflow.compat.v1 as tf\n'), ((20669, 20788), 'tensorflow.compat.v1.nn.dynamic_rnn', 'tf.nn.dynamic_rnn', (['layers', 'inputs', 'seq_len_batch'], {'initial_state': 'current_hidden', 'dtype': 'tf.float32', 'time_major': '(False)'}), '(layers, inputs, seq_len_batch, initial_state=\n current_hidden, dtype=tf.float32, time_major=False)\n', (20686, 20788), True, 'import tensorflow.compat.v1 as tf\n'), ((21236, 21275), 'tensor2tensor.layers.common_layers.shape_list', 'common_layers.shape_list', (['logits_so_far'], {}), '(logits_so_far)\n', (21260, 21275), False, 'from tensor2tensor.layers import common_layers\n'), ((25363, 25392), 'tensorflow.compat.v1.fill', 'tf.fill', (['[batch_size]', 'length'], {}), '([batch_size], length)\n', (25370, 25392), True, 'import tensorflow.compat.v1 as tf\n'), ((25726, 25744), 'tensorflow.compat.v1.to_float', 'tf.to_float', (['train'], {}), '(train)\n', (25737, 25744), True, 'import tensorflow.compat.v1 as tf\n'), ((3993, 4012), 'tensorflow.compat.v1.name_scope', 'tf.name_scope', (['None'], {}), '(None)\n', (4006, 4012), True, 'import tensorflow.compat.v1 as tf\n'), ((4014, 4057), 'tensorflow.compat.v1.name_scope', 'tf.name_scope', (["('train' if train else 'test')"], {}), "('train' if train else 'test')\n", (4027, 4057), True, 'import tensorflow.compat.v1 as tf\n'), ((4148, 4187), 'tensorflow.compat.v1.summary.image', 'tf.summary.image', (['"""rendered_og"""', 'inputs'], {}), "('rendered_og', inputs)\n", (4164, 4187), True, 'import tensorflow.compat.v1 as tf\n'), ((4965, 4986), 'tensorflow.compat.v1.exp', 'tf.exp', (['(log_sigma / 2)'], {}), '(log_sigma / 2)\n', (4971, 4986), True, 'import tensorflow.compat.v1 as tf\n'), ((16254, 16296), 'tensor2tensor.layers.common_layers.flatten4d3d', 'common_layers.flatten4d3d', (['shifted_targets'], {}), '(shifted_targets)\n', (16279, 16296), False, 'from tensor2tensor.layers import common_layers\n'), ((16581, 16623), 'tensor2tensor.layers.common_layers.flatten4d3d', 'common_layers.flatten4d3d', (['shifted_targets'], {}), '(shifted_targets)\n', (16606, 16623), False, 'from tensor2tensor.layers import common_layers\n'), ((17872, 17912), 'tensorflow.compat.v1.one_hot', 'tf.one_hot', (['clss', 'hparams.num_categories'], {}), '(clss, hparams.num_categories)\n', (17882, 17912), True, 'import tensorflow.compat.v1 as tf\n'), ((17950, 17984), 'tensorflow.compat.v1.one_hot', 'tf.one_hot', (['sln', 'max_decode_length'], {}), '(sln, max_decode_length)\n', (17960, 17984), True, 'import tensorflow.compat.v1 as tf\n'), ((18089, 18129), 'tensorflow.compat.v1.one_hot', 'tf.one_hot', (['clss', 'hparams.num_categories'], {}), '(clss, hparams.num_categories)\n', (18099, 18129), True, 'import tensorflow.compat.v1 as tf\n'), ((18312, 18357), 'tensorflow.compat.v1.nn.rnn_cell.LSTMStateTuple', 'tf.nn.rnn_cell.LSTMStateTuple', ([], {'c': 's[0]', 'h': 's[1]'}), '(c=s[0], h=s[1])\n', (18341, 18357), True, 'import tensorflow.compat.v1 as tf\n'), ((21647, 21699), 'tensorflow.compat.v1.TensorShape', 'tf.TensorShape', (['[None, None, 1, hparams.hidden_size]'], {}), '([None, None, 1, hparams.hidden_size])\n', (21661, 21699), True, 'import tensorflow.compat.v1 as tf\n'), ((22726, 22766), 'tensorflow.compat.v1.one_hot', 'tf.one_hot', (['clss', 'hparams.num_categories'], {}), '(clss, hparams.num_categories)\n', (22736, 22766), True, 'import tensorflow.compat.v1 as tf\n'), ((22804, 22823), 'tensorflow.compat.v1.one_hot', 'tf.one_hot', (['sln', '(51)'], {}), '(sln, 51)\n', (22814, 22823), True, 'import tensorflow.compat.v1 as tf\n'), ((22928, 22968), 'tensorflow.compat.v1.one_hot', 'tf.one_hot', (['clss', 'hparams.num_categories'], {}), '(clss, hparams.num_categories)\n', (22938, 22968), True, 'import tensorflow.compat.v1 as tf\n'), ((23185, 23221), 'tensor2tensor.layers.common_layers.shape_list', 'common_layers.shape_list', (['bottleneck'], {}), '(bottleneck)\n', (23209, 23221), False, 'from tensor2tensor.layers import common_layers\n'), ((23363, 23395), 'tensor2tensor.layers.common_layers.shape_list', 'common_layers.shape_list', (['inputs'], {}), '(inputs)\n', (23387, 23395), False, 'from tensor2tensor.layers import common_layers\n'), ((2839, 2887), 'tensor2tensor.layers.common_layers.shape_list', 'common_layers.shape_list', (["features['bottleneck']"], {}), "(features['bottleneck'])\n", (2863, 2887), False, 'from tensor2tensor.layers import common_layers\n'), ((13670, 13718), 'tensor2tensor.layers.common_layers.shape_list', 'common_layers.shape_list', (["features['bottleneck']"], {}), "(features['bottleneck'])\n", (13694, 13718), False, 'from tensor2tensor.layers import common_layers\n'), ((19882, 19918), 'tensor2tensor.layers.common_layers.shape_list', 'common_layers.shape_list', (['bottleneck'], {}), '(bottleneck)\n', (19906, 19918), False, 'from tensor2tensor.layers import common_layers\n'), ((20106, 20147), 'tensor2tensor.layers.common_layers.shape_list', 'common_layers.shape_list', (['shifted_targets'], {}), '(shifted_targets)\n', (20130, 20147), False, 'from tensor2tensor.layers import common_layers\n'), ((20259, 20291), 'tensor2tensor.layers.common_layers.shape_list', 'common_layers.shape_list', (['inputs'], {}), '(inputs)\n', (20283, 20291), False, 'from tensor2tensor.layers import common_layers\n'), ((15357, 15415), 'tensorflow.compat.v1.concat', 'tf.concat', (['[unbottleneck[:, :unbottleneck_dim // 2], c]', '(1)'], {}), '([unbottleneck[:, :unbottleneck_dim // 2], c], 1)\n', (15366, 15415), True, 'import tensorflow.compat.v1 as tf\n'), ((15443, 15501), 'tensorflow.compat.v1.concat', 'tf.concat', (['[unbottleneck[:, unbottleneck_dim // 2:], h]', '(1)'], {}), '([unbottleneck[:, unbottleneck_dim // 2:], h], 1)\n', (15452, 15501), True, 'import tensorflow.compat.v1 as tf\n'), ((5035, 5052), 'tensorflow.compat.v1.exp', 'tf.exp', (['log_sigma'], {}), '(log_sigma)\n', (5041, 5052), True, 'import tensorflow.compat.v1 as tf\n'), ((5055, 5068), 'tensorflow.compat.v1.square', 'tf.square', (['mu'], {}), '(mu)\n', (5064, 5068), True, 'import tensorflow.compat.v1 as tf\n')]
import csv import numpy as np import time from pathlib import Path from Panalyzer.utils.wr_extractor import wr_extractor from Panalyzer.TraceParser.logic_masking import * def arm32buffered_csv2np(fcsv, buffersize, num_reg): detailded_info = {'wr': None, 'regval': None, 'tick': None, 'masking': None, 'src1': None, 'src2': None, 'op': None} tick_list = np.zeros([buffersize], dtype=np.int64) wr_list = np.full([num_reg, 2, buffersize], False, dtype=bool) reg_val_table = np.zeros([num_reg, buffersize], dtype=np.int64) op_list = [] src1_list = [] src2_list = [] with open(fcsv, mode='r') as infocsv: info_reader = csv.reader(infocsv) buffer_idx = 0 chunk_counter = 0 for idx, row in enumerate(info_reader): if idx % buffersize == 0: buffer_idx = 0 chunk_counter = chunk_counter + 1 print(chunk_counter) tick_list = np.zeros([buffersize], dtype=np.int64) wr_list = np.full([num_reg, 2, buffersize], False, dtype=bool) reg_val_table = np.zeros([num_reg, buffersize], dtype=np.int64) op_list = [] src1_list = [] src2_list = [] else: buffer_idx = buffer_idx + 1 tick_list[buffer_idx] = row[0] # Tick number list: an 1 x line_number np array op_id = row[3] op_list.append(op_id) # Opname is just a simple list of strings # Variables required for utility.wr_extractor, feed into the function, then abstract the required # data structure op_dst1 = row[4] op_dst2 = row[5] op_src1 = row[6] op_src2 = row[7] src1_list.append(op_src1) src2_list.append(op_src2) data = row[-1] for k in range(num_reg): # kth register val_prev = reg_val_table[k, buffer_idx - 1] reg_name = 'r' + str(k) # fp, lr, sp ,pc are renamed, simply wr_list[k, 0, buffer_idx] = \ wr_extractor(reg_name, op_dst1, op_dst2, op_src1, op_src2, op_id, data, val_prev)[0] wr_list[k, 1, buffer_idx] = \ wr_extractor(reg_name, op_dst1, op_dst2, op_src1, op_src2, op_id, data, val_prev)[1] reg_val_table[k, buffer_idx] = \ wr_extractor(reg_name, op_dst1, op_dst2, op_src1, op_src2, op_id, data, val_prev)[2] return tick_list if __name__ == "__main__": project_dir = Path(__file__).resolve().parent.parent csv_dir = project_dir.joinpath('tempcsv') fname = "fftbaseline.csv" start_time = time.perf_counter() # Time counter starts T = arm32buffered_csv2np(csv_dir / fname, 10000, 16) elapsed_time_pandas = time.perf_counter() - start_time # Stop point of the timer # tickexample = T['tick'] # wrexample = T['wr'] # regvalexample = T['regval'] # masking_table = T['masking'] # ops_list = T['op'] # # # print('tick \n', tickexample, '\n wr: \n', wrexample, '\n regval:\n', regvalexample) # print(ops_list)
[ "pathlib.Path", "Panalyzer.utils.wr_extractor.wr_extractor", "time.perf_counter", "numpy.zeros", "numpy.full", "csv.reader" ]
[((402, 440), 'numpy.zeros', 'np.zeros', (['[buffersize]'], {'dtype': 'np.int64'}), '([buffersize], dtype=np.int64)\n', (410, 440), True, 'import numpy as np\n'), ((456, 508), 'numpy.full', 'np.full', (['[num_reg, 2, buffersize]', '(False)'], {'dtype': 'bool'}), '([num_reg, 2, buffersize], False, dtype=bool)\n', (463, 508), True, 'import numpy as np\n'), ((530, 577), 'numpy.zeros', 'np.zeros', (['[num_reg, buffersize]'], {'dtype': 'np.int64'}), '([num_reg, buffersize], dtype=np.int64)\n', (538, 577), True, 'import numpy as np\n'), ((2905, 2924), 'time.perf_counter', 'time.perf_counter', ([], {}), '()\n', (2922, 2924), False, 'import time\n'), ((706, 725), 'csv.reader', 'csv.reader', (['infocsv'], {}), '(infocsv)\n', (716, 725), False, 'import csv\n'), ((3033, 3052), 'time.perf_counter', 'time.perf_counter', ([], {}), '()\n', (3050, 3052), False, 'import time\n'), ((1017, 1055), 'numpy.zeros', 'np.zeros', (['[buffersize]'], {'dtype': 'np.int64'}), '([buffersize], dtype=np.int64)\n', (1025, 1055), True, 'import numpy as np\n'), ((1083, 1135), 'numpy.full', 'np.full', (['[num_reg, 2, buffersize]', '(False)'], {'dtype': 'bool'}), '([num_reg, 2, buffersize], False, dtype=bool)\n', (1090, 1135), True, 'import numpy as np\n'), ((1169, 1216), 'numpy.zeros', 'np.zeros', (['[num_reg, buffersize]'], {'dtype': 'np.int64'}), '([num_reg, buffersize], dtype=np.int64)\n', (1177, 1216), True, 'import numpy as np\n'), ((2768, 2782), 'pathlib.Path', 'Path', (['__file__'], {}), '(__file__)\n', (2772, 2782), False, 'from pathlib import Path\n'), ((2281, 2366), 'Panalyzer.utils.wr_extractor.wr_extractor', 'wr_extractor', (['reg_name', 'op_dst1', 'op_dst2', 'op_src1', 'op_src2', 'op_id', 'data', 'val_prev'], {}), '(reg_name, op_dst1, op_dst2, op_src1, op_src2, op_id, data,\n val_prev)\n', (2293, 2366), False, 'from Panalyzer.utils.wr_extractor import wr_extractor\n'), ((2442, 2527), 'Panalyzer.utils.wr_extractor.wr_extractor', 'wr_extractor', (['reg_name', 'op_dst1', 'op_dst2', 'op_src1', 'op_src2', 'op_id', 'data', 'val_prev'], {}), '(reg_name, op_dst1, op_dst2, op_src1, op_src2, op_id, data,\n val_prev)\n', (2454, 2527), False, 'from Panalyzer.utils.wr_extractor import wr_extractor\n'), ((2606, 2691), 'Panalyzer.utils.wr_extractor.wr_extractor', 'wr_extractor', (['reg_name', 'op_dst1', 'op_dst2', 'op_src1', 'op_src2', 'op_id', 'data', 'val_prev'], {}), '(reg_name, op_dst1, op_dst2, op_src1, op_src2, op_id, data,\n val_prev)\n', (2618, 2691), False, 'from Panalyzer.utils.wr_extractor import wr_extractor\n')]
""" define model for gp """ # from threading import Thread # from queue import Queue from multiprocessing import Pool from random import random, randint from math import floor import operator from autoprover.gp.gene import Gene from autoprover.gp.rule import GeneRule from autoprover.gp.action import GeneAction from autoprover.gp.trigger import GeneTrigger from autoprover.gp.restriction import Restriction #TODO fox too many instant class GPModel: """ gp model """ #TODO fix too many args def __init__(self, args=None, populationSize=None, maxGeneration=None, mutateRate=None, eliteRate=None, crossRate=None, crossType=None, verifyNum=None, proof=None, tactics=None, limit_hyp=None, limit_goal=None): self.population_size = populationSize or args.populationSize self.max_generation = maxGeneration or args.maxGeneration self.mutate_rate = mutateRate or args.mutateRate self.elite_rate = eliteRate or args.eliteRate self.cross_rate = crossRate or args.crossRate self.cross_type = crossType or args.crossType self.verify_num = verifyNum or args.verifyNum self.limit_hyp = limit_hyp or args.limit_hyp self.limit_goal = limit_goal or args.limit_goal self.debug = args.debug self.proof = proof self.tactics = tactics self.population = None self.current_generation = 1 self.proofs = [] self.rules = [] self.init_population(self.population_size) self.pre_process() def show_prop(self): """ display property for model """ print(self.population_size) print(self.max_generation) print(self.mutate_rate) print(self.elite_rate) print(self.cross_rate) print(self.cross_type) print(self.verify_num) print(self.proof) def init_population(self, size): """ create population by size """ print("Initializing population.") self.population = [] for _ in range(size): self.population.append(Gene(self.tactics)) def pre_process(self): """ run before start """ self.current_generation = 1 self.update_fitness_for_population() self.fitness_sharing() self.sort_population() self.update_tactic_usage() self.check_proof() def is_proved(self): """ check population has a proof """ return len(self.proofs) > 0 def start(self, gen=None): """ run the model """ if gen is None: # if gen is not set local_gen_limit = self.max_generation + 1 else: local_gen_limit = gen if self.current_generation > self.max_generation: return for _ in range(local_gen_limit): print("Generation No.{0}".format(self.current_generation)) if self.debug: self.sort_population() for index in range(0, 30): self.population[index].print_lastest() self.crossover() self.update_fitness_for_population() self.apply_rules() self.fitness_sharing() self.next_generation() if self.current_generation > self.max_generation: break # self.printGeneByIndex(0, True) def fitness_sharing(self): """Use fitness sharing to re-evaluate fitness""" for gene in self.population: gene.fitness = gene.raw_fitness return def dist(gene1, gene2): """Return distence between two gene""" return abs(len(gene1.goal) - len(gene2.goal)) for gene in self.population: raw_fitnesses = [e.raw_fitness for e in self.population if dist(e, gene) <= 5] gene.fitness = sum(raw_fitnesses) / len(raw_fitnesses) def next_generation(self): """ next generation """ print("Avg. fitness\tAvg. length") print("{0:.8f}\t{1}".format(self.average_fitness(), self.average_length_of_gene())) self.current_generation += 1 self.sort_population() self.update_tactic_usage() self.check_proof() def check_proof(self): """Check if there is a proof in population """ for gene in self.population: if gene.is_proof: print(gene.chromosome) for state in gene.coq_states: print(state) self.proofs.append(Gene(chromosome=gene.valid_tactics)) def update_fitness_for_population(self): """ return individual if theorem is proved, o.w return None """ def wrapper(func, *args, **kwargs): """func wrapper""" return func, args, kwargs with Pool(processes=4) as pool: for gene in self.population: func, args, kargs = wrapper(gene.update_fitness_for_proof, self.proof, self.limit_hyp, self.limit_goal) pool.apply_async(func(*args, **kargs)) def apply_rules(self): """Perform action by rules""" if len(self.rules) == 0: return for gene in self.population: for rule in self.rules: if rule.type == "gene": rule.check_and_apply(gene) def crossover(self): """ the crossover operation for gp """ self.sort_population() elite_amount = round(self.elite_rate * self.population_size) # preserve from the top new_population = [ele for ele in self.population if ele.ttl > 0] for individual in new_population: if individual.ttl > 0: individual.ttl -= 1 new_population += self.population[:elite_amount] while len(new_population) < self.population_size: # newGene = self.crossBelowCrossRate() new_gene, new_gene2 = self.cross_on_arb_seq() if random() <= self.mutate_rate: self.mutate_append(new_gene) new_population.append(new_gene) if len(new_population) == self.population_size: break if random() <= self.mutate_rate: self.mutate_append(new_gene2) new_population.append(new_gene2) self.population = new_population def sort_population(self): """ sort population by length and fitness """ self.population.sort(key=lambda x: x.fitness, reverse=True) def update_tactic_usage(self): """update tactic statistic usage""" usage = {e: 0 for e in self.tactics.all_tactics} count = 0 for gene in self.population: for tactic in gene.chromosome: count += 1 try: usage[tactic] += 1 except KeyError: usage[tactic] = 1 for tactic in usage: usage[tactic] = usage[tactic]/count self.proof.tactics.usage = usage def cross_below_cross_rate(self): """ select two parent by cross rate, crossover on random point """ p1_index = randint(0, floor(self.population_size * self.cross_rate)-1) p2_index = randint(0, floor(self.population_size * self.cross_rate)-1) gene_of_p1 = self.population[p1_index] gene_of_p2 = self.population[p2_index] cross_point = randint(0, int_min(len(gene_of_p1), len(gene_of_p2))-1) new_chromosome = [] new_chromosome += gene_of_p1.chromosome[:cross_point] new_chromosome += gene_of_p2.chromosome[cross_point:] if (self.tactics.is_unrepeatable(new_chromosome[cross_point]) and cross_point < len(new_chromosome)-1): if new_chromosome[cross_point] == new_chromosome[cross_point+1]: del new_chromosome[cross_point] return Gene(chromosome=new_chromosome) def cross_on_arb_seq(self, slmax=6): """ select two parent by cross_rate, crossover by some seqence """ p1_index = randint(0, floor(self.population_size * self.cross_rate)-1) p2_index = randint(0, floor(self.population_size * self.cross_rate)-1) gene_of_p1 = self.population[p1_index] gene_of_p2 = self.population[p2_index] p1_begin = myrandint(0, len(gene_of_p1)-1) p1_end = p1_begin + myrandint(1, int_min(slmax, len(gene_of_p1)-p1_begin)) p2_begin = myrandint(0, len(gene_of_p2)-1) p2_end = p2_begin + myrandint(1, int_min(slmax, len(gene_of_p2)-p2_begin)) new_chromosome = [] new_chromosome += gene_of_p1.chromosome[:p1_begin] new_chromosome += gene_of_p2.chromosome[p2_begin:p2_end] new_chromosome += gene_of_p1.chromosome[p1_end:] new_chromosome2 = [] new_chromosome2 += gene_of_p2.chromosome[:p2_begin] new_chromosome2 += gene_of_p1.chromosome[p1_begin:p1_end] new_chromosome2 += gene_of_p2.chromosome[p2_end:] self.remove_repeatable(new_chromosome) self.remove_repeatable(new_chromosome2) return Gene(chromosome=new_chromosome), Gene(chromosome=new_chromosome2) def remove_repeatable(self, chromosome): """ remove repeatable tactic """ tactic_set = set() for index, tactic in enumerate(chromosome): if self.tactics.is_unrepeatable(tactic): if tactic in tactic_set: del chromosome[index] else: tactic_set.add(tactic) def mutate(self, gene): """ the mutate operation """ if len(gene) == 1: gene.chromosome[0] = self.tactics.mutate_select() else: index = randint(len(gene)//2, len(gene)-1) gene.chromosome[index] = self.tactics.mutate_select() def mutate_append(self, gene): """append a tactic to chromosome""" gene.chromosome.append(self.tactics.mutate_select()) def average_fitness(self): """Calculate the average fitness for population. Returns: double: avg. fitness """ return sum([e.fitness for e in self.population]) / len(self.population) def average_length_of_gene(self): """Calculate the average fitness for population. Returns: double: avg. fitness """ return sum([len(e) for e in self.population]) / len(self.population) def edit(self, index, data=None): """Human involved modification of some gene of the population """ if self.current_generation > self.max_generation: return print("Edit Gene {} now.".format(index)) gene = self.population[index] gene.modification(data=data) gene.update_fitness_for_proof(self.proof) if gene.is_proof: self.proofs.append(Gene(chromosome=gene.valid_tactics)) return def show_proofs(self): """Show proofs found """ if self.proofs: for gene in self.proofs: print(gene.format_output(self.proof)) else: print("There is not proof for now.") def list(self, argv): """List property of some individual. Args: argv(list): sub command of list function. """ def get_interval(interval): """Get begin and end of interval """ interval_list = interval.split("-") if len(interval_list) == 1: return (int(interval_list[0]), int(interval_list[0])+1) else: return (int(interval_list[0]), int(interval_list[1])+1) if not argv or not argv[0]: return (begin, end) = get_interval(argv[0]) if len(argv) == 1: for index, gene in enumerate(self.population[begin:end]): print("{0}: {1:.8f}".format(index, gene.fitness)) gene.print_progress() elif argv[1] == "fitness": for index, gene in enumerate(self.population[begin:end]): print("{0}: {1:.8f}".format(index, gene.fitness)) elif argv[1] == "chromosome": for index, gene in enumerate(self.population[begin:end]): print("{0}: {1}".format(index, gene.chromosome)) elif argv[1] == "ttl": for index, gene in enumerate(self.population[begin:end]): print("{0}: {1}".format(index, gene.ttl)) def read_rule_from_file(self, file_name): """Read a rule from a JSON file""" try: self.rules.append(GeneRule(file_name=file_name, proof=self.proof)) except FileNotFoundError: return def delete_rule(self, index): """delete a rule from rule_list""" del self.rules[index] def remove_tactic(self): """remove a tactic in tactic_set and population""" tactic_removed = input("Enter a tactic to be removed: ") self.proof.tactics.remove(tactic_removed) for gene in self.population: gene.chromosome = [e for e in gene.chromosome if e != tactic_removed] def set_attributes(self, argv): """Set attributes of population """ if argv[0] == "population" or argv[0] == "pop": if argv[1] == "ttl": self.population[int(argv[2])].ttl = int(argv[3]) def defrag(self, index_list): """Defrag some gene""" for index in index_list: self.population[index].defrag(self.proof) def print_stats(self): """print tactic usage""" sorted_stats = sorted(self.proof.tactics.usage.items(), key=operator.itemgetter(1), reverse=True) for tactic, usage in sorted_stats: print("{0}: {1:.4f}%".format(tactic, usage*100)) def myrandint(begin, end): """ randint warrper for begin == end """ if begin == end: return begin else: return randint(begin, end) def int_max(int_a, int_b): """ max(a, b) """ if int_a > int_b: return int_a else: return int_b def int_min(int_a, int_b): """ min(a, b) """ if int_a < int_b: return int_a else: return int_b
[ "math.floor", "autoprover.gp.gene.Gene", "multiprocessing.Pool", "autoprover.gp.rule.GeneRule", "operator.itemgetter", "random.random", "random.randint" ]
[((8178, 8209), 'autoprover.gp.gene.Gene', 'Gene', ([], {'chromosome': 'new_chromosome'}), '(chromosome=new_chromosome)\n', (8182, 8209), False, 'from autoprover.gp.gene import Gene\n'), ((14306, 14325), 'random.randint', 'randint', (['begin', 'end'], {}), '(begin, end)\n', (14313, 14325), False, 'from random import random, randint\n'), ((4982, 4999), 'multiprocessing.Pool', 'Pool', ([], {'processes': '(4)'}), '(processes=4)\n', (4986, 4999), False, 'from multiprocessing import Pool\n'), ((9396, 9427), 'autoprover.gp.gene.Gene', 'Gene', ([], {'chromosome': 'new_chromosome'}), '(chromosome=new_chromosome)\n', (9400, 9427), False, 'from autoprover.gp.gene import Gene\n'), ((9429, 9461), 'autoprover.gp.gene.Gene', 'Gene', ([], {'chromosome': 'new_chromosome2'}), '(chromosome=new_chromosome2)\n', (9433, 9461), False, 'from autoprover.gp.gene import Gene\n'), ((2146, 2164), 'autoprover.gp.gene.Gene', 'Gene', (['self.tactics'], {}), '(self.tactics)\n', (2150, 2164), False, 'from autoprover.gp.gene import Gene\n'), ((6238, 6246), 'random.random', 'random', ([], {}), '()\n', (6244, 6246), False, 'from random import random, randint\n'), ((6455, 6463), 'random.random', 'random', ([], {}), '()\n', (6461, 6463), False, 'from random import random, randint\n'), ((7458, 7503), 'math.floor', 'floor', (['(self.population_size * self.cross_rate)'], {}), '(self.population_size * self.cross_rate)\n', (7463, 7503), False, 'from math import floor\n'), ((7537, 7582), 'math.floor', 'floor', (['(self.population_size * self.cross_rate)'], {}), '(self.population_size * self.cross_rate)\n', (7542, 7582), False, 'from math import floor\n'), ((8373, 8418), 'math.floor', 'floor', (['(self.population_size * self.cross_rate)'], {}), '(self.population_size * self.cross_rate)\n', (8378, 8418), False, 'from math import floor\n'), ((8452, 8497), 'math.floor', 'floor', (['(self.population_size * self.cross_rate)'], {}), '(self.population_size * self.cross_rate)\n', (8457, 8497), False, 'from math import floor\n'), ((11191, 11226), 'autoprover.gp.gene.Gene', 'Gene', ([], {'chromosome': 'gene.valid_tactics'}), '(chromosome=gene.valid_tactics)\n', (11195, 11226), False, 'from autoprover.gp.gene import Gene\n'), ((12928, 12975), 'autoprover.gp.rule.GeneRule', 'GeneRule', ([], {'file_name': 'file_name', 'proof': 'self.proof'}), '(file_name=file_name, proof=self.proof)\n', (12936, 12975), False, 'from autoprover.gp.rule import GeneRule\n'), ((14016, 14038), 'operator.itemgetter', 'operator.itemgetter', (['(1)'], {}), '(1)\n', (14035, 14038), False, 'import operator\n'), ((4684, 4719), 'autoprover.gp.gene.Gene', 'Gene', ([], {'chromosome': 'gene.valid_tactics'}), '(chromosome=gene.valid_tactics)\n', (4688, 4719), False, 'from autoprover.gp.gene import Gene\n')]
import sys sys.exit(1)
[ "sys.exit" ]
[((12, 23), 'sys.exit', 'sys.exit', (['(1)'], {}), '(1)\n', (20, 23), False, 'import sys\n')]
# -*- coding: utf-8 -*- # Generated by Django 1.11.15 on 2018-11-26 10:42 from __future__ import unicode_literals from django.db import migrations, models import jsonfield.fields class Migration(migrations.Migration): initial = True dependencies = [ ] operations = [ migrations.CreateModel( name='Mapping', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), ('label', models.CharField(help_text='Label for your reference', max_length=255, verbose_name='label')), ('source', models.CharField(help_text='The source feed for your data', max_length=255, verbose_name='source')), ('parser', models.CharField(choices=[(b'feedmapper.parsers.AtomParser', b'Atom'), (b'feedmapper.parsers.XMLParser', b'XML')], help_text='Which parser to use when synchronizing', max_length=255, verbose_name='parser')), ('purge', models.BooleanField(default=False, help_text='Purge existing items on sync?', verbose_name='purge')), ('data_map', jsonfield.fields.JSONField(default=dict, verbose_name='data map')), ('notification_recipients', models.TextField(blank=True, help_text='Specify one email address per line to be notified of parsing errors.', verbose_name='notification recipients')), ('parse_attempted', models.DateTimeField(blank=True, null=True, verbose_name='parse attempted')), ('parse_succeeded', models.BooleanField(verbose_name='parse succeeded')), ('parse_log', models.TextField(blank=True, verbose_name='parse log')), ], ), ]
[ "django.db.models.TextField", "django.db.models.BooleanField", "django.db.models.AutoField", "django.db.models.DateTimeField", "django.db.models.CharField" ]
[((393, 486), 'django.db.models.AutoField', 'models.AutoField', ([], {'auto_created': '(True)', 'primary_key': '(True)', 'serialize': '(False)', 'verbose_name': '"""ID"""'}), "(auto_created=True, primary_key=True, serialize=False,\n verbose_name='ID')\n", (409, 486), False, 'from django.db import migrations, models\n'), ((511, 607), 'django.db.models.CharField', 'models.CharField', ([], {'help_text': '"""Label for your reference"""', 'max_length': '(255)', 'verbose_name': '"""label"""'}), "(help_text='Label for your reference', max_length=255,\n verbose_name='label')\n", (527, 607), False, 'from django.db import migrations, models\n'), ((633, 735), 'django.db.models.CharField', 'models.CharField', ([], {'help_text': '"""The source feed for your data"""', 'max_length': '(255)', 'verbose_name': '"""source"""'}), "(help_text='The source feed for your data', max_length=255,\n verbose_name='source')\n", (649, 735), False, 'from django.db import migrations, models\n'), ((761, 981), 'django.db.models.CharField', 'models.CharField', ([], {'choices': "[(b'feedmapper.parsers.AtomParser', b'Atom'), (\n b'feedmapper.parsers.XMLParser', b'XML')]", 'help_text': '"""Which parser to use when synchronizing"""', 'max_length': '(255)', 'verbose_name': '"""parser"""'}), "(choices=[(b'feedmapper.parsers.AtomParser', b'Atom'), (\n b'feedmapper.parsers.XMLParser', b'XML')], help_text=\n 'Which parser to use when synchronizing', max_length=255, verbose_name=\n 'parser')\n", (777, 981), False, 'from django.db import migrations, models\n'), ((995, 1099), 'django.db.models.BooleanField', 'models.BooleanField', ([], {'default': '(False)', 'help_text': '"""Purge existing items on sync?"""', 'verbose_name': '"""purge"""'}), "(default=False, help_text=\n 'Purge existing items on sync?', verbose_name='purge')\n", (1014, 1099), False, 'from django.db import migrations, models\n'), ((1238, 1397), 'django.db.models.TextField', 'models.TextField', ([], {'blank': '(True)', 'help_text': '"""Specify one email address per line to be notified of parsing errors."""', 'verbose_name': '"""notification recipients"""'}), "(blank=True, help_text=\n 'Specify one email address per line to be notified of parsing errors.',\n verbose_name='notification recipients')\n", (1254, 1397), False, 'from django.db import migrations, models\n'), ((1427, 1502), 'django.db.models.DateTimeField', 'models.DateTimeField', ([], {'blank': '(True)', 'null': '(True)', 'verbose_name': '"""parse attempted"""'}), "(blank=True, null=True, verbose_name='parse attempted')\n", (1447, 1502), False, 'from django.db import migrations, models\n'), ((1541, 1592), 'django.db.models.BooleanField', 'models.BooleanField', ([], {'verbose_name': '"""parse succeeded"""'}), "(verbose_name='parse succeeded')\n", (1560, 1592), False, 'from django.db import migrations, models\n'), ((1625, 1679), 'django.db.models.TextField', 'models.TextField', ([], {'blank': '(True)', 'verbose_name': '"""parse log"""'}), "(blank=True, verbose_name='parse log')\n", (1641, 1679), False, 'from django.db import migrations, models\n')]
""" Markdownreveal local module tests. """ import json import time from hashlib import sha1 from pathlib import Path from shutil import rmtree from tarfile import TarInfo from tempfile import mkdtemp import pytest from markdownreveal.local import clean_tar_members from markdownreveal.local import initialize_localdir from markdownreveal.local import latest_project_release def test_latest_project_release(): """ Test `latest_project_release()` function. """ # Latest reveal.js version latest = latest_project_release(github='hakimel/reveal.js') numbers = latest.split('.') assert len(numbers) == 3 assert all(n.isdecimal() for n in numbers) # Latest KaTeX version latest = latest_project_release(github='khan/katex') assert latest[0] == 'v' latest = latest[1:] numbers = latest.split('.') assert len(numbers) == 3 assert all(n.isdecimal() for n in numbers) def test_clean_tar_members(): """ Test `clean_tar_members()` function. """ members = [ TarInfo('toplevel'), TarInfo('toplevel/index.html'), TarInfo('toplevel/foo/bar.xyz'), ] output = [TarInfo('index.html'), TarInfo('foo/bar.xyz')] result = clean_tar_members(members) assert all(x.name == y.name for x, y in zip(output, result)) @pytest.mark.parametrize( 'reveal_version,katex_version,reveal_tag,katex_tag,style', [ ( 'latest', 'latest', latest_project_release(github='hakimel/reveal.js'), latest_project_release(github='khan/katex'), '', ), ( '3.4.0', 'v0.7.1', '3.4.0', 'v0.7.1', 'https://github.com/markdownreveal/style-default/' + 'archive/master.tar.gz', ), ], ) def test_initialize_localdir( reveal_version, katex_version, reveal_tag, katex_tag, style ): """ Test `initialize_localdir()` function. """ localdir = Path(mkdtemp()) config = { 'local_path': localdir, 'reveal_version': reveal_version, 'katex_version': katex_version, 'style': style, } out = initialize_localdir(config) package = json.load(open(str(out / 'revealjs' / 'package.json'))) assert package['version'] == reveal_tag katex_readme = out / 'katex' / 'README.md' assert katex_tag[1:] in katex_readme.read_text() style_out = out / 'markdownrevealstyle' if style: style_dir = localdir / sha1(style.encode('utf')).hexdigest() assert style_dir.exists() assert style_out.exists() else: assert not style_out.exists() # If version is already downloaded initialization should be pretty fast t0 = time.time() out = initialize_localdir(config) assert time.time() - t0 < 0.01 rmtree(str(localdir))
[ "markdownreveal.local.clean_tar_members", "markdownreveal.local.initialize_localdir", "tempfile.mkdtemp", "markdownreveal.local.latest_project_release", "time.time", "tarfile.TarInfo" ]
[((518, 568), 'markdownreveal.local.latest_project_release', 'latest_project_release', ([], {'github': '"""hakimel/reveal.js"""'}), "(github='hakimel/reveal.js')\n", (540, 568), False, 'from markdownreveal.local import latest_project_release\n'), ((717, 760), 'markdownreveal.local.latest_project_release', 'latest_project_release', ([], {'github': '"""khan/katex"""'}), "(github='khan/katex')\n", (739, 760), False, 'from markdownreveal.local import latest_project_release\n'), ((1216, 1242), 'markdownreveal.local.clean_tar_members', 'clean_tar_members', (['members'], {}), '(members)\n', (1233, 1242), False, 'from markdownreveal.local import clean_tar_members\n'), ((2181, 2208), 'markdownreveal.local.initialize_localdir', 'initialize_localdir', (['config'], {}), '(config)\n', (2200, 2208), False, 'from markdownreveal.local import initialize_localdir\n'), ((2751, 2762), 'time.time', 'time.time', ([], {}), '()\n', (2760, 2762), False, 'import time\n'), ((2773, 2800), 'markdownreveal.local.initialize_localdir', 'initialize_localdir', (['config'], {}), '(config)\n', (2792, 2800), False, 'from markdownreveal.local import initialize_localdir\n'), ((1034, 1053), 'tarfile.TarInfo', 'TarInfo', (['"""toplevel"""'], {}), "('toplevel')\n", (1041, 1053), False, 'from tarfile import TarInfo\n'), ((1063, 1093), 'tarfile.TarInfo', 'TarInfo', (['"""toplevel/index.html"""'], {}), "('toplevel/index.html')\n", (1070, 1093), False, 'from tarfile import TarInfo\n'), ((1103, 1134), 'tarfile.TarInfo', 'TarInfo', (['"""toplevel/foo/bar.xyz"""'], {}), "('toplevel/foo/bar.xyz')\n", (1110, 1134), False, 'from tarfile import TarInfo\n'), ((1156, 1177), 'tarfile.TarInfo', 'TarInfo', (['"""index.html"""'], {}), "('index.html')\n", (1163, 1177), False, 'from tarfile import TarInfo\n'), ((1179, 1201), 'tarfile.TarInfo', 'TarInfo', (['"""foo/bar.xyz"""'], {}), "('foo/bar.xyz')\n", (1186, 1201), False, 'from tarfile import TarInfo\n'), ((2001, 2010), 'tempfile.mkdtemp', 'mkdtemp', ([], {}), '()\n', (2008, 2010), False, 'from tempfile import mkdtemp\n'), ((2812, 2823), 'time.time', 'time.time', ([], {}), '()\n', (2821, 2823), False, 'import time\n'), ((1471, 1521), 'markdownreveal.local.latest_project_release', 'latest_project_release', ([], {'github': '"""hakimel/reveal.js"""'}), "(github='hakimel/reveal.js')\n", (1493, 1521), False, 'from markdownreveal.local import latest_project_release\n'), ((1535, 1578), 'markdownreveal.local.latest_project_release', 'latest_project_release', ([], {'github': '"""khan/katex"""'}), "(github='khan/katex')\n", (1557, 1578), False, 'from markdownreveal.local import latest_project_release\n')]
from models.MultiGMPmodel import MultiCopyGMPmodel from models.MultiGGHPmodel import MultiCopyGGHPmodel from util.calculatedCRBrateAndEstimationAccuracy import calculatedCRBrateAndEstimationAccuracy from util.cutCircularChromosomes import cutCircularChromosomes """ Inferring ancestor species for Gramineae species. Ancestor 4: Multi-copy GGHP model, result in outdutdata/Gramineae/Ancestor4 Ancestor 3: Multi-copy GMP model, result in outdutdata/Gramineae/Ancestor3 Ancestor 2: Multi-copy GMP model, result in outdutdata/Gramineae/Ancestor2 Ancestor 1: Multi-copy GMP model, result in outdutdata/Gramineae/Ancestor1 """ path = 'D:/InferAncestorGenome/realData' workdir = path + '/IAGS_version1.0/inputdata/Gramineae/' """ Inferring ancestor species for Gramineae species. Ancestor 4 using Multi-copy GGHP model """ dup_child_file = workdir + 'Zmays.final.block' outgroup_file = workdir + 'Sbicolor.final.block' outAncestor4dir = path + '/IAGS_version1.0/outputdata/Gramineae/Ancestor4/' dup_copy_number = 4 out_copy_number = 2 ancestor_target_copy_number = 2 ancestor_name = 'Ancestor4' MultiCopyGGHPmodel(dup_child_file, outgroup_file, outAncestor4dir, ancestor_name, dup_copy_number, out_copy_number, ancestor_target_copy_number) # speciesAndCopyList = [ # [workdir + 'Zmays.final.block',dup_copy_number,'Z.mays'], # [workdir + 'Sbicolor.final.block',out_copy_number,'S.bicolor'] # ] # cutCircularChromosomes(outAncestor4dir + 'Ancestor4.block', # ancestor_target_copy_number, # ancestor_name,speciesAndCopyList,outAncestor4dir) # Evaluation matching_target_file = workdir + 'Sbicolor.final.block' matching_target_copy_number = out_copy_number matching_target_name = 'S.bicolor' speciesAndCopyList = [ [workdir + 'Zmays.final.block',dup_copy_number,'Z.mays'], [workdir + 'Sbicolor.final.block',out_copy_number,'S.bicolor'] ] model_type = 'MultiCopyGGHP' calculatedCRBrateAndEstimationAccuracy(matching_target_file, matching_target_copy_number, matching_target_name, speciesAndCopyList, outAncestor4dir, model_type) """ Inferring ancestor species for Gramineae species. Ancestor 3 using Multi-copy GMP model """ species_file_list = [workdir + 'Sbicolor.final.block', outAncestor4dir + 'Ancestor4.block', workdir + 'Osativa.final.block'] guided_species_for_matching = workdir + 'Sbicolor.final.block' ancestor_target_copy_number = 2 outAncestor3dir = path + '/IAGS_version1.0/outputdata/Gramineae/Ancestor3/' ancestor_name = 'Ancestor3' MultiCopyGMPmodel(species_file_list, outAncestor3dir, guided_species_for_matching, ancestor_name, ancestor_target_copy_number) # Evaluation matching_target_file = workdir + 'Sbicolor.final.block' matching_target_copy_number = ancestor_target_copy_number matching_target_name = 'S.bicolor' speciesAndCopyList = [ [workdir + 'Sbicolor.final.block',ancestor_target_copy_number,'S.bicolor'], [outAncestor4dir + 'Ancestor4.block',ancestor_target_copy_number,'Ancestor4'], [workdir + 'Osativa.final.block',ancestor_target_copy_number,'O.sativa'] ] model_type = 'MultiCopyGMP' calculatedCRBrateAndEstimationAccuracy(matching_target_file, matching_target_copy_number, matching_target_name, speciesAndCopyList, outAncestor3dir, model_type) """ Inferring ancestor species for Gramineae species. Ancestor 2 using Multi-copy GMP model """ species_file_list = [workdir + 'Bdistachyon.final.block', workdir + 'Telongatum.final.block', workdir + 'Osativa.final.block'] guided_species_for_matching = workdir + 'Bdistachyon.final.block' ancestor_target_copy_number = 2 outAncestor2dir = path + '/IAGS_version1.0/outputdata/Gramineae/Ancestor2/' ancestor_name = 'Ancestor2' MultiCopyGMPmodel(species_file_list, outAncestor2dir, guided_species_for_matching, ancestor_name, ancestor_target_copy_number) # Evaluation matching_target_file = workdir + 'Bdistachyon.final.block' matching_target_copy_number = ancestor_target_copy_number matching_target_name = 'B.distachyon' speciesAndCopyList = [ [workdir + 'Bdistachyon.final.block',ancestor_target_copy_number,'B.distachyon'], [workdir + 'Telongatum.final.block',ancestor_target_copy_number,'T.elongatum'], [workdir + 'Osativa.final.block',ancestor_target_copy_number,'O.sativa'] ] model_type = 'MultiCopyGMP' calculatedCRBrateAndEstimationAccuracy(matching_target_file, matching_target_copy_number, matching_target_name, speciesAndCopyList, outAncestor2dir, model_type) """ Inferring ancestor species for Gramineae species. Ancestor 1 using Multi-copy GMP model """ species_file_list = [workdir + 'Osativa.final.block', outAncestor2dir + 'Ancestor2.block', outAncestor3dir + 'Ancestor3.block'] guided_species_for_matching = workdir + 'Osativa.final.block' ancestor_target_copy_number = 2 outAncestor1dir = path + '/IAGS_version1.0/outputdata/Gramineae/Ancestor1/' ancestor_name = 'Ancestor1' MultiCopyGMPmodel(species_file_list, outAncestor1dir, guided_species_for_matching, ancestor_name, ancestor_target_copy_number) # Evaluation matching_target_file = workdir + 'Osativa.final.block' matching_target_copy_number = ancestor_target_copy_number matching_target_name = 'O.sativa' speciesAndCopyList = [ [workdir + 'Osativa.final.block',ancestor_target_copy_number,'O.sativa'], [outAncestor2dir + 'Ancestor2.block',ancestor_target_copy_number,'Ancestor2'], [outAncestor3dir + 'Ancestor3.block',ancestor_target_copy_number,'Ancestor3'] ] model_type = 'MultiCopyGMP' calculatedCRBrateAndEstimationAccuracy(matching_target_file, matching_target_copy_number, matching_target_name, speciesAndCopyList, outAncestor1dir, model_type)
[ "util.calculatedCRBrateAndEstimationAccuracy.calculatedCRBrateAndEstimationAccuracy", "models.MultiGMPmodel.MultiCopyGMPmodel", "models.MultiGGHPmodel.MultiCopyGGHPmodel" ]
[((1094, 1246), 'models.MultiGGHPmodel.MultiCopyGGHPmodel', 'MultiCopyGGHPmodel', (['dup_child_file', 'outgroup_file', 'outAncestor4dir', 'ancestor_name', 'dup_copy_number', 'out_copy_number', 'ancestor_target_copy_number'], {}), '(dup_child_file, outgroup_file, outAncestor4dir,\n ancestor_name, dup_copy_number, out_copy_number,\n ancestor_target_copy_number)\n', (1112, 1246), False, 'from models.MultiGGHPmodel import MultiCopyGGHPmodel\n'), ((1949, 2117), 'util.calculatedCRBrateAndEstimationAccuracy.calculatedCRBrateAndEstimationAccuracy', 'calculatedCRBrateAndEstimationAccuracy', (['matching_target_file', 'matching_target_copy_number', 'matching_target_name', 'speciesAndCopyList', 'outAncestor4dir', 'model_type'], {}), '(matching_target_file,\n matching_target_copy_number, matching_target_name, speciesAndCopyList,\n outAncestor4dir, model_type)\n', (1987, 2117), False, 'from util.calculatedCRBrateAndEstimationAccuracy import calculatedCRBrateAndEstimationAccuracy\n'), ((2613, 2743), 'models.MultiGMPmodel.MultiCopyGMPmodel', 'MultiCopyGMPmodel', (['species_file_list', 'outAncestor3dir', 'guided_species_for_matching', 'ancestor_name', 'ancestor_target_copy_number'], {}), '(species_file_list, outAncestor3dir,\n guided_species_for_matching, ancestor_name, ancestor_target_copy_number)\n', (2630, 2743), False, 'from models.MultiGMPmodel import MultiCopyGMPmodel\n'), ((3197, 3365), 'util.calculatedCRBrateAndEstimationAccuracy.calculatedCRBrateAndEstimationAccuracy', 'calculatedCRBrateAndEstimationAccuracy', (['matching_target_file', 'matching_target_copy_number', 'matching_target_name', 'speciesAndCopyList', 'outAncestor3dir', 'model_type'], {}), '(matching_target_file,\n matching_target_copy_number, matching_target_name, speciesAndCopyList,\n outAncestor3dir, model_type)\n', (3235, 3365), False, 'from util.calculatedCRBrateAndEstimationAccuracy import calculatedCRBrateAndEstimationAccuracy\n'), ((3866, 3996), 'models.MultiGMPmodel.MultiCopyGMPmodel', 'MultiCopyGMPmodel', (['species_file_list', 'outAncestor2dir', 'guided_species_for_matching', 'ancestor_name', 'ancestor_target_copy_number'], {}), '(species_file_list, outAncestor2dir,\n guided_species_for_matching, ancestor_name, ancestor_target_copy_number)\n', (3883, 3996), False, 'from models.MultiGMPmodel import MultiCopyGMPmodel\n'), ((4463, 4631), 'util.calculatedCRBrateAndEstimationAccuracy.calculatedCRBrateAndEstimationAccuracy', 'calculatedCRBrateAndEstimationAccuracy', (['matching_target_file', 'matching_target_copy_number', 'matching_target_name', 'speciesAndCopyList', 'outAncestor2dir', 'model_type'], {}), '(matching_target_file,\n matching_target_copy_number, matching_target_name, speciesAndCopyList,\n outAncestor2dir, model_type)\n', (4501, 4631), False, 'from util.calculatedCRBrateAndEstimationAccuracy import calculatedCRBrateAndEstimationAccuracy\n'), ((5129, 5259), 'models.MultiGMPmodel.MultiCopyGMPmodel', 'MultiCopyGMPmodel', (['species_file_list', 'outAncestor1dir', 'guided_species_for_matching', 'ancestor_name', 'ancestor_target_copy_number'], {}), '(species_file_list, outAncestor1dir,\n guided_species_for_matching, ancestor_name, ancestor_target_copy_number)\n', (5146, 5259), False, 'from models.MultiGMPmodel import MultiCopyGMPmodel\n'), ((5714, 5882), 'util.calculatedCRBrateAndEstimationAccuracy.calculatedCRBrateAndEstimationAccuracy', 'calculatedCRBrateAndEstimationAccuracy', (['matching_target_file', 'matching_target_copy_number', 'matching_target_name', 'speciesAndCopyList', 'outAncestor1dir', 'model_type'], {}), '(matching_target_file,\n matching_target_copy_number, matching_target_name, speciesAndCopyList,\n outAncestor1dir, model_type)\n', (5752, 5882), False, 'from util.calculatedCRBrateAndEstimationAccuracy import calculatedCRBrateAndEstimationAccuracy\n')]
from django.conf.urls import url import views urlpatterns = [ url(r'^user/', views.get_user, name='user'), url(r'^all/', views.get_all, name='all'), url(r'^fix/$', views.get_fix, name='fix'), ]
[ "django.conf.urls.url" ]
[((67, 109), 'django.conf.urls.url', 'url', (['"""^user/"""', 'views.get_user'], {'name': '"""user"""'}), "('^user/', views.get_user, name='user')\n", (70, 109), False, 'from django.conf.urls import url\n'), ((116, 155), 'django.conf.urls.url', 'url', (['"""^all/"""', 'views.get_all'], {'name': '"""all"""'}), "('^all/', views.get_all, name='all')\n", (119, 155), False, 'from django.conf.urls import url\n'), ((162, 202), 'django.conf.urls.url', 'url', (['"""^fix/$"""', 'views.get_fix'], {'name': '"""fix"""'}), "('^fix/$', views.get_fix, name='fix')\n", (165, 202), False, 'from django.conf.urls import url\n')]
import os import argparse import json import tqdm import numpy as np from program_synthesis.datasets import dataset, executor parser = argparse.ArgumentParser() parser.add_argument("--data-pickle", default='data/karel/val.pkl') parser.add_argument("--input-file", required=True, help="file containing list of beams, each of which is a list of programs") parser.add_argument("--output-file", required=True) args = parser.parse_args() assert not os.path.exists(args.output_file) with open(args.input_file) as f: programs = json.load(f) examples = dataset.KarelTorchDataset( args.data_pickle, lambda x: x) def evaluate_code(eg, beam): exe = executor.KarelExecutor() tests = [] tests += list(eg.input_tests) tests += list(eg.tests) stats = executor.evaluate_code(beam[0], eg.schema.args, tests, exe.execute) prediction = dict( output=beam[0], beams=beam, beams_correct=[executor.evaluate_code(hypothesis, eg.schema.args, tests, exe.execute) for hypothesis in beam], is_correct=stats['correct'] == stats['total'], individual=stats['individual'], passes_given_tests=all(stats['individual'][:len(eg.input_tests)]) ) return prediction result = [evaluate_code(eg, beams) for eg, beams in zip(examples, tqdm.tqdm(programs))] with open(args.output_file, "w") as f: json.dump(result, f)
[ "os.path.exists", "program_synthesis.datasets.executor.KarelExecutor", "argparse.ArgumentParser", "program_synthesis.datasets.dataset.KarelTorchDataset", "tqdm.tqdm", "program_synthesis.datasets.executor.evaluate_code", "json.load", "json.dump" ]
[((140, 165), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {}), '()\n', (163, 165), False, 'import argparse\n'), ((558, 614), 'program_synthesis.datasets.dataset.KarelTorchDataset', 'dataset.KarelTorchDataset', (['args.data_pickle', '(lambda x: x)'], {}), '(args.data_pickle, lambda x: x)\n', (583, 614), False, 'from program_synthesis.datasets import dataset, executor\n'), ((451, 483), 'os.path.exists', 'os.path.exists', (['args.output_file'], {}), '(args.output_file)\n', (465, 483), False, 'import os\n'), ((533, 545), 'json.load', 'json.load', (['f'], {}), '(f)\n', (542, 545), False, 'import json\n'), ((665, 689), 'program_synthesis.datasets.executor.KarelExecutor', 'executor.KarelExecutor', ([], {}), '()\n', (687, 689), False, 'from program_synthesis.datasets import dataset, executor\n'), ((780, 847), 'program_synthesis.datasets.executor.evaluate_code', 'executor.evaluate_code', (['beam[0]', 'eg.schema.args', 'tests', 'exe.execute'], {}), '(beam[0], eg.schema.args, tests, exe.execute)\n', (802, 847), False, 'from program_synthesis.datasets import dataset, executor\n'), ((1365, 1385), 'json.dump', 'json.dump', (['result', 'f'], {}), '(result, f)\n', (1374, 1385), False, 'import json\n'), ((1299, 1318), 'tqdm.tqdm', 'tqdm.tqdm', (['programs'], {}), '(programs)\n', (1308, 1318), False, 'import tqdm\n'), ((938, 1008), 'program_synthesis.datasets.executor.evaluate_code', 'executor.evaluate_code', (['hypothesis', 'eg.schema.args', 'tests', 'exe.execute'], {}), '(hypothesis, eg.schema.args, tests, exe.execute)\n', (960, 1008), False, 'from program_synthesis.datasets import dataset, executor\n')]
# Copyright 2013 The Chromium Authors. All rights reserved. # Use of this source code is governed by a BSD-style license that can be # found in the LICENSE file. from contrib.cluster_telemetry import ct_benchmarks_util from contrib.cluster_telemetry import page_set from contrib.cluster_telemetry import repaint_helpers from benchmarks import rasterize_and_record_micro # pylint: disable=protected-access class RasterizeAndRecordMicroCT( rasterize_and_record_micro._RasterizeAndRecordMicro): """Measures rasterize and record performance for Cluster Telemetry.""" @classmethod def Name(cls): return 'rasterize_and_record_micro_ct' @classmethod def AddBenchmarkCommandLineArgs(cls, parser): (rasterize_and_record_micro._RasterizeAndRecordMicro. AddBenchmarkCommandLineArgs(parser)) ct_benchmarks_util.AddBenchmarkCommandLineArgs(parser) @classmethod def ProcessCommandLineArgs(cls, parser, args): ct_benchmarks_util.ValidateCommandLineArgs(parser, args) def CreateStorySet(self, options): return page_set.CTPageSet( options.urls_list, options.user_agent, options.archive_data_file, run_page_interaction_callback=repaint_helpers.WaitThenRepaint)
[ "contrib.cluster_telemetry.ct_benchmarks_util.ValidateCommandLineArgs", "contrib.cluster_telemetry.page_set.CTPageSet", "benchmarks.rasterize_and_record_micro._RasterizeAndRecordMicro.AddBenchmarkCommandLineArgs", "contrib.cluster_telemetry.ct_benchmarks_util.AddBenchmarkCommandLineArgs" ]
[((718, 810), 'benchmarks.rasterize_and_record_micro._RasterizeAndRecordMicro.AddBenchmarkCommandLineArgs', 'rasterize_and_record_micro._RasterizeAndRecordMicro.AddBenchmarkCommandLineArgs', (['parser'], {}), '(\n parser)\n', (797, 810), False, 'from benchmarks import rasterize_and_record_micro\n'), ((820, 874), 'contrib.cluster_telemetry.ct_benchmarks_util.AddBenchmarkCommandLineArgs', 'ct_benchmarks_util.AddBenchmarkCommandLineArgs', (['parser'], {}), '(parser)\n', (866, 874), False, 'from contrib.cluster_telemetry import ct_benchmarks_util\n'), ((944, 1000), 'contrib.cluster_telemetry.ct_benchmarks_util.ValidateCommandLineArgs', 'ct_benchmarks_util.ValidateCommandLineArgs', (['parser', 'args'], {}), '(parser, args)\n', (986, 1000), False, 'from contrib.cluster_telemetry import ct_benchmarks_util\n'), ((1050, 1207), 'contrib.cluster_telemetry.page_set.CTPageSet', 'page_set.CTPageSet', (['options.urls_list', 'options.user_agent', 'options.archive_data_file'], {'run_page_interaction_callback': 'repaint_helpers.WaitThenRepaint'}), '(options.urls_list, options.user_agent, options.\n archive_data_file, run_page_interaction_callback=repaint_helpers.\n WaitThenRepaint)\n', (1068, 1207), False, 'from contrib.cluster_telemetry import page_set\n')]
import queue ROOM_ENTERED = 'roomEntered' class Event: def __init__(self, eventType='', userParam=dict()): self.type = eventType self.userParam = userParam class EventSystem: def __init__(self): self._eventQueue = queue.Queue() self._eventHandlers = dict() def registerEventHander(self, eventType, callback): ''' Register a handler to be called on the given event type. eventType specifies the type of event the handler should process. callback specifies the function that should be called on the event. Its function header should look like "def myCallback(event):" Returns the ID of the handler. ''' if not eventType in self._eventHandlers: self._eventHandlers[eventType] = [] handlerID = len(self._eventHandlers[eventType]) self._eventHandlers[eventType].append(callback) return handlerID def unregisterEventHandler(self, eventType, handlerID): ''' Unregister a handler, so it won't be called on the specified event. eventType specifies the type of event the handler should process. handlerID specifies the ID of the handler, which should be unregistered. The ID was returned by the corresponding register-function. Returns True on success, else False. ''' if not eventType in self._eventHandlers: return False if handlerID >= len(self._eventHandlers[eventType]): return False self._eventHandlers[eventType].pop(handlerID) return True def createEvent(self, event): self._eventQueue.put_nowait(event) def processEvents(self): while not self._eventQueue.empty(): event = self._eventQueue.get_nowait() # check if eventhandler wants to process event if not event.type in self._eventHandlers: continue for cb in self._eventHandlers[event.type]: cb(event)
[ "queue.Queue" ]
[((251, 264), 'queue.Queue', 'queue.Queue', ([], {}), '()\n', (262, 264), False, 'import queue\n')]
import os import FWCore.ParameterSet.Config as cms process = cms.Process('TauDQMOffline') process.load("FWCore.MessageLogger.MessageLogger_cfi") process.MessageLogger.cerr.FwkReport.reportEvery = 1000 # import of standard configurations from Configuration.StandardSequences.GeometryRecoDB_cff import * process.load('Configuration.StandardSequences.GeometryRecoDB_cff') process.load('Configuration.StandardSequences.MagneticField_cff') process.load('Configuration.StandardSequences.Reconstruction_cff') process.load('Configuration.StandardSequences.FrontierConditions_GlobalTag_cff') process.load('Configuration.StandardSequences.Services_cff') process.load('Configuration.EventContent.EventContent_cff') process.GlobalTag.globaltag = '94X_dataRun2_ReReco_EOY17_v6' #process.load("DQMServices.Components.DQMStoreStats_cfi") #process.load('DQMOffline.Configuration.DQMOffline_cff') readFiles = cms.untracked.vstring() secFiles = cms.untracked.vstring() process.source = cms.Source ("PoolSource",fileNames = readFiles, secondaryFileNames = secFiles) readFiles.extend( [ '/store/data/Run2017D/Tau/MINIAOD/31Mar2018-v1/00000/02FE19AF-3837-E811-B3FF-44A842B4520B.root' ] ); process.maxEvents = cms.untracked.PSet( input = cms.untracked.int32( -1 ) ) process.load('Validation.RecoTau.RecoTauValidation_cff') #process.load('Validation.RecoTau.DQMSequences_cfi') #process.load('Validation.RecoTau.RecoTauValidationMiniAOD_cfi') process.dump = cms.EDAnalyzer("EventContentAnalyzer") process.DQMoutput = cms.OutputModule("DQMRootOutputModule", dataset = cms.untracked.PSet( dataTier = cms.untracked.string('DQMIO'), filterName = cms.untracked.string('') ), fileName = cms.untracked.string('RECO_RAW2DIGI_L1Reco_RECO_EI_PAT_DQM_inDQM.root'), outputCommands = process.DQMEventContent.outputCommands, splitLevel = cms.untracked.int32(0) ) #Setup FWK for multithreaded process.options.numberOfThreads=cms.untracked.uint32(8) process.options.numberOfStreams=cms.untracked.uint32(0) #Have logErrorHarvester wait for the same EDProducers to finish as those providing data for the OutputModule from FWCore.Modules.logErrorHarvester_cff import customiseLogErrorHarvesterUsingOutputCommands process = customiseLogErrorHarvesterUsingOutputCommands(process)
[ "FWCore.ParameterSet.Config.untracked.string", "FWCore.Modules.logErrorHarvester_cff.customiseLogErrorHarvesterUsingOutputCommands", "FWCore.ParameterSet.Config.Source", "FWCore.ParameterSet.Config.untracked.int32", "FWCore.ParameterSet.Config.untracked.uint32", "FWCore.ParameterSet.Config.Process", "FW...
[((63, 91), 'FWCore.ParameterSet.Config.Process', 'cms.Process', (['"""TauDQMOffline"""'], {}), "('TauDQMOffline')\n", (74, 91), True, 'import FWCore.ParameterSet.Config as cms\n'), ((902, 925), 'FWCore.ParameterSet.Config.untracked.vstring', 'cms.untracked.vstring', ([], {}), '()\n', (923, 925), True, 'import FWCore.ParameterSet.Config as cms\n'), ((937, 960), 'FWCore.ParameterSet.Config.untracked.vstring', 'cms.untracked.vstring', ([], {}), '()\n', (958, 960), True, 'import FWCore.ParameterSet.Config as cms\n'), ((979, 1053), 'FWCore.ParameterSet.Config.Source', 'cms.Source', (['"""PoolSource"""'], {'fileNames': 'readFiles', 'secondaryFileNames': 'secFiles'}), "('PoolSource', fileNames=readFiles, secondaryFileNames=secFiles)\n", (989, 1053), True, 'import FWCore.ParameterSet.Config as cms\n'), ((1461, 1499), 'FWCore.ParameterSet.Config.EDAnalyzer', 'cms.EDAnalyzer', (['"""EventContentAnalyzer"""'], {}), "('EventContentAnalyzer')\n", (1475, 1499), True, 'import FWCore.ParameterSet.Config as cms\n'), ((1953, 1976), 'FWCore.ParameterSet.Config.untracked.uint32', 'cms.untracked.uint32', (['(8)'], {}), '(8)\n', (1973, 1976), True, 'import FWCore.ParameterSet.Config as cms\n'), ((2009, 2032), 'FWCore.ParameterSet.Config.untracked.uint32', 'cms.untracked.uint32', (['(0)'], {}), '(0)\n', (2029, 2032), True, 'import FWCore.ParameterSet.Config as cms\n'), ((2248, 2302), 'FWCore.Modules.logErrorHarvester_cff.customiseLogErrorHarvesterUsingOutputCommands', 'customiseLogErrorHarvesterUsingOutputCommands', (['process'], {}), '(process)\n', (2293, 2302), False, 'from FWCore.Modules.logErrorHarvester_cff import customiseLogErrorHarvesterUsingOutputCommands\n'), ((1241, 1264), 'FWCore.ParameterSet.Config.untracked.int32', 'cms.untracked.int32', (['(-1)'], {}), '(-1)\n', (1260, 1264), True, 'import FWCore.ParameterSet.Config as cms\n'), ((1714, 1785), 'FWCore.ParameterSet.Config.untracked.string', 'cms.untracked.string', (['"""RECO_RAW2DIGI_L1Reco_RECO_EI_PAT_DQM_inDQM.root"""'], {}), "('RECO_RAW2DIGI_L1Reco_RECO_EI_PAT_DQM_inDQM.root')\n", (1734, 1785), True, 'import FWCore.ParameterSet.Config as cms\n'), ((1865, 1887), 'FWCore.ParameterSet.Config.untracked.int32', 'cms.untracked.int32', (['(0)'], {}), '(0)\n', (1884, 1887), True, 'import FWCore.ParameterSet.Config as cms\n'), ((1615, 1644), 'FWCore.ParameterSet.Config.untracked.string', 'cms.untracked.string', (['"""DQMIO"""'], {}), "('DQMIO')\n", (1635, 1644), True, 'import FWCore.ParameterSet.Config as cms\n'), ((1667, 1691), 'FWCore.ParameterSet.Config.untracked.string', 'cms.untracked.string', (['""""""'], {}), "('')\n", (1687, 1691), True, 'import FWCore.ParameterSet.Config as cms\n')]
import os import numpy as np import tensorflow as tf import tensorflow_addons as tfa from loguru import logger import config from Train import train from Model import EEGNet class OptunaTrainer: def __init__(self, checkpointPath, epochs, batchsize, logPath=None): self.checkpointPath = checkpointPath self.logpath = logPath self.epochs = epochs self.batchsize = batchsize def __call__(self, trial, dataset, crossVal=False, **kwargs): if isinstance(dataset, tuple): dataset = { "noname": dataset } info = "Trial #{} metric values:\n".format(trial.number) metrics = [] for key, value in dataset.items(): if "augmenter" in kwargs: kwargs["augmenter"].setState(False) shape = list(value[0].shape[-2:]) shape[1] = int(config.window[1] * config.sampleRate) - int(config.window[0] * config.sampleRate) model = self.buildModel(trial, shape) auc, precision = train( model=model, dataset=value, weightsPath=self.checkpointPath, epochs=self.epochs, batchsize=self.batchsize, crossVal=crossVal, **kwargs ) info += "{}: auc {:.2f} pr {:.2f}\t".format(key, auc, precision) metrics.append((auc, precision)) metrics = np.array(metrics) mean = np.mean(metrics, axis=0).round(2) median = np.median(metrics, axis=0).round(2) for i, metric in enumerate(["auc", "precision"]): info += "\nMetric - {}. Mean: {}\tMedian: {}".format(metric, mean[i], median[i]) logger.info(info) logger.info(trial.params) return mean[0] @staticmethod def chooseOptimizer(trial): kwargs = {} optimizer_options = ["RMSprop", "Adam"] optimizer_selected = trial.suggest_categorical("optimizer", optimizer_options) if optimizer_selected == "RMSprop": kwargs["learning_rate"] = trial.suggest_loguniform("rmsprop_learning_rate", 1e-5, 1e-1) kwargs["decay"] = trial.suggest_discrete_uniform("rmsprop_decay", 0.85, 0.99, 0.01) kwargs["momentum"] = trial.suggest_loguniform("rmsprop_momentum", 1e-5, 1e-1) elif optimizer_selected == "Adam": kwargs["learning_rate"] = trial.suggest_loguniform("adam_learning_rate", 1e-5, 1e-1) optimizer = getattr(tf.optimizers, optimizer_selected)(**kwargs) return optimizer @staticmethod def chooseLoss(trial): loss_functions = { "binary_crossentropy": tf, "sigmoid_focal_crossentropy": tfa } loss_selected = trial.suggest_categorical("loss", list(loss_functions.keys())) loss = getattr(loss_functions[loss_selected].losses, loss_selected) return loss def buildModel(self, trial, shape): samples = shape[-1] assert samples // 2 > 16 temporalLength = int(trial.suggest_discrete_uniform("temporal_length", 16, samples // 2, 4)) dropoutRate = trial.suggest_discrete_uniform("dropout_rate", 0.1, 0.5, 0.05) D = trial.suggest_int("depth_multiplier", 1, 4) poolKernel = int(trial.suggest_discrete_uniform("pool_kernel", 4, 16, 2)) model = EEGNet( categoriesN=2, electrodes=shape[0], samples=shape[1], temporalLength=temporalLength, dropoutRate=dropoutRate, D=D, poolPad="same", poolKernel=poolKernel ) optimizer = self.chooseOptimizer(trial) loss = self.chooseLoss(trial) model.compile( loss=loss, optimizer=optimizer, metrics=["accuracy"] ) return model def studyInfo(study, bestN=7, file=None): logger.info("Number of finished trials: {}", len(study.trials)) logger.info("Best {} trials:".format(bestN)) trials = sorted(study.trials, key=lambda elem: elem.value, reverse=True)[:bestN] for i, trial in enumerate(trials): logger.info("Trial {}", i) logger.info("\tValue: {:.2f}", trial.value) logger.info("\tParams: ") for key, value in trial.params.items(): logger.info("\t\t{}: {}", key, value) if file is not None: os.makedirs(os.path.dirname(file), exist_ok=True) studyDF = study.trials_dataframe() studyDF.to_csv(file) logger.info("Study file has been written to {}", file)
[ "numpy.mean", "numpy.median", "loguru.logger.info", "Train.train", "os.path.dirname", "numpy.array", "Model.EEGNet" ]
[((1200, 1217), 'numpy.array', 'np.array', (['metrics'], {}), '(metrics)\n', (1208, 1217), True, 'import numpy as np\n'), ((1449, 1466), 'loguru.logger.info', 'logger.info', (['info'], {}), '(info)\n', (1460, 1466), False, 'from loguru import logger\n'), ((1469, 1494), 'loguru.logger.info', 'logger.info', (['trial.params'], {}), '(trial.params)\n', (1480, 1494), False, 'from loguru import logger\n'), ((2908, 3077), 'Model.EEGNet', 'EEGNet', ([], {'categoriesN': '(2)', 'electrodes': 'shape[0]', 'samples': 'shape[1]', 'temporalLength': 'temporalLength', 'dropoutRate': 'dropoutRate', 'D': 'D', 'poolPad': '"""same"""', 'poolKernel': 'poolKernel'}), "(categoriesN=2, electrodes=shape[0], samples=shape[1], temporalLength\n =temporalLength, dropoutRate=dropoutRate, D=D, poolPad='same',\n poolKernel=poolKernel)\n", (2914, 3077), False, 'from Model import EEGNet\n'), ((3549, 3575), 'loguru.logger.info', 'logger.info', (['"""Trial {}"""', 'i'], {}), "('Trial {}', i)\n", (3560, 3575), False, 'from loguru import logger\n'), ((3578, 3621), 'loguru.logger.info', 'logger.info', (['"""\tValue: {:.2f}"""', 'trial.value'], {}), "('\\tValue: {:.2f}', trial.value)\n", (3589, 3621), False, 'from loguru import logger\n'), ((3625, 3650), 'loguru.logger.info', 'logger.info', (['"""\tParams: """'], {}), "('\\tParams: ')\n", (3636, 3650), False, 'from loguru import logger\n'), ((3872, 3926), 'loguru.logger.info', 'logger.info', (['"""Study file has been written to {}"""', 'file'], {}), "('Study file has been written to {}', file)\n", (3883, 3926), False, 'from loguru import logger\n'), ((907, 1053), 'Train.train', 'train', ([], {'model': 'model', 'dataset': 'value', 'weightsPath': 'self.checkpointPath', 'epochs': 'self.epochs', 'batchsize': 'self.batchsize', 'crossVal': 'crossVal'}), '(model=model, dataset=value, weightsPath=self.checkpointPath, epochs=\n self.epochs, batchsize=self.batchsize, crossVal=crossVal, **kwargs)\n', (912, 1053), False, 'from Train import train\n'), ((3696, 3733), 'loguru.logger.info', 'logger.info', (['"""\t\t{}: {}"""', 'key', 'value'], {}), "('\\t\\t{}: {}', key, value)\n", (3707, 3733), False, 'from loguru import logger\n'), ((3771, 3792), 'os.path.dirname', 'os.path.dirname', (['file'], {}), '(file)\n', (3786, 3792), False, 'import os\n'), ((1228, 1252), 'numpy.mean', 'np.mean', (['metrics'], {'axis': '(0)'}), '(metrics, axis=0)\n', (1235, 1252), True, 'import numpy as np\n'), ((1273, 1299), 'numpy.median', 'np.median', (['metrics'], {'axis': '(0)'}), '(metrics, axis=0)\n', (1282, 1299), True, 'import numpy as np\n')]
from PyQt5 import QtGui, QtWidgets, QtCore class RoundPixmapStyle(QtWidgets.QProxyStyle): def __init__(self, radius=10, *args, **kwargs): super(RoundPixmapStyle, self).__init__(*args, **kwargs) self._radius = radius def drawItemPixmap(self, painter, rectangle, alignment, pixmap): painter.save() pix = QtGui.QPixmap(pixmap.size()) pix.fill(QtCore.Qt.transparent) p = QtGui.QPainter(pix) p.setBrush(QtGui.QBrush(pixmap)) p.setPen(QtCore.Qt.NoPen) p.drawRoundedRect(pixmap.rect(), self._radius, self._radius) p.end() super(RoundPixmapStyle, self).drawItemPixmap(painter, rectangle, alignment, pix) painter.restore()
[ "PyQt5.QtGui.QBrush", "PyQt5.QtGui.QPainter" ]
[((426, 445), 'PyQt5.QtGui.QPainter', 'QtGui.QPainter', (['pix'], {}), '(pix)\n', (440, 445), False, 'from PyQt5 import QtGui, QtWidgets, QtCore\n'), ((465, 485), 'PyQt5.QtGui.QBrush', 'QtGui.QBrush', (['pixmap'], {}), '(pixmap)\n', (477, 485), False, 'from PyQt5 import QtGui, QtWidgets, QtCore\n')]
from qcodes import Instrument try: from spirack import D4_module except ImportError: raise ImportError(('The D4_module class could not be found. ' 'Try installing it using pip install spirack')) from functools import partial class D4(Instrument): """ Qcodes driver for the D4 ADC SPI-rack module. Requires installation of the 'spirack' package using pip. Args: name (str): name of the instrument. spi_rack (SPI_rack): instance of the SPI_rack class as defined in the spirack package. This class manages communication with the individual modules. module (int): module number as set on the hardware. """ def __init__(self, name, spi_rack, module, **kwargs): super().__init__(name, **kwargs) self.d4 = D4_module(spi_rack, module) self.add_parameter('mode', label='Mode', get_cmd=self.get_mode) self.add_parameter('filter_value', label='Filter value', get_cmd=self.get_filter_value) self.add_parameter('buffers_enabled', label='Buffers enabled', get_cmd=self.get_buffers_enabled) for i in range(2): self.add_parameter('adc{}'.format(i + 1), label='ADC {}'.format(i + 1), get_cmd=partial(self.d4.singleConversion, i), units='V') def get_mode(self): return self.d4.mode def get_filter_value(self): return self.d4.filter_val def get_buffers_enabled(self): return self.d4.buf_en
[ "functools.partial", "spirack.D4_module" ]
[((827, 854), 'spirack.D4_module', 'D4_module', (['spi_rack', 'module'], {}), '(spi_rack, module)\n', (836, 854), False, 'from spirack import D4_module\n'), ((1475, 1511), 'functools.partial', 'partial', (['self.d4.singleConversion', 'i'], {}), '(self.d4.singleConversion, i)\n', (1482, 1511), False, 'from functools import partial\n')]
from typing import Optional import yaml from returns.result import Result, Success, Failure from typistry.protos.invalid_object import InvalidObject, IgnorableObject from typistry.protos.typed_dict import TypedDict def safe_parse_yaml(file: str, include_file: bool = False, filter_type: Optional[str] = None) -> Result[TypedDict, InvalidObject]: try: with open(file, 'r') as stream: try: yaml_load = yaml.safe_load(stream) if isinstance(yaml_load, dict): to_type = yaml_load.get("type") if isinstance(to_type, str): if (filter_type == None) or (to_type == filter_type): yaml_load.pop("type") if include_file: yaml_load["source"] = file return Success(TypedDict(yaml_load, type=to_type)) else: return Failure(IgnorableObject(f"Parsed object type: {to_type} does not match specified filter_type {filter_type}", file)) else: return Failure(IgnorableObject("Invalid YAML {file}: {yaml_load}. Parsed object must contain 'type'", file)) else: return Failure(InvalidObject(f"\nInvalid YAML {file}: {yaml_load}. Parsed object must be a dict", file)) except yaml.YAMLError as exc: return Failure(InvalidObject(f"\nInvalid YAML {file}: {exc}\n", file)) except FileNotFoundError as e: return Failure(InvalidObject(f"Specified YAML does not exist: {e}", file))
[ "yaml.safe_load", "typistry.protos.typed_dict.TypedDict", "typistry.protos.invalid_object.IgnorableObject", "typistry.protos.invalid_object.InvalidObject" ]
[((444, 466), 'yaml.safe_load', 'yaml.safe_load', (['stream'], {}), '(stream)\n', (458, 466), False, 'import yaml\n'), ((1603, 1661), 'typistry.protos.invalid_object.InvalidObject', 'InvalidObject', (['f"""Specified YAML does not exist: {e}"""', 'file'], {}), "(f'Specified YAML does not exist: {e}', file)\n", (1616, 1661), False, 'from typistry.protos.invalid_object import InvalidObject, IgnorableObject\n'), ((1325, 1426), 'typistry.protos.invalid_object.InvalidObject', 'InvalidObject', (['f"""\nInvalid YAML {file}: {yaml_load}. Parsed object must be a dict"""', 'file'], {}), '(\n f"""\nInvalid YAML {file}: {yaml_load}. Parsed object must be a dict""",\n file)\n', (1338, 1426), False, 'from typistry.protos.invalid_object import InvalidObject, IgnorableObject\n'), ((1489, 1545), 'typistry.protos.invalid_object.InvalidObject', 'InvalidObject', (['f"""\nInvalid YAML {file}: {exc}\n"""', 'file'], {}), '(f"""\nInvalid YAML {file}: {exc}\n""", file)\n', (1502, 1545), False, 'from typistry.protos.invalid_object import InvalidObject, IgnorableObject\n'), ((1173, 1275), 'typistry.protos.invalid_object.IgnorableObject', 'IgnorableObject', (['"""Invalid YAML {file}: {yaml_load}. Parsed object must contain \'type\'"""', 'file'], {}), '(\n "Invalid YAML {file}: {yaml_load}. Parsed object must contain \'type\'",\n file)\n', (1188, 1275), False, 'from typistry.protos.invalid_object import InvalidObject, IgnorableObject\n'), ((891, 925), 'typistry.protos.typed_dict.TypedDict', 'TypedDict', (['yaml_load'], {'type': 'to_type'}), '(yaml_load, type=to_type)\n', (900, 925), False, 'from typistry.protos.typed_dict import TypedDict\n'), ((1000, 1116), 'typistry.protos.invalid_object.IgnorableObject', 'IgnorableObject', (['f"""Parsed object type: {to_type} does not match specified filter_type {filter_type}"""', 'file'], {}), "(\n f'Parsed object type: {to_type} does not match specified filter_type {filter_type}'\n , file)\n", (1015, 1116), False, 'from typistry.protos.invalid_object import InvalidObject, IgnorableObject\n')]
from datetime import datetime import logging from bs4 import BeautifulSoup from db.models import Victim from net.proxy import Proxy from .sitecrawler import SiteCrawler import time class Nefilim(SiteCrawler): actor = "Nefilim" def _handle_page(self, soup): victim_list = soup.find_all("header", class_="entry-header") for victim in victim_list: victim_title = victim.find("h2", class_="entry-title").text.strip() victim_name = victim_title[0:victim_title.find(". Part")] meta = victim.find("div", class_="entry-meta") published = meta.find("time", class_="entry-date").attrs["datetime"] published_dt = datetime.strptime( published.strip()[:-6], "%Y-%m-%dT%H:%M:%S") victim_leak_site = meta.find("span", class_="posted-on").find("a").attrs["href"] q = self.session.query(Victim).filter_by( url=victim_leak_site, site=self.site) if q.count() == 0: # new victim v = Victim(name=victim_name, url=victim_leak_site, published=published_dt, first_seen=datetime.utcnow(), last_seen=datetime.utcnow(), site=self.site) self.session.add(v) self.new_victims.append(v) else: # already seen, update last_seen v = q.first() v.last_seen = datetime.utcnow() self.current_victims.append(v) self.session.commit() # server was timing out so slows it down a bit time.sleep(1.0) def scrape_victims(self): with Proxy() as p: r = p.get(f"{self.url}", headers=self.headers) soup = BeautifulSoup(r.content.decode(), "html.parser") page_count = 0 while True: page_nav = soup.find("div", class_="nav-previous") if page_nav is None: break url = page_nav.find("a").attrs["href"] r = p.get(f"{url}", headers=self.headers) soup = BeautifulSoup(r.content.decode(), "html.parser") self._handle_page(soup)
[ "net.proxy.Proxy", "time.sleep", "datetime.datetime.utcnow" ]
[((1674, 1689), 'time.sleep', 'time.sleep', (['(1.0)'], {}), '(1.0)\n', (1684, 1689), False, 'import time\n'), ((1742, 1749), 'net.proxy.Proxy', 'Proxy', ([], {}), '()\n', (1747, 1749), False, 'from net.proxy import Proxy\n'), ((1511, 1528), 'datetime.datetime.utcnow', 'datetime.utcnow', ([], {}), '()\n', (1526, 1528), False, 'from datetime import datetime\n'), ((1241, 1258), 'datetime.datetime.utcnow', 'datetime.utcnow', ([], {}), '()\n', (1256, 1258), False, 'from datetime import datetime\n'), ((1270, 1287), 'datetime.datetime.utcnow', 'datetime.utcnow', ([], {}), '()\n', (1285, 1287), False, 'from datetime import datetime\n')]
import numpy import sympy from matplotlib import pyplot from sympy.utilities.lambdify import lambdify # Set the font family and size to use for Matplotlib figures. pyplot.rcParams['font.family'] = 'serif' pyplot.rcParams['font.size'] = 16 sympy.init_printing() x, nu, t = sympy.symbols('x nu t') phi = (sympy.exp(-(x - 4 * t)**2 / (4 * nu * (t + 1))) + sympy.exp(-(x - 4 * t - 2 * numpy.pi)**2 / (4 * nu * (t + 1)))) phiprime = phi.diff(x) u = -2 * nu * (phiprime / phi) + 4 u_lamb = lambdify((t, x, nu), u) # Set parameters. nx = 101 # number of spatial grid points L = 2.0 * numpy.pi # length of the domain dx = L / (nx - 1) # spatial grid size nu = 0.07 # viscosity nt = 100 # number of time steps to compute sigma = 0.1 # CFL limit dt = sigma * dx**2 / nu # time-step size # Discretize the domain. x = numpy.linspace(0.0, L, num=nx) # Set initial conditions. t = 0.0 u0 = numpy.array([u_lamb(t, xi, nu) for xi in x]) # Integrate the Burgers' equation in time. u = u0.copy() for n in range(nt): un = u.copy() # Update all interior points. u[1:-1] = (un[1:-1] - un[1:-1] * dt / dx * (un[1:-1] - un[:-2]) + nu * dt / dx**2 * (un[2:] - 2 * un[1:-1] + un[:-2])) # Update boundary points. u[0] = (un[0] - un[0] * dt / dx * (un[0] - un[-1]) + nu * dt / dx**2 * (un[1] - 2 * un[0] + un[-1])) u[-1] = (un[-1] - un[-1] * dt / dx * (un[-1] - un[-2]) + nu * dt / dx**2 * (un[0] - 2 * un[-1] + un[-2])) # Compute the analytical solution. u_analytical = numpy.array([u_lamb(nt * dt, xi, nu) for xi in x]) # Plot the numerical solution along with the analytical solution. pyplot.figure(figsize=(6.0, 4.0)) pyplot.xlabel('x') pyplot.ylabel('u') pyplot.grid() pyplot.plot(x, u, label='Numerical', color='C0', linestyle='-', linewidth=2) pyplot.plot(x, u_analytical, label='Analytical', color='C1', linestyle='--', linewidth=2) pyplot.legend() pyplot.xlim(0.0, L) pyplot.ylim(0.0, 10.0); pyplot.show() pyplot.clf()
[ "sympy.utilities.lambdify.lambdify", "matplotlib.pyplot.grid", "matplotlib.pyplot.ylabel", "matplotlib.pyplot.xlabel", "matplotlib.pyplot.plot", "matplotlib.pyplot.clf", "sympy.init_printing", "sympy.symbols", "numpy.linspace", "matplotlib.pyplot.figure", "sympy.exp", "matplotlib.pyplot.ylim",...
[((241, 262), 'sympy.init_printing', 'sympy.init_printing', ([], {}), '()\n', (260, 262), False, 'import sympy\n'), ((275, 298), 'sympy.symbols', 'sympy.symbols', (['"""x nu t"""'], {}), "('x nu t')\n", (288, 298), False, 'import sympy\n'), ((496, 519), 'sympy.utilities.lambdify.lambdify', 'lambdify', (['(t, x, nu)', 'u'], {}), '((t, x, nu), u)\n', (504, 519), False, 'from sympy.utilities.lambdify import lambdify\n'), ((827, 857), 'numpy.linspace', 'numpy.linspace', (['(0.0)', 'L'], {'num': 'nx'}), '(0.0, L, num=nx)\n', (841, 857), False, 'import numpy\n'), ((1690, 1723), 'matplotlib.pyplot.figure', 'pyplot.figure', ([], {'figsize': '(6.0, 4.0)'}), '(figsize=(6.0, 4.0))\n', (1703, 1723), False, 'from matplotlib import pyplot\n'), ((1724, 1742), 'matplotlib.pyplot.xlabel', 'pyplot.xlabel', (['"""x"""'], {}), "('x')\n", (1737, 1742), False, 'from matplotlib import pyplot\n'), ((1743, 1761), 'matplotlib.pyplot.ylabel', 'pyplot.ylabel', (['"""u"""'], {}), "('u')\n", (1756, 1761), False, 'from matplotlib import pyplot\n'), ((1762, 1775), 'matplotlib.pyplot.grid', 'pyplot.grid', ([], {}), '()\n', (1773, 1775), False, 'from matplotlib import pyplot\n'), ((1776, 1852), 'matplotlib.pyplot.plot', 'pyplot.plot', (['x', 'u'], {'label': '"""Numerical"""', 'color': '"""C0"""', 'linestyle': '"""-"""', 'linewidth': '(2)'}), "(x, u, label='Numerical', color='C0', linestyle='-', linewidth=2)\n", (1787, 1852), False, 'from matplotlib import pyplot\n'), ((1865, 1958), 'matplotlib.pyplot.plot', 'pyplot.plot', (['x', 'u_analytical'], {'label': '"""Analytical"""', 'color': '"""C1"""', 'linestyle': '"""--"""', 'linewidth': '(2)'}), "(x, u_analytical, label='Analytical', color='C1', linestyle='--',\n linewidth=2)\n", (1876, 1958), False, 'from matplotlib import pyplot\n'), ((1967, 1982), 'matplotlib.pyplot.legend', 'pyplot.legend', ([], {}), '()\n', (1980, 1982), False, 'from matplotlib import pyplot\n'), ((1983, 2002), 'matplotlib.pyplot.xlim', 'pyplot.xlim', (['(0.0)', 'L'], {}), '(0.0, L)\n', (1994, 2002), False, 'from matplotlib import pyplot\n'), ((2003, 2025), 'matplotlib.pyplot.ylim', 'pyplot.ylim', (['(0.0)', '(10.0)'], {}), '(0.0, 10.0)\n', (2014, 2025), False, 'from matplotlib import pyplot\n'), ((2027, 2040), 'matplotlib.pyplot.show', 'pyplot.show', ([], {}), '()\n', (2038, 2040), False, 'from matplotlib import pyplot\n'), ((2041, 2053), 'matplotlib.pyplot.clf', 'pyplot.clf', ([], {}), '()\n', (2051, 2053), False, 'from matplotlib import pyplot\n'), ((306, 355), 'sympy.exp', 'sympy.exp', (['(-(x - 4 * t) ** 2 / (4 * nu * (t + 1)))'], {}), '(-(x - 4 * t) ** 2 / (4 * nu * (t + 1)))\n', (315, 355), False, 'import sympy\n'), ((363, 427), 'sympy.exp', 'sympy.exp', (['(-(x - 4 * t - 2 * numpy.pi) ** 2 / (4 * nu * (t + 1)))'], {}), '(-(x - 4 * t - 2 * numpy.pi) ** 2 / (4 * nu * (t + 1)))\n', (372, 427), False, 'import sympy\n')]
from django import forms from django.forms import ModelForm from .models import Category,Image, Location class ImageForm(forms.ModelForm): class Meta: model = Image fields = '__all__' CATEGORIES =( ("1", "Cars"), ("2", "Food"), ("3", "Travel"), ("4", "Animals"), ("5", "Nature"), ("6", "Sports"), ) class ImagesForm(forms.Form): image = forms.ImageField(required=True) pic_name = forms.CharField(required=True) description = forms.CharField(required=True) location = forms.CharField(required=True) pic_category = forms.ChoiceField(choices=CATEGORIES, required=True)
[ "django.forms.ImageField", "django.forms.ChoiceField", "django.forms.CharField" ]
[((398, 429), 'django.forms.ImageField', 'forms.ImageField', ([], {'required': '(True)'}), '(required=True)\n', (414, 429), False, 'from django import forms\n'), ((446, 476), 'django.forms.CharField', 'forms.CharField', ([], {'required': '(True)'}), '(required=True)\n', (461, 476), False, 'from django import forms\n'), ((495, 525), 'django.forms.CharField', 'forms.CharField', ([], {'required': '(True)'}), '(required=True)\n', (510, 525), False, 'from django import forms\n'), ((541, 571), 'django.forms.CharField', 'forms.CharField', ([], {'required': '(True)'}), '(required=True)\n', (556, 571), False, 'from django import forms\n'), ((591, 643), 'django.forms.ChoiceField', 'forms.ChoiceField', ([], {'choices': 'CATEGORIES', 'required': '(True)'}), '(choices=CATEGORIES, required=True)\n', (608, 643), False, 'from django import forms\n')]
from .base import FunctionalTest from selenium import webdriver from selenium.webdriver.common.keys import Keys class NewVisitorTest(FunctionalTest): def test_can_start_a_list_and_retrieve_it_later(self): self.browser.get(self.server_url) self.assertIn("To-Do", self.browser.title) header_text = self.browser.find_element_by_tag_name("h1").text self.assertIn("To-Do", header_text) inputbox = self.get_item_input_box() self.assertEqual( inputbox.get_attribute("placeholder"), "Enter a to-do item" ) inputbox.send_keys("Buy peacock feathers") inputbox.send_keys(Keys.ENTER) edith_list_url = self.browser.current_url self.assertRegex(edith_list_url, "/lists/.+") self.check_for_row_in_list_table('1: Buy peacock feathers') inputbox = self.get_item_input_box() inputbox.send_keys("Use peacock feathers to make a fly") inputbox.send_keys(Keys.ENTER) self.check_for_row_in_list_table("1: Buy peacock feathers") self.check_for_row_in_list_table( "2: Use peacock feathers to make a fly" ) self.browser.quit() self.browser = webdriver.Firefox() self.browser.get(self.server_url) page_text = self.browser.find_element_by_tag_name("body").text self.assertNotIn("Buy peacock feathers", page_text) self.assertNotIn("make a fly", page_text) inputbox = self.get_item_input_box() inputbox.send_keys("Buy milk") inputbox.send_keys(Keys.ENTER) francis_list_url = self.browser.current_url self.assertRegex(francis_list_url, "/lists/.+") self.assertNotEqual(francis_list_url, edith_list_url) page_text = self.browser.find_element_by_tag_name("body").text self.assertNotIn("Buy peacock feathers", page_text) self.assertNotIn("make a fly", page_text)
[ "selenium.webdriver.Firefox" ]
[((1226, 1245), 'selenium.webdriver.Firefox', 'webdriver.Firefox', ([], {}), '()\n', (1243, 1245), False, 'from selenium import webdriver\n')]
# Copyright 2021 Google LLC # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Tests for reservoir_nn.keras.rewiring.""" from absl.testing import absltest from absl.testing import parameterized import numpy as np from reservoir_nn.keras import rewiring import tensorflow as tf class AdaptiveSparseReservoirTest(parameterized.TestCase): def test_layer_with_num_connections_works(self): x = tf.constant([1.]) layer = rewiring.AdaptiveSparseReservoir( units=10, reservoir_initializer=10, ) layer(x) def test_layer_with_large_num_connections_fails(self): x = tf.constant([1.]) layer = rewiring.AdaptiveSparseReservoir( units=10, reservoir_initializer=100, ) with self.assertRaisesRegex(ValueError, "Cannot build layer.*(100).*(1).*(10)"): layer(x) def test_layer_with_reservoir_works(self): initial_reservoir = np.arange(10).reshape(1, 10) x = tf.constant([1.]) layer = rewiring.AdaptiveSparseReservoir( units=10, reservoir_initializer=initial_reservoir, ) np.testing.assert_array_equal(layer(x), np.arange(10)) def test_layer_with_misshaped_reservoir_fails(self): initial_reservoir = np.arange(10).reshape(2, 5) x = tf.constant([1.]) layer = rewiring.AdaptiveSparseReservoir( units=10, reservoir_initializer=initial_reservoir, ) with self.assertRaisesRegex( ValueError, r"Reservoir has a shape of \(2, 5\), but the layer expects \(1, 10\)"): layer(x) def test_get_coo_weight_matrix_works(self): initial_reservoir = np.arange(10).reshape(2, 5) x = tf.keras.Input(shape=(2,)) layer = rewiring.AdaptiveSparseReservoir( units=5, reservoir_initializer=initial_reservoir, ) layer(x) coo = layer.get_coo_weight_matrix().toarray() np.testing.assert_array_equal(coo, initial_reservoir) def test_get_coo_age_matrix_works(self): initial_reservoir = np.arange(10).reshape(2, 5) x = tf.keras.Input(shape=(2,)) layer = rewiring.AdaptiveSparseReservoir( units=5, reservoir_initializer=initial_reservoir, ) layer(x) coo = layer.get_coo_age_matrix().toarray() np.testing.assert_array_equal(coo, np.zeros((2, 5))) policy = rewiring.MutationPolicy( candidate_fraction=0.0, candidate_mutation_rate=1.0, ) policy.mutation_step(layer) coo = layer.get_coo_age_matrix().toarray() np.testing.assert_array_equal( coo, [[0., 1, 1, 1, 1], [1., 1, 1, 1, 1]], ) def test_apply_global_policy_works(self): policy = rewiring.MutationPolicy( candidate_fraction=0.5, candidate_mutation_rate=0.5, ) gpolicy = rewiring.GlobalPolicy( scale_candidate_fraction=0.5, scale_candidate_mutation_rate=0.5) policy = policy.apply_global_policy(gpolicy) self.assertEqual( policy, rewiring.MutationPolicy( candidate_fraction=0.25, candidate_mutation_rate=0.25)) def test_compute_mutation_probability_works(self): initial_reservoir = np.arange(10).reshape(2, 5) x = tf.keras.Input(shape=(2,)) layer = rewiring.AdaptiveSparseReservoir( units=5, reservoir_initializer=initial_reservoir, ) layer(x) policy = rewiring.MutationPolicy( candidate_fraction=1.0, candidate_mutation_rate=1.0, ) p = policy.compute_mutation_probability( sparse_values=layer.sparse_values.value(), sparse_ages=layer.sparse_ages.value(), ) coo = layer.get_coo_weight_matrix().copy() coo.data[:] = p np.testing.assert_allclose( coo.toarray(), [ [0., 0.7, 0.3, 0, 0], [0., 0, 0, 0, 0], ], atol=0.1) def test_mutation_works(self): initial_reservoir = np.arange(10).reshape(2, 5) x = tf.keras.Input(shape=(2,)) layer = rewiring.AdaptiveSparseReservoir( units=5, reservoir_initializer=initial_reservoir, ) layer(x) coo = layer.get_coo_weight_matrix() np.testing.assert_allclose(coo.toarray(), [ [0., 1, 2, 3, 4], [5., 6, 7, 8, 9], ]) rng = np.random.RandomState(1234) policy = rewiring.MutationPolicy( candidate_fraction=0.2, candidate_mutation_rate=1.0, ) policy.mutation_step(layer, rng) coo = layer.get_coo_weight_matrix() # least active connections are replenished with zeros with 100% # probability. np.testing.assert_allclose(coo.toarray(), [ [0., 0, 2, 3, 4], [5., 6, 7, 8, 9], ]) class SparseEvolutionEnd2EndTest(parameterized.TestCase): def test_fit_small_model_works(self): """Testing fitting a 4x4 sparse reservoir with 4 truth connections.""" nunits = 4 inputs = tf.keras.Input(shape=(nunits,)) policy = rewiring.MutationPolicy( candidate_fraction=0.5, candidate_mutation_rate=0.1, ) layer = rewiring.AdaptiveSparseReservoir( units=nunits, reservoir_initializer=2 * nunits, # regularizer helps sparsify the redundant connections. kernel_regularizer=tf.keras.regularizers.l2(1e-2), ) outputs = layer(inputs) rng = np.random.RandomState(1333) model = tf.keras.Model(inputs, outputs) # Low dimension model prefers SGD: model.compile( optimizer=tf.keras.optimizers.SGD(lr=0.1, nesterov=True), loss="mse", metrics=["mse"]) x = rng.uniform(size=(1000000, nunits)) - 0.5 y = x[:, ::-1].copy() truth = np.eye(4)[::-1] def mutation_schedule(epoch): del epoch return rewiring.GlobalPolicy() model.fit( x, y, batch_size=int(len(x) / 100), epochs=10, verbose=False, callbacks=rewiring.MutationCallback( policy={layer: policy}, mutation_schedule=mutation_schedule, rng=rng, verbose=1)) connection = layer.get_coo_weight_matrix().toarray() # Use truth * 0.65 here because L2 biases the fit towards zero. np.testing.assert_allclose(connection, truth * 0.65, atol=0.10) def test_fit_large_model_works(self): """Testing fitting a 100x100 sparse reservoir with 100 truth connections.""" nunits = 100 inputs = tf.keras.Input(shape=(nunits,)) policy = rewiring.MutationPolicy( candidate_fraction=0.2, candidate_mutation_rate=0.8, ) layer = rewiring.AdaptiveSparseReservoir( units=nunits, reservoir_initializer=2 * nunits, ) outputs = layer(inputs) model = tf.keras.Model(inputs, outputs) model.compile(loss="mse", metrics=["mse"]) x = np.random.uniform(size=(100000, nunits)) - 0.5 truth = np.eye(nunits)[::-1].copy() y = np.einsum("ij,jk->ik", x, truth) class Reporter(tf.keras.callbacks.Callback): def on_epoch_end(self, epoch, logs): connection = layer.get_coo_weight_matrix().toarray() ages = layer.get_coo_age_matrix().toarray() cross = np.abs(connection) * truth print(sorted(zip(*np.nonzero(cross)))) significant_elements = np.sum(cross > 0.02) print(significant_elements) print(ages[cross > 0.02]) model.fit( x, y, batch_size=1600, epochs=20, verbose=True, callbacks=[ rewiring.MutationCallback(policy=policy, verbose=1), Reporter() ]) connection = layer.get_coo_weight_matrix().todense() # We shall have some correlation with the truth after training for a while. cross = np.abs(connection) * truth np.testing.assert_allclose(np.sum(cross > 0.02), 80, atol=20) if __name__ == "__main__": absltest.main()
[ "numpy.einsum", "reservoir_nn.keras.rewiring.AdaptiveSparseReservoir", "numpy.arange", "numpy.random.RandomState", "numpy.testing.assert_allclose", "tensorflow.keras.optimizers.SGD", "reservoir_nn.keras.rewiring.GlobalPolicy", "reservoir_nn.keras.rewiring.MutationPolicy", "numpy.testing.assert_array...
[((8276, 8291), 'absl.testing.absltest.main', 'absltest.main', ([], {}), '()\n', (8289, 8291), False, 'from absl.testing import absltest\n'), ((899, 917), 'tensorflow.constant', 'tf.constant', (['[1.0]'], {}), '([1.0])\n', (910, 917), True, 'import tensorflow as tf\n'), ((929, 997), 'reservoir_nn.keras.rewiring.AdaptiveSparseReservoir', 'rewiring.AdaptiveSparseReservoir', ([], {'units': '(10)', 'reservoir_initializer': '(10)'}), '(units=10, reservoir_initializer=10)\n', (961, 997), False, 'from reservoir_nn.keras import rewiring\n'), ((1100, 1118), 'tensorflow.constant', 'tf.constant', (['[1.0]'], {}), '([1.0])\n', (1111, 1118), True, 'import tensorflow as tf\n'), ((1130, 1199), 'reservoir_nn.keras.rewiring.AdaptiveSparseReservoir', 'rewiring.AdaptiveSparseReservoir', ([], {'units': '(10)', 'reservoir_initializer': '(100)'}), '(units=10, reservoir_initializer=100)\n', (1162, 1199), False, 'from reservoir_nn.keras import rewiring\n'), ((1464, 1482), 'tensorflow.constant', 'tf.constant', (['[1.0]'], {}), '([1.0])\n', (1475, 1482), True, 'import tensorflow as tf\n'), ((1494, 1582), 'reservoir_nn.keras.rewiring.AdaptiveSparseReservoir', 'rewiring.AdaptiveSparseReservoir', ([], {'units': '(10)', 'reservoir_initializer': 'initial_reservoir'}), '(units=10, reservoir_initializer=\n initial_reservoir)\n', (1526, 1582), False, 'from reservoir_nn.keras import rewiring\n'), ((1778, 1796), 'tensorflow.constant', 'tf.constant', (['[1.0]'], {}), '([1.0])\n', (1789, 1796), True, 'import tensorflow as tf\n'), ((1808, 1896), 'reservoir_nn.keras.rewiring.AdaptiveSparseReservoir', 'rewiring.AdaptiveSparseReservoir', ([], {'units': '(10)', 'reservoir_initializer': 'initial_reservoir'}), '(units=10, reservoir_initializer=\n initial_reservoir)\n', (1840, 1896), False, 'from reservoir_nn.keras import rewiring\n'), ((2172, 2198), 'tensorflow.keras.Input', 'tf.keras.Input', ([], {'shape': '(2,)'}), '(shape=(2,))\n', (2186, 2198), True, 'import tensorflow as tf\n'), ((2211, 2298), 'reservoir_nn.keras.rewiring.AdaptiveSparseReservoir', 'rewiring.AdaptiveSparseReservoir', ([], {'units': '(5)', 'reservoir_initializer': 'initial_reservoir'}), '(units=5, reservoir_initializer=\n initial_reservoir)\n', (2243, 2298), False, 'from reservoir_nn.keras import rewiring\n'), ((2385, 2438), 'numpy.testing.assert_array_equal', 'np.testing.assert_array_equal', (['coo', 'initial_reservoir'], {}), '(coo, initial_reservoir)\n', (2414, 2438), True, 'import numpy as np\n'), ((2544, 2570), 'tensorflow.keras.Input', 'tf.keras.Input', ([], {'shape': '(2,)'}), '(shape=(2,))\n', (2558, 2570), True, 'import tensorflow as tf\n'), ((2583, 2670), 'reservoir_nn.keras.rewiring.AdaptiveSparseReservoir', 'rewiring.AdaptiveSparseReservoir', ([], {'units': '(5)', 'reservoir_initializer': 'initial_reservoir'}), '(units=5, reservoir_initializer=\n initial_reservoir)\n', (2615, 2670), False, 'from reservoir_nn.keras import rewiring\n'), ((2821, 2897), 'reservoir_nn.keras.rewiring.MutationPolicy', 'rewiring.MutationPolicy', ([], {'candidate_fraction': '(0.0)', 'candidate_mutation_rate': '(1.0)'}), '(candidate_fraction=0.0, candidate_mutation_rate=1.0)\n', (2844, 2897), False, 'from reservoir_nn.keras import rewiring\n'), ((3004, 3078), 'numpy.testing.assert_array_equal', 'np.testing.assert_array_equal', (['coo', '[[0.0, 1, 1, 1, 1], [1.0, 1, 1, 1, 1]]'], {}), '(coo, [[0.0, 1, 1, 1, 1], [1.0, 1, 1, 1, 1]])\n', (3033, 3078), True, 'import numpy as np\n'), ((3158, 3234), 'reservoir_nn.keras.rewiring.MutationPolicy', 'rewiring.MutationPolicy', ([], {'candidate_fraction': '(0.5)', 'candidate_mutation_rate': '(0.5)'}), '(candidate_fraction=0.5, candidate_mutation_rate=0.5)\n', (3181, 3234), False, 'from reservoir_nn.keras import rewiring\n'), ((3272, 3362), 'reservoir_nn.keras.rewiring.GlobalPolicy', 'rewiring.GlobalPolicy', ([], {'scale_candidate_fraction': '(0.5)', 'scale_candidate_mutation_rate': '(0.5)'}), '(scale_candidate_fraction=0.5,\n scale_candidate_mutation_rate=0.5)\n', (3293, 3362), False, 'from reservoir_nn.keras import rewiring\n'), ((3673, 3699), 'tensorflow.keras.Input', 'tf.keras.Input', ([], {'shape': '(2,)'}), '(shape=(2,))\n', (3687, 3699), True, 'import tensorflow as tf\n'), ((3712, 3799), 'reservoir_nn.keras.rewiring.AdaptiveSparseReservoir', 'rewiring.AdaptiveSparseReservoir', ([], {'units': '(5)', 'reservoir_initializer': 'initial_reservoir'}), '(units=5, reservoir_initializer=\n initial_reservoir)\n', (3744, 3799), False, 'from reservoir_nn.keras import rewiring\n'), ((3846, 3922), 'reservoir_nn.keras.rewiring.MutationPolicy', 'rewiring.MutationPolicy', ([], {'candidate_fraction': '(1.0)', 'candidate_mutation_rate': '(1.0)'}), '(candidate_fraction=1.0, candidate_mutation_rate=1.0)\n', (3869, 3922), False, 'from reservoir_nn.keras import rewiring\n'), ((4402, 4428), 'tensorflow.keras.Input', 'tf.keras.Input', ([], {'shape': '(2,)'}), '(shape=(2,))\n', (4416, 4428), True, 'import tensorflow as tf\n'), ((4441, 4528), 'reservoir_nn.keras.rewiring.AdaptiveSparseReservoir', 'rewiring.AdaptiveSparseReservoir', ([], {'units': '(5)', 'reservoir_initializer': 'initial_reservoir'}), '(units=5, reservoir_initializer=\n initial_reservoir)\n', (4473, 4528), False, 'from reservoir_nn.keras import rewiring\n'), ((4718, 4745), 'numpy.random.RandomState', 'np.random.RandomState', (['(1234)'], {}), '(1234)\n', (4739, 4745), True, 'import numpy as np\n'), ((4760, 4836), 'reservoir_nn.keras.rewiring.MutationPolicy', 'rewiring.MutationPolicy', ([], {'candidate_fraction': '(0.2)', 'candidate_mutation_rate': '(1.0)'}), '(candidate_fraction=0.2, candidate_mutation_rate=1.0)\n', (4783, 4836), False, 'from reservoir_nn.keras import rewiring\n'), ((5338, 5369), 'tensorflow.keras.Input', 'tf.keras.Input', ([], {'shape': '(nunits,)'}), '(shape=(nunits,))\n', (5352, 5369), True, 'import tensorflow as tf\n'), ((5384, 5460), 'reservoir_nn.keras.rewiring.MutationPolicy', 'rewiring.MutationPolicy', ([], {'candidate_fraction': '(0.5)', 'candidate_mutation_rate': '(0.1)'}), '(candidate_fraction=0.5, candidate_mutation_rate=0.1)\n', (5407, 5460), False, 'from reservoir_nn.keras import rewiring\n'), ((5762, 5789), 'numpy.random.RandomState', 'np.random.RandomState', (['(1333)'], {}), '(1333)\n', (5783, 5789), True, 'import numpy as np\n'), ((5803, 5834), 'tensorflow.keras.Model', 'tf.keras.Model', (['inputs', 'outputs'], {}), '(inputs, outputs)\n', (5817, 5834), True, 'import tensorflow as tf\n'), ((6623, 6685), 'numpy.testing.assert_allclose', 'np.testing.assert_allclose', (['connection', '(truth * 0.65)'], {'atol': '(0.1)'}), '(connection, truth * 0.65, atol=0.1)\n', (6649, 6685), True, 'import numpy as np\n'), ((6839, 6870), 'tensorflow.keras.Input', 'tf.keras.Input', ([], {'shape': '(nunits,)'}), '(shape=(nunits,))\n', (6853, 6870), True, 'import tensorflow as tf\n'), ((6884, 6960), 'reservoir_nn.keras.rewiring.MutationPolicy', 'rewiring.MutationPolicy', ([], {'candidate_fraction': '(0.2)', 'candidate_mutation_rate': '(0.8)'}), '(candidate_fraction=0.2, candidate_mutation_rate=0.8)\n', (6907, 6960), False, 'from reservoir_nn.keras import rewiring\n'), ((6996, 7081), 'reservoir_nn.keras.rewiring.AdaptiveSparseReservoir', 'rewiring.AdaptiveSparseReservoir', ([], {'units': 'nunits', 'reservoir_initializer': '(2 * nunits)'}), '(units=nunits, reservoir_initializer=2 * nunits\n )\n', (7028, 7081), False, 'from reservoir_nn.keras import rewiring\n'), ((7141, 7172), 'tensorflow.keras.Model', 'tf.keras.Model', (['inputs', 'outputs'], {}), '(inputs, outputs)\n', (7155, 7172), True, 'import tensorflow as tf\n'), ((7326, 7358), 'numpy.einsum', 'np.einsum', (['"""ij,jk->ik"""', 'x', 'truth'], {}), "('ij,jk->ik', x, truth)\n", (7335, 7358), True, 'import numpy as np\n'), ((1646, 1659), 'numpy.arange', 'np.arange', (['(10)'], {}), '(10)\n', (1655, 1659), True, 'import numpy as np\n'), ((2789, 2805), 'numpy.zeros', 'np.zeros', (['(2, 5)'], {}), '((2, 5))\n', (2797, 2805), True, 'import numpy as np\n'), ((3465, 3543), 'reservoir_nn.keras.rewiring.MutationPolicy', 'rewiring.MutationPolicy', ([], {'candidate_fraction': '(0.25)', 'candidate_mutation_rate': '(0.25)'}), '(candidate_fraction=0.25, candidate_mutation_rate=0.25)\n', (3488, 3543), False, 'from reservoir_nn.keras import rewiring\n'), ((6095, 6104), 'numpy.eye', 'np.eye', (['(4)'], {}), '(4)\n', (6101, 6104), True, 'import numpy as np\n'), ((6175, 6198), 'reservoir_nn.keras.rewiring.GlobalPolicy', 'rewiring.GlobalPolicy', ([], {}), '()\n', (6196, 6198), False, 'from reservoir_nn.keras import rewiring\n'), ((7230, 7270), 'numpy.random.uniform', 'np.random.uniform', ([], {'size': '(100000, nunits)'}), '(size=(100000, nunits))\n', (7247, 7270), True, 'import numpy as np\n'), ((8152, 8170), 'numpy.abs', 'np.abs', (['connection'], {}), '(connection)\n', (8158, 8170), True, 'import numpy as np\n'), ((8210, 8230), 'numpy.sum', 'np.sum', (['(cross > 0.02)'], {}), '(cross > 0.02)\n', (8216, 8230), True, 'import numpy as np\n'), ((1426, 1439), 'numpy.arange', 'np.arange', (['(10)'], {}), '(10)\n', (1435, 1439), True, 'import numpy as np\n'), ((1741, 1754), 'numpy.arange', 'np.arange', (['(10)'], {}), '(10)\n', (1750, 1754), True, 'import numpy as np\n'), ((2135, 2148), 'numpy.arange', 'np.arange', (['(10)'], {}), '(10)\n', (2144, 2148), True, 'import numpy as np\n'), ((2507, 2520), 'numpy.arange', 'np.arange', (['(10)'], {}), '(10)\n', (2516, 2520), True, 'import numpy as np\n'), ((3636, 3649), 'numpy.arange', 'np.arange', (['(10)'], {}), '(10)\n', (3645, 3649), True, 'import numpy as np\n'), ((4365, 4378), 'numpy.arange', 'np.arange', (['(10)'], {}), '(10)\n', (4374, 4378), True, 'import numpy as np\n'), ((5685, 5715), 'tensorflow.keras.regularizers.l2', 'tf.keras.regularizers.l2', (['(0.01)'], {}), '(0.01)\n', (5709, 5715), True, 'import tensorflow as tf\n'), ((5912, 5958), 'tensorflow.keras.optimizers.SGD', 'tf.keras.optimizers.SGD', ([], {'lr': '(0.1)', 'nesterov': '(True)'}), '(lr=0.1, nesterov=True)\n', (5935, 5958), True, 'import tensorflow as tf\n'), ((6335, 6446), 'reservoir_nn.keras.rewiring.MutationCallback', 'rewiring.MutationCallback', ([], {'policy': '{layer: policy}', 'mutation_schedule': 'mutation_schedule', 'rng': 'rng', 'verbose': '(1)'}), '(policy={layer: policy}, mutation_schedule=\n mutation_schedule, rng=rng, verbose=1)\n', (6360, 6446), False, 'from reservoir_nn.keras import rewiring\n'), ((7687, 7707), 'numpy.sum', 'np.sum', (['(cross > 0.02)'], {}), '(cross > 0.02)\n', (7693, 7707), True, 'import numpy as np\n'), ((7289, 7303), 'numpy.eye', 'np.eye', (['nunits'], {}), '(nunits)\n', (7295, 7303), True, 'import numpy as np\n'), ((7582, 7600), 'numpy.abs', 'np.abs', (['connection'], {}), '(connection)\n', (7588, 7600), True, 'import numpy as np\n'), ((7914, 7965), 'reservoir_nn.keras.rewiring.MutationCallback', 'rewiring.MutationCallback', ([], {'policy': 'policy', 'verbose': '(1)'}), '(policy=policy, verbose=1)\n', (7939, 7965), False, 'from reservoir_nn.keras import rewiring\n'), ((7635, 7652), 'numpy.nonzero', 'np.nonzero', (['cross'], {}), '(cross)\n', (7645, 7652), True, 'import numpy as np\n')]
from __future__ import absolute_import from __future__ import division from __future__ import print_function import cv2 import numpy as np from progress.bar import Bar import torch import math import copy from .base_debugger import BaseDebugger from models.utils import _tranpose_and_gather_feat, _gather_feat from models.decode import _topk_original, _topk, _topk_channel, _nms from datasets.dataset.utils import _bbox_overlaps from utils.image import transform_preds class TriCtdetDebugger(BaseDebugger): def __init__(self, opt): super(TriCtdetDebugger, self).__init__(opt) def forward(self, images): with torch.no_grad(): output = self.model(images)[-1] tl = output['tl'].sigmoid_() bl = output['bl'].sigmoid_() br = output['br'].sigmoid_() ct = output['ct'].sigmoid_() tl_tag = output['tl_tag'] bl_tag = output['bl_tag'] br_tag = output['br_tag'] tl_reg = output['tl_reg'] bl_reg = output['bl_reg'] br_reg = output['br_reg'] ct_reg = output['ct_reg'] detections = {'tl_heatmap':tl, 'bl_heatmap':bl, 'br_heatmap':br, 'ct_heatmap':ct, 'tl_reg':tl_reg, 'bl_reg':bl_reg, 'br_reg':br_reg, 'ct_reg':ct_reg, 'tl_tag':tl_tag, 'bl_tag':bl_tag, 'br_tag':br_tag} return detections def debug(self, detections, targets, ae_threshold): tl_heat = detections['tl_heatmap'] bl_heat = detections['bl_heatmap'] br_heat = detections['br_heatmap'] ct_heat = detections['ct_heatmap'] targets['tl_tag'] = targets['tl_tag'][targets['reg_mask']].unsqueeze(0) targets['bl_tag'] = targets['bl_tag'][targets['reg_mask']].unsqueeze(0) targets['br_tag'] = targets['br_tag'][targets['reg_mask']].unsqueeze(0) targets['ct_tag'] = targets['ct_tag'][targets['reg_mask']].unsqueeze(0) targets['tl_reg'] = targets['tl_reg'][targets['reg_mask']].unsqueeze(0) targets['bl_reg'] = targets['bl_reg'][targets['reg_mask']].unsqueeze(0) targets['br_reg'] = targets['br_reg'][targets['reg_mask']].unsqueeze(0) targets['ct_reg'] = targets['ct_reg'][targets['reg_mask']].unsqueeze(0) batch, cat, height, width = tl_heat.size() # tl_scores, tl_inds, tl_clses, tl_ys, tl_xs = _topk(tl_heat, K=256) # bl_scores, bl_inds, bl_clses, bl_ys, bl_xs = _topk(bl_heat, K=256) # br_scores, br_inds, br_clses, br_ys, br_xs = _topk(br_heat, K=256) # ct_scores, ct_inds, ct_clses, ct_ys, ct_xs = _topk(ct_heat, K=256) tl_tag = detections['tl_tag'] bl_tag = detections['bl_tag'] br_tag = detections['br_tag'] tl_reg = detections['tl_reg'] bl_reg = detections['bl_reg'] br_reg = detections['br_reg'] ct_reg = detections['ct_reg'] # gather by gt tl_tag = _tranpose_and_gather_feat(tl_tag, targets['tl_tag'].to(torch.device("cuda"))) bl_tag = _tranpose_and_gather_feat(bl_tag, targets['bl_tag'].to(torch.device("cuda"))) br_tag = _tranpose_and_gather_feat(br_tag, targets['br_tag'].to(torch.device("cuda"))) # gather by top k # tl_tag = _tranpose_and_gather_feat(tl_tag, tl_inds) # bl_tag = _tranpose_and_gather_feat(bl_tag, bl_inds) # br_tag = _tranpose_and_gather_feat(br_tag, br_inds) avg_tag = (tl_tag + bl_tag + br_tag) / 3 dists_tl = torch.abs(avg_tag - tl_tag).to(torch.device("cpu")).numpy() dists_bl = torch.abs(bl_tag - avg_tag).to(torch.device("cpu")).numpy() dists_br = torch.abs(avg_tag - br_tag).to(torch.device("cpu")).numpy() dists_avg = (dists_tl.sum() + dists_bl.sum() + dists_br.sum()) / dists_tl.shape[1] / 3 min_tl = dists_tl.min() max_tl = dists_tl.max() min_bl = dists_bl.min() max_bl = dists_bl.max() min_br = dists_br.min() max_br = dists_br.max() # gather by gt tl_reg = _tranpose_and_gather_feat(tl_reg, targets['tl_tag'].to(torch.device("cuda"))) bl_reg = _tranpose_and_gather_feat(bl_reg, targets['bl_tag'].to(torch.device("cuda"))) br_reg = _tranpose_and_gather_feat(br_reg, targets['br_tag'].to(torch.device("cuda"))) ct_reg = _tranpose_and_gather_feat(ct_reg, targets['ct_tag'].to(torch.device("cuda"))) # reg_diff_tl = tl_reg - targets['tl_reg'].to(torch.device("cuda")) # reg_diff_tl = torch.sqrt(reg_diff_tl[..., 0]*reg_diff_tl[..., 0] + reg_diff_tl[..., 1]*reg_diff_tl[..., 1]) # reg_diff_bl = bl_reg - targets['bl_reg'].to(torch.device("cuda")) # reg_diff_bl = torch.sqrt(reg_diff_bl[..., 0] * reg_diff_bl[..., 0] + reg_diff_bl[..., 1] * reg_diff_bl[..., 1]) # reg_diff_br = br_reg - targets['br_reg'].to(torch.device("cuda")) # reg_diff_br = torch.sqrt(reg_diff_br[..., 0] * reg_diff_br[..., 0] + reg_diff_br[..., 1] * reg_diff_br[..., 1]) # reg_diff_ct = ct_reg - targets['ct_reg'].to(torch.device("cuda")) # reg_diff_ct = torch.sqrt(reg_diff_ct[..., 0] * reg_diff_ct[..., 0] + reg_diff_ct[..., 1] * reg_diff_ct[..., 1]) tl_xs = ((targets['tl_tag'] % (width * height)) % width).int().float().to(torch.device("cuda")) tl_ys = ((targets['tl_tag'] % (width * height)) / width).int().float().to(torch.device("cuda")) bl_xs = ((targets['bl_tag'] % (width * height)) % width).int().float().to(torch.device("cuda")) bl_ys = ((targets['bl_tag'] % (width * height)) / width).int().float().to(torch.device("cuda")) br_xs = ((targets['br_tag'] % (width * height)) % width).int().float().to(torch.device("cuda")) br_ys = ((targets['br_tag'] % (width * height)) / width).int().float().to(torch.device("cuda")) ct_xs = ((targets['ct_tag'] % (width * height)) % width).int().float().to(torch.device("cuda")) ct_ys = ((targets['ct_tag'] % (width * height)) / width).int().float().to(torch.device("cuda")) tl_xs_pr = (tl_xs + tl_reg[..., 0]).squeeze(0).to(torch.device("cpu")).numpy() tl_ys_pr = (tl_ys + tl_reg[..., 1]).squeeze(0).to(torch.device("cpu")).numpy() bl_xs_pr = (bl_xs + bl_reg[..., 0]).squeeze(0).to(torch.device("cpu")).numpy() bl_ys_pr = (bl_ys + bl_reg[..., 1]).squeeze(0).to(torch.device("cpu")).numpy() br_xs_pr = (br_xs + br_reg[..., 0]).squeeze(0).to(torch.device("cpu")).numpy() br_ys_pr = (br_ys + br_reg[..., 1]).squeeze(0).to(torch.device("cpu")).numpy() ct_xs_pr = (ct_xs + ct_reg[..., 0]).squeeze(0).to(torch.device("cpu")).numpy() ct_ys_pr = (ct_ys + ct_reg[..., 1]).squeeze(0).to(torch.device("cpu")).numpy() tl_xs_gt = (tl_xs + targets['tl_reg'][..., 0].to(torch.device("cuda"))).squeeze(0).to(torch.device("cpu")).numpy() tl_ys_gt = (tl_ys + targets['tl_reg'][..., 1].to(torch.device("cuda"))).squeeze(0).to(torch.device("cpu")).numpy() bl_xs_gt = (bl_xs + targets['bl_reg'][..., 0].to(torch.device("cuda"))).squeeze(0).to(torch.device("cpu")).numpy() bl_ys_gt = (bl_ys + targets['bl_reg'][..., 1].to(torch.device("cuda"))).squeeze(0).to(torch.device("cpu")).numpy() br_xs_gt = (br_xs + targets['br_reg'][..., 0].to(torch.device("cuda"))).squeeze(0).to(torch.device("cpu")).numpy() br_ys_gt = (br_ys + targets['br_reg'][..., 1].to(torch.device("cuda"))).squeeze(0).to(torch.device("cpu")).numpy() ct_xs_gt = (ct_xs + targets['ct_reg'][..., 0].to(torch.device("cuda"))).squeeze(0).to(torch.device("cpu")).numpy() ct_ys_gt = (ct_ys + targets['ct_reg'][..., 1].to(torch.device("cuda"))).squeeze(0).to(torch.device("cpu")).numpy() bboxes_gt = targets['bbox'][targets['reg_mask']] nm_instances =tl_xs_pr.shape[0] for i in range(nm_instances): bbox_gt = bboxes_gt[i, :] # prediction bbox_coord_pr = [] tl_x_pr = tl_xs_pr[i] tl_y_pr = tl_ys_pr[i] bl_x_pr = bl_xs_pr[i] bl_y_pr = bl_ys_pr[i] br_x_pr = br_xs_pr[i] br_y_pr = br_ys_pr[i] # center x_c = (tl_x_pr + br_x_pr) / 2. y_c = (tl_y_pr + br_y_pr) / 2. if bl_x_pr == br_x_pr: p_y = tl_y_pr p_x = br_x_pr if br_y_pr > bl_y_pr: angle = np.pi / 2. else: angle = -np.pi / 2. elif bl_y_pr == br_y_pr: p_x = tl_x_pr p_y = br_y_pr angle = 0. else: # angle angle = math.atan2(-(br_y_pr - bl_y_pr), br_x_pr - bl_x_pr) # find intersected point a = (br_x_pr - bl_x_pr) / (br_y_pr - bl_y_pr) b = br_y_pr - a * br_x_pr delta_x = br_x_pr - bl_x_pr delta_y = br_y_pr - bl_y_pr p_x = (delta_x * tl_x_pr + delta_y * tl_y_pr - delta_x * b) / (delta_x + delta_x * a) p_y = a * p_x + b # w, h w = np.sqrt((br_x_pr - p_x) * (br_x_pr - p_x) + (br_y_pr - p_y) * (br_y_pr - p_y)) h = np.sqrt((tl_x_pr - p_x) * (tl_x_pr - p_x) + (tl_y_pr - p_y) * (tl_y_pr - p_y)) bbox_coord_pr.append([x_c - w / 2, y_c - h / 2, x_c + w / 2, y_c + h / 2, angle]) bbox_coord_pr = np.array(bbox_coord_pr) # groundtruth boxes_coord_gt = [] tl_x_gt = tl_xs_gt[i] tl_y_gt = tl_ys_gt[i] bl_x_gt = bl_xs_gt[i] bl_y_gt = bl_ys_gt[i] br_x_gt = br_xs_gt[i] br_y_gt = br_ys_gt[i] if bl_x_gt == br_x_gt: p_y = tl_y_gt p_x = bl_x_gt if br_y_gt > bl_y_gt: angle = np.pi / 4 else: angle = -np.pi / 4 else: # center x_c = (tl_x_gt + br_x_gt) / 2. y_c = (tl_y_gt + br_y_gt) / 2. # angle angle = math.atan(-(br_y_gt - bl_y_gt)/(br_x_gt - bl_x_gt)) # find intersected point a = (br_y_gt - bl_y_gt) / (br_x_gt - bl_x_gt) b = br_y_gt - a * br_x_gt delta_x = br_x_gt - bl_x_gt delta_y = br_y_gt - bl_y_gt p_x = (delta_x * tl_x_gt + delta_y * tl_y_gt - delta_y * b) / (delta_x + delta_y * a) p_y = a * p_x + b # w, h w = np.sqrt((br_x_gt - p_x) * (br_x_gt - p_x) + (br_y_gt - p_y) * (br_y_gt - p_y)) h = np.sqrt((tl_x_gt - p_x) * (tl_x_gt - p_x) + (tl_y_gt - p_y) * (tl_y_gt - p_y)) boxes_coord_gt.append([x_c - w / 2, y_c - h / 2, x_c + w / 2, y_c + h / 2, angle]) boxes_coord_gt = np.array(boxes_coord_gt) # print(np.array_equal(bbox_gt, boxes_coord_gt)) overlaps = _bbox_overlaps(np.ascontiguousarray(bbox_coord_pr[:, :4], dtype=np.float32), np.ascontiguousarray(boxes_coord_gt[:, :4], dtype=np.float32), bbox_coord_pr[:, -1], boxes_coord_gt[:, -1], 128, 128) flag_suc = False flag_exit = 0 for i in range(overlaps.shape[0]): for j in range(overlaps.shape[1]): value_overlap = overlaps[i, j] angle_diff = math.fabs(bbox_coord_pr[i, -1] - boxes_coord_gt[j, -1]) if value_overlap > 0.25 and angle_diff < np.pi / 6: flag_suc = True flag_exit = 1 break if flag_exit: break if flag_exit: break return min_tl, max_tl, min_bl, max_bl, min_br, max_br, dists_avg, flag_suc def process(self, images, kernel=1, ae_threshold=1, K=100, num_dets=100): with torch.no_grad(): output = self.model(images)[-1] tl_heat = output['tl'].sigmoid_() bl_heat = output['bl'].sigmoid_() br_heat = output['br'].sigmoid_() ct_heat = output['ct'].sigmoid_() tl_tag = output['tl_tag'] bl_tag = output['bl_tag'] br_tag = output['br_tag'] tl_reg = output['tl_reg'] bl_reg = output['bl_reg'] br_reg = output['br_reg'] ct_reg = output['ct_reg'] batch, cat, height, width = tl_heat.size() tl_heat = _nms(tl_heat, kernel=3) bl_heat = _nms(bl_heat, kernel=3) br_heat = _nms(br_heat, kernel=3) ct_heat = _nms(ct_heat, kernel=3) tl_scores, tl_inds, tl_clses, tl_ys, tl_xs = _topk(tl_heat, K=K) bl_scores, bl_inds, bl_clses, bl_ys, bl_xs = _topk(bl_heat, K=K) br_scores, br_inds, br_clses, br_ys, br_xs = _topk(br_heat, K=K) ct_scores, ct_inds, ct_clses, ct_ys, ct_xs = _topk(ct_heat, K=K) tl_ys = tl_ys.view(batch, K, 1, 1).expand(batch, K, K, K) tl_xs = tl_xs.view(batch, K, 1, 1).expand(batch, K, K, K) bl_ys = bl_ys.view(batch, 1, K, 1).expand(batch, K, K, K) bl_xs = bl_xs.view(batch, 1, K, 1).expand(batch, K, K, K) br_ys = br_ys.view(batch, 1, 1, K).expand(batch, K, K, K) br_xs = br_xs.view(batch, 1, 1, K).expand(batch, K, K, K) ct_ys = ct_ys.view(batch, 1, K).expand(batch, K, K) ct_xs = ct_xs.view(batch, 1, K).expand(batch, K, K) if tl_reg is not None and bl_reg is not None and br_reg is not None: tl_reg = _tranpose_and_gather_feat(tl_reg, tl_inds) tl_reg = tl_reg.view(batch, K, 1, 1, 2) bl_reg = _tranpose_and_gather_feat(bl_reg, bl_inds) bl_reg = bl_reg.view(batch, 1, K, 1, 2) br_reg = _tranpose_and_gather_feat(br_reg, br_inds) br_reg = br_reg.view(batch, 1, 1, K, 2) ct_reg = _tranpose_and_gather_feat(ct_reg, ct_inds) ct_reg = ct_reg.view(batch, 1, K, 2) tl_xs = tl_xs + tl_reg[..., 0] tl_ys = tl_ys + tl_reg[..., 1] bl_xs = bl_xs + bl_reg[..., 0] bl_ys = bl_ys + bl_reg[..., 1] br_xs = br_xs + br_reg[..., 0] br_ys = br_ys + br_reg[..., 1] ct_xs = ct_xs + ct_reg[..., 0] ct_ys = ct_ys + ct_reg[..., 1] # all possible boxes based on top k corners (ignoring class) bboxes = torch.stack((tl_xs, tl_ys, bl_xs, bl_ys, br_xs, br_ys), dim=4) tl_tag = _tranpose_and_gather_feat(tl_tag, tl_inds) tl_tag = tl_tag.view(batch, K, 1, 1) bl_tag = _tranpose_and_gather_feat(bl_tag, bl_inds) bl_tag = bl_tag.view(batch, 1, K, 1) br_tag = _tranpose_and_gather_feat(br_tag, br_inds) br_tag = br_tag.view(batch, 1, 1, K) avg_tag = (tl_tag + bl_tag + br_tag) / 3 dists = (torch.abs(tl_tag - avg_tag) + torch.abs(bl_tag - avg_tag) + torch.abs(br_tag - avg_tag)) / 3 tl_scores = tl_scores.view(batch, K, 1, 1).expand(batch, K, K, K) bl_scores = bl_scores.view(batch, 1, K, 1).expand(batch, K, K, K) br_scores = br_scores.view(batch, 1, 1, K).expand(batch, K, K, K) # reject boxes based on corner scores # sc_inds = (tl_scores < scores_thresh) | (bl_scores < scores_thresh) | (br_scores < scores_thresh) scores = (tl_scores + bl_scores + br_scores) / 3 # reject boxes based on classes tl_clses = tl_clses.view(batch, K, 1, 1).expand(batch, K, K, K) bl_clses = bl_clses.view(batch, 1, K, 1).expand(batch, K, K, K) br_clses = br_clses.view(batch, 1, 1, K).expand(batch, K, K, K) cls_inds = (tl_clses != bl_clses) | (bl_clses != br_clses) | (tl_clses != br_clses) # reject boxes based on distances dist_inds = (dists > ae_threshold) scores[cls_inds] = -1 scores[dist_inds] = -1 # scores[sc_inds] = -1 scores = scores.view(batch, -1) scores, inds = torch.topk(scores, num_dets) scores = scores.unsqueeze(2) bboxes = bboxes.view(batch, -1, 6) bboxes = _gather_feat(bboxes, inds) clses = bl_clses.contiguous().view(batch, -1, 1) clses = _gather_feat(clses, inds).float() tl_scores = tl_scores.contiguous().view(batch, -1, 1) tl_scores = _gather_feat(tl_scores, inds).float() bl_scores = bl_scores.contiguous().view(batch, -1, 1) bl_scores = _gather_feat(bl_scores, inds).float() br_scores = br_scores.contiguous().view(batch, -1, 1) br_scores = _gather_feat(br_scores, inds).float() ct_xs = ct_xs[:, 0, :] ct_ys = ct_ys[:, 0, :] centers = torch.cat([ct_xs.unsqueeze(2), ct_ys.unsqueeze(2), ct_clses.float().unsqueeze(2), ct_scores.unsqueeze(2)], dim=2) detections = torch.cat([bboxes, scores, tl_scores, bl_scores, br_scores, clses], dim=2) # tl_heat = output['tl'].sigmoid_() # bl_heat = output['bl'].sigmoid_() # br_heat = output['br'].sigmoid_() # ct_heat = output['ct'].sigmoid_() # # tl_tag = output['tl_tag'] # bl_tag = output['bl_tag'] # br_tag = output['br_tag'] # # tl_reg = output['tl_reg'] # bl_reg = output['bl_reg'] # br_reg = output['br_reg'] # ct_reg = output['ct_reg'] # # kernel = self.opt.nms_kernel # ae_threshold = self.opt.ae_threshold # K = self.opt.K # # batch, cat, height, width = tl_heat.size() # # # perform nms on heatmaps # tl_heat = _nms(tl_heat, kernel=kernel) # bl_heat = _nms(bl_heat, kernel=kernel) # br_heat = _nms(br_heat, kernel=kernel) # ct_heat = _nms(ct_heat, kernel=kernel) # # tl_scores, tl_inds, tl_clses, tl_ys, tl_xs = _topk(tl_heat, K=K) # bl_scores, bl_inds, bl_clses, bl_ys, bl_xs = _topk(bl_heat, K=K) # br_scores, br_inds, br_clses, br_ys, br_xs = _topk(br_heat, K=K) # ct_scores, ct_inds, ct_clses, ct_ys, ct_xs = _topk(ct_heat, K=K) # # tl_ys = tl_ys.view(batch, K, 1, 1).expand(batch, K, K, K) # tl_xs = tl_xs.view(batch, K, 1, 1).expand(batch, K, K, K) # bl_ys = bl_ys.view(batch, 1, K, 1).expand(batch, K, K, K) # bl_xs = bl_xs.view(batch, 1, K, 1).expand(batch, K, K, K) # br_ys = br_ys.view(batch, 1, 1, K).expand(batch, K, K, K) # br_xs = br_xs.view(batch, 1, 1, K).expand(batch, K, K, K) # ct_ys = ct_ys.view(batch, 1, K).expand(batch, K, K) # ct_xs = ct_xs.view(batch, 1, K).expand(batch, K, K) # # if tl_reg is not None and bl_reg is not None and br_reg is not None: # tl_reg = _tranpose_and_gather_feat(tl_reg, tl_inds) # tl_reg = tl_reg.view(batch, K, 1, 1, 2) # bl_reg = _tranpose_and_gather_feat(bl_reg, bl_inds) # bl_reg = bl_reg.view(batch, 1, K, 1, 2) # br_reg = _tranpose_and_gather_feat(br_reg, br_inds) # br_reg = br_reg.view(batch, 1, 1, K, 2) # ct_reg = _tranpose_and_gather_feat(ct_reg, ct_inds) # ct_reg = ct_reg.view(batch, 1, K, 2) # # tl_xs = tl_xs + tl_reg[..., 0] # tl_ys = tl_ys + tl_reg[..., 1] # bl_xs = bl_xs + bl_reg[..., 0] # bl_ys = bl_ys + bl_reg[..., 1] # br_xs = br_xs + br_reg[..., 0] # br_ys = br_ys + br_reg[..., 1] # ct_xs = ct_xs + ct_reg[..., 0] # ct_ys = ct_ys + ct_reg[..., 1] # # # all possible boxes based on top k corners (ignoring class) # bboxes = torch.stack((tl_xs, tl_ys, bl_xs, bl_ys, br_xs, br_ys), dim=4) # # tl_tag = _tranpose_and_gather_feat(tl_tag, tl_inds) # tl_tag = tl_tag.view(batch, K, 1, 1).expand(batch, K, K, K) # bl_tag = _tranpose_and_gather_feat(bl_tag, bl_inds) # bl_tag = bl_tag.view(batch, 1, K, 1).expand(batch, K, K, K) # br_tag = _tranpose_and_gather_feat(br_tag, br_inds) # br_tag = br_tag.view(batch, 1, 1, K).expand(batch, K, K, K) # avg_tag = (tl_tag + bl_tag + br_tag) / 3 # dists = (torch.abs(tl_tag - avg_tag) + torch.abs(bl_tag - avg_tag) + torch.abs(br_tag - avg_tag)) / 3 # # tl_scores = tl_scores.view(batch, K, 1, 1).expand(batch, K, K, K) # bl_scores = bl_scores.view(batch, 1, K, 1).expand(batch, K, K, K) # br_scores = br_scores.view(batch, 1, 1, K).expand(batch, K, K, K) # scores = (tl_scores + bl_scores + br_scores) / 3 # # # reject boxes based on classes # tl_clses = tl_clses.view(batch, K, 1, 1).expand(batch, K, K, K) # bl_clses = bl_clses.view(batch, 1, K, 1).expand(batch, K, K, K) # br_clses = br_clses.view(batch, 1, 1, K).expand(batch, K, K, K) # cls_inds = (tl_clses != bl_clses) | (bl_clses != br_clses) | (tl_clses != br_clses) # # # reject boxes based on distances # dist_inds = (dists > ae_threshold) # # # instead of filtering prediction according to the out-of-bound rotation, do data augmentation to mirror groundtruth # # scores[cls_inds] = -1 # scores[dist_inds] = -1 # # scores = scores.view(batch, -1) # scores, inds = torch.topk(scores, 100) # scores = scores.unsqueeze(2) # # bboxes = bboxes.view(batch, -1, 6) # bboxes = _gather_feat(bboxes, inds) # # tl_tag = tl_tag.contiguous().view(batch, -1, 1) # tl_tag = _gather_feat(tl_tag, inds) # bl_tag = bl_tag.contiguous().view(batch, -1, 1) # bl_tag = _gather_feat(bl_tag, inds) # br_tag = br_tag.contiguous().view(batch, -1, 1) # br_tag = _gather_feat(br_tag, inds) # avg_tag = avg_tag.contiguous().view(batch, -1, 1) # avg_tag = _gather_feat(avg_tag, inds) # # clses = bl_clses.contiguous().view(batch, -1, 1) # clses = _gather_feat(clses, inds).float() # # tl_scores = tl_scores.contiguous().view(batch, -1, 1) # tl_scores = _gather_feat(tl_scores, inds).float() # bl_scores = bl_scores.contiguous().view(batch, -1, 1) # bl_scores = _gather_feat(bl_scores, inds).float() # br_scores = br_scores.contiguous().view(batch, -1, 1) # br_scores = _gather_feat(br_scores, inds).float() # # ct_xs = ct_xs[:, 0, :] # ct_ys = ct_ys[:, 0, :] # # centers = torch.cat( # [ct_xs.unsqueeze(2), ct_ys.unsqueeze(2), ct_clses.float().unsqueeze(2), ct_scores.unsqueeze(2)], dim=2) # detections = torch.cat([bboxes, scores, tl_scores, bl_scores, br_scores, clses, tl_tag, bl_tag, br_tag, avg_tag], dim=2) return detections, centers def post_process(self, detections, centers, num_classes, bbox_size_threshold, ori_threshold): detections = detections.detach().cpu().numpy() centers = centers.detach().cpu().numpy() detections = detections.reshape(1, -1, detections.shape[2]) centers = centers.reshape(1, -1, centers.shape[2]) ret = [] for i in range(detections.shape[0]): top_preds = {} detections[i, :, 0:2] *= 4. detections[i, :, 2:4] *= 4. detections[i, :, 4:6] *= 4. centers[i, :, 0:2] *= 4. # Dump bbox whose central region has no center point detections = np.concatenate(detections, axis=1) centers = np.concatenate(centers, axis=1) # filter by orientation distance between quantized and continuous predicted angle classes = detections[..., -1] quant_ori = (5.0 * classes - 85.0) / 180 * np.pi bl_x = detections[..., 2] bl_y = detections[..., 3] br_x = detections[..., 4] br_y = detections[..., 5] cont_ori = np.arctan(-(br_y - bl_y) / (br_x - bl_x)) dist_ori = np.fabs(quant_ori - cont_ori) ori_ind = dist_ori < ori_threshold valid_detections = detections[ori_ind] valid_ind = valid_detections[:, 6] > -1 valid_detections = valid_detections[valid_ind] # valid_ind = detections[:, 6] > -1 # valid_detections = detections[valid_ind] box_width = np.sqrt(np.power(valid_detections[:, 2] - valid_detections[:, 4], 2) + \ np.power(valid_detections[:, 3] - valid_detections[:, 5], 2)) box_height = np.sqrt(np.power(valid_detections[:, 2] - valid_detections[:, 0], 2) + \ np.power(valid_detections[:, 3] - valid_detections[:, 1], 2)) s_ind = (box_width * box_height <= bbox_size_threshold) l_ind = (box_width * box_height > bbox_size_threshold) s_detections = valid_detections[s_ind] l_detections = valid_detections[l_ind] # pro-process for small bounding box s_tl_x = (2 * s_detections[:, 0] + s_detections[:, 4]) / 3 s_br_x = (s_detections[:, 0] + 2 * s_detections[:, 4]) / 3 s_tl_y = (2 * s_detections[:, 1] + s_detections[:, 5]) / 3 s_br_y = (s_detections[:, 1] + 2 * s_detections[:, 5]) / 3 s_temp_score = copy.copy(s_detections[:, 6]) s_detections[:, 6] = -1 center_x = centers[:, 0][:, np.newaxis] center_y = centers[:, 1][:, np.newaxis] s_tl_x = s_tl_x[np.newaxis, :] s_br_x = s_br_x[np.newaxis, :] s_tl_y = s_tl_y[np.newaxis, :] s_br_y = s_br_y[np.newaxis, :] ind_x1 = (center_x > s_tl_x) & (center_x < s_br_x) ind_x2 = (center_x < s_tl_x) & (center_x > s_br_x) ind_y1 = (center_y > s_tl_y) & (center_y < s_br_y) ind_y2 = (center_y < s_tl_y) & (center_y > s_br_y) ind_cls = (centers[:, 2][:, np.newaxis] - s_detections[:, -1][np.newaxis, :]) == 0 ind_s_new_score = np.max((((ind_x1 + 0) & (ind_y1 + 0) & (ind_cls + 0)) | ((ind_x1 + 0) & (ind_y2 + 0) & (ind_cls + 0)) | ((ind_x2 + 0) & (ind_y2 + 0) & (ind_cls + 0))), axis=0) == 1 index_s_new_score = np.argmax((((ind_x1 + 0) & (ind_y1 + 0) & (ind_cls + 0)) | ((ind_x1 + 0) & (ind_y2 + 0) & (ind_cls + 0)) | ((ind_x2 + 0) & (ind_y2 + 0) & (ind_cls + 0)))[:, ind_s_new_score], axis=0) s_corner_score = s_temp_score[ind_s_new_score] s_center_score = centers[index_s_new_score, 3] s_detections[:, 6][ind_s_new_score] = (s_corner_score * 3 + s_center_score) / 4 # pro-process for large bounding box l_tl_x = (2 * l_detections[:, 0] + l_detections[:, 4]) / 3 l_br_x = (l_detections[:, 0] + 2 * l_detections[:, 4]) / 3 l_tl_y = (2 * l_detections[:, 1] + l_detections[:, 5]) / 3 l_br_y = (l_detections[:, 1] + 2 * l_detections[:, 5]) / 3 l_temp_score = copy.copy(l_detections[:, 6]) l_detections[:, 6] = -1 center_x = centers[:, 0][:, np.newaxis] center_y = centers[:, 1][:, np.newaxis] l_tl_x = l_tl_x[np.newaxis, :] l_br_x = l_br_x[np.newaxis, :] l_tl_y = l_tl_y[np.newaxis, :] l_br_y = l_br_y[np.newaxis, :] ind_x1 = (center_x > l_tl_x) & (center_x < l_br_x) ind_x2 = (center_x < l_tl_x) & (center_x > l_br_x) ind_y1 = (center_y > l_tl_y) & (center_y < l_br_y) ind_y2 = (center_y < l_tl_y) & (center_y > l_br_y) ind_cls = (centers[:, 2][:, np.newaxis] - l_detections[:, -1][np.newaxis, :]) == 0 ind_l_new_score = np.max((((ind_x1 + 0) & (ind_y1 + 0) & (ind_cls + 0)) | ((ind_x1 + 0) & (ind_y2 + 0) & (ind_cls + 0)) | ((ind_x2 + 0) & (ind_y2 + 0) & (ind_cls + 0))), axis=0) == 1 index_l_new_score = np.argmax((((ind_x1 + 0) & (ind_y1 + 0) & (ind_cls + 0)) | ((ind_x1 + 0) & (ind_y2 + 0) & (ind_cls + 0)) | ((ind_x2 + 0) & (ind_y2 + 0) & (ind_cls + 0)))[:, ind_l_new_score], axis=0) l_corner_score = l_temp_score[ind_l_new_score] l_center_score = centers[index_l_new_score, 3] l_detections[:, 6][ind_l_new_score] = (l_corner_score * 3 + l_center_score) / 4 detections = np.concatenate([l_detections, s_detections], axis=0) detections = detections[np.argsort(-detections[:, 6])] classes = detections[..., -1] # reject detections with negative scores keep_inds = (detections[:, 6] > -1) detections = detections[keep_inds] classes = classes[keep_inds] detections = np.expand_dims(detections, axis=0) for j in range(num_classes): inds = (classes == j) top_preds[j + 1] = detections[i, inds, :].astype(np.float32).tolist() ret.append(top_preds) for j in range(1, num_classes + 1): ret[0][j] = np.array(ret[0][j], dtype=np.float32).reshape(-1, 11) return ret[0] def merge_outputs(self, detections): results = {} for j in range(1, self.num_classes + 1): results[j] = np.concatenate( [detection[j] for detection in detections], axis=0).astype(np.float32) if len(self.scales) > 1 or self.opt.nms: soft_nms(results[j], Nt=0.5, method=2) scores = np.hstack( [results[j][:, 6] for j in range(1, self.num_classes + 1)]) if len(scores) > self.max_per_image: kth = len(scores) - self.max_per_image thresh = np.partition(scores, kth)[kth] for j in range(1, self.num_classes + 1): keep_inds = (results[j][:, 6] >= thresh) results[j] = results[j][keep_inds] return results
[ "numpy.sqrt", "models.decode._topk", "models.utils._tranpose_and_gather_feat", "numpy.ascontiguousarray", "numpy.argsort", "numpy.array", "copy.copy", "math.atan", "numpy.partition", "numpy.max", "math.fabs", "numpy.concatenate", "numpy.arctan", "torch.abs", "torch.topk", "numpy.argmax...
[((626, 641), 'torch.no_grad', 'torch.no_grad', ([], {}), '()\n', (639, 641), False, 'import torch\n'), ((5076, 5096), 'torch.device', 'torch.device', (['"""cuda"""'], {}), "('cuda')\n", (5088, 5096), False, 'import torch\n'), ((5178, 5198), 'torch.device', 'torch.device', (['"""cuda"""'], {}), "('cuda')\n", (5190, 5198), False, 'import torch\n'), ((5280, 5300), 'torch.device', 'torch.device', (['"""cuda"""'], {}), "('cuda')\n", (5292, 5300), False, 'import torch\n'), ((5382, 5402), 'torch.device', 'torch.device', (['"""cuda"""'], {}), "('cuda')\n", (5394, 5402), False, 'import torch\n'), ((5484, 5504), 'torch.device', 'torch.device', (['"""cuda"""'], {}), "('cuda')\n", (5496, 5504), False, 'import torch\n'), ((5586, 5606), 'torch.device', 'torch.device', (['"""cuda"""'], {}), "('cuda')\n", (5598, 5606), False, 'import torch\n'), ((5688, 5708), 'torch.device', 'torch.device', (['"""cuda"""'], {}), "('cuda')\n", (5700, 5708), False, 'import torch\n'), ((5790, 5810), 'torch.device', 'torch.device', (['"""cuda"""'], {}), "('cuda')\n", (5802, 5810), False, 'import torch\n'), ((8815, 8893), 'numpy.sqrt', 'np.sqrt', (['((br_x_pr - p_x) * (br_x_pr - p_x) + (br_y_pr - p_y) * (br_y_pr - p_y))'], {}), '((br_x_pr - p_x) * (br_x_pr - p_x) + (br_y_pr - p_y) * (br_y_pr - p_y))\n', (8822, 8893), True, 'import numpy as np\n'), ((8908, 8986), 'numpy.sqrt', 'np.sqrt', (['((tl_x_pr - p_x) * (tl_x_pr - p_x) + (tl_y_pr - p_y) * (tl_y_pr - p_y))'], {}), '((tl_x_pr - p_x) * (tl_x_pr - p_x) + (tl_y_pr - p_y) * (tl_y_pr - p_y))\n', (8915, 8986), True, 'import numpy as np\n'), ((9106, 9129), 'numpy.array', 'np.array', (['bbox_coord_pr'], {}), '(bbox_coord_pr)\n', (9114, 9129), True, 'import numpy as np\n'), ((10210, 10288), 'numpy.sqrt', 'np.sqrt', (['((br_x_gt - p_x) * (br_x_gt - p_x) + (br_y_gt - p_y) * (br_y_gt - p_y))'], {}), '((br_x_gt - p_x) * (br_x_gt - p_x) + (br_y_gt - p_y) * (br_y_gt - p_y))\n', (10217, 10288), True, 'import numpy as np\n'), ((10303, 10381), 'numpy.sqrt', 'np.sqrt', (['((tl_x_gt - p_x) * (tl_x_gt - p_x) + (tl_y_gt - p_y) * (tl_y_gt - p_y))'], {}), '((tl_x_gt - p_x) * (tl_x_gt - p_x) + (tl_y_gt - p_y) * (tl_y_gt - p_y))\n', (10310, 10381), True, 'import numpy as np\n'), ((10502, 10526), 'numpy.array', 'np.array', (['boxes_coord_gt'], {}), '(boxes_coord_gt)\n', (10510, 10526), True, 'import numpy as np\n'), ((11632, 11647), 'torch.no_grad', 'torch.no_grad', ([], {}), '()\n', (11645, 11647), False, 'import torch\n'), ((12197, 12220), 'models.decode._nms', '_nms', (['tl_heat'], {'kernel': '(3)'}), '(tl_heat, kernel=3)\n', (12201, 12220), False, 'from models.decode import _topk_original, _topk, _topk_channel, _nms\n'), ((12241, 12264), 'models.decode._nms', '_nms', (['bl_heat'], {'kernel': '(3)'}), '(bl_heat, kernel=3)\n', (12245, 12264), False, 'from models.decode import _topk_original, _topk, _topk_channel, _nms\n'), ((12285, 12308), 'models.decode._nms', '_nms', (['br_heat'], {'kernel': '(3)'}), '(br_heat, kernel=3)\n', (12289, 12308), False, 'from models.decode import _topk_original, _topk, _topk_channel, _nms\n'), ((12329, 12352), 'models.decode._nms', '_nms', (['ct_heat'], {'kernel': '(3)'}), '(ct_heat, kernel=3)\n', (12333, 12352), False, 'from models.decode import _topk_original, _topk, _topk_channel, _nms\n'), ((12409, 12428), 'models.decode._topk', '_topk', (['tl_heat'], {'K': 'K'}), '(tl_heat, K=K)\n', (12414, 12428), False, 'from models.decode import _topk_original, _topk, _topk_channel, _nms\n'), ((12484, 12503), 'models.decode._topk', '_topk', (['bl_heat'], {'K': 'K'}), '(bl_heat, K=K)\n', (12489, 12503), False, 'from models.decode import _topk_original, _topk, _topk_channel, _nms\n'), ((12559, 12578), 'models.decode._topk', '_topk', (['br_heat'], {'K': 'K'}), '(br_heat, K=K)\n', (12564, 12578), False, 'from models.decode import _topk_original, _topk, _topk_channel, _nms\n'), ((12634, 12653), 'models.decode._topk', '_topk', (['ct_heat'], {'K': 'K'}), '(ct_heat, K=K)\n', (12639, 12653), False, 'from models.decode import _topk_original, _topk, _topk_channel, _nms\n'), ((14196, 14258), 'torch.stack', 'torch.stack', (['(tl_xs, tl_ys, bl_xs, bl_ys, br_xs, br_ys)'], {'dim': '(4)'}), '((tl_xs, tl_ys, bl_xs, bl_ys, br_xs, br_ys), dim=4)\n', (14207, 14258), False, 'import torch\n'), ((14279, 14321), 'models.utils._tranpose_and_gather_feat', '_tranpose_and_gather_feat', (['tl_tag', 'tl_inds'], {}), '(tl_tag, tl_inds)\n', (14304, 14321), False, 'from models.utils import _tranpose_and_gather_feat, _gather_feat\n'), ((14388, 14430), 'models.utils._tranpose_and_gather_feat', '_tranpose_and_gather_feat', (['bl_tag', 'bl_inds'], {}), '(bl_tag, bl_inds)\n', (14413, 14430), False, 'from models.utils import _tranpose_and_gather_feat, _gather_feat\n'), ((14497, 14539), 'models.utils._tranpose_and_gather_feat', '_tranpose_and_gather_feat', (['br_tag', 'br_inds'], {}), '(br_tag, br_inds)\n', (14522, 14539), False, 'from models.utils import _tranpose_and_gather_feat, _gather_feat\n'), ((15812, 15840), 'torch.topk', 'torch.topk', (['scores', 'num_dets'], {}), '(scores, num_dets)\n', (15822, 15840), False, 'import torch\n'), ((15945, 15971), 'models.utils._gather_feat', '_gather_feat', (['bboxes', 'inds'], {}), '(bboxes, inds)\n', (15957, 15971), False, 'from models.utils import _tranpose_and_gather_feat, _gather_feat\n'), ((16681, 16755), 'torch.cat', 'torch.cat', (['[bboxes, scores, tl_scores, bl_scores, br_scores, clses]'], {'dim': '(2)'}), '([bboxes, scores, tl_scores, bl_scores, br_scores, clses], dim=2)\n', (16690, 16755), False, 'import torch\n'), ((23561, 23595), 'numpy.concatenate', 'np.concatenate', (['detections'], {'axis': '(1)'}), '(detections, axis=1)\n', (23575, 23595), True, 'import numpy as np\n'), ((23614, 23645), 'numpy.concatenate', 'np.concatenate', (['centers'], {'axis': '(1)'}), '(centers, axis=1)\n', (23628, 23645), True, 'import numpy as np\n'), ((23987, 24028), 'numpy.arctan', 'np.arctan', (['(-(br_y - bl_y) / (br_x - bl_x))'], {}), '(-(br_y - bl_y) / (br_x - bl_x))\n', (23996, 24028), True, 'import numpy as np\n'), ((24048, 24077), 'numpy.fabs', 'np.fabs', (['(quant_ori - cont_ori)'], {}), '(quant_ori - cont_ori)\n', (24055, 24077), True, 'import numpy as np\n'), ((25298, 25327), 'copy.copy', 'copy.copy', (['s_detections[:, 6]'], {}), '(s_detections[:, 6])\n', (25307, 25327), False, 'import copy\n'), ((26228, 26395), 'numpy.argmax', 'np.argmax', (['(ind_x1 + 0 & ind_y1 + 0 & ind_cls + 0 | ind_x1 + 0 & ind_y2 + 0 & ind_cls +\n 0 | ind_x2 + 0 & ind_y2 + 0 & ind_cls + 0)[:, ind_s_new_score]'], {'axis': '(0)'}), '((ind_x1 + 0 & ind_y1 + 0 & ind_cls + 0 | ind_x1 + 0 & ind_y2 + 0 &\n ind_cls + 0 | ind_x2 + 0 & ind_y2 + 0 & ind_cls + 0)[:, ind_s_new_score\n ], axis=0)\n', (26237, 26395), True, 'import numpy as np\n'), ((27025, 27054), 'copy.copy', 'copy.copy', (['l_detections[:, 6]'], {}), '(l_detections[:, 6])\n', (27034, 27054), False, 'import copy\n'), ((27955, 28122), 'numpy.argmax', 'np.argmax', (['(ind_x1 + 0 & ind_y1 + 0 & ind_cls + 0 | ind_x1 + 0 & ind_y2 + 0 & ind_cls +\n 0 | ind_x2 + 0 & ind_y2 + 0 & ind_cls + 0)[:, ind_l_new_score]'], {'axis': '(0)'}), '((ind_x1 + 0 & ind_y1 + 0 & ind_cls + 0 | ind_x1 + 0 & ind_y2 + 0 &\n ind_cls + 0 | ind_x2 + 0 & ind_y2 + 0 & ind_cls + 0)[:, ind_l_new_score\n ], axis=0)\n', (27964, 28122), True, 'import numpy as np\n'), ((28436, 28488), 'numpy.concatenate', 'np.concatenate', (['[l_detections, s_detections]'], {'axis': '(0)'}), '([l_detections, s_detections], axis=0)\n', (28450, 28488), True, 'import numpy as np\n'), ((28786, 28820), 'numpy.expand_dims', 'np.expand_dims', (['detections'], {'axis': '(0)'}), '(detections, axis=0)\n', (28800, 28820), True, 'import numpy as np\n'), ((2861, 2881), 'torch.device', 'torch.device', (['"""cuda"""'], {}), "('cuda')\n", (2873, 2881), False, 'import torch\n'), ((2954, 2974), 'torch.device', 'torch.device', (['"""cuda"""'], {}), "('cuda')\n", (2966, 2974), False, 'import torch\n'), ((3047, 3067), 'torch.device', 'torch.device', (['"""cuda"""'], {}), "('cuda')\n", (3059, 3067), False, 'import torch\n'), ((3919, 3939), 'torch.device', 'torch.device', (['"""cuda"""'], {}), "('cuda')\n", (3931, 3939), False, 'import torch\n'), ((4012, 4032), 'torch.device', 'torch.device', (['"""cuda"""'], {}), "('cuda')\n", (4024, 4032), False, 'import torch\n'), ((4105, 4125), 'torch.device', 'torch.device', (['"""cuda"""'], {}), "('cuda')\n", (4117, 4125), False, 'import torch\n'), ((4198, 4218), 'torch.device', 'torch.device', (['"""cuda"""'], {}), "('cuda')\n", (4210, 4218), False, 'import torch\n'), ((9768, 9821), 'math.atan', 'math.atan', (['(-(br_y_gt - bl_y_gt) / (br_x_gt - bl_x_gt))'], {}), '(-(br_y_gt - bl_y_gt) / (br_x_gt - bl_x_gt))\n', (9777, 9821), False, 'import math\n'), ((10623, 10683), 'numpy.ascontiguousarray', 'np.ascontiguousarray', (['bbox_coord_pr[:, :4]'], {'dtype': 'np.float32'}), '(bbox_coord_pr[:, :4], dtype=np.float32)\n', (10643, 10683), True, 'import numpy as np\n'), ((10721, 10782), 'numpy.ascontiguousarray', 'np.ascontiguousarray', (['boxes_coord_gt[:, :4]'], {'dtype': 'np.float32'}), '(boxes_coord_gt[:, :4], dtype=np.float32)\n', (10741, 10782), True, 'import numpy as np\n'), ((13290, 13332), 'models.utils._tranpose_and_gather_feat', '_tranpose_and_gather_feat', (['tl_reg', 'tl_inds'], {}), '(tl_reg, tl_inds)\n', (13315, 13332), False, 'from models.utils import _tranpose_and_gather_feat, _gather_feat\n'), ((13410, 13452), 'models.utils._tranpose_and_gather_feat', '_tranpose_and_gather_feat', (['bl_reg', 'bl_inds'], {}), '(bl_reg, bl_inds)\n', (13435, 13452), False, 'from models.utils import _tranpose_and_gather_feat, _gather_feat\n'), ((13530, 13572), 'models.utils._tranpose_and_gather_feat', '_tranpose_and_gather_feat', (['br_reg', 'br_inds'], {}), '(br_reg, br_inds)\n', (13555, 13572), False, 'from models.utils import _tranpose_and_gather_feat, _gather_feat\n'), ((13650, 13692), 'models.utils._tranpose_and_gather_feat', '_tranpose_and_gather_feat', (['ct_reg', 'ct_inds'], {}), '(ct_reg, ct_inds)\n', (13675, 13692), False, 'from models.utils import _tranpose_and_gather_feat, _gather_feat\n'), ((25967, 26105), 'numpy.max', 'np.max', (['(ind_x1 + 0 & ind_y1 + 0 & ind_cls + 0 | ind_x1 + 0 & ind_y2 + 0 & ind_cls +\n 0 | ind_x2 + 0 & ind_y2 + 0 & ind_cls + 0)'], {'axis': '(0)'}), '(ind_x1 + 0 & ind_y1 + 0 & ind_cls + 0 | ind_x1 + 0 & ind_y2 + 0 & \n ind_cls + 0 | ind_x2 + 0 & ind_y2 + 0 & ind_cls + 0, axis=0)\n', (25973, 26105), True, 'import numpy as np\n'), ((27694, 27832), 'numpy.max', 'np.max', (['(ind_x1 + 0 & ind_y1 + 0 & ind_cls + 0 | ind_x1 + 0 & ind_y2 + 0 & ind_cls +\n 0 | ind_x2 + 0 & ind_y2 + 0 & ind_cls + 0)'], {'axis': '(0)'}), '(ind_x1 + 0 & ind_y1 + 0 & ind_cls + 0 | ind_x1 + 0 & ind_y2 + 0 & \n ind_cls + 0 | ind_x2 + 0 & ind_y2 + 0 & ind_cls + 0, axis=0)\n', (27700, 27832), True, 'import numpy as np\n'), ((28521, 28550), 'numpy.argsort', 'np.argsort', (['(-detections[:, 6])'], {}), '(-detections[:, 6])\n', (28531, 28550), True, 'import numpy as np\n'), ((29678, 29703), 'numpy.partition', 'np.partition', (['scores', 'kth'], {}), '(scores, kth)\n', (29690, 29703), True, 'import numpy as np\n'), ((3371, 3390), 'torch.device', 'torch.device', (['"""cpu"""'], {}), "('cpu')\n", (3383, 3390), False, 'import torch\n'), ((3448, 3467), 'torch.device', 'torch.device', (['"""cpu"""'], {}), "('cpu')\n", (3460, 3467), False, 'import torch\n'), ((3525, 3544), 'torch.device', 'torch.device', (['"""cpu"""'], {}), "('cpu')\n", (3537, 3544), False, 'import torch\n'), ((5869, 5888), 'torch.device', 'torch.device', (['"""cpu"""'], {}), "('cpu')\n", (5881, 5888), False, 'import torch\n'), ((5954, 5973), 'torch.device', 'torch.device', (['"""cpu"""'], {}), "('cpu')\n", (5966, 5973), False, 'import torch\n'), ((6039, 6058), 'torch.device', 'torch.device', (['"""cpu"""'], {}), "('cpu')\n", (6051, 6058), False, 'import torch\n'), ((6124, 6143), 'torch.device', 'torch.device', (['"""cpu"""'], {}), "('cpu')\n", (6136, 6143), False, 'import torch\n'), ((6209, 6228), 'torch.device', 'torch.device', (['"""cpu"""'], {}), "('cpu')\n", (6221, 6228), False, 'import torch\n'), ((6294, 6313), 'torch.device', 'torch.device', (['"""cpu"""'], {}), "('cpu')\n", (6306, 6313), False, 'import torch\n'), ((6379, 6398), 'torch.device', 'torch.device', (['"""cpu"""'], {}), "('cpu')\n", (6391, 6398), False, 'import torch\n'), ((6464, 6483), 'torch.device', 'torch.device', (['"""cpu"""'], {}), "('cpu')\n", (6476, 6483), False, 'import torch\n'), ((6586, 6605), 'torch.device', 'torch.device', (['"""cpu"""'], {}), "('cpu')\n", (6598, 6605), False, 'import torch\n'), ((6707, 6726), 'torch.device', 'torch.device', (['"""cpu"""'], {}), "('cpu')\n", (6719, 6726), False, 'import torch\n'), ((6828, 6847), 'torch.device', 'torch.device', (['"""cpu"""'], {}), "('cpu')\n", (6840, 6847), False, 'import torch\n'), ((6949, 6968), 'torch.device', 'torch.device', (['"""cpu"""'], {}), "('cpu')\n", (6961, 6968), False, 'import torch\n'), ((7070, 7089), 'torch.device', 'torch.device', (['"""cpu"""'], {}), "('cpu')\n", (7082, 7089), False, 'import torch\n'), ((7191, 7210), 'torch.device', 'torch.device', (['"""cpu"""'], {}), "('cpu')\n", (7203, 7210), False, 'import torch\n'), ((7312, 7331), 'torch.device', 'torch.device', (['"""cpu"""'], {}), "('cpu')\n", (7324, 7331), False, 'import torch\n'), ((7433, 7452), 'torch.device', 'torch.device', (['"""cpu"""'], {}), "('cpu')\n", (7445, 7452), False, 'import torch\n'), ((8373, 8424), 'math.atan2', 'math.atan2', (['(-(br_y_pr - bl_y_pr))', '(br_x_pr - bl_x_pr)'], {}), '(-(br_y_pr - bl_y_pr), br_x_pr - bl_x_pr)\n', (8383, 8424), False, 'import math\n'), ((11137, 11192), 'math.fabs', 'math.fabs', (['(bbox_coord_pr[i, -1] - boxes_coord_gt[j, -1])'], {}), '(bbox_coord_pr[i, -1] - boxes_coord_gt[j, -1])\n', (11146, 11192), False, 'import math\n'), ((14717, 14744), 'torch.abs', 'torch.abs', (['(br_tag - avg_tag)'], {}), '(br_tag - avg_tag)\n', (14726, 14744), False, 'import torch\n'), ((16050, 16075), 'models.utils._gather_feat', '_gather_feat', (['clses', 'inds'], {}), '(clses, inds)\n', (16062, 16075), False, 'from models.utils import _tranpose_and_gather_feat, _gather_feat\n'), ((16170, 16199), 'models.utils._gather_feat', '_gather_feat', (['tl_scores', 'inds'], {}), '(tl_scores, inds)\n', (16182, 16199), False, 'from models.utils import _tranpose_and_gather_feat, _gather_feat\n'), ((16294, 16323), 'models.utils._gather_feat', '_gather_feat', (['bl_scores', 'inds'], {}), '(bl_scores, inds)\n', (16306, 16323), False, 'from models.utils import _tranpose_and_gather_feat, _gather_feat\n'), ((16418, 16447), 'models.utils._gather_feat', '_gather_feat', (['br_scores', 'inds'], {}), '(br_scores, inds)\n', (16430, 16447), False, 'from models.utils import _tranpose_and_gather_feat, _gather_feat\n'), ((24397, 24457), 'numpy.power', 'np.power', (['(valid_detections[:, 2] - valid_detections[:, 4])', '(2)'], {}), '(valid_detections[:, 2] - valid_detections[:, 4], 2)\n', (24405, 24457), True, 'import numpy as np\n'), ((24490, 24550), 'numpy.power', 'np.power', (['(valid_detections[:, 3] - valid_detections[:, 5])', '(2)'], {}), '(valid_detections[:, 3] - valid_detections[:, 5], 2)\n', (24498, 24550), True, 'import numpy as np\n'), ((24581, 24641), 'numpy.power', 'np.power', (['(valid_detections[:, 2] - valid_detections[:, 0])', '(2)'], {}), '(valid_detections[:, 2] - valid_detections[:, 0], 2)\n', (24589, 24641), True, 'import numpy as np\n'), ((24675, 24735), 'numpy.power', 'np.power', (['(valid_detections[:, 3] - valid_detections[:, 1])', '(2)'], {}), '(valid_detections[:, 3] - valid_detections[:, 1], 2)\n', (24683, 24735), True, 'import numpy as np\n'), ((29064, 29101), 'numpy.array', 'np.array', (['ret[0][j]'], {'dtype': 'np.float32'}), '(ret[0][j], dtype=np.float32)\n', (29072, 29101), True, 'import numpy as np\n'), ((29266, 29332), 'numpy.concatenate', 'np.concatenate', (['[detection[j] for detection in detections]'], {'axis': '(0)'}), '([detection[j] for detection in detections], axis=0)\n', (29280, 29332), True, 'import numpy as np\n'), ((3340, 3367), 'torch.abs', 'torch.abs', (['(avg_tag - tl_tag)'], {}), '(avg_tag - tl_tag)\n', (3349, 3367), False, 'import torch\n'), ((3417, 3444), 'torch.abs', 'torch.abs', (['(bl_tag - avg_tag)'], {}), '(bl_tag - avg_tag)\n', (3426, 3444), False, 'import torch\n'), ((3494, 3521), 'torch.abs', 'torch.abs', (['(avg_tag - br_tag)'], {}), '(avg_tag - br_tag)\n', (3503, 3521), False, 'import torch\n'), ((14657, 14684), 'torch.abs', 'torch.abs', (['(tl_tag - avg_tag)'], {}), '(tl_tag - avg_tag)\n', (14666, 14684), False, 'import torch\n'), ((14687, 14714), 'torch.abs', 'torch.abs', (['(bl_tag - avg_tag)'], {}), '(bl_tag - avg_tag)\n', (14696, 14714), False, 'import torch\n'), ((6549, 6569), 'torch.device', 'torch.device', (['"""cuda"""'], {}), "('cuda')\n", (6561, 6569), False, 'import torch\n'), ((6670, 6690), 'torch.device', 'torch.device', (['"""cuda"""'], {}), "('cuda')\n", (6682, 6690), False, 'import torch\n'), ((6791, 6811), 'torch.device', 'torch.device', (['"""cuda"""'], {}), "('cuda')\n", (6803, 6811), False, 'import torch\n'), ((6912, 6932), 'torch.device', 'torch.device', (['"""cuda"""'], {}), "('cuda')\n", (6924, 6932), False, 'import torch\n'), ((7033, 7053), 'torch.device', 'torch.device', (['"""cuda"""'], {}), "('cuda')\n", (7045, 7053), False, 'import torch\n'), ((7154, 7174), 'torch.device', 'torch.device', (['"""cuda"""'], {}), "('cuda')\n", (7166, 7174), False, 'import torch\n'), ((7275, 7295), 'torch.device', 'torch.device', (['"""cuda"""'], {}), "('cuda')\n", (7287, 7295), False, 'import torch\n'), ((7396, 7416), 'torch.device', 'torch.device', (['"""cuda"""'], {}), "('cuda')\n", (7408, 7416), False, 'import torch\n')]
import logbook import pandas as pd import zipline as zl from datetime import datetime, timedelta import pathlib import azul import numpy as np from typing import List, Tuple log = logbook.Logger('BasePriceManager') class BasePriceManager(object): def __init__(self, calendar_name='NYSE'): self._calendar = zl.get_calendar(name=calendar_name) self._cols = ['open', 'high', 'low', 'close', 'volume', 'dividend', 'split'] # The number of days the price manager will keep trying to pull data for a symbol that is not returning data. self.MISSING_DATE_THRESHOLD = 5 def get_price_data(self, symbols: List[str], output_dir: str, start_date: datetime, end_date: datetime) -> None: minute_dir_path = pathlib.Path(output_dir, 'minute') daily_dir_path = pathlib.Path(output_dir, 'daily') for ticker in symbols: self._download_and_process_data( ticker, start_date, end_date, minute_dir_path, daily_dir_path) def _download_and_process_data( self, ticker: str, start_date: datetime, end_date: datetime, minute_dir_path: pathlib.Path, daily_dir_path: pathlib.Path ) -> None: df = self._minute_dataframe_for_dates(ticker, start_date, end_date) if df.empty: return df = self._check_sessions(df, ticker, frequency='minute') minute_dir_path.mkdir(parents=True, exist_ok=True) filename = pathlib.Path(minute_dir_path, ticker + '.csv') df.to_csv(filename) daily_df = self._resample_minute_data_to_daily_data(df) daily_df = self._check_sessions(daily_df, ticker, frequency='daily') daily_dir_path.mkdir(parents=True, exist_ok=True) filename = pathlib.Path(daily_dir_path, ticker + '.csv') daily_df.to_csv(filename) log.notice('Retrieved: {}'.format(ticker)) def _resample_minute_data_to_daily_data(self, df): ohlc_dict = { 'open': 'first', 'high': 'max', 'low': 'min', 'close': 'last', 'volume': 'sum', 'dividend': 'last', 'split': 'last' } if df.empty: daily_df = df else: daily_df = df.resample('D', closed='left', label='left').agg(ohlc_dict).dropna(how='any') # Resample mixes the columns so lets re-arrange them daily_df = daily_df[self._cols] return daily_df def _validated_start_and_end_dates( self, start_date: datetime, end_date: datetime ) -> Tuple[datetime, datetime]: """ Creates valid start and end dates. Defaults to a start date of 30 calendar days ago and end date of today. Args: start_date (datetime): The start date. end_date (datetime): The end date. Returns: start_date (datetime): The validated start date. end_date (datetime): The validated end date. """ today = datetime.today() if start_date is None: start_date = today - timedelta(days=30) if end_date is None: end_date = today if start_date > end_date: temp_date = start_date start_date = end_date end_date = temp_date return start_date, end_date def _minute_dataframe_for_dates( self, ticker: str, start_date: datetime, end_date: datetime ) -> pd.DataFrame: """ Returns a DataFrame containing the all the minute bars for stock between the start and end dates. Args: ticker (str): Ticker symbol for the stock. start_date (datetime): Date to start pulling data. end_date (datetime): Date to stop pulling data. Returns: combined_df (DataFrame): Contains the all the minute bars for a stock between the start and end dates. """ start_date, end_date = self._validated_start_and_end_dates(start_date, end_date) combined_df = pd.DataFrame(columns=self._cols) combined_df.index.name = 'date' # # # Get the date the symbol was listed on the exchange. # list_date = self._list_date(ticker) # # if list_date is not None: # # If the we are asking for data from before the stock was listed, then set the start date to the day # # the stock was listed. # if list_date > start_date: # log.info('The symbol {} was not listed until: {}. Adjusting start time.', ticker, list_date) # start_date = list_date # Build a list of the trading days from the dates passed in. session_dates = self._calendar.sessions_in_range(start_date, end_date) if session_dates.empty: log.info('The symbol {} did not trade between {} and {} ', ticker, start_date, end_date) return combined_df # Iterate over the trading dates backwards. This means we don't need to know exactly # when the stock started trading. Note: this won't pull data for stocks that have been delisted. # TODO: Add code to capture data for delisted stocks. num_missing_dates = 0 for timestamp in reversed(session_dates): df = self._minute_dataframe_for_date(ticker, timestamp) if df.empty: # Start counting the number of consecutive trading dates we are missing data. num_missing_dates += 1 log.info('No minute data for {} on {}'.format(ticker, timestamp.date())) else: # reset missing date counter num_missing_dates = 0 log.info('Retrieved minute data for {} on {}'.format(ticker, timestamp.date())) combined_df = pd.concat([combined_df, df]) if num_missing_dates >= self.MISSING_DATE_THRESHOLD: log.info('No minute data for {} for {} days. Quitting.'.format(ticker, self.MISSING_DATE_THRESHOLD)) break # Sort the dataframe oldest first, newest last. combined_df.sort_index(inplace=True) return combined_df # def _list_date(self, ticker: str) -> datetime: # return None def _minute_dataframe_for_date(self, ticker: str, start_timestamp: pd.Timestamp) -> pd.DataFrame: raise NotImplementedError def _fixna(self, df, symbol): cols = ['close', 'high', 'low', 'open'] df[cols] = df[cols].replace({0: np.nan}) df[cols] = df[cols].replace({-1.0: np.nan}) if df.isnull().sum().sum() > 0: # fixna_list.append(symbol) df['open'] = df['open'].bfill().ffill() df['close'] = df['close'].bfill().ffill() df.loc[df['low'].isnull(), 'low'] = df['open'] df.loc[df['high'].isnull(), 'high'] = df['open'] df.loc[df['close'].isnull(), 'close'] = df['open'] return df def _check_sessions(self, df, ticker, frequency='daily'): # Remove any data that are outside of the trading sessions for the calendar. if df.empty: return df asset_first_day = df.index[0] asset_last_day = df.index[-1] sessions = self._calendar.sessions_in_range(asset_first_day, asset_last_day) asset_sessions = sessions[sessions.slice_indexer(asset_first_day, asset_last_day)] if frequency == 'minute': minutes_passed = len(df) asset_first_day = self._calendar.minute_to_session_label(asset_first_day, direction='next') asset_last_day = self._calendar.minute_to_session_label(asset_last_day, direction='previous') minutes_in_session = self._calendar.minutes_for_sessions_in_range(asset_first_day, asset_last_day) df = df[df.index.isin(minutes_in_session)] if (minutes_passed) > len(minutes_in_session): # print('Removed ' + str((minutes_passed) - len(minutes_in_session)) + ' minutes') pass elif minutes_passed < len(minutes_in_session): num_missing_sessions = len(minutes_in_session) - minutes_passed log.info('Missing sessions for {}'.format(ticker)) elif frequency == 'daily' and len(df) != len(asset_sessions): missing_sessions = asset_sessions.difference( pd.to_datetime(np.array(df.index), unit='s', utc=True, )).tolist() extra_sessions = pd.to_datetime(np.array(df.index), unit='s', utc=True, ).difference( asset_sessions).tolist() if len(missing_sessions) > 0: # missing_sessions_list.append(symbol) # print('Adding ' + str(len(missing_sessions)) + ' sessions for ' + str(ticker)) pass if len(extra_sessions) > 0: # extra_sessions_list.append(symbol) # print('Removing ' + str(len(extra_sessions)) + ' sessions for ' + str(symbol)) pass for missing_session in missing_sessions: prev_date = self._calendar.previous_session_label(missing_session) row_to_copy = df[(df.index == prev_date)] row_to_copy_val = row_to_copy.values # from IPython import embed; embed() df.loc[missing_session] = row_to_copy_val[0] df.loc[missing_session].volume = 0 # row = row_to_copy # table.append(row) for extra_session in extra_sessions: # delete stuff df.drop(extra_session) if frequency == 'minute': log.info('Downloaded and processed {} minute bars for {}', len(df), ticker) else: log.info('Downsampled {} daily bars for {}', len(df), ticker) return df
[ "logbook.Logger", "pandas.DataFrame", "pathlib.Path", "numpy.array", "zipline.get_calendar", "datetime.datetime.today", "datetime.timedelta", "pandas.concat" ]
[((181, 215), 'logbook.Logger', 'logbook.Logger', (['"""BasePriceManager"""'], {}), "('BasePriceManager')\n", (195, 215), False, 'import logbook\n'), ((322, 357), 'zipline.get_calendar', 'zl.get_calendar', ([], {'name': 'calendar_name'}), '(name=calendar_name)\n', (337, 357), True, 'import zipline as zl\n'), ((746, 780), 'pathlib.Path', 'pathlib.Path', (['output_dir', '"""minute"""'], {}), "(output_dir, 'minute')\n", (758, 780), False, 'import pathlib\n'), ((806, 839), 'pathlib.Path', 'pathlib.Path', (['output_dir', '"""daily"""'], {}), "(output_dir, 'daily')\n", (818, 839), False, 'import pathlib\n'), ((1503, 1549), 'pathlib.Path', 'pathlib.Path', (['minute_dir_path', "(ticker + '.csv')"], {}), "(minute_dir_path, ticker + '.csv')\n", (1515, 1549), False, 'import pathlib\n'), ((1797, 1842), 'pathlib.Path', 'pathlib.Path', (['daily_dir_path', "(ticker + '.csv')"], {}), "(daily_dir_path, ticker + '.csv')\n", (1809, 1842), False, 'import pathlib\n'), ((3071, 3087), 'datetime.datetime.today', 'datetime.today', ([], {}), '()\n', (3085, 3087), False, 'from datetime import datetime, timedelta\n'), ((4143, 4175), 'pandas.DataFrame', 'pd.DataFrame', ([], {'columns': 'self._cols'}), '(columns=self._cols)\n', (4155, 4175), True, 'import pandas as pd\n'), ((3152, 3170), 'datetime.timedelta', 'timedelta', ([], {'days': '(30)'}), '(days=30)\n', (3161, 3170), False, 'from datetime import datetime, timedelta\n'), ((5926, 5954), 'pandas.concat', 'pd.concat', (['[combined_df, df]'], {}), '([combined_df, df])\n', (5935, 5954), True, 'import pandas as pd\n'), ((8511, 8529), 'numpy.array', 'np.array', (['df.index'], {}), '(df.index)\n', (8519, 8529), True, 'import numpy as np\n'), ((8607, 8625), 'numpy.array', 'np.array', (['df.index'], {}), '(df.index)\n', (8615, 8625), True, 'import numpy as np\n')]
from unittest import TestCase from unittest import skip from cs3api4lab.api.cs3_share_api import Cs3ShareApi from cs3api4lab.api.cs3_file_api import Cs3FileApi from cs3api4lab.config.config_manager import Cs3ConfigManager from traitlets.config import LoggingConfigurable class TestCs3ShareApi(TestCase, LoggingConfigurable): api = None config = None share_id = None receiver_id = 'f7fbf8c8-139b-4376-b307-cf0a8c2d0d9c' receiver_idp = '<EMAIL>' receiver2_id = '932b4540-8d16-481e-8ef4-588e4b6b151c' receiver2_idp = 'example.<EMAIL>' receiver_role = 'viewer' receiver_grantee_type = 'user' file_path = '/test.txt' file_name = 'test.txt' storage_id = '123e4567-e89b-12d3-a456-426655440000' def setUp(self): self.config = Cs3ConfigManager.get_config() self.storage = Cs3FileApi(self.log) self.api = Cs3ShareApi(self.log) def test_create_and_list_directory_model(self): created_share = self._create_share() self.share_id = created_share['opaque_id'] share_list = self.api.list() try: if not list(filter(lambda s: s['name'] == self.file_name, share_list['content'])): raise Exception("Share not created") finally: self._clear_shares() def test_create_duplicate_and_list_directory_model(self): created_share = self._create_share() self.share_id = created_share['opaque_id'] self._create_test_share(self.receiver2_id, self.receiver2_idp) share_list = self.api.list() self.assertEqual(len(share_list['content']), 1) try: if not list(filter(lambda s: s['name'] == self.file_name, share_list['content'])): raise Exception("Share not created") finally: self._clear_shares() def test_list_grantees_for_file(self): self._create_share() response = self.api.list_grantees_for_file(self.file_path) try: if not response: raise Exception("Failed to retrieve grantees of the file") if response['shares'][0]['grantee']['opaque_id'] != self.receiver_id: raise Exception("Incorrect grantee") if response['shares'][0]['grantee']['permissions'] != self.receiver_role: raise Exception("Incorrect permissions") finally: self._clear_shares() def test_remove(self): created_share = self._create_share() self.share_id = created_share['opaque_id'] share_list = self.api.list() try: if not list(filter(lambda s: s['name'] == self.file_name, share_list['content'])): raise Exception("Share not created") finally: self.api.remove(self.share_id) share_list = self.api.list() if list(filter(lambda s: s['name'] == self.file_name, share_list['content'])): raise Exception("Share not removed") def test_update(self): created_share = self._create_share() self.share_id = created_share['opaque_id'] self.api.update(self.share_id, 'editor') share_list = self.api.list_grantees_for_file(self.file_path) try: if not list(filter( lambda s: s['grantee']['opaque_id'] == self.receiver_id and s['grantee']['permissions'] == 'editor', share_list['shares'])): raise Exception("Share not updated") finally: self._clear_shares() @skip def test_list_received(self): self.api.list_received() @skip def test_update_received(self): self.api.update_received("1", "accepted") def _create_share(self): self._create_test_file() return self._create_test_share() def _clear_shares(self): shares = self.api.list_grantees_for_file(self.file_path) for share in shares['shares']: self._remove_test_share(share['opaque_id']) try: self._remove_test_file() except IOError as e: print("Error remove file:", e) def _create_test_share(self, receiver_id='f7fbf8c8-139b-4376-b307-cf0a8c2d0d9c', receiver_idp='<EMAIL>'): file_path = self.config['home_dir'] + self.file_path return self.api.create(self.config['endpoint'], file_path, receiver_id, receiver_idp, self.receiver_role, self.receiver_grantee_type) def _remove_test_share(self, share_id): self.api.remove(share_id) def _create_test_file(self): self.storage.write_file(self.file_path, "Lorem ipsum dolor sit amet...", self.config['endpoint']) def _remove_test_file(self): self.storage.remove(self.file_path, self.config['endpoint'])
[ "cs3api4lab.config.config_manager.Cs3ConfigManager.get_config", "cs3api4lab.api.cs3_share_api.Cs3ShareApi", "cs3api4lab.api.cs3_file_api.Cs3FileApi" ]
[((784, 813), 'cs3api4lab.config.config_manager.Cs3ConfigManager.get_config', 'Cs3ConfigManager.get_config', ([], {}), '()\n', (811, 813), False, 'from cs3api4lab.config.config_manager import Cs3ConfigManager\n'), ((837, 857), 'cs3api4lab.api.cs3_file_api.Cs3FileApi', 'Cs3FileApi', (['self.log'], {}), '(self.log)\n', (847, 857), False, 'from cs3api4lab.api.cs3_file_api import Cs3FileApi\n'), ((877, 898), 'cs3api4lab.api.cs3_share_api.Cs3ShareApi', 'Cs3ShareApi', (['self.log'], {}), '(self.log)\n', (888, 898), False, 'from cs3api4lab.api.cs3_share_api import Cs3ShareApi\n')]
# -*- coding: utf-8 -*- """ Created on Mon Jul 1 23:11:03 2019 @author: Relieak """ ''' DCARD 熱門前30文章搜尋 純用 BeautifulSoup 寫法 ''' from bs4 import BeautifulSoup import requests import re url = "https://www.dcard.tw/f/" html = requests.get(url) html.encoding = "utf-8" #print(html.text) sp = BeautifulSoup(html.text,"html.parser") #print(sp.prettify()) #排版後更容易分析 #data = sp.select(".PostEntry_root_V6g0rd") data = sp.select(".PostList_entry_1rq5Lf") #若要搜尋標簽中的內容 必須先搜尋上一個標籤 否則會找不到 ex : href 先<div 不能先<a #print(data[0]) #for link in data[1].find_all("a",{"class" : "PostEntry_root_V6g0rd"}) : # # #print(link) # # http = link.get("href") # # print("https://www.dcard.tw%s" %(http)) while True : try : number = input("~~~歡迎來到Dcard 前30熱門文章搜尋~~~\n""\n你想看第幾則文章(離開 請按Enter)>> ") if number == "" : break elif 0 < int(number) <= 30 : try : for i in range(int(number)-1,int(number)) : data1 = data[i].find_all("h3",{"class" : "Title__Text-v196i6-0 gmfDU"}) data2 = data[i].find_all("div",{"class" : "PostEntry_excerpt_2eHlNn"}) data3 = data[i].find_all("div",{"class" : "ActionBar__LikeCount-pwz3il-1 cGEHtj"}) data4 = data[i].find_all("span",{"class" : "ActionBar__CommentWrapper-pwz3il-5 hkpJwJ"}) data5 = data[i].find_all("span",{"class" : "Header__PublishedDateTime-xvcbwe-3 MDszy"}) print("") print("第%d則" %(int(number))) print("標題 :",data1[0].text) print("簡介 :",data2[0].text) print("表達心情數 :",data3[0].text) # .strip() 方法用於移除字符串頭尾指定的字符(默認為空格或換行符)或字符序列 -->文字<-- print("回應數 :",data4[0].text.strip("回應")) print("發表時間 :",data5[0].text) # 搜尋標簽中的內容的方法 #href for link in data[i].find_all("a",{"class" : "PostEntry_root_V6g0rd"}) : http = link.get("href") #print(http) #正規表示法 A = re.compile("[0-9]+") #表示任意數字串 B = re.compile("[A-Za-z0-9\./_]+") #表達任意數字,任意英文字母和底線字元的組合,也可寫成 \w # search(string)的用法是傳回第一組符合正規表示法的字串 Search1 = A.search(http) Search2 = B.search(http) #去掉網址後面的中文字 #print(Search1) #print(Search2) print("文章ID :",Search1.group()) #傳回儲存在match物件中的值 group() print("網頁 :https://www.dcard.tw%s" %(Search2.group())) print("") except : """ 若文章簡介開頭是 "前情提要" 會搜尋不到內文 "PostEntry_excerpt_2eHlNn" 產生錯誤 須去搜尋 "PostEntry_reply_1oU-6z" """ data2 = data[i].find_all("div",{"class" : "PostEntry_reply_1oU-6z"}) data3 = data[i].find_all("div",{"class" : "ActionBar__LikeCount-pwz3il-1 cGEHtj"}) data4 = data[i].find_all("span",{"class" : "ActionBar__CommentWrapper-pwz3il-5 hkpJwJ"}) data5 = data[i].find_all("span",{"class" : "Header__PublishedDateTime-xvcbwe-3 MDszy"}) print("簡介 :",data2[0].text) print("表達心情數 :",data3[0].text) print("回應數 :",data4[0].text.strip("回應")) print("發表時間 :",data5[0].text) for link in data[i].find_all("a",{"class" : "PostEntry_root_V6g0rd"}) : http = link.get("href") A = re.compile("[0-9]+") B = re.compile("[A-Za-z0-9\./_]+") Search1 = A.search(http) Search2 = B.search(http) print("文章ID :",Search1.group()) print("網頁 :https://www.dcard.tw%s" %(Search2.group())) print("") else : print("") print("只能搜尋前30篇文章喔!!") except : print("") print("輸入錯誤請重新輸入!!!")
[ "bs4.BeautifulSoup", "requests.get", "re.compile" ]
[((230, 247), 'requests.get', 'requests.get', (['url'], {}), '(url)\n', (242, 247), False, 'import requests\n'), ((297, 336), 'bs4.BeautifulSoup', 'BeautifulSoup', (['html.text', '"""html.parser"""'], {}), "(html.text, 'html.parser')\n", (310, 336), False, 'from bs4 import BeautifulSoup\n'), ((2656, 2676), 're.compile', 're.compile', (['"""[0-9]+"""'], {}), "('[0-9]+')\n", (2666, 2676), False, 'import re\n'), ((2746, 2777), 're.compile', 're.compile', (['"""[A-Za-z0-9\\\\./_]+"""'], {}), "('[A-Za-z0-9\\\\./_]+')\n", (2756, 2777), False, 'import re\n'), ((4575, 4595), 're.compile', 're.compile', (['"""[0-9]+"""'], {}), "('[0-9]+')\n", (4585, 4595), False, 'import re\n'), ((4620, 4651), 're.compile', 're.compile', (['"""[A-Za-z0-9\\\\./_]+"""'], {}), "('[A-Za-z0-9\\\\./_]+')\n", (4630, 4651), False, 'import re\n')]
import random import sys print("") print("4d6, drop lowest:") print("") def roll4d6(): global threshold target_met = False total = [] for attr in range(0,6): result = [] raw = [] for roll in range(0,4): dice = random.randint(1,6) result.append(dice) raw.append(dice) result.remove(min(result)) print(raw, result, sum(result)) total.append(sum(result)) if sum(result) >= int(threshold): target_met = True print("------------------") print("Total: " + str(sum(total))) return sum(total), target_met try: global target_met, threshold target = 0 count = 1 if "-" in sys.argv[1]: # count tag, threshold = sys.argv[1].split("-") for i in range(int(threshold)): print("") print("*******************") target, target_met = roll4d6() print("4d6 Total: " + str(target) + " Total " + str(threshold) + " Sets: " + str(threshold) + " Rolls: " + str(count)) count = count + 1 print("") elif "+" in sys.argv[1]: # at least one stat matches this number tag, threshold = sys.argv[1].split("+") target_met = False while target_met == False: print("") print("*******************") target, target_met = roll4d6() print("4d6 Total: " + str(target) + " Threshold (At Least): " + str(threshold) + " Rolls: " + str(count)) count = count + 1 print("") else: threshold = sys.argv[1] while int(target) <= int(threshold): print("") print("*******************") target, result = roll4d6() print("4d6 Total: " + str(target) + " Threshold: " + str(threshold) + " Rolls: " + str(count)) count = count + 1 print("") except: threshold = 0 roll4d6() print("4d6.py +<num> for at least one stat in a set at num or above") print("4d6.py -<num> will roll up that many stat sets") print("4d6.py <num> will roll until the total of a set meets or exceeds num")
[ "random.randint" ]
[((265, 285), 'random.randint', 'random.randint', (['(1)', '(6)'], {}), '(1, 6)\n', (279, 285), False, 'import random\n')]
from __future__ import print_function import os import os.path from xml.etree import ElementTree def read_machine_file(): """ Reads the machine definition xml ROS launch file MCT_CONFIG/machine/mct.machine Returns a list containg a dictionary of attributes for each machine in the machine file. """ mct_config_dir = os.environ['MCT_CONFIG'] mct_machine_file = os.path.join(mct_config_dir, 'machine', 'mct.machine') tree = ElementTree.parse(mct_machine_file) machine_elem_list = tree.findall('machine') machine_list = [] for machine_elem in machine_elem_list: machine_list.append(machine_elem.attrib) return machine_list # ----------------------------------------------------------------------------- if __name__ == '__main__': machine_list = read_machine_file() print(machine_list)
[ "os.path.join", "xml.etree.ElementTree.parse" ]
[((393, 447), 'os.path.join', 'os.path.join', (['mct_config_dir', '"""machine"""', '"""mct.machine"""'], {}), "(mct_config_dir, 'machine', 'mct.machine')\n", (405, 447), False, 'import os\n'), ((459, 494), 'xml.etree.ElementTree.parse', 'ElementTree.parse', (['mct_machine_file'], {}), '(mct_machine_file)\n', (476, 494), False, 'from xml.etree import ElementTree\n')]
#!/usr/local/python/bin/python # script to check the previously unsolved files # # to do: # Sanity check all the image_ids in the table actually have a png # quick check shows 1349 in DB and 1353 pngs, 4 out, not bad # from create_movie import create_movie import os,sys,getpass,time import glob as g from astropy.io import fits from collections import defaultdict import pyds9 import argparse as ap me=getpass.getuser() if me=='ops': w_dir="/ngts/staging/archive/minisurvey/junk" astrom_loc="/usr/local/astrometry.net/bin/" elif me=='James': w_dir='/Users/James/Desktop/junk' astrom_loc="/usr/local/bin/" else: print("WHOAMI?") sys.exit(1) # check for w_dir if os.path.exists(w_dir)==False: print("I'm dying... (no w_dir)") sys.exit(1) # get command line args def argParse(): parser=ap.ArgumentParser(description="A script to redo the failed minisurvey publishing step") parser.add_argument('--astrometry', help = "try redoing the astrometry?", action='store_true') parser.add_argument('--manual', help = "manually analyse the images with DS9/DSS", action='store_true') parser.add_argument('--yes2all', help = "select this to skip prompting - used if pass already has been made through the imagaes and all are good", action='store_true') args=parser.parse_args() return args # read in astrometry.net log files def getAstromFromFile(astromfile): f = open(astromfile, "r").readlines() ra=None dec=None for i in f: if i.startswith("Field center: (RA H:M:S"): tmp = i.split('=') ra,dec = tmp[1].split(',') ra = ra.strip() ra = ra.replace('(','') dec = dec.strip() dec = dec.replace(').','') break return ra, dec # do the astrometry def astrometry(image,scale_l,scale_h,ra=None,dec=None,radius=5.0,cpulimit=90): astromfile="astrometry_%s.log" % (image) command = "%s/solve-field %s --scale-low %s --scale-high %s --cpulimit %s --no-plots --overwrite" % (astrom_loc,image, scale_l, scale_h, cpulimit) command = "%s > %s" % (command,astromfile) os.system(command) ra,dec=getAstromFromFile(astromfile) return ra,dec args=argParse() os.chdir(w_dir) t=sorted(g.glob('*.fits')) if args.astrometry: RA,DEC=[],[] for i in t: ra,dec=astrometry(i,2.83,2.93,cpulimit=2) if ra: RA.append(ra) DEC.append(dec) else: RA.append("0") DEC.append("0") if args.manual: fields=defaultdict(list) done=defaultdict(list) # loop over the and check for multiples # of the same field, if so work on the last one only for i in t: h=fits.open(i)[0].header['FIELD'] fields[h].append(i) d=pyds9.DS9() time.sleep(5) d.set('scale zscale') d.set('preserve scale') d.set('preserve pan') print("Remeber to DELETE duplicate images") rm_string="" for i in fields: image=fields[i][-1] d.set('frame clear all') h=fits.open(image)[0] ra=h.header['CMD_RA'] dec=h.header['CMD_DEC'] # print this so we can see which have duplicates to delete print(fields[i]) if len(fields[i])>1: for k in range(0,len(fields[i])-1): rm_string=rm_string+"%s " % (fields[i][k]) # display the image in DS9 and load the correct region of sky beside it d.set('tile yes') d.set('frame 1') d.set('file %s' % (image)) d.set('zoom 2') d.set('wcs align yes') d.set('cmap invert yes') d.set('frame 2') d.set('dsseso coord %.6f %.6f degrees size 30 30 arcmin' % (ra,dec)) d.set('zoom to fit') d.set('wcs align yes') d.set('cmap invert yes') d.set('frame center all') if args.yes2all: done[i].append(image) else: yn=input("Do the fields match? (y/n): ") if yn.lower().startswith('y'): done[i].append(image) else: continue print(rm_string) # need to make an astrometry* log file for the manually solved images? # and also a png too, then update the database as with the others, manually? table_update_string="" if len(done)> 0: print("Check the UPDATE strings as use them to UPDATE the minisurvey table") for i in done: create_movie(done[i],images_directory="%s/" % (w_dir),no_time_series=True,include_increment=False,clobber_images_directory=False,resize_factor=4,multiprocess=False) table_update_string=table_update_string+"UPDATE mini_survey SET checked_out=0,astrometry=1,done=1,png=1,fails=0 where image_id=\"%s\";\n" % (done[i][0][5:-5]) # image name minus IMAGE and .fits print(table_update_string)
[ "os.path.exists", "pyds9.DS9", "argparse.ArgumentParser", "time.sleep", "astropy.io.fits.open", "os.chdir", "create_movie.create_movie", "collections.defaultdict", "sys.exit", "getpass.getuser", "os.system", "glob.glob" ]
[((408, 425), 'getpass.getuser', 'getpass.getuser', ([], {}), '()\n', (423, 425), False, 'import os, sys, getpass, time\n'), ((2255, 2270), 'os.chdir', 'os.chdir', (['w_dir'], {}), '(w_dir)\n', (2263, 2270), False, 'import os, sys, getpass, time\n'), ((692, 713), 'os.path.exists', 'os.path.exists', (['w_dir'], {}), '(w_dir)\n', (706, 713), False, 'import os, sys, getpass, time\n'), ((763, 774), 'sys.exit', 'sys.exit', (['(1)'], {}), '(1)\n', (771, 774), False, 'import os, sys, getpass, time\n'), ((827, 919), 'argparse.ArgumentParser', 'ap.ArgumentParser', ([], {'description': '"""A script to redo the failed minisurvey publishing step"""'}), "(description=\n 'A script to redo the failed minisurvey publishing step')\n", (844, 919), True, 'import argparse as ap\n'), ((2132, 2150), 'os.system', 'os.system', (['command'], {}), '(command)\n', (2141, 2150), False, 'import os, sys, getpass, time\n'), ((2280, 2296), 'glob.glob', 'g.glob', (['"""*.fits"""'], {}), "('*.fits')\n", (2286, 2296), True, 'import glob as g\n'), ((2569, 2586), 'collections.defaultdict', 'defaultdict', (['list'], {}), '(list)\n', (2580, 2586), False, 'from collections import defaultdict\n'), ((2596, 2613), 'collections.defaultdict', 'defaultdict', (['list'], {}), '(list)\n', (2607, 2613), False, 'from collections import defaultdict\n'), ((2808, 2819), 'pyds9.DS9', 'pyds9.DS9', ([], {}), '()\n', (2817, 2819), False, 'import pyds9\n'), ((2824, 2837), 'time.sleep', 'time.sleep', (['(5)'], {}), '(5)\n', (2834, 2837), False, 'import os, sys, getpass, time\n'), ((658, 669), 'sys.exit', 'sys.exit', (['(1)'], {}), '(1)\n', (666, 669), False, 'import os, sys, getpass, time\n'), ((3076, 3092), 'astropy.io.fits.open', 'fits.open', (['image'], {}), '(image)\n', (3085, 3092), False, 'from astropy.io import fits\n'), ((4485, 4662), 'create_movie.create_movie', 'create_movie', (['done[i]'], {'images_directory': "('%s/' % w_dir)", 'no_time_series': '(True)', 'include_increment': '(False)', 'clobber_images_directory': '(False)', 'resize_factor': '(4)', 'multiprocess': '(False)'}), "(done[i], images_directory='%s/' % w_dir, no_time_series=True,\n include_increment=False, clobber_images_directory=False, resize_factor=\n 4, multiprocess=False)\n", (4497, 4662), False, 'from create_movie import create_movie\n'), ((2741, 2753), 'astropy.io.fits.open', 'fits.open', (['i'], {}), '(i)\n', (2750, 2753), False, 'from astropy.io import fits\n')]
from csv import DictReader from os.path import dirname, join, exists from os import makedirs, stat from concurrent.futures import ThreadPoolExecutor, as_completed import urllib.request from time import time def get_file_by_case(case): D = r'C:\Users\andy\supreme-court-scrap\documents-from-student-decoding' parts = case.split('/') year = parts[-1] case = parts[0].zfill(6) a, b, c = case[:2], case[2:4], case[4:6] dir = join(D, year, a, b) outfile = join(dir, f'{c}.html') return outfile def write_out_file(case, text): outfile = get_file_by_case(case) makedirs(dirname(outfile), 777, True) open(outfile, 'w', encoding='Windows-1255').write(text) print(f'Wrote to {outfile}') def parse_case(case): url = case['url'] case_id = case['casename'].split()[1] return case_id, url def exists_case(case_id): return exists(get_file_by_case(case_id)) def download(case): case_id, url = parse_case(case) print(f'Downloading {case_id}') s = time() response = urllib.request.urlopen(url) data = response.read() # a `bytes` object text = data.decode('Windows-1255') # a `str`; this elapsed = time() -s print(f'Downloaded {case_id} in {elapsed}') write_out_file(case_id, text) return f = r'C:\Users\andy\supreme-court-scrap\answers-13.csv' f = open(f, 'r', encoding='utf-8') f = DictReader(f) futures = [] with ThreadPoolExecutor(max_workers=20) as tpe: for case in f: case_id, url = parse_case(case) if exists_case(case_id): print(f'Skipping {case_id}') continue future = tpe.submit(download, case) futures.append(future) for future in as_completed(futures): future.result()
[ "csv.DictReader", "concurrent.futures.ThreadPoolExecutor", "os.path.join", "concurrent.futures.as_completed", "os.path.dirname", "time.time" ]
[((1378, 1391), 'csv.DictReader', 'DictReader', (['f'], {}), '(f)\n', (1388, 1391), False, 'from csv import DictReader\n'), ((447, 466), 'os.path.join', 'join', (['D', 'year', 'a', 'b'], {}), '(D, year, a, b)\n', (451, 466), False, 'from os.path import dirname, join, exists\n'), ((481, 503), 'os.path.join', 'join', (['dir', 'f"""{c}.html"""'], {}), "(dir, f'{c}.html')\n", (485, 503), False, 'from os.path import dirname, join, exists\n'), ((1012, 1018), 'time.time', 'time', ([], {}), '()\n', (1016, 1018), False, 'from time import time\n'), ((1411, 1445), 'concurrent.futures.ThreadPoolExecutor', 'ThreadPoolExecutor', ([], {'max_workers': '(20)'}), '(max_workers=20)\n', (1429, 1445), False, 'from concurrent.futures import ThreadPoolExecutor, as_completed\n'), ((1703, 1724), 'concurrent.futures.as_completed', 'as_completed', (['futures'], {}), '(futures)\n', (1715, 1724), False, 'from concurrent.futures import ThreadPoolExecutor, as_completed\n'), ((606, 622), 'os.path.dirname', 'dirname', (['outfile'], {}), '(outfile)\n', (613, 622), False, 'from os.path import dirname, join, exists\n'), ((1179, 1185), 'time.time', 'time', ([], {}), '()\n', (1183, 1185), False, 'from time import time\n')]
from pathlib import Path from shutil import copyfile import logging def move_files(dir_from: Path, dir_to: Path): logging.info(f"Moving files from '{dir_from}' to '{dir_to}'") p = dir_from.glob("**/*") input_paths = [x for x in p if x.is_file()] for input_path in input_paths: filename = input_path.name output_path = dir_to / filename copyfile(input_path, output_path) logging.info(f"Moved file: {filename}")
[ "shutil.copyfile", "logging.info" ]
[((120, 181), 'logging.info', 'logging.info', (['f"""Moving files from \'{dir_from}\' to \'{dir_to}\'"""'], {}), '(f"Moving files from \'{dir_from}\' to \'{dir_to}\'")\n', (132, 181), False, 'import logging\n'), ((378, 411), 'shutil.copyfile', 'copyfile', (['input_path', 'output_path'], {}), '(input_path, output_path)\n', (386, 411), False, 'from shutil import copyfile\n'), ((420, 459), 'logging.info', 'logging.info', (['f"""Moved file: {filename}"""'], {}), "(f'Moved file: {filename}')\n", (432, 459), False, 'import logging\n')]
#!/usr/bin/python #-*-coding=utf-8-*- import json import jieba class emo: dict = {} emoVal = {} emodir = "emo_dict" def __init__(self,dictname): self.dict = self.loadDict(dictname) def loadDict(self,dictname): path = "./"+self.emodir+"/"+dictname+".json" print ("loading "+ path +"......") f = open(path, mode='r') try: js = json.load(f) print (["new dictionary loaded!",js]) f.close() return js except Exception : print (Exception) def update(self,dictname): new_dict = self.loadDict(dictname) for w in new_dict: try: dict[w].append(new_dict[w]) except KeyError: dict[w] = (dict[w]+new_dict[w])/2.0 def getEmo(self, txts, name): Eval = 0 wordsNum = 0 for w in txts: if w in self.dict: Eval += self.dict[w] wordsNum+=1 if name in self.emoVal: self.emoVal[name]['totalV'] += Eval self.emoVal[name]['wnum'] += wordsNum else: self.emoVal[name] = dict(wnum = wordsNum, totalV = Eval) return self.emoVal[name] def clear(self): self.emoVal = {} #test #emo = emo("emodata") #print emo.dict #txt = "2011年底泽熙在重庆啤酒的抢反弹广为人知,徐翔事后说:“重庆啤酒不是股票,是彩票。第一次刮出来‘谢谢你’,第二次刮出来还是‘谢谢你’,这时候大家都把它当废纸扔了,但彩票还没刮" #test_txts = jieba.cut(txt) #print emo.getEmo(test_txts,"重庆啤酒")
[ "json.load" ]
[((412, 424), 'json.load', 'json.load', (['f'], {}), '(f)\n', (421, 424), False, 'import json\n')]
import random import numpy as np from gym_multigrid.multigrid import World from gym_multigrid.multigrid import DIR_TO_VEC from gym_multigrid.multigrid import Actions class Agent: def __init__(self, agent_id, agent_type=0): self.id = agent_id self.total_reward = 0 self.action_probabilities = [0.1, 0.2, 0.2, 0.4, 0.1] self.agent_type = agent_type self.observation = None def next_action(self, observation, reward, round_id): pass def start_simulation(self, observation, rounds): pass def end_simulation(self, observation, reward, round_id): pass def random_action(self): action = random.choices(np.arange(5), weights=self.action_probabilities, k=1)[0] return action def get_my_position(self): width = len(self.observation) height = len(self.observation[0]) for x in range(width): for y in range(height): if self.observation[x][y][0] == World.OBJECT_TO_IDX["agent"] and self.observation[x][y][2] == self.id: return x, y return -1, -1 def get_all_ball_positions(self): width = len(self.observation) height = len(self.observation[0]) positions_x = [] positions_y = [] for x in range(width): for y in range(height): if self.observation[x][y][0] == World.OBJECT_TO_IDX["ball"]: positions_x.append(x) positions_y.append(y) return positions_x, positions_y """ actions: still = 0 left = 1 right = 2 forward = 3 pickup = 4 """ class RandomAgent(Agent): def __init__(self, agent_id): super().__init__(agent_id, agent_type=1) def start_simulation(self, observation, rounds): """ Nothing to be done """ def next_action(self, observation, reward, round_id): #print("random index: ", self.id, " type: ", self.agent_type) return self.random_action() def end_simulation(self, observation, reward, round_id): """ Nothing to be done """ class GreedyAgent(Agent): def __init__(self, agent_id): super().__init__(agent_id, agent_type=2) self.width = 0 self.height = 0 def get_ball_positions(self): positions = [] for x in range(self.width): for y in range(self.height): if self.observation[x][y][0] == World.OBJECT_TO_IDX["ball"]: positions.append([x, y]) return positions def greedy_action(self): pos_x, pos_y = self.get_my_position() direction = self.observation[pos_x][pos_y][1] ball_positions = self.get_ball_positions() target_ball_positions = get_closest_balls(pos_x, pos_y, direction, ball_positions) target_ball_position = random.choice(target_ball_positions) return move_towards_ball(pos_x, pos_y, direction, target_ball_position[0], target_ball_position[1]) def start_simulation(self, observation, rounds): self.width = len(observation) self.height = len(observation[0]) def next_action(self, observation, reward, round_id): self.observation = observation x, y = self.get_my_position() #print("greedy index: ", self.id, " type: ", x, " ", y) return self.greedy_action() def end_simulation(self, observation, reward, round_id): """ Nothing to be done """ def sign(x): if x > 0: return 1 elif x < 0: return -1 else: return 0 def distance_from_ball(pos_x, pos_y, direction, ball_x, ball_y): dx = ball_x - pos_x dy = ball_y - pos_y turns_x = abs(sign(dx) - DIR_TO_VEC[direction][0]) turns_y = abs(sign(dy) - DIR_TO_VEC[direction][1]) return abs(dx) + abs(dy) + max(turns_x, turns_y) def get_closest_balls(pos_x, pos_y, direction, ball_positions): if len(ball_positions) == 0: return [[0, 0]] best_positions = [] best_distance = -1 for index, [x, y] in enumerate(ball_positions): current_distance = distance_from_ball(pos_x, pos_y, direction, x, y) if best_distance == -1 or current_distance < best_distance: best_distance = current_distance best_positions = [[x, y]] elif current_distance == best_distance: best_positions.append([x, y]) return best_positions def get_next_state(pos_x, pos_y, direction, action): if action == Actions.still: return pos_x, pos_y, direction if action == Actions.left: new_direction = (direction + 3) % 4 return pos_x, pos_y, new_direction if action == Actions.right: new_direction = (direction + 1) % 4 return pos_x, pos_y, new_direction if action == Actions.forward: return pos_x + DIR_TO_VEC[direction][0], pos_y + DIR_TO_VEC[direction][1], direction def move_towards_ball(pos_x, pos_y, direction, ball_x, ball_y): distance = distance_from_ball(pos_x, pos_y, direction, ball_x, ball_y) if distance == 1: return Actions.pickup best_action = Actions.still best_next_distance = distance for action in [Actions.left, Actions.right, Actions.forward]: next_x, next_y, next_direction = get_next_state(pos_x, pos_y, direction, action) current_next_distance = distance_from_ball(next_x, next_y, next_direction, ball_x, ball_y) if current_next_distance < best_next_distance: best_next_distance = current_next_distance best_action = action return best_action
[ "random.choice", "numpy.arange" ]
[((2837, 2873), 'random.choice', 'random.choice', (['target_ball_positions'], {}), '(target_ball_positions)\n', (2850, 2873), False, 'import random\n'), ((693, 705), 'numpy.arange', 'np.arange', (['(5)'], {}), '(5)\n', (702, 705), True, 'import numpy as np\n')]
# -*- coding: UTF-8 -*- import glob import numpy as np import pandas as pd from PIL import Image import random # h,w = 60,50 h, w = (60, 50) size = h * w # Receding_Hairline Wearing_Necktie Rosy_Cheeks Eyeglasses Goatee Chubby # Sideburns Blurry Wearing_Hat Double_Chin Pale_Skin Gray_Hair Mustache Bald label_cls = 'Eyeglasses' pngs = sorted(glob.glob('./data/img_align_celeba/*.jpg')) data = pd.read_table('./data/list_attr_celeba.txt', delim_whitespace=True, error_bad_lines=False) eyeglasses = np.array(data[label_cls]) eyeglasses_cls = (eyeglasses + 1)/2 label_glasses = np.zeros((202599, 2)) correct_list = [] correct_list_test = [] false_list = [] false_list_test = [] for i in range(len(label_glasses)): if eyeglasses_cls[i] == 1: label_glasses[i][1] = 1 if i < 160000: correct_list.append(i) else: correct_list_test.append(i) else: label_glasses[i][0] = 1 if i < 160000: false_list.append(i) else: false_list_test.append(i) print(len(correct_list_test), len(false_list_test)) training_set_label = label_glasses[0:160000, :] test_set_label = label_glasses[160000:, :] training_set_cls = eyeglasses_cls[0:160000] test_set_cls = eyeglasses_cls[160000:] def create_trainbatch(num=10, channel=0): train_num = random.sample(false_list, num) if channel == 0: train_set = np.zeros((num, h, w)) else: train_set = np.zeros((num, h, w, 3)) train_set_label_ = [] train_set_cls_ = [] for i in range(num): img = Image.open(pngs[train_num[i]]) img_grey = img.resize((w, h)) if channel == 0: img_grey = np.array(img_grey.convert('L')) train_set[i, :, :] = img_grey else: img_grey = np.array(img_grey) train_set[i, :, :, :] = img_grey train_set_label_.append(training_set_label[train_num[i]]) train_set_cls_.append(training_set_cls[train_num[i]]) # if channel == 0: # train_set = train_set.reshape(size,num).T train_set_label_new = np.array(train_set_label_) train_set_cls_new = np.array(train_set_cls_) return train_set/255, train_set_label_new, train_set_cls_new def create_trainbatch_all_correct(num=10, channel=0): train_num = random.sample(correct_list, num) if channel == 0: train_set = np.zeros((num, h, w)) else: train_set = np.zeros((num, h, w, 3)) train_set_label_ = [] train_set_cls_ = [] n = 0 for i in range(num): img = Image.open(pngs[train_num[i]]) img_grey = img.resize((w, h)) if channel == 0: img_grey = np.array(img_grey.convert('L')) train_set[i, :, :] = img_grey else: img_grey = np.array(img_grey) train_set[i, :, :, :] = img_grey train_set_label_.append(training_set_label[train_num[i]]) train_set_cls_.append(training_set_cls[train_num[i]]) # if channel == 0: # train_set = train_set.reshape(size,num).T train_set_label_new = np.array(train_set_label_) train_set_cls_new = np.array(train_set_cls_) return train_set/255, train_set_label_new, train_set_cls_new def create_trainbatch_(num=10, channel=0): train_num1 = random.sample(correct_list, int(num/2)) train_num2 = random.sample(false_list, int(num/2)) train_num = train_num1+train_num2 if channel == 0: train_set = np.zeros((num, h, w)) else: train_set = np.zeros((num, h, w, 3)) train_set_label_ = [] train_set_cls_ = [] n = 0 for i in range(num): img = Image.open(pngs[train_num[i]]) img_grey = img.resize((w, h)) if channel == 0: img_grey = np.array(img_grey.convert('L')) train_set[i, :, :] = img_grey else: img_grey = np.array(img_grey) train_set[i, :, :, :] = img_grey train_set_label_.append(training_set_label[train_num[i]]) train_set_cls_.append(training_set_cls[train_num[i]]) # if channel == 0: # train_set = train_set.reshape(size,num).T train_set_label_new = np.array(train_set_label_) train_set_cls_new = np.array(train_set_cls_) return train_set/255, train_set_label_new, train_set_cls_new def create_trainbatch_grad(num=200, channel=0): train_num1 = random.sample(correct_list, int(10)) train_num2 = random.sample(false_list, int(190)) train_num = train_num1+train_num2 if channel == 0: train_set = np.zeros((num, h, w)) else: train_set = np.zeros((num, h, w, 3)) train_set_label_ = [] train_set_cls_ = [] n = 0 for i in range(num): img = Image.open(pngs[train_num[i]]) img_grey = img.resize((w, h)) if channel == 0: img_grey = np.array(img_grey.convert('L')) train_set[i, :, :] = img_grey else: img_grey = np.array(img_grey) train_set[i, :, :, :] = img_grey train_set_label_.append(training_set_label[train_num[i]]) train_set_cls_.append(training_set_cls[train_num[i]]) # if channel == 0: # train_set = train_set.reshape(size,num).T train_set_label_new = np.array(train_set_label_) train_set_cls_new = np.array(train_set_cls_) return train_set/255, train_set_label_new, train_set_cls_new def create_testset(num=100, channel=0): test_num1 = random.sample(correct_list_test, num) test_num2 = random.sample(false_list_test, num) test_num = test_num1 + test_num2 if channel == 0: test_set = np.zeros((num*2, h, w)) else: test_set = np.zeros((num*2, h, w, 3)) test_set_label_ = [] test_set_cls_ = [] for i in range(num*2): img = Image.open(pngs[test_num[i]]) img_grey = img.resize((w, h)) if channel == 0: img_grey = np.array(img_grey.convert('L')) test_set[i, :, :] = img_grey else: img_grey = np.array(img_grey) test_set[i, :, :, :] = img_grey test_set_label_.append(label_glasses[test_num[i]]) test_set_cls_.append(eyeglasses_cls[test_num[i]]) # if channel == 0: # test_set = test_set.reshape(size,num*2).T test_set_label_new = np.array(test_set_label_) test_set_cls_new = np.array(test_set_cls_) return test_set/255, test_set_label_new, test_set_cls_new, test_set_cls_new.mean()*100 def create_testset_all(channel=0): test_num1 = random.sample(correct_list_test, len(correct_list_test)) test_num2 = random.sample(false_list_test, len(false_list_test)) test_num = test_num1 + test_num2 # test_num = num = len(test_num) if channel == 0: test_set = np.zeros((num, h, w)) else: test_set = np.zeros((num, h, w, 3)) test_set_label_ = [] test_set_cls_ = [] for i in range(num): img = Image.open(pngs[test_num[i]]) img_grey = img.resize((w, h)) if channel == 0: img_grey = np.array(img_grey.convert('L')) test_set[i, :, :] = img_grey else: img_grey = np.array(img_grey) test_set[i, :, :, :] = img_grey test_set_label_.append(label_glasses[test_num[i]]) test_set_cls_.append(eyeglasses_cls[test_num[i]]) # if channel == 0: # test_set = test_set.reshape(size,num).T test_set_label_new = np.array(test_set_label_) test_set_cls_new = np.array(test_set_cls_) return test_set/255, test_set_label_new, test_set_cls_new, test_set_cls_new.mean()*100 def create_testset_unbalanced(channel=0): test_num1 = random.sample(correct_list_test, 10) test_num2 = random.sample(false_list_test, 190) test_num = test_num1 + test_num2 # test_num = num = len(test_num) if channel == 0: test_set = np.zeros((num, h, w)) else: test_set = np.zeros((num, h, w, 3)) test_set_label_ = [] test_set_cls_ = [] for i in range(num): img = Image.open(pngs[test_num[i]]) img_grey = img.resize((w, h)) if channel == 0: img_grey = np.array(img_grey.convert('L')) test_set[i, :, :] = img_grey else: img_grey = np.array(img_grey) test_set[i, :, :, :] = img_grey test_set_label_.append(label_glasses[test_num[i]]) test_set_cls_.append(eyeglasses_cls[test_num[i]]) # if channel == 0: # test_set = test_set.reshape(size,num).T test_set_label_new = np.array(test_set_label_) test_set_cls_new = np.array(test_set_cls_) return test_set/255, test_set_label_new, test_set_cls_new, test_set_cls_new.mean()*100
[ "random.sample", "PIL.Image.open", "numpy.array", "numpy.zeros", "pandas.read_table", "glob.glob" ]
[((411, 505), 'pandas.read_table', 'pd.read_table', (['"""./data/list_attr_celeba.txt"""'], {'delim_whitespace': '(True)', 'error_bad_lines': '(False)'}), "('./data/list_attr_celeba.txt', delim_whitespace=True,\n error_bad_lines=False)\n", (424, 505), True, 'import pandas as pd\n'), ((537, 562), 'numpy.array', 'np.array', (['data[label_cls]'], {}), '(data[label_cls])\n', (545, 562), True, 'import numpy as np\n'), ((616, 637), 'numpy.zeros', 'np.zeros', (['(202599, 2)'], {}), '((202599, 2))\n', (624, 637), True, 'import numpy as np\n'), ((360, 402), 'glob.glob', 'glob.glob', (['"""./data/img_align_celeba/*.jpg"""'], {}), "('./data/img_align_celeba/*.jpg')\n", (369, 402), False, 'import glob\n'), ((1365, 1395), 'random.sample', 'random.sample', (['false_list', 'num'], {}), '(false_list, num)\n', (1378, 1395), False, 'import random\n'), ((2127, 2153), 'numpy.array', 'np.array', (['train_set_label_'], {}), '(train_set_label_)\n', (2135, 2153), True, 'import numpy as np\n'), ((2178, 2202), 'numpy.array', 'np.array', (['train_set_cls_'], {}), '(train_set_cls_)\n', (2186, 2202), True, 'import numpy as np\n'), ((2342, 2374), 'random.sample', 'random.sample', (['correct_list', 'num'], {}), '(correct_list, num)\n', (2355, 2374), False, 'import random\n'), ((3114, 3140), 'numpy.array', 'np.array', (['train_set_label_'], {}), '(train_set_label_)\n', (3122, 3140), True, 'import numpy as np\n'), ((3165, 3189), 'numpy.array', 'np.array', (['train_set_cls_'], {}), '(train_set_cls_)\n', (3173, 3189), True, 'import numpy as np\n'), ((4192, 4218), 'numpy.array', 'np.array', (['train_set_label_'], {}), '(train_set_label_)\n', (4200, 4218), True, 'import numpy as np\n'), ((4243, 4267), 'numpy.array', 'np.array', (['train_set_cls_'], {}), '(train_set_cls_)\n', (4251, 4267), True, 'import numpy as np\n'), ((5272, 5298), 'numpy.array', 'np.array', (['train_set_label_'], {}), '(train_set_label_)\n', (5280, 5298), True, 'import numpy as np\n'), ((5323, 5347), 'numpy.array', 'np.array', (['train_set_cls_'], {}), '(train_set_cls_)\n', (5331, 5347), True, 'import numpy as np\n'), ((5473, 5510), 'random.sample', 'random.sample', (['correct_list_test', 'num'], {}), '(correct_list_test, num)\n', (5486, 5510), False, 'import random\n'), ((5527, 5562), 'random.sample', 'random.sample', (['false_list_test', 'num'], {}), '(false_list_test, num)\n', (5540, 5562), False, 'import random\n'), ((6318, 6343), 'numpy.array', 'np.array', (['test_set_label_'], {}), '(test_set_label_)\n', (6326, 6343), True, 'import numpy as np\n'), ((6367, 6390), 'numpy.array', 'np.array', (['test_set_cls_'], {}), '(test_set_cls_)\n', (6375, 6390), True, 'import numpy as np\n'), ((7451, 7476), 'numpy.array', 'np.array', (['test_set_label_'], {}), '(test_set_label_)\n', (7459, 7476), True, 'import numpy as np\n'), ((7500, 7523), 'numpy.array', 'np.array', (['test_set_cls_'], {}), '(test_set_cls_)\n', (7508, 7523), True, 'import numpy as np\n'), ((7677, 7713), 'random.sample', 'random.sample', (['correct_list_test', '(10)'], {}), '(correct_list_test, 10)\n', (7690, 7713), False, 'import random\n'), ((7730, 7765), 'random.sample', 'random.sample', (['false_list_test', '(190)'], {}), '(false_list_test, 190)\n', (7743, 7765), False, 'import random\n'), ((8554, 8579), 'numpy.array', 'np.array', (['test_set_label_'], {}), '(test_set_label_)\n', (8562, 8579), True, 'import numpy as np\n'), ((8603, 8626), 'numpy.array', 'np.array', (['test_set_cls_'], {}), '(test_set_cls_)\n', (8611, 8626), True, 'import numpy as np\n'), ((1437, 1458), 'numpy.zeros', 'np.zeros', (['(num, h, w)'], {}), '((num, h, w))\n', (1445, 1458), True, 'import numpy as np\n'), ((1489, 1513), 'numpy.zeros', 'np.zeros', (['(num, h, w, 3)'], {}), '((num, h, w, 3))\n', (1497, 1513), True, 'import numpy as np\n'), ((1605, 1635), 'PIL.Image.open', 'Image.open', (['pngs[train_num[i]]'], {}), '(pngs[train_num[i]])\n', (1615, 1635), False, 'from PIL import Image\n'), ((2416, 2437), 'numpy.zeros', 'np.zeros', (['(num, h, w)'], {}), '((num, h, w))\n', (2424, 2437), True, 'import numpy as np\n'), ((2468, 2492), 'numpy.zeros', 'np.zeros', (['(num, h, w, 3)'], {}), '((num, h, w, 3))\n', (2476, 2492), True, 'import numpy as np\n'), ((2592, 2622), 'PIL.Image.open', 'Image.open', (['pngs[train_num[i]]'], {}), '(pngs[train_num[i]])\n', (2602, 2622), False, 'from PIL import Image\n'), ((3493, 3514), 'numpy.zeros', 'np.zeros', (['(num, h, w)'], {}), '((num, h, w))\n', (3501, 3514), True, 'import numpy as np\n'), ((3545, 3569), 'numpy.zeros', 'np.zeros', (['(num, h, w, 3)'], {}), '((num, h, w, 3))\n', (3553, 3569), True, 'import numpy as np\n'), ((3669, 3699), 'PIL.Image.open', 'Image.open', (['pngs[train_num[i]]'], {}), '(pngs[train_num[i]])\n', (3679, 3699), False, 'from PIL import Image\n'), ((4572, 4593), 'numpy.zeros', 'np.zeros', (['(num, h, w)'], {}), '((num, h, w))\n', (4580, 4593), True, 'import numpy as np\n'), ((4624, 4648), 'numpy.zeros', 'np.zeros', (['(num, h, w, 3)'], {}), '((num, h, w, 3))\n', (4632, 4648), True, 'import numpy as np\n'), ((4749, 4779), 'PIL.Image.open', 'Image.open', (['pngs[train_num[i]]'], {}), '(pngs[train_num[i]])\n', (4759, 4779), False, 'from PIL import Image\n'), ((5640, 5665), 'numpy.zeros', 'np.zeros', (['(num * 2, h, w)'], {}), '((num * 2, h, w))\n', (5648, 5665), True, 'import numpy as np\n'), ((5693, 5721), 'numpy.zeros', 'np.zeros', (['(num * 2, h, w, 3)'], {}), '((num * 2, h, w, 3))\n', (5701, 5721), True, 'import numpy as np\n'), ((5811, 5840), 'PIL.Image.open', 'Image.open', (['pngs[test_num[i]]'], {}), '(pngs[test_num[i]])\n', (5821, 5840), False, 'from PIL import Image\n'), ((6781, 6802), 'numpy.zeros', 'np.zeros', (['(num, h, w)'], {}), '((num, h, w))\n', (6789, 6802), True, 'import numpy as np\n'), ((6832, 6856), 'numpy.zeros', 'np.zeros', (['(num, h, w, 3)'], {}), '((num, h, w, 3))\n', (6840, 6856), True, 'import numpy as np\n'), ((6946, 6975), 'PIL.Image.open', 'Image.open', (['pngs[test_num[i]]'], {}), '(pngs[test_num[i]])\n', (6956, 6975), False, 'from PIL import Image\n'), ((7884, 7905), 'numpy.zeros', 'np.zeros', (['(num, h, w)'], {}), '((num, h, w))\n', (7892, 7905), True, 'import numpy as np\n'), ((7935, 7959), 'numpy.zeros', 'np.zeros', (['(num, h, w, 3)'], {}), '((num, h, w, 3))\n', (7943, 7959), True, 'import numpy as np\n'), ((8049, 8078), 'PIL.Image.open', 'Image.open', (['pngs[test_num[i]]'], {}), '(pngs[test_num[i]])\n', (8059, 8078), False, 'from PIL import Image\n'), ((1833, 1851), 'numpy.array', 'np.array', (['img_grey'], {}), '(img_grey)\n', (1841, 1851), True, 'import numpy as np\n'), ((2820, 2838), 'numpy.array', 'np.array', (['img_grey'], {}), '(img_grey)\n', (2828, 2838), True, 'import numpy as np\n'), ((3898, 3916), 'numpy.array', 'np.array', (['img_grey'], {}), '(img_grey)\n', (3906, 3916), True, 'import numpy as np\n'), ((4978, 4996), 'numpy.array', 'np.array', (['img_grey'], {}), '(img_grey)\n', (4986, 4996), True, 'import numpy as np\n'), ((6037, 6055), 'numpy.array', 'np.array', (['img_grey'], {}), '(img_grey)\n', (6045, 6055), True, 'import numpy as np\n'), ((7172, 7190), 'numpy.array', 'np.array', (['img_grey'], {}), '(img_grey)\n', (7180, 7190), True, 'import numpy as np\n'), ((8275, 8293), 'numpy.array', 'np.array', (['img_grey'], {}), '(img_grey)\n', (8283, 8293), True, 'import numpy as np\n')]
# Generated by Django 3.2.4 on 2021-08-03 15:02 from django.contrib.sites.management import create_default_site from django.db import migrations def create_privacy_policy_flatpage(apps, schema_editor): Site = apps.get_model("sites", "Site") site = Site.objects.first() if not site: create_default_site(None) site = Site.objects.first() FlatPage = apps.get_model("flatpages", "FlatPage") page, created = \ FlatPage.objects.get_or_create(url='/privacy-policy/', defaults={'title': 'Privacy Policy'}) if created: page.sites.add(site) def delete_privacy_policy_flatpage(apps, schema_editor): FlatPage = apps.get_model("flatpages", "FlatPage") FlatPage.objects.filter(url='/privacy-policy/').delete() class Migration(migrations.Migration): dependencies = [ ('playlist_creation', '0001_initial'), ("sites", "0002_alter_domain_unique") ] operations = [ migrations.RunPython(create_privacy_policy_flatpage, delete_privacy_policy_flatpage), ]
[ "django.contrib.sites.management.create_default_site", "django.db.migrations.RunPython" ]
[((304, 329), 'django.contrib.sites.management.create_default_site', 'create_default_site', (['None'], {}), '(None)\n', (323, 329), False, 'from django.contrib.sites.management import create_default_site\n'), ((994, 1082), 'django.db.migrations.RunPython', 'migrations.RunPython', (['create_privacy_policy_flatpage', 'delete_privacy_policy_flatpage'], {}), '(create_privacy_policy_flatpage,\n delete_privacy_policy_flatpage)\n', (1014, 1082), False, 'from django.db import migrations\n')]
# Generated by Django 3.2.8 on 2021-10-15 16:58 from django.db import migrations class Migration(migrations.Migration): dependencies = [ ('app', '0038_auto_20211012_1634'), ] operations = [ migrations.RenameField( model_name='device', old_name='erase_wifi_credentials_at', new_name='last_event_sent_at', ), migrations.RemoveField( model_name='device', name='zero_config_at', ), migrations.RenameField( model_name='devicelogentry', old_name='erase_wifi_credentials_at', new_name='last_event_sent_at', ), migrations.RemoveField( model_name='devicelogentry', name='zero_config_at', ), ]
[ "django.db.migrations.RemoveField", "django.db.migrations.RenameField" ]
[((223, 340), 'django.db.migrations.RenameField', 'migrations.RenameField', ([], {'model_name': '"""device"""', 'old_name': '"""erase_wifi_credentials_at"""', 'new_name': '"""last_event_sent_at"""'}), "(model_name='device', old_name=\n 'erase_wifi_credentials_at', new_name='last_event_sent_at')\n", (245, 340), False, 'from django.db import migrations\n'), ((392, 458), 'django.db.migrations.RemoveField', 'migrations.RemoveField', ([], {'model_name': '"""device"""', 'name': '"""zero_config_at"""'}), "(model_name='device', name='zero_config_at')\n", (414, 458), False, 'from django.db import migrations\n'), ((503, 628), 'django.db.migrations.RenameField', 'migrations.RenameField', ([], {'model_name': '"""devicelogentry"""', 'old_name': '"""erase_wifi_credentials_at"""', 'new_name': '"""last_event_sent_at"""'}), "(model_name='devicelogentry', old_name=\n 'erase_wifi_credentials_at', new_name='last_event_sent_at')\n", (525, 628), False, 'from django.db import migrations\n'), ((680, 754), 'django.db.migrations.RemoveField', 'migrations.RemoveField', ([], {'model_name': '"""devicelogentry"""', 'name': '"""zero_config_at"""'}), "(model_name='devicelogentry', name='zero_config_at')\n", (702, 754), False, 'from django.db import migrations\n')]
from rest_framework.views import APIView from rest_framework.response import Response from rest_framework import status from rest_framework import viewsets from rest_framework.authentication import TokenAuthentication from rest_framework import filters from profiles_api import serializers from profiles_api import models from profiles_api import permissions class HelloApiView(APIView): """Test API View""" serializer_class = serializers.HelloSerializer def get(self,request, format=None): """returns a list of APIView features""" an_apiview = [ 'uses HTTTP mothod as function(get, post, patch, put, delete) ', 'Is simmilar to Traditional Django View', 'Gives you the most control over your application logic', 'Is mapped manually to URLs', ] return Response({'message':'Hello!','an_apiview':an_apiview}) def post(self, request): """Create a hello message with our name""" serializer = self.serializer_class(data=request.data) if serializer.is_valid(): name = serializer.validated_data.get('name') message = f'Hello {name}' return Response({'message':message}) else: return Response( serializer.errors, status = status.HTTP_400_BAD_REQUEST ) def put(self, request, pk=None): """Handle updating an object""" return Response({'method': 'PUT'}) def patch(self, request, pk=None): """Handle partial update of object""" return Response({'method': 'PATCH'}) def delete(self, request, pk=None): """Delete an object""" return Response({'method': 'DELETE'}) class HelloViewSet(viewsets.ViewSet): """Test API ViewSet""" serializer_class = serializers.HelloSerializer def list(self, request): """Return a hello message.""" a_viewset = [ 'Uses actions (list, create, retrieve, update, partial_update)', 'Automatically maps to URLS using Routers', 'Provides more functionality with less code', ] return Response({'message': 'Hello!', 'a_viewset': a_viewset}) def create(self, request): """create a new hello mesage""" serializer= self.serializer_class(data=request.data) if serializer.is_valid(): name = serializer.validated_data.get('name') message = f'hello{name}!' return Response({'message':message}) else: return Response( serializer.errors, status = status.HTTP_400_BAD_REQUEST ) def retrieve(self, request,pk=None): """habdle getting an object by its id""" return Response({'http_method': 'GET'}) def update(self, request, pk=None): """handle updateing an object""" return Response({'http_method':'PUT'}) def partial_update(self, request, pk=None): """handle updating part of an object""" return Response({'http_method':'PATCH'}) def destroy(self, request, pk=None): """handle removing an object""" return Response({'http_method':'DELETE'}) class UserProfileViewSet(viewsets.ModelViewSet): """handle creating and updating profiles""" serializer_class = serializers.UserProfileSerializer queryset = models.UserProfile.objects.all() authentication_classes = (TokenAuthentication,) permission_class = (permissions.UpdateOwnProfile,) filter_backends = (filters.SearchFilter,) search_fields = ('name', 'email',)
[ "rest_framework.response.Response", "profiles_api.models.UserProfile.objects.all" ]
[((3382, 3414), 'profiles_api.models.UserProfile.objects.all', 'models.UserProfile.objects.all', ([], {}), '()\n', (3412, 3414), False, 'from profiles_api import models\n'), ((834, 891), 'rest_framework.response.Response', 'Response', (["{'message': 'Hello!', 'an_apiview': an_apiview}"], {}), "({'message': 'Hello!', 'an_apiview': an_apiview})\n", (842, 891), False, 'from rest_framework.response import Response\n'), ((1449, 1476), 'rest_framework.response.Response', 'Response', (["{'method': 'PUT'}"], {}), "({'method': 'PUT'})\n", (1457, 1476), False, 'from rest_framework.response import Response\n'), ((1578, 1607), 'rest_framework.response.Response', 'Response', (["{'method': 'PATCH'}"], {}), "({'method': 'PATCH'})\n", (1586, 1607), False, 'from rest_framework.response import Response\n'), ((1695, 1725), 'rest_framework.response.Response', 'Response', (["{'method': 'DELETE'}"], {}), "({'method': 'DELETE'})\n", (1703, 1725), False, 'from rest_framework.response import Response\n'), ((2158, 2213), 'rest_framework.response.Response', 'Response', (["{'message': 'Hello!', 'a_viewset': a_viewset}"], {}), "({'message': 'Hello!', 'a_viewset': a_viewset})\n", (2166, 2213), False, 'from rest_framework.response import Response\n'), ((2771, 2803), 'rest_framework.response.Response', 'Response', (["{'http_method': 'GET'}"], {}), "({'http_method': 'GET'})\n", (2779, 2803), False, 'from rest_framework.response import Response\n'), ((2901, 2933), 'rest_framework.response.Response', 'Response', (["{'http_method': 'PUT'}"], {}), "({'http_method': 'PUT'})\n", (2909, 2933), False, 'from rest_framework.response import Response\n'), ((3045, 3079), 'rest_framework.response.Response', 'Response', (["{'http_method': 'PATCH'}"], {}), "({'http_method': 'PATCH'})\n", (3053, 3079), False, 'from rest_framework.response import Response\n'), ((3176, 3211), 'rest_framework.response.Response', 'Response', (["{'http_method': 'DELETE'}"], {}), "({'http_method': 'DELETE'})\n", (3184, 3211), False, 'from rest_framework.response import Response\n'), ((1181, 1211), 'rest_framework.response.Response', 'Response', (["{'message': message}"], {}), "({'message': message})\n", (1189, 1211), False, 'from rest_framework.response import Response\n'), ((1244, 1307), 'rest_framework.response.Response', 'Response', (['serializer.errors'], {'status': 'status.HTTP_400_BAD_REQUEST'}), '(serializer.errors, status=status.HTTP_400_BAD_REQUEST)\n', (1252, 1307), False, 'from rest_framework.response import Response\n'), ((2496, 2526), 'rest_framework.response.Response', 'Response', (["{'message': message}"], {}), "({'message': message})\n", (2504, 2526), False, 'from rest_framework.response import Response\n'), ((2560, 2623), 'rest_framework.response.Response', 'Response', (['serializer.errors'], {'status': 'status.HTTP_400_BAD_REQUEST'}), '(serializer.errors, status=status.HTTP_400_BAD_REQUEST)\n', (2568, 2623), False, 'from rest_framework.response import Response\n')]
# -*- coding: utf-8 -*- import random, json, string, warnings from typing import Union, List, Set, Dict from simalign import SentenceAligner from nltk.translate.phrase_based import phrase_extraction class CodeMixer(object): def __init__(self, matrix_lg: str, embedded_lgs: Union[List[str],Set[str]], device='cuda', precomputed_phrases=False): warnings.filterwarnings("ignore", category=FutureWarning) if not precomputed_phrases: self.aligner = SentenceAligner(model="xlmr", token_type="bpe", matching_methods="m", device=device) self.rtl_lgs = {'ar', 'he'} self.matrix_lg = matrix_lg self.embedded_lgs = set(embedded_lgs) def get_phrases(self, matrix_sentence: str, translations: Dict[str,str], sample_lgs: Union[List[str],Set[str]]=None): if not sample_lgs: sample_lgs = self.embedded_lgs filtered_translations = {k: v for k,v in translations.items() if k in sample_lgs} matrix_tokens = matrix_sentence.split() phrases = [] for lg, embedded_sentence in filtered_translations.items(): tokenized_embedded = embedded_sentence.split() alignments = self.aligner.get_word_aligns(matrix_tokens, tokenized_embedded) candidate_phrases = phrase_extraction(matrix_sentence, embedded_sentence, alignments['mwmf'], 4) candidate_phrases = [(p[0], p[1], p[2], p[3], lg) for p in candidate_phrases if p[2][0] not in string.punctuation] if lg == 'zh': candidate_phrases = [(p[0], p[1], p[2], p[3].replace(' ', ''), p[4]) for p in candidate_phrases] if lg == 'th': candidate_phrases = [(p[0], p[1], p[2], p[3].replace(' ', ' ').replace(' ,', ','), p[4]) for p in candidate_phrases] phrases += candidate_phrases sorted_phrases = sorted(phrases, key=lambda x: (x[0][0], -x[0][1])) grouped_phrases = {i:[] for i in range(len(matrix_tokens))} for phrase in sorted_phrases: grouped_phrases[phrase[0][0]].append(phrase) return grouped_phrases def swap_phrase(self, tokens, replace_start_idx, replace_end_idx, to_replace): return tokens[0:replace_start_idx] + [to_replace] + tokens[replace_end_idx:] def get_weights(self, lg_counts): filtered_lg_counts = {k: v for k,v in lg_counts.items() if k in self.embedded_lgs or k == self.matrix_lg} total_count = sum(filtered_lg_counts.values()) return {k: v/total_count for k,v in filtered_lg_counts.items()} def generate(self, sentence, reference_translations, probability=0.15, lg_counts: Dict[str,int]=None): phrases = self.get_phrases(sentence, reference_translations) return generate_precomputed_alignments(sentence, phrases, probability, lg_counts) def generate_precomputed_alignments(self, sentence, phrase_alignments, probability=0.15):#, lg_counts: Dict[str,int]=None): tokens = sentence.split() token_length = len(tokens) pos = 0 prev_lg = self.matrix_lg prev_replacement_pos = pos while pos < token_length: candidates = phrase_alignments.get(pos) pos += 1 if random.random() >= probability or not candidates: prev_lg = self.matrix_lg continue eligible_candidates = [] for candidate in candidates: phrase_to_replace = candidate[2] replacement = candidate[3] replacement_lg = candidate[4] replace_start_idx = candidate[0][0] - token_length replace_end_idx = candidate[0][1] - token_length if phrase_to_replace.split() != tokens[replace_start_idx:replace_end_idx]: continue if replacement_lg not in self.rtl_lgs and replacement_lg == prev_lg and candidate[1][1] <= prev_replacement_pos: continue eligible_candidates.append(candidate) if eligible_candidates: chosen_candidate = random.choice(eligible_candidates) replacement_lg = chosen_candidate[4] replacement = chosen_candidate[3] replace_start_idx = chosen_candidate[0][0] - token_length replace_end_idx = chosen_candidate[0][1] - token_length tokens = self.swap_phrase(tokens, replace_start_idx, replace_end_idx, replacement) prev_lg = replacement_lg prev_replacement_pos = pos pos = max(replace_end_idx, pos) else: prev_lg = self.matrix_lg continue return ' '.join(tokens)
[ "random.choice", "nltk.translate.phrase_based.phrase_extraction", "random.random", "warnings.filterwarnings", "simalign.SentenceAligner" ]
[((357, 414), 'warnings.filterwarnings', 'warnings.filterwarnings', (['"""ignore"""'], {'category': 'FutureWarning'}), "('ignore', category=FutureWarning)\n", (380, 414), False, 'import random, json, string, warnings\n'), ((478, 566), 'simalign.SentenceAligner', 'SentenceAligner', ([], {'model': '"""xlmr"""', 'token_type': '"""bpe"""', 'matching_methods': '"""m"""', 'device': 'device'}), "(model='xlmr', token_type='bpe', matching_methods='m',\n device=device)\n", (493, 566), False, 'from simalign import SentenceAligner\n'), ((1301, 1377), 'nltk.translate.phrase_based.phrase_extraction', 'phrase_extraction', (['matrix_sentence', 'embedded_sentence', "alignments['mwmf']", '(4)'], {}), "(matrix_sentence, embedded_sentence, alignments['mwmf'], 4)\n", (1318, 1377), False, 'from nltk.translate.phrase_based import phrase_extraction\n'), ((4118, 4152), 'random.choice', 'random.choice', (['eligible_candidates'], {}), '(eligible_candidates)\n', (4131, 4152), False, 'import random, json, string, warnings\n'), ((3246, 3261), 'random.random', 'random.random', ([], {}), '()\n', (3259, 3261), False, 'import random, json, string, warnings\n')]
#! python3 # To write files and make a directory import os # To filter string import re # For sleeping import time # Googles Text to speech lib from gtts import gTTS # File info reader, in this case, it is for determening the length of a mp3 from mutagen.mp3 import MP3 def format_text(text: str): """ Removes unwanted characters Replaces spaces with underscores """ whitelist = re.compile(r'[^a-zA-Z ]+') return re.sub(whitelist, '', text).replace(' ', '_') def create_dir(dir_name: str): """Creates directory when not made yet""" if not os.path.isdir(dir_name): os.makedirs(dir_name) def create_spell_out(text: str): """Maps the characters to the NATO words""" words_for_letters = ['Alfa', 'Bravo', 'Charlie', 'Delta', 'Echo', 'Foxtrot', 'Golf', 'Hotel', 'India', 'Juliett', 'Kilo', 'Lima', 'Mike', 'November', 'Oscar', 'Papa', 'Quebec', 'Romeo', 'Sierra', 'Tango', 'Uniform', 'Victor', 'Whiskey', 'X-ray', 'Yankee', 'Zulu', 'underscore'] # ord('{') - 97 == words_for_letters.index('underscore') words = [word.replace('_', '{') for word in text.split(' ')] return [words_for_letters[ord( letter.lower()) - 97] for word in words for letter in word] def save_audio(text: str, filename: str, dir: str): """ Converts text to audio and saves Notes ----- If the .mp3 file extension is missing in the filename, it will be added If a file with the same name exists, it will not save, only notify the user Returns _______ Path : str """ # Make the path to the folder path = '{0}/{1}'.format(dir, filename) if not filename.endswith('.mp3'): path += '.mp3' # Generates and saves audio file tts = gTTS(text=text, lang='en') # Only saves when file does not exist if os.path.isfile(path): print("File named {0} already exist, will not safe".format(path)) else: tts.save(path) return path def play_audio(path: str): os.startfile(os.getcwd() + path[1:]) duration = MP3(path).info.length time.sleep(duration) # ---------- MAIN-PROGRAM ---------- if __name__ == '__main__': print('Welcome!') output_dir = './audio' create_dir(output_dir) while True: text = format_text(input('What is the sentence?')) output_words = 'Message incoming' + ', '.join(create_spell_out(text)) path = save_audio(output_words, text, output_dir) play_audio(path) if(input('Continue? y/n') not in ('y', 'yes')): break
[ "os.makedirs", "re.compile", "time.sleep", "os.getcwd", "os.path.isfile", "mutagen.mp3.MP3", "os.path.isdir", "gtts.gTTS", "re.sub" ]
[((406, 431), 're.compile', 're.compile', (['"""[^a-zA-Z ]+"""'], {}), "('[^a-zA-Z ]+')\n", (416, 431), False, 'import re\n'), ((1843, 1869), 'gtts.gTTS', 'gTTS', ([], {'text': 'text', 'lang': '"""en"""'}), "(text=text, lang='en')\n", (1847, 1869), False, 'from gtts import gTTS\n'), ((1920, 1940), 'os.path.isfile', 'os.path.isfile', (['path'], {}), '(path)\n', (1934, 1940), False, 'import os\n'), ((2176, 2196), 'time.sleep', 'time.sleep', (['duration'], {}), '(duration)\n', (2186, 2196), False, 'import time\n'), ((580, 603), 'os.path.isdir', 'os.path.isdir', (['dir_name'], {}), '(dir_name)\n', (593, 603), False, 'import os\n'), ((613, 634), 'os.makedirs', 'os.makedirs', (['dir_name'], {}), '(dir_name)\n', (624, 634), False, 'import os\n'), ((444, 471), 're.sub', 're.sub', (['whitelist', '""""""', 'text'], {}), "(whitelist, '', text)\n", (450, 471), False, 'import re\n'), ((2111, 2122), 'os.getcwd', 'os.getcwd', ([], {}), '()\n', (2120, 2122), False, 'import os\n'), ((2150, 2159), 'mutagen.mp3.MP3', 'MP3', (['path'], {}), '(path)\n', (2153, 2159), False, 'from mutagen.mp3 import MP3\n')]
#!/usr/bin/env python import json jsonFile = open('temp.json', 'r') jsonData = json.load(jsonFile) places = jsonData["places"] for item in places: originalArray = places[item] newArray = ["default"] for originalPiece in originalArray: newArray.append(originalPiece) places[item] = newArray jsonData["places"] = places newJsonFile = open('newStuff.json', 'w') json.dump(jsonData, newJsonFile) jsonFile.close() newJsonFile.close()
[ "json.load", "json.dump" ]
[((81, 100), 'json.load', 'json.load', (['jsonFile'], {}), '(jsonFile)\n', (90, 100), False, 'import json\n'), ((372, 404), 'json.dump', 'json.dump', (['jsonData', 'newJsonFile'], {}), '(jsonData, newJsonFile)\n', (381, 404), False, 'import json\n')]
from setuptools import setup, Extension setup(ext_modules=[ Extension('_module', sources=["module_wrap.c"]) ])
[ "setuptools.Extension" ]
[((65, 112), 'setuptools.Extension', 'Extension', (['"""_module"""'], {'sources': "['module_wrap.c']"}), "('_module', sources=['module_wrap.c'])\n", (74, 112), False, 'from setuptools import setup, Extension\n')]
from gaesessions import SessionMiddleware # suggestion: generate your own random key using os.urandom(64) # WARNING: Make sure you run os.urandom(64) OFFLINE and copy/paste the output to # this file. If you use os.urandom() to *dynamically* generate your key at # runtime then any existing sessions will become junk every time you start, # deploy, or update your app! import os # https://cloud.google.com/appengine/docs/standard/python/tools/using-libraries-python-27 # appengine_config.py from google.appengine.ext import vendor # Add any libraries install in the "lib" folder. # vendor.add('lib') vendor.add(os.path.join(os.path.dirname(os.path.realpath(__file__)), 'lib')) # def add_vendor_packages(vendor_folder): # """ # Adds our vendor packages folder to sys.path so that third-party # packages can be imported. # """ # import site # import os.path # import sys # # # Use site.addsitedir() because it appropriately reads .pth # # files for namespaced packages. Unfortunately, there's not an # # option to choose where addsitedir() puts its paths in sys.path # # so we have to do a little bit of magic to make it play along. # # # We're going to grab the current sys.path and split it up into # # the first entry and then the rest. Essentially turning # # ['.', '/site-packages/x', 'site-packages/y'] # # into # # ['.'] and ['/site-packages/x', 'site-packages/y'] # # The reason for this is we want '.' to remain at the top of the # # list but we want our vendor files to override everything else. # sys.path, remainder = sys.path[:1], sys.path[1:] # # # Now we call addsitedir which will append our vendor directories # # to sys.path (which was truncated by the last step.) # site.addsitedir(os.path.join(os.path.dirname(__file__), vendor_folder)) # # # Finally, we'll add the paths we removed back. # sys.path.extend(remainder) # # # Change 'lib' to whichever directory you use for your vendored packages. # add_vendor_packages('lib') COOKIE_KEY = "" # TODO: Generate new COOKIE_KEY using os.urandom(64) See above def webapp_add_wsgi_middleware(app): from google.appengine.ext.appstats import recording app = SessionMiddleware(app, cookie_key=COOKIE_KEY) app = recording.appstats_wsgi_middleware(app) return app
[ "os.path.realpath", "google.appengine.ext.appstats.recording.appstats_wsgi_middleware", "gaesessions.SessionMiddleware" ]
[((2228, 2273), 'gaesessions.SessionMiddleware', 'SessionMiddleware', (['app'], {'cookie_key': 'COOKIE_KEY'}), '(app, cookie_key=COOKIE_KEY)\n', (2245, 2273), False, 'from gaesessions import SessionMiddleware\n'), ((2282, 2321), 'google.appengine.ext.appstats.recording.appstats_wsgi_middleware', 'recording.appstats_wsgi_middleware', (['app'], {}), '(app)\n', (2316, 2321), False, 'from google.appengine.ext.appstats import recording\n'), ((642, 668), 'os.path.realpath', 'os.path.realpath', (['__file__'], {}), '(__file__)\n', (658, 668), False, 'import os\n')]
import flask blueprint = flask.Blueprint('rooms', __name__, template_folder='templates') from . import routes
[ "flask.Blueprint" ]
[((26, 89), 'flask.Blueprint', 'flask.Blueprint', (['"""rooms"""', '__name__'], {'template_folder': '"""templates"""'}), "('rooms', __name__, template_folder='templates')\n", (41, 89), False, 'import flask\n')]
# Debugging #T# Table of contents #C# Python debugger (pdb) #T# Beginning of content #C# Python debugger (pdb) # |------------------------------------------------------------- #T# pdb is the builtin Python debugger, it has breakpoints, stepping through the code, printing the values of variables, post-mortem debugging, debugging of modules, functions, scripts, among other features #T# pdb can be executed with an script argument to debug said script, the following syntax is done in the operating system shell # SYNTAX python3 -m pdb script1.py #T# python3 is the Python executable, -m pdb script1.py runs pdb to debug script1.py (see the file titled Interpreter), this automatically enters post-mortem if script1 crashes #T# the pdb module is imported to use the pdb debugger as part of a script import pdb #T# the run function of the pdb module allows debugging the execution of a Python string # SYNTAX pdb.run('string1') #T# the pdb debugger is started right before the execution of string1, and is used to debug whatever string1 executes pdb.run('import S01_Basic_syntax') # this debugs the S01_Basic_syntax.py file, because the import statement executes the imported module #T# the following code is used to show the syntax of the pdb debugger in its interactive mode output_var1 = "help variable to show the different output of the pdb debugger" var1 = [5, 2, 3] var2 = 7 def func1(num1, num2): num3 = num1 + num2 print("func1_string1") return num3 def func2(): func1(var1[0], var1[2]) loc1 = 72 for i1 in [1, 2, 3]: print("i1 is", i1) func2() #T# create a breakpoint with the breakpoint function, this starts the (Pdb) interpreter to do interactive debugging breakpoint() # |--------------------------------------------------\ #T# the following syntaxes are written in the pdb debugger language, so they can't be written outside of a comment because they are not valid Python syntax and this .py file would show errors in an IDE (IDE stands for Integrated Development Environment) #T# the 'output_var1' variable used in the following is used as a helper to signal and display the output of the pdb debugger commands #T# the pdb debugger prompt is (Pdb), so anything shown after a (Pdb) means that it was typed in said prompt, e.g. '(Pdb) prompt_typings1' prompt_typings1 was typed directly in the pdb debugger prompt #T# when source code is printed, the current line is shown with '->' after the line number # SYNTAX next #T# the next command executes code up to the next line of code (not entering functions) output_var1 # (Pdb) next # this shows output of the script or program under debugging # SYNTAX step #T# the step command steps into functions or the next line output_var1 # (Pdb) step # this shows output of the script or program under debugging, possibly inside a function # SYNTAX continue #T# the continue command continues execution until a breakpoint is found output_var1 # (Pdb) continue # this shows output of the script or program under debugging, up to the next breakpoint or the end of the file # SYNTAX until int1 #T# the until command continues execution until a line of number int1 or greater is reached, without int1 it continues until the next bigger line number output_var1 # (Pdb) until # this shows output of the script or program under debugging, up to the next bigger line number # SYNTAX return #T# the return command continues execution until arriving at the return keyword of the current function, so this is used inside functions output_var1 # (Pdb) return # inside a function, this shows output of the script or program under debugging, up to the return keyword of the current function # SYNTAX run # SYNTAX restart #T# the run command and its alias the restart command restart the script or program under debugging, preserving the options and created breakpoints output_var1 # (Pdb) restart # the script or program restarts # SYNTAX p var1 #T# the p command (for print) prints the value of var1, if the name var1 is defined, this syntax is an alias for print(var1) output_var1 # (Pdb) p var1 # [5, 2, 3] output_var1 # (Pdb) print(var1) # [5, 2, 3] # SYNTAX p func1 #T# same as before, but when used with a function, its address is printed output_var1 # (Pdb) p func1 # <function func1 at 0x7f35210401f0> # or similar # SYNTAX p func1(arg1, arg2) #T# same as before, but this prints any output from func1 and its return value using arg1, arg2 as arguments, and any other arguments present output_var1 # (Pdb) p func1(2, 3) #T# the former prints # func1_string1 # 5 # SYNTAX args #T# the args command displays the arguments passed to a function output_var1 # (Pdb) args # inside func1(5, 3) #T# the former prints # num1 = 5 # num2 = 3 # SYNTAX display var1 #T# the display command prints a variable each time it changes # SYNTAX undisplay var1 #T# stop displaying a variable var1 with the undisplay command # SYNTAX l int1 #T# the l command lists 11 source code lines, this is done around line int1, 5 lines above and 5 lines below it, this syntax is an alias for list int1 output_var1 # (Pdb) l 7 # (Pdb) list 7 #T# the former prints # 2 # Debugging # 3 # 4 #T# Table of contents # 5 # 6 #C# Python debugger (pdb) # 7 # 8 #T# Beginning of content # 9 # 10 #C# Python debugger (pdb) # 11 # 12 # |------------------------------------------------------------- # SYNTAX l . #T# same as before, but list source code lines around the current line output_var1 # (Pdb) l . # this prints similar as before # SYNTAX ll #T# the ll commands does a long list of the source code local to the current line output_var1 # (Pdb) ll # the output is too large to put here, more than 150 lines # SYNTAX break #T# the break command alone displays all breakpoints output_var1 # (Pdb) break # with two breakpoints already created #T# the former prints # Num Type Disp Enb Where # 1 breakpoint keep yes at /path/to/S13_Debugging.py:20 # 2 breakpoint keep yes at /path/to/S13_Debugging.py:35 # stop only if var2 < 10 # SYNTAX break file1:int1 #T# the break command is used to create breakpoints in file1 (the current python script name without the .py extension), in line int1 output_var1 # (Pdb) break S13_Debugging:20 # Breakpoint 1 at /path/to/S13_Debugging.py:20 # SYNTAX break file1.func1, condition1 #T# same as before, but the breakpoint is created in the first line of func1 (its def line), and the breakpoint only activates if condition1 evaluates to True using Python boolean syntax output_var1 # (Pdb) break S13_Debugging.func1, var2 < 10 # Breakpoint 2 at /path/to/S13_Debugging.py:35 # SYNTAX disable int1 #T# the disable command disables the breakpoint numbered with the number int1 output_var1 # (Pdb) disable 1 # Disabled breakpoint 1 at /path/to/S13_Debugging.py:20 # SYNTAX enable int1 #T# the enable command enables the breakpoint numbered with the number int1 output_var1 # (Pdb) enable 1 # Enabled breakpoint 1 at /path/to/S13_Debugging.py:20 # SYNTAX clear int1 #T# the clear command completely deletes a breakpoint output_var1 # (Pdb) clear 1 # Deleted breakpoint 1 at /path/to/S13_Debugging.py:20 # SYNTAX where #T# the where command prints the stack_frame trace output_var1 # (Pdb) where #T# the former prints # /path/to/S13_Debugging.py(46)<module>() #-> func2() # /path/to/S13_Debugging.py(42)func2() #-> func1(var1[0], var1[2]) #> /path/to/S13_Debugging.py(40)func1()->8 #-> return num3 # SYNTAX up int1 #T# the up command goes up to an older frame in the stack trace, the amount of frames that go up is int1 output_var1 # (Pdb) up 1 #T# the former prints #> /path/to/S13_Debugging.py(42)func2() #-> func1(var1[0], var1[2]) # SYNTAX down int1 #T# the down command goes down to a newer frame in the stack trace, the amount of frames that go down is int1 output_var1 # (Pdb) down 1 # *** Newest frame # this is the output at the lowest frame # SYNTAX help #T# print the debugger pdb help with the help command # SYNTAX quit #T# quit the debugger with the quit command # |--------------------------------------------------/ # |-------------------------------------------------------------
[ "pdb.run" ]
[((1058, 1092), 'pdb.run', 'pdb.run', (['"""import S01_Basic_syntax"""'], {}), "('import S01_Basic_syntax')\n", (1065, 1092), False, 'import pdb\n')]
""" Created on Sat Mar 23 00:23:27 2019 @author: nahid """ #https://docs.scipy.org/doc/numpy/reference/generated/numpy.absolute.html import numpy as np import matplotlib.pyplot as plt x = np.array([-1.2, 1.2]) x = np.absolute(x) print(x) print(np.absolute(1 + 2j)) #Plot the function over [-10, 10]: x = np.linspace(-10, 10, 101); #start, end, totalElements you want to create plt.plot(np.absolute(x)) plt.show() plt.plot(x) plt.show() xx = x + 1j * x[:, np.newaxis] plt.imshow(np.abs(xx), extent=[-10, 10, -10, 10], cmap='gray') plt.show()
[ "numpy.abs", "numpy.absolute", "matplotlib.pyplot.plot", "numpy.array", "numpy.linspace", "matplotlib.pyplot.show" ]
[((189, 210), 'numpy.array', 'np.array', (['[-1.2, 1.2]'], {}), '([-1.2, 1.2])\n', (197, 210), True, 'import numpy as np\n'), ((215, 229), 'numpy.absolute', 'np.absolute', (['x'], {}), '(x)\n', (226, 229), True, 'import numpy as np\n'), ((306, 331), 'numpy.linspace', 'np.linspace', (['(-10)', '(10)', '(101)'], {}), '(-10, 10, 101)\n', (317, 331), True, 'import numpy as np\n'), ((404, 414), 'matplotlib.pyplot.show', 'plt.show', ([], {}), '()\n', (412, 414), True, 'import matplotlib.pyplot as plt\n'), ((415, 426), 'matplotlib.pyplot.plot', 'plt.plot', (['x'], {}), '(x)\n', (423, 426), True, 'import matplotlib.pyplot as plt\n'), ((427, 437), 'matplotlib.pyplot.show', 'plt.show', ([], {}), '()\n', (435, 437), True, 'import matplotlib.pyplot as plt\n'), ((533, 543), 'matplotlib.pyplot.show', 'plt.show', ([], {}), '()\n', (541, 543), True, 'import matplotlib.pyplot as plt\n'), ((245, 266), 'numpy.absolute', 'np.absolute', (['(1 + 2.0j)'], {}), '(1 + 2.0j)\n', (256, 266), True, 'import numpy as np\n'), ((388, 402), 'numpy.absolute', 'np.absolute', (['x'], {}), '(x)\n', (399, 402), True, 'import numpy as np\n'), ((481, 491), 'numpy.abs', 'np.abs', (['xx'], {}), '(xx)\n', (487, 491), True, 'import numpy as np\n')]
from __future__ import annotations import asyncio import typing import types import pandas as pd import tooltime from ctc import evm from ctc import spec async def async_get_lending_flows( wallet: spec.Address, pool_token: spec.ERC20Reference, protocol: typing.Literal['aave', 'compound', 'rari'], wallet_deposits: spec.DataFrame | None = None, deposits: spec.DataFrame | None = None, wallet_withdrawals: spec.DataFrame | None = None, withdrawals: spec.DataFrame | None = None, include_latest: bool = True, provider: spec.ProviderSpec = None, replace_symbols: bool = True, normalize: bool = True, include_rewards: bool = True, ) -> spec.DataFrame: if protocol == 'aave': from ctc.protocols import aave_v2_utils protocol_module: types.ModuleType = aave_v2_utils elif protocol == 'compound': from ctc.protocols import compound_utils protocol_module = compound_utils elif protocol == 'rari': from ctc.protocols import rari_utils protocol_module = rari_utils else: raise Exception('unknown protocol: ' + str(protocol)) df = await _async_create_raw_wallet_flows_df( wallet=wallet, wallet_deposits=wallet_deposits, deposits=deposits, wallet_withdrawals=wallet_withdrawals, withdrawals=withdrawals, include_latest=include_latest, provider=provider, ) underlying = await protocol_module.async_get_underlying_asset( pool_token=pool_token, provider=provider, ) # add time data blocks = df.index.values blocks_before = blocks - 1 # queue tasks timestamps_coroutine = evm.async_get_block_timestamps( blocks=blocks, provider=provider, ) timestamps_task = asyncio.create_task(timestamps_coroutine) pool_token_balances_before_coroutine = ( evm.async_get_erc20_balance_of_by_block( token=pool_token, address=wallet, blocks=blocks_before, provider=provider, ) ) pool_token_balances_before_task = asyncio.create_task( pool_token_balances_before_coroutine ) pool_token_balances_after_coroutine = ( evm.async_get_erc20_balance_of_by_block( token=pool_token, address=wallet, blocks=blocks, provider=provider, ) ) pool_token_balances_after_task = asyncio.create_task( pool_token_balances_after_coroutine ) asset_prices_coroutine = protocol_module.async_get_asset_price_by_block( asset=underlying, blocks=blocks, provider=provider, ) asset_prices_task = asyncio.create_task(asset_prices_coroutine) # queue optional tasks if include_rewards: reward_coroutine = protocol_module.async_compute_wallet_rewards( wallet=wallet, blocks=blocks, provider=provider, replace_symbol=replace_symbols, ) reward_task = asyncio.create_task(reward_coroutine) if normalize: decimals_coroutine = evm.async_get_erc20_decimals( underlying, provider=provider, ) decimals_task = asyncio.create_task(decimals_coroutine) if replace_symbols: underlying_symbol_coroutine = evm.async_get_erc20_symbol( underlying, provider=provider, ) underlying_symbol_task = asyncio.create_task( underlying_symbol_coroutine ) pool_token_coroutine = evm.async_get_erc20_symbol( pool_token, provider=provider, ) pool_token_symbol_task = asyncio.create_task(pool_token_coroutine) # normalize deposits and withdrawals if normalize: decimals = await decimals_task df['asset_deposit'] /= 10 ** decimals df['asset_withdrawal'] /= 10 ** decimals # compute time columns timestamps = await timestamps_task df.insert(loc=0, column='timestamp', value=timestamps) # type: ignore df.insert( loc=1, column='time', value=df['timestamp'].map(tooltime.timestamp_to_iso), ) # add pool token balances df['pool_token_balance_before'] = await pool_token_balances_before_task df['pool_token_balance_after'] = await pool_token_balances_after_task # add underlying balances df['asset_balance_before'] = df['pool_token_balance_before'] df['asset_balance_after'] = df['pool_token_balance_after'] # add asset price df['asset_price'] = await asset_prices_task df['asset_balance_usd'] = df['asset_balance_after'] * df['asset_price'] # add rewards rewards = await reward_task for key, value in rewards.items(): df[key] = value # replace symbols if replace_symbols: rename_columns = {} underlying_symbol = await underlying_symbol_task pool_token_symbol = await pool_token_symbol_task for column in df.columns: if 'asset' in column: rename_columns[column] = column.replace( 'asset', underlying_symbol ) if 'pool_token' in column: rename_columns[column] = column.replace( 'pool_token', pool_token_symbol ) df = df.rename(columns=rename_columns) return df async def _async_create_raw_wallet_flows_df( wallet: spec.Address, wallet_deposits: spec.DataFrame | None = None, deposits: spec.DataFrame | None = None, wallet_withdrawals: spec.DataFrame | None = None, withdrawals: spec.DataFrame | None = None, include_latest: bool = True, provider: spec.ProviderSpec = None, ) -> spec.DataFrame: from ctc.protocols import aave_v2_utils no_deposits = wallet_deposits is None and deposits is None no_withdrawals = wallet_withdrawals is None and withdrawals is None if no_deposits and not no_withdrawals: deposits = await aave_v2_utils.async_get_deposits() elif not no_deposits and no_withdrawals: withdrawals = await aave_v2_utils.async_get_withdrawals() elif no_deposits and no_withdrawals: deposits, withdrawals = await asyncio.gather( aave_v2_utils.async_get_deposits(provider=provider), aave_v2_utils.async_get_withdrawals(provider=provider), ) wallet = wallet.lower() if wallet_deposits is None: if deposits is None: raise Exception('could not determine deposits') wallet_deposits = deposits[deposits['arg__user'] == wallet] if isinstance(wallet_deposits.index, pd.MultiIndex): wallet_deposits = wallet_deposits.groupby(level='block_number').sum() if isinstance(wallet_deposits, pd.DataFrame): wallet_deposits_series = wallet_deposits['arg__amount'] if wallet_withdrawals is None: if withdrawals is None: raise Exception('could not determine withdrawals') wallet_withdrawals = withdrawals[withdrawals['arg__user'] == wallet] if isinstance(wallet_withdrawals.index, pd.MultiIndex): wallet_withdrawals = wallet_withdrawals.groupby( level='block_number' ).sum() if isinstance(wallet_withdrawals, pd.DataFrame): wallet_withdrawals_series = wallet_withdrawals['arg__amount'] raw_data = { 'asset_deposit': wallet_deposits_series, 'asset_withdrawal': wallet_withdrawals_series, } raw_df = pd.DataFrame(raw_data) raw_df = raw_df.fillna(0) if include_latest: block = await evm.async_get_latest_block_number(provider=provider) raw_df.loc[block] = [0, 0] return raw_df
[ "ctc.evm.async_get_erc20_decimals", "ctc.protocols.aave_v2_utils.async_get_deposits", "ctc.protocols.aave_v2_utils.async_get_withdrawals", "ctc.evm.async_get_block_timestamps", "ctc.evm.async_get_erc20_balance_of_by_block", "ctc.evm.async_get_latest_block_number", "pandas.DataFrame", "ctc.evm.async_ge...
[((1697, 1761), 'ctc.evm.async_get_block_timestamps', 'evm.async_get_block_timestamps', ([], {'blocks': 'blocks', 'provider': 'provider'}), '(blocks=blocks, provider=provider)\n', (1727, 1761), False, 'from ctc import evm\n'), ((1807, 1848), 'asyncio.create_task', 'asyncio.create_task', (['timestamps_coroutine'], {}), '(timestamps_coroutine)\n', (1826, 1848), False, 'import asyncio\n'), ((1902, 2020), 'ctc.evm.async_get_erc20_balance_of_by_block', 'evm.async_get_erc20_balance_of_by_block', ([], {'token': 'pool_token', 'address': 'wallet', 'blocks': 'blocks_before', 'provider': 'provider'}), '(token=pool_token, address=wallet,\n blocks=blocks_before, provider=provider)\n', (1941, 2020), False, 'from ctc import evm\n'), ((2120, 2177), 'asyncio.create_task', 'asyncio.create_task', (['pool_token_balances_before_coroutine'], {}), '(pool_token_balances_before_coroutine)\n', (2139, 2177), False, 'import asyncio\n'), ((2244, 2355), 'ctc.evm.async_get_erc20_balance_of_by_block', 'evm.async_get_erc20_balance_of_by_block', ([], {'token': 'pool_token', 'address': 'wallet', 'blocks': 'blocks', 'provider': 'provider'}), '(token=pool_token, address=wallet,\n blocks=blocks, provider=provider)\n', (2283, 2355), False, 'from ctc import evm\n'), ((2454, 2510), 'asyncio.create_task', 'asyncio.create_task', (['pool_token_balances_after_coroutine'], {}), '(pool_token_balances_after_coroutine)\n', (2473, 2510), False, 'import asyncio\n'), ((2708, 2751), 'asyncio.create_task', 'asyncio.create_task', (['asset_prices_coroutine'], {}), '(asset_prices_coroutine)\n', (2727, 2751), False, 'import asyncio\n'), ((7505, 7527), 'pandas.DataFrame', 'pd.DataFrame', (['raw_data'], {}), '(raw_data)\n', (7517, 7527), True, 'import pandas as pd\n'), ((3038, 3075), 'asyncio.create_task', 'asyncio.create_task', (['reward_coroutine'], {}), '(reward_coroutine)\n', (3057, 3075), False, 'import asyncio\n'), ((3123, 3182), 'ctc.evm.async_get_erc20_decimals', 'evm.async_get_erc20_decimals', (['underlying'], {'provider': 'provider'}), '(underlying, provider=provider)\n', (3151, 3182), False, 'from ctc import evm\n'), ((3242, 3281), 'asyncio.create_task', 'asyncio.create_task', (['decimals_coroutine'], {}), '(decimals_coroutine)\n', (3261, 3281), False, 'import asyncio\n'), ((3344, 3401), 'ctc.evm.async_get_erc20_symbol', 'evm.async_get_erc20_symbol', (['underlying'], {'provider': 'provider'}), '(underlying, provider=provider)\n', (3370, 3401), False, 'from ctc import evm\n'), ((3470, 3518), 'asyncio.create_task', 'asyncio.create_task', (['underlying_symbol_coroutine'], {}), '(underlying_symbol_coroutine)\n', (3489, 3518), False, 'import asyncio\n'), ((3572, 3629), 'ctc.evm.async_get_erc20_symbol', 'evm.async_get_erc20_symbol', (['pool_token'], {'provider': 'provider'}), '(pool_token, provider=provider)\n', (3598, 3629), False, 'from ctc import evm\n'), ((3698, 3739), 'asyncio.create_task', 'asyncio.create_task', (['pool_token_coroutine'], {}), '(pool_token_coroutine)\n', (3717, 3739), False, 'import asyncio\n'), ((6017, 6051), 'ctc.protocols.aave_v2_utils.async_get_deposits', 'aave_v2_utils.async_get_deposits', ([], {}), '()\n', (6049, 6051), False, 'from ctc.protocols import aave_v2_utils\n'), ((7604, 7656), 'ctc.evm.async_get_latest_block_number', 'evm.async_get_latest_block_number', ([], {'provider': 'provider'}), '(provider=provider)\n', (7637, 7656), False, 'from ctc import evm\n'), ((6125, 6162), 'ctc.protocols.aave_v2_utils.async_get_withdrawals', 'aave_v2_utils.async_get_withdrawals', ([], {}), '()\n', (6160, 6162), False, 'from ctc.protocols import aave_v2_utils\n'), ((6270, 6321), 'ctc.protocols.aave_v2_utils.async_get_deposits', 'aave_v2_utils.async_get_deposits', ([], {'provider': 'provider'}), '(provider=provider)\n', (6302, 6321), False, 'from ctc.protocols import aave_v2_utils\n'), ((6335, 6389), 'ctc.protocols.aave_v2_utils.async_get_withdrawals', 'aave_v2_utils.async_get_withdrawals', ([], {'provider': 'provider'}), '(provider=provider)\n', (6370, 6389), False, 'from ctc.protocols import aave_v2_utils\n')]
from PIL import Image, ImageDraw im = Image.open("a.jpg") im = im.resize((150,150)) result = im.convert('P', palette=Image.ADAPTIVE, colors=8) result.putalpha(0) colors = result.getcolors(150*150) newimg = Image.new('RGB', (64*8, 64)) draw = ImageDraw.Draw(newimg) posx = 0 for row, col in colors: draw.rectangle([posx, 0, posx+64, 64], fill=col) posx = posx + 64 del draw newimg.save("swatch", "JPEG")
[ "PIL.Image.new", "PIL.ImageDraw.Draw", "PIL.Image.open" ]
[((39, 58), 'PIL.Image.open', 'Image.open', (['"""a.jpg"""'], {}), "('a.jpg')\n", (49, 58), False, 'from PIL import Image, ImageDraw\n'), ((208, 238), 'PIL.Image.new', 'Image.new', (['"""RGB"""', '(64 * 8, 64)'], {}), "('RGB', (64 * 8, 64))\n", (217, 238), False, 'from PIL import Image, ImageDraw\n'), ((244, 266), 'PIL.ImageDraw.Draw', 'ImageDraw.Draw', (['newimg'], {}), '(newimg)\n', (258, 266), False, 'from PIL import Image, ImageDraw\n')]
# Copyright 2020 Google LLC # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Handles account deletion logic. Deletes all objects that are related to a user. This includes: NDB objects: - User by user_id - PageRating by user_id. - Connection by publisher_id - Connection by subscriber_id - PastRecommendation by user_id - RecommendationSession by user_id - Category by parent - export.ExportRatingResult by key Non NDB objects: - Memcache: "ri:<user_id>" - Clear text search indexes: - rating_history:<user_id> - saved_for_later:<user_id> """ from google.appengine.api import memcache from google.appengine.ext import deferred from google.appengine.ext import ndb from recommender import models def _DeleteUser(user_id): models.UserKey(user_id).delete() def _DeleteAll(keys, continuation_fn, user_id): if keys: ndb.delete_multi(keys) deferred.defer(continuation_fn, user_id) def _DeletePageRating(user_id): _DeleteAll( models.PageRating.query(ancestor=models.UserKey(user_id)).fetch( keys_only=True, limit=500), _DeletePageRating, user_id) def _DeleteConnectionPublisher(user_id): _DeleteAll( models.Connection.query(models.Connection.publisher_id == user_id).fetch( keys_only=True, limit=500), _DeleteConnectionPublisher, user_id) def _DeleteConnectionSubscriber(user_id): _DeleteAll( models.Connection.query(models.Connection.subscriber_id == user_id).fetch( keys_only=True, limit=500), _DeleteConnectionSubscriber, user_id) def _DeletePastRecommendation(user_id): _DeleteAll( models.PastRecommendation.query( models.PastRecommendation.user_id == user_id).fetch( keys_only=True, limit=500), _DeletePastRecommendation, user_id) def _DeleteRecommendationSession(user_id): _DeleteAll( models.RecommendationSession.query( models.RecommendationSession.user_id == user_id).fetch( keys_only=True, limit=500), _DeleteRecommendationSession, user_id) def _DeleteCategory(user_id): _DeleteAll( models.Category.query(ancestor=models.UserKey(user_id)).fetch( keys_only=True, limit=500), _DeleteCategory, user_id) def _DeleteCachedRatings(user_id): memcache.delete(models.GetUserRatedItemsCacheKey(user_id)) HANDLERS = [ _DeleteUser, _DeletePageRating, _DeleteConnectionPublisher, _DeleteConnectionSubscriber, _DeletePastRecommendation, _DeleteRecommendationSession, _DeleteCategory, _DeleteCachedRatings, ] def DeleteAccount(user_id): for handler in HANDLERS: deferred.defer(handler, user_id)
[ "recommender.models.GetUserRatedItemsCacheKey", "google.appengine.ext.deferred.defer", "recommender.models.UserKey", "recommender.models.PastRecommendation.query", "recommender.models.RecommendationSession.query", "google.appengine.ext.ndb.delete_multi", "recommender.models.Connection.query" ]
[((1330, 1352), 'google.appengine.ext.ndb.delete_multi', 'ndb.delete_multi', (['keys'], {}), '(keys)\n', (1346, 1352), False, 'from google.appengine.ext import ndb\n'), ((1357, 1397), 'google.appengine.ext.deferred.defer', 'deferred.defer', (['continuation_fn', 'user_id'], {}), '(continuation_fn, user_id)\n', (1371, 1397), False, 'from google.appengine.ext import deferred\n'), ((2728, 2769), 'recommender.models.GetUserRatedItemsCacheKey', 'models.GetUserRatedItemsCacheKey', (['user_id'], {}), '(user_id)\n', (2760, 2769), False, 'from recommender import models\n'), ((3066, 3098), 'google.appengine.ext.deferred.defer', 'deferred.defer', (['handler', 'user_id'], {}), '(handler, user_id)\n', (3080, 3098), False, 'from google.appengine.ext import deferred\n'), ((1232, 1255), 'recommender.models.UserKey', 'models.UserKey', (['user_id'], {}), '(user_id)\n', (1246, 1255), False, 'from recommender import models\n'), ((1646, 1712), 'recommender.models.Connection.query', 'models.Connection.query', (['(models.Connection.publisher_id == user_id)'], {}), '(models.Connection.publisher_id == user_id)\n', (1669, 1712), False, 'from recommender import models\n'), ((1859, 1926), 'recommender.models.Connection.query', 'models.Connection.query', (['(models.Connection.subscriber_id == user_id)'], {}), '(models.Connection.subscriber_id == user_id)\n', (1882, 1926), False, 'from recommender import models\n'), ((2072, 2149), 'recommender.models.PastRecommendation.query', 'models.PastRecommendation.query', (['(models.PastRecommendation.user_id == user_id)'], {}), '(models.PastRecommendation.user_id == user_id)\n', (2103, 2149), False, 'from recommender import models\n'), ((2311, 2398), 'recommender.models.RecommendationSession.query', 'models.RecommendationSession.query', (['(models.RecommendationSession.user_id == user_id)'], {}), '(models.RecommendationSession.user_id ==\n user_id)\n', (2345, 2398), False, 'from recommender import models\n'), ((1485, 1508), 'recommender.models.UserKey', 'models.UserKey', (['user_id'], {}), '(user_id)\n', (1499, 1508), False, 'from recommender import models\n'), ((2577, 2600), 'recommender.models.UserKey', 'models.UserKey', (['user_id'], {}), '(user_id)\n', (2591, 2600), False, 'from recommender import models\n')]
from os.path import abspath, dirname, join WORLDGEN_ROOT_PATH = abspath(join(dirname(__file__), '..', '..')) def worldgen_path(*args): """ Returns an absolute path from a path relative to the mujoco_worldgen repository root directory. """ return join(WORLDGEN_ROOT_PATH, *args)
[ "os.path.dirname", "os.path.join" ]
[((269, 300), 'os.path.join', 'join', (['WORLDGEN_ROOT_PATH', '*args'], {}), '(WORLDGEN_ROOT_PATH, *args)\n', (273, 300), False, 'from os.path import abspath, dirname, join\n'), ((78, 95), 'os.path.dirname', 'dirname', (['__file__'], {}), '(__file__)\n', (85, 95), False, 'from os.path import abspath, dirname, join\n')]
# encoding: utf-8 import os import re import sys import gzip import time import json import socket import random import weakref import datetime import functools import threading import collections import urllib.error import urllib.parse import urllib.request import collections.abc import json_dict from . import utils class ProxyURLRefreshError(Exception): pass class AliveProxiesNotFound(Exception): pass class NoFreeProxies(Exception): pass def _get_missing(target, source): """Возвращает присутствующие в `target`, но отсутствующие в `source` элементы """ old_target = set(target) new_target = old_target.intersection(source) return old_target.difference(new_target) def _build_opener(proxy=None): if proxy is not None: parsed = urllib.parse.urlparse(proxy) handler = urllib.request.ProxyHandler({parsed.scheme: proxy}) return urllib.request.build_opener(handler) else: return urllib.request.build_opener() class Proxies: default_opener = _build_opener() def __init__( self, proxies=None, proxies_url=None, proxies_url_gateway=None, proxies_file=None, options=None, ): """ @param proxies: список адресов прокси-серверов @param proxies_url: ссылка на список прокси-серверов @param proxies_file: путь до файла со списком прокси-серверов @param options: доп. параметры """ if options is None: options = {} shuffle = options.get('shuffle', False) if proxies is not None: proxies = list(proxies) if shuffle: random.shuffle(proxies) auto_refresh_period = options.get('auto_refresh_period') if auto_refresh_period: auto_refresh_period = datetime.timedelta(**auto_refresh_period) blacklist = utils.get_json_dict(json_dict.JsonLastUpdatedOrderedDict, filename=options.get('blacklist')) cooling_down = utils.get_json_dict(json_dict.JsonOrderedDict, filename=options.get('cooldown')) stats = utils.get_json_dict(json_dict.JsonDict, filename=options.get('stats')) if proxies_url_gateway: url_opener = _build_opener(proxies_url_gateway) else: url_opener = None self._url_opener = url_opener self._proxies = proxies self.proxies_url = proxies_url self.proxies_file = proxies_file self._shuffle = shuffle self.slice = options.get('slice') self.force_type = options.get('type') self.auto_refresh_period = auto_refresh_period self._blacklist = blacklist self._cooling_down = cooling_down self._stats = stats self._cleanup_lock = threading.RLock() self._last_auto_refresh = None self._auto_refresh_lock = threading.Lock() self._load_lock = threading.Lock() self._modified_at = time.perf_counter() self.__pool = None self._smart_holdout_start = options.get('smart_holdout_start') self._options = options if self._proxies is not None: proxies = set(self._proxies) self._cleanup_internals(proxies) @property def proxies(self): if self._proxies is None: with self._load_lock: # Вышли из состояния гонки, теперь можно удостовериться в реальной необходимости if self._proxies is None: self._proxies = self._load() self._cleanup_internals(self._proxies) self._modified_at = time.perf_counter() return self._proxies def _load(self): if self.proxies_url: proxies = self.read_url(self.proxies_url, opener=self._url_opener) elif self.proxies_file: proxies = self.read_file(self.proxies_file) else: raise NotImplementedError( "Can't load proxies: " "please specify one of the sources ('proxies_url' or 'proxies_file')" ) if self.slice: proxies = proxies[slice(*self.slice)] if self.force_type: new_type = self.force_type + '://' # `socks` format proxies = [ re.sub(r'^(?:(.*?)://)?', new_type, proxy) for proxy in proxies ] if self._shuffle: random.shuffle(proxies) return proxies def _cleanup_internals(self, proxies): with self._cleanup_lock: self._cleanup_blacklist(proxies) self._cleanup_cooling_down(proxies) self._cleanup_stats(proxies) def _cleanup_cooling_down(self, proxies): for proxy in _get_missing(self._cooling_down, proxies): self._cooling_down.pop(proxy) def _cleanup_blacklist(self, proxies): for proxy in _get_missing(self._blacklist, proxies): self._blacklist.pop(proxy) def _cleanup_stats(self, proxies): for proxy in _get_missing(self._stats, proxies): self._stats.pop(proxy) def _get_options(self, *options, missing_ok=True): if missing_ok: return {k: self._options.get(k) for k in options} else: return {k: self._options[k] for k in options} @classmethod def read_string(cls, string, sep=','): return list(x for x in map(str.strip, string.split(sep)) if x) @classmethod def read_url(cls, url, sep='\n', retry=10, sleep_range=(2, 10), timeout=2, opener=None): if opener is None: opener = cls.default_opener while True: try: resp = opener.open(url, timeout=timeout) break except (urllib.error.HTTPError, socket.timeout): if not retry: raise retry -= 1 time.sleep(random.randint(*sleep_range)) content = resp.read() if resp.headers.get('Content-Encoding', 'identity') == 'gzip': content = gzip.decompress(content) charset = resp.headers.get_content_charset('utf-8') content = content.decode(charset) return cls.read_string(content, sep=sep) @classmethod def read_file(cls, file_name, sep='\n'): with open(file_name) as f: return cls.read_string(f.read(), sep=sep) def refresh(self): if not self.proxies_url and not self.proxies_file: return try: self._proxies = self._load() self._cleanup_internals(self._proxies) except urllib.error.HTTPError: import problems problems.handle(ProxyURLRefreshError, extra={'url': self.proxies_url}) else: self._modified_at = time.perf_counter() def _auto_refresh(self): if self.proxies_file: with self._auto_refresh_lock: modification_time = datetime.datetime.fromtimestamp(os.stat(self.proxies_file).st_mtime) if modification_time == self._last_auto_refresh: return self.refresh() self._last_auto_refresh = modification_time elif self.proxies_url: if self.auto_refresh_period is None: return with self._auto_refresh_lock: now = datetime.datetime.now() if self._last_auto_refresh is not None: if now - self._last_auto_refresh < self.auto_refresh_period: return self.refresh() self._last_auto_refresh = now def get_random_address(self): self._auto_refresh() return random.choice(self.proxies) def get_pool(self): if self.__pool is None: with self._cleanup_lock: # оптимизация: используем уже существующий лок # Вышли из состояния гонки, теперь можно удостовериться в реальной необходимости if self.__pool is None: options = self._get_options('default_holdout', 'default_bad_holdout', 'force_defaults') if self._smart_holdout_start is not None: options['smart_holdout'] = True options['smart_holdout_start'] = self._smart_holdout_start options.update(self._get_options('smart_holdout_min', 'smart_holdout_max')) self.__pool = _Pool( self, self._cooling_down, self._blacklist, self._stats, self._cleanup_lock, **options ) return self.__pool @classmethod def from_cfg_string(cls, cfg_string): """Возвращает список прокси с тем исключением что список опций берется автоматически. Формат: json Доступные опции: type ('socks5', 'http'; для полного списка типов см. модуль socks): все прокси будут автоматически промаркированы этип типом slice (tuple c аргументами для builtins.slice): будет взят только указанный фрагмент списка прокси-серверов auto_refresh_period (dict): {'days': ..., 'hours': ..., 'minutes': ...} как часто необходимо обновлять список прокси-серверов (только для `url` и `file`) url_gateway: адрес proxy, через которые будет загружаться список прокси по url (url, file, list) - может быть именем файла, ссылкой или списком в формате json Параметры slice и force_type являются необязательными Примеры: option = {"list": ["127.0.0.1:3128"]} option = {"list": ["127.0.0.1:3128", "127.0.0.1:9999"]} option = {"file": "./my_new_proxies.txt", "type": "socks5"} option = {"url": "http://example.com/get/proxy_list/", "slice": [35, null], "type": "http"} option = {"url": "http://example.com/get/proxy_list/", "auto_refresh_period": {"days": 1}} option = {"url": "http://example.com/get/proxy_list/", "url_gateway": "http://proxy.example.com:9999"} """ cfg = json.loads(cfg_string) proxies = cfg.pop('list', None) proxies_url = cfg.pop('url', None) proxies_url_gateway = cfg.pop('url_gateway', None) proxies_file = cfg.pop('file', None) return cls( proxies=proxies, proxies_url=proxies_url, proxies_url_gateway=proxies_url_gateway, proxies_file=proxies_file, options=cfg ) class _Pool: def __init__( self, proxies: "`Proxies` instance", cooling_down, blacklist, stats, _cleanup_lock=None, smart_holdout=False, smart_holdout_start=None, smart_holdout_min=None, smart_holdout_max=None, default_holdout=None, default_bad_holdout=None, force_defaults=False, ): if smart_holdout: if smart_holdout_start in (None, 0): raise RuntimeError("Вы должны указать начальное время охлаждения") if smart_holdout_max is None: smart_holdout_max = float('inf') self._used = set() self._cond = threading.Condition(lock=_cleanup_lock) self._free = collections.deque( p for p in proxies.proxies if ( p not in blacklist and p not in cooling_down ) ) self._proxies = proxies self._cooling_down = cooling_down self._blacklist = blacklist self._stats = stats self._smart_holdout = smart_holdout self._smart_holdout_start = smart_holdout_start self._smart_holdout_min = smart_holdout_min or 0 self._smart_holdout_max = smart_holdout_max self._default_holdout = default_holdout self._default_bad_holdout = default_bad_holdout self._force_defaults = force_defaults self._proxies_modified_at = proxies._modified_at @property def _size(self): return len(self._free) + len(self._used) + len(self._cooling_down) + len(self._blacklist) def _cool_released(self): now = time.time() cooled = [] for proxy, holdout in self._cooling_down.items(): if now >= holdout: cooled.append(proxy) for proxy in cooled: self._cooling_down.pop(proxy, None) if proxy not in self._blacklist: self._free.append(proxy) def _is_proxies_changed(self): self._proxies._auto_refresh() return self._proxies._modified_at != self._proxies_modified_at def _remove_outdated(self): # список прокси изменился, оставляем только актуальные full_list = set(self._proxies.proxies) for proxy in _get_missing(self._blacklist, full_list): self._blacklist.pop(proxy, None) for proxy in _get_missing(self._cooling_down, full_list): self._cooling_down.pop(proxy, None) for proxy in _get_missing(self._used, full_list): self._used.remove(proxy) for proxy in _get_missing(self._stats, full_list): self._stats.pop(proxy, None) free = set( p for p in full_list if ( p not in self._used and p not in self._blacklist and p not in self._cooling_down ) ) old_free = set(self._free) new_free = old_free.intersection(free) if old_free.difference(new_free): self._free.clear() self._free.extend(new_free) self._proxies_modified_at = self._proxies._modified_at def _update_stats(self, proxy, bad=False, holdout=None): proxy_stat = self._stats.get(proxy) or {} ok, fail = proxy_stat.get('uptime', (0, 0)) if not bad: ok += 1 else: fail += 1 proxy_stat['uptime'] = ok, fail proxy_stat['last_holdout'] = holdout if ( not bad or ( holdout is not None and holdout >= (proxy_stat.get('last_good_holdout') or 0) ) ): proxy_stat['last_good_holdout'] = holdout # универсальный способ сказать что статистика обновилась # тк без вызова метода .save будет работать и с обычным словарем (не только с JsonDict) self._stats[proxy] = proxy_stat def _get_next_holdout(self, proxy, bad=False): """Рассчитывает время охлаждения. @param proxy: прокси, для которого необходимо вычислить @param bad: True - вычисляем охлаждение для неудачи, иначе False @return: рекомендуемое время охлаждения в секундах или None, если недостаточно данных """ # Алгоритм основан на бинарном поиске, # в отличии от которого нам не известна верхняя граница proxy_stat = self._stats.get(proxy) if proxy_stat is None: return None last_holdout = proxy_stat['last_holdout'] last_good_holdout = proxy_stat.get('last_good_holdout', 0) lo = last_holdout # предыдущее время охлаждения (нижняя граница) if bad: # Мы получили "бан" ... if lo < last_good_holdout: # ... возвращаемся к предыдущему хорошему значению ... holdout = last_good_holdout else: # ... или сдвигаем границу дальше holdout = lo * 2 else: # возвращаемся к предыдущей границе (lo / 2) # но с небольшим отступом - на середину отрезка [(lo / 2), lo] holdout = lo * 0.75 return holdout def acquire(self, timeout=None): start = time.perf_counter() with self._cond: while True: if self._is_proxies_changed(): self._remove_outdated() self._cool_released() if self._free: proxy = self._free.popleft() self._used.add(proxy) return proxy if self._blacklist: # Возвращаем самый стабильный из блеклиста. Возможно бан снят. def _uptime(p): uptime = float('inf') p_stat = self._stats.get(p) if p_stat is not None: ok, failed = p_stat.get('uptime', (0, 0)) if failed != 0: uptime = ok // failed else: uptime = ok return uptime proxy = next(( p for p in sorted(self._blacklist, key=_uptime, reverse=True) if p not in self._cooling_down ), None) if proxy is not None: self._blacklist.pop(proxy) self._used.add(proxy) return proxy else: # Все прокси из блеклиста находятся на охлаждении pass if self._cooling_down: self._cond.wait(1) else: self._cond.wait(timeout) if timeout is not None: if time.perf_counter() - start > timeout: raise NoFreeProxies def release(self, proxy, bad=False, holdout=None, bad_reason=None): """Возвращает прокси в пул @param proxy: прокси @param holdout (сек): None - вернуть сразу, иначе прокси не будет использован до истечения указанного интервала """ with self._cond: is_outdated = proxy not in self._used if is_outdated: # Скорее всего прокси уже не актуален # И был удален из списка return self._used.remove(proxy) if holdout is None or self._force_defaults: holdout = self._default_holdout if not bad else self._default_bad_holdout if self._smart_holdout: _holdout = ( self._get_next_holdout(proxy, bad=bad) or holdout or self._smart_holdout_start ) # Не позволяем границе опуститься слишком низко if _holdout < self._smart_holdout_min: holdout = self._smart_holdout_min elif _holdout > self._smart_holdout_max: holdout = self._smart_holdout_max else: holdout = max(self._smart_holdout_min, _holdout) if holdout is not None: self._cooling_down[proxy] = time.time() + holdout if bad: self._blacklist[proxy] = bad_reason elif holdout is None: # прокси не требует остывания self._free.append(proxy) self._cond.notify() self._update_stats(proxy, bad=bad, holdout=holdout) class IChain: def switch(self, bad=False, holdout=None, bad_reason=None, lazy=False): raise NotImplementedError def get_adapter(self): raise NotImplementedError def get_handler(self): raise NotImplementedError def get_path(self): raise NotImplementedError def wrap_session(self, session): raise NotImplementedError def wrap_module(self, module, all_threads=False): """ Attempts to replace a module's socket library with a SOCKS socket. This will only work on modules that import socket directly into the namespace; most of the Python Standard Library falls into this category. """ import socks import socks.monkey_socket routes = socks.RoutingTable.from_addresses(self.get_path()) if not all_threads: socks.monkey_socket.socks_wrap_module_thread(routes, module) else: socks.monkey_socket.socks_wrap_module_global(routes, module) class Chain(IChain): """ Не является потокобезопасным. """ def __init__(self, proxies, proxy_gw=None, use_pool=False, pool_acquire_timeout=None): """ @param proxies: список адресов прокси-серверов @param proxy_gw: прокси-сервер, который должен стоять во главе цепочки (все запросы к другим прокси-серверам будут проходить через него) @param use_pool: использовать список прокси в качестве пула @param pool_acquire_timeout (сек.): если за указанный период не удастся получить свободный прокси будет брошено исключение `NoFreeProxies`, None - ждать до появления свободного адреса """ if not isinstance(proxies, Proxies) and isinstance(proxies, collections.Sequence): proxies = Proxies(proxies) if use_pool: pool = proxies.get_pool() else: pool = None self.proxies = proxies self.proxy_gw = proxy_gw self._proxies_pool = pool self._current_pool_proxy = None self._pool_acquire_timeout = pool_acquire_timeout self.__path = [] # fix http://bugs.python.org/issue23841 if sys.version_info >= (3, 4, 0): self.finalizer = weakref.finalize(self, self.finalize) else: self.finalizer = None def __del__(self): if self.finalizer is None: self.finalize() def finalize(self): if self._proxies_pool is not None: self._release_pool_proxy() def _build_path(self, proxy): path = [] if self.proxy_gw: path.append(self.proxy_gw) path.append(proxy) return path def _release_pool_proxy(self, bad=False, holdout=None, bad_reason=None): if self._current_pool_proxy: proxy = self._current_pool_proxy self._current_pool_proxy = None self._proxies_pool.release(proxy, bad=bad, holdout=holdout, bad_reason=bad_reason) def _acquire_pool_proxy(self): proxy = self._proxies_pool.acquire(timeout=self._pool_acquire_timeout) self._current_pool_proxy = proxy return proxy def _get_proxy(self): if self._proxies_pool is not None: return self._acquire_pool_proxy() else: return self.proxies.get_random_address() @property def _path(self): if not self.__path: self.__path = self._build_path(self._get_proxy()) return self.__path def get_path(self): return self._path def switch(self, bad=False, holdout=None, bad_reason=None, lazy=False): self.__path.clear() if self._proxies_pool is not None: self._release_pool_proxy(bad, holdout, bad_reason) if not lazy: self.__path = self._build_path(self._get_proxy()) def get_adapter(self): import socks.adapters return socks.adapters.ChainedProxyHTTPAdapter(chain=self._path) def get_handler(self): import socks.handlers return socks.handlers.ChainProxyHandler(chain=self._path) def wrap_session(self, session): adapter = self.get_adapter() session.mount('http://', adapter) session.mount('https://', adapter) return session @classmethod def from_config(cls, cfg): proxy_cfg_string = cfg.get('Прокси') if proxy_cfg_string is None: return None proxy_gw = cfg.get('Шлюз') proxies = Proxies.from_cfg_string(proxy_cfg_string) return cls(proxies, proxy_gw=proxy_gw) class MultiChain(IChain): def __init__(self, *proxies_all, use_pool=True, pool_acquire_timeout=None): if use_pool: pool_kw = {'use_pool': True, 'pool_acquire_timeout': 1} else: pool_kw = {} self._pool_acquire_timeout = pool_acquire_timeout self._chains = collections.deque( Chain(p, gw, **pool_kw) for p, gw in self._unwrap_proxies_all(proxies_all) ) @staticmethod def _unwrap_proxies_all(proxies_all): for p in proxies_all: if isinstance(p, tuple): # (Proxies, Gateway) p, gw = p else: # Proxies p, gw = p, None yield p, gw def _self_auto_rotate(func): @functools.wraps(func) def wrapped(self, *args, **kw): start = time.perf_counter() while True: try: return func(self, *args, **kw) except NoFreeProxies: self._rotate() # FIXME: cycle rotate is normal? if ( self._pool_acquire_timeout is not None and time.perf_counter() - start > self._pool_acquire_timeout ): raise return wrapped @property def _current(self): return self._chains[-1] def get_path(self): return self._current.get_path() def _rotate(self): self._chains.rotate(1) def switch(self, bad=False, holdout=None, bad_reason=None, lazy=False): self._current.switch(bad=bad, holdout=holdout, bad_reason=bad_reason, lazy=True) self._rotate() if not lazy: self._enforce_current_path_build() @_self_auto_rotate def _enforce_current_path_build(self): _ = self._current._path # FIXME: ugly enforce path building after switching @_self_auto_rotate def get_adapter(self): return self._current.get_adapter() @_self_auto_rotate def get_handler(self): return self._current.get_handler() @_self_auto_rotate def wrap_session(self, session): return self._current.wrap_session(session) @_self_auto_rotate def wrap_module(self, module): return self._current.wrap_module(module)
[ "gzip.decompress", "datetime.timedelta", "weakref.finalize", "collections.deque", "threading.Lock", "threading.RLock", "time.perf_counter", "functools.wraps", "threading.Condition", "random.randint", "json.loads", "random.choice", "random.shuffle", "socks.handlers.ChainProxyHandler", "so...
[((2788, 2805), 'threading.RLock', 'threading.RLock', ([], {}), '()\n', (2803, 2805), False, 'import threading\n'), ((2880, 2896), 'threading.Lock', 'threading.Lock', ([], {}), '()\n', (2894, 2896), False, 'import threading\n'), ((2924, 2940), 'threading.Lock', 'threading.Lock', ([], {}), '()\n', (2938, 2940), False, 'import threading\n'), ((2969, 2988), 'time.perf_counter', 'time.perf_counter', ([], {}), '()\n', (2986, 2988), False, 'import time\n'), ((7759, 7786), 'random.choice', 'random.choice', (['self.proxies'], {}), '(self.proxies)\n', (7772, 7786), False, 'import random\n'), ((10186, 10208), 'json.loads', 'json.loads', (['cfg_string'], {}), '(cfg_string)\n', (10196, 10208), False, 'import json\n'), ((11239, 11278), 'threading.Condition', 'threading.Condition', ([], {'lock': '_cleanup_lock'}), '(lock=_cleanup_lock)\n', (11258, 11278), False, 'import threading\n'), ((11301, 11398), 'collections.deque', 'collections.deque', (['(p for p in proxies.proxies if p not in blacklist and p not in cooling_down)'], {}), '(p for p in proxies.proxies if p not in blacklist and p not in\n cooling_down)\n', (11318, 11398), False, 'import collections\n'), ((12214, 12225), 'time.time', 'time.time', ([], {}), '()\n', (12223, 12225), False, 'import time\n'), ((15809, 15828), 'time.perf_counter', 'time.perf_counter', ([], {}), '()\n', (15826, 15828), False, 'import time\n'), ((23158, 23214), 'socks.adapters.ChainedProxyHTTPAdapter', 'socks.adapters.ChainedProxyHTTPAdapter', ([], {'chain': 'self._path'}), '(chain=self._path)\n', (23196, 23214), False, 'import socks\n'), ((23288, 23338), 'socks.handlers.ChainProxyHandler', 'socks.handlers.ChainProxyHandler', ([], {'chain': 'self._path'}), '(chain=self._path)\n', (23320, 23338), False, 'import socks\n'), ((24602, 24623), 'functools.wraps', 'functools.wraps', (['func'], {}), '(func)\n', (24617, 24623), False, 'import functools\n'), ((1840, 1881), 'datetime.timedelta', 'datetime.timedelta', ([], {}), '(**auto_refresh_period)\n', (1858, 1881), False, 'import datetime\n'), ((4440, 4463), 'random.shuffle', 'random.shuffle', (['proxies'], {}), '(proxies)\n', (4454, 4463), False, 'import random\n'), ((6095, 6119), 'gzip.decompress', 'gzip.decompress', (['content'], {}), '(content)\n', (6110, 6119), False, 'import gzip\n'), ((6829, 6848), 'time.perf_counter', 'time.perf_counter', ([], {}), '()\n', (6846, 6848), False, 'import time\n'), ((20090, 20150), 'socks.monkey_socket.socks_wrap_module_thread', 'socks.monkey_socket.socks_wrap_module_thread', (['routes', 'module'], {}), '(routes, module)\n', (20134, 20150), False, 'import socks\n'), ((20177, 20237), 'socks.monkey_socket.socks_wrap_module_global', 'socks.monkey_socket.socks_wrap_module_global', (['routes', 'module'], {}), '(routes, module)\n', (20221, 20237), False, 'import socks\n'), ((21477, 21514), 'weakref.finalize', 'weakref.finalize', (['self', 'self.finalize'], {}), '(self, self.finalize)\n', (21493, 21514), False, 'import weakref\n'), ((24684, 24703), 'time.perf_counter', 'time.perf_counter', ([], {}), '()\n', (24701, 24703), False, 'import time\n'), ((1684, 1707), 'random.shuffle', 'random.shuffle', (['proxies'], {}), '(proxies)\n', (1698, 1707), False, 'import random\n'), ((4307, 4348), 're.sub', 're.sub', (['"""^(?:(.*?)://)?"""', 'new_type', 'proxy'], {}), "('^(?:(.*?)://)?', new_type, proxy)\n", (4313, 4348), False, 'import re\n'), ((6712, 6782), 'problems.handle', 'problems.handle', (['ProxyURLRefreshError'], {'extra': "{'url': self.proxies_url}"}), "(ProxyURLRefreshError, extra={'url': self.proxies_url})\n", (6727, 6782), False, 'import problems\n'), ((3639, 3658), 'time.perf_counter', 'time.perf_counter', ([], {}), '()\n', (3656, 3658), False, 'import time\n'), ((7409, 7432), 'datetime.datetime.now', 'datetime.datetime.now', ([], {}), '()\n', (7430, 7432), False, 'import datetime\n'), ((18915, 18926), 'time.time', 'time.time', ([], {}), '()\n', (18924, 18926), False, 'import time\n'), ((5940, 5968), 'random.randint', 'random.randint', (['*sleep_range'], {}), '(*sleep_range)\n', (5954, 5968), False, 'import random\n'), ((7019, 7045), 'os.stat', 'os.stat', (['self.proxies_file'], {}), '(self.proxies_file)\n', (7026, 7045), False, 'import os\n'), ((17475, 17494), 'time.perf_counter', 'time.perf_counter', ([], {}), '()\n', (17492, 17494), False, 'import time\n'), ((25023, 25042), 'time.perf_counter', 'time.perf_counter', ([], {}), '()\n', (25040, 25042), False, 'import time\n')]
from project.aquarium.freshwater_aquarium import FreshwaterAquarium # noqa from project.aquarium.saltwater_aquarium import SaltwaterAquarium # noqa from project.decoration.decoration_repository import DecorationRepository from project.decoration.ornament import Ornament # noqa from project.decoration.plant import Plant # noqa from project.fish.freshwater_fish import FreshwaterFish # noqa from project.fish.saltwater_fish import SaltwaterFish # noqa class Controller: def __init__(self): self.decorations_repository = DecorationRepository() self.aquariums = [] def add_aquarium(self, aquarium_type, aquarium_name): try: aquarium = globals()[aquarium_type] aquarium = aquarium(aquarium_name) except KeyError: return "Invalid aquarium type." self.aquariums.append(aquarium) return f"Successfully added {aquarium_type}." def add_decoration(self, decoration_type): try: decoration = globals()[decoration_type] decoration = decoration() except KeyError: return "Invalid decoration type." self.decorations_repository.add(decoration) return f"Successfully added {decoration_type}." def insert_decoration(self, aquarium_name, decoration_type): decoration = self.decorations_repository.find_by_type(decoration_type) if decoration == "None": return f"There isn't a decoration of type {decoration_type}." aquarium = self.__find_aquarium(aquarium_name) if aquarium: aquarium.add_decoration(decoration) self.decorations_repository.remove(decoration) return f"Successfully added {decoration_type} to {aquarium_name}." def add_fish(self, aquarium_name, fish_type, fish_name, fish_species, price): try: fish = globals()[fish_type] fish = fish(fish_name, fish_species, price) except KeyError: return f"There isn't a fish of type {fish_type}." aquarium = self.__find_aquarium(aquarium_name) result = aquarium.add_fish(fish) return result def feed_fish(self, aquarium_name): aquarium = self.__find_aquarium(aquarium_name) aquarium.feed() return f"Fish fed: {len(aquarium.fish)}" def calculate_value(self, aquarium_name): aquarium = self.__find_aquarium(aquarium_name) value = aquarium.total_fish_price + aquarium.total_decoration_price return f"The value of Aquarium {aquarium_name} is {value:.2f}." def report(self): return "\n".join(str(a) for a in self.aquariums) def __find_aquarium(self, aquarium_name): for aquarium in self.aquariums: if aquarium.name == aquarium_name: return aquarium
[ "project.decoration.decoration_repository.DecorationRepository" ]
[((540, 562), 'project.decoration.decoration_repository.DecorationRepository', 'DecorationRepository', ([], {}), '()\n', (560, 562), False, 'from project.decoration.decoration_repository import DecorationRepository\n')]
import json from Note import Note n = Note("Friction", "introduction to friction", "UC Berkeley", 0, True, False, False, "https://google.com", 'lec.pdf') print(n.toJSON())
[ "Note.Note" ]
[((39, 158), 'Note.Note', 'Note', (['"""Friction"""', '"""introduction to friction"""', '"""UC Berkeley"""', '(0)', '(True)', '(False)', '(False)', '"""https://google.com"""', '"""lec.pdf"""'], {}), "('Friction', 'introduction to friction', 'UC Berkeley', 0, True, False,\n False, 'https://google.com', 'lec.pdf')\n", (43, 158), False, 'from Note import Note\n')]
# Reference/tutorial to take a look at # https://help.autodesk.com/view/MOBPRO/2019/ENU/?guid=__py_ref__tasks_2_assign_rigid_body_8py_example_html # https://help.autodesk.com/view/MOBPRO/2019/ENU/?guid=__files_GUID_A1189AA0_3816_4350_B8F3_5383DEC25A33_htm # https://mocappys.com/complete-guide-to-poses-in-motionbuilder/#07_Using_the_Pose_Controls # https://github.com/eksod/Retargeter from pyfbsdk import * import os import re import random def addJoint(jointMap, newJoint, parent, localTrans, adjChildList=[], adjTrans=False, jointOrder=list()): jointMap[newJoint] = (parent, localTrans) for k in jointMap: v = jointMap[k] if v[0] == parent and k!=newJoint: if k in adjChildList: if adjTrans: adjTrans = [v[1][0]- localTrans[0], v[1][1]- localTrans[1], v[1][2]- localTrans[2]] jointMap[k] = (newJoint, adjTrans) else: jointMap[k] = (newJoint, v[1]) if parent in jointOrder: n = jointOrder.index(parent) jointOrder.insert(n+1, newJoint) else: jointOrder.insert(0, newJoint) # Create a skeleton in a T-pose facing along the positive Z axis def createSkeleton(pNamespace): # reference jointMap is copied from mixamo character tpose. jointMap = { #jointName, (parentName, translation ) #'Reference': (None, ( 0, 0, 0)), 'Hips': ('None', ( 0, 0, 0)), 'LeftUpLeg': ('Hips', (8.20778, -6.75171, -1.59956)), 'LeftLeg': ('LeftUpLeg', (0, -44.3705, 0.284643)), 'LeftFoot': ('LeftLeg', (0, -44.4279, -2.98219)), 'LeftToeBase': ('LeftFoot', (0, -8.72867, 10.7106)), 'RightUpLeg': ('Hips', (-8.2078, -6.75166, -1.59956)), 'RightLeg': ('RightUpLeg', (0, -44.3705, 0.286156)), 'RightFoot': ('RightLeg', (0, -44.4277, -2.98379)), 'RightToeBase': ('RightFoot', (0, -8.72867, 10.7106)), 'Spine': ('Hips', (0. , 29.36511 , -2.242699)), 'LeftArm': ('Spine', (15.16277, 10.421402, -4.86012)), 'LeftForeArm': ('LeftArm', (27.8415, -8.94286e-05, 3.74589e-05)), 'LeftHand': ('LeftForeArm', (28.3288, -1.74407e-05, 3.78045e-05)), 'RightArm': ('Spine', (-15.16277, 10.421402, -4.86012)), 'RightForeArm': ('RightArm', (-27.8415, -3.30792e-05, 1.16763e-05)), 'RightHand': ('RightForeArm', (-28.3288, 0, 5.5816e-05)), 'Neck': ('Spine', (0, 16.6717, -2.51617)), 'Head': ('Neck', (0, 9.61788, 1.68501)), } skeleton = {} headOrder, legOrder, armOrder= ['Hips', 'Spine', 'Neck', 'Head'], ['LeftUpLeg', 'LeftLeg', 'LeftFoot', 'LeftToeBase'], ['LeftArm', 'LeftForeArm', 'LeftHand'] if random.random() < 0.3: addJoint(jointMap, "LeftHipJoint", "Hips", (0,0,0), ["LeftUpLeg"], False, legOrder) addJoint(jointMap, "RightHipJoint", "Hips", (0,0,0), ["RightUpLeg"], False) if random.random() < 0.3: addJoint(jointMap, "LowerBack", "Hips", (0,0,0), ["Spine"], False, headOrder) spineCntRange = range(1, 7) weights=[1.5+1/x for x in spineCntRange] sum_weights= sum(weights) weights =[x/sum_weights for x in weights] spineRnd = random.random() sum_w = 0 spineCnt = -1 for i, w in enumerate(weights): sum_w += w if spineRnd < sum_w: spineCnt = i+1 break if spineCnt == -1: spineCnt = 6 lastSpine = 'Spine' if spineCnt > 1: origTrans = jointMap['Neck'][1] unitTrans = [origTrans[0]/spineCnt, origTrans[1]/spineCnt, origTrans[2]/spineCnt] jointMap['Spine'] = (jointMap['Spine'][0], unitTrans) for i in range(1, spineCnt): addJoint(jointMap, 'Spine'+str(i), lastSpine, unitTrans, ['LeftArm', 'RightArm', 'Neck'], False, headOrder) lastSpine= 'Spine'+str(i) if random.random() < 0.5: addJoint(jointMap, 'LeftShoulder', lastSpine, (4.57045, 10.946, -2.62799), ['LeftArm'], True, armOrder) addJoint(jointMap, 'RightShoulder', lastSpine, (-4.56997, 10.9462, -2.62802), ['RightArm'], True) elif random.random() < 0.3 : addJoint(jointMap, 'LeftShoulder', lastSpine, (0, 0, 0), ['LeftArm'], True, armOrder) addJoint(jointMap, 'RightShoulder', lastSpine, (0, 0, 0), ['RightArm'], True) if random.random() < 0.5: neckHeadCnt = 2 origTrans = jointMap['Head'][1] unitTrans = [origTrans[0]/neckHeadCnt, origTrans[1]/neckHeadCnt, origTrans[2]/neckHeadCnt] addJoint(jointMap, 'Neck1', 'Neck', unitTrans, ['Head'], True, headOrder) #End Sites #Head if random.random() < 0.5: headTrans = (0, 19, 1) addJoint(jointMap, 'Head_End', 'Head', headTrans, [], False, headOrder) # Leg if random.random() < 1: toeTrans = (0, 0, 6) addJoint(jointMap, 'LeftToeBase_End', 'LeftToeBase', toeTrans, [], False, legOrder) addJoint(jointMap, 'RightToeBase_End', 'RightToeBase', toeTrans, [], False) # Arm if random.random() < 1: addJoint(jointMap, 'LeftHand_End', 'LeftHand', (6, 0, 0), [], False, armOrder) addJoint(jointMap, 'RightHand_End', 'RightHand', (-6, 0, 0), [], False) scale = 0.6+ 0.8*random.random() link_scale = {} # Populate the skeleton with joints. other_root = findRoot() for jointName, (parentName, translation) in jointMap.iteritems(): jointName_ = jointName if not skelExists(other_root, jointName) else jointName+"_" if jointName == 'Reference' or jointName == 'Hips': # If it is the reference node, create an FBModelRoot. joint = FBModelRoot(jointName_) else: # Otherwise, create an FBModelSkeleton. joint = FBModelSkeleton(jointName_) joint.LongName = pNamespace + ':' + joint.Name # Apply the specified namespace to each joint. joint.Color = FBColor(0.3, 0.8, 1) # Cyan joint.Size = 150 # Arbitrary size: big enough to see in viewport joint.Show = True # Make the joint visible in the scene. # Add the joint to our skeleton. skeleton[jointName] = joint if "Right" not in jointName: ls = [0.6+ 0.8*random.random(), 0.6+ 0.8*random.random(), 0.6+ 0.8*random.random()] jointMap[jointName] = (parentName, (translation[0]*ls[0], translation[1]*ls[1], translation[2]*ls[2])) #print(jointName, ls) if "Left" in jointName: cor_right = "Right"+jointName[4:] r_parentName, r_translation = jointMap[cor_right] jointMap[cor_right] = (r_parentName, (r_translation[0]*ls[0], r_translation[1]*ls[1], r_translation[2]*ls[2])) # move Hips so that foot touches the ground (2cm above considering foot depth(?)) joint_lower= 'LeftToeBase' lower_sum= 0 footOffset = 2 while joint_lower != 'Hips': translation = jointMap[joint_lower][1] lower_sum= lower_sum +translation[1] joint_lower= jointMap[joint_lower][0] hips_trans = jointMap['Hips'][1] jointMap['Hips'] = ('None', (hips_trans[0], -lower_sum+footOffset, hips_trans[2])) # Once all the joints have been created, apply the parent/child # relationships to each of the skeleton's joints. jointOrder= headOrder+legOrder+armOrder def connectPlaceJoint(jointName): (parentName, translation)= jointMap[jointName] # Only assign a parent if it exists. if parentName != None and parentName in jointMap.keys(): skeleton[jointName].Parent = skeleton[parentName] # The translation should be set after the parent has been assigned. skeleton[jointName].Translation = FBVector3d(translation)*scale for jointName in reversed(jointOrder): connectPlaceJoint(jointName) if "Left" in jointName: cor_right = "Right"+jointName[4:] connectPlaceJoint(cor_right) return skeleton # Characterize the skeleton and create a control rig. def characterizeSkeleton(pCharacterName, pSkeleton, ctrlRig=False): # Create a new character. character = FBCharacter(pCharacterName) FBApplication().CurrentCharacter = character # Add each joint in our skeleton to the character. for jointName, joint in pSkeleton.iteritems(): slot = character.PropertyList.Find(jointName + 'Link') # skip dummy nodes (e.g.: LeftHipJoint) if slot is not None: slot.append(joint) # Flag that the character has been characterized. character.SetCharacterizeOn(True) if ctrlRig: # Create a control rig using Forward and Inverse Kinematics, # as specified by the "True" parameter. character.CreateControlRig(True) # Set the control rig to active. character.ActiveInput = True return character # This is the Motionbuilder mapping to use the same function. Edit this list or create your own. mobuMap = {'Reference' : 'reference', 'Hips':'Hips', 'LeftUpLeg' : 'LeftUpLeg', 'LeftLeg' : 'LeftLeg', 'LeftFoot' : 'LeftFoot', 'LeftToeBase': 'LeftToeBase', 'RightUpLeg' : 'RightUpLeg', 'RightLeg' : 'RightLeg', 'RightFoot' : 'RightFoot', 'RightToeBase': 'RightToeBase', 'Spine' : 'Spine', 'LeftArm' : 'LeftArm', 'LeftForeArm' : 'LeftForeArm', 'LeftHand' : 'LeftHand', 'RightArm' : 'RightArm', 'RightForeArm' : 'RightForeArm', 'RightHand' : 'RightHand', 'Head' : 'Head', 'LeftShoulder' : 'LeftShoulder', 'RightShoulder' : 'RightShoulder', 'Neck' : 'Neck', 'Spine1' : 'Spine1', 'Spine2' : 'Spine2', 'Spine3' : 'Spine3', 'Spine4' : 'Spine4', 'Spine5' : 'Spine5', 'Spine6' : 'Spine6', 'Spine7' : 'Spine7', 'Spine8' : 'Spine8', 'Spine9' : 'Spine9', 'Neck1' : 'Neck1', 'Neck2' : 'Neck2', 'Neck3' : 'Neck3', 'Neck4' : 'Neck4', 'Neck5' : 'Neck5', 'Neck6' : 'Neck6', 'Neck7' : 'Neck7', 'Neck8' : 'Neck8', 'Neck9' : 'Neck9', 'LeftHandThumb1' : 'LeftHandThumb1', 'LeftHandThumb2' : 'LeftHandThumb2', 'LeftHandThumb3' : 'LeftHandThumb3', 'LeftHandIndex1' : 'LeftHandIndex1', 'LeftHandIndex2' : 'LeftHandIndex2', 'LeftHandIndex3' : 'LeftHandIndex3', 'LeftHandMiddle1' : 'LeftHandMiddle1', 'LeftHandMiddle2' : 'LeftHandMiddle2', 'LeftHandMiddle3' : 'LeftHandMiddle3', 'LeftHandRing1' : 'LeftHandRing1', 'LeftHandRing2' : 'LeftHandRing2', 'LeftHandRing3' : 'LeftHandRing3', 'LeftHandPinky1' : 'LeftHandPinky1', 'LeftHandPinky2' : 'LeftHandPinky2', 'LeftHandPinky3' : 'LeftHandPinky3', 'RightHandThumb1' : 'RightHandThumb1', 'RightHandThumb2' : 'RightHandThumb2', 'RightHandThumb3' : 'RightHandThumb3', 'RightHandIndex1' : 'RightHandIndex1', 'RightHandIndex2' : 'RightHandIndex2', 'RightHandIndex3' : 'RightHandIndex3', 'RightHandMiddle1' : 'RightHandMiddle1', 'RightHandMiddle2' : 'RightHandMiddle2', 'RightHandMiddle3' : 'RightHandMiddle3', 'RightHandRing1' : 'RightHandRing1', 'RightHandRing2' : 'RightHandRing2', 'RightHandRing3' : 'RightHandRing3', 'RightHandPinky1' : 'RightHandPinky1', 'RightHandPinky2' : 'RightHandPinky2', 'RightHandPinky3' : 'RightHandPinky3', 'LeftFootThumb1' : 'LeftFootThumb1', 'LeftFootThumb2' : 'LeftFootThumb2', 'LeftFootThumb3' : 'LeftFootThumb3', 'LeftFootIndex1' : 'LeftFootIndex1', 'LeftFootIndex2' : 'LeftFootIndex2', 'LeftFootIndex3' : 'LeftFootIndex3', 'LeftFootMiddle1' : 'LeftFootMiddle1', 'LeftFootMiddle2' : 'LeftFootMiddle2', 'LeftFootMiddle3' : 'LeftFootMiddle3', 'LeftFootRing1' : 'LeftFootRing1', 'LeftFootRing2' : 'LeftFootRing2', 'LeftFootRing3' : 'LeftFootRing3', 'LeftFootPinky1' : 'LeftFootPinky1', 'LeftFootPinky2' : 'LeftFootPinky2', 'LeftFootPinky3' : 'LeftFootPinky3', 'RightFootThumb1' : 'RightFootThumb1', 'RightFootThumb2' : 'RightFootThumb2', 'RightFootThumb3' : 'RightFootThumb3', 'RightFootIndex1' : 'RightFootIndex1', 'RightFootIndex2' : 'RightFootIndex2', 'RightFootIndex3' : 'RightFootIndex3', 'RightFootMiddle1' : 'RightFootMiddle1', 'RightFootMiddle2' : 'RightFootMiddle2', 'RightFootMiddle3' : 'RightFootMiddle3', 'RightFootRing1' : 'RightFootRing1', 'RightFootRing2' : 'RightFootRing2', 'RightFootRing3' : 'RightFootRing3', 'RightFootPinky1' : 'RightFootPinky1', 'RightFootPinky2' : 'RightFootPinky2', 'RightFootPinky3' : 'RightFootPinky3', 'LeftUpLegRoll' : 'LeftUpLegRoll', 'LeftLegRoll' : 'LeftLegRoll', 'RightUpLegRoll' : 'RightUpLegRoll', 'RightLegRoll' : 'RightLegRoll', 'LeftArmRoll' : 'LeftArmRoll', 'LeftForeArmRoll' : 'LeftForeArmRoll', 'RightArmRoll' : 'RightArmRoll', 'RightForeArmRoll' : 'RightForeArmRoll' } def addJointToCharacter ( characterObject, slot, jointName ): myJoint = FBFindModelByLabelName(jointName) if myJoint: proplist = characterObject.PropertyList.Find(slot + "Link") proplist.append (myJoint) def CharacterizeBiped(rootname, useBipedPrefixNamingScheme, nameprefix, boneMap, models): system = FBSystem() app = FBApplication() longname = models.LongName namespaceindex = longname.rfind(":") if namespaceindex != -1: namespace = longname[0:namespaceindex+1] name = longname[namespaceindex + 1:] else: namespace = "" name = longname myBiped = FBCharacter("mycharacter") app.CurrentCharacter = myBiped # If in Biped mode, extract the character prefix name if useBipedPrefixNamingScheme: splitname = name.split() nameprefix = splitname[0] + " " # Override the rootname so it is the character orefix name rootname = splitname[0] myBiped.LongName = namespace + rootname else: myBiped.LongName = namespace + nameprefix + rootname # Create a FBProgress object and set default values for the caption and text. fbp = FBProgress() fbp.Caption = "" fbp.Text = " ----------------------------------- Creating Biped character" progress = 0.0 progresssteps = len(boneMap) # assign Biped to Character Mapping. for pslot, pjointName in boneMap.iteritems(): if not pjointName: addJointToCharacter(myBiped, pslot, namespace + rootname) else: addJointToCharacter(myBiped, pslot, namespace + nameprefix + pjointName) progress += 1 val = progress / len(boneMap) * 100 fbp.Percent = int(val) switchOn = myBiped.SetCharacterizeOn( True ) # print "Character mapping created for " + (myBiped.LongName) # We must call FBDelete when the FBProgress object is no longer needed. fbp.FBDelete() return myBiped def createControlRig(character, activeInput=True): # Create a control rig using Forward and Inverse Kinematics, # as specified by the "True" parameter. character.CreateControlRig(True) # Set the control rig to active. if activeInput: character.ActiveInput = True def plotAnim(char, animChar): """ Receives two characters, sets the input of the first character to the second and plot. Return ploted character. """ #if char.GetCharacterize: # switchOn = char.SetCharacterizeOn(True) plotoBla = FBPlotOptions() plotoBla.ConstantKeyReducerKeepOneKey = True plotoBla.PlotAllTakes = True plotoBla.PlotOnFrame = True plotoBla.PlotPeriod = FBTime( 0, 0, 0, 1 ) plotoBla.PlotTranslationOnRootOnly = True plotoBla.PreciseTimeDiscontinuities = True #plotoBla.RotationFilterToApply = FBRotationFilter.kFBRotationFilterGimbleKiller plotoBla.UseConstantKeyReducer = False plotoBla.ConstantKeyReducerKeepOneKey = True char.InputCharacter = animChar char.InputType = FBCharacterInputType.kFBCharacterInputCharacter char.ActiveInput = True if (not char.PlotAnimation(FBCharacterPlotWhere.kFBCharacterPlotOnSkeleton, plotoBla)): FBMessageBox( "Something went wrong", "Plot animation returned false, cannot continue", "OK", None, None ) return False return char def SwitchTake( pTakeName ): iDestName= pTakeName for iTake in FBSystem().Scene.Takes: if iTake.Name == iDestName: FBSystem().CurrentTake = iTake def findRoot(): for child in FBSystem().Scene.RootModel.Children: if "Hips" in child.Name: return child return None def skelExists(root, name): if root == None: return False if root.Name == name: return True for child in root.Children: if skelExists(child, name): return True return False ############################################################### # Main. # ############################################################### rootDir = "C:/Users/sunmin/Documents/GitHub/MBScript/" bvhFileDir = os.path.join(rootDir, "bvh/") dirList = [os.path.join(bvhFileDir, subDir) for subDir in os.listdir(bvhFileDir) if os.path.isdir(os.path.join(bvhFileDir, subDir))] fileList = [] ## filter short clips for subDir in dirList: allList = [os.path.join(subDir, f) for f in os.listdir(subDir) if f.endswith('.bvh')] fileList += allList #dirFileList.append(allList) system = FBSystem() app = FBApplication() newFileDir = os.path.join(rootDir, "result/") if not os.path.exists(newFileDir): os.makedirs(newFileDir) logPath = os.path.join(newFileDir, "result_log.txt") logFile = open(logPath, 'w') for idx, filePath in enumerate(fileList): print(idx, filePath) fileName = filePath.split(os.sep)[-1] fileType = filePath.split(os.sep)[-2] logStr = str(idx)+','+fileType+','+fileName+"\n" logFile.write(logStr) userRoot = "Hips" boneMap = mobuMap bipedPrefixNamingScheme = False prefix = "" app.FileNew() success = app.FileImport(filePath, True) poseOptions = FBCharacterPoseOptions() poseOptions.mCharacterPoseKeyingMode = FBCharacterPoseKeyingMode.kFBCharacterPoseKeyingModeFullBody animRoot = findRoot() animChar = CharacterizeBiped(userRoot, bipedPrefixNamingScheme, prefix, boneMap, animRoot) animChar.SelectModels(True, True, True, False) lPlayer = FBPlayerControl() lPlayer.Goto(FBTime(0, 0, 0, 0)) createControlRig(animChar, False) # first key all frame for animation bvh to prevent unwanted interpolation between frames lEndTime = system.CurrentTake.LocalTimeSpan.GetStop() lEndFrame = system.CurrentTake.LocalTimeSpan.GetStop().GetFrame() lStartFrameTime = system.CurrentTake.LocalTimeSpan.GetStart() lStartFrame = system.CurrentTake.LocalTimeSpan.GetStart().GetFrame() lRange = min(int(lEndFrame)+1, 50) lPlayer = FBPlayerControl() for i in range(lRange): lPlayer.Goto(FBTime(0, 0, 0, i)) system.Scene.Evaluate() lPlayer.Key() FBSystem().Scene.Evaluate() # Create a new skeleton. characterName = 'SkelTestCharacter' skeleton = createSkeleton(characterName) character = characterizeSkeleton(characterName, skeleton, False) plotAnim(character, animChar) animChar.SelectModels(False, True, True, True) character.SelectModels(True, True, True, True) # setup save options (for some reason, they were not working outside this loop...) sOptions = FBFbxOptions(False) # false = save options sOptions.SaveCharacter = True sOptions.SaveControlSet = False sOptions.SaveCharacterExtension = False sOptions.ShowFileDialog = False sOptions.ShowOptionsDialog = False fileName = str(idx)+'.bvh' newPath = os.path.join(newFileDir, fileName) FBApplication().FileExport(newPath) logFile.close()
[ "os.path.exists", "os.listdir", "os.makedirs", "os.path.join", "random.random" ]
[((18487, 18516), 'os.path.join', 'os.path.join', (['rootDir', '"""bvh/"""'], {}), "(rootDir, 'bvh/')\n", (18499, 18516), False, 'import os\n'), ((18919, 18951), 'os.path.join', 'os.path.join', (['rootDir', '"""result/"""'], {}), "(rootDir, 'result/')\n", (18931, 18951), False, 'import os\n'), ((19026, 19068), 'os.path.join', 'os.path.join', (['newFileDir', '"""result_log.txt"""'], {}), "(newFileDir, 'result_log.txt')\n", (19038, 19068), False, 'import os\n'), ((3424, 3439), 'random.random', 'random.random', ([], {}), '()\n', (3437, 3439), False, 'import random\n'), ((18528, 18560), 'os.path.join', 'os.path.join', (['bvhFileDir', 'subDir'], {}), '(bvhFileDir, subDir)\n', (18540, 18560), False, 'import os\n'), ((18959, 18985), 'os.path.exists', 'os.path.exists', (['newFileDir'], {}), '(newFileDir)\n', (18973, 18985), False, 'import os\n'), ((18991, 19014), 'os.makedirs', 'os.makedirs', (['newFileDir'], {}), '(newFileDir)\n', (19002, 19014), False, 'import os\n'), ((21226, 21260), 'os.path.join', 'os.path.join', (['newFileDir', 'fileName'], {}), '(newFileDir, fileName)\n', (21238, 21260), False, 'import os\n'), ((2935, 2950), 'random.random', 'random.random', ([], {}), '()\n', (2948, 2950), False, 'import random\n'), ((3141, 3156), 'random.random', 'random.random', ([], {}), '()\n', (3154, 3156), False, 'import random\n'), ((4083, 4098), 'random.random', 'random.random', ([], {}), '()\n', (4096, 4098), False, 'import random\n'), ((4564, 4579), 'random.random', 'random.random', ([], {}), '()\n', (4577, 4579), False, 'import random\n'), ((4888, 4903), 'random.random', 'random.random', ([], {}), '()\n', (4901, 4903), False, 'import random\n'), ((5040, 5055), 'random.random', 'random.random', ([], {}), '()\n', (5053, 5055), False, 'import random\n'), ((5284, 5299), 'random.random', 'random.random', ([], {}), '()\n', (5297, 5299), False, 'import random\n'), ((18575, 18597), 'os.listdir', 'os.listdir', (['bvhFileDir'], {}), '(bvhFileDir)\n', (18585, 18597), False, 'import os\n'), ((18725, 18748), 'os.path.join', 'os.path.join', (['subDir', 'f'], {}), '(subDir, f)\n', (18737, 18748), False, 'import os\n'), ((4343, 4358), 'random.random', 'random.random', ([], {}), '()\n', (4356, 4358), False, 'import random\n'), ((5497, 5512), 'random.random', 'random.random', ([], {}), '()\n', (5510, 5512), False, 'import random\n'), ((18615, 18647), 'os.path.join', 'os.path.join', (['bvhFileDir', 'subDir'], {}), '(bvhFileDir, subDir)\n', (18627, 18647), False, 'import os\n'), ((18758, 18776), 'os.listdir', 'os.listdir', (['subDir'], {}), '(subDir)\n', (18768, 18776), False, 'import os\n'), ((6599, 6614), 'random.random', 'random.random', ([], {}), '()\n', (6612, 6614), False, 'import random\n'), ((6625, 6640), 'random.random', 'random.random', ([], {}), '()\n', (6638, 6640), False, 'import random\n'), ((6651, 6666), 'random.random', 'random.random', ([], {}), '()\n', (6664, 6666), False, 'import random\n')]
import logging import pytest from ophyd.sim import make_fake_device from pcdsdevices.gon import (BaseGon, Goniometer, GonWithDetArm, Kappa, SamPhi, XYZStage) logger = logging.getLogger(__name__) def test_gon_factory(): logger.debug('test_gon_factory') assert isinstance(Goniometer(name='gon', prefix_hor='a', prefix_ver='b', prefix_rot='c', prefix_tip='d', prefix_tilt='e'), BaseGon) assert isinstance(Goniometer(name='gon', prefix_hor='a', prefix_ver='b', prefix_rot='c', prefix_tip='d', prefix_tilt='e', prefix_detver='i', prefix_dettilt='j', prefix_2theta='k'), GonWithDetArm) @pytest.mark.timeout(5) def test_gon_init(): logger.debug('test_gon_init') FakeGon = make_fake_device(BaseGon) FakeGon(name='test', prefix_hor='hor', prefix_ver='ver', prefix_rot='rot', prefix_tip='tip', prefix_tilt='tilt') FakeGon = make_fake_device(GonWithDetArm) FakeGon(name='test', prefix_hor='hor', prefix_ver='ver', prefix_rot='rot', prefix_tip='tip', prefix_tilt='tilt', prefix_detver='detver', prefix_dettilt='dettilt', prefix_2theta='2theta') FakeGon = make_fake_device(XYZStage) FakeGon(name='test', prefix_x='x', prefix_y='y', prefix_z='z') FakeGon = make_fake_device(SamPhi) FakeGon(name='test', prefix_samz='samz', prefix_samphi='samphi') FakeGon = make_fake_device(Kappa) FakeGon(name='test', prefix_x='x', prefix_y='y', prefix_z='z', prefix_eta='eta', prefix_kappa='kappa', prefix_phi='phi') @pytest.mark.timeout(5) def test_gon_disconnected(): logger.debug('test_gon_disconnected') BaseGon(name='test1', prefix_hor='hor', prefix_ver='ver', prefix_rot='rot', prefix_tip='tip', prefix_tilt='tilt') GonWithDetArm(name='test2', prefix_hor='hor', prefix_ver='ver', prefix_rot='rot', prefix_tip='tip', prefix_tilt='tilt', prefix_detver='detver', prefix_dettilt='dettilt', prefix_2theta='2theta') XYZStage(name='test3', prefix_x='x', prefix_y='y', prefix_z='z') SamPhi(name='test4', prefix_samz='samz', prefix_samphi='samphi') Kappa(name='test5', prefix_x='x', prefix_y='y', prefix_z='z', prefix_eta='eta', prefix_kappa='kappa', prefix_phi='phi')
[ "logging.getLogger", "pcdsdevices.gon.BaseGon", "pcdsdevices.gon.SamPhi", "pcdsdevices.gon.Kappa", "pcdsdevices.gon.XYZStage", "pcdsdevices.gon.Goniometer", "pytest.mark.timeout", "pcdsdevices.gon.GonWithDetArm", "ophyd.sim.make_fake_device" ]
[((199, 226), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (216, 226), False, 'import logging\n'), ((816, 838), 'pytest.mark.timeout', 'pytest.mark.timeout', (['(5)'], {}), '(5)\n', (835, 838), False, 'import pytest\n'), ((1730, 1752), 'pytest.mark.timeout', 'pytest.mark.timeout', (['(5)'], {}), '(5)\n', (1749, 1752), False, 'import pytest\n'), ((908, 933), 'ophyd.sim.make_fake_device', 'make_fake_device', (['BaseGon'], {}), '(BaseGon)\n', (924, 933), False, 'from ophyd.sim import make_fake_device\n'), ((1077, 1108), 'ophyd.sim.make_fake_device', 'make_fake_device', (['GonWithDetArm'], {}), '(GonWithDetArm)\n', (1093, 1108), False, 'from ophyd.sim import make_fake_device\n'), ((1350, 1376), 'ophyd.sim.make_fake_device', 'make_fake_device', (['XYZStage'], {}), '(XYZStage)\n', (1366, 1376), False, 'from ophyd.sim import make_fake_device\n'), ((1458, 1482), 'ophyd.sim.make_fake_device', 'make_fake_device', (['SamPhi'], {}), '(SamPhi)\n', (1474, 1482), False, 'from ophyd.sim import make_fake_device\n'), ((1566, 1589), 'ophyd.sim.make_fake_device', 'make_fake_device', (['Kappa'], {}), '(Kappa)\n', (1582, 1589), False, 'from ophyd.sim import make_fake_device\n'), ((1828, 1945), 'pcdsdevices.gon.BaseGon', 'BaseGon', ([], {'name': '"""test1"""', 'prefix_hor': '"""hor"""', 'prefix_ver': '"""ver"""', 'prefix_rot': '"""rot"""', 'prefix_tip': '"""tip"""', 'prefix_tilt': '"""tilt"""'}), "(name='test1', prefix_hor='hor', prefix_ver='ver', prefix_rot='rot',\n prefix_tip='tip', prefix_tilt='tilt')\n", (1835, 1945), False, 'from pcdsdevices.gon import BaseGon, Goniometer, GonWithDetArm, Kappa, SamPhi, XYZStage\n'), ((1958, 2160), 'pcdsdevices.gon.GonWithDetArm', 'GonWithDetArm', ([], {'name': '"""test2"""', 'prefix_hor': '"""hor"""', 'prefix_ver': '"""ver"""', 'prefix_rot': '"""rot"""', 'prefix_tip': '"""tip"""', 'prefix_tilt': '"""tilt"""', 'prefix_detver': '"""detver"""', 'prefix_dettilt': '"""dettilt"""', 'prefix_2theta': '"""2theta"""'}), "(name='test2', prefix_hor='hor', prefix_ver='ver', prefix_rot=\n 'rot', prefix_tip='tip', prefix_tilt='tilt', prefix_detver='detver',\n prefix_dettilt='dettilt', prefix_2theta='2theta')\n", (1971, 2160), False, 'from pcdsdevices.gon import BaseGon, Goniometer, GonWithDetArm, Kappa, SamPhi, XYZStage\n'), ((2210, 2274), 'pcdsdevices.gon.XYZStage', 'XYZStage', ([], {'name': '"""test3"""', 'prefix_x': '"""x"""', 'prefix_y': '"""y"""', 'prefix_z': '"""z"""'}), "(name='test3', prefix_x='x', prefix_y='y', prefix_z='z')\n", (2218, 2274), False, 'from pcdsdevices.gon import BaseGon, Goniometer, GonWithDetArm, Kappa, SamPhi, XYZStage\n'), ((2279, 2343), 'pcdsdevices.gon.SamPhi', 'SamPhi', ([], {'name': '"""test4"""', 'prefix_samz': '"""samz"""', 'prefix_samphi': '"""samphi"""'}), "(name='test4', prefix_samz='samz', prefix_samphi='samphi')\n", (2285, 2343), False, 'from pcdsdevices.gon import BaseGon, Goniometer, GonWithDetArm, Kappa, SamPhi, XYZStage\n'), ((2348, 2472), 'pcdsdevices.gon.Kappa', 'Kappa', ([], {'name': '"""test5"""', 'prefix_x': '"""x"""', 'prefix_y': '"""y"""', 'prefix_z': '"""z"""', 'prefix_eta': '"""eta"""', 'prefix_kappa': '"""kappa"""', 'prefix_phi': '"""phi"""'}), "(name='test5', prefix_x='x', prefix_y='y', prefix_z='z', prefix_eta=\n 'eta', prefix_kappa='kappa', prefix_phi='phi')\n", (2353, 2472), False, 'from pcdsdevices.gon import BaseGon, Goniometer, GonWithDetArm, Kappa, SamPhi, XYZStage\n'), ((312, 419), 'pcdsdevices.gon.Goniometer', 'Goniometer', ([], {'name': '"""gon"""', 'prefix_hor': '"""a"""', 'prefix_ver': '"""b"""', 'prefix_rot': '"""c"""', 'prefix_tip': '"""d"""', 'prefix_tilt': '"""e"""'}), "(name='gon', prefix_hor='a', prefix_ver='b', prefix_rot='c',\n prefix_tip='d', prefix_tilt='e')\n", (322, 419), False, 'from pcdsdevices.gon import BaseGon, Goniometer, GonWithDetArm, Kappa, SamPhi, XYZStage\n'), ((514, 683), 'pcdsdevices.gon.Goniometer', 'Goniometer', ([], {'name': '"""gon"""', 'prefix_hor': '"""a"""', 'prefix_ver': '"""b"""', 'prefix_rot': '"""c"""', 'prefix_tip': '"""d"""', 'prefix_tilt': '"""e"""', 'prefix_detver': '"""i"""', 'prefix_dettilt': '"""j"""', 'prefix_2theta': '"""k"""'}), "(name='gon', prefix_hor='a', prefix_ver='b', prefix_rot='c',\n prefix_tip='d', prefix_tilt='e', prefix_detver='i', prefix_dettilt='j',\n prefix_2theta='k')\n", (524, 683), False, 'from pcdsdevices.gon import BaseGon, Goniometer, GonWithDetArm, Kappa, SamPhi, XYZStage\n')]
#!/usr/bin/env python3 import sys import os import argparse import time import serial import csv import math import pickle from collections import defaultdict import numpy as np from sklearn.decomposition import PCA, FastICA from sklearn.svm import SVC # Graph WINDOW_WIDTH = 800 WINDOW_HEIGHT = 800 PLOT_SCROLL = 3 # higher is faster CHANNELS = 8 FONT_SIZE = 25 # Data FREQUENCY = 200 # Hz CSV_HEADER_EMG = ["timestamp", "emg1", "emg2", "emg3", "emg4", "emg5", "emg6", "emg7", "emg8"] CSV_HEADER_CA = ["timestamp", "ca1", "ca2", "ca3", "ca4", "ca5", "ca6", "ca7", "ca8"] # Processing RMS_WINDOW_SIZE = 50 SVM_WINDOW_SIZE = 5 # higher is smoother but more delay SVM_IDLE_WEIGHT_FACTOR = 100.0 # higher makes "idle" move more important VERBOSE = False # Plotting (Pygame) window interface class Plotter(): def __init__(self, live=False): if "pygame" not in sys.modules: print("Error! pygame not loaded! Plotter not available for library use.") return None self.screen = pygame.display.set_mode((WINDOW_WIDTH, WINDOW_HEIGHT)) pygame.display.set_caption("Electromyography Processor") self.font = pygame.font.Font(None, FONT_SIZE) self.live = live self.last_values = None self.last_rms_values = None self.last_ca_values = None self.plots = 0 def plot(self, values, rms_values=[], ca_values=[], ca="", gesture="", frequency=None, recording=False): if self.last_values is None: self.last_values = values self.last_rms_values = rms_values self.last_ca_values = ca_values self.plots = len(values) + len(ca_values) return self.screen.scroll(-PLOT_SCROLL) self.screen.fill(pygame.Color("black"), (WINDOW_WIDTH - PLOT_SCROLL, 0, WINDOW_WIDTH, WINDOW_HEIGHT)) self.screen.fill(pygame.Color("black"), (0, 0, 60, WINDOW_HEIGHT)) self.clear_info() # Subplot base for i in range(self.plots): base_height = self.subplot_height(i) pygame.draw.line(self.screen, pygame.Color("darkgrey"), (WINDOW_WIDTH - PLOT_SCROLL, base_height), (WINDOW_WIDTH, base_height)) if i < 8 and self.plots >= 8: # Raw / RMS plot_text = self.font.render(f"RAW {i}", True, pygame.Color("darkgrey")) rms_offset = 10 if rms_values else 0 if rms_values: plot_rms = self.font.render(f"RMS {i}", True, pygame.Color("blue")) self.screen.blit(plot_rms, (0, base_height - rms_offset - FONT_SIZE // 2)) self.screen.blit(plot_text, (0, base_height + rms_offset - FONT_SIZE // 2)) else: # PCA/ICA plot_text = self.font.render(f" {ca.upper()} {i - len(values)}", True, pygame.Color("green")) self.screen.blit(plot_text, (0, base_height - FONT_SIZE // 2)) # Raw signal for i, (u, v) in enumerate(zip(self.last_values, values)): pygame.draw.line(self.screen, pygame.Color("darkslategrey"), (WINDOW_WIDTH - PLOT_SCROLL, self.subplot_height(i, u)), (WINDOW_WIDTH, self.subplot_height(i, v))) # Processed signals if rms_values: for i, (u, v) in enumerate(zip(self.last_rms_values, rms_values)): pygame.draw.line(self.screen, pygame.Color("blue"), (WINDOW_WIDTH - PLOT_SCROLL, self.subplot_height(i, u)), (WINDOW_WIDTH, self.subplot_height(i, v))) if ca_values: for i, (u, v) in enumerate(zip(self.last_ca_values, ca_values)): pygame.draw.line(self.screen, pygame.Color("green"), (WINDOW_WIDTH - PLOT_SCROLL, self.subplot_height(i + len(rms_values), u)), (WINDOW_WIDTH, self.subplot_height(i + len(rms_values), v))) # Information if frequency: self.render_frequency(frequency) self.render_mode() self.render_controls(recording) if gesture: self.render_classification(gesture) pygame.display.flip() self.last_values = values self.last_rms_values = rms_values self.last_ca_values = ca_values def subplot_height(self, i, value=0): scaled_value = value * 1.5 return int(WINDOW_HEIGHT / (self.plots + 1) * (i + 1 - scaled_value)) def clear_info(self): self.screen.fill(pygame.Color("black"), (0, 0, WINDOW_WIDTH, FONT_SIZE)) self.screen.fill(pygame.Color("black"), (0, WINDOW_HEIGHT - FONT_SIZE, WINDOW_WIDTH, WINDOW_HEIGHT)) def render_mode(self): mode_text = "LIVE" if self.live else "PLAYBACK" mode = self.font.render("LIVE" if self.live else "PLAYBACK", True, pygame.Color("green")) self.screen.blit(mode, (WINDOW_WIDTH // 2 - len(mode_text) * FONT_SIZE // 2, 0)) def render_frequency(self, frequency): framerate = self.font.render(f"{frequency} Hz", True, pygame.Color("green") if frequency > 180 else pygame.Color("red")) self.screen.fill(pygame.Color("black"), (0, 0, 75, FONT_SIZE)) # Clear old framerate self.screen.blit(framerate, (0, 0)) def render_controls(self, recording): pause = self.font.render("P (pause)", True, pygame.Color("white")) self.screen.blit(pause, (WINDOW_WIDTH - 250, 0)) if self.live: # Can only record live record = self.font.render("R (stop rec)" if recording else "R (record)", True, pygame.Color("red") if recording else pygame.Color("white")) self.screen.blit(record, (WINDOW_WIDTH - 150, 0)) def render_classification(self, gesture): plot_gesture = self.font.render(f"Classification: {gesture}", True, pygame.Color("green")) self.screen.blit(plot_gesture, (WINDOW_WIDTH // 2 - 225, WINDOW_HEIGHT - FONT_SIZE)) def pause(self): self.clear_info() pause = self.font.render("P (resume)", True, pygame.Color("red")) self.screen.blit(pause, (WINDOW_WIDTH - 250, 0)) self.render_mode() pygame.display.flip() def end(self): self.clear_info() pause = self.font.render("END", True, pygame.Color("red")) self.screen.blit(pause, (WINDOW_WIDTH - 250, 0)) self.render_mode() pygame.display.flip() # Interface for data streaming from either live Myo device or recorded playback class Stream(): def __init__(self, do_rms=False, pca_train_set=[], ica_train_set=[], ca_components=3, svm_train_set=[]): self.plotter = None # Late setup (display modes) self.reset() # Processing self.do_rms = do_rms self.ca_components = ca_components self.pca = self.init_pca(pca_train_set) if pca_train_set else None self.ica = self.init_ica(ica_train_set) if ica_train_set else None self.svm = self.init_svm(svm_train_set) if svm_train_set else None self.gesture = "" def create_plot(self, live=False): self.plotter = Plotter(live=live) def plot(self, data, ca=False, recording=False): self.calc_frequency() # Processing rms_data, ca_data = [], [] if ca: ca_data, data = data, [] else: if self.do_rms or self.pca is not None or self.ica is not None: rms_data = self.calc_rms(data) ca_data = [] if self.pca is not None: ca_data = self.calc_pca(rms_data) elif self.ica is not None: ca_data = self.calc_ica(rms_data) if self.svm is not None: self.gesture = self.class_svm(ca_data) if not self.paused and self.plotter is not None: self.plotter.plot([x / 500. for x in data], rms_values=[x / 500. for x in rms_data], ca_values=[x / 500. for x in ca_data], ca=self.current_model()[1], gesture=self.gesture, frequency=self.frequency, recording=recording) return rms_data, ca_data, self.gesture def calc_frequency(self): self.times.append(time.time()) if len(self.times) >= 100: self.frequency = int((len(self.times) - 1) / (self.times[-1] - self.times[0])) self.times.clear() def pause(self, state=False, toggle=False): if toggle: self.paused = not self.paused else: self.paused = state if self.paused and not self.ended: self.plotter.pause() def end(self): self.ended = True if self.plotter is not None: self.plotter.end() def reset(self): self.paused = False self.ended = False # Frequency measuring self.times = [] self.frequency = 0 # Processing self.rms_window = [] self.svm_window = [] # Processing def calc_rms(self, data): # Gather samples, up to RMS_WINDOW_SIZE self.rms_window.append(data) if len(self.rms_window) >= RMS_WINDOW_SIZE: self.rms_window.pop(0) # Calculate RMS for each channel rms_data = [0] * CHANNELS for channel in range(CHANNELS): samples = [item[channel] for item in self.rms_window] total = sum([sample ** 2 for sample in samples]) rms_data[channel] = math.sqrt(1.0 / RMS_WINDOW_SIZE * total) if VERBOSE: print(f"rms: {rms_data}") return rms_data def read_ca_train_set(self, train_set, stype="?"): emg_data = [] for file in train_set: print(f"Training {stype.upper()} with '{file}'...") emg_file = open(file, "r", newline="") emg_reader = csv.reader(emg_file, csv.unix_dialect, quoting=csv.QUOTE_MINIMAL) # Read file header = next(emg_reader) if header == CSV_HEADER_EMG: try: while True: data = next(emg_reader) _, emg = data[0], list(map(int, data[1:])) emg_data.append(self.calc_rms(emg)) except StopIteration: pass else: print("-> Error! Incorrect header! Expected 'RAW'.") self.rms_window.clear() emg_file.close() emg_data = np.array(emg_data) return emg_data def read_model(self, model, stype="?"): print(f"Reading {stype.upper()} model '{model}'...") with open(model, "rb") as f: return pickle.load(f) def init_pca(self, train_set): if isinstance(train_set, list): emg_data = self.read_ca_train_set(train_set, "pca") # Initialize and train pca = PCA(n_components=self.ca_components) pca.fit(emg_data) else: pca = self.read_model(train_set, "pca") return pca def calc_pca(self, rms_data): emg_data = np.array(rms_data).reshape(1, -1) # Reshape to 1 sample, N features pca_data = self.pca.transform(emg_data)[0] # Take 1 sample from array of samples (contains only one) if VERBOSE: print(f"pca: {pca_data}") return pca_data def init_ica(self, train_set): if isinstance(train_set, list): emg_data = self.read_ca_train_set(train_set, "ica") # Initialize and train ica = FastICA(n_components=self.ca_components, random_state=0) ica.fit(emg_data) else: ica = self.read_model(train_set, "ica") return ica def calc_ica(self, rms_data): emg_data = np.array(rms_data).reshape(1, -1) # Reshape to 1 sample, N features ica_data = self.ica.transform(emg_data)[0] # Take 1 sample from array of samples (contains only one) ica_data *= 5000 # Scale up if VERBOSE: print(f"ica: {ica_data}") return ica_data def read_class_train_set(self, train_set, stype="?"): emg_data = [] classes = [] for file in train_set: classification = os.path.basename(file).split("_")[0] print(f"Training {stype.upper()} '{classification}' with '{file}'...") emg_file = open(file, "r", newline="") emg_reader = csv.reader(emg_file, csv.unix_dialect, quoting=csv.QUOTE_MINIMAL) # Read file header = next(emg_reader) if header == CSV_HEADER_CA[:self.ca_components + 1]: try: while True: data = next(emg_reader) _, emg = data[0], list(map(float, data[1:])) emg_data.append(emg) classes.append(classification) except StopIteration: pass else: print("-> Error! Incorrect header! Expected 'PCA/ICA'.") emg_file.close() if "idle" not in classes: print("Warning! No 'idle' move trained!") emg_data, classes = np.array(emg_data), np.array(classes) return emg_data, classes def init_svm(self, train_set): if isinstance(train_set, list): emg_data, classes = self.read_class_train_set(train_set, "svm") svm = SVC(random_state=0, kernel="rbf", class_weight={"idle": SVM_IDLE_WEIGHT_FACTOR}) svm.fit(emg_data, classes) else: svm = self.read_model(train_set, "svm") return svm def class_svm(self, ca_data): # Gather samples, up to SVM_WINDOW_SIZE to smooth classification self.svm_window.append(ca_data) if len(self.svm_window) > SVM_WINDOW_SIZE: self.svm_window.pop(0) window = np.array(self.svm_window) svm_classes = self.svm.predict(window) # predict each sample in window # Take classification with most occurences in the window d = defaultdict(int) for svm_class in svm_classes: d[svm_class] += 1 svm_class = max(d.items(), key=lambda x: x[1])[0] if VERBOSE: print(f"svm: {svm_class}") return svm_class return "" def current_model(self): if self.svm is not None: return self.svm, "svm" elif self.pca is not None: return self.pca, "pca" elif self.ica is not None: return self.ica, "ica" return None, "" # Live Myo device interface class Myo(): def __init__(self, stream, tty, native, mac): # Instantiate self.myo = MyoRaw(tty, native, mac) self.stream = stream self.recording = False self.recording_type = self.init_recording() # Recording self.emg_file = None self.emg_writer = None # Setup self.setup() def close(self): self.myo.disconnect() self.record(False) def setup(self): # Add handles to process EMG and battery level data self.myo.add_handler(DataCategory.EMG, self.handle_emg) self.myo.add_handler(DataCategory.BATTERY, self.handle_battery) # Subscribe to all data services in full RAW mode (200 Hz) self.myo.subscribe(EMGMode.RAW) # Disable sleep to a void disconnects while retrieving data self.myo.set_sleep_mode(1) # Vibrate to signalise a successful setup # myo.vibrate(1) def run(self): self.myo.run(1) def disconnect(self): self.myo.disconnect() def sleep(self): self.myo.deep_sleep() def handle_emg(self, timestamp, emg, moving, characteristic_num): emg = list(emg) _, ca_data, _ = self.stream.plot(emg, recording=self.recording) record_data = ca_data if len(ca_data) > 0 else emg if self.recording: csv_data = [timestamp] csv_data.extend(record_data) try: self.emg_writer.writerow(csv_data) except AttributeError: print("Error! Unable to write to CSV!") if VERBOSE: print(f"[myo] {self.recording_type}: {timestamp}, {record_data}") def handle_battery(self, timestamp, battery_level): if battery_level < 5: self.myo.set_leds([255, 0, 0], [255, 0, 0]) # red logo, red bar else: self.myo.set_leds([128, 128, 255], [128, 128, 255]) # purple logo, purple bar if VERBOSE: print(f"[myo] battery level: {timestamp}, {battery_level}") def init_recording(self): if self.stream.pca is not None: return "pca" elif self.stream.ica is not None: return "ica" return "raw" def record(self, state=False, toggle=False): if toggle: recording = not self.recording else: recording = state if recording: filename = f"recordings/{self.recording_type}/{time.strftime('%Y%m%d-%H%M%S')}.csv" os.makedirs(os.path.dirname(filename), exist_ok=True) self.emg_file = open(filename, "w", newline="") self.emg_writer = csv.writer(self.emg_file, csv.unix_dialect, quoting=csv.QUOTE_MINIMAL) if self.recording_type == "raw": self.emg_writer.writerow(CSV_HEADER_EMG) else: self.emg_writer.writerow(CSV_HEADER_CA[:self.stream.ca_components + 1]) elif self.emg_file is not None: self.emg_file.close() self.emg_file = None self.emg_writer = None self.recording = recording # Recorded Myo data playback interface class Playback(): def __init__(self, stream, filename): self.stream = stream self.valid = False self.type = "" try: self.emg_file = open(filename, "r", newline="") self.emg_reader = csv.reader(self.emg_file, csv.unix_dialect, quoting=csv.QUOTE_MINIMAL) self.read_header() except FileNotFoundError: self.emg_file = None def close(self): if self.emg_file: self.emg_file.close() def read_header(self): try: header = next(self.emg_reader) if header == CSV_HEADER_EMG: self.valid = True self.type = "raw" if header[:2] == CSV_HEADER_CA[:2]: self.valid = True self.type = "ca" except StopIteration: pass def is_valid(self): return self.valid # Plays a frame from the recording and indicating end of recording on subsequent calls def play_frame(self): if not self.stream.paused: try: data = next(self.emg_reader) if self.type == "raw": timestamp, emg = data[0], list(map(int, data[1:])) rms_data, ca_data, gesture = self.stream.plot(emg) else: timestamp, emg = data[0], list(map(float, data[1:])) rms_data, ca_data, gesture = self.stream.plot(emg, ca=True) if VERBOSE: print(f"[playback] emg: {timestamp}, {emg}") return timestamp, rms_data, ca_data, gesture except StopIteration: self.stream.end() return 0, [], [], "" def main(): # Parse arguments parser = argparse.ArgumentParser(description="Electromyography Processor") group1 = parser.add_mutually_exclusive_group() group1.add_argument("-r", "--recording", default=None, metavar="REC", help="playback recorded Myo data stream") group1.add_argument("-s", "--sleep", default=False, action="store_true", help="put Myo into deep sleep (turn off)") parser.add_argument("--rms", default=False, action="store_true", help="process stream using RMS smoothing") group2 = parser.add_mutually_exclusive_group() group2.add_argument("--pca", nargs="+", metavar="REC", help="process stream using RAW training set or PCA model") group2.add_argument("--ica", nargs="+", metavar="REC", help="process stream using RAW training set or ICA model") parser.add_argument("-c", "--components", default=3, type=int, help="PCA/ICA components to use") group3 = parser.add_mutually_exclusive_group() group3.add_argument("--svm", nargs="+", metavar="REC", help="classify using PCA/ICA training set or SVM model") group4 = parser.add_mutually_exclusive_group() group4.add_argument("--tty", default=None, help="Myo dongle device (autodetected if omitted)") group4.add_argument("--native", default=False, action="store_true", help="use a native Bluetooth stack") parser.add_argument("--mac", default=None, help="Myo MAC address (arbitrarily detected if omitted)") parser.add_argument("-v", "--verbose", default=False, action="store_true", help="verbose output") args = parser.parse_args() if args.svm and not args.pca and not args.ica: parser.error("the following arguments are required for 'svm': 'pca' or 'ica'") # Model was given instead of trainining set if args.pca is not None and len(args.pca) == 1 and not args.pca[0].endswith(".csv"): args.pca = args.pca[0] if args.ica is not None and len(args.ica) == 1 and not args.ica[0].endswith(".csv"): args.ica = args.ica[0] if args.svm is not None and len(args.svm) == 1 and not args.svm[0].endswith(".csv"): args.svm = args.svm[0] if args.verbose: global VERBOSE VERBOSE = args.verbose live_myo = args.recording is None # Setup common stream interface for Myo or Playback stream = Stream(do_rms=args.rms, pca_train_set=args.pca, ica_train_set=args.ica, svm_train_set=args.svm, ca_components=args.components) # Setup Myo or Playback if live_myo: try: print("Connecting to Myo...") myo = Myo(stream, args.tty, args.native, args.mac) print("Connected to Myo!") except (ValueError, KeyboardInterrupt) as e: print(f"Error! Unable to connect!\n{e}") return 1 else: playback = Playback(stream, args.recording) if not playback.is_valid(): print("Error! Invalid CSV file!") return 2 # Run main logic if args.sleep: if live_myo: myo.sleep() else: pygame.init() stream.create_plot(live=live_myo) # Run until terminated by user or recording ended try: starttime = time.time() while True: if live_myo: try: myo.run() except serial.serialutil.SerialException: print("Error! Myo exception! Attempting reboot...") myo.disconnect() myo = Myo(stream, args.tty, args.native, args.mac) else: playback.play_frame() # Delay by (1 second / FREQUENCY Hz) including execution time delay = 1 / FREQUENCY diff = min(time.time() - starttime, 1 / FREQUENCY) time.sleep(delay - diff) starttime = time.time() # Handle Pygame events for ev in pygame.event.get(): if ev.type == pygame.QUIT: raise KeyboardInterrupt() elif ev.type == pygame.KEYDOWN: if ev.key == pygame.K_q: raise KeyboardInterrupt() elif ev.key == pygame.K_p: stream.pause(toggle=True) elif ev.key == pygame.K_r: if live_myo: myo.record(toggle=True) except KeyboardInterrupt: pass if live_myo: myo.close() else: playback.close() return 0 # Conditional imports if __name__ == "__main__" or os.environ.get("EMGPROC_LOAD_GAME", False): import pygame if __name__ == "__main__" or os.environ.get("EMGPROC_LOAD_MYO", False): from myo_raw import MyoRaw, DataCategory, EMGMode if __name__ == "__main__": sys.exit(main())
[ "pygame.init", "math.sqrt", "time.sleep", "numpy.array", "sklearn.decomposition.FastICA", "pygame.font.Font", "argparse.ArgumentParser", "sklearn.decomposition.PCA", "pygame.display.set_mode", "pygame.display.flip", "csv.reader", "csv.writer", "pickle.load", "os.path.dirname", "time.time...
[((19921, 19986), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {'description': '"""Electromyography Processor"""'}), "(description='Electromyography Processor')\n", (19944, 19986), False, 'import argparse\n'), ((24584, 24626), 'os.environ.get', 'os.environ.get', (['"""EMGPROC_LOAD_GAME"""', '(False)'], {}), "('EMGPROC_LOAD_GAME', False)\n", (24598, 24626), False, 'import os\n'), ((24675, 24716), 'os.environ.get', 'os.environ.get', (['"""EMGPROC_LOAD_MYO"""', '(False)'], {}), "('EMGPROC_LOAD_MYO', False)\n", (24689, 24716), False, 'import os\n'), ((1025, 1079), 'pygame.display.set_mode', 'pygame.display.set_mode', (['(WINDOW_WIDTH, WINDOW_HEIGHT)'], {}), '((WINDOW_WIDTH, WINDOW_HEIGHT))\n', (1048, 1079), False, 'import pygame\n'), ((1088, 1144), 'pygame.display.set_caption', 'pygame.display.set_caption', (['"""Electromyography Processor"""'], {}), "('Electromyography Processor')\n", (1114, 1144), False, 'import pygame\n'), ((1165, 1198), 'pygame.font.Font', 'pygame.font.Font', (['None', 'FONT_SIZE'], {}), '(None, FONT_SIZE)\n', (1181, 1198), False, 'import pygame\n'), ((4272, 4293), 'pygame.display.flip', 'pygame.display.flip', ([], {}), '()\n', (4291, 4293), False, 'import pygame\n'), ((6363, 6384), 'pygame.display.flip', 'pygame.display.flip', ([], {}), '()\n', (6382, 6384), False, 'import pygame\n'), ((6591, 6612), 'pygame.display.flip', 'pygame.display.flip', ([], {}), '()\n', (6610, 6612), False, 'import pygame\n'), ((10782, 10800), 'numpy.array', 'np.array', (['emg_data'], {}), '(emg_data)\n', (10790, 10800), True, 'import numpy as np\n'), ((15081, 15105), 'myo_raw.MyoRaw', 'MyoRaw', (['tty', 'native', 'mac'], {}), '(tty, native, mac)\n', (15087, 15105), False, 'from myo_raw import MyoRaw, DataCategory, EMGMode\n'), ((22923, 22936), 'pygame.init', 'pygame.init', ([], {}), '()\n', (22934, 22936), False, 'import pygame\n'), ((1767, 1788), 'pygame.Color', 'pygame.Color', (['"""black"""'], {}), "('black')\n", (1779, 1788), False, 'import pygame\n'), ((1877, 1898), 'pygame.Color', 'pygame.Color', (['"""black"""'], {}), "('black')\n", (1889, 1898), False, 'import pygame\n'), ((4619, 4640), 'pygame.Color', 'pygame.Color', (['"""black"""'], {}), "('black')\n", (4631, 4640), False, 'import pygame\n'), ((4700, 4721), 'pygame.Color', 'pygame.Color', (['"""black"""'], {}), "('black')\n", (4712, 4721), False, 'import pygame\n'), ((4975, 4996), 'pygame.Color', 'pygame.Color', (['"""green"""'], {}), "('green')\n", (4987, 4996), False, 'import pygame\n'), ((5322, 5343), 'pygame.Color', 'pygame.Color', (['"""black"""'], {}), "('black')\n", (5334, 5343), False, 'import pygame\n'), ((5530, 5551), 'pygame.Color', 'pygame.Color', (['"""white"""'], {}), "('white')\n", (5542, 5551), False, 'import pygame\n'), ((6032, 6053), 'pygame.Color', 'pygame.Color', (['"""green"""'], {}), "('green')\n", (6044, 6053), False, 'import pygame\n'), ((6249, 6268), 'pygame.Color', 'pygame.Color', (['"""red"""'], {}), "('red')\n", (6261, 6268), False, 'import pygame\n'), ((6477, 6496), 'pygame.Color', 'pygame.Color', (['"""red"""'], {}), "('red')\n", (6489, 6496), False, 'import pygame\n'), ((8520, 8531), 'time.time', 'time.time', ([], {}), '()\n', (8529, 8531), False, 'import time\n'), ((9771, 9811), 'math.sqrt', 'math.sqrt', (['(1.0 / RMS_WINDOW_SIZE * total)'], {}), '(1.0 / RMS_WINDOW_SIZE * total)\n', (9780, 9811), False, 'import math\n'), ((10147, 10212), 'csv.reader', 'csv.reader', (['emg_file', 'csv.unix_dialect'], {'quoting': 'csv.QUOTE_MINIMAL'}), '(emg_file, csv.unix_dialect, quoting=csv.QUOTE_MINIMAL)\n', (10157, 10212), False, 'import csv\n'), ((10987, 11001), 'pickle.load', 'pickle.load', (['f'], {}), '(f)\n', (10998, 11001), False, 'import pickle\n'), ((11196, 11232), 'sklearn.decomposition.PCA', 'PCA', ([], {'n_components': 'self.ca_components'}), '(n_components=self.ca_components)\n', (11199, 11232), False, 'from sklearn.decomposition import PCA, FastICA\n'), ((11860, 11916), 'sklearn.decomposition.FastICA', 'FastICA', ([], {'n_components': 'self.ca_components', 'random_state': '(0)'}), '(n_components=self.ca_components, random_state=0)\n', (11867, 11916), False, 'from sklearn.decomposition import PCA, FastICA\n'), ((12747, 12812), 'csv.reader', 'csv.reader', (['emg_file', 'csv.unix_dialect'], {'quoting': 'csv.QUOTE_MINIMAL'}), '(emg_file, csv.unix_dialect, quoting=csv.QUOTE_MINIMAL)\n', (12757, 12812), False, 'import csv\n'), ((13514, 13532), 'numpy.array', 'np.array', (['emg_data'], {}), '(emg_data)\n', (13522, 13532), True, 'import numpy as np\n'), ((13534, 13551), 'numpy.array', 'np.array', (['classes'], {}), '(classes)\n', (13542, 13551), True, 'import numpy as np\n'), ((13756, 13841), 'sklearn.svm.SVC', 'SVC', ([], {'random_state': '(0)', 'kernel': '"""rbf"""', 'class_weight': "{'idle': SVM_IDLE_WEIGHT_FACTOR}"}), "(random_state=0, kernel='rbf', class_weight={'idle': SVM_IDLE_WEIGHT_FACTOR}\n )\n", (13759, 13841), False, 'from sklearn.svm import SVC\n'), ((14218, 14243), 'numpy.array', 'np.array', (['self.svm_window'], {}), '(self.svm_window)\n', (14226, 14243), True, 'import numpy as np\n'), ((14414, 14430), 'collections.defaultdict', 'defaultdict', (['int'], {}), '(int)\n', (14425, 14430), False, 'from collections import defaultdict\n'), ((17642, 17712), 'csv.writer', 'csv.writer', (['self.emg_file', 'csv.unix_dialect'], {'quoting': 'csv.QUOTE_MINIMAL'}), '(self.emg_file, csv.unix_dialect, quoting=csv.QUOTE_MINIMAL)\n', (17652, 17712), False, 'import csv\n'), ((18383, 18453), 'csv.reader', 'csv.reader', (['self.emg_file', 'csv.unix_dialect'], {'quoting': 'csv.QUOTE_MINIMAL'}), '(self.emg_file, csv.unix_dialect, quoting=csv.QUOTE_MINIMAL)\n', (18393, 18453), False, 'import csv\n'), ((23075, 23086), 'time.time', 'time.time', ([], {}), '()\n', (23084, 23086), False, 'import time\n'), ((2104, 2128), 'pygame.Color', 'pygame.Color', (['"""darkgrey"""'], {}), "('darkgrey')\n", (2116, 2128), False, 'import pygame\n'), ((3113, 3142), 'pygame.Color', 'pygame.Color', (['"""darkslategrey"""'], {}), "('darkslategrey')\n", (3125, 3142), False, 'import pygame\n'), ((5230, 5251), 'pygame.Color', 'pygame.Color', (['"""green"""'], {}), "('green')\n", (5242, 5251), False, 'import pygame\n'), ((5276, 5295), 'pygame.Color', 'pygame.Color', (['"""red"""'], {}), "('red')\n", (5288, 5295), False, 'import pygame\n'), ((11403, 11421), 'numpy.array', 'np.array', (['rms_data'], {}), '(rms_data)\n', (11411, 11421), True, 'import numpy as np\n'), ((12087, 12105), 'numpy.array', 'np.array', (['rms_data'], {}), '(rms_data)\n', (12095, 12105), True, 'import numpy as np\n'), ((17510, 17535), 'os.path.dirname', 'os.path.dirname', (['filename'], {}), '(filename)\n', (17525, 17535), False, 'import os\n'), ((23868, 23886), 'pygame.event.get', 'pygame.event.get', ([], {}), '()\n', (23884, 23886), False, 'import pygame\n'), ((2379, 2403), 'pygame.Color', 'pygame.Color', (['"""darkgrey"""'], {}), "('darkgrey')\n", (2391, 2403), False, 'import pygame\n'), ((2880, 2901), 'pygame.Color', 'pygame.Color', (['"""green"""'], {}), "('green')\n", (2892, 2901), False, 'import pygame\n'), ((3479, 3499), 'pygame.Color', 'pygame.Color', (['"""blue"""'], {}), "('blue')\n", (3491, 3499), False, 'import pygame\n'), ((3813, 3834), 'pygame.Color', 'pygame.Color', (['"""green"""'], {}), "('green')\n", (3825, 3834), False, 'import pygame\n'), ((5786, 5805), 'pygame.Color', 'pygame.Color', (['"""red"""'], {}), "('red')\n", (5798, 5805), False, 'import pygame\n'), ((5824, 5845), 'pygame.Color', 'pygame.Color', (['"""white"""'], {}), "('white')\n", (5836, 5845), False, 'import pygame\n'), ((17449, 17479), 'time.strftime', 'time.strftime', (['"""%Y%m%d-%H%M%S"""'], {}), "('%Y%m%d-%H%M%S')\n", (17462, 17479), False, 'import time\n'), ((23733, 23757), 'time.sleep', 'time.sleep', (['(delay - diff)'], {}), '(delay - diff)\n', (23743, 23757), False, 'import time\n'), ((23790, 23801), 'time.time', 'time.time', ([], {}), '()\n', (23799, 23801), False, 'import time\n'), ((2555, 2575), 'pygame.Color', 'pygame.Color', (['"""blue"""'], {}), "('blue')\n", (2567, 2575), False, 'import pygame\n'), ((12550, 12572), 'os.path.basename', 'os.path.basename', (['file'], {}), '(file)\n', (12566, 12572), False, 'import os\n'), ((23673, 23684), 'time.time', 'time.time', ([], {}), '()\n', (23682, 23684), False, 'import time\n')]
import numpy as np from preprocess import Vectorizer from flask import render_template, make_response from google.oauth2.id_token import verify_oauth2_token from google.auth.transport.requests import Request from google.cloud import firestore from os.path import join, abspath, dirname from random import randint from pickle import load from scipy.sparse import load_npz database = firestore.Client() hedgy_path = dirname(abspath(__file__)) with open(join(hedgy_path, 'chapters.p'), 'rb') as chapters_f, open(join(hedgy_path, 'vectorizer.p'), 'rb') as vectorizer_f: chapters = load(chapters_f) vectorizer = load(vectorizer_f) tfidf_matrix = load_npz(join(hedgy_path, 'tfidf.npz')) similarity_matrix = np.load(join(hedgy_path, 'similarity.npy')) def hedgy(request): ranking, sliced, max_request, seed, token = [], False, 50, None, None credential, clicks = None, None if request.method == 'POST': credential = request.form['credential'] if 'credential' in request.form else None elif request.cookies.get('__session'): credential, clicks = request.cookies.get('__session').split('#') if credential: try: token = verify_oauth2_token(credential, Request(), '1080182836213-psdjtgo2u10a1fb6e4sbdfpdlmco5i63.apps.googleusercontent.com') except: pass if token: user_doc = database.collection('users').document(token['sub']) if not user_doc.get().exists: user_doc.set({'email': token['email'], 'given_name': token['given_name'], 'family_name': token['family_name'], 'picture': token['picture'], 'clicks': []}) if clicks: user_doc.update({'clicks': firestore.ArrayUnion(clicks[:-1].split(','))}) if 'max' in request.args: max_request = int(request.args.get('max')) if 'query' in request.args or 'similar' in request.args: if 'query' in request.args: query_vector = vectorizer.transform([request.args.get('query')]) similarity_vector = (tfidf_matrix @ query_vector.T).toarray().squeeze() else: similarity_vector = similarity_matrix[int(request.args.get('similar'))] if np.any(similarity_vector): max_chapters = np.count_nonzero(similarity_vector) if max_request < max_chapters: max_chapters = max_request sliced = True ranking = np.argsort(similarity_vector)[::-1][:max_chapters].tolist() elif 'seed' in request.args: seed = int(request.args.get('seed')) else: seed = randint(1, 1000000) if seed: np.random.seed(seed) ranking = np.random.permutation(len(chapters))[:max_request].tolist() sliced = True response = make_response(render_template('hedgy.html', chapters=chapters, ranking=ranking, sliced=sliced, max_request=max_request, seed=seed, token=token, args=request.args)) if token: response.set_cookie('__session', credential + '#', secure=True) response.headers['Cache-Control'] = 'private' elif '__session' in request.cookies: response.set_cookie('__session', '', expires=0) response.headers['Cache-Control'] = 'private' return response
[ "flask.render_template", "google.cloud.firestore.Client", "google.auth.transport.requests.Request", "pickle.load", "os.path.join", "numpy.any", "numpy.count_nonzero", "numpy.argsort", "numpy.random.seed", "os.path.abspath", "random.randint" ]
[((383, 401), 'google.cloud.firestore.Client', 'firestore.Client', ([], {}), '()\n', (399, 401), False, 'from google.cloud import firestore\n'), ((424, 441), 'os.path.abspath', 'abspath', (['__file__'], {}), '(__file__)\n', (431, 441), False, 'from os.path import join, abspath, dirname\n'), ((583, 599), 'pickle.load', 'load', (['chapters_f'], {}), '(chapters_f)\n', (587, 599), False, 'from pickle import load\n'), ((617, 635), 'pickle.load', 'load', (['vectorizer_f'], {}), '(vectorizer_f)\n', (621, 635), False, 'from pickle import load\n'), ((660, 689), 'os.path.join', 'join', (['hedgy_path', '"""tfidf.npz"""'], {}), "(hedgy_path, 'tfidf.npz')\n", (664, 689), False, 'from os.path import join, abspath, dirname\n'), ((719, 753), 'os.path.join', 'join', (['hedgy_path', '"""similarity.npy"""'], {}), "(hedgy_path, 'similarity.npy')\n", (723, 753), False, 'from os.path import join, abspath, dirname\n'), ((453, 483), 'os.path.join', 'join', (['hedgy_path', '"""chapters.p"""'], {}), "(hedgy_path, 'chapters.p')\n", (457, 483), False, 'from os.path import join, abspath, dirname\n'), ((511, 543), 'os.path.join', 'join', (['hedgy_path', '"""vectorizer.p"""'], {}), "(hedgy_path, 'vectorizer.p')\n", (515, 543), False, 'from os.path import join, abspath, dirname\n'), ((2621, 2640), 'random.randint', 'randint', (['(1)', '(1000000)'], {}), '(1, 1000000)\n', (2628, 2640), False, 'from random import randint\n'), ((2662, 2682), 'numpy.random.seed', 'np.random.seed', (['seed'], {}), '(seed)\n', (2676, 2682), True, 'import numpy as np\n'), ((2812, 2965), 'flask.render_template', 'render_template', (['"""hedgy.html"""'], {'chapters': 'chapters', 'ranking': 'ranking', 'sliced': 'sliced', 'max_request': 'max_request', 'seed': 'seed', 'token': 'token', 'args': 'request.args'}), "('hedgy.html', chapters=chapters, ranking=ranking, sliced=\n sliced, max_request=max_request, seed=seed, token=token, args=request.args)\n", (2827, 2965), False, 'from flask import render_template, make_response\n'), ((2202, 2227), 'numpy.any', 'np.any', (['similarity_vector'], {}), '(similarity_vector)\n', (2208, 2227), True, 'import numpy as np\n'), ((1210, 1219), 'google.auth.transport.requests.Request', 'Request', ([], {}), '()\n', (1217, 1219), False, 'from google.auth.transport.requests import Request\n'), ((2260, 2295), 'numpy.count_nonzero', 'np.count_nonzero', (['similarity_vector'], {}), '(similarity_vector)\n', (2276, 2295), True, 'import numpy as np\n'), ((2450, 2479), 'numpy.argsort', 'np.argsort', (['similarity_vector'], {}), '(similarity_vector)\n', (2460, 2479), True, 'import numpy as np\n')]
#-*- coding:utf-8 -*- import sys, traceback from core_backend import context from core_backend.libs.exception import Error import logging #logger = Log.getDebugLogger() #logger.setLevel(logging.INFO) logger = logging.getLogger(__name__) class plugin(object): def __init__(self, handler, session): self.handler = handler self.session = session self.context = self.handler.context self.request = self.context.request self.service = self.handler._service def process(self): pass def post_process(self): pass class PluginHandler(object): def __init__(self, handler, session): self.handler = handler self.session = session self.plg_modules = [] self.plg_inst_list = [] def import_module(self, module, fromlist): # return __import__(self.get_module(m), fromlist=["plugins"]) return __import__(module, fromlist=fromlist) def load_plugins(self, plg_module): plgconfig = self.import_module(plg_module, [plg_module]) module_files = plgconfig.plugins_modules for f in module_files: m = self.import_module(plg_module + '.' + f, [plg_module]) self.plg_modules.append(m) ins = m.Plugin(self.handler, self.session) self.plg_inst_list.append(ins) def run_plugins(self): for ins in self.plg_inst_list: ins.process() def run_post_plugins(self): for ins in self.plg_inst_list: ins.post_process()
[ "logging.getLogger" ]
[((210, 237), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (227, 237), False, 'import logging\n')]
''' Filename: predict.py Python Version: 3.6.5 Project: Neutrophil Identifier Author: <NAME> Created date: Sep 5, 2018 4:13 PM ----- Last Modified: Oct 9, 2018 3:48 PM Modified By: <NAME> ----- License: MIT http://www.opensource.org/licenses/MIT ''' import os import sys import logging from math import ceil from keras.models import load_model import numpy as np import tables as tb from paths import Paths def read_hdf5(hdf5_file, dataset="pred", batch_size=32): """ """ data_img = dataset + "_img" m_data = hdf5_file.root.__getitem__(data_img).shape[0] batch_list = list(range(int(ceil(m_data / batch_size)))) while True: for num in batch_list: n_start = num * batch_size n_end = min((num + 1) * batch_size, m_data) inputs = hdf5_file.root.__getitem__(data_img)[n_start:n_end, ...] yield inputs def predict(model_path, hdf5_file_path=None): """ """ try: BATCH_SIZE = 32 model = load_model(model_path) default_path = Paths.tiles_80 if hdf5_file_path: hdf5_file = tb.open_file(hdf5_file_path, mode='r') else: hdf5_file = tb.open_file(default_path, mode='r') m_samples = hdf5_file.root.__getitem__("pred_img").shape[0] logging.debug(f'# of samples: {m_samples}') steps = int(ceil(m_samples / BATCH_SIZE)) generator = read_hdf5(hdf5_file, dataset="pred", batch_size=BATCH_SIZE) preds = model.predict_generator(generator, steps=steps, verbose=1) logging.info(preds[0:100]) base_name = os.path.basename(model_path).split('.')[0] save_path = os.path.join( Paths.data_test, base_name + "_tiles_80_preds.csv") np.savetxt(save_path, preds, delimiter=',') except Exception as e: hdf5_file.close() logging.debug(e.with_traceback()) finally: hdf5_file.close() if __name__ == "__main__": logging.basicConfig(level=logging.DEBUG) predict(sys.argv[1])
[ "logging.basicConfig", "keras.models.load_model", "logging.debug", "math.ceil", "os.path.join", "tables.open_file", "os.path.basename", "numpy.savetxt", "logging.info" ]
[((1965, 2005), 'logging.basicConfig', 'logging.basicConfig', ([], {'level': 'logging.DEBUG'}), '(level=logging.DEBUG)\n', (1984, 2005), False, 'import logging\n'), ((996, 1018), 'keras.models.load_model', 'load_model', (['model_path'], {}), '(model_path)\n', (1006, 1018), False, 'from keras.models import load_model\n'), ((1300, 1343), 'logging.debug', 'logging.debug', (['f"""# of samples: {m_samples}"""'], {}), "(f'# of samples: {m_samples}')\n", (1313, 1343), False, 'import logging\n'), ((1557, 1583), 'logging.info', 'logging.info', (['preds[0:100]'], {}), '(preds[0:100])\n', (1569, 1583), False, 'import logging\n'), ((1668, 1732), 'os.path.join', 'os.path.join', (['Paths.data_test', "(base_name + '_tiles_80_preds.csv')"], {}), "(Paths.data_test, base_name + '_tiles_80_preds.csv')\n", (1680, 1732), False, 'import os\n'), ((1754, 1797), 'numpy.savetxt', 'np.savetxt', (['save_path', 'preds'], {'delimiter': '""","""'}), "(save_path, preds, delimiter=',')\n", (1764, 1797), True, 'import numpy as np\n'), ((1108, 1146), 'tables.open_file', 'tb.open_file', (['hdf5_file_path'], {'mode': '"""r"""'}), "(hdf5_file_path, mode='r')\n", (1120, 1146), True, 'import tables as tb\n'), ((1186, 1222), 'tables.open_file', 'tb.open_file', (['default_path'], {'mode': '"""r"""'}), "(default_path, mode='r')\n", (1198, 1222), True, 'import tables as tb\n'), ((1364, 1392), 'math.ceil', 'ceil', (['(m_samples / BATCH_SIZE)'], {}), '(m_samples / BATCH_SIZE)\n', (1368, 1392), False, 'from math import ceil\n'), ((608, 633), 'math.ceil', 'ceil', (['(m_data / batch_size)'], {}), '(m_data / batch_size)\n', (612, 633), False, 'from math import ceil\n'), ((1605, 1633), 'os.path.basename', 'os.path.basename', (['model_path'], {}), '(model_path)\n', (1621, 1633), False, 'import os\n')]
import argparse import json from tqdm import tqdm from common.dataset.reader import JSONLineReader from common.util.log_helper import LogHelper def _sent_to_str(sent): return sent[-2] + "$$$" + str(sent[-1]) def _replace_sent_with_str(sent, string): segments = string.split(r"$$$") if len(segments) != 2: raise Exception("Illegal string: " + string) sent[-2] = segments[0] sent[-1] = int(segments[1]) return sent def _build_new_sent_with_str(string, num_of_segments=2): if num_of_segments == 2: sent = ["", -1] elif num_of_segments == 4: sent = [-1, -1, "", -1] else: raise Exception("Illegal num_of_segments: " + str(num_of_segments)) return _replace_sent_with_str(sent, string) def _sents_from_evidences(evidences): sents = set() for evidence in evidences: for s in evidence: sent = _sent_to_str(s) sents.add(sent) return sents def _fill_pred_sents_with_gold(pred_sents, gold_sents, max_sent): selected_sents = pred_sents[:max_sent] neg_indices = [] for i, selected in enumerate(selected_sents): key_selected = _sent_to_str(selected) if key_selected in gold_sents: gold_sents.remove(key_selected) else: neg_indices.append(i) if len(gold_sents) == 0: return selected_sents if len(selected_sents) <= max_sent: for _ in range(max_sent - len(selected_sents)): selected_sents.append(_build_new_sent_with_str(gold_sents.pop())) if len(gold_sents) == 0: return selected_sents if len(neg_indices) > 0: neg_indices = reversed(neg_indices) for i in neg_indices: sent = selected_sents[i] selected_sents[i] = _replace_sent_with_str(sent, gold_sents.pop()) if len(gold_sents) == 0: return selected_sents if len(gold_sents) > 0: logger.warn(str(len(gold_sents)) + " gold sentences cannot be filled into prediction") return selected_sents if __name__ == '__main__': LogHelper.setup() logger = LogHelper.get_logger('fill_gold_sentences') parser = argparse.ArgumentParser() parser.add_argument( '--input', help='/path/to/input/file', required=True) parser.add_argument( '--output', help='/path/to/output/file', required=True) parser.add_argument( '--max-sent', type=int, help='Maximal number of sentences per claim', default=10) args = parser.parse_args() jlr = JSONLineReader() data = jlr.read(args.input) with open(args.output, "w+") as output_file: for data in tqdm(data): if data['verifiable'] != 'NOT VERIFIABLE': pred_sents = data['predicted_sentences'] gold_evidences = data['evidence'] gold_sents = _sents_from_evidences(gold_evidences) filled_pred_sents = _fill_pred_sents_with_gold( pred_sents, gold_sents, args.max_sent) data['predicted_sentences'] = filled_pred_sents output_file.write(json.dumps(data) + "\n")
[ "argparse.ArgumentParser", "common.util.log_helper.LogHelper.setup", "common.dataset.reader.JSONLineReader", "common.util.log_helper.LogHelper.get_logger", "tqdm.tqdm", "json.dumps" ]
[((2117, 2134), 'common.util.log_helper.LogHelper.setup', 'LogHelper.setup', ([], {}), '()\n', (2132, 2134), False, 'from common.util.log_helper import LogHelper\n'), ((2148, 2191), 'common.util.log_helper.LogHelper.get_logger', 'LogHelper.get_logger', (['"""fill_gold_sentences"""'], {}), "('fill_gold_sentences')\n", (2168, 2191), False, 'from common.util.log_helper import LogHelper\n'), ((2205, 2230), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {}), '()\n', (2228, 2230), False, 'import argparse\n'), ((2563, 2579), 'common.dataset.reader.JSONLineReader', 'JSONLineReader', ([], {}), '()\n', (2577, 2579), False, 'from common.dataset.reader import JSONLineReader\n'), ((2681, 2691), 'tqdm.tqdm', 'tqdm', (['data'], {}), '(data)\n', (2685, 2691), False, 'from tqdm import tqdm\n'), ((3139, 3155), 'json.dumps', 'json.dumps', (['data'], {}), '(data)\n', (3149, 3155), False, 'import json\n')]
# Code generated by `typeddictgen`. DO NOT EDIT. """V1ISCSIVolumeSourceDict generated type.""" from typing import TypedDict, List from kubernetes_typed.client import V1LocalObjectReferenceDict V1ISCSIVolumeSourceDict = TypedDict( "V1ISCSIVolumeSourceDict", { "chapAuthDiscovery": bool, "chapAuthSession": bool, "fsType": str, "initiatorName": str, "iqn": str, "iscsiInterface": str, "lun": int, "portals": List[str], "readOnly": bool, "secretRef": V1LocalObjectReferenceDict, "targetPortal": str, }, total=False, )
[ "typing.TypedDict" ]
[((221, 526), 'typing.TypedDict', 'TypedDict', (['"""V1ISCSIVolumeSourceDict"""', "{'chapAuthDiscovery': bool, 'chapAuthSession': bool, 'fsType': str,\n 'initiatorName': str, 'iqn': str, 'iscsiInterface': str, 'lun': int,\n 'portals': List[str], 'readOnly': bool, 'secretRef':\n V1LocalObjectReferenceDict, 'targetPortal': str}"], {'total': '(False)'}), "('V1ISCSIVolumeSourceDict', {'chapAuthDiscovery': bool,\n 'chapAuthSession': bool, 'fsType': str, 'initiatorName': str, 'iqn':\n str, 'iscsiInterface': str, 'lun': int, 'portals': List[str],\n 'readOnly': bool, 'secretRef': V1LocalObjectReferenceDict,\n 'targetPortal': str}, total=False)\n", (230, 526), False, 'from typing import TypedDict, List\n')]
from pytpp.attributes._helper import IterableMeta, Attribute from pytpp.attributes.workflow import WorkflowAttributes class AdaptableWorkflowAttributes(WorkflowAttributes, metaclass=IterableMeta): __config_class__ = "Adaptable Workflow" adaptable_workflow_text_field_1 = Attribute('Adaptable Workflow Text Field 1', min_version='18.3') adaptable_workflow_text_field_10 = Attribute('Adaptable Workflow Text Field 10', min_version='18.3') adaptable_workflow_text_field_11 = Attribute('Adaptable Workflow Text Field 11', min_version='18.3') adaptable_workflow_text_field_12 = Attribute('Adaptable Workflow Text Field 12', min_version='18.3') adaptable_workflow_text_field_2 = Attribute('Adaptable Workflow Text Field 2', min_version='18.3') adaptable_workflow_text_field_3 = Attribute('Adaptable Workflow Text Field 3', min_version='18.3') adaptable_workflow_text_field_4 = Attribute('Adaptable Workflow Text Field 4', min_version='18.3') adaptable_workflow_text_field_5 = Attribute('Adaptable Workflow Text Field 5', min_version='18.3') adaptable_workflow_text_field_6 = Attribute('Adaptable Workflow Text Field 6', min_version='18.3') adaptable_workflow_text_field_7 = Attribute('Adaptable Workflow Text Field 7', min_version='18.3') adaptable_workflow_text_field_8 = Attribute('Adaptable Workflow Text Field 8', min_version='18.3') adaptable_workflow_text_field_9 = Attribute('Adaptable Workflow Text Field 9', min_version='18.3') credential = Attribute('Credential', min_version='18.3') log_debug = Attribute('Log Debug', min_version='19.3') oauth_token_application_id = Attribute('OAuth Token Application Id', min_version='21.3') oauth_token_credential = Attribute('OAuth Token Credential', min_version='21.3') oauth_token_scope = Attribute('OAuth Token Scope', min_version='21.3') powershell_script = Attribute('PowerShell Script', min_version='18.3') powershell_script_hash_vault_id = Attribute('PowerShell Script Hash Vault Id', min_version='19.1') script_execution_timeout = Attribute('Script Execution Timeout', min_version='20.2') secondary_credential = Attribute('Secondary Credential', min_version='18.3') service_address = Attribute('Service Address', min_version='18.3')
[ "pytpp.attributes._helper.Attribute" ]
[((275, 339), 'pytpp.attributes._helper.Attribute', 'Attribute', (['"""Adaptable Workflow Text Field 1"""'], {'min_version': '"""18.3"""'}), "('Adaptable Workflow Text Field 1', min_version='18.3')\n", (284, 339), False, 'from pytpp.attributes._helper import IterableMeta, Attribute\n'), ((376, 441), 'pytpp.attributes._helper.Attribute', 'Attribute', (['"""Adaptable Workflow Text Field 10"""'], {'min_version': '"""18.3"""'}), "('Adaptable Workflow Text Field 10', min_version='18.3')\n", (385, 441), False, 'from pytpp.attributes._helper import IterableMeta, Attribute\n'), ((478, 543), 'pytpp.attributes._helper.Attribute', 'Attribute', (['"""Adaptable Workflow Text Field 11"""'], {'min_version': '"""18.3"""'}), "('Adaptable Workflow Text Field 11', min_version='18.3')\n", (487, 543), False, 'from pytpp.attributes._helper import IterableMeta, Attribute\n'), ((580, 645), 'pytpp.attributes._helper.Attribute', 'Attribute', (['"""Adaptable Workflow Text Field 12"""'], {'min_version': '"""18.3"""'}), "('Adaptable Workflow Text Field 12', min_version='18.3')\n", (589, 645), False, 'from pytpp.attributes._helper import IterableMeta, Attribute\n'), ((681, 745), 'pytpp.attributes._helper.Attribute', 'Attribute', (['"""Adaptable Workflow Text Field 2"""'], {'min_version': '"""18.3"""'}), "('Adaptable Workflow Text Field 2', min_version='18.3')\n", (690, 745), False, 'from pytpp.attributes._helper import IterableMeta, Attribute\n'), ((781, 845), 'pytpp.attributes._helper.Attribute', 'Attribute', (['"""Adaptable Workflow Text Field 3"""'], {'min_version': '"""18.3"""'}), "('Adaptable Workflow Text Field 3', min_version='18.3')\n", (790, 845), False, 'from pytpp.attributes._helper import IterableMeta, Attribute\n'), ((881, 945), 'pytpp.attributes._helper.Attribute', 'Attribute', (['"""Adaptable Workflow Text Field 4"""'], {'min_version': '"""18.3"""'}), "('Adaptable Workflow Text Field 4', min_version='18.3')\n", (890, 945), False, 'from pytpp.attributes._helper import IterableMeta, Attribute\n'), ((981, 1045), 'pytpp.attributes._helper.Attribute', 'Attribute', (['"""Adaptable Workflow Text Field 5"""'], {'min_version': '"""18.3"""'}), "('Adaptable Workflow Text Field 5', min_version='18.3')\n", (990, 1045), False, 'from pytpp.attributes._helper import IterableMeta, Attribute\n'), ((1081, 1145), 'pytpp.attributes._helper.Attribute', 'Attribute', (['"""Adaptable Workflow Text Field 6"""'], {'min_version': '"""18.3"""'}), "('Adaptable Workflow Text Field 6', min_version='18.3')\n", (1090, 1145), False, 'from pytpp.attributes._helper import IterableMeta, Attribute\n'), ((1181, 1245), 'pytpp.attributes._helper.Attribute', 'Attribute', (['"""Adaptable Workflow Text Field 7"""'], {'min_version': '"""18.3"""'}), "('Adaptable Workflow Text Field 7', min_version='18.3')\n", (1190, 1245), False, 'from pytpp.attributes._helper import IterableMeta, Attribute\n'), ((1281, 1345), 'pytpp.attributes._helper.Attribute', 'Attribute', (['"""Adaptable Workflow Text Field 8"""'], {'min_version': '"""18.3"""'}), "('Adaptable Workflow Text Field 8', min_version='18.3')\n", (1290, 1345), False, 'from pytpp.attributes._helper import IterableMeta, Attribute\n'), ((1381, 1445), 'pytpp.attributes._helper.Attribute', 'Attribute', (['"""Adaptable Workflow Text Field 9"""'], {'min_version': '"""18.3"""'}), "('Adaptable Workflow Text Field 9', min_version='18.3')\n", (1390, 1445), False, 'from pytpp.attributes._helper import IterableMeta, Attribute\n'), ((1460, 1503), 'pytpp.attributes._helper.Attribute', 'Attribute', (['"""Credential"""'], {'min_version': '"""18.3"""'}), "('Credential', min_version='18.3')\n", (1469, 1503), False, 'from pytpp.attributes._helper import IterableMeta, Attribute\n'), ((1517, 1559), 'pytpp.attributes._helper.Attribute', 'Attribute', (['"""Log Debug"""'], {'min_version': '"""19.3"""'}), "('Log Debug', min_version='19.3')\n", (1526, 1559), False, 'from pytpp.attributes._helper import IterableMeta, Attribute\n'), ((1590, 1649), 'pytpp.attributes._helper.Attribute', 'Attribute', (['"""OAuth Token Application Id"""'], {'min_version': '"""21.3"""'}), "('OAuth Token Application Id', min_version='21.3')\n", (1599, 1649), False, 'from pytpp.attributes._helper import IterableMeta, Attribute\n'), ((1676, 1731), 'pytpp.attributes._helper.Attribute', 'Attribute', (['"""OAuth Token Credential"""'], {'min_version': '"""21.3"""'}), "('OAuth Token Credential', min_version='21.3')\n", (1685, 1731), False, 'from pytpp.attributes._helper import IterableMeta, Attribute\n'), ((1753, 1803), 'pytpp.attributes._helper.Attribute', 'Attribute', (['"""OAuth Token Scope"""'], {'min_version': '"""21.3"""'}), "('OAuth Token Scope', min_version='21.3')\n", (1762, 1803), False, 'from pytpp.attributes._helper import IterableMeta, Attribute\n'), ((1825, 1875), 'pytpp.attributes._helper.Attribute', 'Attribute', (['"""PowerShell Script"""'], {'min_version': '"""18.3"""'}), "('PowerShell Script', min_version='18.3')\n", (1834, 1875), False, 'from pytpp.attributes._helper import IterableMeta, Attribute\n'), ((1911, 1975), 'pytpp.attributes._helper.Attribute', 'Attribute', (['"""PowerShell Script Hash Vault Id"""'], {'min_version': '"""19.1"""'}), "('PowerShell Script Hash Vault Id', min_version='19.1')\n", (1920, 1975), False, 'from pytpp.attributes._helper import IterableMeta, Attribute\n'), ((2004, 2061), 'pytpp.attributes._helper.Attribute', 'Attribute', (['"""Script Execution Timeout"""'], {'min_version': '"""20.2"""'}), "('Script Execution Timeout', min_version='20.2')\n", (2013, 2061), False, 'from pytpp.attributes._helper import IterableMeta, Attribute\n'), ((2086, 2139), 'pytpp.attributes._helper.Attribute', 'Attribute', (['"""Secondary Credential"""'], {'min_version': '"""18.3"""'}), "('Secondary Credential', min_version='18.3')\n", (2095, 2139), False, 'from pytpp.attributes._helper import IterableMeta, Attribute\n'), ((2159, 2207), 'pytpp.attributes._helper.Attribute', 'Attribute', (['"""Service Address"""'], {'min_version': '"""18.3"""'}), "('Service Address', min_version='18.3')\n", (2168, 2207), False, 'from pytpp.attributes._helper import IterableMeta, Attribute\n')]
#!/usr/env/python """ Generalized Cyclic Multi-Pass Sweep-line Algorithm The cyclic multi-pass sweep-line algorithm, simply repeatedly sweeps across the same Timeline, until a specified number of passes has been completed or some signal is given to stop sweeping. Classes: - CycleSweep """ from typing import TypeVar, Union from sources.abstract import Subscriber, Timeline from .basesweep import Sweepline T = TypeVar('T') class CycleSweep(Sweepline[T]): """ The generalized cyclic multi-pass sweep-line algorithm. Generics: T: Objects type within the Timeline. Extends: Sweepline[T] """ ### Methods: Evaluation def evaluate(self, iterations: int = -1, *args, evparams_kw = {}, **kwargs): """ Execute the cyclic multi-pass sweep-line algorithm over the attached Timeline. Broadcast Events to the Observers. Args: iterations: The number of passes to run the algorithm over the Timeline. If negative value, the algorithm will continue until one of the subscriber calls the added stopiteration() method in the Events as a parameter. evparams_kw: Arguments for event.setparams(). args, kwargs: Arguments for timeline.events(). """ N = 0 status = {'stop': False} def stopiter(): status['stop'] = True while 0 > iterations or iterations > N: # cycle through each event in the timeline, one-pass for event in self.timeline.events(*args, **kwargs): event.setparams(iteration=N, stopiteration=stopiter, **evparams_kw) self.on_next(event) # stopiteration called if status['stop']: iterations = 0 # increment sweep counter N += 1 self.on_completed()
[ "typing.TypeVar" ]
[((419, 431), 'typing.TypeVar', 'TypeVar', (['"""T"""'], {}), "('T')\n", (426, 431), False, 'from typing import TypeVar, Union\n')]
import numpy import pandas as pd from boo.columns import SHORT_COLUMNS from boo.errors import UnclassifiableCodeError QUOTE_CHAR = '"' EMPTY = int(0) NUMERIC_COLUMNS = SHORT_COLUMNS.numeric def adjust_rub(df, cols=NUMERIC_COLUMNS): rows = (df.unit == "385") df.loc[rows, cols] = df.loc[rows, cols].multiply(1000) df.loc[rows, "unit"] = "384" rows = (df.unit == "383") df.loc[rows, cols] = df.loc[rows, cols].divide(1000).round(0).astype(int) df.loc[rows, "unit"] = "384" return df def dequote(name: str): """Split company *name* to organisation and title.""" # Warning: will not work well on company names with more than 4 quotechars parts = name.split(QUOTE_CHAR) org = parts[0].strip() cnt = name.count(QUOTE_CHAR) if cnt == 2: title = parts[1].strip() elif cnt > 2: title = QUOTE_CHAR.join(parts[1:]) else: title = name return org, title.strip() def replace_names(title: str): return title .replace( "ПУБЛИЧНОЕ АКЦИОНЕРНОЕ ОБЩЕСТВО", "ПАО") .replace( "ОТКРЫТОЕ АКЦИОНЕРНОЕ ОБЩЕСТВО", "ОАО") .replace( "АКЦИОНЕРНОЕ ОБЩЕСТВО ЭНЕРГЕТИКИ И ЭЛЕКТРИФИКАЦИИ", "AO энерго") .replace( "НЕФТЕПЕРЕРАБАТЫВАЮЩИЙ ЗАВОД", "НПЗ") .replace( "ГЕНЕРИРУЮЩАЯ КОМПАНИЯ ОПТОВОГО РЫНКА ЭЛЕКТРОЭНЕРГИИ", "ОГК") .replace( "ГОРНО-ОБОГАТИТЕЛЬНЫЙ КОМБИНАТ", "ГОК") def add_title(df): s_ = df.name.apply(dequote) df['org'] = s_.apply(lambda x: x[0]) df['title'] = s_.apply(lambda x: replace_names(x[1])) return df def rename_rows(df): RENAME_DICT = { '2460066195': "РусГидро", '4716016979': "ФСК ЕЭС", '7702038150': "Московский метрополитен", '7721632827': "Концерн Росэнергоатом", '7706664260': "Атомэнергопром", '7703683145': "Холдинг ВТБ Капитал АЙ БИ", '9102048801': "Черноморнефтегаз", '7736036626': "РИТЭК" } keys = RENAME_DICT.keys() ix = df.index.isin(keys) if not ix.any(): return df sub = df.loc[ix, 'title'] for k, v in RENAME_DICT.items(): if k in sub.index: sub.loc[k] = v df.loc[ix, 'title'] = sub return df def split_okved(code_string: str): """Get 3 levels of OKVED codes from *code_string*.""" if code_string.count(".") > 2: raise UnclassifiableCodeError(code_string) try: codes = [int(x) for x in code_string.split(".")] except ValueError: raise UnclassifiableCodeError(code_string) return codes + [0] * (3 - len(codes)) def add_okved_subcode(df): df['ok1'], df['ok2'], df['ok3'] = zip(*df.okved.apply(split_okved)) return df def fst(x): try: return int(x[0:2]) except TypeError: return 0 def add_region(df): df['region'] = df.inn.apply(fst) return df def more_columns(df): return add_okved_subcode(add_region(add_title(df))) def canonic_df(df): """Преобразовать данные внтури датафрейма: - Привести все строки к одинаковым единицам измерения (тыс. руб.) - Убрать неиспользуемые колонки (date_revised, report_type) - Новые колонки: * короткое название компании * три уровня кода ОКВЭД * регион (по ИНН) """ df_ = add_okved_subcode(add_region(add_title(df))) df_ = adjust_rub(df_) df_ = df_.set_index('inn') df_ = rename_rows(df_) return df_[canonic_columns()] def canonic_columns(numeric=SHORT_COLUMNS.numeric): return (['title', 'org', 'okpo', 'okopf', 'okfs', 'okved'] + ['unit'] + ['ok1', 'ok2', 'ok3', 'region'] + numeric) def is_numeric_column(name, numeric=SHORT_COLUMNS.numeric): return name in numeric def columns_typed_as_integer(numeric=SHORT_COLUMNS.numeric): return numeric + ['ok1', 'ok2', 'ok3', 'region'] def canonic_dtypes(): def switch(col): int_columns = columns_typed_as_integer() return numpy.int64 if (col in int_columns) else str result = {col: switch(col) for col in canonic_columns()} result['inn'] = str return result
[ "boo.errors.UnclassifiableCodeError" ]
[((2447, 2483), 'boo.errors.UnclassifiableCodeError', 'UnclassifiableCodeError', (['code_string'], {}), '(code_string)\n', (2470, 2483), False, 'from boo.errors import UnclassifiableCodeError\n'), ((2587, 2623), 'boo.errors.UnclassifiableCodeError', 'UnclassifiableCodeError', (['code_string'], {}), '(code_string)\n', (2610, 2623), False, 'from boo.errors import UnclassifiableCodeError\n')]
# 3rd party import pytest from coincidence.regressions import AdvancedFileRegressionFixture from consolekit.testing import CliRunner, Result from domdf_python_tools.paths import in_directory from github3.exceptions import UnprocessableEntity # this package from repo_helper_github.cli import new @pytest.mark.usefixtures("module_cassette") def test_create_repo(github_manager, temp_github_repo): with in_directory(temp_github_repo): github_manager.new() github_manager.github.repository("domdfcoding", "repo_helper_demo") with pytest.raises(UnprocessableEntity, match="422 Repository creation failed."): github_manager.new() @pytest.mark.usefixtures("betamax_github_session", "module_cassette") def test_via_cli( temp_github_repo, advanced_file_regression: AdvancedFileRegressionFixture, github_manager, ): with in_directory(temp_github_repo): runner = CliRunner() result: Result = runner.invoke(new) assert result.exit_code == 0 result.check_stdout(advanced_file_regression, extension=".md") # Check the repository now exists github_manager.github.repository("domdfcoding", "repo_helper_demo") with pytest.raises(UnprocessableEntity, match="422 Repository creation failed."): github_manager.new()
[ "domdf_python_tools.paths.in_directory", "pytest.raises", "pytest.mark.usefixtures", "consolekit.testing.CliRunner" ]
[((300, 342), 'pytest.mark.usefixtures', 'pytest.mark.usefixtures', (['"""module_cassette"""'], {}), "('module_cassette')\n", (323, 342), False, 'import pytest\n'), ((643, 711), 'pytest.mark.usefixtures', 'pytest.mark.usefixtures', (['"""betamax_github_session"""', '"""module_cassette"""'], {}), "('betamax_github_session', 'module_cassette')\n", (666, 711), False, 'import pytest\n'), ((405, 435), 'domdf_python_tools.paths.in_directory', 'in_directory', (['temp_github_repo'], {}), '(temp_github_repo)\n', (417, 435), False, 'from domdf_python_tools.paths import in_directory\n'), ((838, 868), 'domdf_python_tools.paths.in_directory', 'in_directory', (['temp_github_repo'], {}), '(temp_github_repo)\n', (850, 868), False, 'from domdf_python_tools.paths import in_directory\n'), ((881, 892), 'consolekit.testing.CliRunner', 'CliRunner', ([], {}), '()\n', (890, 892), False, 'from consolekit.testing import CliRunner, Result\n'), ((1138, 1213), 'pytest.raises', 'pytest.raises', (['UnprocessableEntity'], {'match': '"""422 Repository creation failed."""'}), "(UnprocessableEntity, match='422 Repository creation failed.')\n", (1151, 1213), False, 'import pytest\n'), ((539, 614), 'pytest.raises', 'pytest.raises', (['UnprocessableEntity'], {'match': '"""422 Repository creation failed."""'}), "(UnprocessableEntity, match='422 Repository creation failed.')\n", (552, 614), False, 'import pytest\n')]
import numpy as np import os import tensorflow as tf EPS = 1e-8 def placeholder(dim=None): return tf.placeholder(dtype=tf.float32, shape=(None,dim) if dim else (None,)) def placeholders(*args): return [placeholder(dim) for dim in args] def mlp(x, hidden_sizes=(32,), activation=tf.tanh, output_activation=None): init_fn = tf.keras.initializers.Orthogonal(1.0) for h in hidden_sizes[:-1]: x = tf.layers.dense(x, units=h, activation=activation, kernel_initializer=init_fn) return tf.layers.dense(x, units=hidden_sizes[-1], activation=output_activation, kernel_initializer=init_fn) def get_vars(scope): return [x for x in tf.global_variables() if scope in x.name] def count_vars(scope): v = get_vars(scope) return sum([np.prod(var.shape.as_list()) for var in v]) def gaussian_likelihood(x, mu, log_std): pre_sum = -0.5 * (((x-mu)/(tf.exp(log_std)+EPS))**2 + 2*log_std + np.log(2*np.pi)) return tf.reduce_sum(pre_sum, axis=1) def clip_but_pass_gradient(x, l=-1., u=1.): clip_up = tf.cast(x > u, tf.float32) clip_low = tf.cast(x < l, tf.float32) return x + tf.stop_gradient((u - x)*clip_up + (l - x)*clip_low) """ Policies """ def gumbel_policy(x, act_dim, hidden_sizes, activation): # policy network outputs net = mlp(x, list(hidden_sizes), activation, activation) logits = tf.layers.dense(net, act_dim, activation='linear') # action and log action probabilites (log_softmax covers numerical problems) action_probs = tf.nn.softmax([logits], axis=-1) log_action_probs = tf.nn.log_softmax([logits], axis=-1) # policy with no noise mu = tf.argmax(logits, axis=-1) # add gumbel noise to action distributions temperature = tf.convert_to_tensor(1.0) # 0 --> argmax, inf --> uniform uniform_noise = tf.random_uniform(shape=tf.shape(logits), minval=np.finfo(np.float32).tiny, # (0,1) range maxval=1.) gumbel_noise = -tf.log(-tf.log(uniform_noise)) noisy_logits = logits + gumbel_noise pi_dist = tf.nn.softmax(noisy_logits / temperature[..., tf.newaxis]) # dont use tf.dist.relaxedCategorical for log_prob, seems to give wrong results logp_pi = -tf.reduce_sum(-pi_dist * tf.nn.log_softmax(logits, axis=-1), axis=1) return mu, pi_dist, logp_pi """ Actor-Critics """ def a_out_mlp_actor_critic(x, a, hidden_sizes=[400,300], activation=tf.nn.relu, policy=gumbel_policy): act_dim = a.shape.as_list()[-1] with tf.variable_scope('pi'): mu, pi_dist, logp_pi = policy(x, act_dim, hidden_sizes, activation) # vfs with tf.variable_scope('q1'): q1 = mlp(x, list(hidden_sizes)+[act_dim], activation, None) q1_a = tf.reduce_sum(tf.multiply(q1, a), axis=1) with tf.variable_scope('q2'): q2 = mlp(x, list(hidden_sizes)+[act_dim], activation, None) q2_a = tf.reduce_sum(tf.multiply(q2, a), axis=1) return mu, pi_dist, logp_pi, q1_a, q2_a def a_in_mlp_actor_critic(x, a, hidden_sizes=[400,300], activation=tf.nn.relu, policy=gumbel_policy): act_dim = a.shape.as_list()[-1] with tf.variable_scope('pi'): mu, pi_dist, logp_pi = policy(x, act_dim, hidden_sizes, activation) # vfs with tf.variable_scope('q1'): q1_a = tf.squeeze(mlp(tf.concat([x,a], axis=-1), list(hidden_sizes)+[1], activation, None), axis=1) with tf.variable_scope('q2'): q2_a = tf.squeeze(mlp(tf.concat([x,a], axis=-1), list(hidden_sizes)+[1], activation, None), axis=1) return mu, pi_dist, logp_pi, q1_a, q2_a
[ "tensorflow.shape", "tensorflow.reduce_sum", "numpy.log", "tensorflow.multiply", "tensorflow.nn.softmax", "tensorflow.keras.initializers.Orthogonal", "tensorflow.cast", "tensorflow.log", "tensorflow.placeholder", "tensorflow.concat", "tensorflow.convert_to_tensor", "tensorflow.variable_scope",...
[((104, 175), 'tensorflow.placeholder', 'tf.placeholder', ([], {'dtype': 'tf.float32', 'shape': '((None, dim) if dim else (None,))'}), '(dtype=tf.float32, shape=(None, dim) if dim else (None,))\n', (118, 175), True, 'import tensorflow as tf\n'), ((338, 375), 'tensorflow.keras.initializers.Orthogonal', 'tf.keras.initializers.Orthogonal', (['(1.0)'], {}), '(1.0)\n', (370, 375), True, 'import tensorflow as tf\n'), ((510, 614), 'tensorflow.layers.dense', 'tf.layers.dense', (['x'], {'units': 'hidden_sizes[-1]', 'activation': 'output_activation', 'kernel_initializer': 'init_fn'}), '(x, units=hidden_sizes[-1], activation=output_activation,\n kernel_initializer=init_fn)\n', (525, 614), True, 'import tensorflow as tf\n'), ((946, 976), 'tensorflow.reduce_sum', 'tf.reduce_sum', (['pre_sum'], {'axis': '(1)'}), '(pre_sum, axis=1)\n', (959, 976), True, 'import tensorflow as tf\n'), ((1036, 1062), 'tensorflow.cast', 'tf.cast', (['(x > u)', 'tf.float32'], {}), '(x > u, tf.float32)\n', (1043, 1062), True, 'import tensorflow as tf\n'), ((1078, 1104), 'tensorflow.cast', 'tf.cast', (['(x < l)', 'tf.float32'], {}), '(x < l, tf.float32)\n', (1085, 1104), True, 'import tensorflow as tf\n'), ((1353, 1403), 'tensorflow.layers.dense', 'tf.layers.dense', (['net', 'act_dim'], {'activation': '"""linear"""'}), "(net, act_dim, activation='linear')\n", (1368, 1403), True, 'import tensorflow as tf\n'), ((1505, 1537), 'tensorflow.nn.softmax', 'tf.nn.softmax', (['[logits]'], {'axis': '(-1)'}), '([logits], axis=-1)\n', (1518, 1537), True, 'import tensorflow as tf\n'), ((1561, 1597), 'tensorflow.nn.log_softmax', 'tf.nn.log_softmax', (['[logits]'], {'axis': '(-1)'}), '([logits], axis=-1)\n', (1578, 1597), True, 'import tensorflow as tf\n'), ((1635, 1661), 'tensorflow.argmax', 'tf.argmax', (['logits'], {'axis': '(-1)'}), '(logits, axis=-1)\n', (1644, 1661), True, 'import tensorflow as tf\n'), ((1728, 1753), 'tensorflow.convert_to_tensor', 'tf.convert_to_tensor', (['(1.0)'], {}), '(1.0)\n', (1748, 1753), True, 'import tensorflow as tf\n'), ((2091, 2149), 'tensorflow.nn.softmax', 'tf.nn.softmax', (['(noisy_logits / temperature[..., tf.newaxis])'], {}), '(noisy_logits / temperature[..., tf.newaxis])\n', (2104, 2149), True, 'import tensorflow as tf\n'), ((420, 498), 'tensorflow.layers.dense', 'tf.layers.dense', (['x'], {'units': 'h', 'activation': 'activation', 'kernel_initializer': 'init_fn'}), '(x, units=h, activation=activation, kernel_initializer=init_fn)\n', (435, 498), True, 'import tensorflow as tf\n'), ((1120, 1176), 'tensorflow.stop_gradient', 'tf.stop_gradient', (['((u - x) * clip_up + (l - x) * clip_low)'], {}), '((u - x) * clip_up + (l - x) * clip_low)\n', (1136, 1176), True, 'import tensorflow as tf\n'), ((2525, 2548), 'tensorflow.variable_scope', 'tf.variable_scope', (['"""pi"""'], {}), "('pi')\n", (2542, 2548), True, 'import tensorflow as tf\n'), ((2646, 2669), 'tensorflow.variable_scope', 'tf.variable_scope', (['"""q1"""'], {}), "('q1')\n", (2663, 2669), True, 'import tensorflow as tf\n'), ((2809, 2832), 'tensorflow.variable_scope', 'tf.variable_scope', (['"""q2"""'], {}), "('q2')\n", (2826, 2832), True, 'import tensorflow as tf\n'), ((3159, 3182), 'tensorflow.variable_scope', 'tf.variable_scope', (['"""pi"""'], {}), "('pi')\n", (3176, 3182), True, 'import tensorflow as tf\n'), ((3280, 3303), 'tensorflow.variable_scope', 'tf.variable_scope', (['"""q1"""'], {}), "('q1')\n", (3297, 3303), True, 'import tensorflow as tf\n'), ((3424, 3447), 'tensorflow.variable_scope', 'tf.variable_scope', (['"""q2"""'], {}), "('q2')\n", (3441, 3447), True, 'import tensorflow as tf\n'), ((656, 677), 'tensorflow.global_variables', 'tf.global_variables', ([], {}), '()\n', (675, 677), True, 'import tensorflow as tf\n'), ((918, 935), 'numpy.log', 'np.log', (['(2 * np.pi)'], {}), '(2 * np.pi)\n', (924, 935), True, 'import numpy as np\n'), ((1830, 1846), 'tensorflow.shape', 'tf.shape', (['logits'], {}), '(logits)\n', (1838, 1846), True, 'import tensorflow as tf\n'), ((2772, 2790), 'tensorflow.multiply', 'tf.multiply', (['q1', 'a'], {}), '(q1, a)\n', (2783, 2790), True, 'import tensorflow as tf\n'), ((2935, 2953), 'tensorflow.multiply', 'tf.multiply', (['q2', 'a'], {}), '(q2, a)\n', (2946, 2953), True, 'import tensorflow as tf\n'), ((1893, 1913), 'numpy.finfo', 'np.finfo', (['np.float32'], {}), '(np.float32)\n', (1901, 1913), True, 'import numpy as np\n'), ((2012, 2033), 'tensorflow.log', 'tf.log', (['uniform_noise'], {}), '(uniform_noise)\n', (2018, 2033), True, 'import tensorflow as tf\n'), ((2275, 2309), 'tensorflow.nn.log_softmax', 'tf.nn.log_softmax', (['logits'], {'axis': '(-1)'}), '(logits, axis=-1)\n', (2292, 2309), True, 'import tensorflow as tf\n'), ((3336, 3362), 'tensorflow.concat', 'tf.concat', (['[x, a]'], {'axis': '(-1)'}), '([x, a], axis=-1)\n', (3345, 3362), True, 'import tensorflow as tf\n'), ((3480, 3506), 'tensorflow.concat', 'tf.concat', (['[x, a]'], {'axis': '(-1)'}), '([x, a], axis=-1)\n', (3489, 3506), True, 'import tensorflow as tf\n'), ((879, 894), 'tensorflow.exp', 'tf.exp', (['log_std'], {}), '(log_std)\n', (885, 894), True, 'import tensorflow as tf\n')]
from objects.lighting.lighting_interactions import SwitchLightImmediateInteraction from objects.object_state_utils import ObjectStateHelper import sims4 logger = sims4.log.Logger('LightingAndObjectState', default_owner='mkartika') class SwitchLightAndStateImmediateInteraction(SwitchLightImmediateInteraction): INSTANCE_TUNABLES = {'state_settings': ObjectStateHelper.TunableFactory(description='\n Find objects in the same room or lot based on the tags and \n set state to the desired state.\n ')} def _run_interaction_gen(self, timeline): yield from super()._run_interaction_gen(timeline) self.state_settings.execute_helper(self)
[ "objects.object_state_utils.ObjectStateHelper.TunableFactory", "sims4.log.Logger" ]
[((162, 230), 'sims4.log.Logger', 'sims4.log.Logger', (['"""LightingAndObjectState"""'], {'default_owner': '"""mkartika"""'}), "('LightingAndObjectState', default_owner='mkartika')\n", (178, 230), False, 'import sims4\n'), ((355, 546), 'objects.object_state_utils.ObjectStateHelper.TunableFactory', 'ObjectStateHelper.TunableFactory', ([], {'description': '"""\n Find objects in the same room or lot based on the tags and \n set state to the desired state.\n """'}), '(description=\n """\n Find objects in the same room or lot based on the tags and \n set state to the desired state.\n """\n )\n', (387, 546), False, 'from objects.object_state_utils import ObjectStateHelper\n')]
import mock def test_setup(GPIO, spidev): from unicornhatmini import UnicornHATMini unicornhatmini = UnicornHATMini() spidev.SpiDev.assert_has_calls(( mock.call(0, 0), mock.call(0, 1) ), any_order=True) GPIO.setwarnings.assert_called_once_with(False) GPIO.setmode.assert_called_once_with(GPIO.BCM) del unicornhatmini def test_shutdown(GPIO, spidev, atexit): from unicornhatmini import UnicornHATMini unicornhatmini = UnicornHATMini() atexit.register.assert_called_once_with(unicornhatmini._exit) unicornhatmini._exit()
[ "mock.call", "unicornhatmini.UnicornHATMini" ]
[((111, 127), 'unicornhatmini.UnicornHATMini', 'UnicornHATMini', ([], {}), '()\n', (125, 127), False, 'from unicornhatmini import UnicornHATMini\n'), ((476, 492), 'unicornhatmini.UnicornHATMini', 'UnicornHATMini', ([], {}), '()\n', (490, 492), False, 'from unicornhatmini import UnicornHATMini\n'), ((174, 189), 'mock.call', 'mock.call', (['(0)', '(0)'], {}), '(0, 0)\n', (183, 189), False, 'import mock\n'), ((199, 214), 'mock.call', 'mock.call', (['(0)', '(1)'], {}), '(0, 1)\n', (208, 214), False, 'import mock\n')]
import time from ..base import order as od from .api import BybitApi class BybitOrderManager(od.OrderManagerBase): def __init__(self, api, ws=None, retention=60): super().__init__(api, ws, retention) self.ws.subscribe('execution', self.__on_events, True) self.ws.subscribe('position', self.__on_events, True) self.ws.subscribe('order', self.__on_events, True) def _generate_order_object(self, e): info = e.info if e.type != od.EVENT_OPEN: self.log.warning(f'event for unknown order: {e}') return None api = BybitApi.ccxt_instance() symbol = api.markets_by_id[info['symbol']]['symbol'] return od.Order( symbol, info['order_type'].lower(), info['side'].lower(), info['qty'], float(info['price'])) def __on_events(self, msg): topic = msg['topic'] for e in msg['data']: oe = od.OrderEvent() oe.info = e oe.ts = time.time() if topic == 'order': oe.id = e['order_id'] st = e['order_status'] if st == 'New': oe.type = od.EVENT_OPEN elif st == 'Filled': oe.type = od.EVENT_CLOSE elif st in ['Cancelled', 'Rejected']: oe.type = od.EVENT_CANCEL else: # ignore(PartiallyFilled, Created, PendingCancel) continue elif topic == 'execution': oe.type = od.EVENT_EXECUTION oe.id = e['order_id'] oe.price = float(e['price']) size = e['exec_qty'] oe.size = -size if e['side'] == 'Sell' else size oe.fee = float(e['exec_fee']) * size elif topic == 'position': break else: assert False self._handle_order_event(oe) class BybitPositionGroup(od.PositionGroupBase): INVERSE = True class BybitOrderGroup(od.OrderGroupBase): PositionGroup = BybitPositionGroup class BybitOrderGroupManager(od.OrderGroupManagerBase): OrderGroup = BybitOrderGroup # Future class BybitUsdtOrderManager(BybitOrderManager): pass class BybitUsdtPositionGroup(BybitPositionGroup): INVERSE = False class BybitUsdtOrderGroup(BybitOrderGroup): PositionGroup = BybitUsdtPositionGroup class BybitUsdtOrderGroupManager(BybitOrderGroupManager): OrderGroup = BybitUsdtOrderGroup
[ "time.time" ]
[((998, 1009), 'time.time', 'time.time', ([], {}), '()\n', (1007, 1009), False, 'import time\n')]
import os import wget import paddle from .tokenizer import Tokenizer from .model import CLIP from paddle.vision.transforms import Compose, Resize, CenterCrop, ToTensor, Normalize tokenizer = Tokenizer() def get_transforms(image_resolution): transforms = Compose([ Resize(image_resolution, interpolation='bicubic'), CenterCrop(image_resolution), lambda image: image.convert("RGB"), ToTensor(), Normalize((0.48145466, 0.4578275, 0.40821073), (0.26862954, 0.26130258, 0.27577711)), ]) return transforms def clip_rn50(): model = CLIP( embed_dim=1024, image_resolution=224, vision_layers=(3, 4, 6, 3), vision_width=64, vision_patch_size=None, context_length=77, vocab_size=49408, transformer_width=512, transformer_heads=8, transformer_layers=12 ) return model, get_transforms(224) def clip_rn101(): model = CLIP( embed_dim=512, image_resolution=224, vision_layers=(3, 4, 23, 3), vision_width=64, vision_patch_size=None, context_length=77, vocab_size=49408, transformer_width=512, transformer_heads=8, transformer_layers=12 ) return model, get_transforms(224) def clip_rn50x4(): model = CLIP( embed_dim=640, image_resolution=288, vision_layers=(4, 6, 10, 6), vision_width=80, vision_patch_size=None, context_length=77, vocab_size=49408, transformer_width=640, transformer_heads=10, transformer_layers=12 ) return model, get_transforms(288) def clip_vit_b_32(): model = CLIP( embed_dim=512, image_resolution=224, vision_layers=12, vision_width=768, vision_patch_size=32, context_length=77, vocab_size=49408, transformer_width=512, transformer_heads=8, transformer_layers=12 ) return model, get_transforms(224) def tokenize(texts, context_length=77): """ Returns the tokenized representation of given input string(s) Parameters ---------- texts : Union[str, List[str]] An input string or a list of input strings to tokenize context_length : int The context length to use; all CLIP models use 77 as the context length Returns ------- A two-dimensional tensor containing the resulting tokens, shape = [number of input strings, context_length] """ if isinstance(texts, str): texts = [texts] sot_token = tokenizer.encoder["<|startoftext|>"] eot_token = tokenizer.encoder["<|endoftext|>"] all_tokens = [[sot_token] + tokenizer.encode(text) + [eot_token] for text in texts] result = paddle.zeros((len(all_tokens), context_length), dtype='int64') for i, tokens in enumerate(all_tokens): if len(tokens) > context_length: raise RuntimeError( f"Input {texts[i]} is too long for context length {context_length}") result[i, :len(tokens)] = paddle.to_tensor(tokens) return result model_dict = { 'RN50': [clip_rn50, r'https://bj.bcebos.com/v1/ai-studio-online/6ffc89246e974a809e6e4b40fdb58063a112a0153e674dae8ed5b6dfe5d46d86?responseContentDisposition=attachment%3B%20filename%3DRN50.pdparams', 'RN50.pdparams'], 'RN50x4': [clip_rn50x4, r'https://bj.bcebos.com/v1/ai-studio-online/9f874e0174da48ffbd7c17e77b1fb278632620a9995e476ba873e334caec9037?responseContentDisposition=attachment%3B%20filename%3DRN50x4.pdparams', 'RN50x4.pdparams'], 'RN101': [clip_rn101, r'https://bj.bcebos.com/v1/ai-studio-online/484592d98c584785bc8f6f9f7badbf4a9fb7a96f6102470697ed974e8eeee2a9?responseContentDisposition=attachment%3B%20filename%3DRN101.pdparams', 'RN101.pdparams'], 'ViT_B_32': [clip_vit_b_32, r'https://bj.bcebos.com/v1/ai-studio-online/eb5e4dbf1ec142caa003a27cefd510ef46a8a6c3932a4d60bfecb3f3ab746c02?responseContentDisposition=attachment%3B%20filename%3DViT-B-32.pdparams', 'ViT-B-32.pdparams'] } def load_model(model_name, pretrained=False): model_fn, url, file_name = model_dict[model_name] model, transforms = model_fn() if pretrained: model_path = os.path.join('pretrained_models', file_name) if not os.path.isfile(model_path): if not os.path.exists('pretrained_models'): os.mkdir('pretrained_models') wget.download(url, out=model_path) params = paddle.load(model_path) model.set_dict(params) model.eval() return model, transforms
[ "paddle.vision.transforms.ToTensor", "wget.download", "os.path.exists", "paddle.vision.transforms.Normalize", "paddle.vision.transforms.CenterCrop", "os.path.join", "os.path.isfile", "paddle.to_tensor", "os.mkdir", "paddle.load", "paddle.vision.transforms.Resize" ]
[((3125, 3149), 'paddle.to_tensor', 'paddle.to_tensor', (['tokens'], {}), '(tokens)\n', (3141, 3149), False, 'import paddle\n'), ((4277, 4321), 'os.path.join', 'os.path.join', (['"""pretrained_models"""', 'file_name'], {}), "('pretrained_models', file_name)\n", (4289, 4321), False, 'import os\n'), ((4531, 4554), 'paddle.load', 'paddle.load', (['model_path'], {}), '(model_path)\n', (4542, 4554), False, 'import paddle\n'), ((280, 329), 'paddle.vision.transforms.Resize', 'Resize', (['image_resolution'], {'interpolation': '"""bicubic"""'}), "(image_resolution, interpolation='bicubic')\n", (286, 329), False, 'from paddle.vision.transforms import Compose, Resize, CenterCrop, ToTensor, Normalize\n'), ((339, 367), 'paddle.vision.transforms.CenterCrop', 'CenterCrop', (['image_resolution'], {}), '(image_resolution)\n', (349, 367), False, 'from paddle.vision.transforms import Compose, Resize, CenterCrop, ToTensor, Normalize\n'), ((421, 431), 'paddle.vision.transforms.ToTensor', 'ToTensor', ([], {}), '()\n', (429, 431), False, 'from paddle.vision.transforms import Compose, Resize, CenterCrop, ToTensor, Normalize\n'), ((441, 530), 'paddle.vision.transforms.Normalize', 'Normalize', (['(0.48145466, 0.4578275, 0.40821073)', '(0.26862954, 0.26130258, 0.27577711)'], {}), '((0.48145466, 0.4578275, 0.40821073), (0.26862954, 0.26130258, \n 0.27577711))\n', (450, 530), False, 'from paddle.vision.transforms import Compose, Resize, CenterCrop, ToTensor, Normalize\n'), ((4337, 4363), 'os.path.isfile', 'os.path.isfile', (['model_path'], {}), '(model_path)\n', (4351, 4363), False, 'import os\n'), ((4479, 4513), 'wget.download', 'wget.download', (['url'], {'out': 'model_path'}), '(url, out=model_path)\n', (4492, 4513), False, 'import wget\n'), ((4384, 4419), 'os.path.exists', 'os.path.exists', (['"""pretrained_models"""'], {}), "('pretrained_models')\n", (4398, 4419), False, 'import os\n'), ((4437, 4466), 'os.mkdir', 'os.mkdir', (['"""pretrained_models"""'], {}), "('pretrained_models')\n", (4445, 4466), False, 'import os\n')]
# coding=utf-8 # *** WARNING: this file was generated by the Pulumi Terraform Bridge (tfgen) Tool. *** # *** Do not edit by hand unless you're certain you know what you are doing! *** import warnings import pulumi import pulumi.runtime from typing import Any, Mapping, Optional, Sequence, Union, overload from . import _utilities __all__ = [ 'AlbBackendGroupGrpcBackendArgs', 'AlbBackendGroupGrpcBackendHealthcheckArgs', 'AlbBackendGroupGrpcBackendHealthcheckGrpcHealthcheckArgs', 'AlbBackendGroupGrpcBackendHealthcheckHttpHealthcheckArgs', 'AlbBackendGroupGrpcBackendHealthcheckStreamHealthcheckArgs', 'AlbBackendGroupGrpcBackendLoadBalancingConfigArgs', 'AlbBackendGroupGrpcBackendTlsArgs', 'AlbBackendGroupGrpcBackendTlsValidationContextArgs', 'AlbBackendGroupHttpBackendArgs', 'AlbBackendGroupHttpBackendHealthcheckArgs', 'AlbBackendGroupHttpBackendHealthcheckGrpcHealthcheckArgs', 'AlbBackendGroupHttpBackendHealthcheckHttpHealthcheckArgs', 'AlbBackendGroupHttpBackendHealthcheckStreamHealthcheckArgs', 'AlbBackendGroupHttpBackendLoadBalancingConfigArgs', 'AlbBackendGroupHttpBackendTlsArgs', 'AlbBackendGroupHttpBackendTlsValidationContextArgs', 'AlbLoadBalancerAllocationPolicyArgs', 'AlbLoadBalancerAllocationPolicyLocationArgs', 'AlbLoadBalancerListenerArgs', 'AlbLoadBalancerListenerEndpointArgs', 'AlbLoadBalancerListenerEndpointAddressArgs', 'AlbLoadBalancerListenerEndpointAddressExternalIpv4AddressArgs', 'AlbLoadBalancerListenerEndpointAddressExternalIpv6AddressArgs', 'AlbLoadBalancerListenerEndpointAddressInternalIpv4AddressArgs', 'AlbLoadBalancerListenerHttpArgs', 'AlbLoadBalancerListenerHttpHandlerArgs', 'AlbLoadBalancerListenerHttpHandlerHttp2OptionsArgs', 'AlbLoadBalancerListenerHttpRedirectsArgs', 'AlbLoadBalancerListenerTlsArgs', 'AlbLoadBalancerListenerTlsDefaultHandlerArgs', 'AlbLoadBalancerListenerTlsDefaultHandlerHttpHandlerArgs', 'AlbLoadBalancerListenerTlsDefaultHandlerHttpHandlerHttp2OptionsArgs', 'AlbLoadBalancerListenerTlsSniHandlerArgs', 'AlbLoadBalancerListenerTlsSniHandlerHandlerArgs', 'AlbLoadBalancerListenerTlsSniHandlerHandlerHttpHandlerArgs', 'AlbLoadBalancerListenerTlsSniHandlerHandlerHttpHandlerHttp2OptionsArgs', 'AlbTargetGroupTargetArgs', 'AlbVirtualHostModifyRequestHeaderArgs', 'AlbVirtualHostModifyResponseHeaderArgs', 'AlbVirtualHostRouteArgs', 'AlbVirtualHostRouteGrpcRouteArgs', 'AlbVirtualHostRouteGrpcRouteGrpcMatchArgs', 'AlbVirtualHostRouteGrpcRouteGrpcMatchFqmnArgs', 'AlbVirtualHostRouteGrpcRouteGrpcRouteActionArgs', 'AlbVirtualHostRouteGrpcRouteGrpcStatusResponseActionArgs', 'AlbVirtualHostRouteHttpRouteArgs', 'AlbVirtualHostRouteHttpRouteDirectResponseActionArgs', 'AlbVirtualHostRouteHttpRouteHttpMatchArgs', 'AlbVirtualHostRouteHttpRouteHttpMatchPathArgs', 'AlbVirtualHostRouteHttpRouteHttpRouteActionArgs', 'AlbVirtualHostRouteHttpRouteRedirectActionArgs', 'ComputeDiskDiskPlacementPolicyArgs', 'ComputeInstanceBootDiskArgs', 'ComputeInstanceBootDiskInitializeParamsArgs', 'ComputeInstanceGroupAllocationPolicyArgs', 'ComputeInstanceGroupApplicationLoadBalancerArgs', 'ComputeInstanceGroupDeployPolicyArgs', 'ComputeInstanceGroupHealthCheckArgs', 'ComputeInstanceGroupHealthCheckHttpOptionArgs', 'ComputeInstanceGroupHealthCheckTcpOptionArgs', 'ComputeInstanceGroupInstanceArgs', 'ComputeInstanceGroupInstanceNetworkInterfaceArgs', 'ComputeInstanceGroupInstanceTemplateArgs', 'ComputeInstanceGroupInstanceTemplateBootDiskArgs', 'ComputeInstanceGroupInstanceTemplateBootDiskInitializeParamsArgs', 'ComputeInstanceGroupInstanceTemplateNetworkInterfaceArgs', 'ComputeInstanceGroupInstanceTemplateNetworkInterfaceDnsRecordArgs', 'ComputeInstanceGroupInstanceTemplateNetworkInterfaceIpv6DnsRecordArgs', 'ComputeInstanceGroupInstanceTemplateNetworkInterfaceNatDnsRecordArgs', 'ComputeInstanceGroupInstanceTemplateNetworkSettingArgs', 'ComputeInstanceGroupInstanceTemplatePlacementPolicyArgs', 'ComputeInstanceGroupInstanceTemplateResourcesArgs', 'ComputeInstanceGroupInstanceTemplateSchedulingPolicyArgs', 'ComputeInstanceGroupInstanceTemplateSecondaryDiskArgs', 'ComputeInstanceGroupInstanceTemplateSecondaryDiskInitializeParamsArgs', 'ComputeInstanceGroupLoadBalancerArgs', 'ComputeInstanceGroupScalePolicyArgs', 'ComputeInstanceGroupScalePolicyAutoScaleArgs', 'ComputeInstanceGroupScalePolicyAutoScaleCustomRuleArgs', 'ComputeInstanceGroupScalePolicyFixedScaleArgs', 'ComputeInstanceGroupScalePolicyTestAutoScaleArgs', 'ComputeInstanceGroupScalePolicyTestAutoScaleCustomRuleArgs', 'ComputeInstanceNetworkInterfaceArgs', 'ComputeInstanceNetworkInterfaceDnsRecordArgs', 'ComputeInstanceNetworkInterfaceIpv6DnsRecordArgs', 'ComputeInstanceNetworkInterfaceNatDnsRecordArgs', 'ComputeInstancePlacementPolicyArgs', 'ComputeInstanceResourcesArgs', 'ComputeInstanceSchedulingPolicyArgs', 'ComputeInstanceSecondaryDiskArgs', 'DataprocClusterClusterConfigArgs', 'DataprocClusterClusterConfigHadoopArgs', 'DataprocClusterClusterConfigSubclusterSpecArgs', 'DataprocClusterClusterConfigSubclusterSpecAutoscalingConfigArgs', 'DataprocClusterClusterConfigSubclusterSpecResourcesArgs', 'FunctionContentArgs', 'FunctionPackageArgs', 'FunctionScalingPolicyPolicyArgs', 'FunctionTriggerDlqArgs', 'FunctionTriggerFunctionArgs', 'FunctionTriggerIotArgs', 'FunctionTriggerLogGroupArgs', 'FunctionTriggerLoggingArgs', 'FunctionTriggerMessageQueueArgs', 'FunctionTriggerObjectStorageArgs', 'FunctionTriggerTimerArgs', 'KubernetesClusterKmsProviderArgs', 'KubernetesClusterMasterArgs', 'KubernetesClusterMasterMaintenancePolicyArgs', 'KubernetesClusterMasterMaintenancePolicyMaintenanceWindowArgs', 'KubernetesClusterMasterRegionalArgs', 'KubernetesClusterMasterRegionalLocationArgs', 'KubernetesClusterMasterVersionInfoArgs', 'KubernetesClusterMasterZonalArgs', 'KubernetesClusterNetworkImplementationArgs', 'KubernetesClusterNetworkImplementationCiliumArgs', 'KubernetesNodeGroupAllocationPolicyArgs', 'KubernetesNodeGroupAllocationPolicyLocationArgs', 'KubernetesNodeGroupDeployPolicyArgs', 'KubernetesNodeGroupInstanceTemplateArgs', 'KubernetesNodeGroupInstanceTemplateBootDiskArgs', 'KubernetesNodeGroupInstanceTemplateNetworkInterfaceArgs', 'KubernetesNodeGroupInstanceTemplatePlacementPolicyArgs', 'KubernetesNodeGroupInstanceTemplateResourcesArgs', 'KubernetesNodeGroupInstanceTemplateSchedulingPolicyArgs', 'KubernetesNodeGroupMaintenancePolicyArgs', 'KubernetesNodeGroupMaintenancePolicyMaintenanceWindowArgs', 'KubernetesNodeGroupScalePolicyArgs', 'KubernetesNodeGroupScalePolicyAutoScaleArgs', 'KubernetesNodeGroupScalePolicyFixedScaleArgs', 'KubernetesNodeGroupVersionInfoArgs', 'LbNetworkLoadBalancerAttachedTargetGroupArgs', 'LbNetworkLoadBalancerAttachedTargetGroupHealthcheckArgs', 'LbNetworkLoadBalancerAttachedTargetGroupHealthcheckHttpOptionsArgs', 'LbNetworkLoadBalancerAttachedTargetGroupHealthcheckTcpOptionsArgs', 'LbNetworkLoadBalancerListenerArgs', 'LbNetworkLoadBalancerListenerExternalAddressSpecArgs', 'LbNetworkLoadBalancerListenerInternalAddressSpecArgs', 'LbTargetGroupTargetArgs', 'MdbClickhouseClusterAccessArgs', 'MdbClickhouseClusterBackupWindowStartArgs', 'MdbClickhouseClusterClickhouseArgs', 'MdbClickhouseClusterClickhouseConfigArgs', 'MdbClickhouseClusterClickhouseConfigCompressionArgs', 'MdbClickhouseClusterClickhouseConfigGraphiteRollupArgs', 'MdbClickhouseClusterClickhouseConfigGraphiteRollupPatternArgs', 'MdbClickhouseClusterClickhouseConfigGraphiteRollupPatternRetentionArgs', 'MdbClickhouseClusterClickhouseConfigKafkaArgs', 'MdbClickhouseClusterClickhouseConfigKafkaTopicArgs', 'MdbClickhouseClusterClickhouseConfigKafkaTopicSettingsArgs', 'MdbClickhouseClusterClickhouseConfigMergeTreeArgs', 'MdbClickhouseClusterClickhouseConfigRabbitmqArgs', 'MdbClickhouseClusterClickhouseResourcesArgs', 'MdbClickhouseClusterCloudStorageArgs', 'MdbClickhouseClusterDatabaseArgs', 'MdbClickhouseClusterFormatSchemaArgs', 'MdbClickhouseClusterHostArgs', 'MdbClickhouseClusterMaintenanceWindowArgs', 'MdbClickhouseClusterMlModelArgs', 'MdbClickhouseClusterShardGroupArgs', 'MdbClickhouseClusterUserArgs', 'MdbClickhouseClusterUserPermissionArgs', 'MdbClickhouseClusterUserQuotaArgs', 'MdbClickhouseClusterUserSettingsArgs', 'MdbClickhouseClusterZookeeperArgs', 'MdbClickhouseClusterZookeeperResourcesArgs', 'MdbElasticSearchClusterConfigArgs', 'MdbElasticSearchClusterConfigDataNodeArgs', 'MdbElasticSearchClusterConfigDataNodeResourcesArgs', 'MdbElasticSearchClusterConfigMasterNodeArgs', 'MdbElasticSearchClusterConfigMasterNodeResourcesArgs', 'MdbElasticSearchClusterHostArgs', 'MdbGreenplumClusterAccessArgs', 'MdbGreenplumClusterBackupWindowStartArgs', 'MdbGreenplumClusterMasterHostArgs', 'MdbGreenplumClusterMasterSubclusterArgs', 'MdbGreenplumClusterMasterSubclusterResourcesArgs', 'MdbGreenplumClusterSegmentHostArgs', 'MdbGreenplumClusterSegmentSubclusterArgs', 'MdbGreenplumClusterSegmentSubclusterResourcesArgs', 'MdbKafkaClusterConfigArgs', 'MdbKafkaClusterConfigKafkaArgs', 'MdbKafkaClusterConfigKafkaKafkaConfigArgs', 'MdbKafkaClusterConfigKafkaResourcesArgs', 'MdbKafkaClusterConfigZookeeperArgs', 'MdbKafkaClusterConfigZookeeperResourcesArgs', 'MdbKafkaClusterHostArgs', 'MdbKafkaClusterTopicArgs', 'MdbKafkaClusterTopicTopicConfigArgs', 'MdbKafkaClusterUserArgs', 'MdbKafkaClusterUserPermissionArgs', 'MdbKafkaTopicTopicConfigArgs', 'MdbMongodbClusterClusterConfigArgs', 'MdbMongodbClusterClusterConfigAccessArgs', 'MdbMongodbClusterClusterConfigBackupWindowStartArgs', 'MdbMongodbClusterDatabaseArgs', 'MdbMongodbClusterHostArgs', 'MdbMongodbClusterMaintenanceWindowArgs', 'MdbMongodbClusterResourcesArgs', 'MdbMongodbClusterUserArgs', 'MdbMongodbClusterUserPermissionArgs', 'MdbMysqlClusterAccessArgs', 'MdbMysqlClusterBackupWindowStartArgs', 'MdbMysqlClusterDatabaseArgs', 'MdbMysqlClusterHostArgs', 'MdbMysqlClusterMaintenanceWindowArgs', 'MdbMysqlClusterResourcesArgs', 'MdbMysqlClusterRestoreArgs', 'MdbMysqlClusterUserArgs', 'MdbMysqlClusterUserConnectionLimitsArgs', 'MdbMysqlClusterUserPermissionArgs', 'MdbRedisClusterConfigArgs', 'MdbRedisClusterHostArgs', 'MdbRedisClusterMaintenanceWindowArgs', 'MdbRedisClusterResourcesArgs', 'MdbSqlServerClusterBackupWindowStartArgs', 'MdbSqlServerClusterDatabaseArgs', 'MdbSqlServerClusterHostArgs', 'MdbSqlServerClusterResourcesArgs', 'MdbSqlServerClusterUserArgs', 'MdbSqlServerClusterUserPermissionArgs', 'StorageBucketCorsRuleArgs', 'StorageBucketGrantArgs', 'StorageBucketLifecycleRuleArgs', 'StorageBucketLifecycleRuleExpirationArgs', 'StorageBucketLifecycleRuleNoncurrentVersionExpirationArgs', 'StorageBucketLifecycleRuleNoncurrentVersionTransitionArgs', 'StorageBucketLifecycleRuleTransitionArgs', 'StorageBucketLoggingArgs', 'StorageBucketServerSideEncryptionConfigurationArgs', 'StorageBucketServerSideEncryptionConfigurationRuleArgs', 'StorageBucketServerSideEncryptionConfigurationRuleApplyServerSideEncryptionByDefaultArgs', 'StorageBucketVersioningArgs', 'StorageBucketWebsiteArgs', 'VpcAddressExternalIpv4AddressArgs', 'VpcDefaultSecurityGroupEgressArgs', 'VpcDefaultSecurityGroupIngressArgs', 'VpcRouteTableStaticRouteArgs', 'VpcSecurityGroupEgressArgs', 'VpcSecurityGroupIngressArgs', 'VpcSubnetDhcpOptionsArgs', 'YdbDatabaseDedicatedLocationArgs', 'YdbDatabaseDedicatedLocationRegionArgs', 'YdbDatabaseDedicatedScalePolicyArgs', 'YdbDatabaseDedicatedScalePolicyFixedScaleArgs', 'YdbDatabaseDedicatedStorageConfigArgs', 'GetAlbBackendGroupGrpcBackendArgs', 'GetAlbBackendGroupGrpcBackendHealthcheckArgs', 'GetAlbBackendGroupGrpcBackendHealthcheckGrpcHealthcheckArgs', 'GetAlbBackendGroupGrpcBackendHealthcheckHttpHealthcheckArgs', 'GetAlbBackendGroupGrpcBackendHealthcheckStreamHealthcheckArgs', 'GetAlbBackendGroupGrpcBackendLoadBalancingConfigArgs', 'GetAlbBackendGroupGrpcBackendTlsArgs', 'GetAlbBackendGroupGrpcBackendTlsValidationContextArgs', 'GetAlbBackendGroupHttpBackendArgs', 'GetAlbBackendGroupHttpBackendHealthcheckArgs', 'GetAlbBackendGroupHttpBackendHealthcheckGrpcHealthcheckArgs', 'GetAlbBackendGroupHttpBackendHealthcheckHttpHealthcheckArgs', 'GetAlbBackendGroupHttpBackendHealthcheckStreamHealthcheckArgs', 'GetAlbBackendGroupHttpBackendLoadBalancingConfigArgs', 'GetAlbBackendGroupHttpBackendTlsArgs', 'GetAlbBackendGroupHttpBackendTlsValidationContextArgs', 'GetComputeDiskDiskPlacementPolicyArgs', 'GetComputeInstancePlacementPolicyArgs', 'GetComputeInstanceSchedulingPolicyArgs', 'GetFunctionScalingPolicyPolicyArgs', 'GetIamPolicyBindingArgs', 'GetMdbClickhouseClusterCloudStorageArgs', 'GetMdbKafkaClusterConfigArgs', 'GetMdbKafkaClusterConfigKafkaArgs', 'GetMdbKafkaClusterConfigKafkaKafkaConfigArgs', 'GetMdbKafkaClusterConfigKafkaResourcesArgs', 'GetMdbKafkaClusterConfigZookeeperArgs', 'GetMdbKafkaClusterConfigZookeeperResourcesArgs', 'GetMdbKafkaClusterTopicArgs', 'GetMdbKafkaClusterTopicTopicConfigArgs', 'GetMdbKafkaClusterUserArgs', 'GetMdbKafkaClusterUserPermissionArgs', 'GetMdbMysqlClusterAccessArgs', ] @pulumi.input_type class AlbBackendGroupGrpcBackendArgs: def __init__(__self__, *, name: pulumi.Input[str], target_group_ids: pulumi.Input[Sequence[pulumi.Input[str]]], healthcheck: Optional[pulumi.Input['AlbBackendGroupGrpcBackendHealthcheckArgs']] = None, load_balancing_config: Optional[pulumi.Input['AlbBackendGroupGrpcBackendLoadBalancingConfigArgs']] = None, port: Optional[pulumi.Input[int]] = None, tls: Optional[pulumi.Input['AlbBackendGroupGrpcBackendTlsArgs']] = None, weight: Optional[pulumi.Input[int]] = None): """ :param pulumi.Input[str] name: Name of the backend. :param pulumi.Input[Sequence[pulumi.Input[str]]] target_group_ids: References target groups for the backend. :param pulumi.Input['AlbBackendGroupGrpcBackendHealthcheckArgs'] healthcheck: Healthcheck specification that will be used by this backend. Structure is documented below. :param pulumi.Input['AlbBackendGroupGrpcBackendLoadBalancingConfigArgs'] load_balancing_config: Load Balancing Config specification that will be used by this backend. Structure is documented below. :param pulumi.Input[int] port: Port for incoming traffic. :param pulumi.Input['AlbBackendGroupGrpcBackendTlsArgs'] tls: Tls specification that will be used by this backend. Structure is documented below. :param pulumi.Input[int] weight: Weight of the backend. Traffic will be split between backends of the same BackendGroup according to their weights. """ pulumi.set(__self__, "name", name) pulumi.set(__self__, "target_group_ids", target_group_ids) if healthcheck is not None: pulumi.set(__self__, "healthcheck", healthcheck) if load_balancing_config is not None: pulumi.set(__self__, "load_balancing_config", load_balancing_config) if port is not None: pulumi.set(__self__, "port", port) if tls is not None: pulumi.set(__self__, "tls", tls) if weight is not None: pulumi.set(__self__, "weight", weight) @property @pulumi.getter def name(self) -> pulumi.Input[str]: """ Name of the backend. """ return pulumi.get(self, "name") @name.setter def name(self, value: pulumi.Input[str]): pulumi.set(self, "name", value) @property @pulumi.getter(name="targetGroupIds") def target_group_ids(self) -> pulumi.Input[Sequence[pulumi.Input[str]]]: """ References target groups for the backend. """ return pulumi.get(self, "target_group_ids") @target_group_ids.setter def target_group_ids(self, value: pulumi.Input[Sequence[pulumi.Input[str]]]): pulumi.set(self, "target_group_ids", value) @property @pulumi.getter def healthcheck(self) -> Optional[pulumi.Input['AlbBackendGroupGrpcBackendHealthcheckArgs']]: """ Healthcheck specification that will be used by this backend. Structure is documented below. """ return pulumi.get(self, "healthcheck") @healthcheck.setter def healthcheck(self, value: Optional[pulumi.Input['AlbBackendGroupGrpcBackendHealthcheckArgs']]): pulumi.set(self, "healthcheck", value) @property @pulumi.getter(name="loadBalancingConfig") def load_balancing_config(self) -> Optional[pulumi.Input['AlbBackendGroupGrpcBackendLoadBalancingConfigArgs']]: """ Load Balancing Config specification that will be used by this backend. Structure is documented below. """ return pulumi.get(self, "load_balancing_config") @load_balancing_config.setter def load_balancing_config(self, value: Optional[pulumi.Input['AlbBackendGroupGrpcBackendLoadBalancingConfigArgs']]): pulumi.set(self, "load_balancing_config", value) @property @pulumi.getter def port(self) -> Optional[pulumi.Input[int]]: """ Port for incoming traffic. """ return pulumi.get(self, "port") @port.setter def port(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "port", value) @property @pulumi.getter def tls(self) -> Optional[pulumi.Input['AlbBackendGroupGrpcBackendTlsArgs']]: """ Tls specification that will be used by this backend. Structure is documented below. """ return pulumi.get(self, "tls") @tls.setter def tls(self, value: Optional[pulumi.Input['AlbBackendGroupGrpcBackendTlsArgs']]): pulumi.set(self, "tls", value) @property @pulumi.getter def weight(self) -> Optional[pulumi.Input[int]]: """ Weight of the backend. Traffic will be split between backends of the same BackendGroup according to their weights. """ return pulumi.get(self, "weight") @weight.setter def weight(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "weight", value) @pulumi.input_type class AlbBackendGroupGrpcBackendHealthcheckArgs: def __init__(__self__, *, interval: pulumi.Input[str], timeout: pulumi.Input[str], grpc_healthcheck: Optional[pulumi.Input['AlbBackendGroupGrpcBackendHealthcheckGrpcHealthcheckArgs']] = None, healthcheck_port: Optional[pulumi.Input[int]] = None, healthy_threshold: Optional[pulumi.Input[int]] = None, http_healthcheck: Optional[pulumi.Input['AlbBackendGroupGrpcBackendHealthcheckHttpHealthcheckArgs']] = None, interval_jitter_percent: Optional[pulumi.Input[float]] = None, stream_healthcheck: Optional[pulumi.Input['AlbBackendGroupGrpcBackendHealthcheckStreamHealthcheckArgs']] = None, unhealthy_threshold: Optional[pulumi.Input[int]] = None): """ :param pulumi.Input[str] interval: Interval between health checks. :param pulumi.Input[str] timeout: Time to wait for a health check response. :param pulumi.Input['AlbBackendGroupGrpcBackendHealthcheckGrpcHealthcheckArgs'] grpc_healthcheck: Grpc Healthcheck specification that will be used by this healthcheck. Structure is documented below. :param pulumi.Input[int] healthcheck_port: Optional alternative port for health checking. :param pulumi.Input[int] healthy_threshold: Number of consecutive successful health checks required to promote endpoint into the healthy state. 0 means 1. Note that during startup, only a single successful health check is required to mark a host healthy. :param pulumi.Input['AlbBackendGroupGrpcBackendHealthcheckHttpHealthcheckArgs'] http_healthcheck: Http Healthcheck specification that will be used by this healthcheck. Structure is documented below. :param pulumi.Input[float] interval_jitter_percent: An optional jitter amount as a percentage of interval. If specified, during every interval value of (interval_ms * interval_jitter_percent / 100) will be added to the wait time. :param pulumi.Input['AlbBackendGroupGrpcBackendHealthcheckStreamHealthcheckArgs'] stream_healthcheck: Stream Healthcheck specification that will be used by this healthcheck. Structure is documented below. :param pulumi.Input[int] unhealthy_threshold: Number of consecutive failed health checks required to demote endpoint into the unhealthy state. 0 means 1. Note that for HTTP health checks, a single 503 immediately makes endpoint unhealthy. """ pulumi.set(__self__, "interval", interval) pulumi.set(__self__, "timeout", timeout) if grpc_healthcheck is not None: pulumi.set(__self__, "grpc_healthcheck", grpc_healthcheck) if healthcheck_port is not None: pulumi.set(__self__, "healthcheck_port", healthcheck_port) if healthy_threshold is not None: pulumi.set(__self__, "healthy_threshold", healthy_threshold) if http_healthcheck is not None: pulumi.set(__self__, "http_healthcheck", http_healthcheck) if interval_jitter_percent is not None: pulumi.set(__self__, "interval_jitter_percent", interval_jitter_percent) if stream_healthcheck is not None: pulumi.set(__self__, "stream_healthcheck", stream_healthcheck) if unhealthy_threshold is not None: pulumi.set(__self__, "unhealthy_threshold", unhealthy_threshold) @property @pulumi.getter def interval(self) -> pulumi.Input[str]: """ Interval between health checks. """ return pulumi.get(self, "interval") @interval.setter def interval(self, value: pulumi.Input[str]): pulumi.set(self, "interval", value) @property @pulumi.getter def timeout(self) -> pulumi.Input[str]: """ Time to wait for a health check response. """ return pulumi.get(self, "timeout") @timeout.setter def timeout(self, value: pulumi.Input[str]): pulumi.set(self, "timeout", value) @property @pulumi.getter(name="grpcHealthcheck") def grpc_healthcheck(self) -> Optional[pulumi.Input['AlbBackendGroupGrpcBackendHealthcheckGrpcHealthcheckArgs']]: """ Grpc Healthcheck specification that will be used by this healthcheck. Structure is documented below. """ return pulumi.get(self, "grpc_healthcheck") @grpc_healthcheck.setter def grpc_healthcheck(self, value: Optional[pulumi.Input['AlbBackendGroupGrpcBackendHealthcheckGrpcHealthcheckArgs']]): pulumi.set(self, "grpc_healthcheck", value) @property @pulumi.getter(name="healthcheckPort") def healthcheck_port(self) -> Optional[pulumi.Input[int]]: """ Optional alternative port for health checking. """ return pulumi.get(self, "healthcheck_port") @healthcheck_port.setter def healthcheck_port(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "healthcheck_port", value) @property @pulumi.getter(name="healthyThreshold") def healthy_threshold(self) -> Optional[pulumi.Input[int]]: """ Number of consecutive successful health checks required to promote endpoint into the healthy state. 0 means 1. Note that during startup, only a single successful health check is required to mark a host healthy. """ return pulumi.get(self, "healthy_threshold") @healthy_threshold.setter def healthy_threshold(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "healthy_threshold", value) @property @pulumi.getter(name="httpHealthcheck") def http_healthcheck(self) -> Optional[pulumi.Input['AlbBackendGroupGrpcBackendHealthcheckHttpHealthcheckArgs']]: """ Http Healthcheck specification that will be used by this healthcheck. Structure is documented below. """ return pulumi.get(self, "http_healthcheck") @http_healthcheck.setter def http_healthcheck(self, value: Optional[pulumi.Input['AlbBackendGroupGrpcBackendHealthcheckHttpHealthcheckArgs']]): pulumi.set(self, "http_healthcheck", value) @property @pulumi.getter(name="intervalJitterPercent") def interval_jitter_percent(self) -> Optional[pulumi.Input[float]]: """ An optional jitter amount as a percentage of interval. If specified, during every interval value of (interval_ms * interval_jitter_percent / 100) will be added to the wait time. """ return pulumi.get(self, "interval_jitter_percent") @interval_jitter_percent.setter def interval_jitter_percent(self, value: Optional[pulumi.Input[float]]): pulumi.set(self, "interval_jitter_percent", value) @property @pulumi.getter(name="streamHealthcheck") def stream_healthcheck(self) -> Optional[pulumi.Input['AlbBackendGroupGrpcBackendHealthcheckStreamHealthcheckArgs']]: """ Stream Healthcheck specification that will be used by this healthcheck. Structure is documented below. """ return pulumi.get(self, "stream_healthcheck") @stream_healthcheck.setter def stream_healthcheck(self, value: Optional[pulumi.Input['AlbBackendGroupGrpcBackendHealthcheckStreamHealthcheckArgs']]): pulumi.set(self, "stream_healthcheck", value) @property @pulumi.getter(name="unhealthyThreshold") def unhealthy_threshold(self) -> Optional[pulumi.Input[int]]: """ Number of consecutive failed health checks required to demote endpoint into the unhealthy state. 0 means 1. Note that for HTTP health checks, a single 503 immediately makes endpoint unhealthy. """ return pulumi.get(self, "unhealthy_threshold") @unhealthy_threshold.setter def unhealthy_threshold(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "unhealthy_threshold", value) @pulumi.input_type class AlbBackendGroupGrpcBackendHealthcheckGrpcHealthcheckArgs: def __init__(__self__, *, service_name: Optional[pulumi.Input[str]] = None): """ :param pulumi.Input[str] service_name: Service name for grpc.health.v1.HealthCheckRequest message. """ if service_name is not None: pulumi.set(__self__, "service_name", service_name) @property @pulumi.getter(name="serviceName") def service_name(self) -> Optional[pulumi.Input[str]]: """ Service name for grpc.health.v1.HealthCheckRequest message. """ return pulumi.get(self, "service_name") @service_name.setter def service_name(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "service_name", value) @pulumi.input_type class AlbBackendGroupGrpcBackendHealthcheckHttpHealthcheckArgs: def __init__(__self__, *, path: pulumi.Input[str], host: Optional[pulumi.Input[str]] = None, http2: Optional[pulumi.Input[bool]] = None): """ :param pulumi.Input[str] path: HTTP path. :param pulumi.Input[str] host: "Host" HTTP header value. :param pulumi.Input[bool] http2: If set, health checks will use HTTP2. """ pulumi.set(__self__, "path", path) if host is not None: pulumi.set(__self__, "host", host) if http2 is not None: pulumi.set(__self__, "http2", http2) @property @pulumi.getter def path(self) -> pulumi.Input[str]: """ HTTP path. """ return pulumi.get(self, "path") @path.setter def path(self, value: pulumi.Input[str]): pulumi.set(self, "path", value) @property @pulumi.getter def host(self) -> Optional[pulumi.Input[str]]: """ "Host" HTTP header value. """ return pulumi.get(self, "host") @host.setter def host(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "host", value) @property @pulumi.getter def http2(self) -> Optional[pulumi.Input[bool]]: """ If set, health checks will use HTTP2. """ return pulumi.get(self, "http2") @http2.setter def http2(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "http2", value) @pulumi.input_type class AlbBackendGroupGrpcBackendHealthcheckStreamHealthcheckArgs: def __init__(__self__, *, receive: Optional[pulumi.Input[str]] = None, send: Optional[pulumi.Input[str]] = None): """ :param pulumi.Input[str] receive: Text to search in reply. :param pulumi.Input[str] send: Message to send. If empty, it's a connect-only health check. """ if receive is not None: pulumi.set(__self__, "receive", receive) if send is not None: pulumi.set(__self__, "send", send) @property @pulumi.getter def receive(self) -> Optional[pulumi.Input[str]]: """ Text to search in reply. """ return pulumi.get(self, "receive") @receive.setter def receive(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "receive", value) @property @pulumi.getter def send(self) -> Optional[pulumi.Input[str]]: """ Message to send. If empty, it's a connect-only health check. """ return pulumi.get(self, "send") @send.setter def send(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "send", value) @pulumi.input_type class AlbBackendGroupGrpcBackendLoadBalancingConfigArgs: def __init__(__self__, *, locality_aware_routing_percent: Optional[pulumi.Input[int]] = None, panic_threshold: Optional[pulumi.Input[int]] = None, strict_locality: Optional[pulumi.Input[bool]] = None): """ :param pulumi.Input[int] locality_aware_routing_percent: Percent of traffic to be sent to the same availability zone. The rest will be equally divided between other zones. :param pulumi.Input[int] panic_threshold: If percentage of healthy hosts in the backend is lower than panic_threshold, traffic will be routed to all backends no matter what the health status is. This helps to avoid healthy backends overloading when everything is bad. Zero means no panic threshold. :param pulumi.Input[bool] strict_locality: If set, will route requests only to the same availability zone. Balancer won't know about endpoints in other zones. """ if locality_aware_routing_percent is not None: pulumi.set(__self__, "locality_aware_routing_percent", locality_aware_routing_percent) if panic_threshold is not None: pulumi.set(__self__, "panic_threshold", panic_threshold) if strict_locality is not None: pulumi.set(__self__, "strict_locality", strict_locality) @property @pulumi.getter(name="localityAwareRoutingPercent") def locality_aware_routing_percent(self) -> Optional[pulumi.Input[int]]: """ Percent of traffic to be sent to the same availability zone. The rest will be equally divided between other zones. """ return pulumi.get(self, "locality_aware_routing_percent") @locality_aware_routing_percent.setter def locality_aware_routing_percent(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "locality_aware_routing_percent", value) @property @pulumi.getter(name="panicThreshold") def panic_threshold(self) -> Optional[pulumi.Input[int]]: """ If percentage of healthy hosts in the backend is lower than panic_threshold, traffic will be routed to all backends no matter what the health status is. This helps to avoid healthy backends overloading when everything is bad. Zero means no panic threshold. """ return pulumi.get(self, "panic_threshold") @panic_threshold.setter def panic_threshold(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "panic_threshold", value) @property @pulumi.getter(name="strictLocality") def strict_locality(self) -> Optional[pulumi.Input[bool]]: """ If set, will route requests only to the same availability zone. Balancer won't know about endpoints in other zones. """ return pulumi.get(self, "strict_locality") @strict_locality.setter def strict_locality(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "strict_locality", value) @pulumi.input_type class AlbBackendGroupGrpcBackendTlsArgs: def __init__(__self__, *, sni: Optional[pulumi.Input[str]] = None, validation_context: Optional[pulumi.Input['AlbBackendGroupGrpcBackendTlsValidationContextArgs']] = None): """ :param pulumi.Input[str] sni: [SNI](https://en.wikipedia.org/wiki/Server_Name_Indication) string for TLS connections. * `validation_context.0.trusted_ca_id` - (Optional) Trusted CA certificate ID in the Certificate Manager. * `validation_context.0.trusted_ca_bytes` - (Optional) PEM-encoded trusted CA certificate chain. """ if sni is not None: pulumi.set(__self__, "sni", sni) if validation_context is not None: pulumi.set(__self__, "validation_context", validation_context) @property @pulumi.getter def sni(self) -> Optional[pulumi.Input[str]]: """ [SNI](https://en.wikipedia.org/wiki/Server_Name_Indication) string for TLS connections. * `validation_context.0.trusted_ca_id` - (Optional) Trusted CA certificate ID in the Certificate Manager. * `validation_context.0.trusted_ca_bytes` - (Optional) PEM-encoded trusted CA certificate chain. """ return pulumi.get(self, "sni") @sni.setter def sni(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "sni", value) @property @pulumi.getter(name="validationContext") def validation_context(self) -> Optional[pulumi.Input['AlbBackendGroupGrpcBackendTlsValidationContextArgs']]: return pulumi.get(self, "validation_context") @validation_context.setter def validation_context(self, value: Optional[pulumi.Input['AlbBackendGroupGrpcBackendTlsValidationContextArgs']]): pulumi.set(self, "validation_context", value) @pulumi.input_type class AlbBackendGroupGrpcBackendTlsValidationContextArgs: def __init__(__self__, *, trusted_ca_bytes: Optional[pulumi.Input[str]] = None, trusted_ca_id: Optional[pulumi.Input[str]] = None): if trusted_ca_bytes is not None: pulumi.set(__self__, "trusted_ca_bytes", trusted_ca_bytes) if trusted_ca_id is not None: pulumi.set(__self__, "trusted_ca_id", trusted_ca_id) @property @pulumi.getter(name="trustedCaBytes") def trusted_ca_bytes(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "trusted_ca_bytes") @trusted_ca_bytes.setter def trusted_ca_bytes(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "trusted_ca_bytes", value) @property @pulumi.getter(name="trustedCaId") def trusted_ca_id(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "trusted_ca_id") @trusted_ca_id.setter def trusted_ca_id(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "trusted_ca_id", value) @pulumi.input_type class AlbBackendGroupHttpBackendArgs: def __init__(__self__, *, name: pulumi.Input[str], target_group_ids: pulumi.Input[Sequence[pulumi.Input[str]]], healthcheck: Optional[pulumi.Input['AlbBackendGroupHttpBackendHealthcheckArgs']] = None, http2: Optional[pulumi.Input[bool]] = None, load_balancing_config: Optional[pulumi.Input['AlbBackendGroupHttpBackendLoadBalancingConfigArgs']] = None, port: Optional[pulumi.Input[int]] = None, tls: Optional[pulumi.Input['AlbBackendGroupHttpBackendTlsArgs']] = None, weight: Optional[pulumi.Input[int]] = None): """ :param pulumi.Input[str] name: Name of the backend. :param pulumi.Input[Sequence[pulumi.Input[str]]] target_group_ids: References target groups for the backend. :param pulumi.Input['AlbBackendGroupHttpBackendHealthcheckArgs'] healthcheck: Healthcheck specification that will be used by this backend. Structure is documented below. :param pulumi.Input[bool] http2: If set, health checks will use HTTP2. :param pulumi.Input['AlbBackendGroupHttpBackendLoadBalancingConfigArgs'] load_balancing_config: Load Balancing Config specification that will be used by this backend. Structure is documented below. :param pulumi.Input[int] port: Port for incoming traffic. :param pulumi.Input['AlbBackendGroupHttpBackendTlsArgs'] tls: Tls specification that will be used by this backend. Structure is documented below. :param pulumi.Input[int] weight: Weight of the backend. Traffic will be split between backends of the same BackendGroup according to their weights. """ pulumi.set(__self__, "name", name) pulumi.set(__self__, "target_group_ids", target_group_ids) if healthcheck is not None: pulumi.set(__self__, "healthcheck", healthcheck) if http2 is not None: pulumi.set(__self__, "http2", http2) if load_balancing_config is not None: pulumi.set(__self__, "load_balancing_config", load_balancing_config) if port is not None: pulumi.set(__self__, "port", port) if tls is not None: pulumi.set(__self__, "tls", tls) if weight is not None: pulumi.set(__self__, "weight", weight) @property @pulumi.getter def name(self) -> pulumi.Input[str]: """ Name of the backend. """ return pulumi.get(self, "name") @name.setter def name(self, value: pulumi.Input[str]): pulumi.set(self, "name", value) @property @pulumi.getter(name="targetGroupIds") def target_group_ids(self) -> pulumi.Input[Sequence[pulumi.Input[str]]]: """ References target groups for the backend. """ return pulumi.get(self, "target_group_ids") @target_group_ids.setter def target_group_ids(self, value: pulumi.Input[Sequence[pulumi.Input[str]]]): pulumi.set(self, "target_group_ids", value) @property @pulumi.getter def healthcheck(self) -> Optional[pulumi.Input['AlbBackendGroupHttpBackendHealthcheckArgs']]: """ Healthcheck specification that will be used by this backend. Structure is documented below. """ return pulumi.get(self, "healthcheck") @healthcheck.setter def healthcheck(self, value: Optional[pulumi.Input['AlbBackendGroupHttpBackendHealthcheckArgs']]): pulumi.set(self, "healthcheck", value) @property @pulumi.getter def http2(self) -> Optional[pulumi.Input[bool]]: """ If set, health checks will use HTTP2. """ return pulumi.get(self, "http2") @http2.setter def http2(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "http2", value) @property @pulumi.getter(name="loadBalancingConfig") def load_balancing_config(self) -> Optional[pulumi.Input['AlbBackendGroupHttpBackendLoadBalancingConfigArgs']]: """ Load Balancing Config specification that will be used by this backend. Structure is documented below. """ return pulumi.get(self, "load_balancing_config") @load_balancing_config.setter def load_balancing_config(self, value: Optional[pulumi.Input['AlbBackendGroupHttpBackendLoadBalancingConfigArgs']]): pulumi.set(self, "load_balancing_config", value) @property @pulumi.getter def port(self) -> Optional[pulumi.Input[int]]: """ Port for incoming traffic. """ return pulumi.get(self, "port") @port.setter def port(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "port", value) @property @pulumi.getter def tls(self) -> Optional[pulumi.Input['AlbBackendGroupHttpBackendTlsArgs']]: """ Tls specification that will be used by this backend. Structure is documented below. """ return pulumi.get(self, "tls") @tls.setter def tls(self, value: Optional[pulumi.Input['AlbBackendGroupHttpBackendTlsArgs']]): pulumi.set(self, "tls", value) @property @pulumi.getter def weight(self) -> Optional[pulumi.Input[int]]: """ Weight of the backend. Traffic will be split between backends of the same BackendGroup according to their weights. """ return pulumi.get(self, "weight") @weight.setter def weight(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "weight", value) @pulumi.input_type class AlbBackendGroupHttpBackendHealthcheckArgs: def __init__(__self__, *, interval: pulumi.Input[str], timeout: pulumi.Input[str], grpc_healthcheck: Optional[pulumi.Input['AlbBackendGroupHttpBackendHealthcheckGrpcHealthcheckArgs']] = None, healthcheck_port: Optional[pulumi.Input[int]] = None, healthy_threshold: Optional[pulumi.Input[int]] = None, http_healthcheck: Optional[pulumi.Input['AlbBackendGroupHttpBackendHealthcheckHttpHealthcheckArgs']] = None, interval_jitter_percent: Optional[pulumi.Input[float]] = None, stream_healthcheck: Optional[pulumi.Input['AlbBackendGroupHttpBackendHealthcheckStreamHealthcheckArgs']] = None, unhealthy_threshold: Optional[pulumi.Input[int]] = None): """ :param pulumi.Input[str] interval: Interval between health checks. :param pulumi.Input[str] timeout: Time to wait for a health check response. :param pulumi.Input['AlbBackendGroupHttpBackendHealthcheckGrpcHealthcheckArgs'] grpc_healthcheck: Grpc Healthcheck specification that will be used by this healthcheck. Structure is documented below. :param pulumi.Input[int] healthcheck_port: Optional alternative port for health checking. :param pulumi.Input[int] healthy_threshold: Number of consecutive successful health checks required to promote endpoint into the healthy state. 0 means 1. Note that during startup, only a single successful health check is required to mark a host healthy. :param pulumi.Input['AlbBackendGroupHttpBackendHealthcheckHttpHealthcheckArgs'] http_healthcheck: Http Healthcheck specification that will be used by this healthcheck. Structure is documented below. :param pulumi.Input[float] interval_jitter_percent: An optional jitter amount as a percentage of interval. If specified, during every interval value of (interval_ms * interval_jitter_percent / 100) will be added to the wait time. :param pulumi.Input['AlbBackendGroupHttpBackendHealthcheckStreamHealthcheckArgs'] stream_healthcheck: Stream Healthcheck specification that will be used by this healthcheck. Structure is documented below. :param pulumi.Input[int] unhealthy_threshold: Number of consecutive failed health checks required to demote endpoint into the unhealthy state. 0 means 1. Note that for HTTP health checks, a single 503 immediately makes endpoint unhealthy. """ pulumi.set(__self__, "interval", interval) pulumi.set(__self__, "timeout", timeout) if grpc_healthcheck is not None: pulumi.set(__self__, "grpc_healthcheck", grpc_healthcheck) if healthcheck_port is not None: pulumi.set(__self__, "healthcheck_port", healthcheck_port) if healthy_threshold is not None: pulumi.set(__self__, "healthy_threshold", healthy_threshold) if http_healthcheck is not None: pulumi.set(__self__, "http_healthcheck", http_healthcheck) if interval_jitter_percent is not None: pulumi.set(__self__, "interval_jitter_percent", interval_jitter_percent) if stream_healthcheck is not None: pulumi.set(__self__, "stream_healthcheck", stream_healthcheck) if unhealthy_threshold is not None: pulumi.set(__self__, "unhealthy_threshold", unhealthy_threshold) @property @pulumi.getter def interval(self) -> pulumi.Input[str]: """ Interval between health checks. """ return pulumi.get(self, "interval") @interval.setter def interval(self, value: pulumi.Input[str]): pulumi.set(self, "interval", value) @property @pulumi.getter def timeout(self) -> pulumi.Input[str]: """ Time to wait for a health check response. """ return pulumi.get(self, "timeout") @timeout.setter def timeout(self, value: pulumi.Input[str]): pulumi.set(self, "timeout", value) @property @pulumi.getter(name="grpcHealthcheck") def grpc_healthcheck(self) -> Optional[pulumi.Input['AlbBackendGroupHttpBackendHealthcheckGrpcHealthcheckArgs']]: """ Grpc Healthcheck specification that will be used by this healthcheck. Structure is documented below. """ return pulumi.get(self, "grpc_healthcheck") @grpc_healthcheck.setter def grpc_healthcheck(self, value: Optional[pulumi.Input['AlbBackendGroupHttpBackendHealthcheckGrpcHealthcheckArgs']]): pulumi.set(self, "grpc_healthcheck", value) @property @pulumi.getter(name="healthcheckPort") def healthcheck_port(self) -> Optional[pulumi.Input[int]]: """ Optional alternative port for health checking. """ return pulumi.get(self, "healthcheck_port") @healthcheck_port.setter def healthcheck_port(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "healthcheck_port", value) @property @pulumi.getter(name="healthyThreshold") def healthy_threshold(self) -> Optional[pulumi.Input[int]]: """ Number of consecutive successful health checks required to promote endpoint into the healthy state. 0 means 1. Note that during startup, only a single successful health check is required to mark a host healthy. """ return pulumi.get(self, "healthy_threshold") @healthy_threshold.setter def healthy_threshold(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "healthy_threshold", value) @property @pulumi.getter(name="httpHealthcheck") def http_healthcheck(self) -> Optional[pulumi.Input['AlbBackendGroupHttpBackendHealthcheckHttpHealthcheckArgs']]: """ Http Healthcheck specification that will be used by this healthcheck. Structure is documented below. """ return pulumi.get(self, "http_healthcheck") @http_healthcheck.setter def http_healthcheck(self, value: Optional[pulumi.Input['AlbBackendGroupHttpBackendHealthcheckHttpHealthcheckArgs']]): pulumi.set(self, "http_healthcheck", value) @property @pulumi.getter(name="intervalJitterPercent") def interval_jitter_percent(self) -> Optional[pulumi.Input[float]]: """ An optional jitter amount as a percentage of interval. If specified, during every interval value of (interval_ms * interval_jitter_percent / 100) will be added to the wait time. """ return pulumi.get(self, "interval_jitter_percent") @interval_jitter_percent.setter def interval_jitter_percent(self, value: Optional[pulumi.Input[float]]): pulumi.set(self, "interval_jitter_percent", value) @property @pulumi.getter(name="streamHealthcheck") def stream_healthcheck(self) -> Optional[pulumi.Input['AlbBackendGroupHttpBackendHealthcheckStreamHealthcheckArgs']]: """ Stream Healthcheck specification that will be used by this healthcheck. Structure is documented below. """ return pulumi.get(self, "stream_healthcheck") @stream_healthcheck.setter def stream_healthcheck(self, value: Optional[pulumi.Input['AlbBackendGroupHttpBackendHealthcheckStreamHealthcheckArgs']]): pulumi.set(self, "stream_healthcheck", value) @property @pulumi.getter(name="unhealthyThreshold") def unhealthy_threshold(self) -> Optional[pulumi.Input[int]]: """ Number of consecutive failed health checks required to demote endpoint into the unhealthy state. 0 means 1. Note that for HTTP health checks, a single 503 immediately makes endpoint unhealthy. """ return pulumi.get(self, "unhealthy_threshold") @unhealthy_threshold.setter def unhealthy_threshold(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "unhealthy_threshold", value) @pulumi.input_type class AlbBackendGroupHttpBackendHealthcheckGrpcHealthcheckArgs: def __init__(__self__, *, service_name: Optional[pulumi.Input[str]] = None): """ :param pulumi.Input[str] service_name: Service name for grpc.health.v1.HealthCheckRequest message. """ if service_name is not None: pulumi.set(__self__, "service_name", service_name) @property @pulumi.getter(name="serviceName") def service_name(self) -> Optional[pulumi.Input[str]]: """ Service name for grpc.health.v1.HealthCheckRequest message. """ return pulumi.get(self, "service_name") @service_name.setter def service_name(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "service_name", value) @pulumi.input_type class AlbBackendGroupHttpBackendHealthcheckHttpHealthcheckArgs: def __init__(__self__, *, path: pulumi.Input[str], host: Optional[pulumi.Input[str]] = None, http2: Optional[pulumi.Input[bool]] = None): """ :param pulumi.Input[str] path: HTTP path. :param pulumi.Input[str] host: "Host" HTTP header value. :param pulumi.Input[bool] http2: If set, health checks will use HTTP2. """ pulumi.set(__self__, "path", path) if host is not None: pulumi.set(__self__, "host", host) if http2 is not None: pulumi.set(__self__, "http2", http2) @property @pulumi.getter def path(self) -> pulumi.Input[str]: """ HTTP path. """ return pulumi.get(self, "path") @path.setter def path(self, value: pulumi.Input[str]): pulumi.set(self, "path", value) @property @pulumi.getter def host(self) -> Optional[pulumi.Input[str]]: """ "Host" HTTP header value. """ return pulumi.get(self, "host") @host.setter def host(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "host", value) @property @pulumi.getter def http2(self) -> Optional[pulumi.Input[bool]]: """ If set, health checks will use HTTP2. """ return pulumi.get(self, "http2") @http2.setter def http2(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "http2", value) @pulumi.input_type class AlbBackendGroupHttpBackendHealthcheckStreamHealthcheckArgs: def __init__(__self__, *, receive: Optional[pulumi.Input[str]] = None, send: Optional[pulumi.Input[str]] = None): """ :param pulumi.Input[str] receive: Text to search in reply. :param pulumi.Input[str] send: Message to send. If empty, it's a connect-only health check. """ if receive is not None: pulumi.set(__self__, "receive", receive) if send is not None: pulumi.set(__self__, "send", send) @property @pulumi.getter def receive(self) -> Optional[pulumi.Input[str]]: """ Text to search in reply. """ return pulumi.get(self, "receive") @receive.setter def receive(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "receive", value) @property @pulumi.getter def send(self) -> Optional[pulumi.Input[str]]: """ Message to send. If empty, it's a connect-only health check. """ return pulumi.get(self, "send") @send.setter def send(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "send", value) @pulumi.input_type class AlbBackendGroupHttpBackendLoadBalancingConfigArgs: def __init__(__self__, *, locality_aware_routing_percent: Optional[pulumi.Input[int]] = None, panic_threshold: Optional[pulumi.Input[int]] = None, strict_locality: Optional[pulumi.Input[bool]] = None): """ :param pulumi.Input[int] locality_aware_routing_percent: Percent of traffic to be sent to the same availability zone. The rest will be equally divided between other zones. :param pulumi.Input[int] panic_threshold: If percentage of healthy hosts in the backend is lower than panic_threshold, traffic will be routed to all backends no matter what the health status is. This helps to avoid healthy backends overloading when everything is bad. Zero means no panic threshold. :param pulumi.Input[bool] strict_locality: If set, will route requests only to the same availability zone. Balancer won't know about endpoints in other zones. """ if locality_aware_routing_percent is not None: pulumi.set(__self__, "locality_aware_routing_percent", locality_aware_routing_percent) if panic_threshold is not None: pulumi.set(__self__, "panic_threshold", panic_threshold) if strict_locality is not None: pulumi.set(__self__, "strict_locality", strict_locality) @property @pulumi.getter(name="localityAwareRoutingPercent") def locality_aware_routing_percent(self) -> Optional[pulumi.Input[int]]: """ Percent of traffic to be sent to the same availability zone. The rest will be equally divided between other zones. """ return pulumi.get(self, "locality_aware_routing_percent") @locality_aware_routing_percent.setter def locality_aware_routing_percent(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "locality_aware_routing_percent", value) @property @pulumi.getter(name="panicThreshold") def panic_threshold(self) -> Optional[pulumi.Input[int]]: """ If percentage of healthy hosts in the backend is lower than panic_threshold, traffic will be routed to all backends no matter what the health status is. This helps to avoid healthy backends overloading when everything is bad. Zero means no panic threshold. """ return pulumi.get(self, "panic_threshold") @panic_threshold.setter def panic_threshold(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "panic_threshold", value) @property @pulumi.getter(name="strictLocality") def strict_locality(self) -> Optional[pulumi.Input[bool]]: """ If set, will route requests only to the same availability zone. Balancer won't know about endpoints in other zones. """ return pulumi.get(self, "strict_locality") @strict_locality.setter def strict_locality(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "strict_locality", value) @pulumi.input_type class AlbBackendGroupHttpBackendTlsArgs: def __init__(__self__, *, sni: Optional[pulumi.Input[str]] = None, validation_context: Optional[pulumi.Input['AlbBackendGroupHttpBackendTlsValidationContextArgs']] = None): """ :param pulumi.Input[str] sni: [SNI](https://en.wikipedia.org/wiki/Server_Name_Indication) string for TLS connections. * `validation_context.0.trusted_ca_id` - (Optional) Trusted CA certificate ID in the Certificate Manager. * `validation_context.0.trusted_ca_bytes` - (Optional) PEM-encoded trusted CA certificate chain. """ if sni is not None: pulumi.set(__self__, "sni", sni) if validation_context is not None: pulumi.set(__self__, "validation_context", validation_context) @property @pulumi.getter def sni(self) -> Optional[pulumi.Input[str]]: """ [SNI](https://en.wikipedia.org/wiki/Server_Name_Indication) string for TLS connections. * `validation_context.0.trusted_ca_id` - (Optional) Trusted CA certificate ID in the Certificate Manager. * `validation_context.0.trusted_ca_bytes` - (Optional) PEM-encoded trusted CA certificate chain. """ return pulumi.get(self, "sni") @sni.setter def sni(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "sni", value) @property @pulumi.getter(name="validationContext") def validation_context(self) -> Optional[pulumi.Input['AlbBackendGroupHttpBackendTlsValidationContextArgs']]: return pulumi.get(self, "validation_context") @validation_context.setter def validation_context(self, value: Optional[pulumi.Input['AlbBackendGroupHttpBackendTlsValidationContextArgs']]): pulumi.set(self, "validation_context", value) @pulumi.input_type class AlbBackendGroupHttpBackendTlsValidationContextArgs: def __init__(__self__, *, trusted_ca_bytes: Optional[pulumi.Input[str]] = None, trusted_ca_id: Optional[pulumi.Input[str]] = None): if trusted_ca_bytes is not None: pulumi.set(__self__, "trusted_ca_bytes", trusted_ca_bytes) if trusted_ca_id is not None: pulumi.set(__self__, "trusted_ca_id", trusted_ca_id) @property @pulumi.getter(name="trustedCaBytes") def trusted_ca_bytes(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "trusted_ca_bytes") @trusted_ca_bytes.setter def trusted_ca_bytes(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "trusted_ca_bytes", value) @property @pulumi.getter(name="trustedCaId") def trusted_ca_id(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "trusted_ca_id") @trusted_ca_id.setter def trusted_ca_id(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "trusted_ca_id", value) @pulumi.input_type class AlbLoadBalancerAllocationPolicyArgs: def __init__(__self__, *, locations: pulumi.Input[Sequence[pulumi.Input['AlbLoadBalancerAllocationPolicyLocationArgs']]]): """ :param pulumi.Input[Sequence[pulumi.Input['AlbLoadBalancerAllocationPolicyLocationArgs']]] locations: Unique set of locations. The structure is documented below. """ pulumi.set(__self__, "locations", locations) @property @pulumi.getter def locations(self) -> pulumi.Input[Sequence[pulumi.Input['AlbLoadBalancerAllocationPolicyLocationArgs']]]: """ Unique set of locations. The structure is documented below. """ return pulumi.get(self, "locations") @locations.setter def locations(self, value: pulumi.Input[Sequence[pulumi.Input['AlbLoadBalancerAllocationPolicyLocationArgs']]]): pulumi.set(self, "locations", value) @pulumi.input_type class AlbLoadBalancerAllocationPolicyLocationArgs: def __init__(__self__, *, subnet_id: pulumi.Input[str], zone_id: pulumi.Input[str], disable_traffic: Optional[pulumi.Input[bool]] = None): """ :param pulumi.Input[str] subnet_id: Provided by the client or computed automatically. :param pulumi.Input[str] zone_id: ID of the zone that location is located at. :param pulumi.Input[bool] disable_traffic: If set, will disable all L7 instances in the zone for request handling. """ pulumi.set(__self__, "subnet_id", subnet_id) pulumi.set(__self__, "zone_id", zone_id) if disable_traffic is not None: pulumi.set(__self__, "disable_traffic", disable_traffic) @property @pulumi.getter(name="subnetId") def subnet_id(self) -> pulumi.Input[str]: """ Provided by the client or computed automatically. """ return pulumi.get(self, "subnet_id") @subnet_id.setter def subnet_id(self, value: pulumi.Input[str]): pulumi.set(self, "subnet_id", value) @property @pulumi.getter(name="zoneId") def zone_id(self) -> pulumi.Input[str]: """ ID of the zone that location is located at. """ return pulumi.get(self, "zone_id") @zone_id.setter def zone_id(self, value: pulumi.Input[str]): pulumi.set(self, "zone_id", value) @property @pulumi.getter(name="disableTraffic") def disable_traffic(self) -> Optional[pulumi.Input[bool]]: """ If set, will disable all L7 instances in the zone for request handling. """ return pulumi.get(self, "disable_traffic") @disable_traffic.setter def disable_traffic(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "disable_traffic", value) @pulumi.input_type class AlbLoadBalancerListenerArgs: def __init__(__self__, *, name: pulumi.Input[str], endpoints: Optional[pulumi.Input[Sequence[pulumi.Input['AlbLoadBalancerListenerEndpointArgs']]]] = None, http: Optional[pulumi.Input['AlbLoadBalancerListenerHttpArgs']] = None, tls: Optional[pulumi.Input['AlbLoadBalancerListenerTlsArgs']] = None): """ :param pulumi.Input[str] name: name of SNI match. :param pulumi.Input[Sequence[pulumi.Input['AlbLoadBalancerListenerEndpointArgs']]] endpoints: Network endpoints (addresses and ports) of the listener. The structure is documented below. :param pulumi.Input['AlbLoadBalancerListenerHttpArgs'] http: HTTP listener resource. The structure is documented below. :param pulumi.Input['AlbLoadBalancerListenerTlsArgs'] tls: TLS listener resource. The structure is documented below. """ pulumi.set(__self__, "name", name) if endpoints is not None: pulumi.set(__self__, "endpoints", endpoints) if http is not None: pulumi.set(__self__, "http", http) if tls is not None: pulumi.set(__self__, "tls", tls) @property @pulumi.getter def name(self) -> pulumi.Input[str]: """ name of SNI match. """ return pulumi.get(self, "name") @name.setter def name(self, value: pulumi.Input[str]): pulumi.set(self, "name", value) @property @pulumi.getter def endpoints(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['AlbLoadBalancerListenerEndpointArgs']]]]: """ Network endpoints (addresses and ports) of the listener. The structure is documented below. """ return pulumi.get(self, "endpoints") @endpoints.setter def endpoints(self, value: Optional[pulumi.Input[Sequence[pulumi.Input['AlbLoadBalancerListenerEndpointArgs']]]]): pulumi.set(self, "endpoints", value) @property @pulumi.getter def http(self) -> Optional[pulumi.Input['AlbLoadBalancerListenerHttpArgs']]: """ HTTP listener resource. The structure is documented below. """ return pulumi.get(self, "http") @http.setter def http(self, value: Optional[pulumi.Input['AlbLoadBalancerListenerHttpArgs']]): pulumi.set(self, "http", value) @property @pulumi.getter def tls(self) -> Optional[pulumi.Input['AlbLoadBalancerListenerTlsArgs']]: """ TLS listener resource. The structure is documented below. """ return pulumi.get(self, "tls") @tls.setter def tls(self, value: Optional[pulumi.Input['AlbLoadBalancerListenerTlsArgs']]): pulumi.set(self, "tls", value) @pulumi.input_type class AlbLoadBalancerListenerEndpointArgs: def __init__(__self__, *, addresses: pulumi.Input[Sequence[pulumi.Input['AlbLoadBalancerListenerEndpointAddressArgs']]], ports: pulumi.Input[Sequence[pulumi.Input[int]]]): """ :param pulumi.Input[Sequence[pulumi.Input['AlbLoadBalancerListenerEndpointAddressArgs']]] addresses: Provided by the client or computed automatically. :param pulumi.Input[Sequence[pulumi.Input[int]]] ports: One or more ports to listen on. """ pulumi.set(__self__, "addresses", addresses) pulumi.set(__self__, "ports", ports) @property @pulumi.getter def addresses(self) -> pulumi.Input[Sequence[pulumi.Input['AlbLoadBalancerListenerEndpointAddressArgs']]]: """ Provided by the client or computed automatically. """ return pulumi.get(self, "addresses") @addresses.setter def addresses(self, value: pulumi.Input[Sequence[pulumi.Input['AlbLoadBalancerListenerEndpointAddressArgs']]]): pulumi.set(self, "addresses", value) @property @pulumi.getter def ports(self) -> pulumi.Input[Sequence[pulumi.Input[int]]]: """ One or more ports to listen on. """ return pulumi.get(self, "ports") @ports.setter def ports(self, value: pulumi.Input[Sequence[pulumi.Input[int]]]): pulumi.set(self, "ports", value) @pulumi.input_type class AlbLoadBalancerListenerEndpointAddressArgs: def __init__(__self__, *, external_ipv4_address: Optional[pulumi.Input['AlbLoadBalancerListenerEndpointAddressExternalIpv4AddressArgs']] = None, external_ipv6_address: Optional[pulumi.Input['AlbLoadBalancerListenerEndpointAddressExternalIpv6AddressArgs']] = None, internal_ipv4_address: Optional[pulumi.Input['AlbLoadBalancerListenerEndpointAddressInternalIpv4AddressArgs']] = None): """ :param pulumi.Input['AlbLoadBalancerListenerEndpointAddressExternalIpv4AddressArgs'] external_ipv4_address: External IPv4 address. The structure is documented below. :param pulumi.Input['AlbLoadBalancerListenerEndpointAddressExternalIpv6AddressArgs'] external_ipv6_address: External IPv6 address. The structure is documented below. :param pulumi.Input['AlbLoadBalancerListenerEndpointAddressInternalIpv4AddressArgs'] internal_ipv4_address: Internal IPv4 address. The structure is documented below. """ if external_ipv4_address is not None: pulumi.set(__self__, "external_ipv4_address", external_ipv4_address) if external_ipv6_address is not None: pulumi.set(__self__, "external_ipv6_address", external_ipv6_address) if internal_ipv4_address is not None: pulumi.set(__self__, "internal_ipv4_address", internal_ipv4_address) @property @pulumi.getter(name="externalIpv4Address") def external_ipv4_address(self) -> Optional[pulumi.Input['AlbLoadBalancerListenerEndpointAddressExternalIpv4AddressArgs']]: """ External IPv4 address. The structure is documented below. """ return pulumi.get(self, "external_ipv4_address") @external_ipv4_address.setter def external_ipv4_address(self, value: Optional[pulumi.Input['AlbLoadBalancerListenerEndpointAddressExternalIpv4AddressArgs']]): pulumi.set(self, "external_ipv4_address", value) @property @pulumi.getter(name="externalIpv6Address") def external_ipv6_address(self) -> Optional[pulumi.Input['AlbLoadBalancerListenerEndpointAddressExternalIpv6AddressArgs']]: """ External IPv6 address. The structure is documented below. """ return pulumi.get(self, "external_ipv6_address") @external_ipv6_address.setter def external_ipv6_address(self, value: Optional[pulumi.Input['AlbLoadBalancerListenerEndpointAddressExternalIpv6AddressArgs']]): pulumi.set(self, "external_ipv6_address", value) @property @pulumi.getter(name="internalIpv4Address") def internal_ipv4_address(self) -> Optional[pulumi.Input['AlbLoadBalancerListenerEndpointAddressInternalIpv4AddressArgs']]: """ Internal IPv4 address. The structure is documented below. """ return pulumi.get(self, "internal_ipv4_address") @internal_ipv4_address.setter def internal_ipv4_address(self, value: Optional[pulumi.Input['AlbLoadBalancerListenerEndpointAddressInternalIpv4AddressArgs']]): pulumi.set(self, "internal_ipv4_address", value) @pulumi.input_type class AlbLoadBalancerListenerEndpointAddressExternalIpv4AddressArgs: def __init__(__self__, *, address: Optional[pulumi.Input[str]] = None): """ :param pulumi.Input[str] address: Provided by the client or computed automatically. """ if address is not None: pulumi.set(__self__, "address", address) @property @pulumi.getter def address(self) -> Optional[pulumi.Input[str]]: """ Provided by the client or computed automatically. """ return pulumi.get(self, "address") @address.setter def address(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "address", value) @pulumi.input_type class AlbLoadBalancerListenerEndpointAddressExternalIpv6AddressArgs: def __init__(__self__, *, address: Optional[pulumi.Input[str]] = None): """ :param pulumi.Input[str] address: Provided by the client or computed automatically. """ if address is not None: pulumi.set(__self__, "address", address) @property @pulumi.getter def address(self) -> Optional[pulumi.Input[str]]: """ Provided by the client or computed automatically. """ return pulumi.get(self, "address") @address.setter def address(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "address", value) @pulumi.input_type class AlbLoadBalancerListenerEndpointAddressInternalIpv4AddressArgs: def __init__(__self__, *, address: Optional[pulumi.Input[str]] = None, subnet_id: Optional[pulumi.Input[str]] = None): """ :param pulumi.Input[str] address: Provided by the client or computed automatically. :param pulumi.Input[str] subnet_id: Provided by the client or computed automatically. """ if address is not None: pulumi.set(__self__, "address", address) if subnet_id is not None: pulumi.set(__self__, "subnet_id", subnet_id) @property @pulumi.getter def address(self) -> Optional[pulumi.Input[str]]: """ Provided by the client or computed automatically. """ return pulumi.get(self, "address") @address.setter def address(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "address", value) @property @pulumi.getter(name="subnetId") def subnet_id(self) -> Optional[pulumi.Input[str]]: """ Provided by the client or computed automatically. """ return pulumi.get(self, "subnet_id") @subnet_id.setter def subnet_id(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "subnet_id", value) @pulumi.input_type class AlbLoadBalancerListenerHttpArgs: def __init__(__self__, *, handler: Optional[pulumi.Input['AlbLoadBalancerListenerHttpHandlerArgs']] = None, redirects: Optional[pulumi.Input['AlbLoadBalancerListenerHttpRedirectsArgs']] = None): """ :param pulumi.Input['AlbLoadBalancerListenerHttpHandlerArgs'] handler: HTTP handler that sets plaintext HTTP router. The structure is documented below. :param pulumi.Input['AlbLoadBalancerListenerHttpRedirectsArgs'] redirects: Shortcut for adding http > https redirects. The structure is documented below. """ if handler is not None: pulumi.set(__self__, "handler", handler) if redirects is not None: pulumi.set(__self__, "redirects", redirects) @property @pulumi.getter def handler(self) -> Optional[pulumi.Input['AlbLoadBalancerListenerHttpHandlerArgs']]: """ HTTP handler that sets plaintext HTTP router. The structure is documented below. """ return pulumi.get(self, "handler") @handler.setter def handler(self, value: Optional[pulumi.Input['AlbLoadBalancerListenerHttpHandlerArgs']]): pulumi.set(self, "handler", value) @property @pulumi.getter def redirects(self) -> Optional[pulumi.Input['AlbLoadBalancerListenerHttpRedirectsArgs']]: """ Shortcut for adding http > https redirects. The structure is documented below. """ return pulumi.get(self, "redirects") @redirects.setter def redirects(self, value: Optional[pulumi.Input['AlbLoadBalancerListenerHttpRedirectsArgs']]): pulumi.set(self, "redirects", value) @pulumi.input_type class AlbLoadBalancerListenerHttpHandlerArgs: def __init__(__self__, *, allow_http10: Optional[pulumi.Input[bool]] = None, http2_options: Optional[pulumi.Input['AlbLoadBalancerListenerHttpHandlerHttp2OptionsArgs']] = None, http_router_id: Optional[pulumi.Input[str]] = None): """ :param pulumi.Input[bool] allow_http10: If set, will enable only HTTP1 protocol with HTTP1.0 support. :param pulumi.Input['AlbLoadBalancerListenerHttpHandlerHttp2OptionsArgs'] http2_options: If set, will enable HTTP2 protocol for the handler. The structure is documented below. :param pulumi.Input[str] http_router_id: HTTP router id. """ if allow_http10 is not None: pulumi.set(__self__, "allow_http10", allow_http10) if http2_options is not None: pulumi.set(__self__, "http2_options", http2_options) if http_router_id is not None: pulumi.set(__self__, "http_router_id", http_router_id) @property @pulumi.getter(name="allowHttp10") def allow_http10(self) -> Optional[pulumi.Input[bool]]: """ If set, will enable only HTTP1 protocol with HTTP1.0 support. """ return pulumi.get(self, "allow_http10") @allow_http10.setter def allow_http10(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "allow_http10", value) @property @pulumi.getter(name="http2Options") def http2_options(self) -> Optional[pulumi.Input['AlbLoadBalancerListenerHttpHandlerHttp2OptionsArgs']]: """ If set, will enable HTTP2 protocol for the handler. The structure is documented below. """ return pulumi.get(self, "http2_options") @http2_options.setter def http2_options(self, value: Optional[pulumi.Input['AlbLoadBalancerListenerHttpHandlerHttp2OptionsArgs']]): pulumi.set(self, "http2_options", value) @property @pulumi.getter(name="httpRouterId") def http_router_id(self) -> Optional[pulumi.Input[str]]: """ HTTP router id. """ return pulumi.get(self, "http_router_id") @http_router_id.setter def http_router_id(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "http_router_id", value) @pulumi.input_type class AlbLoadBalancerListenerHttpHandlerHttp2OptionsArgs: def __init__(__self__, *, max_concurrent_streams: Optional[pulumi.Input[int]] = None): """ :param pulumi.Input[int] max_concurrent_streams: Maximum number of concurrent streams. """ if max_concurrent_streams is not None: pulumi.set(__self__, "max_concurrent_streams", max_concurrent_streams) @property @pulumi.getter(name="maxConcurrentStreams") def max_concurrent_streams(self) -> Optional[pulumi.Input[int]]: """ Maximum number of concurrent streams. """ return pulumi.get(self, "max_concurrent_streams") @max_concurrent_streams.setter def max_concurrent_streams(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "max_concurrent_streams", value) @pulumi.input_type class AlbLoadBalancerListenerHttpRedirectsArgs: def __init__(__self__, *, http_to_https: Optional[pulumi.Input[bool]] = None): if http_to_https is not None: pulumi.set(__self__, "http_to_https", http_to_https) @property @pulumi.getter(name="httpToHttps") def http_to_https(self) -> Optional[pulumi.Input[bool]]: return pulumi.get(self, "http_to_https") @http_to_https.setter def http_to_https(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "http_to_https", value) @pulumi.input_type class AlbLoadBalancerListenerTlsArgs: def __init__(__self__, *, default_handler: pulumi.Input['AlbLoadBalancerListenerTlsDefaultHandlerArgs'], sni_handlers: Optional[pulumi.Input[Sequence[pulumi.Input['AlbLoadBalancerListenerTlsSniHandlerArgs']]]] = None): """ :param pulumi.Input['AlbLoadBalancerListenerTlsDefaultHandlerArgs'] default_handler: TLS handler resource. The structure is documented below. :param pulumi.Input[Sequence[pulumi.Input['AlbLoadBalancerListenerTlsSniHandlerArgs']]] sni_handlers: SNI match resource. The structure is documented below. """ pulumi.set(__self__, "default_handler", default_handler) if sni_handlers is not None: pulumi.set(__self__, "sni_handlers", sni_handlers) @property @pulumi.getter(name="defaultHandler") def default_handler(self) -> pulumi.Input['AlbLoadBalancerListenerTlsDefaultHandlerArgs']: """ TLS handler resource. The structure is documented below. """ return pulumi.get(self, "default_handler") @default_handler.setter def default_handler(self, value: pulumi.Input['AlbLoadBalancerListenerTlsDefaultHandlerArgs']): pulumi.set(self, "default_handler", value) @property @pulumi.getter(name="sniHandlers") def sni_handlers(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['AlbLoadBalancerListenerTlsSniHandlerArgs']]]]: """ SNI match resource. The structure is documented below. """ return pulumi.get(self, "sni_handlers") @sni_handlers.setter def sni_handlers(self, value: Optional[pulumi.Input[Sequence[pulumi.Input['AlbLoadBalancerListenerTlsSniHandlerArgs']]]]): pulumi.set(self, "sni_handlers", value) @pulumi.input_type class AlbLoadBalancerListenerTlsDefaultHandlerArgs: def __init__(__self__, *, certificate_ids: pulumi.Input[Sequence[pulumi.Input[str]]], http_handler: Optional[pulumi.Input['AlbLoadBalancerListenerTlsDefaultHandlerHttpHandlerArgs']] = None): """ :param pulumi.Input[Sequence[pulumi.Input[str]]] certificate_ids: Certificate IDs in the Certificate Manager. Multiple TLS certificates can be associated with the same context to allow both RSA and ECDSA certificates. Only the first certificate of each type will be used. :param pulumi.Input['AlbLoadBalancerListenerTlsDefaultHandlerHttpHandlerArgs'] http_handler: HTTP handler resource. The structure is documented below. """ pulumi.set(__self__, "certificate_ids", certificate_ids) if http_handler is not None: pulumi.set(__self__, "http_handler", http_handler) @property @pulumi.getter(name="certificateIds") def certificate_ids(self) -> pulumi.Input[Sequence[pulumi.Input[str]]]: """ Certificate IDs in the Certificate Manager. Multiple TLS certificates can be associated with the same context to allow both RSA and ECDSA certificates. Only the first certificate of each type will be used. """ return pulumi.get(self, "certificate_ids") @certificate_ids.setter def certificate_ids(self, value: pulumi.Input[Sequence[pulumi.Input[str]]]): pulumi.set(self, "certificate_ids", value) @property @pulumi.getter(name="httpHandler") def http_handler(self) -> Optional[pulumi.Input['AlbLoadBalancerListenerTlsDefaultHandlerHttpHandlerArgs']]: """ HTTP handler resource. The structure is documented below. """ return pulumi.get(self, "http_handler") @http_handler.setter def http_handler(self, value: Optional[pulumi.Input['AlbLoadBalancerListenerTlsDefaultHandlerHttpHandlerArgs']]): pulumi.set(self, "http_handler", value) @pulumi.input_type class AlbLoadBalancerListenerTlsDefaultHandlerHttpHandlerArgs: def __init__(__self__, *, allow_http10: Optional[pulumi.Input[bool]] = None, http2_options: Optional[pulumi.Input['AlbLoadBalancerListenerTlsDefaultHandlerHttpHandlerHttp2OptionsArgs']] = None, http_router_id: Optional[pulumi.Input[str]] = None): """ :param pulumi.Input[bool] allow_http10: If set, will enable only HTTP1 protocol with HTTP1.0 support. :param pulumi.Input['AlbLoadBalancerListenerTlsDefaultHandlerHttpHandlerHttp2OptionsArgs'] http2_options: If set, will enable HTTP2 protocol for the handler. The structure is documented below. :param pulumi.Input[str] http_router_id: HTTP router id. """ if allow_http10 is not None: pulumi.set(__self__, "allow_http10", allow_http10) if http2_options is not None: pulumi.set(__self__, "http2_options", http2_options) if http_router_id is not None: pulumi.set(__self__, "http_router_id", http_router_id) @property @pulumi.getter(name="allowHttp10") def allow_http10(self) -> Optional[pulumi.Input[bool]]: """ If set, will enable only HTTP1 protocol with HTTP1.0 support. """ return pulumi.get(self, "allow_http10") @allow_http10.setter def allow_http10(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "allow_http10", value) @property @pulumi.getter(name="http2Options") def http2_options(self) -> Optional[pulumi.Input['AlbLoadBalancerListenerTlsDefaultHandlerHttpHandlerHttp2OptionsArgs']]: """ If set, will enable HTTP2 protocol for the handler. The structure is documented below. """ return pulumi.get(self, "http2_options") @http2_options.setter def http2_options(self, value: Optional[pulumi.Input['AlbLoadBalancerListenerTlsDefaultHandlerHttpHandlerHttp2OptionsArgs']]): pulumi.set(self, "http2_options", value) @property @pulumi.getter(name="httpRouterId") def http_router_id(self) -> Optional[pulumi.Input[str]]: """ HTTP router id. """ return pulumi.get(self, "http_router_id") @http_router_id.setter def http_router_id(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "http_router_id", value) @pulumi.input_type class AlbLoadBalancerListenerTlsDefaultHandlerHttpHandlerHttp2OptionsArgs: def __init__(__self__, *, max_concurrent_streams: Optional[pulumi.Input[int]] = None): """ :param pulumi.Input[int] max_concurrent_streams: Maximum number of concurrent streams. """ if max_concurrent_streams is not None: pulumi.set(__self__, "max_concurrent_streams", max_concurrent_streams) @property @pulumi.getter(name="maxConcurrentStreams") def max_concurrent_streams(self) -> Optional[pulumi.Input[int]]: """ Maximum number of concurrent streams. """ return pulumi.get(self, "max_concurrent_streams") @max_concurrent_streams.setter def max_concurrent_streams(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "max_concurrent_streams", value) @pulumi.input_type class AlbLoadBalancerListenerTlsSniHandlerArgs: def __init__(__self__, *, handler: pulumi.Input['AlbLoadBalancerListenerTlsSniHandlerHandlerArgs'], name: pulumi.Input[str], server_names: pulumi.Input[Sequence[pulumi.Input[str]]]): """ :param pulumi.Input['AlbLoadBalancerListenerTlsSniHandlerHandlerArgs'] handler: HTTP handler that sets plaintext HTTP router. The structure is documented below. :param pulumi.Input[str] name: name of SNI match. :param pulumi.Input[Sequence[pulumi.Input[str]]] server_names: A set of server names. """ pulumi.set(__self__, "handler", handler) pulumi.set(__self__, "name", name) pulumi.set(__self__, "server_names", server_names) @property @pulumi.getter def handler(self) -> pulumi.Input['AlbLoadBalancerListenerTlsSniHandlerHandlerArgs']: """ HTTP handler that sets plaintext HTTP router. The structure is documented below. """ return pulumi.get(self, "handler") @handler.setter def handler(self, value: pulumi.Input['AlbLoadBalancerListenerTlsSniHandlerHandlerArgs']): pulumi.set(self, "handler", value) @property @pulumi.getter def name(self) -> pulumi.Input[str]: """ name of SNI match. """ return pulumi.get(self, "name") @name.setter def name(self, value: pulumi.Input[str]): pulumi.set(self, "name", value) @property @pulumi.getter(name="serverNames") def server_names(self) -> pulumi.Input[Sequence[pulumi.Input[str]]]: """ A set of server names. """ return pulumi.get(self, "server_names") @server_names.setter def server_names(self, value: pulumi.Input[Sequence[pulumi.Input[str]]]): pulumi.set(self, "server_names", value) @pulumi.input_type class AlbLoadBalancerListenerTlsSniHandlerHandlerArgs: def __init__(__self__, *, certificate_ids: pulumi.Input[Sequence[pulumi.Input[str]]], http_handler: Optional[pulumi.Input['AlbLoadBalancerListenerTlsSniHandlerHandlerHttpHandlerArgs']] = None): """ :param pulumi.Input[Sequence[pulumi.Input[str]]] certificate_ids: Certificate IDs in the Certificate Manager. Multiple TLS certificates can be associated with the same context to allow both RSA and ECDSA certificates. Only the first certificate of each type will be used. :param pulumi.Input['AlbLoadBalancerListenerTlsSniHandlerHandlerHttpHandlerArgs'] http_handler: HTTP handler resource. The structure is documented below. """ pulumi.set(__self__, "certificate_ids", certificate_ids) if http_handler is not None: pulumi.set(__self__, "http_handler", http_handler) @property @pulumi.getter(name="certificateIds") def certificate_ids(self) -> pulumi.Input[Sequence[pulumi.Input[str]]]: """ Certificate IDs in the Certificate Manager. Multiple TLS certificates can be associated with the same context to allow both RSA and ECDSA certificates. Only the first certificate of each type will be used. """ return pulumi.get(self, "certificate_ids") @certificate_ids.setter def certificate_ids(self, value: pulumi.Input[Sequence[pulumi.Input[str]]]): pulumi.set(self, "certificate_ids", value) @property @pulumi.getter(name="httpHandler") def http_handler(self) -> Optional[pulumi.Input['AlbLoadBalancerListenerTlsSniHandlerHandlerHttpHandlerArgs']]: """ HTTP handler resource. The structure is documented below. """ return pulumi.get(self, "http_handler") @http_handler.setter def http_handler(self, value: Optional[pulumi.Input['AlbLoadBalancerListenerTlsSniHandlerHandlerHttpHandlerArgs']]): pulumi.set(self, "http_handler", value) @pulumi.input_type class AlbLoadBalancerListenerTlsSniHandlerHandlerHttpHandlerArgs: def __init__(__self__, *, allow_http10: Optional[pulumi.Input[bool]] = None, http2_options: Optional[pulumi.Input['AlbLoadBalancerListenerTlsSniHandlerHandlerHttpHandlerHttp2OptionsArgs']] = None, http_router_id: Optional[pulumi.Input[str]] = None): """ :param pulumi.Input[bool] allow_http10: If set, will enable only HTTP1 protocol with HTTP1.0 support. :param pulumi.Input['AlbLoadBalancerListenerTlsSniHandlerHandlerHttpHandlerHttp2OptionsArgs'] http2_options: If set, will enable HTTP2 protocol for the handler. The structure is documented below. :param pulumi.Input[str] http_router_id: HTTP router id. """ if allow_http10 is not None: pulumi.set(__self__, "allow_http10", allow_http10) if http2_options is not None: pulumi.set(__self__, "http2_options", http2_options) if http_router_id is not None: pulumi.set(__self__, "http_router_id", http_router_id) @property @pulumi.getter(name="allowHttp10") def allow_http10(self) -> Optional[pulumi.Input[bool]]: """ If set, will enable only HTTP1 protocol with HTTP1.0 support. """ return pulumi.get(self, "allow_http10") @allow_http10.setter def allow_http10(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "allow_http10", value) @property @pulumi.getter(name="http2Options") def http2_options(self) -> Optional[pulumi.Input['AlbLoadBalancerListenerTlsSniHandlerHandlerHttpHandlerHttp2OptionsArgs']]: """ If set, will enable HTTP2 protocol for the handler. The structure is documented below. """ return pulumi.get(self, "http2_options") @http2_options.setter def http2_options(self, value: Optional[pulumi.Input['AlbLoadBalancerListenerTlsSniHandlerHandlerHttpHandlerHttp2OptionsArgs']]): pulumi.set(self, "http2_options", value) @property @pulumi.getter(name="httpRouterId") def http_router_id(self) -> Optional[pulumi.Input[str]]: """ HTTP router id. """ return pulumi.get(self, "http_router_id") @http_router_id.setter def http_router_id(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "http_router_id", value) @pulumi.input_type class AlbLoadBalancerListenerTlsSniHandlerHandlerHttpHandlerHttp2OptionsArgs: def __init__(__self__, *, max_concurrent_streams: Optional[pulumi.Input[int]] = None): """ :param pulumi.Input[int] max_concurrent_streams: Maximum number of concurrent streams. """ if max_concurrent_streams is not None: pulumi.set(__self__, "max_concurrent_streams", max_concurrent_streams) @property @pulumi.getter(name="maxConcurrentStreams") def max_concurrent_streams(self) -> Optional[pulumi.Input[int]]: """ Maximum number of concurrent streams. """ return pulumi.get(self, "max_concurrent_streams") @max_concurrent_streams.setter def max_concurrent_streams(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "max_concurrent_streams", value) @pulumi.input_type class AlbTargetGroupTargetArgs: def __init__(__self__, *, ip_address: pulumi.Input[str], subnet_id: pulumi.Input[str]): """ :param pulumi.Input[str] ip_address: IP address of the target. :param pulumi.Input[str] subnet_id: ID of the subnet that targets are connected to. All targets in the target group must be connected to the same subnet within a single availability zone. """ pulumi.set(__self__, "ip_address", ip_address) pulumi.set(__self__, "subnet_id", subnet_id) @property @pulumi.getter(name="ipAddress") def ip_address(self) -> pulumi.Input[str]: """ IP address of the target. """ return pulumi.get(self, "ip_address") @ip_address.setter def ip_address(self, value: pulumi.Input[str]): pulumi.set(self, "ip_address", value) @property @pulumi.getter(name="subnetId") def subnet_id(self) -> pulumi.Input[str]: """ ID of the subnet that targets are connected to. All targets in the target group must be connected to the same subnet within a single availability zone. """ return pulumi.get(self, "subnet_id") @subnet_id.setter def subnet_id(self, value: pulumi.Input[str]): pulumi.set(self, "subnet_id", value) @pulumi.input_type class AlbVirtualHostModifyRequestHeaderArgs: def __init__(__self__, *, append: Optional[pulumi.Input[str]] = None, name: Optional[pulumi.Input[str]] = None, remove: Optional[pulumi.Input[bool]] = None, replace: Optional[pulumi.Input[str]] = None): """ :param pulumi.Input[str] append: Append string to the header value. :param pulumi.Input[str] name: name of the route. :param pulumi.Input[bool] remove: If set, remove the header. :param pulumi.Input[str] replace: New value for a header. Header values support the following [formatters](https://www.envoyproxy.io/docs/envoy/latest/configuration/http/http_conn_man/headers#custom-request-response-headers). """ if append is not None: pulumi.set(__self__, "append", append) if name is not None: pulumi.set(__self__, "name", name) if remove is not None: pulumi.set(__self__, "remove", remove) if replace is not None: pulumi.set(__self__, "replace", replace) @property @pulumi.getter def append(self) -> Optional[pulumi.Input[str]]: """ Append string to the header value. """ return pulumi.get(self, "append") @append.setter def append(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "append", value) @property @pulumi.getter def name(self) -> Optional[pulumi.Input[str]]: """ name of the route. """ return pulumi.get(self, "name") @name.setter def name(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "name", value) @property @pulumi.getter def remove(self) -> Optional[pulumi.Input[bool]]: """ If set, remove the header. """ return pulumi.get(self, "remove") @remove.setter def remove(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "remove", value) @property @pulumi.getter def replace(self) -> Optional[pulumi.Input[str]]: """ New value for a header. Header values support the following [formatters](https://www.envoyproxy.io/docs/envoy/latest/configuration/http/http_conn_man/headers#custom-request-response-headers). """ return pulumi.get(self, "replace") @replace.setter def replace(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "replace", value) @pulumi.input_type class AlbVirtualHostModifyResponseHeaderArgs: def __init__(__self__, *, append: Optional[pulumi.Input[str]] = None, name: Optional[pulumi.Input[str]] = None, remove: Optional[pulumi.Input[bool]] = None, replace: Optional[pulumi.Input[str]] = None): """ :param pulumi.Input[str] append: Append string to the header value. :param pulumi.Input[str] name: name of the route. :param pulumi.Input[bool] remove: If set, remove the header. :param pulumi.Input[str] replace: New value for a header. Header values support the following [formatters](https://www.envoyproxy.io/docs/envoy/latest/configuration/http/http_conn_man/headers#custom-request-response-headers). """ if append is not None: pulumi.set(__self__, "append", append) if name is not None: pulumi.set(__self__, "name", name) if remove is not None: pulumi.set(__self__, "remove", remove) if replace is not None: pulumi.set(__self__, "replace", replace) @property @pulumi.getter def append(self) -> Optional[pulumi.Input[str]]: """ Append string to the header value. """ return pulumi.get(self, "append") @append.setter def append(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "append", value) @property @pulumi.getter def name(self) -> Optional[pulumi.Input[str]]: """ name of the route. """ return pulumi.get(self, "name") @name.setter def name(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "name", value) @property @pulumi.getter def remove(self) -> Optional[pulumi.Input[bool]]: """ If set, remove the header. """ return pulumi.get(self, "remove") @remove.setter def remove(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "remove", value) @property @pulumi.getter def replace(self) -> Optional[pulumi.Input[str]]: """ New value for a header. Header values support the following [formatters](https://www.envoyproxy.io/docs/envoy/latest/configuration/http/http_conn_man/headers#custom-request-response-headers). """ return pulumi.get(self, "replace") @replace.setter def replace(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "replace", value) @pulumi.input_type class AlbVirtualHostRouteArgs: def __init__(__self__, *, grpc_route: Optional[pulumi.Input['AlbVirtualHostRouteGrpcRouteArgs']] = None, http_route: Optional[pulumi.Input['AlbVirtualHostRouteHttpRouteArgs']] = None, name: Optional[pulumi.Input[str]] = None): """ :param pulumi.Input['AlbVirtualHostRouteGrpcRouteArgs'] grpc_route: GRPC route resource. The structure is documented below. :param pulumi.Input['AlbVirtualHostRouteHttpRouteArgs'] http_route: HTTP route resource. The structure is documented below. :param pulumi.Input[str] name: name of the route. """ if grpc_route is not None: pulumi.set(__self__, "grpc_route", grpc_route) if http_route is not None: pulumi.set(__self__, "http_route", http_route) if name is not None: pulumi.set(__self__, "name", name) @property @pulumi.getter(name="grpcRoute") def grpc_route(self) -> Optional[pulumi.Input['AlbVirtualHostRouteGrpcRouteArgs']]: """ GRPC route resource. The structure is documented below. """ return pulumi.get(self, "grpc_route") @grpc_route.setter def grpc_route(self, value: Optional[pulumi.Input['AlbVirtualHostRouteGrpcRouteArgs']]): pulumi.set(self, "grpc_route", value) @property @pulumi.getter(name="httpRoute") def http_route(self) -> Optional[pulumi.Input['AlbVirtualHostRouteHttpRouteArgs']]: """ HTTP route resource. The structure is documented below. """ return pulumi.get(self, "http_route") @http_route.setter def http_route(self, value: Optional[pulumi.Input['AlbVirtualHostRouteHttpRouteArgs']]): pulumi.set(self, "http_route", value) @property @pulumi.getter def name(self) -> Optional[pulumi.Input[str]]: """ name of the route. """ return pulumi.get(self, "name") @name.setter def name(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "name", value) @pulumi.input_type class AlbVirtualHostRouteGrpcRouteArgs: def __init__(__self__, *, grpc_matches: Optional[pulumi.Input[Sequence[pulumi.Input['AlbVirtualHostRouteGrpcRouteGrpcMatchArgs']]]] = None, grpc_route_action: Optional[pulumi.Input['AlbVirtualHostRouteGrpcRouteGrpcRouteActionArgs']] = None, grpc_status_response_action: Optional[pulumi.Input['AlbVirtualHostRouteGrpcRouteGrpcStatusResponseActionArgs']] = None): """ :param pulumi.Input[Sequence[pulumi.Input['AlbVirtualHostRouteGrpcRouteGrpcMatchArgs']]] grpc_matches: Checks "/" prefix by default. The structure is documented below. :param pulumi.Input['AlbVirtualHostRouteGrpcRouteGrpcRouteActionArgs'] grpc_route_action: GRPC route action resource. The structure is documented below. :param pulumi.Input['AlbVirtualHostRouteGrpcRouteGrpcStatusResponseActionArgs'] grpc_status_response_action: GRPC status response action resource. The structure is documented below. """ if grpc_matches is not None: pulumi.set(__self__, "grpc_matches", grpc_matches) if grpc_route_action is not None: pulumi.set(__self__, "grpc_route_action", grpc_route_action) if grpc_status_response_action is not None: pulumi.set(__self__, "grpc_status_response_action", grpc_status_response_action) @property @pulumi.getter(name="grpcMatches") def grpc_matches(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['AlbVirtualHostRouteGrpcRouteGrpcMatchArgs']]]]: """ Checks "/" prefix by default. The structure is documented below. """ return pulumi.get(self, "grpc_matches") @grpc_matches.setter def grpc_matches(self, value: Optional[pulumi.Input[Sequence[pulumi.Input['AlbVirtualHostRouteGrpcRouteGrpcMatchArgs']]]]): pulumi.set(self, "grpc_matches", value) @property @pulumi.getter(name="grpcRouteAction") def grpc_route_action(self) -> Optional[pulumi.Input['AlbVirtualHostRouteGrpcRouteGrpcRouteActionArgs']]: """ GRPC route action resource. The structure is documented below. """ return pulumi.get(self, "grpc_route_action") @grpc_route_action.setter def grpc_route_action(self, value: Optional[pulumi.Input['AlbVirtualHostRouteGrpcRouteGrpcRouteActionArgs']]): pulumi.set(self, "grpc_route_action", value) @property @pulumi.getter(name="grpcStatusResponseAction") def grpc_status_response_action(self) -> Optional[pulumi.Input['AlbVirtualHostRouteGrpcRouteGrpcStatusResponseActionArgs']]: """ GRPC status response action resource. The structure is documented below. """ return pulumi.get(self, "grpc_status_response_action") @grpc_status_response_action.setter def grpc_status_response_action(self, value: Optional[pulumi.Input['AlbVirtualHostRouteGrpcRouteGrpcStatusResponseActionArgs']]): pulumi.set(self, "grpc_status_response_action", value) @pulumi.input_type class AlbVirtualHostRouteGrpcRouteGrpcMatchArgs: def __init__(__self__, *, fqmn: Optional[pulumi.Input['AlbVirtualHostRouteGrpcRouteGrpcMatchFqmnArgs']] = None): """ :param pulumi.Input['AlbVirtualHostRouteGrpcRouteGrpcMatchFqmnArgs'] fqmn: If not set, all services/methods are assumed. The structure is documented below. """ if fqmn is not None: pulumi.set(__self__, "fqmn", fqmn) @property @pulumi.getter def fqmn(self) -> Optional[pulumi.Input['AlbVirtualHostRouteGrpcRouteGrpcMatchFqmnArgs']]: """ If not set, all services/methods are assumed. The structure is documented below. """ return pulumi.get(self, "fqmn") @fqmn.setter def fqmn(self, value: Optional[pulumi.Input['AlbVirtualHostRouteGrpcRouteGrpcMatchFqmnArgs']]): pulumi.set(self, "fqmn", value) @pulumi.input_type class AlbVirtualHostRouteGrpcRouteGrpcMatchFqmnArgs: def __init__(__self__, *, exact: Optional[pulumi.Input[str]] = None, prefix: Optional[pulumi.Input[str]] = None): if exact is not None: pulumi.set(__self__, "exact", exact) if prefix is not None: pulumi.set(__self__, "prefix", prefix) @property @pulumi.getter def exact(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "exact") @exact.setter def exact(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "exact", value) @property @pulumi.getter def prefix(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "prefix") @prefix.setter def prefix(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "prefix", value) @pulumi.input_type class AlbVirtualHostRouteGrpcRouteGrpcRouteActionArgs: def __init__(__self__, *, backend_group_id: pulumi.Input[str], auto_host_rewrite: Optional[pulumi.Input[bool]] = None, host_rewrite: Optional[pulumi.Input[str]] = None, idle_timeout: Optional[pulumi.Input[str]] = None, max_timeout: Optional[pulumi.Input[str]] = None): """ :param pulumi.Input[str] backend_group_id: Backend group to route requests. :param pulumi.Input[bool] auto_host_rewrite: If set, will automatically rewrite host. :param pulumi.Input[str] host_rewrite: Host rewrite specifier. :param pulumi.Input[str] idle_timeout: Specifies the idle timeout (time without any data transfer for the active request) for the route. It is useful for streaming scenarios - one should set idle_timeout to something meaningful and max_timeout to the maximum time the stream is allowed to be alive. If not specified, there is no per-route idle timeout. :param pulumi.Input[str] max_timeout: Lower timeout may be specified by the client (using grpc-timeout header). If not set, default is 60 seconds. """ pulumi.set(__self__, "backend_group_id", backend_group_id) if auto_host_rewrite is not None: pulumi.set(__self__, "auto_host_rewrite", auto_host_rewrite) if host_rewrite is not None: pulumi.set(__self__, "host_rewrite", host_rewrite) if idle_timeout is not None: pulumi.set(__self__, "idle_timeout", idle_timeout) if max_timeout is not None: pulumi.set(__self__, "max_timeout", max_timeout) @property @pulumi.getter(name="backendGroupId") def backend_group_id(self) -> pulumi.Input[str]: """ Backend group to route requests. """ return pulumi.get(self, "backend_group_id") @backend_group_id.setter def backend_group_id(self, value: pulumi.Input[str]): pulumi.set(self, "backend_group_id", value) @property @pulumi.getter(name="autoHostRewrite") def auto_host_rewrite(self) -> Optional[pulumi.Input[bool]]: """ If set, will automatically rewrite host. """ return pulumi.get(self, "auto_host_rewrite") @auto_host_rewrite.setter def auto_host_rewrite(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "auto_host_rewrite", value) @property @pulumi.getter(name="hostRewrite") def host_rewrite(self) -> Optional[pulumi.Input[str]]: """ Host rewrite specifier. """ return pulumi.get(self, "host_rewrite") @host_rewrite.setter def host_rewrite(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "host_rewrite", value) @property @pulumi.getter(name="idleTimeout") def idle_timeout(self) -> Optional[pulumi.Input[str]]: """ Specifies the idle timeout (time without any data transfer for the active request) for the route. It is useful for streaming scenarios - one should set idle_timeout to something meaningful and max_timeout to the maximum time the stream is allowed to be alive. If not specified, there is no per-route idle timeout. """ return pulumi.get(self, "idle_timeout") @idle_timeout.setter def idle_timeout(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "idle_timeout", value) @property @pulumi.getter(name="maxTimeout") def max_timeout(self) -> Optional[pulumi.Input[str]]: """ Lower timeout may be specified by the client (using grpc-timeout header). If not set, default is 60 seconds. """ return pulumi.get(self, "max_timeout") @max_timeout.setter def max_timeout(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "max_timeout", value) @pulumi.input_type class AlbVirtualHostRouteGrpcRouteGrpcStatusResponseActionArgs: def __init__(__self__, *, status: Optional[pulumi.Input[str]] = None): """ :param pulumi.Input[str] status: The status of the response. Supported values are: ok, invalid_argumet, not_found, permission_denied, unauthenticated, unimplemented, internal, unavailable. """ if status is not None: pulumi.set(__self__, "status", status) @property @pulumi.getter def status(self) -> Optional[pulumi.Input[str]]: """ The status of the response. Supported values are: ok, invalid_argumet, not_found, permission_denied, unauthenticated, unimplemented, internal, unavailable. """ return pulumi.get(self, "status") @status.setter def status(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "status", value) @pulumi.input_type class AlbVirtualHostRouteHttpRouteArgs: def __init__(__self__, *, direct_response_action: Optional[pulumi.Input['AlbVirtualHostRouteHttpRouteDirectResponseActionArgs']] = None, http_matches: Optional[pulumi.Input[Sequence[pulumi.Input['AlbVirtualHostRouteHttpRouteHttpMatchArgs']]]] = None, http_route_action: Optional[pulumi.Input['AlbVirtualHostRouteHttpRouteHttpRouteActionArgs']] = None, redirect_action: Optional[pulumi.Input['AlbVirtualHostRouteHttpRouteRedirectActionArgs']] = None): """ :param pulumi.Input['AlbVirtualHostRouteHttpRouteDirectResponseActionArgs'] direct_response_action: Direct response action resource. The structure is documented below. :param pulumi.Input[Sequence[pulumi.Input['AlbVirtualHostRouteHttpRouteHttpMatchArgs']]] http_matches: Checks "/" prefix by default. The structure is documented below. :param pulumi.Input['AlbVirtualHostRouteHttpRouteHttpRouteActionArgs'] http_route_action: HTTP route action resource. The structure is documented below. :param pulumi.Input['AlbVirtualHostRouteHttpRouteRedirectActionArgs'] redirect_action: Redirect action resource. The structure is documented below. """ if direct_response_action is not None: pulumi.set(__self__, "direct_response_action", direct_response_action) if http_matches is not None: pulumi.set(__self__, "http_matches", http_matches) if http_route_action is not None: pulumi.set(__self__, "http_route_action", http_route_action) if redirect_action is not None: pulumi.set(__self__, "redirect_action", redirect_action) @property @pulumi.getter(name="directResponseAction") def direct_response_action(self) -> Optional[pulumi.Input['AlbVirtualHostRouteHttpRouteDirectResponseActionArgs']]: """ Direct response action resource. The structure is documented below. """ return pulumi.get(self, "direct_response_action") @direct_response_action.setter def direct_response_action(self, value: Optional[pulumi.Input['AlbVirtualHostRouteHttpRouteDirectResponseActionArgs']]): pulumi.set(self, "direct_response_action", value) @property @pulumi.getter(name="httpMatches") def http_matches(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['AlbVirtualHostRouteHttpRouteHttpMatchArgs']]]]: """ Checks "/" prefix by default. The structure is documented below. """ return pulumi.get(self, "http_matches") @http_matches.setter def http_matches(self, value: Optional[pulumi.Input[Sequence[pulumi.Input['AlbVirtualHostRouteHttpRouteHttpMatchArgs']]]]): pulumi.set(self, "http_matches", value) @property @pulumi.getter(name="httpRouteAction") def http_route_action(self) -> Optional[pulumi.Input['AlbVirtualHostRouteHttpRouteHttpRouteActionArgs']]: """ HTTP route action resource. The structure is documented below. """ return pulumi.get(self, "http_route_action") @http_route_action.setter def http_route_action(self, value: Optional[pulumi.Input['AlbVirtualHostRouteHttpRouteHttpRouteActionArgs']]): pulumi.set(self, "http_route_action", value) @property @pulumi.getter(name="redirectAction") def redirect_action(self) -> Optional[pulumi.Input['AlbVirtualHostRouteHttpRouteRedirectActionArgs']]: """ Redirect action resource. The structure is documented below. """ return pulumi.get(self, "redirect_action") @redirect_action.setter def redirect_action(self, value: Optional[pulumi.Input['AlbVirtualHostRouteHttpRouteRedirectActionArgs']]): pulumi.set(self, "redirect_action", value) @pulumi.input_type class AlbVirtualHostRouteHttpRouteDirectResponseActionArgs: def __init__(__self__, *, body: Optional[pulumi.Input[str]] = None, status: Optional[pulumi.Input[int]] = None): """ :param pulumi.Input[str] body: Response body text. :param pulumi.Input[int] status: The status of the response. Supported values are: ok, invalid_argumet, not_found, permission_denied, unauthenticated, unimplemented, internal, unavailable. """ if body is not None: pulumi.set(__self__, "body", body) if status is not None: pulumi.set(__self__, "status", status) @property @pulumi.getter def body(self) -> Optional[pulumi.Input[str]]: """ Response body text. """ return pulumi.get(self, "body") @body.setter def body(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "body", value) @property @pulumi.getter def status(self) -> Optional[pulumi.Input[int]]: """ The status of the response. Supported values are: ok, invalid_argumet, not_found, permission_denied, unauthenticated, unimplemented, internal, unavailable. """ return pulumi.get(self, "status") @status.setter def status(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "status", value) @pulumi.input_type class AlbVirtualHostRouteHttpRouteHttpMatchArgs: def __init__(__self__, *, http_methods: Optional[pulumi.Input[Sequence[Any]]] = None, path: Optional[pulumi.Input['AlbVirtualHostRouteHttpRouteHttpMatchPathArgs']] = None): """ :param pulumi.Input[Sequence[Any]] http_methods: List of methods(strings). :param pulumi.Input['AlbVirtualHostRouteHttpRouteHttpMatchPathArgs'] path: If not set, '/' is assumed. The structure is documented below. """ if http_methods is not None: pulumi.set(__self__, "http_methods", http_methods) if path is not None: pulumi.set(__self__, "path", path) @property @pulumi.getter(name="httpMethods") def http_methods(self) -> Optional[pulumi.Input[Sequence[Any]]]: """ List of methods(strings). """ return pulumi.get(self, "http_methods") @http_methods.setter def http_methods(self, value: Optional[pulumi.Input[Sequence[Any]]]): pulumi.set(self, "http_methods", value) @property @pulumi.getter def path(self) -> Optional[pulumi.Input['AlbVirtualHostRouteHttpRouteHttpMatchPathArgs']]: """ If not set, '/' is assumed. The structure is documented below. """ return pulumi.get(self, "path") @path.setter def path(self, value: Optional[pulumi.Input['AlbVirtualHostRouteHttpRouteHttpMatchPathArgs']]): pulumi.set(self, "path", value) @pulumi.input_type class AlbVirtualHostRouteHttpRouteHttpMatchPathArgs: def __init__(__self__, *, exact: Optional[pulumi.Input[str]] = None, prefix: Optional[pulumi.Input[str]] = None): if exact is not None: pulumi.set(__self__, "exact", exact) if prefix is not None: pulumi.set(__self__, "prefix", prefix) @property @pulumi.getter def exact(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "exact") @exact.setter def exact(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "exact", value) @property @pulumi.getter def prefix(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "prefix") @prefix.setter def prefix(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "prefix", value) @pulumi.input_type class AlbVirtualHostRouteHttpRouteHttpRouteActionArgs: def __init__(__self__, *, backend_group_id: pulumi.Input[str], auto_host_rewrite: Optional[pulumi.Input[bool]] = None, host_rewrite: Optional[pulumi.Input[str]] = None, idle_timeout: Optional[pulumi.Input[str]] = None, prefix_rewrite: Optional[pulumi.Input[str]] = None, timeout: Optional[pulumi.Input[str]] = None, upgrade_types: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None): """ :param pulumi.Input[str] backend_group_id: Backend group to route requests. :param pulumi.Input[bool] auto_host_rewrite: If set, will automatically rewrite host. :param pulumi.Input[str] host_rewrite: Host rewrite specifier. :param pulumi.Input[str] idle_timeout: Specifies the idle timeout (time without any data transfer for the active request) for the route. It is useful for streaming scenarios - one should set idle_timeout to something meaningful and max_timeout to the maximum time the stream is allowed to be alive. If not specified, there is no per-route idle timeout. :param pulumi.Input[str] prefix_rewrite: If not empty, matched path prefix will be replaced by this value. :param pulumi.Input[str] timeout: Specifies the request timeout (overall time request processing is allowed to take) for the route. If not set, default is 60 seconds. :param pulumi.Input[Sequence[pulumi.Input[str]]] upgrade_types: List of upgrade types. Only specified upgrade types will be allowed. For example, "websocket". """ pulumi.set(__self__, "backend_group_id", backend_group_id) if auto_host_rewrite is not None: pulumi.set(__self__, "auto_host_rewrite", auto_host_rewrite) if host_rewrite is not None: pulumi.set(__self__, "host_rewrite", host_rewrite) if idle_timeout is not None: pulumi.set(__self__, "idle_timeout", idle_timeout) if prefix_rewrite is not None: pulumi.set(__self__, "prefix_rewrite", prefix_rewrite) if timeout is not None: pulumi.set(__self__, "timeout", timeout) if upgrade_types is not None: pulumi.set(__self__, "upgrade_types", upgrade_types) @property @pulumi.getter(name="backendGroupId") def backend_group_id(self) -> pulumi.Input[str]: """ Backend group to route requests. """ return pulumi.get(self, "backend_group_id") @backend_group_id.setter def backend_group_id(self, value: pulumi.Input[str]): pulumi.set(self, "backend_group_id", value) @property @pulumi.getter(name="autoHostRewrite") def auto_host_rewrite(self) -> Optional[pulumi.Input[bool]]: """ If set, will automatically rewrite host. """ return pulumi.get(self, "auto_host_rewrite") @auto_host_rewrite.setter def auto_host_rewrite(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "auto_host_rewrite", value) @property @pulumi.getter(name="hostRewrite") def host_rewrite(self) -> Optional[pulumi.Input[str]]: """ Host rewrite specifier. """ return pulumi.get(self, "host_rewrite") @host_rewrite.setter def host_rewrite(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "host_rewrite", value) @property @pulumi.getter(name="idleTimeout") def idle_timeout(self) -> Optional[pulumi.Input[str]]: """ Specifies the idle timeout (time without any data transfer for the active request) for the route. It is useful for streaming scenarios - one should set idle_timeout to something meaningful and max_timeout to the maximum time the stream is allowed to be alive. If not specified, there is no per-route idle timeout. """ return pulumi.get(self, "idle_timeout") @idle_timeout.setter def idle_timeout(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "idle_timeout", value) @property @pulumi.getter(name="prefixRewrite") def prefix_rewrite(self) -> Optional[pulumi.Input[str]]: """ If not empty, matched path prefix will be replaced by this value. """ return pulumi.get(self, "prefix_rewrite") @prefix_rewrite.setter def prefix_rewrite(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "prefix_rewrite", value) @property @pulumi.getter def timeout(self) -> Optional[pulumi.Input[str]]: """ Specifies the request timeout (overall time request processing is allowed to take) for the route. If not set, default is 60 seconds. """ return pulumi.get(self, "timeout") @timeout.setter def timeout(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "timeout", value) @property @pulumi.getter(name="upgradeTypes") def upgrade_types(self) -> Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]: """ List of upgrade types. Only specified upgrade types will be allowed. For example, "websocket". """ return pulumi.get(self, "upgrade_types") @upgrade_types.setter def upgrade_types(self, value: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]): pulumi.set(self, "upgrade_types", value) @pulumi.input_type class AlbVirtualHostRouteHttpRouteRedirectActionArgs: def __init__(__self__, *, remove_query: Optional[pulumi.Input[bool]] = None, replace_host: Optional[pulumi.Input[str]] = None, replace_path: Optional[pulumi.Input[str]] = None, replace_port: Optional[pulumi.Input[int]] = None, replace_prefix: Optional[pulumi.Input[str]] = None, replace_scheme: Optional[pulumi.Input[str]] = None, response_code: Optional[pulumi.Input[str]] = None): """ :param pulumi.Input[str] replace_host: Replaces hostname. :param pulumi.Input[str] replace_path: Replace path. :param pulumi.Input[int] replace_port: Replaces port. :param pulumi.Input[str] replace_prefix: Replace only matched prefix. Example:<br/> match:{ prefix_match: "/some" } <br/> redirect: { replace_prefix: "/other" } <br/> will redirect "/something" to "/otherthing". :param pulumi.Input[str] replace_scheme: Replaces scheme. If the original scheme is `http` or `https`, will also remove the 80 or 443 port, if present. :param pulumi.Input[str] response_code: The HTTP status code to use in the redirect response. Supported values are: moved_permanently, found, see_other, temporary_redirect, permanent_redirect. """ if remove_query is not None: pulumi.set(__self__, "remove_query", remove_query) if replace_host is not None: pulumi.set(__self__, "replace_host", replace_host) if replace_path is not None: pulumi.set(__self__, "replace_path", replace_path) if replace_port is not None: pulumi.set(__self__, "replace_port", replace_port) if replace_prefix is not None: pulumi.set(__self__, "replace_prefix", replace_prefix) if replace_scheme is not None: pulumi.set(__self__, "replace_scheme", replace_scheme) if response_code is not None: pulumi.set(__self__, "response_code", response_code) @property @pulumi.getter(name="removeQuery") def remove_query(self) -> Optional[pulumi.Input[bool]]: return pulumi.get(self, "remove_query") @remove_query.setter def remove_query(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "remove_query", value) @property @pulumi.getter(name="replaceHost") def replace_host(self) -> Optional[pulumi.Input[str]]: """ Replaces hostname. """ return pulumi.get(self, "replace_host") @replace_host.setter def replace_host(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "replace_host", value) @property @pulumi.getter(name="replacePath") def replace_path(self) -> Optional[pulumi.Input[str]]: """ Replace path. """ return pulumi.get(self, "replace_path") @replace_path.setter def replace_path(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "replace_path", value) @property @pulumi.getter(name="replacePort") def replace_port(self) -> Optional[pulumi.Input[int]]: """ Replaces port. """ return pulumi.get(self, "replace_port") @replace_port.setter def replace_port(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "replace_port", value) @property @pulumi.getter(name="replacePrefix") def replace_prefix(self) -> Optional[pulumi.Input[str]]: """ Replace only matched prefix. Example:<br/> match:{ prefix_match: "/some" } <br/> redirect: { replace_prefix: "/other" } <br/> will redirect "/something" to "/otherthing". """ return pulumi.get(self, "replace_prefix") @replace_prefix.setter def replace_prefix(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "replace_prefix", value) @property @pulumi.getter(name="replaceScheme") def replace_scheme(self) -> Optional[pulumi.Input[str]]: """ Replaces scheme. If the original scheme is `http` or `https`, will also remove the 80 or 443 port, if present. """ return pulumi.get(self, "replace_scheme") @replace_scheme.setter def replace_scheme(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "replace_scheme", value) @property @pulumi.getter(name="responseCode") def response_code(self) -> Optional[pulumi.Input[str]]: """ The HTTP status code to use in the redirect response. Supported values are: moved_permanently, found, see_other, temporary_redirect, permanent_redirect. """ return pulumi.get(self, "response_code") @response_code.setter def response_code(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "response_code", value) @pulumi.input_type class ComputeDiskDiskPlacementPolicyArgs: def __init__(__self__, *, disk_placement_group_id: pulumi.Input[str]): """ :param pulumi.Input[str] disk_placement_group_id: Specifies Disk Placement Group id. """ pulumi.set(__self__, "disk_placement_group_id", disk_placement_group_id) @property @pulumi.getter(name="diskPlacementGroupId") def disk_placement_group_id(self) -> pulumi.Input[str]: """ Specifies Disk Placement Group id. """ return pulumi.get(self, "disk_placement_group_id") @disk_placement_group_id.setter def disk_placement_group_id(self, value: pulumi.Input[str]): pulumi.set(self, "disk_placement_group_id", value) @pulumi.input_type class ComputeInstanceBootDiskArgs: def __init__(__self__, *, auto_delete: Optional[pulumi.Input[bool]] = None, device_name: Optional[pulumi.Input[str]] = None, disk_id: Optional[pulumi.Input[str]] = None, initialize_params: Optional[pulumi.Input['ComputeInstanceBootDiskInitializeParamsArgs']] = None, mode: Optional[pulumi.Input[str]] = None): """ :param pulumi.Input[bool] auto_delete: Whether the disk is auto-deleted when the instance is deleted. The default value is false. :param pulumi.Input[str] device_name: Name that can be used to access an attached disk under `/dev/disk/by-id/`. :param pulumi.Input[str] disk_id: ID of the disk that is attached to the instance. :param pulumi.Input['ComputeInstanceBootDiskInitializeParamsArgs'] initialize_params: Parameters for a new disk that will be created alongside the new instance. Either `initialize_params` or `disk_id` must be set. The structure is documented below. :param pulumi.Input[str] mode: Type of access to the disk resource. By default, a disk is attached in `READ_WRITE` mode. """ if auto_delete is not None: pulumi.set(__self__, "auto_delete", auto_delete) if device_name is not None: pulumi.set(__self__, "device_name", device_name) if disk_id is not None: pulumi.set(__self__, "disk_id", disk_id) if initialize_params is not None: pulumi.set(__self__, "initialize_params", initialize_params) if mode is not None: pulumi.set(__self__, "mode", mode) @property @pulumi.getter(name="autoDelete") def auto_delete(self) -> Optional[pulumi.Input[bool]]: """ Whether the disk is auto-deleted when the instance is deleted. The default value is false. """ return pulumi.get(self, "auto_delete") @auto_delete.setter def auto_delete(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "auto_delete", value) @property @pulumi.getter(name="deviceName") def device_name(self) -> Optional[pulumi.Input[str]]: """ Name that can be used to access an attached disk under `/dev/disk/by-id/`. """ return pulumi.get(self, "device_name") @device_name.setter def device_name(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "device_name", value) @property @pulumi.getter(name="diskId") def disk_id(self) -> Optional[pulumi.Input[str]]: """ ID of the disk that is attached to the instance. """ return pulumi.get(self, "disk_id") @disk_id.setter def disk_id(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "disk_id", value) @property @pulumi.getter(name="initializeParams") def initialize_params(self) -> Optional[pulumi.Input['ComputeInstanceBootDiskInitializeParamsArgs']]: """ Parameters for a new disk that will be created alongside the new instance. Either `initialize_params` or `disk_id` must be set. The structure is documented below. """ return pulumi.get(self, "initialize_params") @initialize_params.setter def initialize_params(self, value: Optional[pulumi.Input['ComputeInstanceBootDiskInitializeParamsArgs']]): pulumi.set(self, "initialize_params", value) @property @pulumi.getter def mode(self) -> Optional[pulumi.Input[str]]: """ Type of access to the disk resource. By default, a disk is attached in `READ_WRITE` mode. """ return pulumi.get(self, "mode") @mode.setter def mode(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "mode", value) @pulumi.input_type class ComputeInstanceBootDiskInitializeParamsArgs: def __init__(__self__, *, description: Optional[pulumi.Input[str]] = None, image_id: Optional[pulumi.Input[str]] = None, name: Optional[pulumi.Input[str]] = None, size: Optional[pulumi.Input[int]] = None, snapshot_id: Optional[pulumi.Input[str]] = None, type: Optional[pulumi.Input[str]] = None): """ :param pulumi.Input[str] description: Description of the boot disk. :param pulumi.Input[str] image_id: A disk image to initialize this disk from. :param pulumi.Input[str] name: Name of the boot disk. :param pulumi.Input[int] size: Size of the disk in GB. :param pulumi.Input[str] snapshot_id: A snapshot to initialize this disk from. :param pulumi.Input[str] type: Disk type. """ if description is not None: pulumi.set(__self__, "description", description) if image_id is not None: pulumi.set(__self__, "image_id", image_id) if name is not None: pulumi.set(__self__, "name", name) if size is not None: pulumi.set(__self__, "size", size) if snapshot_id is not None: pulumi.set(__self__, "snapshot_id", snapshot_id) if type is not None: pulumi.set(__self__, "type", type) @property @pulumi.getter def description(self) -> Optional[pulumi.Input[str]]: """ Description of the boot disk. """ return pulumi.get(self, "description") @description.setter def description(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "description", value) @property @pulumi.getter(name="imageId") def image_id(self) -> Optional[pulumi.Input[str]]: """ A disk image to initialize this disk from. """ return pulumi.get(self, "image_id") @image_id.setter def image_id(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "image_id", value) @property @pulumi.getter def name(self) -> Optional[pulumi.Input[str]]: """ Name of the boot disk. """ return pulumi.get(self, "name") @name.setter def name(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "name", value) @property @pulumi.getter def size(self) -> Optional[pulumi.Input[int]]: """ Size of the disk in GB. """ return pulumi.get(self, "size") @size.setter def size(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "size", value) @property @pulumi.getter(name="snapshotId") def snapshot_id(self) -> Optional[pulumi.Input[str]]: """ A snapshot to initialize this disk from. """ return pulumi.get(self, "snapshot_id") @snapshot_id.setter def snapshot_id(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "snapshot_id", value) @property @pulumi.getter def type(self) -> Optional[pulumi.Input[str]]: """ Disk type. """ return pulumi.get(self, "type") @type.setter def type(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "type", value) @pulumi.input_type class ComputeInstanceGroupAllocationPolicyArgs: def __init__(__self__, *, zones: pulumi.Input[Sequence[pulumi.Input[str]]]): """ :param pulumi.Input[Sequence[pulumi.Input[str]]] zones: A list of availability zones. """ pulumi.set(__self__, "zones", zones) @property @pulumi.getter def zones(self) -> pulumi.Input[Sequence[pulumi.Input[str]]]: """ A list of availability zones. """ return pulumi.get(self, "zones") @zones.setter def zones(self, value: pulumi.Input[Sequence[pulumi.Input[str]]]): pulumi.set(self, "zones", value) @pulumi.input_type class ComputeInstanceGroupApplicationLoadBalancerArgs: def __init__(__self__, *, max_opening_traffic_duration: Optional[pulumi.Input[int]] = None, status_message: Optional[pulumi.Input[str]] = None, target_group_description: Optional[pulumi.Input[str]] = None, target_group_id: Optional[pulumi.Input[str]] = None, target_group_labels: Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]] = None, target_group_name: Optional[pulumi.Input[str]] = None): """ :param pulumi.Input[int] max_opening_traffic_duration: Timeout for waiting for the VM to be checked by the load balancer. If the timeout is exceeded, the VM will be turned off based on the deployment policy. Specified in seconds. :param pulumi.Input[str] status_message: The status message of the instance. :param pulumi.Input[str] target_group_description: A description of the target group. :param pulumi.Input[Mapping[str, pulumi.Input[str]]] target_group_labels: A set of key/value label pairs. :param pulumi.Input[str] target_group_name: The name of the target group. """ if max_opening_traffic_duration is not None: pulumi.set(__self__, "max_opening_traffic_duration", max_opening_traffic_duration) if status_message is not None: pulumi.set(__self__, "status_message", status_message) if target_group_description is not None: pulumi.set(__self__, "target_group_description", target_group_description) if target_group_id is not None: pulumi.set(__self__, "target_group_id", target_group_id) if target_group_labels is not None: pulumi.set(__self__, "target_group_labels", target_group_labels) if target_group_name is not None: pulumi.set(__self__, "target_group_name", target_group_name) @property @pulumi.getter(name="maxOpeningTrafficDuration") def max_opening_traffic_duration(self) -> Optional[pulumi.Input[int]]: """ Timeout for waiting for the VM to be checked by the load balancer. If the timeout is exceeded, the VM will be turned off based on the deployment policy. Specified in seconds. """ return pulumi.get(self, "max_opening_traffic_duration") @max_opening_traffic_duration.setter def max_opening_traffic_duration(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "max_opening_traffic_duration", value) @property @pulumi.getter(name="statusMessage") def status_message(self) -> Optional[pulumi.Input[str]]: """ The status message of the instance. """ return pulumi.get(self, "status_message") @status_message.setter def status_message(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "status_message", value) @property @pulumi.getter(name="targetGroupDescription") def target_group_description(self) -> Optional[pulumi.Input[str]]: """ A description of the target group. """ return pulumi.get(self, "target_group_description") @target_group_description.setter def target_group_description(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "target_group_description", value) @property @pulumi.getter(name="targetGroupId") def target_group_id(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "target_group_id") @target_group_id.setter def target_group_id(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "target_group_id", value) @property @pulumi.getter(name="targetGroupLabels") def target_group_labels(self) -> Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]]: """ A set of key/value label pairs. """ return pulumi.get(self, "target_group_labels") @target_group_labels.setter def target_group_labels(self, value: Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]]): pulumi.set(self, "target_group_labels", value) @property @pulumi.getter(name="targetGroupName") def target_group_name(self) -> Optional[pulumi.Input[str]]: """ The name of the target group. """ return pulumi.get(self, "target_group_name") @target_group_name.setter def target_group_name(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "target_group_name", value) @pulumi.input_type class ComputeInstanceGroupDeployPolicyArgs: def __init__(__self__, *, max_expansion: pulumi.Input[int], max_unavailable: pulumi.Input[int], max_creating: Optional[pulumi.Input[int]] = None, max_deleting: Optional[pulumi.Input[int]] = None, startup_duration: Optional[pulumi.Input[int]] = None, strategy: Optional[pulumi.Input[str]] = None): """ :param pulumi.Input[int] max_expansion: The maximum number of instances that can be temporarily allocated above the group's target size during the update process. :param pulumi.Input[int] max_unavailable: The maximum number of running instances that can be taken offline (stopped or deleted) at the same time during the update process. :param pulumi.Input[int] max_creating: The maximum number of instances that can be created at the same time. :param pulumi.Input[int] max_deleting: The maximum number of instances that can be deleted at the same time. :param pulumi.Input[int] startup_duration: The amount of time in seconds to allow for an instance to start. Instance will be considered up and running (and start receiving traffic) only after the startup_duration has elapsed and all health checks are passed. :param pulumi.Input[str] strategy: Affects the lifecycle of the instance during deployment. If set to `proactive` (default), Instance Groups can forcefully stop a running instance. If `opportunistic`, Instance Groups does not stop a running instance. Instead, it will wait until the instance stops itself or becomes unhealthy. """ pulumi.set(__self__, "max_expansion", max_expansion) pulumi.set(__self__, "max_unavailable", max_unavailable) if max_creating is not None: pulumi.set(__self__, "max_creating", max_creating) if max_deleting is not None: pulumi.set(__self__, "max_deleting", max_deleting) if startup_duration is not None: pulumi.set(__self__, "startup_duration", startup_duration) if strategy is not None: pulumi.set(__self__, "strategy", strategy) @property @pulumi.getter(name="maxExpansion") def max_expansion(self) -> pulumi.Input[int]: """ The maximum number of instances that can be temporarily allocated above the group's target size during the update process. """ return pulumi.get(self, "max_expansion") @max_expansion.setter def max_expansion(self, value: pulumi.Input[int]): pulumi.set(self, "max_expansion", value) @property @pulumi.getter(name="maxUnavailable") def max_unavailable(self) -> pulumi.Input[int]: """ The maximum number of running instances that can be taken offline (stopped or deleted) at the same time during the update process. """ return pulumi.get(self, "max_unavailable") @max_unavailable.setter def max_unavailable(self, value: pulumi.Input[int]): pulumi.set(self, "max_unavailable", value) @property @pulumi.getter(name="maxCreating") def max_creating(self) -> Optional[pulumi.Input[int]]: """ The maximum number of instances that can be created at the same time. """ return pulumi.get(self, "max_creating") @max_creating.setter def max_creating(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "max_creating", value) @property @pulumi.getter(name="maxDeleting") def max_deleting(self) -> Optional[pulumi.Input[int]]: """ The maximum number of instances that can be deleted at the same time. """ return pulumi.get(self, "max_deleting") @max_deleting.setter def max_deleting(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "max_deleting", value) @property @pulumi.getter(name="startupDuration") def startup_duration(self) -> Optional[pulumi.Input[int]]: """ The amount of time in seconds to allow for an instance to start. Instance will be considered up and running (and start receiving traffic) only after the startup_duration has elapsed and all health checks are passed. """ return pulumi.get(self, "startup_duration") @startup_duration.setter def startup_duration(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "startup_duration", value) @property @pulumi.getter def strategy(self) -> Optional[pulumi.Input[str]]: """ Affects the lifecycle of the instance during deployment. If set to `proactive` (default), Instance Groups can forcefully stop a running instance. If `opportunistic`, Instance Groups does not stop a running instance. Instead, it will wait until the instance stops itself or becomes unhealthy. """ return pulumi.get(self, "strategy") @strategy.setter def strategy(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "strategy", value) @pulumi.input_type class ComputeInstanceGroupHealthCheckArgs: def __init__(__self__, *, healthy_threshold: Optional[pulumi.Input[int]] = None, http_options: Optional[pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupHealthCheckHttpOptionArgs']]]] = None, interval: Optional[pulumi.Input[int]] = None, tcp_options: Optional[pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupHealthCheckTcpOptionArgs']]]] = None, timeout: Optional[pulumi.Input[int]] = None, unhealthy_threshold: Optional[pulumi.Input[int]] = None): """ :param pulumi.Input[int] healthy_threshold: The number of successful health checks before the managed instance is declared healthy. :param pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupHealthCheckHttpOptionArgs']]] http_options: HTTP check options. The structure is documented below. :param pulumi.Input[int] interval: The interval to wait between health checks in seconds. :param pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupHealthCheckTcpOptionArgs']]] tcp_options: TCP check options. The structure is documented below. :param pulumi.Input[int] timeout: The length of time to wait for a response before the health check times out in seconds. :param pulumi.Input[int] unhealthy_threshold: The number of failed health checks before the managed instance is declared unhealthy. """ if healthy_threshold is not None: pulumi.set(__self__, "healthy_threshold", healthy_threshold) if http_options is not None: pulumi.set(__self__, "http_options", http_options) if interval is not None: pulumi.set(__self__, "interval", interval) if tcp_options is not None: pulumi.set(__self__, "tcp_options", tcp_options) if timeout is not None: pulumi.set(__self__, "timeout", timeout) if unhealthy_threshold is not None: pulumi.set(__self__, "unhealthy_threshold", unhealthy_threshold) @property @pulumi.getter(name="healthyThreshold") def healthy_threshold(self) -> Optional[pulumi.Input[int]]: """ The number of successful health checks before the managed instance is declared healthy. """ return pulumi.get(self, "healthy_threshold") @healthy_threshold.setter def healthy_threshold(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "healthy_threshold", value) @property @pulumi.getter(name="httpOptions") def http_options(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupHealthCheckHttpOptionArgs']]]]: """ HTTP check options. The structure is documented below. """ return pulumi.get(self, "http_options") @http_options.setter def http_options(self, value: Optional[pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupHealthCheckHttpOptionArgs']]]]): pulumi.set(self, "http_options", value) @property @pulumi.getter def interval(self) -> Optional[pulumi.Input[int]]: """ The interval to wait between health checks in seconds. """ return pulumi.get(self, "interval") @interval.setter def interval(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "interval", value) @property @pulumi.getter(name="tcpOptions") def tcp_options(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupHealthCheckTcpOptionArgs']]]]: """ TCP check options. The structure is documented below. """ return pulumi.get(self, "tcp_options") @tcp_options.setter def tcp_options(self, value: Optional[pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupHealthCheckTcpOptionArgs']]]]): pulumi.set(self, "tcp_options", value) @property @pulumi.getter def timeout(self) -> Optional[pulumi.Input[int]]: """ The length of time to wait for a response before the health check times out in seconds. """ return pulumi.get(self, "timeout") @timeout.setter def timeout(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "timeout", value) @property @pulumi.getter(name="unhealthyThreshold") def unhealthy_threshold(self) -> Optional[pulumi.Input[int]]: """ The number of failed health checks before the managed instance is declared unhealthy. """ return pulumi.get(self, "unhealthy_threshold") @unhealthy_threshold.setter def unhealthy_threshold(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "unhealthy_threshold", value) @pulumi.input_type class ComputeInstanceGroupHealthCheckHttpOptionArgs: def __init__(__self__, *, path: pulumi.Input[str], port: pulumi.Input[int]): """ :param pulumi.Input[str] path: The URL path used for health check requests. :param pulumi.Input[int] port: The port used for TCP health checks. """ pulumi.set(__self__, "path", path) pulumi.set(__self__, "port", port) @property @pulumi.getter def path(self) -> pulumi.Input[str]: """ The URL path used for health check requests. """ return pulumi.get(self, "path") @path.setter def path(self, value: pulumi.Input[str]): pulumi.set(self, "path", value) @property @pulumi.getter def port(self) -> pulumi.Input[int]: """ The port used for TCP health checks. """ return pulumi.get(self, "port") @port.setter def port(self, value: pulumi.Input[int]): pulumi.set(self, "port", value) @pulumi.input_type class ComputeInstanceGroupHealthCheckTcpOptionArgs: def __init__(__self__, *, port: pulumi.Input[int]): """ :param pulumi.Input[int] port: The port used for TCP health checks. """ pulumi.set(__self__, "port", port) @property @pulumi.getter def port(self) -> pulumi.Input[int]: """ The port used for TCP health checks. """ return pulumi.get(self, "port") @port.setter def port(self, value: pulumi.Input[int]): pulumi.set(self, "port", value) @pulumi.input_type class ComputeInstanceGroupInstanceArgs: def __init__(__self__, *, fqdn: Optional[pulumi.Input[str]] = None, instance_id: Optional[pulumi.Input[str]] = None, name: Optional[pulumi.Input[str]] = None, network_interfaces: Optional[pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupInstanceNetworkInterfaceArgs']]]] = None, status: Optional[pulumi.Input[str]] = None, status_changed_at: Optional[pulumi.Input[str]] = None, status_message: Optional[pulumi.Input[str]] = None, zone_id: Optional[pulumi.Input[str]] = None): """ :param pulumi.Input[str] fqdn: DNS record fqdn (must have dot at the end). :param pulumi.Input[str] instance_id: The ID of the instance. :param pulumi.Input[str] name: Name template of the instance. In order to be unique it must contain at least one of instance unique placeholders: {instance.short_id} {instance.index} combination of {instance.zone_id} and {instance.index_in_zone} Example: my-instance-{instance.index} If not set, default is used: {instance_group.id}-{instance.short_id} It may also contain another placeholders, see metadata doc for full list. :param pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupInstanceNetworkInterfaceArgs']]] network_interfaces: Network specifications for the instance. This can be used multiple times for adding multiple interfaces. The structure is documented below. :param pulumi.Input[str] status: The status of the instance. :param pulumi.Input[str] status_message: The status message of the instance. :param pulumi.Input[str] zone_id: The ID of the availability zone where the instance resides. """ if fqdn is not None: pulumi.set(__self__, "fqdn", fqdn) if instance_id is not None: pulumi.set(__self__, "instance_id", instance_id) if name is not None: pulumi.set(__self__, "name", name) if network_interfaces is not None: pulumi.set(__self__, "network_interfaces", network_interfaces) if status is not None: pulumi.set(__self__, "status", status) if status_changed_at is not None: pulumi.set(__self__, "status_changed_at", status_changed_at) if status_message is not None: pulumi.set(__self__, "status_message", status_message) if zone_id is not None: pulumi.set(__self__, "zone_id", zone_id) @property @pulumi.getter def fqdn(self) -> Optional[pulumi.Input[str]]: """ DNS record fqdn (must have dot at the end). """ return pulumi.get(self, "fqdn") @fqdn.setter def fqdn(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "fqdn", value) @property @pulumi.getter(name="instanceId") def instance_id(self) -> Optional[pulumi.Input[str]]: """ The ID of the instance. """ return pulumi.get(self, "instance_id") @instance_id.setter def instance_id(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "instance_id", value) @property @pulumi.getter def name(self) -> Optional[pulumi.Input[str]]: """ Name template of the instance. In order to be unique it must contain at least one of instance unique placeholders: {instance.short_id} {instance.index} combination of {instance.zone_id} and {instance.index_in_zone} Example: my-instance-{instance.index} If not set, default is used: {instance_group.id}-{instance.short_id} It may also contain another placeholders, see metadata doc for full list. """ return pulumi.get(self, "name") @name.setter def name(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "name", value) @property @pulumi.getter(name="networkInterfaces") def network_interfaces(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupInstanceNetworkInterfaceArgs']]]]: """ Network specifications for the instance. This can be used multiple times for adding multiple interfaces. The structure is documented below. """ return pulumi.get(self, "network_interfaces") @network_interfaces.setter def network_interfaces(self, value: Optional[pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupInstanceNetworkInterfaceArgs']]]]): pulumi.set(self, "network_interfaces", value) @property @pulumi.getter def status(self) -> Optional[pulumi.Input[str]]: """ The status of the instance. """ return pulumi.get(self, "status") @status.setter def status(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "status", value) @property @pulumi.getter(name="statusChangedAt") def status_changed_at(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "status_changed_at") @status_changed_at.setter def status_changed_at(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "status_changed_at", value) @property @pulumi.getter(name="statusMessage") def status_message(self) -> Optional[pulumi.Input[str]]: """ The status message of the instance. """ return pulumi.get(self, "status_message") @status_message.setter def status_message(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "status_message", value) @property @pulumi.getter(name="zoneId") def zone_id(self) -> Optional[pulumi.Input[str]]: """ The ID of the availability zone where the instance resides. """ return pulumi.get(self, "zone_id") @zone_id.setter def zone_id(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "zone_id", value) @pulumi.input_type class ComputeInstanceGroupInstanceNetworkInterfaceArgs: def __init__(__self__, *, index: Optional[pulumi.Input[int]] = None, ip_address: Optional[pulumi.Input[str]] = None, ipv4: Optional[pulumi.Input[bool]] = None, ipv6: Optional[pulumi.Input[bool]] = None, ipv6_address: Optional[pulumi.Input[str]] = None, mac_address: Optional[pulumi.Input[str]] = None, nat: Optional[pulumi.Input[bool]] = None, nat_ip_address: Optional[pulumi.Input[str]] = None, nat_ip_version: Optional[pulumi.Input[str]] = None, subnet_id: Optional[pulumi.Input[str]] = None): """ :param pulumi.Input[int] index: The index of the network interface as generated by the server. :param pulumi.Input[str] ip_address: Manual set static IP address. :param pulumi.Input[bool] ipv4: True if IPv4 address allocated for the network interface. :param pulumi.Input[str] ipv6_address: Manual set static IPv6 address. :param pulumi.Input[str] mac_address: The MAC address assigned to the network interface. :param pulumi.Input[bool] nat: Flag for using NAT. :param pulumi.Input[str] nat_ip_address: A public address that can be used to access the internet over NAT. Use `variables` to set. :param pulumi.Input[str] nat_ip_version: The IP version for the public address. :param pulumi.Input[str] subnet_id: The ID of the subnet to attach this interface to. The subnet must reside in the same zone where this instance was created. """ if index is not None: pulumi.set(__self__, "index", index) if ip_address is not None: pulumi.set(__self__, "ip_address", ip_address) if ipv4 is not None: pulumi.set(__self__, "ipv4", ipv4) if ipv6 is not None: pulumi.set(__self__, "ipv6", ipv6) if ipv6_address is not None: pulumi.set(__self__, "ipv6_address", ipv6_address) if mac_address is not None: pulumi.set(__self__, "mac_address", mac_address) if nat is not None: pulumi.set(__self__, "nat", nat) if nat_ip_address is not None: pulumi.set(__self__, "nat_ip_address", nat_ip_address) if nat_ip_version is not None: pulumi.set(__self__, "nat_ip_version", nat_ip_version) if subnet_id is not None: pulumi.set(__self__, "subnet_id", subnet_id) @property @pulumi.getter def index(self) -> Optional[pulumi.Input[int]]: """ The index of the network interface as generated by the server. """ return pulumi.get(self, "index") @index.setter def index(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "index", value) @property @pulumi.getter(name="ipAddress") def ip_address(self) -> Optional[pulumi.Input[str]]: """ Manual set static IP address. """ return pulumi.get(self, "ip_address") @ip_address.setter def ip_address(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "ip_address", value) @property @pulumi.getter def ipv4(self) -> Optional[pulumi.Input[bool]]: """ True if IPv4 address allocated for the network interface. """ return pulumi.get(self, "ipv4") @ipv4.setter def ipv4(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "ipv4", value) @property @pulumi.getter def ipv6(self) -> Optional[pulumi.Input[bool]]: return pulumi.get(self, "ipv6") @ipv6.setter def ipv6(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "ipv6", value) @property @pulumi.getter(name="ipv6Address") def ipv6_address(self) -> Optional[pulumi.Input[str]]: """ Manual set static IPv6 address. """ return pulumi.get(self, "ipv6_address") @ipv6_address.setter def ipv6_address(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "ipv6_address", value) @property @pulumi.getter(name="macAddress") def mac_address(self) -> Optional[pulumi.Input[str]]: """ The MAC address assigned to the network interface. """ return pulumi.get(self, "mac_address") @mac_address.setter def mac_address(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "mac_address", value) @property @pulumi.getter def nat(self) -> Optional[pulumi.Input[bool]]: """ Flag for using NAT. """ return pulumi.get(self, "nat") @nat.setter def nat(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "nat", value) @property @pulumi.getter(name="natIpAddress") def nat_ip_address(self) -> Optional[pulumi.Input[str]]: """ A public address that can be used to access the internet over NAT. Use `variables` to set. """ return pulumi.get(self, "nat_ip_address") @nat_ip_address.setter def nat_ip_address(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "nat_ip_address", value) @property @pulumi.getter(name="natIpVersion") def nat_ip_version(self) -> Optional[pulumi.Input[str]]: """ The IP version for the public address. """ return pulumi.get(self, "nat_ip_version") @nat_ip_version.setter def nat_ip_version(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "nat_ip_version", value) @property @pulumi.getter(name="subnetId") def subnet_id(self) -> Optional[pulumi.Input[str]]: """ The ID of the subnet to attach this interface to. The subnet must reside in the same zone where this instance was created. """ return pulumi.get(self, "subnet_id") @subnet_id.setter def subnet_id(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "subnet_id", value) @pulumi.input_type class ComputeInstanceGroupInstanceTemplateArgs: def __init__(__self__, *, boot_disk: pulumi.Input['ComputeInstanceGroupInstanceTemplateBootDiskArgs'], network_interfaces: pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupInstanceTemplateNetworkInterfaceArgs']]], resources: pulumi.Input['ComputeInstanceGroupInstanceTemplateResourcesArgs'], description: Optional[pulumi.Input[str]] = None, hostname: Optional[pulumi.Input[str]] = None, labels: Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]] = None, metadata: Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]] = None, name: Optional[pulumi.Input[str]] = None, network_settings: Optional[pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupInstanceTemplateNetworkSettingArgs']]]] = None, placement_policy: Optional[pulumi.Input['ComputeInstanceGroupInstanceTemplatePlacementPolicyArgs']] = None, platform_id: Optional[pulumi.Input[str]] = None, scheduling_policy: Optional[pulumi.Input['ComputeInstanceGroupInstanceTemplateSchedulingPolicyArgs']] = None, secondary_disks: Optional[pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupInstanceTemplateSecondaryDiskArgs']]]] = None, service_account_id: Optional[pulumi.Input[str]] = None): """ :param pulumi.Input['ComputeInstanceGroupInstanceTemplateBootDiskArgs'] boot_disk: Boot disk specifications for the instance. The structure is documented below. :param pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupInstanceTemplateNetworkInterfaceArgs']]] network_interfaces: Network specifications for the instance. This can be used multiple times for adding multiple interfaces. The structure is documented below. :param pulumi.Input['ComputeInstanceGroupInstanceTemplateResourcesArgs'] resources: Compute resource specifications for the instance. The structure is documented below. :param pulumi.Input[str] description: A description of the boot disk. :param pulumi.Input[str] hostname: Hostname template for the instance. This field is used to generate the FQDN value of instance. The hostname must be unique within the network and region. If not specified, the hostname will be equal to id of the instance and FQDN will be `<id>.auto.internal`. Otherwise FQDN will be `<hostname>.<region_id>.internal`. In order to be unique it must contain at least on of instance unique placeholders: {instance.short_id} {instance.index} combination of {instance.zone_id} and {instance.index_in_zone} Example: my-instance-{instance.index} If not set, `name` value will be used It may also contain another placeholders, see metadata doc for full list. :param pulumi.Input[Mapping[str, pulumi.Input[str]]] labels: A map of labels of metric. :param pulumi.Input[Mapping[str, pulumi.Input[str]]] metadata: A set of metadata key/value pairs to make available from within the instance. :param pulumi.Input[str] name: Name template of the instance. In order to be unique it must contain at least one of instance unique placeholders: {instance.short_id} {instance.index} combination of {instance.zone_id} and {instance.index_in_zone} Example: my-instance-{instance.index} If not set, default is used: {instance_group.id}-{instance.short_id} It may also contain another placeholders, see metadata doc for full list. :param pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupInstanceTemplateNetworkSettingArgs']]] network_settings: Network acceleration type for instance. The structure is documented below. :param pulumi.Input['ComputeInstanceGroupInstanceTemplatePlacementPolicyArgs'] placement_policy: The placement policy configuration. The structure is documented below. :param pulumi.Input[str] platform_id: The ID of the hardware platform configuration for the instance. The default is 'standard-v1'. :param pulumi.Input['ComputeInstanceGroupInstanceTemplateSchedulingPolicyArgs'] scheduling_policy: The scheduling policy configuration. The structure is documented below. :param pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupInstanceTemplateSecondaryDiskArgs']]] secondary_disks: A list of disks to attach to the instance. The structure is documented below. :param pulumi.Input[str] service_account_id: The ID of the service account authorized for this instance. """ pulumi.set(__self__, "boot_disk", boot_disk) pulumi.set(__self__, "network_interfaces", network_interfaces) pulumi.set(__self__, "resources", resources) if description is not None: pulumi.set(__self__, "description", description) if hostname is not None: pulumi.set(__self__, "hostname", hostname) if labels is not None: pulumi.set(__self__, "labels", labels) if metadata is not None: pulumi.set(__self__, "metadata", metadata) if name is not None: pulumi.set(__self__, "name", name) if network_settings is not None: pulumi.set(__self__, "network_settings", network_settings) if placement_policy is not None: pulumi.set(__self__, "placement_policy", placement_policy) if platform_id is not None: pulumi.set(__self__, "platform_id", platform_id) if scheduling_policy is not None: pulumi.set(__self__, "scheduling_policy", scheduling_policy) if secondary_disks is not None: pulumi.set(__self__, "secondary_disks", secondary_disks) if service_account_id is not None: pulumi.set(__self__, "service_account_id", service_account_id) @property @pulumi.getter(name="bootDisk") def boot_disk(self) -> pulumi.Input['ComputeInstanceGroupInstanceTemplateBootDiskArgs']: """ Boot disk specifications for the instance. The structure is documented below. """ return pulumi.get(self, "boot_disk") @boot_disk.setter def boot_disk(self, value: pulumi.Input['ComputeInstanceGroupInstanceTemplateBootDiskArgs']): pulumi.set(self, "boot_disk", value) @property @pulumi.getter(name="networkInterfaces") def network_interfaces(self) -> pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupInstanceTemplateNetworkInterfaceArgs']]]: """ Network specifications for the instance. This can be used multiple times for adding multiple interfaces. The structure is documented below. """ return pulumi.get(self, "network_interfaces") @network_interfaces.setter def network_interfaces(self, value: pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupInstanceTemplateNetworkInterfaceArgs']]]): pulumi.set(self, "network_interfaces", value) @property @pulumi.getter def resources(self) -> pulumi.Input['ComputeInstanceGroupInstanceTemplateResourcesArgs']: """ Compute resource specifications for the instance. The structure is documented below. """ return pulumi.get(self, "resources") @resources.setter def resources(self, value: pulumi.Input['ComputeInstanceGroupInstanceTemplateResourcesArgs']): pulumi.set(self, "resources", value) @property @pulumi.getter def description(self) -> Optional[pulumi.Input[str]]: """ A description of the boot disk. """ return pulumi.get(self, "description") @description.setter def description(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "description", value) @property @pulumi.getter def hostname(self) -> Optional[pulumi.Input[str]]: """ Hostname template for the instance. This field is used to generate the FQDN value of instance. The hostname must be unique within the network and region. If not specified, the hostname will be equal to id of the instance and FQDN will be `<id>.auto.internal`. Otherwise FQDN will be `<hostname>.<region_id>.internal`. In order to be unique it must contain at least on of instance unique placeholders: {instance.short_id} {instance.index} combination of {instance.zone_id} and {instance.index_in_zone} Example: my-instance-{instance.index} If not set, `name` value will be used It may also contain another placeholders, see metadata doc for full list. """ return pulumi.get(self, "hostname") @hostname.setter def hostname(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "hostname", value) @property @pulumi.getter def labels(self) -> Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]]: """ A map of labels of metric. """ return pulumi.get(self, "labels") @labels.setter def labels(self, value: Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]]): pulumi.set(self, "labels", value) @property @pulumi.getter def metadata(self) -> Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]]: """ A set of metadata key/value pairs to make available from within the instance. """ return pulumi.get(self, "metadata") @metadata.setter def metadata(self, value: Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]]): pulumi.set(self, "metadata", value) @property @pulumi.getter def name(self) -> Optional[pulumi.Input[str]]: """ Name template of the instance. In order to be unique it must contain at least one of instance unique placeholders: {instance.short_id} {instance.index} combination of {instance.zone_id} and {instance.index_in_zone} Example: my-instance-{instance.index} If not set, default is used: {instance_group.id}-{instance.short_id} It may also contain another placeholders, see metadata doc for full list. """ return pulumi.get(self, "name") @name.setter def name(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "name", value) @property @pulumi.getter(name="networkSettings") def network_settings(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupInstanceTemplateNetworkSettingArgs']]]]: """ Network acceleration type for instance. The structure is documented below. """ return pulumi.get(self, "network_settings") @network_settings.setter def network_settings(self, value: Optional[pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupInstanceTemplateNetworkSettingArgs']]]]): pulumi.set(self, "network_settings", value) @property @pulumi.getter(name="placementPolicy") def placement_policy(self) -> Optional[pulumi.Input['ComputeInstanceGroupInstanceTemplatePlacementPolicyArgs']]: """ The placement policy configuration. The structure is documented below. """ return pulumi.get(self, "placement_policy") @placement_policy.setter def placement_policy(self, value: Optional[pulumi.Input['ComputeInstanceGroupInstanceTemplatePlacementPolicyArgs']]): pulumi.set(self, "placement_policy", value) @property @pulumi.getter(name="platformId") def platform_id(self) -> Optional[pulumi.Input[str]]: """ The ID of the hardware platform configuration for the instance. The default is 'standard-v1'. """ return pulumi.get(self, "platform_id") @platform_id.setter def platform_id(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "platform_id", value) @property @pulumi.getter(name="schedulingPolicy") def scheduling_policy(self) -> Optional[pulumi.Input['ComputeInstanceGroupInstanceTemplateSchedulingPolicyArgs']]: """ The scheduling policy configuration. The structure is documented below. """ return pulumi.get(self, "scheduling_policy") @scheduling_policy.setter def scheduling_policy(self, value: Optional[pulumi.Input['ComputeInstanceGroupInstanceTemplateSchedulingPolicyArgs']]): pulumi.set(self, "scheduling_policy", value) @property @pulumi.getter(name="secondaryDisks") def secondary_disks(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupInstanceTemplateSecondaryDiskArgs']]]]: """ A list of disks to attach to the instance. The structure is documented below. """ return pulumi.get(self, "secondary_disks") @secondary_disks.setter def secondary_disks(self, value: Optional[pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupInstanceTemplateSecondaryDiskArgs']]]]): pulumi.set(self, "secondary_disks", value) @property @pulumi.getter(name="serviceAccountId") def service_account_id(self) -> Optional[pulumi.Input[str]]: """ The ID of the service account authorized for this instance. """ return pulumi.get(self, "service_account_id") @service_account_id.setter def service_account_id(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "service_account_id", value) @pulumi.input_type class ComputeInstanceGroupInstanceTemplateBootDiskArgs: def __init__(__self__, *, device_name: Optional[pulumi.Input[str]] = None, disk_id: Optional[pulumi.Input[str]] = None, initialize_params: Optional[pulumi.Input['ComputeInstanceGroupInstanceTemplateBootDiskInitializeParamsArgs']] = None, mode: Optional[pulumi.Input[str]] = None): """ :param pulumi.Input[str] device_name: This value can be used to reference the device under `/dev/disk/by-id/`. :param pulumi.Input[str] disk_id: ID of the existing disk. To set use variables. :param pulumi.Input['ComputeInstanceGroupInstanceTemplateBootDiskInitializeParamsArgs'] initialize_params: Parameters for creating a disk alongside the instance. The structure is documented below. :param pulumi.Input[str] mode: The access mode to the disk resource. By default a disk is attached in `READ_WRITE` mode. """ if device_name is not None: pulumi.set(__self__, "device_name", device_name) if disk_id is not None: pulumi.set(__self__, "disk_id", disk_id) if initialize_params is not None: pulumi.set(__self__, "initialize_params", initialize_params) if mode is not None: pulumi.set(__self__, "mode", mode) @property @pulumi.getter(name="deviceName") def device_name(self) -> Optional[pulumi.Input[str]]: """ This value can be used to reference the device under `/dev/disk/by-id/`. """ return pulumi.get(self, "device_name") @device_name.setter def device_name(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "device_name", value) @property @pulumi.getter(name="diskId") def disk_id(self) -> Optional[pulumi.Input[str]]: """ ID of the existing disk. To set use variables. """ return pulumi.get(self, "disk_id") @disk_id.setter def disk_id(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "disk_id", value) @property @pulumi.getter(name="initializeParams") def initialize_params(self) -> Optional[pulumi.Input['ComputeInstanceGroupInstanceTemplateBootDiskInitializeParamsArgs']]: """ Parameters for creating a disk alongside the instance. The structure is documented below. """ return pulumi.get(self, "initialize_params") @initialize_params.setter def initialize_params(self, value: Optional[pulumi.Input['ComputeInstanceGroupInstanceTemplateBootDiskInitializeParamsArgs']]): pulumi.set(self, "initialize_params", value) @property @pulumi.getter def mode(self) -> Optional[pulumi.Input[str]]: """ The access mode to the disk resource. By default a disk is attached in `READ_WRITE` mode. """ return pulumi.get(self, "mode") @mode.setter def mode(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "mode", value) @pulumi.input_type class ComputeInstanceGroupInstanceTemplateBootDiskInitializeParamsArgs: def __init__(__self__, *, description: Optional[pulumi.Input[str]] = None, image_id: Optional[pulumi.Input[str]] = None, size: Optional[pulumi.Input[int]] = None, snapshot_id: Optional[pulumi.Input[str]] = None, type: Optional[pulumi.Input[str]] = None): """ :param pulumi.Input[str] description: A description of the boot disk. :param pulumi.Input[str] image_id: The disk image to initialize this disk from. :param pulumi.Input[int] size: The number of instances in the instance group. :param pulumi.Input[str] snapshot_id: The snapshot to initialize this disk from. :param pulumi.Input[str] type: Network acceleration type. By default a network is in `STANDARD` mode. """ if description is not None: pulumi.set(__self__, "description", description) if image_id is not None: pulumi.set(__self__, "image_id", image_id) if size is not None: pulumi.set(__self__, "size", size) if snapshot_id is not None: pulumi.set(__self__, "snapshot_id", snapshot_id) if type is not None: pulumi.set(__self__, "type", type) @property @pulumi.getter def description(self) -> Optional[pulumi.Input[str]]: """ A description of the boot disk. """ return pulumi.get(self, "description") @description.setter def description(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "description", value) @property @pulumi.getter(name="imageId") def image_id(self) -> Optional[pulumi.Input[str]]: """ The disk image to initialize this disk from. """ return pulumi.get(self, "image_id") @image_id.setter def image_id(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "image_id", value) @property @pulumi.getter def size(self) -> Optional[pulumi.Input[int]]: """ The number of instances in the instance group. """ return pulumi.get(self, "size") @size.setter def size(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "size", value) @property @pulumi.getter(name="snapshotId") def snapshot_id(self) -> Optional[pulumi.Input[str]]: """ The snapshot to initialize this disk from. """ return pulumi.get(self, "snapshot_id") @snapshot_id.setter def snapshot_id(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "snapshot_id", value) @property @pulumi.getter def type(self) -> Optional[pulumi.Input[str]]: """ Network acceleration type. By default a network is in `STANDARD` mode. """ return pulumi.get(self, "type") @type.setter def type(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "type", value) @pulumi.input_type class ComputeInstanceGroupInstanceTemplateNetworkInterfaceArgs: def __init__(__self__, *, dns_records: Optional[pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupInstanceTemplateNetworkInterfaceDnsRecordArgs']]]] = None, ip_address: Optional[pulumi.Input[str]] = None, ipv4: Optional[pulumi.Input[bool]] = None, ipv6: Optional[pulumi.Input[bool]] = None, ipv6_address: Optional[pulumi.Input[str]] = None, ipv6_dns_records: Optional[pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupInstanceTemplateNetworkInterfaceIpv6DnsRecordArgs']]]] = None, nat: Optional[pulumi.Input[bool]] = None, nat_dns_records: Optional[pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupInstanceTemplateNetworkInterfaceNatDnsRecordArgs']]]] = None, nat_ip_address: Optional[pulumi.Input[str]] = None, network_id: Optional[pulumi.Input[str]] = None, security_group_ids: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, subnet_ids: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None): """ :param pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupInstanceTemplateNetworkInterfaceDnsRecordArgs']]] dns_records: List of dns records. The structure is documented below. :param pulumi.Input[str] ip_address: Manual set static IP address. :param pulumi.Input[bool] ipv4: True if IPv4 address allocated for the network interface. :param pulumi.Input[str] ipv6_address: Manual set static IPv6 address. :param pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupInstanceTemplateNetworkInterfaceIpv6DnsRecordArgs']]] ipv6_dns_records: List of ipv6 dns records. The structure is documented below. :param pulumi.Input[bool] nat: Flag for using NAT. :param pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupInstanceTemplateNetworkInterfaceNatDnsRecordArgs']]] nat_dns_records: List of nat dns records. The structure is documented below. :param pulumi.Input[str] nat_ip_address: A public address that can be used to access the internet over NAT. Use `variables` to set. :param pulumi.Input[str] network_id: The ID of the network. :param pulumi.Input[Sequence[pulumi.Input[str]]] security_group_ids: Security group ids for network interface. :param pulumi.Input[Sequence[pulumi.Input[str]]] subnet_ids: The ID of the subnets to attach this interface to. """ if dns_records is not None: pulumi.set(__self__, "dns_records", dns_records) if ip_address is not None: pulumi.set(__self__, "ip_address", ip_address) if ipv4 is not None: pulumi.set(__self__, "ipv4", ipv4) if ipv6 is not None: pulumi.set(__self__, "ipv6", ipv6) if ipv6_address is not None: pulumi.set(__self__, "ipv6_address", ipv6_address) if ipv6_dns_records is not None: pulumi.set(__self__, "ipv6_dns_records", ipv6_dns_records) if nat is not None: pulumi.set(__self__, "nat", nat) if nat_dns_records is not None: pulumi.set(__self__, "nat_dns_records", nat_dns_records) if nat_ip_address is not None: pulumi.set(__self__, "nat_ip_address", nat_ip_address) if network_id is not None: pulumi.set(__self__, "network_id", network_id) if security_group_ids is not None: pulumi.set(__self__, "security_group_ids", security_group_ids) if subnet_ids is not None: pulumi.set(__self__, "subnet_ids", subnet_ids) @property @pulumi.getter(name="dnsRecords") def dns_records(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupInstanceTemplateNetworkInterfaceDnsRecordArgs']]]]: """ List of dns records. The structure is documented below. """ return pulumi.get(self, "dns_records") @dns_records.setter def dns_records(self, value: Optional[pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupInstanceTemplateNetworkInterfaceDnsRecordArgs']]]]): pulumi.set(self, "dns_records", value) @property @pulumi.getter(name="ipAddress") def ip_address(self) -> Optional[pulumi.Input[str]]: """ Manual set static IP address. """ return pulumi.get(self, "ip_address") @ip_address.setter def ip_address(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "ip_address", value) @property @pulumi.getter def ipv4(self) -> Optional[pulumi.Input[bool]]: """ True if IPv4 address allocated for the network interface. """ return pulumi.get(self, "ipv4") @ipv4.setter def ipv4(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "ipv4", value) @property @pulumi.getter def ipv6(self) -> Optional[pulumi.Input[bool]]: return pulumi.get(self, "ipv6") @ipv6.setter def ipv6(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "ipv6", value) @property @pulumi.getter(name="ipv6Address") def ipv6_address(self) -> Optional[pulumi.Input[str]]: """ Manual set static IPv6 address. """ return pulumi.get(self, "ipv6_address") @ipv6_address.setter def ipv6_address(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "ipv6_address", value) @property @pulumi.getter(name="ipv6DnsRecords") def ipv6_dns_records(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupInstanceTemplateNetworkInterfaceIpv6DnsRecordArgs']]]]: """ List of ipv6 dns records. The structure is documented below. """ return pulumi.get(self, "ipv6_dns_records") @ipv6_dns_records.setter def ipv6_dns_records(self, value: Optional[pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupInstanceTemplateNetworkInterfaceIpv6DnsRecordArgs']]]]): pulumi.set(self, "ipv6_dns_records", value) @property @pulumi.getter def nat(self) -> Optional[pulumi.Input[bool]]: """ Flag for using NAT. """ return pulumi.get(self, "nat") @nat.setter def nat(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "nat", value) @property @pulumi.getter(name="natDnsRecords") def nat_dns_records(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupInstanceTemplateNetworkInterfaceNatDnsRecordArgs']]]]: """ List of nat dns records. The structure is documented below. """ return pulumi.get(self, "nat_dns_records") @nat_dns_records.setter def nat_dns_records(self, value: Optional[pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupInstanceTemplateNetworkInterfaceNatDnsRecordArgs']]]]): pulumi.set(self, "nat_dns_records", value) @property @pulumi.getter(name="natIpAddress") def nat_ip_address(self) -> Optional[pulumi.Input[str]]: """ A public address that can be used to access the internet over NAT. Use `variables` to set. """ return pulumi.get(self, "nat_ip_address") @nat_ip_address.setter def nat_ip_address(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "nat_ip_address", value) @property @pulumi.getter(name="networkId") def network_id(self) -> Optional[pulumi.Input[str]]: """ The ID of the network. """ return pulumi.get(self, "network_id") @network_id.setter def network_id(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "network_id", value) @property @pulumi.getter(name="securityGroupIds") def security_group_ids(self) -> Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]: """ Security group ids for network interface. """ return pulumi.get(self, "security_group_ids") @security_group_ids.setter def security_group_ids(self, value: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]): pulumi.set(self, "security_group_ids", value) @property @pulumi.getter(name="subnetIds") def subnet_ids(self) -> Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]: """ The ID of the subnets to attach this interface to. """ return pulumi.get(self, "subnet_ids") @subnet_ids.setter def subnet_ids(self, value: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]): pulumi.set(self, "subnet_ids", value) @pulumi.input_type class ComputeInstanceGroupInstanceTemplateNetworkInterfaceDnsRecordArgs: def __init__(__self__, *, fqdn: pulumi.Input[str], dns_zone_id: Optional[pulumi.Input[str]] = None, ptr: Optional[pulumi.Input[bool]] = None, ttl: Optional[pulumi.Input[int]] = None): """ :param pulumi.Input[str] fqdn: DNS record fqdn (must have dot at the end). :param pulumi.Input[str] dns_zone_id: DNS zone id (if not set, private zone used). :param pulumi.Input[bool] ptr: When set to true, also create PTR DNS record. :param pulumi.Input[int] ttl: DNS record TTL. """ pulumi.set(__self__, "fqdn", fqdn) if dns_zone_id is not None: pulumi.set(__self__, "dns_zone_id", dns_zone_id) if ptr is not None: pulumi.set(__self__, "ptr", ptr) if ttl is not None: pulumi.set(__self__, "ttl", ttl) @property @pulumi.getter def fqdn(self) -> pulumi.Input[str]: """ DNS record fqdn (must have dot at the end). """ return pulumi.get(self, "fqdn") @fqdn.setter def fqdn(self, value: pulumi.Input[str]): pulumi.set(self, "fqdn", value) @property @pulumi.getter(name="dnsZoneId") def dns_zone_id(self) -> Optional[pulumi.Input[str]]: """ DNS zone id (if not set, private zone used). """ return pulumi.get(self, "dns_zone_id") @dns_zone_id.setter def dns_zone_id(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "dns_zone_id", value) @property @pulumi.getter def ptr(self) -> Optional[pulumi.Input[bool]]: """ When set to true, also create PTR DNS record. """ return pulumi.get(self, "ptr") @ptr.setter def ptr(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "ptr", value) @property @pulumi.getter def ttl(self) -> Optional[pulumi.Input[int]]: """ DNS record TTL. """ return pulumi.get(self, "ttl") @ttl.setter def ttl(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "ttl", value) @pulumi.input_type class ComputeInstanceGroupInstanceTemplateNetworkInterfaceIpv6DnsRecordArgs: def __init__(__self__, *, fqdn: pulumi.Input[str], dns_zone_id: Optional[pulumi.Input[str]] = None, ptr: Optional[pulumi.Input[bool]] = None, ttl: Optional[pulumi.Input[int]] = None): """ :param pulumi.Input[str] fqdn: DNS record fqdn (must have dot at the end). :param pulumi.Input[str] dns_zone_id: DNS zone id (if not set, private zone used). :param pulumi.Input[bool] ptr: When set to true, also create PTR DNS record. :param pulumi.Input[int] ttl: DNS record TTL. """ pulumi.set(__self__, "fqdn", fqdn) if dns_zone_id is not None: pulumi.set(__self__, "dns_zone_id", dns_zone_id) if ptr is not None: pulumi.set(__self__, "ptr", ptr) if ttl is not None: pulumi.set(__self__, "ttl", ttl) @property @pulumi.getter def fqdn(self) -> pulumi.Input[str]: """ DNS record fqdn (must have dot at the end). """ return pulumi.get(self, "fqdn") @fqdn.setter def fqdn(self, value: pulumi.Input[str]): pulumi.set(self, "fqdn", value) @property @pulumi.getter(name="dnsZoneId") def dns_zone_id(self) -> Optional[pulumi.Input[str]]: """ DNS zone id (if not set, private zone used). """ return pulumi.get(self, "dns_zone_id") @dns_zone_id.setter def dns_zone_id(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "dns_zone_id", value) @property @pulumi.getter def ptr(self) -> Optional[pulumi.Input[bool]]: """ When set to true, also create PTR DNS record. """ return pulumi.get(self, "ptr") @ptr.setter def ptr(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "ptr", value) @property @pulumi.getter def ttl(self) -> Optional[pulumi.Input[int]]: """ DNS record TTL. """ return pulumi.get(self, "ttl") @ttl.setter def ttl(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "ttl", value) @pulumi.input_type class ComputeInstanceGroupInstanceTemplateNetworkInterfaceNatDnsRecordArgs: def __init__(__self__, *, fqdn: pulumi.Input[str], dns_zone_id: Optional[pulumi.Input[str]] = None, ptr: Optional[pulumi.Input[bool]] = None, ttl: Optional[pulumi.Input[int]] = None): """ :param pulumi.Input[str] fqdn: DNS record fqdn (must have dot at the end). :param pulumi.Input[str] dns_zone_id: DNS zone id (if not set, private zone used). :param pulumi.Input[bool] ptr: When set to true, also create PTR DNS record. :param pulumi.Input[int] ttl: DNS record TTL. """ pulumi.set(__self__, "fqdn", fqdn) if dns_zone_id is not None: pulumi.set(__self__, "dns_zone_id", dns_zone_id) if ptr is not None: pulumi.set(__self__, "ptr", ptr) if ttl is not None: pulumi.set(__self__, "ttl", ttl) @property @pulumi.getter def fqdn(self) -> pulumi.Input[str]: """ DNS record fqdn (must have dot at the end). """ return pulumi.get(self, "fqdn") @fqdn.setter def fqdn(self, value: pulumi.Input[str]): pulumi.set(self, "fqdn", value) @property @pulumi.getter(name="dnsZoneId") def dns_zone_id(self) -> Optional[pulumi.Input[str]]: """ DNS zone id (if not set, private zone used). """ return pulumi.get(self, "dns_zone_id") @dns_zone_id.setter def dns_zone_id(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "dns_zone_id", value) @property @pulumi.getter def ptr(self) -> Optional[pulumi.Input[bool]]: """ When set to true, also create PTR DNS record. """ return pulumi.get(self, "ptr") @ptr.setter def ptr(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "ptr", value) @property @pulumi.getter def ttl(self) -> Optional[pulumi.Input[int]]: """ DNS record TTL. """ return pulumi.get(self, "ttl") @ttl.setter def ttl(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "ttl", value) @pulumi.input_type class ComputeInstanceGroupInstanceTemplateNetworkSettingArgs: def __init__(__self__, *, type: Optional[pulumi.Input[str]] = None): """ :param pulumi.Input[str] type: Network acceleration type. By default a network is in `STANDARD` mode. """ if type is not None: pulumi.set(__self__, "type", type) @property @pulumi.getter def type(self) -> Optional[pulumi.Input[str]]: """ Network acceleration type. By default a network is in `STANDARD` mode. """ return pulumi.get(self, "type") @type.setter def type(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "type", value) @pulumi.input_type class ComputeInstanceGroupInstanceTemplatePlacementPolicyArgs: def __init__(__self__, *, placement_group_id: pulumi.Input[str]): """ :param pulumi.Input[str] placement_group_id: Specifies the id of the Placement Group to assign to the instances. """ pulumi.set(__self__, "placement_group_id", placement_group_id) @property @pulumi.getter(name="placementGroupId") def placement_group_id(self) -> pulumi.Input[str]: """ Specifies the id of the Placement Group to assign to the instances. """ return pulumi.get(self, "placement_group_id") @placement_group_id.setter def placement_group_id(self, value: pulumi.Input[str]): pulumi.set(self, "placement_group_id", value) @pulumi.input_type class ComputeInstanceGroupInstanceTemplateResourcesArgs: def __init__(__self__, *, cores: pulumi.Input[int], memory: pulumi.Input[float], core_fraction: Optional[pulumi.Input[int]] = None, gpus: Optional[pulumi.Input[int]] = None): """ :param pulumi.Input[int] cores: The number of CPU cores for the instance. :param pulumi.Input[float] memory: The memory size in GB. :param pulumi.Input[int] core_fraction: If provided, specifies baseline core performance as a percent. """ pulumi.set(__self__, "cores", cores) pulumi.set(__self__, "memory", memory) if core_fraction is not None: pulumi.set(__self__, "core_fraction", core_fraction) if gpus is not None: pulumi.set(__self__, "gpus", gpus) @property @pulumi.getter def cores(self) -> pulumi.Input[int]: """ The number of CPU cores for the instance. """ return pulumi.get(self, "cores") @cores.setter def cores(self, value: pulumi.Input[int]): pulumi.set(self, "cores", value) @property @pulumi.getter def memory(self) -> pulumi.Input[float]: """ The memory size in GB. """ return pulumi.get(self, "memory") @memory.setter def memory(self, value: pulumi.Input[float]): pulumi.set(self, "memory", value) @property @pulumi.getter(name="coreFraction") def core_fraction(self) -> Optional[pulumi.Input[int]]: """ If provided, specifies baseline core performance as a percent. """ return pulumi.get(self, "core_fraction") @core_fraction.setter def core_fraction(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "core_fraction", value) @property @pulumi.getter def gpus(self) -> Optional[pulumi.Input[int]]: return pulumi.get(self, "gpus") @gpus.setter def gpus(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "gpus", value) @pulumi.input_type class ComputeInstanceGroupInstanceTemplateSchedulingPolicyArgs: def __init__(__self__, *, preemptible: Optional[pulumi.Input[bool]] = None): """ :param pulumi.Input[bool] preemptible: Specifies if the instance is preemptible. Defaults to false. """ if preemptible is not None: pulumi.set(__self__, "preemptible", preemptible) @property @pulumi.getter def preemptible(self) -> Optional[pulumi.Input[bool]]: """ Specifies if the instance is preemptible. Defaults to false. """ return pulumi.get(self, "preemptible") @preemptible.setter def preemptible(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "preemptible", value) @pulumi.input_type class ComputeInstanceGroupInstanceTemplateSecondaryDiskArgs: def __init__(__self__, *, device_name: Optional[pulumi.Input[str]] = None, disk_id: Optional[pulumi.Input[str]] = None, initialize_params: Optional[pulumi.Input['ComputeInstanceGroupInstanceTemplateSecondaryDiskInitializeParamsArgs']] = None, mode: Optional[pulumi.Input[str]] = None): """ :param pulumi.Input[str] device_name: This value can be used to reference the device under `/dev/disk/by-id/`. :param pulumi.Input[str] disk_id: ID of the existing disk. To set use variables. :param pulumi.Input['ComputeInstanceGroupInstanceTemplateSecondaryDiskInitializeParamsArgs'] initialize_params: Parameters for creating a disk alongside the instance. The structure is documented below. :param pulumi.Input[str] mode: The access mode to the disk resource. By default a disk is attached in `READ_WRITE` mode. """ if device_name is not None: pulumi.set(__self__, "device_name", device_name) if disk_id is not None: pulumi.set(__self__, "disk_id", disk_id) if initialize_params is not None: pulumi.set(__self__, "initialize_params", initialize_params) if mode is not None: pulumi.set(__self__, "mode", mode) @property @pulumi.getter(name="deviceName") def device_name(self) -> Optional[pulumi.Input[str]]: """ This value can be used to reference the device under `/dev/disk/by-id/`. """ return pulumi.get(self, "device_name") @device_name.setter def device_name(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "device_name", value) @property @pulumi.getter(name="diskId") def disk_id(self) -> Optional[pulumi.Input[str]]: """ ID of the existing disk. To set use variables. """ return pulumi.get(self, "disk_id") @disk_id.setter def disk_id(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "disk_id", value) @property @pulumi.getter(name="initializeParams") def initialize_params(self) -> Optional[pulumi.Input['ComputeInstanceGroupInstanceTemplateSecondaryDiskInitializeParamsArgs']]: """ Parameters for creating a disk alongside the instance. The structure is documented below. """ return pulumi.get(self, "initialize_params") @initialize_params.setter def initialize_params(self, value: Optional[pulumi.Input['ComputeInstanceGroupInstanceTemplateSecondaryDiskInitializeParamsArgs']]): pulumi.set(self, "initialize_params", value) @property @pulumi.getter def mode(self) -> Optional[pulumi.Input[str]]: """ The access mode to the disk resource. By default a disk is attached in `READ_WRITE` mode. """ return pulumi.get(self, "mode") @mode.setter def mode(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "mode", value) @pulumi.input_type class ComputeInstanceGroupInstanceTemplateSecondaryDiskInitializeParamsArgs: def __init__(__self__, *, description: Optional[pulumi.Input[str]] = None, image_id: Optional[pulumi.Input[str]] = None, size: Optional[pulumi.Input[int]] = None, snapshot_id: Optional[pulumi.Input[str]] = None, type: Optional[pulumi.Input[str]] = None): """ :param pulumi.Input[str] description: A description of the boot disk. :param pulumi.Input[str] image_id: The disk image to initialize this disk from. :param pulumi.Input[int] size: The number of instances in the instance group. :param pulumi.Input[str] snapshot_id: The snapshot to initialize this disk from. :param pulumi.Input[str] type: Network acceleration type. By default a network is in `STANDARD` mode. """ if description is not None: pulumi.set(__self__, "description", description) if image_id is not None: pulumi.set(__self__, "image_id", image_id) if size is not None: pulumi.set(__self__, "size", size) if snapshot_id is not None: pulumi.set(__self__, "snapshot_id", snapshot_id) if type is not None: pulumi.set(__self__, "type", type) @property @pulumi.getter def description(self) -> Optional[pulumi.Input[str]]: """ A description of the boot disk. """ return pulumi.get(self, "description") @description.setter def description(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "description", value) @property @pulumi.getter(name="imageId") def image_id(self) -> Optional[pulumi.Input[str]]: """ The disk image to initialize this disk from. """ return pulumi.get(self, "image_id") @image_id.setter def image_id(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "image_id", value) @property @pulumi.getter def size(self) -> Optional[pulumi.Input[int]]: """ The number of instances in the instance group. """ return pulumi.get(self, "size") @size.setter def size(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "size", value) @property @pulumi.getter(name="snapshotId") def snapshot_id(self) -> Optional[pulumi.Input[str]]: """ The snapshot to initialize this disk from. """ return pulumi.get(self, "snapshot_id") @snapshot_id.setter def snapshot_id(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "snapshot_id", value) @property @pulumi.getter def type(self) -> Optional[pulumi.Input[str]]: """ Network acceleration type. By default a network is in `STANDARD` mode. """ return pulumi.get(self, "type") @type.setter def type(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "type", value) @pulumi.input_type class ComputeInstanceGroupLoadBalancerArgs: def __init__(__self__, *, max_opening_traffic_duration: Optional[pulumi.Input[int]] = None, status_message: Optional[pulumi.Input[str]] = None, target_group_description: Optional[pulumi.Input[str]] = None, target_group_id: Optional[pulumi.Input[str]] = None, target_group_labels: Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]] = None, target_group_name: Optional[pulumi.Input[str]] = None): """ :param pulumi.Input[int] max_opening_traffic_duration: Timeout for waiting for the VM to be checked by the load balancer. If the timeout is exceeded, the VM will be turned off based on the deployment policy. Specified in seconds. :param pulumi.Input[str] status_message: The status message of the instance. :param pulumi.Input[str] target_group_description: A description of the target group. :param pulumi.Input[Mapping[str, pulumi.Input[str]]] target_group_labels: A set of key/value label pairs. :param pulumi.Input[str] target_group_name: The name of the target group. """ if max_opening_traffic_duration is not None: pulumi.set(__self__, "max_opening_traffic_duration", max_opening_traffic_duration) if status_message is not None: pulumi.set(__self__, "status_message", status_message) if target_group_description is not None: pulumi.set(__self__, "target_group_description", target_group_description) if target_group_id is not None: pulumi.set(__self__, "target_group_id", target_group_id) if target_group_labels is not None: pulumi.set(__self__, "target_group_labels", target_group_labels) if target_group_name is not None: pulumi.set(__self__, "target_group_name", target_group_name) @property @pulumi.getter(name="maxOpeningTrafficDuration") def max_opening_traffic_duration(self) -> Optional[pulumi.Input[int]]: """ Timeout for waiting for the VM to be checked by the load balancer. If the timeout is exceeded, the VM will be turned off based on the deployment policy. Specified in seconds. """ return pulumi.get(self, "max_opening_traffic_duration") @max_opening_traffic_duration.setter def max_opening_traffic_duration(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "max_opening_traffic_duration", value) @property @pulumi.getter(name="statusMessage") def status_message(self) -> Optional[pulumi.Input[str]]: """ The status message of the instance. """ return pulumi.get(self, "status_message") @status_message.setter def status_message(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "status_message", value) @property @pulumi.getter(name="targetGroupDescription") def target_group_description(self) -> Optional[pulumi.Input[str]]: """ A description of the target group. """ return pulumi.get(self, "target_group_description") @target_group_description.setter def target_group_description(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "target_group_description", value) @property @pulumi.getter(name="targetGroupId") def target_group_id(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "target_group_id") @target_group_id.setter def target_group_id(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "target_group_id", value) @property @pulumi.getter(name="targetGroupLabels") def target_group_labels(self) -> Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]]: """ A set of key/value label pairs. """ return pulumi.get(self, "target_group_labels") @target_group_labels.setter def target_group_labels(self, value: Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]]): pulumi.set(self, "target_group_labels", value) @property @pulumi.getter(name="targetGroupName") def target_group_name(self) -> Optional[pulumi.Input[str]]: """ The name of the target group. """ return pulumi.get(self, "target_group_name") @target_group_name.setter def target_group_name(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "target_group_name", value) @pulumi.input_type class ComputeInstanceGroupScalePolicyArgs: def __init__(__self__, *, auto_scale: Optional[pulumi.Input['ComputeInstanceGroupScalePolicyAutoScaleArgs']] = None, fixed_scale: Optional[pulumi.Input['ComputeInstanceGroupScalePolicyFixedScaleArgs']] = None, test_auto_scale: Optional[pulumi.Input['ComputeInstanceGroupScalePolicyTestAutoScaleArgs']] = None): """ :param pulumi.Input['ComputeInstanceGroupScalePolicyAutoScaleArgs'] auto_scale: The auto scaling policy of the instance group. The structure is documented below. :param pulumi.Input['ComputeInstanceGroupScalePolicyFixedScaleArgs'] fixed_scale: The fixed scaling policy of the instance group. The structure is documented below. :param pulumi.Input['ComputeInstanceGroupScalePolicyTestAutoScaleArgs'] test_auto_scale: The test auto scaling policy of the instance group. Use it to test how the auto scale works. The structure is documented below. """ if auto_scale is not None: pulumi.set(__self__, "auto_scale", auto_scale) if fixed_scale is not None: pulumi.set(__self__, "fixed_scale", fixed_scale) if test_auto_scale is not None: pulumi.set(__self__, "test_auto_scale", test_auto_scale) @property @pulumi.getter(name="autoScale") def auto_scale(self) -> Optional[pulumi.Input['ComputeInstanceGroupScalePolicyAutoScaleArgs']]: """ The auto scaling policy of the instance group. The structure is documented below. """ return pulumi.get(self, "auto_scale") @auto_scale.setter def auto_scale(self, value: Optional[pulumi.Input['ComputeInstanceGroupScalePolicyAutoScaleArgs']]): pulumi.set(self, "auto_scale", value) @property @pulumi.getter(name="fixedScale") def fixed_scale(self) -> Optional[pulumi.Input['ComputeInstanceGroupScalePolicyFixedScaleArgs']]: """ The fixed scaling policy of the instance group. The structure is documented below. """ return pulumi.get(self, "fixed_scale") @fixed_scale.setter def fixed_scale(self, value: Optional[pulumi.Input['ComputeInstanceGroupScalePolicyFixedScaleArgs']]): pulumi.set(self, "fixed_scale", value) @property @pulumi.getter(name="testAutoScale") def test_auto_scale(self) -> Optional[pulumi.Input['ComputeInstanceGroupScalePolicyTestAutoScaleArgs']]: """ The test auto scaling policy of the instance group. Use it to test how the auto scale works. The structure is documented below. """ return pulumi.get(self, "test_auto_scale") @test_auto_scale.setter def test_auto_scale(self, value: Optional[pulumi.Input['ComputeInstanceGroupScalePolicyTestAutoScaleArgs']]): pulumi.set(self, "test_auto_scale", value) @pulumi.input_type class ComputeInstanceGroupScalePolicyAutoScaleArgs: def __init__(__self__, *, initial_size: pulumi.Input[int], measurement_duration: pulumi.Input[int], cpu_utilization_target: Optional[pulumi.Input[float]] = None, custom_rules: Optional[pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupScalePolicyAutoScaleCustomRuleArgs']]]] = None, max_size: Optional[pulumi.Input[int]] = None, min_zone_size: Optional[pulumi.Input[int]] = None, stabilization_duration: Optional[pulumi.Input[int]] = None, warmup_duration: Optional[pulumi.Input[int]] = None): """ :param pulumi.Input[int] initial_size: The initial number of instances in the instance group. :param pulumi.Input[int] measurement_duration: The amount of time, in seconds, that metrics are averaged for. If the average value at the end of the interval is higher than the `cpu_utilization_target`, the instance group will increase the number of virtual machines in the group. :param pulumi.Input[float] cpu_utilization_target: Target CPU load level. :param pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupScalePolicyAutoScaleCustomRuleArgs']]] custom_rules: A list of custom rules. The structure is documented below. :param pulumi.Input[int] max_size: The maximum number of virtual machines in the group. :param pulumi.Input[int] min_zone_size: The minimum number of virtual machines in a single availability zone. :param pulumi.Input[int] stabilization_duration: The minimum time interval, in seconds, to monitor the load before an instance group can reduce the number of virtual machines in the group. During this time, the group will not decrease even if the average load falls below the value of `cpu_utilization_target`. :param pulumi.Input[int] warmup_duration: The warm-up time of the virtual machine, in seconds. During this time, traffic is fed to the virtual machine, but load metrics are not taken into account. """ pulumi.set(__self__, "initial_size", initial_size) pulumi.set(__self__, "measurement_duration", measurement_duration) if cpu_utilization_target is not None: pulumi.set(__self__, "cpu_utilization_target", cpu_utilization_target) if custom_rules is not None: pulumi.set(__self__, "custom_rules", custom_rules) if max_size is not None: pulumi.set(__self__, "max_size", max_size) if min_zone_size is not None: pulumi.set(__self__, "min_zone_size", min_zone_size) if stabilization_duration is not None: pulumi.set(__self__, "stabilization_duration", stabilization_duration) if warmup_duration is not None: pulumi.set(__self__, "warmup_duration", warmup_duration) @property @pulumi.getter(name="initialSize") def initial_size(self) -> pulumi.Input[int]: """ The initial number of instances in the instance group. """ return pulumi.get(self, "initial_size") @initial_size.setter def initial_size(self, value: pulumi.Input[int]): pulumi.set(self, "initial_size", value) @property @pulumi.getter(name="measurementDuration") def measurement_duration(self) -> pulumi.Input[int]: """ The amount of time, in seconds, that metrics are averaged for. If the average value at the end of the interval is higher than the `cpu_utilization_target`, the instance group will increase the number of virtual machines in the group. """ return pulumi.get(self, "measurement_duration") @measurement_duration.setter def measurement_duration(self, value: pulumi.Input[int]): pulumi.set(self, "measurement_duration", value) @property @pulumi.getter(name="cpuUtilizationTarget") def cpu_utilization_target(self) -> Optional[pulumi.Input[float]]: """ Target CPU load level. """ return pulumi.get(self, "cpu_utilization_target") @cpu_utilization_target.setter def cpu_utilization_target(self, value: Optional[pulumi.Input[float]]): pulumi.set(self, "cpu_utilization_target", value) @property @pulumi.getter(name="customRules") def custom_rules(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupScalePolicyAutoScaleCustomRuleArgs']]]]: """ A list of custom rules. The structure is documented below. """ return pulumi.get(self, "custom_rules") @custom_rules.setter def custom_rules(self, value: Optional[pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupScalePolicyAutoScaleCustomRuleArgs']]]]): pulumi.set(self, "custom_rules", value) @property @pulumi.getter(name="maxSize") def max_size(self) -> Optional[pulumi.Input[int]]: """ The maximum number of virtual machines in the group. """ return pulumi.get(self, "max_size") @max_size.setter def max_size(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "max_size", value) @property @pulumi.getter(name="minZoneSize") def min_zone_size(self) -> Optional[pulumi.Input[int]]: """ The minimum number of virtual machines in a single availability zone. """ return pulumi.get(self, "min_zone_size") @min_zone_size.setter def min_zone_size(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "min_zone_size", value) @property @pulumi.getter(name="stabilizationDuration") def stabilization_duration(self) -> Optional[pulumi.Input[int]]: """ The minimum time interval, in seconds, to monitor the load before an instance group can reduce the number of virtual machines in the group. During this time, the group will not decrease even if the average load falls below the value of `cpu_utilization_target`. """ return pulumi.get(self, "stabilization_duration") @stabilization_duration.setter def stabilization_duration(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "stabilization_duration", value) @property @pulumi.getter(name="warmupDuration") def warmup_duration(self) -> Optional[pulumi.Input[int]]: """ The warm-up time of the virtual machine, in seconds. During this time, traffic is fed to the virtual machine, but load metrics are not taken into account. """ return pulumi.get(self, "warmup_duration") @warmup_duration.setter def warmup_duration(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "warmup_duration", value) @pulumi.input_type class ComputeInstanceGroupScalePolicyAutoScaleCustomRuleArgs: def __init__(__self__, *, metric_name: pulumi.Input[str], metric_type: pulumi.Input[str], rule_type: pulumi.Input[str], target: pulumi.Input[float], folder_id: Optional[pulumi.Input[str]] = None, labels: Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]] = None, service: Optional[pulumi.Input[str]] = None): """ :param pulumi.Input[str] metric_name: The name of metric. :param pulumi.Input[str] metric_type: Metric type, `GAUGE` or `COUNTER`. :param pulumi.Input[str] rule_type: Rule type: `UTILIZATION` - This type means that the metric applies to one instance. First, Instance Groups calculates the average metric value for each instance, then averages the values for instances in one availability zone. This type of metric must have the `instance_id` label. `WORKLOAD` - This type means that the metric applies to instances in one availability zone. This type of metric must have the `zone_id` label. :param pulumi.Input[float] target: Target metric value level. :param pulumi.Input[str] folder_id: Folder ID of custom metric in Yandex Monitoring that should be used for scaling. :param pulumi.Input[Mapping[str, pulumi.Input[str]]] labels: A map of labels of metric. :param pulumi.Input[str] service: Service of custom metric in Yandex Monitoring that should be used for scaling. """ pulumi.set(__self__, "metric_name", metric_name) pulumi.set(__self__, "metric_type", metric_type) pulumi.set(__self__, "rule_type", rule_type) pulumi.set(__self__, "target", target) if folder_id is not None: pulumi.set(__self__, "folder_id", folder_id) if labels is not None: pulumi.set(__self__, "labels", labels) if service is not None: pulumi.set(__self__, "service", service) @property @pulumi.getter(name="metricName") def metric_name(self) -> pulumi.Input[str]: """ The name of metric. """ return pulumi.get(self, "metric_name") @metric_name.setter def metric_name(self, value: pulumi.Input[str]): pulumi.set(self, "metric_name", value) @property @pulumi.getter(name="metricType") def metric_type(self) -> pulumi.Input[str]: """ Metric type, `GAUGE` or `COUNTER`. """ return pulumi.get(self, "metric_type") @metric_type.setter def metric_type(self, value: pulumi.Input[str]): pulumi.set(self, "metric_type", value) @property @pulumi.getter(name="ruleType") def rule_type(self) -> pulumi.Input[str]: """ Rule type: `UTILIZATION` - This type means that the metric applies to one instance. First, Instance Groups calculates the average metric value for each instance, then averages the values for instances in one availability zone. This type of metric must have the `instance_id` label. `WORKLOAD` - This type means that the metric applies to instances in one availability zone. This type of metric must have the `zone_id` label. """ return pulumi.get(self, "rule_type") @rule_type.setter def rule_type(self, value: pulumi.Input[str]): pulumi.set(self, "rule_type", value) @property @pulumi.getter def target(self) -> pulumi.Input[float]: """ Target metric value level. """ return pulumi.get(self, "target") @target.setter def target(self, value: pulumi.Input[float]): pulumi.set(self, "target", value) @property @pulumi.getter(name="folderId") def folder_id(self) -> Optional[pulumi.Input[str]]: """ Folder ID of custom metric in Yandex Monitoring that should be used for scaling. """ return pulumi.get(self, "folder_id") @folder_id.setter def folder_id(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "folder_id", value) @property @pulumi.getter def labels(self) -> Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]]: """ A map of labels of metric. """ return pulumi.get(self, "labels") @labels.setter def labels(self, value: Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]]): pulumi.set(self, "labels", value) @property @pulumi.getter def service(self) -> Optional[pulumi.Input[str]]: """ Service of custom metric in Yandex Monitoring that should be used for scaling. """ return pulumi.get(self, "service") @service.setter def service(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "service", value) @pulumi.input_type class ComputeInstanceGroupScalePolicyFixedScaleArgs: def __init__(__self__, *, size: pulumi.Input[int]): """ :param pulumi.Input[int] size: The number of instances in the instance group. """ pulumi.set(__self__, "size", size) @property @pulumi.getter def size(self) -> pulumi.Input[int]: """ The number of instances in the instance group. """ return pulumi.get(self, "size") @size.setter def size(self, value: pulumi.Input[int]): pulumi.set(self, "size", value) @pulumi.input_type class ComputeInstanceGroupScalePolicyTestAutoScaleArgs: def __init__(__self__, *, initial_size: pulumi.Input[int], measurement_duration: pulumi.Input[int], cpu_utilization_target: Optional[pulumi.Input[float]] = None, custom_rules: Optional[pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupScalePolicyTestAutoScaleCustomRuleArgs']]]] = None, max_size: Optional[pulumi.Input[int]] = None, min_zone_size: Optional[pulumi.Input[int]] = None, stabilization_duration: Optional[pulumi.Input[int]] = None, warmup_duration: Optional[pulumi.Input[int]] = None): """ :param pulumi.Input[int] initial_size: The initial number of instances in the instance group. :param pulumi.Input[int] measurement_duration: The amount of time, in seconds, that metrics are averaged for. If the average value at the end of the interval is higher than the `cpu_utilization_target`, the instance group will increase the number of virtual machines in the group. :param pulumi.Input[float] cpu_utilization_target: Target CPU load level. :param pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupScalePolicyTestAutoScaleCustomRuleArgs']]] custom_rules: A list of custom rules. The structure is documented below. :param pulumi.Input[int] max_size: The maximum number of virtual machines in the group. :param pulumi.Input[int] min_zone_size: The minimum number of virtual machines in a single availability zone. :param pulumi.Input[int] stabilization_duration: The minimum time interval, in seconds, to monitor the load before an instance group can reduce the number of virtual machines in the group. During this time, the group will not decrease even if the average load falls below the value of `cpu_utilization_target`. :param pulumi.Input[int] warmup_duration: The warm-up time of the virtual machine, in seconds. During this time, traffic is fed to the virtual machine, but load metrics are not taken into account. """ pulumi.set(__self__, "initial_size", initial_size) pulumi.set(__self__, "measurement_duration", measurement_duration) if cpu_utilization_target is not None: pulumi.set(__self__, "cpu_utilization_target", cpu_utilization_target) if custom_rules is not None: pulumi.set(__self__, "custom_rules", custom_rules) if max_size is not None: pulumi.set(__self__, "max_size", max_size) if min_zone_size is not None: pulumi.set(__self__, "min_zone_size", min_zone_size) if stabilization_duration is not None: pulumi.set(__self__, "stabilization_duration", stabilization_duration) if warmup_duration is not None: pulumi.set(__self__, "warmup_duration", warmup_duration) @property @pulumi.getter(name="initialSize") def initial_size(self) -> pulumi.Input[int]: """ The initial number of instances in the instance group. """ return pulumi.get(self, "initial_size") @initial_size.setter def initial_size(self, value: pulumi.Input[int]): pulumi.set(self, "initial_size", value) @property @pulumi.getter(name="measurementDuration") def measurement_duration(self) -> pulumi.Input[int]: """ The amount of time, in seconds, that metrics are averaged for. If the average value at the end of the interval is higher than the `cpu_utilization_target`, the instance group will increase the number of virtual machines in the group. """ return pulumi.get(self, "measurement_duration") @measurement_duration.setter def measurement_duration(self, value: pulumi.Input[int]): pulumi.set(self, "measurement_duration", value) @property @pulumi.getter(name="cpuUtilizationTarget") def cpu_utilization_target(self) -> Optional[pulumi.Input[float]]: """ Target CPU load level. """ return pulumi.get(self, "cpu_utilization_target") @cpu_utilization_target.setter def cpu_utilization_target(self, value: Optional[pulumi.Input[float]]): pulumi.set(self, "cpu_utilization_target", value) @property @pulumi.getter(name="customRules") def custom_rules(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupScalePolicyTestAutoScaleCustomRuleArgs']]]]: """ A list of custom rules. The structure is documented below. """ return pulumi.get(self, "custom_rules") @custom_rules.setter def custom_rules(self, value: Optional[pulumi.Input[Sequence[pulumi.Input['ComputeInstanceGroupScalePolicyTestAutoScaleCustomRuleArgs']]]]): pulumi.set(self, "custom_rules", value) @property @pulumi.getter(name="maxSize") def max_size(self) -> Optional[pulumi.Input[int]]: """ The maximum number of virtual machines in the group. """ return pulumi.get(self, "max_size") @max_size.setter def max_size(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "max_size", value) @property @pulumi.getter(name="minZoneSize") def min_zone_size(self) -> Optional[pulumi.Input[int]]: """ The minimum number of virtual machines in a single availability zone. """ return pulumi.get(self, "min_zone_size") @min_zone_size.setter def min_zone_size(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "min_zone_size", value) @property @pulumi.getter(name="stabilizationDuration") def stabilization_duration(self) -> Optional[pulumi.Input[int]]: """ The minimum time interval, in seconds, to monitor the load before an instance group can reduce the number of virtual machines in the group. During this time, the group will not decrease even if the average load falls below the value of `cpu_utilization_target`. """ return pulumi.get(self, "stabilization_duration") @stabilization_duration.setter def stabilization_duration(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "stabilization_duration", value) @property @pulumi.getter(name="warmupDuration") def warmup_duration(self) -> Optional[pulumi.Input[int]]: """ The warm-up time of the virtual machine, in seconds. During this time, traffic is fed to the virtual machine, but load metrics are not taken into account. """ return pulumi.get(self, "warmup_duration") @warmup_duration.setter def warmup_duration(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "warmup_duration", value) @pulumi.input_type class ComputeInstanceGroupScalePolicyTestAutoScaleCustomRuleArgs: def __init__(__self__, *, metric_name: pulumi.Input[str], metric_type: pulumi.Input[str], rule_type: pulumi.Input[str], target: pulumi.Input[float], folder_id: Optional[pulumi.Input[str]] = None, labels: Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]] = None, service: Optional[pulumi.Input[str]] = None): """ :param pulumi.Input[str] metric_name: The name of metric. :param pulumi.Input[str] metric_type: Metric type, `GAUGE` or `COUNTER`. :param pulumi.Input[str] rule_type: Rule type: `UTILIZATION` - This type means that the metric applies to one instance. First, Instance Groups calculates the average metric value for each instance, then averages the values for instances in one availability zone. This type of metric must have the `instance_id` label. `WORKLOAD` - This type means that the metric applies to instances in one availability zone. This type of metric must have the `zone_id` label. :param pulumi.Input[float] target: Target metric value level. :param pulumi.Input[str] folder_id: Folder ID of custom metric in Yandex Monitoring that should be used for scaling. :param pulumi.Input[Mapping[str, pulumi.Input[str]]] labels: A map of labels of metric. :param pulumi.Input[str] service: Service of custom metric in Yandex Monitoring that should be used for scaling. """ pulumi.set(__self__, "metric_name", metric_name) pulumi.set(__self__, "metric_type", metric_type) pulumi.set(__self__, "rule_type", rule_type) pulumi.set(__self__, "target", target) if folder_id is not None: pulumi.set(__self__, "folder_id", folder_id) if labels is not None: pulumi.set(__self__, "labels", labels) if service is not None: pulumi.set(__self__, "service", service) @property @pulumi.getter(name="metricName") def metric_name(self) -> pulumi.Input[str]: """ The name of metric. """ return pulumi.get(self, "metric_name") @metric_name.setter def metric_name(self, value: pulumi.Input[str]): pulumi.set(self, "metric_name", value) @property @pulumi.getter(name="metricType") def metric_type(self) -> pulumi.Input[str]: """ Metric type, `GAUGE` or `COUNTER`. """ return pulumi.get(self, "metric_type") @metric_type.setter def metric_type(self, value: pulumi.Input[str]): pulumi.set(self, "metric_type", value) @property @pulumi.getter(name="ruleType") def rule_type(self) -> pulumi.Input[str]: """ Rule type: `UTILIZATION` - This type means that the metric applies to one instance. First, Instance Groups calculates the average metric value for each instance, then averages the values for instances in one availability zone. This type of metric must have the `instance_id` label. `WORKLOAD` - This type means that the metric applies to instances in one availability zone. This type of metric must have the `zone_id` label. """ return pulumi.get(self, "rule_type") @rule_type.setter def rule_type(self, value: pulumi.Input[str]): pulumi.set(self, "rule_type", value) @property @pulumi.getter def target(self) -> pulumi.Input[float]: """ Target metric value level. """ return pulumi.get(self, "target") @target.setter def target(self, value: pulumi.Input[float]): pulumi.set(self, "target", value) @property @pulumi.getter(name="folderId") def folder_id(self) -> Optional[pulumi.Input[str]]: """ Folder ID of custom metric in Yandex Monitoring that should be used for scaling. """ return pulumi.get(self, "folder_id") @folder_id.setter def folder_id(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "folder_id", value) @property @pulumi.getter def labels(self) -> Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]]: """ A map of labels of metric. """ return pulumi.get(self, "labels") @labels.setter def labels(self, value: Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]]): pulumi.set(self, "labels", value) @property @pulumi.getter def service(self) -> Optional[pulumi.Input[str]]: """ Service of custom metric in Yandex Monitoring that should be used for scaling. """ return pulumi.get(self, "service") @service.setter def service(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "service", value) @pulumi.input_type class ComputeInstanceNetworkInterfaceArgs: def __init__(__self__, *, subnet_id: pulumi.Input[str], dns_records: Optional[pulumi.Input[Sequence[pulumi.Input['ComputeInstanceNetworkInterfaceDnsRecordArgs']]]] = None, index: Optional[pulumi.Input[int]] = None, ip_address: Optional[pulumi.Input[str]] = None, ipv4: Optional[pulumi.Input[bool]] = None, ipv6: Optional[pulumi.Input[bool]] = None, ipv6_address: Optional[pulumi.Input[str]] = None, ipv6_dns_records: Optional[pulumi.Input[Sequence[pulumi.Input['ComputeInstanceNetworkInterfaceIpv6DnsRecordArgs']]]] = None, mac_address: Optional[pulumi.Input[str]] = None, nat: Optional[pulumi.Input[bool]] = None, nat_dns_records: Optional[pulumi.Input[Sequence[pulumi.Input['ComputeInstanceNetworkInterfaceNatDnsRecordArgs']]]] = None, nat_ip_address: Optional[pulumi.Input[str]] = None, nat_ip_version: Optional[pulumi.Input[str]] = None, security_group_ids: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None): """ :param pulumi.Input[str] subnet_id: ID of the subnet to attach this interface to. The subnet must exist in the same zone where this instance will be created. :param pulumi.Input[Sequence[pulumi.Input['ComputeInstanceNetworkInterfaceDnsRecordArgs']]] dns_records: List of configurations for creating ipv4 DNS records. The structure is documented below. :param pulumi.Input[str] ip_address: The private IP address to assign to the instance. If empty, the address will be automatically assigned from the specified subnet. :param pulumi.Input[bool] ipv4: Allocate an IPv4 address for the interface. The default value is `true`. :param pulumi.Input[bool] ipv6: If true, allocate an IPv6 address for the interface. The address will be automatically assigned from the specified subnet. :param pulumi.Input[str] ipv6_address: The private IPv6 address to assign to the instance. :param pulumi.Input[Sequence[pulumi.Input['ComputeInstanceNetworkInterfaceIpv6DnsRecordArgs']]] ipv6_dns_records: List of configurations for creating ipv6 DNS records. The structure is documented below. :param pulumi.Input[bool] nat: Provide a public address, for instance, to access the internet over NAT. :param pulumi.Input[Sequence[pulumi.Input['ComputeInstanceNetworkInterfaceNatDnsRecordArgs']]] nat_dns_records: List of configurations for creating ipv4 NAT DNS records. The structure is documented below. :param pulumi.Input[str] nat_ip_address: Provide a public address, for instance, to access the internet over NAT. Address should be already reserved in web UI. :param pulumi.Input[Sequence[pulumi.Input[str]]] security_group_ids: Security group ids for network interface. """ pulumi.set(__self__, "subnet_id", subnet_id) if dns_records is not None: pulumi.set(__self__, "dns_records", dns_records) if index is not None: pulumi.set(__self__, "index", index) if ip_address is not None: pulumi.set(__self__, "ip_address", ip_address) if ipv4 is not None: pulumi.set(__self__, "ipv4", ipv4) if ipv6 is not None: pulumi.set(__self__, "ipv6", ipv6) if ipv6_address is not None: pulumi.set(__self__, "ipv6_address", ipv6_address) if ipv6_dns_records is not None: pulumi.set(__self__, "ipv6_dns_records", ipv6_dns_records) if mac_address is not None: pulumi.set(__self__, "mac_address", mac_address) if nat is not None: pulumi.set(__self__, "nat", nat) if nat_dns_records is not None: pulumi.set(__self__, "nat_dns_records", nat_dns_records) if nat_ip_address is not None: pulumi.set(__self__, "nat_ip_address", nat_ip_address) if nat_ip_version is not None: pulumi.set(__self__, "nat_ip_version", nat_ip_version) if security_group_ids is not None: pulumi.set(__self__, "security_group_ids", security_group_ids) @property @pulumi.getter(name="subnetId") def subnet_id(self) -> pulumi.Input[str]: """ ID of the subnet to attach this interface to. The subnet must exist in the same zone where this instance will be created. """ return pulumi.get(self, "subnet_id") @subnet_id.setter def subnet_id(self, value: pulumi.Input[str]): pulumi.set(self, "subnet_id", value) @property @pulumi.getter(name="dnsRecords") def dns_records(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['ComputeInstanceNetworkInterfaceDnsRecordArgs']]]]: """ List of configurations for creating ipv4 DNS records. The structure is documented below. """ return pulumi.get(self, "dns_records") @dns_records.setter def dns_records(self, value: Optional[pulumi.Input[Sequence[pulumi.Input['ComputeInstanceNetworkInterfaceDnsRecordArgs']]]]): pulumi.set(self, "dns_records", value) @property @pulumi.getter def index(self) -> Optional[pulumi.Input[int]]: return pulumi.get(self, "index") @index.setter def index(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "index", value) @property @pulumi.getter(name="ipAddress") def ip_address(self) -> Optional[pulumi.Input[str]]: """ The private IP address to assign to the instance. If empty, the address will be automatically assigned from the specified subnet. """ return pulumi.get(self, "ip_address") @ip_address.setter def ip_address(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "ip_address", value) @property @pulumi.getter def ipv4(self) -> Optional[pulumi.Input[bool]]: """ Allocate an IPv4 address for the interface. The default value is `true`. """ return pulumi.get(self, "ipv4") @ipv4.setter def ipv4(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "ipv4", value) @property @pulumi.getter def ipv6(self) -> Optional[pulumi.Input[bool]]: """ If true, allocate an IPv6 address for the interface. The address will be automatically assigned from the specified subnet. """ return pulumi.get(self, "ipv6") @ipv6.setter def ipv6(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "ipv6", value) @property @pulumi.getter(name="ipv6Address") def ipv6_address(self) -> Optional[pulumi.Input[str]]: """ The private IPv6 address to assign to the instance. """ return pulumi.get(self, "ipv6_address") @ipv6_address.setter def ipv6_address(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "ipv6_address", value) @property @pulumi.getter(name="ipv6DnsRecords") def ipv6_dns_records(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['ComputeInstanceNetworkInterfaceIpv6DnsRecordArgs']]]]: """ List of configurations for creating ipv6 DNS records. The structure is documented below. """ return pulumi.get(self, "ipv6_dns_records") @ipv6_dns_records.setter def ipv6_dns_records(self, value: Optional[pulumi.Input[Sequence[pulumi.Input['ComputeInstanceNetworkInterfaceIpv6DnsRecordArgs']]]]): pulumi.set(self, "ipv6_dns_records", value) @property @pulumi.getter(name="macAddress") def mac_address(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "mac_address") @mac_address.setter def mac_address(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "mac_address", value) @property @pulumi.getter def nat(self) -> Optional[pulumi.Input[bool]]: """ Provide a public address, for instance, to access the internet over NAT. """ return pulumi.get(self, "nat") @nat.setter def nat(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "nat", value) @property @pulumi.getter(name="natDnsRecords") def nat_dns_records(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['ComputeInstanceNetworkInterfaceNatDnsRecordArgs']]]]: """ List of configurations for creating ipv4 NAT DNS records. The structure is documented below. """ return pulumi.get(self, "nat_dns_records") @nat_dns_records.setter def nat_dns_records(self, value: Optional[pulumi.Input[Sequence[pulumi.Input['ComputeInstanceNetworkInterfaceNatDnsRecordArgs']]]]): pulumi.set(self, "nat_dns_records", value) @property @pulumi.getter(name="natIpAddress") def nat_ip_address(self) -> Optional[pulumi.Input[str]]: """ Provide a public address, for instance, to access the internet over NAT. Address should be already reserved in web UI. """ return pulumi.get(self, "nat_ip_address") @nat_ip_address.setter def nat_ip_address(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "nat_ip_address", value) @property @pulumi.getter(name="natIpVersion") def nat_ip_version(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "nat_ip_version") @nat_ip_version.setter def nat_ip_version(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "nat_ip_version", value) @property @pulumi.getter(name="securityGroupIds") def security_group_ids(self) -> Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]: """ Security group ids for network interface. """ return pulumi.get(self, "security_group_ids") @security_group_ids.setter def security_group_ids(self, value: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]): pulumi.set(self, "security_group_ids", value) @pulumi.input_type class ComputeInstanceNetworkInterfaceDnsRecordArgs: def __init__(__self__, *, fqdn: pulumi.Input[str], dns_zone_id: Optional[pulumi.Input[str]] = None, ptr: Optional[pulumi.Input[bool]] = None, ttl: Optional[pulumi.Input[int]] = None): """ :param pulumi.Input[str] fqdn: DNS record FQDN (must have a dot at the end). :param pulumi.Input[str] dns_zone_id: DNS zone ID (if not set, private zone used). :param pulumi.Input[bool] ptr: When set to true, also create a PTR DNS record. :param pulumi.Input[int] ttl: DNS record TTL. in seconds """ pulumi.set(__self__, "fqdn", fqdn) if dns_zone_id is not None: pulumi.set(__self__, "dns_zone_id", dns_zone_id) if ptr is not None: pulumi.set(__self__, "ptr", ptr) if ttl is not None: pulumi.set(__self__, "ttl", ttl) @property @pulumi.getter def fqdn(self) -> pulumi.Input[str]: """ DNS record FQDN (must have a dot at the end). """ return pulumi.get(self, "fqdn") @fqdn.setter def fqdn(self, value: pulumi.Input[str]): pulumi.set(self, "fqdn", value) @property @pulumi.getter(name="dnsZoneId") def dns_zone_id(self) -> Optional[pulumi.Input[str]]: """ DNS zone ID (if not set, private zone used). """ return pulumi.get(self, "dns_zone_id") @dns_zone_id.setter def dns_zone_id(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "dns_zone_id", value) @property @pulumi.getter def ptr(self) -> Optional[pulumi.Input[bool]]: """ When set to true, also create a PTR DNS record. """ return pulumi.get(self, "ptr") @ptr.setter def ptr(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "ptr", value) @property @pulumi.getter def ttl(self) -> Optional[pulumi.Input[int]]: """ DNS record TTL. in seconds """ return pulumi.get(self, "ttl") @ttl.setter def ttl(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "ttl", value) @pulumi.input_type class ComputeInstanceNetworkInterfaceIpv6DnsRecordArgs: def __init__(__self__, *, fqdn: pulumi.Input[str], dns_zone_id: Optional[pulumi.Input[str]] = None, ptr: Optional[pulumi.Input[bool]] = None, ttl: Optional[pulumi.Input[int]] = None): """ :param pulumi.Input[str] fqdn: DNS record FQDN (must have a dot at the end). :param pulumi.Input[str] dns_zone_id: DNS zone ID (if not set, private zone used). :param pulumi.Input[bool] ptr: When set to true, also create a PTR DNS record. :param pulumi.Input[int] ttl: DNS record TTL. in seconds """ pulumi.set(__self__, "fqdn", fqdn) if dns_zone_id is not None: pulumi.set(__self__, "dns_zone_id", dns_zone_id) if ptr is not None: pulumi.set(__self__, "ptr", ptr) if ttl is not None: pulumi.set(__self__, "ttl", ttl) @property @pulumi.getter def fqdn(self) -> pulumi.Input[str]: """ DNS record FQDN (must have a dot at the end). """ return pulumi.get(self, "fqdn") @fqdn.setter def fqdn(self, value: pulumi.Input[str]): pulumi.set(self, "fqdn", value) @property @pulumi.getter(name="dnsZoneId") def dns_zone_id(self) -> Optional[pulumi.Input[str]]: """ DNS zone ID (if not set, private zone used). """ return pulumi.get(self, "dns_zone_id") @dns_zone_id.setter def dns_zone_id(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "dns_zone_id", value) @property @pulumi.getter def ptr(self) -> Optional[pulumi.Input[bool]]: """ When set to true, also create a PTR DNS record. """ return pulumi.get(self, "ptr") @ptr.setter def ptr(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "ptr", value) @property @pulumi.getter def ttl(self) -> Optional[pulumi.Input[int]]: """ DNS record TTL. in seconds """ return pulumi.get(self, "ttl") @ttl.setter def ttl(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "ttl", value) @pulumi.input_type class ComputeInstanceNetworkInterfaceNatDnsRecordArgs: def __init__(__self__, *, fqdn: pulumi.Input[str], dns_zone_id: Optional[pulumi.Input[str]] = None, ptr: Optional[pulumi.Input[bool]] = None, ttl: Optional[pulumi.Input[int]] = None): """ :param pulumi.Input[str] fqdn: DNS record FQDN (must have a dot at the end). :param pulumi.Input[str] dns_zone_id: DNS zone ID (if not set, private zone used). :param pulumi.Input[bool] ptr: When set to true, also create a PTR DNS record. :param pulumi.Input[int] ttl: DNS record TTL. in seconds """ pulumi.set(__self__, "fqdn", fqdn) if dns_zone_id is not None: pulumi.set(__self__, "dns_zone_id", dns_zone_id) if ptr is not None: pulumi.set(__self__, "ptr", ptr) if ttl is not None: pulumi.set(__self__, "ttl", ttl) @property @pulumi.getter def fqdn(self) -> pulumi.Input[str]: """ DNS record FQDN (must have a dot at the end). """ return pulumi.get(self, "fqdn") @fqdn.setter def fqdn(self, value: pulumi.Input[str]): pulumi.set(self, "fqdn", value) @property @pulumi.getter(name="dnsZoneId") def dns_zone_id(self) -> Optional[pulumi.Input[str]]: """ DNS zone ID (if not set, private zone used). """ return pulumi.get(self, "dns_zone_id") @dns_zone_id.setter def dns_zone_id(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "dns_zone_id", value) @property @pulumi.getter def ptr(self) -> Optional[pulumi.Input[bool]]: """ When set to true, also create a PTR DNS record. """ return pulumi.get(self, "ptr") @ptr.setter def ptr(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "ptr", value) @property @pulumi.getter def ttl(self) -> Optional[pulumi.Input[int]]: """ DNS record TTL. in seconds """ return pulumi.get(self, "ttl") @ttl.setter def ttl(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "ttl", value) @pulumi.input_type class ComputeInstancePlacementPolicyArgs: def __init__(__self__, *, placement_group_id: pulumi.Input[str]): """ :param pulumi.Input[str] placement_group_id: Specifies the id of the Placement Group to assign to the instance. """ pulumi.set(__self__, "placement_group_id", placement_group_id) @property @pulumi.getter(name="placementGroupId") def placement_group_id(self) -> pulumi.Input[str]: """ Specifies the id of the Placement Group to assign to the instance. """ return pulumi.get(self, "placement_group_id") @placement_group_id.setter def placement_group_id(self, value: pulumi.Input[str]): pulumi.set(self, "placement_group_id", value) @pulumi.input_type class ComputeInstanceResourcesArgs: def __init__(__self__, *, cores: pulumi.Input[int], memory: pulumi.Input[float], core_fraction: Optional[pulumi.Input[int]] = None, gpus: Optional[pulumi.Input[int]] = None): """ :param pulumi.Input[int] cores: CPU cores for the instance. :param pulumi.Input[float] memory: Memory size in GB. :param pulumi.Input[int] core_fraction: If provided, specifies baseline performance for a core as a percent. """ pulumi.set(__self__, "cores", cores) pulumi.set(__self__, "memory", memory) if core_fraction is not None: pulumi.set(__self__, "core_fraction", core_fraction) if gpus is not None: pulumi.set(__self__, "gpus", gpus) @property @pulumi.getter def cores(self) -> pulumi.Input[int]: """ CPU cores for the instance. """ return pulumi.get(self, "cores") @cores.setter def cores(self, value: pulumi.Input[int]): pulumi.set(self, "cores", value) @property @pulumi.getter def memory(self) -> pulumi.Input[float]: """ Memory size in GB. """ return pulumi.get(self, "memory") @memory.setter def memory(self, value: pulumi.Input[float]): pulumi.set(self, "memory", value) @property @pulumi.getter(name="coreFraction") def core_fraction(self) -> Optional[pulumi.Input[int]]: """ If provided, specifies baseline performance for a core as a percent. """ return pulumi.get(self, "core_fraction") @core_fraction.setter def core_fraction(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "core_fraction", value) @property @pulumi.getter def gpus(self) -> Optional[pulumi.Input[int]]: return pulumi.get(self, "gpus") @gpus.setter def gpus(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "gpus", value) @pulumi.input_type class ComputeInstanceSchedulingPolicyArgs: def __init__(__self__, *, preemptible: Optional[pulumi.Input[bool]] = None): """ :param pulumi.Input[bool] preemptible: Specifies if the instance is preemptible. Defaults to false. """ if preemptible is not None: pulumi.set(__self__, "preemptible", preemptible) @property @pulumi.getter def preemptible(self) -> Optional[pulumi.Input[bool]]: """ Specifies if the instance is preemptible. Defaults to false. """ return pulumi.get(self, "preemptible") @preemptible.setter def preemptible(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "preemptible", value) @pulumi.input_type class ComputeInstanceSecondaryDiskArgs: def __init__(__self__, *, disk_id: pulumi.Input[str], auto_delete: Optional[pulumi.Input[bool]] = None, device_name: Optional[pulumi.Input[str]] = None, mode: Optional[pulumi.Input[str]] = None): """ :param pulumi.Input[str] disk_id: ID of the disk that is attached to the instance. :param pulumi.Input[bool] auto_delete: Whether the disk is auto-deleted when the instance is deleted. The default value is false. :param pulumi.Input[str] device_name: Name that can be used to access an attached disk under `/dev/disk/by-id/`. :param pulumi.Input[str] mode: Type of access to the disk resource. By default, a disk is attached in `READ_WRITE` mode. """ pulumi.set(__self__, "disk_id", disk_id) if auto_delete is not None: pulumi.set(__self__, "auto_delete", auto_delete) if device_name is not None: pulumi.set(__self__, "device_name", device_name) if mode is not None: pulumi.set(__self__, "mode", mode) @property @pulumi.getter(name="diskId") def disk_id(self) -> pulumi.Input[str]: """ ID of the disk that is attached to the instance. """ return pulumi.get(self, "disk_id") @disk_id.setter def disk_id(self, value: pulumi.Input[str]): pulumi.set(self, "disk_id", value) @property @pulumi.getter(name="autoDelete") def auto_delete(self) -> Optional[pulumi.Input[bool]]: """ Whether the disk is auto-deleted when the instance is deleted. The default value is false. """ return pulumi.get(self, "auto_delete") @auto_delete.setter def auto_delete(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "auto_delete", value) @property @pulumi.getter(name="deviceName") def device_name(self) -> Optional[pulumi.Input[str]]: """ Name that can be used to access an attached disk under `/dev/disk/by-id/`. """ return pulumi.get(self, "device_name") @device_name.setter def device_name(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "device_name", value) @property @pulumi.getter def mode(self) -> Optional[pulumi.Input[str]]: """ Type of access to the disk resource. By default, a disk is attached in `READ_WRITE` mode. """ return pulumi.get(self, "mode") @mode.setter def mode(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "mode", value) @pulumi.input_type class DataprocClusterClusterConfigArgs: def __init__(__self__, *, subcluster_specs: pulumi.Input[Sequence[pulumi.Input['DataprocClusterClusterConfigSubclusterSpecArgs']]], hadoop: Optional[pulumi.Input['DataprocClusterClusterConfigHadoopArgs']] = None, version_id: Optional[pulumi.Input[str]] = None): """ :param pulumi.Input[Sequence[pulumi.Input['DataprocClusterClusterConfigSubclusterSpecArgs']]] subcluster_specs: Configuration of the Data Proc subcluster. The structure is documented below. :param pulumi.Input['DataprocClusterClusterConfigHadoopArgs'] hadoop: Data Proc specific options. The structure is documented below. :param pulumi.Input[str] version_id: Version of Data Proc image. """ pulumi.set(__self__, "subcluster_specs", subcluster_specs) if hadoop is not None: pulumi.set(__self__, "hadoop", hadoop) if version_id is not None: pulumi.set(__self__, "version_id", version_id) @property @pulumi.getter(name="subclusterSpecs") def subcluster_specs(self) -> pulumi.Input[Sequence[pulumi.Input['DataprocClusterClusterConfigSubclusterSpecArgs']]]: """ Configuration of the Data Proc subcluster. The structure is documented below. """ return pulumi.get(self, "subcluster_specs") @subcluster_specs.setter def subcluster_specs(self, value: pulumi.Input[Sequence[pulumi.Input['DataprocClusterClusterConfigSubclusterSpecArgs']]]): pulumi.set(self, "subcluster_specs", value) @property @pulumi.getter def hadoop(self) -> Optional[pulumi.Input['DataprocClusterClusterConfigHadoopArgs']]: """ Data Proc specific options. The structure is documented below. """ return pulumi.get(self, "hadoop") @hadoop.setter def hadoop(self, value: Optional[pulumi.Input['DataprocClusterClusterConfigHadoopArgs']]): pulumi.set(self, "hadoop", value) @property @pulumi.getter(name="versionId") def version_id(self) -> Optional[pulumi.Input[str]]: """ Version of Data Proc image. """ return pulumi.get(self, "version_id") @version_id.setter def version_id(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "version_id", value) @pulumi.input_type class DataprocClusterClusterConfigHadoopArgs: def __init__(__self__, *, properties: Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]] = None, services: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, ssh_public_keys: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None): """ :param pulumi.Input[Mapping[str, pulumi.Input[str]]] properties: A set of key/value pairs that are used to configure cluster services. :param pulumi.Input[Sequence[pulumi.Input[str]]] services: List of services to run on Data Proc cluster. :param pulumi.Input[Sequence[pulumi.Input[str]]] ssh_public_keys: List of SSH public keys to put to the hosts of the cluster. For information on how to connect to the cluster, see [the official documentation](https://cloud.yandex.com/docs/data-proc/operations/connect). """ if properties is not None: pulumi.set(__self__, "properties", properties) if services is not None: pulumi.set(__self__, "services", services) if ssh_public_keys is not None: pulumi.set(__self__, "ssh_public_keys", ssh_public_keys) @property @pulumi.getter def properties(self) -> Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]]: """ A set of key/value pairs that are used to configure cluster services. """ return pulumi.get(self, "properties") @properties.setter def properties(self, value: Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]]): pulumi.set(self, "properties", value) @property @pulumi.getter def services(self) -> Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]: """ List of services to run on Data Proc cluster. """ return pulumi.get(self, "services") @services.setter def services(self, value: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]): pulumi.set(self, "services", value) @property @pulumi.getter(name="sshPublicKeys") def ssh_public_keys(self) -> Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]: """ List of SSH public keys to put to the hosts of the cluster. For information on how to connect to the cluster, see [the official documentation](https://cloud.yandex.com/docs/data-proc/operations/connect). """ return pulumi.get(self, "ssh_public_keys") @ssh_public_keys.setter def ssh_public_keys(self, value: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]): pulumi.set(self, "ssh_public_keys", value) @pulumi.input_type class DataprocClusterClusterConfigSubclusterSpecArgs: def __init__(__self__, *, hosts_count: pulumi.Input[int], name: pulumi.Input[str], resources: pulumi.Input['DataprocClusterClusterConfigSubclusterSpecResourcesArgs'], role: pulumi.Input[str], subnet_id: pulumi.Input[str], autoscaling_config: Optional[pulumi.Input['DataprocClusterClusterConfigSubclusterSpecAutoscalingConfigArgs']] = None, id: Optional[pulumi.Input[str]] = None): """ :param pulumi.Input[int] hosts_count: Number of hosts within Data Proc subcluster. :param pulumi.Input[str] name: Name of the Data Proc subcluster. :param pulumi.Input['DataprocClusterClusterConfigSubclusterSpecResourcesArgs'] resources: Resources allocated to each host of the Data Proc subcluster. The structure is documented below. :param pulumi.Input[str] role: Role of the subcluster in the Data Proc cluster. :param pulumi.Input[str] subnet_id: The ID of the subnet, to which hosts of the subcluster belong. Subnets of all the subclusters must belong to the same VPC network. :param pulumi.Input['DataprocClusterClusterConfigSubclusterSpecAutoscalingConfigArgs'] autoscaling_config: Autoscaling configuration for compute subclusters. :param pulumi.Input[str] id: (Computed) ID of a new Data Proc cluster. """ pulumi.set(__self__, "hosts_count", hosts_count) pulumi.set(__self__, "name", name) pulumi.set(__self__, "resources", resources) pulumi.set(__self__, "role", role) pulumi.set(__self__, "subnet_id", subnet_id) if autoscaling_config is not None: pulumi.set(__self__, "autoscaling_config", autoscaling_config) if id is not None: pulumi.set(__self__, "id", id) @property @pulumi.getter(name="hostsCount") def hosts_count(self) -> pulumi.Input[int]: """ Number of hosts within Data Proc subcluster. """ return pulumi.get(self, "hosts_count") @hosts_count.setter def hosts_count(self, value: pulumi.Input[int]): pulumi.set(self, "hosts_count", value) @property @pulumi.getter def name(self) -> pulumi.Input[str]: """ Name of the Data Proc subcluster. """ return pulumi.get(self, "name") @name.setter def name(self, value: pulumi.Input[str]): pulumi.set(self, "name", value) @property @pulumi.getter def resources(self) -> pulumi.Input['DataprocClusterClusterConfigSubclusterSpecResourcesArgs']: """ Resources allocated to each host of the Data Proc subcluster. The structure is documented below. """ return pulumi.get(self, "resources") @resources.setter def resources(self, value: pulumi.Input['DataprocClusterClusterConfigSubclusterSpecResourcesArgs']): pulumi.set(self, "resources", value) @property @pulumi.getter def role(self) -> pulumi.Input[str]: """ Role of the subcluster in the Data Proc cluster. """ return pulumi.get(self, "role") @role.setter def role(self, value: pulumi.Input[str]): pulumi.set(self, "role", value) @property @pulumi.getter(name="subnetId") def subnet_id(self) -> pulumi.Input[str]: """ The ID of the subnet, to which hosts of the subcluster belong. Subnets of all the subclusters must belong to the same VPC network. """ return pulumi.get(self, "subnet_id") @subnet_id.setter def subnet_id(self, value: pulumi.Input[str]): pulumi.set(self, "subnet_id", value) @property @pulumi.getter(name="autoscalingConfig") def autoscaling_config(self) -> Optional[pulumi.Input['DataprocClusterClusterConfigSubclusterSpecAutoscalingConfigArgs']]: """ Autoscaling configuration for compute subclusters. """ return pulumi.get(self, "autoscaling_config") @autoscaling_config.setter def autoscaling_config(self, value: Optional[pulumi.Input['DataprocClusterClusterConfigSubclusterSpecAutoscalingConfigArgs']]): pulumi.set(self, "autoscaling_config", value) @property @pulumi.getter def id(self) -> Optional[pulumi.Input[str]]: """ (Computed) ID of a new Data Proc cluster. """ return pulumi.get(self, "id") @id.setter def id(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "id", value) @pulumi.input_type class DataprocClusterClusterConfigSubclusterSpecAutoscalingConfigArgs: def __init__(__self__, *, max_hosts_count: pulumi.Input[int], cpu_utilization_target: Optional[pulumi.Input[float]] = None, decommission_timeout: Optional[pulumi.Input[int]] = None, measurement_duration: Optional[pulumi.Input[int]] = None, preemptible: Optional[pulumi.Input[bool]] = None, stabilization_duration: Optional[pulumi.Input[int]] = None, warmup_duration: Optional[pulumi.Input[int]] = None): """ :param pulumi.Input[int] max_hosts_count: Maximum number of nodes in autoscaling subclusters. :param pulumi.Input[float] cpu_utilization_target: Defines an autoscaling rule based on the average CPU utilization of the instance group. If not set default autoscaling metric will be used. :param pulumi.Input[int] decommission_timeout: Timeout to gracefully decommission nodes during downscaling. In seconds. :param pulumi.Input[int] measurement_duration: Time in seconds allotted for averaging metrics. :param pulumi.Input[bool] preemptible: Bool flag -- whether to use preemptible compute instances. Preemptible instances are stopped at least once every 24 hours, and can be stopped at any time if their resources are needed by Compute. For more information, see [Preemptible Virtual Machines](https://cloud.yandex.com/docs/compute/concepts/preemptible-vm). :param pulumi.Input[int] stabilization_duration: Minimum amount of time in seconds allotted for monitoring before Instance Groups can reduce the number of instances in the group. During this time, the group size doesn't decrease, even if the new metric values indicate that it should. :param pulumi.Input[int] warmup_duration: The warmup time of the instance in seconds. During this time, traffic is sent to the instance, but instance metrics are not collected. """ pulumi.set(__self__, "max_hosts_count", max_hosts_count) if cpu_utilization_target is not None: pulumi.set(__self__, "cpu_utilization_target", cpu_utilization_target) if decommission_timeout is not None: pulumi.set(__self__, "decommission_timeout", decommission_timeout) if measurement_duration is not None: pulumi.set(__self__, "measurement_duration", measurement_duration) if preemptible is not None: pulumi.set(__self__, "preemptible", preemptible) if stabilization_duration is not None: pulumi.set(__self__, "stabilization_duration", stabilization_duration) if warmup_duration is not None: pulumi.set(__self__, "warmup_duration", warmup_duration) @property @pulumi.getter(name="maxHostsCount") def max_hosts_count(self) -> pulumi.Input[int]: """ Maximum number of nodes in autoscaling subclusters. """ return pulumi.get(self, "max_hosts_count") @max_hosts_count.setter def max_hosts_count(self, value: pulumi.Input[int]): pulumi.set(self, "max_hosts_count", value) @property @pulumi.getter(name="cpuUtilizationTarget") def cpu_utilization_target(self) -> Optional[pulumi.Input[float]]: """ Defines an autoscaling rule based on the average CPU utilization of the instance group. If not set default autoscaling metric will be used. """ return pulumi.get(self, "cpu_utilization_target") @cpu_utilization_target.setter def cpu_utilization_target(self, value: Optional[pulumi.Input[float]]): pulumi.set(self, "cpu_utilization_target", value) @property @pulumi.getter(name="decommissionTimeout") def decommission_timeout(self) -> Optional[pulumi.Input[int]]: """ Timeout to gracefully decommission nodes during downscaling. In seconds. """ return pulumi.get(self, "decommission_timeout") @decommission_timeout.setter def decommission_timeout(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "decommission_timeout", value) @property @pulumi.getter(name="measurementDuration") def measurement_duration(self) -> Optional[pulumi.Input[int]]: """ Time in seconds allotted for averaging metrics. """ return pulumi.get(self, "measurement_duration") @measurement_duration.setter def measurement_duration(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "measurement_duration", value) @property @pulumi.getter def preemptible(self) -> Optional[pulumi.Input[bool]]: """ Bool flag -- whether to use preemptible compute instances. Preemptible instances are stopped at least once every 24 hours, and can be stopped at any time if their resources are needed by Compute. For more information, see [Preemptible Virtual Machines](https://cloud.yandex.com/docs/compute/concepts/preemptible-vm). """ return pulumi.get(self, "preemptible") @preemptible.setter def preemptible(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "preemptible", value) @property @pulumi.getter(name="stabilizationDuration") def stabilization_duration(self) -> Optional[pulumi.Input[int]]: """ Minimum amount of time in seconds allotted for monitoring before Instance Groups can reduce the number of instances in the group. During this time, the group size doesn't decrease, even if the new metric values indicate that it should. """ return pulumi.get(self, "stabilization_duration") @stabilization_duration.setter def stabilization_duration(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "stabilization_duration", value) @property @pulumi.getter(name="warmupDuration") def warmup_duration(self) -> Optional[pulumi.Input[int]]: """ The warmup time of the instance in seconds. During this time, traffic is sent to the instance, but instance metrics are not collected. """ return pulumi.get(self, "warmup_duration") @warmup_duration.setter def warmup_duration(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "warmup_duration", value) @pulumi.input_type class DataprocClusterClusterConfigSubclusterSpecResourcesArgs: def __init__(__self__, *, disk_size: pulumi.Input[int], resource_preset_id: pulumi.Input[str], disk_type_id: Optional[pulumi.Input[str]] = None): """ :param pulumi.Input[int] disk_size: Volume of the storage available to a host, in gigabytes. :param pulumi.Input[str] resource_preset_id: The ID of the preset for computational resources available to a host. All available presets are listed in the [documentation](https://cloud.yandex.com/docs/data-proc/concepts/instance-types). :param pulumi.Input[str] disk_type_id: Type of the storage of a host. One of `network-hdd` (default) or `network-ssd`. """ pulumi.set(__self__, "disk_size", disk_size) pulumi.set(__self__, "resource_preset_id", resource_preset_id) if disk_type_id is not None: pulumi.set(__self__, "disk_type_id", disk_type_id) @property @pulumi.getter(name="diskSize") def disk_size(self) -> pulumi.Input[int]: """ Volume of the storage available to a host, in gigabytes. """ return pulumi.get(self, "disk_size") @disk_size.setter def disk_size(self, value: pulumi.Input[int]): pulumi.set(self, "disk_size", value) @property @pulumi.getter(name="resourcePresetId") def resource_preset_id(self) -> pulumi.Input[str]: """ The ID of the preset for computational resources available to a host. All available presets are listed in the [documentation](https://cloud.yandex.com/docs/data-proc/concepts/instance-types). """ return pulumi.get(self, "resource_preset_id") @resource_preset_id.setter def resource_preset_id(self, value: pulumi.Input[str]): pulumi.set(self, "resource_preset_id", value) @property @pulumi.getter(name="diskTypeId") def disk_type_id(self) -> Optional[pulumi.Input[str]]: """ Type of the storage of a host. One of `network-hdd` (default) or `network-ssd`. """ return pulumi.get(self, "disk_type_id") @disk_type_id.setter def disk_type_id(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "disk_type_id", value) @pulumi.input_type class FunctionContentArgs: def __init__(__self__, *, zip_filename: pulumi.Input[str]): pulumi.set(__self__, "zip_filename", zip_filename) @property @pulumi.getter(name="zipFilename") def zip_filename(self) -> pulumi.Input[str]: return pulumi.get(self, "zip_filename") @zip_filename.setter def zip_filename(self, value: pulumi.Input[str]): pulumi.set(self, "zip_filename", value) @pulumi.input_type class FunctionPackageArgs: def __init__(__self__, *, bucket_name: pulumi.Input[str], object_name: pulumi.Input[str], sha256: Optional[pulumi.Input[str]] = None): pulumi.set(__self__, "bucket_name", bucket_name) pulumi.set(__self__, "object_name", object_name) if sha256 is not None: pulumi.set(__self__, "sha256", sha256) @property @pulumi.getter(name="bucketName") def bucket_name(self) -> pulumi.Input[str]: return pulumi.get(self, "bucket_name") @bucket_name.setter def bucket_name(self, value: pulumi.Input[str]): pulumi.set(self, "bucket_name", value) @property @pulumi.getter(name="objectName") def object_name(self) -> pulumi.Input[str]: return pulumi.get(self, "object_name") @object_name.setter def object_name(self, value: pulumi.Input[str]): pulumi.set(self, "object_name", value) @property @pulumi.getter def sha256(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "sha256") @sha256.setter def sha256(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "sha256", value) @pulumi.input_type class FunctionScalingPolicyPolicyArgs: def __init__(__self__, *, tag: pulumi.Input[str], zone_instances_limit: Optional[pulumi.Input[int]] = None, zone_requests_limit: Optional[pulumi.Input[int]] = None): pulumi.set(__self__, "tag", tag) if zone_instances_limit is not None: pulumi.set(__self__, "zone_instances_limit", zone_instances_limit) if zone_requests_limit is not None: pulumi.set(__self__, "zone_requests_limit", zone_requests_limit) @property @pulumi.getter def tag(self) -> pulumi.Input[str]: return pulumi.get(self, "tag") @tag.setter def tag(self, value: pulumi.Input[str]): pulumi.set(self, "tag", value) @property @pulumi.getter(name="zoneInstancesLimit") def zone_instances_limit(self) -> Optional[pulumi.Input[int]]: return pulumi.get(self, "zone_instances_limit") @zone_instances_limit.setter def zone_instances_limit(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "zone_instances_limit", value) @property @pulumi.getter(name="zoneRequestsLimit") def zone_requests_limit(self) -> Optional[pulumi.Input[int]]: return pulumi.get(self, "zone_requests_limit") @zone_requests_limit.setter def zone_requests_limit(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "zone_requests_limit", value) @pulumi.input_type class FunctionTriggerDlqArgs: def __init__(__self__, *, queue_id: pulumi.Input[str], service_account_id: pulumi.Input[str]): pulumi.set(__self__, "queue_id", queue_id) pulumi.set(__self__, "service_account_id", service_account_id) @property @pulumi.getter(name="queueId") def queue_id(self) -> pulumi.Input[str]: return pulumi.get(self, "queue_id") @queue_id.setter def queue_id(self, value: pulumi.Input[str]): pulumi.set(self, "queue_id", value) @property @pulumi.getter(name="serviceAccountId") def service_account_id(self) -> pulumi.Input[str]: return pulumi.get(self, "service_account_id") @service_account_id.setter def service_account_id(self, value: pulumi.Input[str]): pulumi.set(self, "service_account_id", value) @pulumi.input_type class FunctionTriggerFunctionArgs: def __init__(__self__, *, id: pulumi.Input[str], retry_attempts: Optional[pulumi.Input[str]] = None, retry_interval: Optional[pulumi.Input[str]] = None, service_account_id: Optional[pulumi.Input[str]] = None, tag: Optional[pulumi.Input[str]] = None): pulumi.set(__self__, "id", id) if retry_attempts is not None: pulumi.set(__self__, "retry_attempts", retry_attempts) if retry_interval is not None: pulumi.set(__self__, "retry_interval", retry_interval) if service_account_id is not None: pulumi.set(__self__, "service_account_id", service_account_id) if tag is not None: pulumi.set(__self__, "tag", tag) @property @pulumi.getter def id(self) -> pulumi.Input[str]: return pulumi.get(self, "id") @id.setter def id(self, value: pulumi.Input[str]): pulumi.set(self, "id", value) @property @pulumi.getter(name="retryAttempts") def retry_attempts(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "retry_attempts") @retry_attempts.setter def retry_attempts(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "retry_attempts", value) @property @pulumi.getter(name="retryInterval") def retry_interval(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "retry_interval") @retry_interval.setter def retry_interval(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "retry_interval", value) @property @pulumi.getter(name="serviceAccountId") def service_account_id(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "service_account_id") @service_account_id.setter def service_account_id(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "service_account_id", value) @property @pulumi.getter def tag(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "tag") @tag.setter def tag(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "tag", value) @pulumi.input_type class FunctionTriggerIotArgs: def __init__(__self__, *, registry_id: pulumi.Input[str], device_id: Optional[pulumi.Input[str]] = None, topic: Optional[pulumi.Input[str]] = None): pulumi.set(__self__, "registry_id", registry_id) if device_id is not None: pulumi.set(__self__, "device_id", device_id) if topic is not None: pulumi.set(__self__, "topic", topic) @property @pulumi.getter(name="registryId") def registry_id(self) -> pulumi.Input[str]: return pulumi.get(self, "registry_id") @registry_id.setter def registry_id(self, value: pulumi.Input[str]): pulumi.set(self, "registry_id", value) @property @pulumi.getter(name="deviceId") def device_id(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "device_id") @device_id.setter def device_id(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "device_id", value) @property @pulumi.getter def topic(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "topic") @topic.setter def topic(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "topic", value) @pulumi.input_type class FunctionTriggerLogGroupArgs: def __init__(__self__, *, batch_cutoff: pulumi.Input[str], log_group_ids: pulumi.Input[Sequence[pulumi.Input[str]]], batch_size: Optional[pulumi.Input[str]] = None): pulumi.set(__self__, "batch_cutoff", batch_cutoff) pulumi.set(__self__, "log_group_ids", log_group_ids) if batch_size is not None: pulumi.set(__self__, "batch_size", batch_size) @property @pulumi.getter(name="batchCutoff") def batch_cutoff(self) -> pulumi.Input[str]: return pulumi.get(self, "batch_cutoff") @batch_cutoff.setter def batch_cutoff(self, value: pulumi.Input[str]): pulumi.set(self, "batch_cutoff", value) @property @pulumi.getter(name="logGroupIds") def log_group_ids(self) -> pulumi.Input[Sequence[pulumi.Input[str]]]: return pulumi.get(self, "log_group_ids") @log_group_ids.setter def log_group_ids(self, value: pulumi.Input[Sequence[pulumi.Input[str]]]): pulumi.set(self, "log_group_ids", value) @property @pulumi.getter(name="batchSize") def batch_size(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "batch_size") @batch_size.setter def batch_size(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "batch_size", value) @pulumi.input_type class FunctionTriggerLoggingArgs: def __init__(__self__, *, batch_cutoff: pulumi.Input[str], group_id: pulumi.Input[str], levels: pulumi.Input[Sequence[pulumi.Input[str]]], resource_ids: pulumi.Input[Sequence[pulumi.Input[str]]], resource_types: pulumi.Input[Sequence[pulumi.Input[str]]], batch_size: Optional[pulumi.Input[str]] = None): pulumi.set(__self__, "batch_cutoff", batch_cutoff) pulumi.set(__self__, "group_id", group_id) pulumi.set(__self__, "levels", levels) pulumi.set(__self__, "resource_ids", resource_ids) pulumi.set(__self__, "resource_types", resource_types) if batch_size is not None: pulumi.set(__self__, "batch_size", batch_size) @property @pulumi.getter(name="batchCutoff") def batch_cutoff(self) -> pulumi.Input[str]: return pulumi.get(self, "batch_cutoff") @batch_cutoff.setter def batch_cutoff(self, value: pulumi.Input[str]): pulumi.set(self, "batch_cutoff", value) @property @pulumi.getter(name="groupId") def group_id(self) -> pulumi.Input[str]: return pulumi.get(self, "group_id") @group_id.setter def group_id(self, value: pulumi.Input[str]): pulumi.set(self, "group_id", value) @property @pulumi.getter def levels(self) -> pulumi.Input[Sequence[pulumi.Input[str]]]: return pulumi.get(self, "levels") @levels.setter def levels(self, value: pulumi.Input[Sequence[pulumi.Input[str]]]): pulumi.set(self, "levels", value) @property @pulumi.getter(name="resourceIds") def resource_ids(self) -> pulumi.Input[Sequence[pulumi.Input[str]]]: return pulumi.get(self, "resource_ids") @resource_ids.setter def resource_ids(self, value: pulumi.Input[Sequence[pulumi.Input[str]]]): pulumi.set(self, "resource_ids", value) @property @pulumi.getter(name="resourceTypes") def resource_types(self) -> pulumi.Input[Sequence[pulumi.Input[str]]]: return pulumi.get(self, "resource_types") @resource_types.setter def resource_types(self, value: pulumi.Input[Sequence[pulumi.Input[str]]]): pulumi.set(self, "resource_types", value) @property @pulumi.getter(name="batchSize") def batch_size(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "batch_size") @batch_size.setter def batch_size(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "batch_size", value) @pulumi.input_type class FunctionTriggerMessageQueueArgs: def __init__(__self__, *, batch_cutoff: pulumi.Input[str], queue_id: pulumi.Input[str], service_account_id: pulumi.Input[str], batch_size: Optional[pulumi.Input[str]] = None, visibility_timeout: Optional[pulumi.Input[str]] = None): pulumi.set(__self__, "batch_cutoff", batch_cutoff) pulumi.set(__self__, "queue_id", queue_id) pulumi.set(__self__, "service_account_id", service_account_id) if batch_size is not None: pulumi.set(__self__, "batch_size", batch_size) if visibility_timeout is not None: pulumi.set(__self__, "visibility_timeout", visibility_timeout) @property @pulumi.getter(name="batchCutoff") def batch_cutoff(self) -> pulumi.Input[str]: return pulumi.get(self, "batch_cutoff") @batch_cutoff.setter def batch_cutoff(self, value: pulumi.Input[str]): pulumi.set(self, "batch_cutoff", value) @property @pulumi.getter(name="queueId") def queue_id(self) -> pulumi.Input[str]: return pulumi.get(self, "queue_id") @queue_id.setter def queue_id(self, value: pulumi.Input[str]): pulumi.set(self, "queue_id", value) @property @pulumi.getter(name="serviceAccountId") def service_account_id(self) -> pulumi.Input[str]: return pulumi.get(self, "service_account_id") @service_account_id.setter def service_account_id(self, value: pulumi.Input[str]): pulumi.set(self, "service_account_id", value) @property @pulumi.getter(name="batchSize") def batch_size(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "batch_size") @batch_size.setter def batch_size(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "batch_size", value) @property @pulumi.getter(name="visibilityTimeout") def visibility_timeout(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "visibility_timeout") @visibility_timeout.setter def visibility_timeout(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "visibility_timeout", value) @pulumi.input_type class FunctionTriggerObjectStorageArgs: def __init__(__self__, *, bucket_id: pulumi.Input[str], create: Optional[pulumi.Input[bool]] = None, delete: Optional[pulumi.Input[bool]] = None, prefix: Optional[pulumi.Input[str]] = None, suffix: Optional[pulumi.Input[str]] = None, update: Optional[pulumi.Input[bool]] = None): pulumi.set(__self__, "bucket_id", bucket_id) if create is not None: pulumi.set(__self__, "create", create) if delete is not None: pulumi.set(__self__, "delete", delete) if prefix is not None: pulumi.set(__self__, "prefix", prefix) if suffix is not None: pulumi.set(__self__, "suffix", suffix) if update is not None: pulumi.set(__self__, "update", update) @property @pulumi.getter(name="bucketId") def bucket_id(self) -> pulumi.Input[str]: return pulumi.get(self, "bucket_id") @bucket_id.setter def bucket_id(self, value: pulumi.Input[str]): pulumi.set(self, "bucket_id", value) @property @pulumi.getter def create(self) -> Optional[pulumi.Input[bool]]: return pulumi.get(self, "create") @create.setter def create(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "create", value) @property @pulumi.getter def delete(self) -> Optional[pulumi.Input[bool]]: return pulumi.get(self, "delete") @delete.setter def delete(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "delete", value) @property @pulumi.getter def prefix(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "prefix") @prefix.setter def prefix(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "prefix", value) @property @pulumi.getter def suffix(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "suffix") @suffix.setter def suffix(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "suffix", value) @property @pulumi.getter def update(self) -> Optional[pulumi.Input[bool]]: return pulumi.get(self, "update") @update.setter def update(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "update", value) @pulumi.input_type class FunctionTriggerTimerArgs: def __init__(__self__, *, cron_expression: pulumi.Input[str]): pulumi.set(__self__, "cron_expression", cron_expression) @property @pulumi.getter(name="cronExpression") def cron_expression(self) -> pulumi.Input[str]: return pulumi.get(self, "cron_expression") @cron_expression.setter def cron_expression(self, value: pulumi.Input[str]): pulumi.set(self, "cron_expression", value) @pulumi.input_type class KubernetesClusterKmsProviderArgs: def __init__(__self__, *, key_id: Optional[pulumi.Input[str]] = None): """ :param pulumi.Input[str] key_id: KMS key ID. """ if key_id is not None: pulumi.set(__self__, "key_id", key_id) @property @pulumi.getter(name="keyId") def key_id(self) -> Optional[pulumi.Input[str]]: """ KMS key ID. """ return pulumi.get(self, "key_id") @key_id.setter def key_id(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "key_id", value) @pulumi.input_type class KubernetesClusterMasterArgs: def __init__(__self__, *, cluster_ca_certificate: Optional[pulumi.Input[str]] = None, external_v4_address: Optional[pulumi.Input[str]] = None, external_v4_endpoint: Optional[pulumi.Input[str]] = None, internal_v4_address: Optional[pulumi.Input[str]] = None, internal_v4_endpoint: Optional[pulumi.Input[str]] = None, maintenance_policy: Optional[pulumi.Input['KubernetesClusterMasterMaintenancePolicyArgs']] = None, public_ip: Optional[pulumi.Input[bool]] = None, regional: Optional[pulumi.Input['KubernetesClusterMasterRegionalArgs']] = None, security_group_ids: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, version: Optional[pulumi.Input[str]] = None, version_info: Optional[pulumi.Input['KubernetesClusterMasterVersionInfoArgs']] = None, zonal: Optional[pulumi.Input['KubernetesClusterMasterZonalArgs']] = None): """ :param pulumi.Input[str] cluster_ca_certificate: (Computed) PEM-encoded public certificate that is the root of trust for the Kubernetes cluster. :param pulumi.Input[str] external_v4_address: (Computed) An IPv4 external network address that is assigned to the master. :param pulumi.Input[str] external_v4_endpoint: (Computed) External endpoint that can be used to access Kubernetes cluster API from the internet (outside of the cloud). :param pulumi.Input[str] internal_v4_address: (Computed) An IPv4 internal network address that is assigned to the master. :param pulumi.Input[str] internal_v4_endpoint: (Computed) Internal endpoint that can be used to connect to the master from cloud networks. :param pulumi.Input['KubernetesClusterMasterMaintenancePolicyArgs'] maintenance_policy: (Optional) (Computed) Maintenance policy for Kubernetes master. If policy is omitted, automatic revision upgrades of the kubernetes master are enabled and could happen at any time. Revision upgrades are performed only within the same minor version, e.g. 1.13. Minor version upgrades (e.g. 1.13->1.14) should be performed manually. The structure is documented below. :param pulumi.Input[bool] public_ip: (Optional) (Computed) Boolean flag. When `true`, Kubernetes master will have visible ipv4 address. :param pulumi.Input['KubernetesClusterMasterRegionalArgs'] regional: (Optional) Initialize parameters for Regional Master (highly available master). The structure is documented below. :param pulumi.Input[Sequence[pulumi.Input[str]]] security_group_ids: (Optional) List of security group IDs to which the Kubernetes cluster belongs. :param pulumi.Input[str] version: (Optional) (Computed) Version of Kubernetes that will be used for master. :param pulumi.Input['KubernetesClusterMasterVersionInfoArgs'] version_info: (Computed) Information about cluster version. The structure is documented below. :param pulumi.Input['KubernetesClusterMasterZonalArgs'] zonal: (Optional) Initialize parameters for Zonal Master (single node master). The structure is documented below. """ if cluster_ca_certificate is not None: pulumi.set(__self__, "cluster_ca_certificate", cluster_ca_certificate) if external_v4_address is not None: pulumi.set(__self__, "external_v4_address", external_v4_address) if external_v4_endpoint is not None: pulumi.set(__self__, "external_v4_endpoint", external_v4_endpoint) if internal_v4_address is not None: pulumi.set(__self__, "internal_v4_address", internal_v4_address) if internal_v4_endpoint is not None: pulumi.set(__self__, "internal_v4_endpoint", internal_v4_endpoint) if maintenance_policy is not None: pulumi.set(__self__, "maintenance_policy", maintenance_policy) if public_ip is not None: pulumi.set(__self__, "public_ip", public_ip) if regional is not None: pulumi.set(__self__, "regional", regional) if security_group_ids is not None: pulumi.set(__self__, "security_group_ids", security_group_ids) if version is not None: pulumi.set(__self__, "version", version) if version_info is not None: pulumi.set(__self__, "version_info", version_info) if zonal is not None: pulumi.set(__self__, "zonal", zonal) @property @pulumi.getter(name="clusterCaCertificate") def cluster_ca_certificate(self) -> Optional[pulumi.Input[str]]: """ (Computed) PEM-encoded public certificate that is the root of trust for the Kubernetes cluster. """ return pulumi.get(self, "cluster_ca_certificate") @cluster_ca_certificate.setter def cluster_ca_certificate(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "cluster_ca_certificate", value) @property @pulumi.getter(name="externalV4Address") def external_v4_address(self) -> Optional[pulumi.Input[str]]: """ (Computed) An IPv4 external network address that is assigned to the master. """ return pulumi.get(self, "external_v4_address") @external_v4_address.setter def external_v4_address(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "external_v4_address", value) @property @pulumi.getter(name="externalV4Endpoint") def external_v4_endpoint(self) -> Optional[pulumi.Input[str]]: """ (Computed) External endpoint that can be used to access Kubernetes cluster API from the internet (outside of the cloud). """ return pulumi.get(self, "external_v4_endpoint") @external_v4_endpoint.setter def external_v4_endpoint(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "external_v4_endpoint", value) @property @pulumi.getter(name="internalV4Address") def internal_v4_address(self) -> Optional[pulumi.Input[str]]: """ (Computed) An IPv4 internal network address that is assigned to the master. """ return pulumi.get(self, "internal_v4_address") @internal_v4_address.setter def internal_v4_address(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "internal_v4_address", value) @property @pulumi.getter(name="internalV4Endpoint") def internal_v4_endpoint(self) -> Optional[pulumi.Input[str]]: """ (Computed) Internal endpoint that can be used to connect to the master from cloud networks. """ return pulumi.get(self, "internal_v4_endpoint") @internal_v4_endpoint.setter def internal_v4_endpoint(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "internal_v4_endpoint", value) @property @pulumi.getter(name="maintenancePolicy") def maintenance_policy(self) -> Optional[pulumi.Input['KubernetesClusterMasterMaintenancePolicyArgs']]: """ (Optional) (Computed) Maintenance policy for Kubernetes master. If policy is omitted, automatic revision upgrades of the kubernetes master are enabled and could happen at any time. Revision upgrades are performed only within the same minor version, e.g. 1.13. Minor version upgrades (e.g. 1.13->1.14) should be performed manually. The structure is documented below. """ return pulumi.get(self, "maintenance_policy") @maintenance_policy.setter def maintenance_policy(self, value: Optional[pulumi.Input['KubernetesClusterMasterMaintenancePolicyArgs']]): pulumi.set(self, "maintenance_policy", value) @property @pulumi.getter(name="publicIp") def public_ip(self) -> Optional[pulumi.Input[bool]]: """ (Optional) (Computed) Boolean flag. When `true`, Kubernetes master will have visible ipv4 address. """ return pulumi.get(self, "public_ip") @public_ip.setter def public_ip(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "public_ip", value) @property @pulumi.getter def regional(self) -> Optional[pulumi.Input['KubernetesClusterMasterRegionalArgs']]: """ (Optional) Initialize parameters for Regional Master (highly available master). The structure is documented below. """ return pulumi.get(self, "regional") @regional.setter def regional(self, value: Optional[pulumi.Input['KubernetesClusterMasterRegionalArgs']]): pulumi.set(self, "regional", value) @property @pulumi.getter(name="securityGroupIds") def security_group_ids(self) -> Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]: """ (Optional) List of security group IDs to which the Kubernetes cluster belongs. """ return pulumi.get(self, "security_group_ids") @security_group_ids.setter def security_group_ids(self, value: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]): pulumi.set(self, "security_group_ids", value) @property @pulumi.getter def version(self) -> Optional[pulumi.Input[str]]: """ (Optional) (Computed) Version of Kubernetes that will be used for master. """ return pulumi.get(self, "version") @version.setter def version(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "version", value) @property @pulumi.getter(name="versionInfo") def version_info(self) -> Optional[pulumi.Input['KubernetesClusterMasterVersionInfoArgs']]: """ (Computed) Information about cluster version. The structure is documented below. """ return pulumi.get(self, "version_info") @version_info.setter def version_info(self, value: Optional[pulumi.Input['KubernetesClusterMasterVersionInfoArgs']]): pulumi.set(self, "version_info", value) @property @pulumi.getter def zonal(self) -> Optional[pulumi.Input['KubernetesClusterMasterZonalArgs']]: """ (Optional) Initialize parameters for Zonal Master (single node master). The structure is documented below. """ return pulumi.get(self, "zonal") @zonal.setter def zonal(self, value: Optional[pulumi.Input['KubernetesClusterMasterZonalArgs']]): pulumi.set(self, "zonal", value) @pulumi.input_type class KubernetesClusterMasterMaintenancePolicyArgs: def __init__(__self__, *, auto_upgrade: pulumi.Input[bool], maintenance_windows: Optional[pulumi.Input[Sequence[pulumi.Input['KubernetesClusterMasterMaintenancePolicyMaintenanceWindowArgs']]]] = None): """ :param pulumi.Input[bool] auto_upgrade: (Required) Boolean flag that specifies if master can be upgraded automatically. When omitted, default value is TRUE. :param pulumi.Input[Sequence[pulumi.Input['KubernetesClusterMasterMaintenancePolicyMaintenanceWindowArgs']]] maintenance_windows: (Optional) (Computed) This structure specifies maintenance window, when update for master is allowed. When omitted, it defaults to any time. To specify time of day interval, for all days, one element should be provided, with two fields set, `start_time` and `duration`. Please see `zonal_cluster_resource_name` config example. """ pulumi.set(__self__, "auto_upgrade", auto_upgrade) if maintenance_windows is not None: pulumi.set(__self__, "maintenance_windows", maintenance_windows) @property @pulumi.getter(name="autoUpgrade") def auto_upgrade(self) -> pulumi.Input[bool]: """ (Required) Boolean flag that specifies if master can be upgraded automatically. When omitted, default value is TRUE. """ return pulumi.get(self, "auto_upgrade") @auto_upgrade.setter def auto_upgrade(self, value: pulumi.Input[bool]): pulumi.set(self, "auto_upgrade", value) @property @pulumi.getter(name="maintenanceWindows") def maintenance_windows(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['KubernetesClusterMasterMaintenancePolicyMaintenanceWindowArgs']]]]: """ (Optional) (Computed) This structure specifies maintenance window, when update for master is allowed. When omitted, it defaults to any time. To specify time of day interval, for all days, one element should be provided, with two fields set, `start_time` and `duration`. Please see `zonal_cluster_resource_name` config example. """ return pulumi.get(self, "maintenance_windows") @maintenance_windows.setter def maintenance_windows(self, value: Optional[pulumi.Input[Sequence[pulumi.Input['KubernetesClusterMasterMaintenancePolicyMaintenanceWindowArgs']]]]): pulumi.set(self, "maintenance_windows", value) @pulumi.input_type class KubernetesClusterMasterMaintenancePolicyMaintenanceWindowArgs: def __init__(__self__, *, duration: pulumi.Input[str], start_time: pulumi.Input[str], day: Optional[pulumi.Input[str]] = None): pulumi.set(__self__, "duration", duration) pulumi.set(__self__, "start_time", start_time) if day is not None: pulumi.set(__self__, "day", day) @property @pulumi.getter def duration(self) -> pulumi.Input[str]: return pulumi.get(self, "duration") @duration.setter def duration(self, value: pulumi.Input[str]): pulumi.set(self, "duration", value) @property @pulumi.getter(name="startTime") def start_time(self) -> pulumi.Input[str]: return pulumi.get(self, "start_time") @start_time.setter def start_time(self, value: pulumi.Input[str]): pulumi.set(self, "start_time", value) @property @pulumi.getter def day(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "day") @day.setter def day(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "day", value) @pulumi.input_type class KubernetesClusterMasterRegionalArgs: def __init__(__self__, *, region: pulumi.Input[str], locations: Optional[pulumi.Input[Sequence[pulumi.Input['KubernetesClusterMasterRegionalLocationArgs']]]] = None): """ :param pulumi.Input[str] region: (Required) Name of availability region (e.g. "ru-central1"), where master instances will be allocated. :param pulumi.Input[Sequence[pulumi.Input['KubernetesClusterMasterRegionalLocationArgs']]] locations: Array of locations, where master instances will be allocated. The structure is documented below. """ pulumi.set(__self__, "region", region) if locations is not None: pulumi.set(__self__, "locations", locations) @property @pulumi.getter def region(self) -> pulumi.Input[str]: """ (Required) Name of availability region (e.g. "ru-central1"), where master instances will be allocated. """ return pulumi.get(self, "region") @region.setter def region(self, value: pulumi.Input[str]): pulumi.set(self, "region", value) @property @pulumi.getter def locations(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['KubernetesClusterMasterRegionalLocationArgs']]]]: """ Array of locations, where master instances will be allocated. The structure is documented below. """ return pulumi.get(self, "locations") @locations.setter def locations(self, value: Optional[pulumi.Input[Sequence[pulumi.Input['KubernetesClusterMasterRegionalLocationArgs']]]]): pulumi.set(self, "locations", value) @pulumi.input_type class KubernetesClusterMasterRegionalLocationArgs: def __init__(__self__, *, subnet_id: Optional[pulumi.Input[str]] = None, zone: Optional[pulumi.Input[str]] = None): """ :param pulumi.Input[str] subnet_id: (Optional) ID of the subnet. :param pulumi.Input[str] zone: (Optional) ID of the availability zone. """ if subnet_id is not None: pulumi.set(__self__, "subnet_id", subnet_id) if zone is not None: pulumi.set(__self__, "zone", zone) @property @pulumi.getter(name="subnetId") def subnet_id(self) -> Optional[pulumi.Input[str]]: """ (Optional) ID of the subnet. """ return pulumi.get(self, "subnet_id") @subnet_id.setter def subnet_id(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "subnet_id", value) @property @pulumi.getter def zone(self) -> Optional[pulumi.Input[str]]: """ (Optional) ID of the availability zone. """ return pulumi.get(self, "zone") @zone.setter def zone(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "zone", value) @pulumi.input_type class KubernetesClusterMasterVersionInfoArgs: def __init__(__self__, *, current_version: Optional[pulumi.Input[str]] = None, new_revision_available: Optional[pulumi.Input[bool]] = None, new_revision_summary: Optional[pulumi.Input[str]] = None, version_deprecated: Optional[pulumi.Input[bool]] = None): """ :param pulumi.Input[str] current_version: Current Kubernetes version, major.minor (e.g. 1.15). :param pulumi.Input[bool] new_revision_available: Boolean flag. Newer revisions may include Kubernetes patches (e.g 1.15.1 > 1.15.2) as well as some internal component updates - new features or bug fixes in yandex-specific components either on the master or nodes. :param pulumi.Input[str] new_revision_summary: Human readable description of the changes to be applied when updating to the latest revision. Empty if new_revision_available is false. :param pulumi.Input[bool] version_deprecated: Boolean flag. The current version is on the deprecation schedule, component (master or node group) should be upgraded. """ if current_version is not None: pulumi.set(__self__, "current_version", current_version) if new_revision_available is not None: pulumi.set(__self__, "new_revision_available", new_revision_available) if new_revision_summary is not None: pulumi.set(__self__, "new_revision_summary", new_revision_summary) if version_deprecated is not None: pulumi.set(__self__, "version_deprecated", version_deprecated) @property @pulumi.getter(name="currentVersion") def current_version(self) -> Optional[pulumi.Input[str]]: """ Current Kubernetes version, major.minor (e.g. 1.15). """ return pulumi.get(self, "current_version") @current_version.setter def current_version(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "current_version", value) @property @pulumi.getter(name="newRevisionAvailable") def new_revision_available(self) -> Optional[pulumi.Input[bool]]: """ Boolean flag. Newer revisions may include Kubernetes patches (e.g 1.15.1 > 1.15.2) as well as some internal component updates - new features or bug fixes in yandex-specific components either on the master or nodes. """ return pulumi.get(self, "new_revision_available") @new_revision_available.setter def new_revision_available(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "new_revision_available", value) @property @pulumi.getter(name="newRevisionSummary") def new_revision_summary(self) -> Optional[pulumi.Input[str]]: """ Human readable description of the changes to be applied when updating to the latest revision. Empty if new_revision_available is false. """ return pulumi.get(self, "new_revision_summary") @new_revision_summary.setter def new_revision_summary(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "new_revision_summary", value) @property @pulumi.getter(name="versionDeprecated") def version_deprecated(self) -> Optional[pulumi.Input[bool]]: """ Boolean flag. The current version is on the deprecation schedule, component (master or node group) should be upgraded. """ return pulumi.get(self, "version_deprecated") @version_deprecated.setter def version_deprecated(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "version_deprecated", value) @pulumi.input_type class KubernetesClusterMasterZonalArgs: def __init__(__self__, *, subnet_id: Optional[pulumi.Input[str]] = None, zone: Optional[pulumi.Input[str]] = None): """ :param pulumi.Input[str] subnet_id: (Optional) ID of the subnet. :param pulumi.Input[str] zone: (Optional) ID of the availability zone. """ if subnet_id is not None: pulumi.set(__self__, "subnet_id", subnet_id) if zone is not None: pulumi.set(__self__, "zone", zone) @property @pulumi.getter(name="subnetId") def subnet_id(self) -> Optional[pulumi.Input[str]]: """ (Optional) ID of the subnet. """ return pulumi.get(self, "subnet_id") @subnet_id.setter def subnet_id(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "subnet_id", value) @property @pulumi.getter def zone(self) -> Optional[pulumi.Input[str]]: """ (Optional) ID of the availability zone. """ return pulumi.get(self, "zone") @zone.setter def zone(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "zone", value) @pulumi.input_type class KubernetesClusterNetworkImplementationArgs: def __init__(__self__, *, cilium: Optional[pulumi.Input['KubernetesClusterNetworkImplementationCiliumArgs']] = None): """ :param pulumi.Input['KubernetesClusterNetworkImplementationCiliumArgs'] cilium: (Optional) Cilium network implementation configuration. No options exist. """ if cilium is not None: pulumi.set(__self__, "cilium", cilium) @property @pulumi.getter def cilium(self) -> Optional[pulumi.Input['KubernetesClusterNetworkImplementationCiliumArgs']]: """ (Optional) Cilium network implementation configuration. No options exist. """ return pulumi.get(self, "cilium") @cilium.setter def cilium(self, value: Optional[pulumi.Input['KubernetesClusterNetworkImplementationCiliumArgs']]): pulumi.set(self, "cilium", value) @pulumi.input_type class KubernetesClusterNetworkImplementationCiliumArgs: def __init__(__self__): pass @pulumi.input_type class KubernetesNodeGroupAllocationPolicyArgs: def __init__(__self__, *, locations: Optional[pulumi.Input[Sequence[pulumi.Input['KubernetesNodeGroupAllocationPolicyLocationArgs']]]] = None): """ :param pulumi.Input[Sequence[pulumi.Input['KubernetesNodeGroupAllocationPolicyLocationArgs']]] locations: Repeated field, that specify subnets (zones), that will be used by node group compute instances. The structure is documented below. """ if locations is not None: pulumi.set(__self__, "locations", locations) @property @pulumi.getter def locations(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['KubernetesNodeGroupAllocationPolicyLocationArgs']]]]: """ Repeated field, that specify subnets (zones), that will be used by node group compute instances. The structure is documented below. """ return pulumi.get(self, "locations") @locations.setter def locations(self, value: Optional[pulumi.Input[Sequence[pulumi.Input['KubernetesNodeGroupAllocationPolicyLocationArgs']]]]): pulumi.set(self, "locations", value) @pulumi.input_type class KubernetesNodeGroupAllocationPolicyLocationArgs: def __init__(__self__, *, subnet_id: Optional[pulumi.Input[str]] = None, zone: Optional[pulumi.Input[str]] = None): """ :param pulumi.Input[str] subnet_id: ID of the subnet, that will be used by one compute instance in node group. :param pulumi.Input[str] zone: ID of the availability zone where for one compute instance in node group. """ if subnet_id is not None: warnings.warn("""The 'subnet_id' field has been deprecated. Please use 'subnet_ids under network_interface' instead.""", DeprecationWarning) pulumi.log.warn("""subnet_id is deprecated: The 'subnet_id' field has been deprecated. Please use 'subnet_ids under network_interface' instead.""") if subnet_id is not None: pulumi.set(__self__, "subnet_id", subnet_id) if zone is not None: pulumi.set(__self__, "zone", zone) @property @pulumi.getter(name="subnetId") def subnet_id(self) -> Optional[pulumi.Input[str]]: """ ID of the subnet, that will be used by one compute instance in node group. """ return pulumi.get(self, "subnet_id") @subnet_id.setter def subnet_id(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "subnet_id", value) @property @pulumi.getter def zone(self) -> Optional[pulumi.Input[str]]: """ ID of the availability zone where for one compute instance in node group. """ return pulumi.get(self, "zone") @zone.setter def zone(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "zone", value) @pulumi.input_type class KubernetesNodeGroupDeployPolicyArgs: def __init__(__self__, *, max_expansion: pulumi.Input[int], max_unavailable: pulumi.Input[int]): """ :param pulumi.Input[int] max_expansion: The maximum number of instances that can be temporarily allocated above the group's target size during the update. :param pulumi.Input[int] max_unavailable: The maximum number of running instances that can be taken offline during update. """ pulumi.set(__self__, "max_expansion", max_expansion) pulumi.set(__self__, "max_unavailable", max_unavailable) @property @pulumi.getter(name="maxExpansion") def max_expansion(self) -> pulumi.Input[int]: """ The maximum number of instances that can be temporarily allocated above the group's target size during the update. """ return pulumi.get(self, "max_expansion") @max_expansion.setter def max_expansion(self, value: pulumi.Input[int]): pulumi.set(self, "max_expansion", value) @property @pulumi.getter(name="maxUnavailable") def max_unavailable(self) -> pulumi.Input[int]: """ The maximum number of running instances that can be taken offline during update. """ return pulumi.get(self, "max_unavailable") @max_unavailable.setter def max_unavailable(self, value: pulumi.Input[int]): pulumi.set(self, "max_unavailable", value) @pulumi.input_type class KubernetesNodeGroupInstanceTemplateArgs: def __init__(__self__, *, boot_disk: Optional[pulumi.Input['KubernetesNodeGroupInstanceTemplateBootDiskArgs']] = None, metadata: Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]] = None, nat: Optional[pulumi.Input[bool]] = None, network_acceleration_type: Optional[pulumi.Input[str]] = None, network_interfaces: Optional[pulumi.Input[Sequence[pulumi.Input['KubernetesNodeGroupInstanceTemplateNetworkInterfaceArgs']]]] = None, placement_policy: Optional[pulumi.Input['KubernetesNodeGroupInstanceTemplatePlacementPolicyArgs']] = None, platform_id: Optional[pulumi.Input[str]] = None, resources: Optional[pulumi.Input['KubernetesNodeGroupInstanceTemplateResourcesArgs']] = None, scheduling_policy: Optional[pulumi.Input['KubernetesNodeGroupInstanceTemplateSchedulingPolicyArgs']] = None): """ :param pulumi.Input['KubernetesNodeGroupInstanceTemplateBootDiskArgs'] boot_disk: The specifications for boot disks that will be attached to the instance. The structure is documented below. :param pulumi.Input[Mapping[str, pulumi.Input[str]]] metadata: The set of metadata `key:value` pairs assigned to this instance template. This includes custom metadata and predefined keys. * `resources.0.memory` - The memory size allocated to the instance. * `resources.0.cores` - Number of CPU cores allocated to the instance. * `resources.0.core_fraction` - Baseline core performance as a percent. * `resources.0.gpus` - Number of GPU cores allocated to the instance. :param pulumi.Input[bool] nat: A public address that can be used to access the internet over NAT. :param pulumi.Input[str] network_acceleration_type: Type of network acceleration. Values: `standard`, `software_accelerated`. :param pulumi.Input[Sequence[pulumi.Input['KubernetesNodeGroupInstanceTemplateNetworkInterfaceArgs']]] network_interfaces: An array with the network interfaces that will be attached to the instance. The structure is documented below. :param pulumi.Input['KubernetesNodeGroupInstanceTemplatePlacementPolicyArgs'] placement_policy: The placement policy configuration. The structure is documented below. :param pulumi.Input[str] platform_id: The ID of the hardware platform configuration for the node group compute instances. :param pulumi.Input['KubernetesNodeGroupInstanceTemplateSchedulingPolicyArgs'] scheduling_policy: The scheduling policy for the instances in node group. The structure is documented below. """ if boot_disk is not None: pulumi.set(__self__, "boot_disk", boot_disk) if metadata is not None: pulumi.set(__self__, "metadata", metadata) if nat is not None: warnings.warn("""The 'nat' field has been deprecated. Please use 'nat under network_interface' instead.""", DeprecationWarning) pulumi.log.warn("""nat is deprecated: The 'nat' field has been deprecated. Please use 'nat under network_interface' instead.""") if nat is not None: pulumi.set(__self__, "nat", nat) if network_acceleration_type is not None: pulumi.set(__self__, "network_acceleration_type", network_acceleration_type) if network_interfaces is not None: pulumi.set(__self__, "network_interfaces", network_interfaces) if placement_policy is not None: pulumi.set(__self__, "placement_policy", placement_policy) if platform_id is not None: pulumi.set(__self__, "platform_id", platform_id) if resources is not None: pulumi.set(__self__, "resources", resources) if scheduling_policy is not None: pulumi.set(__self__, "scheduling_policy", scheduling_policy) @property @pulumi.getter(name="bootDisk") def boot_disk(self) -> Optional[pulumi.Input['KubernetesNodeGroupInstanceTemplateBootDiskArgs']]: """ The specifications for boot disks that will be attached to the instance. The structure is documented below. """ return pulumi.get(self, "boot_disk") @boot_disk.setter def boot_disk(self, value: Optional[pulumi.Input['KubernetesNodeGroupInstanceTemplateBootDiskArgs']]): pulumi.set(self, "boot_disk", value) @property @pulumi.getter def metadata(self) -> Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]]: """ The set of metadata `key:value` pairs assigned to this instance template. This includes custom metadata and predefined keys. * `resources.0.memory` - The memory size allocated to the instance. * `resources.0.cores` - Number of CPU cores allocated to the instance. * `resources.0.core_fraction` - Baseline core performance as a percent. * `resources.0.gpus` - Number of GPU cores allocated to the instance. """ return pulumi.get(self, "metadata") @metadata.setter def metadata(self, value: Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]]): pulumi.set(self, "metadata", value) @property @pulumi.getter def nat(self) -> Optional[pulumi.Input[bool]]: """ A public address that can be used to access the internet over NAT. """ return pulumi.get(self, "nat") @nat.setter def nat(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "nat", value) @property @pulumi.getter(name="networkAccelerationType") def network_acceleration_type(self) -> Optional[pulumi.Input[str]]: """ Type of network acceleration. Values: `standard`, `software_accelerated`. """ return pulumi.get(self, "network_acceleration_type") @network_acceleration_type.setter def network_acceleration_type(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "network_acceleration_type", value) @property @pulumi.getter(name="networkInterfaces") def network_interfaces(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['KubernetesNodeGroupInstanceTemplateNetworkInterfaceArgs']]]]: """ An array with the network interfaces that will be attached to the instance. The structure is documented below. """ return pulumi.get(self, "network_interfaces") @network_interfaces.setter def network_interfaces(self, value: Optional[pulumi.Input[Sequence[pulumi.Input['KubernetesNodeGroupInstanceTemplateNetworkInterfaceArgs']]]]): pulumi.set(self, "network_interfaces", value) @property @pulumi.getter(name="placementPolicy") def placement_policy(self) -> Optional[pulumi.Input['KubernetesNodeGroupInstanceTemplatePlacementPolicyArgs']]: """ The placement policy configuration. The structure is documented below. """ return pulumi.get(self, "placement_policy") @placement_policy.setter def placement_policy(self, value: Optional[pulumi.Input['KubernetesNodeGroupInstanceTemplatePlacementPolicyArgs']]): pulumi.set(self, "placement_policy", value) @property @pulumi.getter(name="platformId") def platform_id(self) -> Optional[pulumi.Input[str]]: """ The ID of the hardware platform configuration for the node group compute instances. """ return pulumi.get(self, "platform_id") @platform_id.setter def platform_id(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "platform_id", value) @property @pulumi.getter def resources(self) -> Optional[pulumi.Input['KubernetesNodeGroupInstanceTemplateResourcesArgs']]: return pulumi.get(self, "resources") @resources.setter def resources(self, value: Optional[pulumi.Input['KubernetesNodeGroupInstanceTemplateResourcesArgs']]): pulumi.set(self, "resources", value) @property @pulumi.getter(name="schedulingPolicy") def scheduling_policy(self) -> Optional[pulumi.Input['KubernetesNodeGroupInstanceTemplateSchedulingPolicyArgs']]: """ The scheduling policy for the instances in node group. The structure is documented below. """ return pulumi.get(self, "scheduling_policy") @scheduling_policy.setter def scheduling_policy(self, value: Optional[pulumi.Input['KubernetesNodeGroupInstanceTemplateSchedulingPolicyArgs']]): pulumi.set(self, "scheduling_policy", value) @pulumi.input_type class KubernetesNodeGroupInstanceTemplateBootDiskArgs: def __init__(__self__, *, size: Optional[pulumi.Input[int]] = None, type: Optional[pulumi.Input[str]] = None): """ :param pulumi.Input[int] size: The number of instances in the node group. :param pulumi.Input[str] type: The disk type. """ if size is not None: pulumi.set(__self__, "size", size) if type is not None: pulumi.set(__self__, "type", type) @property @pulumi.getter def size(self) -> Optional[pulumi.Input[int]]: """ The number of instances in the node group. """ return pulumi.get(self, "size") @size.setter def size(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "size", value) @property @pulumi.getter def type(self) -> Optional[pulumi.Input[str]]: """ The disk type. """ return pulumi.get(self, "type") @type.setter def type(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "type", value) @pulumi.input_type class KubernetesNodeGroupInstanceTemplateNetworkInterfaceArgs: def __init__(__self__, *, subnet_ids: pulumi.Input[Sequence[pulumi.Input[str]]], ipv4: Optional[pulumi.Input[bool]] = None, ipv6: Optional[pulumi.Input[bool]] = None, nat: Optional[pulumi.Input[bool]] = None, security_group_ids: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None): """ :param pulumi.Input[Sequence[pulumi.Input[str]]] subnet_ids: The IDs of the subnets. :param pulumi.Input[bool] ipv4: Allocate an IPv4 address for the interface. The default value is `true`. :param pulumi.Input[bool] ipv6: If true, allocate an IPv6 address for the interface. The address will be automatically assigned from the specified subnet. :param pulumi.Input[bool] nat: A public address that can be used to access the internet over NAT. :param pulumi.Input[Sequence[pulumi.Input[str]]] security_group_ids: Security group ids for network interface. """ pulumi.set(__self__, "subnet_ids", subnet_ids) if ipv4 is not None: pulumi.set(__self__, "ipv4", ipv4) if ipv6 is not None: pulumi.set(__self__, "ipv6", ipv6) if nat is not None: pulumi.set(__self__, "nat", nat) if security_group_ids is not None: pulumi.set(__self__, "security_group_ids", security_group_ids) @property @pulumi.getter(name="subnetIds") def subnet_ids(self) -> pulumi.Input[Sequence[pulumi.Input[str]]]: """ The IDs of the subnets. """ return pulumi.get(self, "subnet_ids") @subnet_ids.setter def subnet_ids(self, value: pulumi.Input[Sequence[pulumi.Input[str]]]): pulumi.set(self, "subnet_ids", value) @property @pulumi.getter def ipv4(self) -> Optional[pulumi.Input[bool]]: """ Allocate an IPv4 address for the interface. The default value is `true`. """ return pulumi.get(self, "ipv4") @ipv4.setter def ipv4(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "ipv4", value) @property @pulumi.getter def ipv6(self) -> Optional[pulumi.Input[bool]]: """ If true, allocate an IPv6 address for the interface. The address will be automatically assigned from the specified subnet. """ return pulumi.get(self, "ipv6") @ipv6.setter def ipv6(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "ipv6", value) @property @pulumi.getter def nat(self) -> Optional[pulumi.Input[bool]]: """ A public address that can be used to access the internet over NAT. """ return pulumi.get(self, "nat") @nat.setter def nat(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "nat", value) @property @pulumi.getter(name="securityGroupIds") def security_group_ids(self) -> Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]: """ Security group ids for network interface. """ return pulumi.get(self, "security_group_ids") @security_group_ids.setter def security_group_ids(self, value: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]): pulumi.set(self, "security_group_ids", value) @pulumi.input_type class KubernetesNodeGroupInstanceTemplatePlacementPolicyArgs: def __init__(__self__, *, placement_group_id: pulumi.Input[str]): """ :param pulumi.Input[str] placement_group_id: Specifies the id of the Placement Group to assign to the instances. """ pulumi.set(__self__, "placement_group_id", placement_group_id) @property @pulumi.getter(name="placementGroupId") def placement_group_id(self) -> pulumi.Input[str]: """ Specifies the id of the Placement Group to assign to the instances. """ return pulumi.get(self, "placement_group_id") @placement_group_id.setter def placement_group_id(self, value: pulumi.Input[str]): pulumi.set(self, "placement_group_id", value) @pulumi.input_type class KubernetesNodeGroupInstanceTemplateResourcesArgs: def __init__(__self__, *, core_fraction: Optional[pulumi.Input[int]] = None, cores: Optional[pulumi.Input[int]] = None, gpus: Optional[pulumi.Input[int]] = None, memory: Optional[pulumi.Input[float]] = None): if core_fraction is not None: pulumi.set(__self__, "core_fraction", core_fraction) if cores is not None: pulumi.set(__self__, "cores", cores) if gpus is not None: pulumi.set(__self__, "gpus", gpus) if memory is not None: pulumi.set(__self__, "memory", memory) @property @pulumi.getter(name="coreFraction") def core_fraction(self) -> Optional[pulumi.Input[int]]: return pulumi.get(self, "core_fraction") @core_fraction.setter def core_fraction(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "core_fraction", value) @property @pulumi.getter def cores(self) -> Optional[pulumi.Input[int]]: return pulumi.get(self, "cores") @cores.setter def cores(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "cores", value) @property @pulumi.getter def gpus(self) -> Optional[pulumi.Input[int]]: return pulumi.get(self, "gpus") @gpus.setter def gpus(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "gpus", value) @property @pulumi.getter def memory(self) -> Optional[pulumi.Input[float]]: return pulumi.get(self, "memory") @memory.setter def memory(self, value: Optional[pulumi.Input[float]]): pulumi.set(self, "memory", value) @pulumi.input_type class KubernetesNodeGroupInstanceTemplateSchedulingPolicyArgs: def __init__(__self__, *, preemptible: Optional[pulumi.Input[bool]] = None): """ :param pulumi.Input[bool] preemptible: Specifies if the instance is preemptible. Defaults to false. --- """ if preemptible is not None: pulumi.set(__self__, "preemptible", preemptible) @property @pulumi.getter def preemptible(self) -> Optional[pulumi.Input[bool]]: """ Specifies if the instance is preemptible. Defaults to false. --- """ return pulumi.get(self, "preemptible") @preemptible.setter def preemptible(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "preemptible", value) @pulumi.input_type class KubernetesNodeGroupMaintenancePolicyArgs: def __init__(__self__, *, auto_repair: pulumi.Input[bool], auto_upgrade: pulumi.Input[bool], maintenance_windows: Optional[pulumi.Input[Sequence[pulumi.Input['KubernetesNodeGroupMaintenancePolicyMaintenanceWindowArgs']]]] = None): """ :param pulumi.Input[bool] auto_repair: Boolean flag that specifies if node group can be repaired automatically. When omitted, default value is TRUE. :param pulumi.Input[bool] auto_upgrade: Boolean flag that specifies if node group can be upgraded automatically. When omitted, default value is TRUE. :param pulumi.Input[Sequence[pulumi.Input['KubernetesNodeGroupMaintenancePolicyMaintenanceWindowArgs']]] maintenance_windows: (Computed) Set of day intervals, when maintenance is allowed for this node group. When omitted, it defaults to any time. """ pulumi.set(__self__, "auto_repair", auto_repair) pulumi.set(__self__, "auto_upgrade", auto_upgrade) if maintenance_windows is not None: pulumi.set(__self__, "maintenance_windows", maintenance_windows) @property @pulumi.getter(name="autoRepair") def auto_repair(self) -> pulumi.Input[bool]: """ Boolean flag that specifies if node group can be repaired automatically. When omitted, default value is TRUE. """ return pulumi.get(self, "auto_repair") @auto_repair.setter def auto_repair(self, value: pulumi.Input[bool]): pulumi.set(self, "auto_repair", value) @property @pulumi.getter(name="autoUpgrade") def auto_upgrade(self) -> pulumi.Input[bool]: """ Boolean flag that specifies if node group can be upgraded automatically. When omitted, default value is TRUE. """ return pulumi.get(self, "auto_upgrade") @auto_upgrade.setter def auto_upgrade(self, value: pulumi.Input[bool]): pulumi.set(self, "auto_upgrade", value) @property @pulumi.getter(name="maintenanceWindows") def maintenance_windows(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['KubernetesNodeGroupMaintenancePolicyMaintenanceWindowArgs']]]]: """ (Computed) Set of day intervals, when maintenance is allowed for this node group. When omitted, it defaults to any time. """ return pulumi.get(self, "maintenance_windows") @maintenance_windows.setter def maintenance_windows(self, value: Optional[pulumi.Input[Sequence[pulumi.Input['KubernetesNodeGroupMaintenancePolicyMaintenanceWindowArgs']]]]): pulumi.set(self, "maintenance_windows", value) @pulumi.input_type class KubernetesNodeGroupMaintenancePolicyMaintenanceWindowArgs: def __init__(__self__, *, duration: pulumi.Input[str], start_time: pulumi.Input[str], day: Optional[pulumi.Input[str]] = None): pulumi.set(__self__, "duration", duration) pulumi.set(__self__, "start_time", start_time) if day is not None: pulumi.set(__self__, "day", day) @property @pulumi.getter def duration(self) -> pulumi.Input[str]: return pulumi.get(self, "duration") @duration.setter def duration(self, value: pulumi.Input[str]): pulumi.set(self, "duration", value) @property @pulumi.getter(name="startTime") def start_time(self) -> pulumi.Input[str]: return pulumi.get(self, "start_time") @start_time.setter def start_time(self, value: pulumi.Input[str]): pulumi.set(self, "start_time", value) @property @pulumi.getter def day(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "day") @day.setter def day(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "day", value) @pulumi.input_type class KubernetesNodeGroupScalePolicyArgs: def __init__(__self__, *, auto_scale: Optional[pulumi.Input['KubernetesNodeGroupScalePolicyAutoScaleArgs']] = None, fixed_scale: Optional[pulumi.Input['KubernetesNodeGroupScalePolicyFixedScaleArgs']] = None): """ :param pulumi.Input['KubernetesNodeGroupScalePolicyAutoScaleArgs'] auto_scale: Scale policy for an autoscaled node group. The structure is documented below. :param pulumi.Input['KubernetesNodeGroupScalePolicyFixedScaleArgs'] fixed_scale: Scale policy for a fixed scale node group. The structure is documented below. """ if auto_scale is not None: pulumi.set(__self__, "auto_scale", auto_scale) if fixed_scale is not None: pulumi.set(__self__, "fixed_scale", fixed_scale) @property @pulumi.getter(name="autoScale") def auto_scale(self) -> Optional[pulumi.Input['KubernetesNodeGroupScalePolicyAutoScaleArgs']]: """ Scale policy for an autoscaled node group. The structure is documented below. """ return pulumi.get(self, "auto_scale") @auto_scale.setter def auto_scale(self, value: Optional[pulumi.Input['KubernetesNodeGroupScalePolicyAutoScaleArgs']]): pulumi.set(self, "auto_scale", value) @property @pulumi.getter(name="fixedScale") def fixed_scale(self) -> Optional[pulumi.Input['KubernetesNodeGroupScalePolicyFixedScaleArgs']]: """ Scale policy for a fixed scale node group. The structure is documented below. """ return pulumi.get(self, "fixed_scale") @fixed_scale.setter def fixed_scale(self, value: Optional[pulumi.Input['KubernetesNodeGroupScalePolicyFixedScaleArgs']]): pulumi.set(self, "fixed_scale", value) @pulumi.input_type class KubernetesNodeGroupScalePolicyAutoScaleArgs: def __init__(__self__, *, initial: pulumi.Input[int], max: pulumi.Input[int], min: pulumi.Input[int]): """ :param pulumi.Input[int] initial: Initial number of instances in the node group. :param pulumi.Input[int] max: Maximum number of instances in the node group. :param pulumi.Input[int] min: Minimum number of instances in the node group. """ pulumi.set(__self__, "initial", initial) pulumi.set(__self__, "max", max) pulumi.set(__self__, "min", min) @property @pulumi.getter def initial(self) -> pulumi.Input[int]: """ Initial number of instances in the node group. """ return pulumi.get(self, "initial") @initial.setter def initial(self, value: pulumi.Input[int]): pulumi.set(self, "initial", value) @property @pulumi.getter def max(self) -> pulumi.Input[int]: """ Maximum number of instances in the node group. """ return pulumi.get(self, "max") @max.setter def max(self, value: pulumi.Input[int]): pulumi.set(self, "max", value) @property @pulumi.getter def min(self) -> pulumi.Input[int]: """ Minimum number of instances in the node group. """ return pulumi.get(self, "min") @min.setter def min(self, value: pulumi.Input[int]): pulumi.set(self, "min", value) @pulumi.input_type class KubernetesNodeGroupScalePolicyFixedScaleArgs: def __init__(__self__, *, size: Optional[pulumi.Input[int]] = None): """ :param pulumi.Input[int] size: The number of instances in the node group. """ if size is not None: pulumi.set(__self__, "size", size) @property @pulumi.getter def size(self) -> Optional[pulumi.Input[int]]: """ The number of instances in the node group. """ return pulumi.get(self, "size") @size.setter def size(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "size", value) @pulumi.input_type class KubernetesNodeGroupVersionInfoArgs: def __init__(__self__, *, current_version: Optional[pulumi.Input[str]] = None, new_revision_available: Optional[pulumi.Input[bool]] = None, new_revision_summary: Optional[pulumi.Input[str]] = None, version_deprecated: Optional[pulumi.Input[bool]] = None): """ :param pulumi.Input[str] current_version: Current Kubernetes version, major.minor (e.g. 1.15). :param pulumi.Input[bool] new_revision_available: True/false flag. Newer revisions may include Kubernetes patches (e.g 1.15.1 > 1.15.2) as well as some internal component updates - new features or bug fixes in yandex-specific components either on the master or nodes. :param pulumi.Input[str] new_revision_summary: Human readable description of the changes to be applied when updating to the latest revision. Empty if new_revision_available is false. :param pulumi.Input[bool] version_deprecated: True/false flag. The current version is on the deprecation schedule, component (master or node group) should be upgraded. """ if current_version is not None: pulumi.set(__self__, "current_version", current_version) if new_revision_available is not None: pulumi.set(__self__, "new_revision_available", new_revision_available) if new_revision_summary is not None: pulumi.set(__self__, "new_revision_summary", new_revision_summary) if version_deprecated is not None: pulumi.set(__self__, "version_deprecated", version_deprecated) @property @pulumi.getter(name="currentVersion") def current_version(self) -> Optional[pulumi.Input[str]]: """ Current Kubernetes version, major.minor (e.g. 1.15). """ return pulumi.get(self, "current_version") @current_version.setter def current_version(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "current_version", value) @property @pulumi.getter(name="newRevisionAvailable") def new_revision_available(self) -> Optional[pulumi.Input[bool]]: """ True/false flag. Newer revisions may include Kubernetes patches (e.g 1.15.1 > 1.15.2) as well as some internal component updates - new features or bug fixes in yandex-specific components either on the master or nodes. """ return pulumi.get(self, "new_revision_available") @new_revision_available.setter def new_revision_available(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "new_revision_available", value) @property @pulumi.getter(name="newRevisionSummary") def new_revision_summary(self) -> Optional[pulumi.Input[str]]: """ Human readable description of the changes to be applied when updating to the latest revision. Empty if new_revision_available is false. """ return pulumi.get(self, "new_revision_summary") @new_revision_summary.setter def new_revision_summary(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "new_revision_summary", value) @property @pulumi.getter(name="versionDeprecated") def version_deprecated(self) -> Optional[pulumi.Input[bool]]: """ True/false flag. The current version is on the deprecation schedule, component (master or node group) should be upgraded. """ return pulumi.get(self, "version_deprecated") @version_deprecated.setter def version_deprecated(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "version_deprecated", value) @pulumi.input_type class LbNetworkLoadBalancerAttachedTargetGroupArgs: def __init__(__self__, *, healthchecks: pulumi.Input[Sequence[pulumi.Input['LbNetworkLoadBalancerAttachedTargetGroupHealthcheckArgs']]], target_group_id: pulumi.Input[str]): """ :param pulumi.Input[Sequence[pulumi.Input['LbNetworkLoadBalancerAttachedTargetGroupHealthcheckArgs']]] healthchecks: A HealthCheck resource. The structure is documented below. :param pulumi.Input[str] target_group_id: ID of the target group. """ pulumi.set(__self__, "healthchecks", healthchecks) pulumi.set(__self__, "target_group_id", target_group_id) @property @pulumi.getter def healthchecks(self) -> pulumi.Input[Sequence[pulumi.Input['LbNetworkLoadBalancerAttachedTargetGroupHealthcheckArgs']]]: """ A HealthCheck resource. The structure is documented below. """ return pulumi.get(self, "healthchecks") @healthchecks.setter def healthchecks(self, value: pulumi.Input[Sequence[pulumi.Input['LbNetworkLoadBalancerAttachedTargetGroupHealthcheckArgs']]]): pulumi.set(self, "healthchecks", value) @property @pulumi.getter(name="targetGroupId") def target_group_id(self) -> pulumi.Input[str]: """ ID of the target group. """ return pulumi.get(self, "target_group_id") @target_group_id.setter def target_group_id(self, value: pulumi.Input[str]): pulumi.set(self, "target_group_id", value) @pulumi.input_type class LbNetworkLoadBalancerAttachedTargetGroupHealthcheckArgs: def __init__(__self__, *, name: pulumi.Input[str], healthy_threshold: Optional[pulumi.Input[int]] = None, http_options: Optional[pulumi.Input['LbNetworkLoadBalancerAttachedTargetGroupHealthcheckHttpOptionsArgs']] = None, interval: Optional[pulumi.Input[int]] = None, tcp_options: Optional[pulumi.Input['LbNetworkLoadBalancerAttachedTargetGroupHealthcheckTcpOptionsArgs']] = None, timeout: Optional[pulumi.Input[int]] = None, unhealthy_threshold: Optional[pulumi.Input[int]] = None): """ :param pulumi.Input[str] name: Name of the listener. The name must be unique for each listener on a single load balancer. :param pulumi.Input[int] healthy_threshold: Number of successful health checks required in order to set the `HEALTHY` status for the target. :param pulumi.Input['LbNetworkLoadBalancerAttachedTargetGroupHealthcheckHttpOptionsArgs'] http_options: Options for HTTP health check. The structure is documented below. :param pulumi.Input[int] interval: The interval between health checks. The default is 2 seconds. :param pulumi.Input['LbNetworkLoadBalancerAttachedTargetGroupHealthcheckTcpOptionsArgs'] tcp_options: Options for TCP health check. The structure is documented below. :param pulumi.Input[int] timeout: Timeout for a target to return a response for the health check. The default is 1 second. :param pulumi.Input[int] unhealthy_threshold: Number of failed health checks before changing the status to `UNHEALTHY`. The default is 2. """ pulumi.set(__self__, "name", name) if healthy_threshold is not None: pulumi.set(__self__, "healthy_threshold", healthy_threshold) if http_options is not None: pulumi.set(__self__, "http_options", http_options) if interval is not None: pulumi.set(__self__, "interval", interval) if tcp_options is not None: pulumi.set(__self__, "tcp_options", tcp_options) if timeout is not None: pulumi.set(__self__, "timeout", timeout) if unhealthy_threshold is not None: pulumi.set(__self__, "unhealthy_threshold", unhealthy_threshold) @property @pulumi.getter def name(self) -> pulumi.Input[str]: """ Name of the listener. The name must be unique for each listener on a single load balancer. """ return pulumi.get(self, "name") @name.setter def name(self, value: pulumi.Input[str]): pulumi.set(self, "name", value) @property @pulumi.getter(name="healthyThreshold") def healthy_threshold(self) -> Optional[pulumi.Input[int]]: """ Number of successful health checks required in order to set the `HEALTHY` status for the target. """ return pulumi.get(self, "healthy_threshold") @healthy_threshold.setter def healthy_threshold(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "healthy_threshold", value) @property @pulumi.getter(name="httpOptions") def http_options(self) -> Optional[pulumi.Input['LbNetworkLoadBalancerAttachedTargetGroupHealthcheckHttpOptionsArgs']]: """ Options for HTTP health check. The structure is documented below. """ return pulumi.get(self, "http_options") @http_options.setter def http_options(self, value: Optional[pulumi.Input['LbNetworkLoadBalancerAttachedTargetGroupHealthcheckHttpOptionsArgs']]): pulumi.set(self, "http_options", value) @property @pulumi.getter def interval(self) -> Optional[pulumi.Input[int]]: """ The interval between health checks. The default is 2 seconds. """ return pulumi.get(self, "interval") @interval.setter def interval(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "interval", value) @property @pulumi.getter(name="tcpOptions") def tcp_options(self) -> Optional[pulumi.Input['LbNetworkLoadBalancerAttachedTargetGroupHealthcheckTcpOptionsArgs']]: """ Options for TCP health check. The structure is documented below. """ return pulumi.get(self, "tcp_options") @tcp_options.setter def tcp_options(self, value: Optional[pulumi.Input['LbNetworkLoadBalancerAttachedTargetGroupHealthcheckTcpOptionsArgs']]): pulumi.set(self, "tcp_options", value) @property @pulumi.getter def timeout(self) -> Optional[pulumi.Input[int]]: """ Timeout for a target to return a response for the health check. The default is 1 second. """ return pulumi.get(self, "timeout") @timeout.setter def timeout(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "timeout", value) @property @pulumi.getter(name="unhealthyThreshold") def unhealthy_threshold(self) -> Optional[pulumi.Input[int]]: """ Number of failed health checks before changing the status to `UNHEALTHY`. The default is 2. """ return pulumi.get(self, "unhealthy_threshold") @unhealthy_threshold.setter def unhealthy_threshold(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "unhealthy_threshold", value) @pulumi.input_type class LbNetworkLoadBalancerAttachedTargetGroupHealthcheckHttpOptionsArgs: def __init__(__self__, *, port: pulumi.Input[int], path: Optional[pulumi.Input[str]] = None): """ :param pulumi.Input[int] port: Port for incoming traffic. :param pulumi.Input[str] path: URL path to set for health checking requests for every target in the target group. For example `/ping`. The default path is `/`. """ pulumi.set(__self__, "port", port) if path is not None: pulumi.set(__self__, "path", path) @property @pulumi.getter def port(self) -> pulumi.Input[int]: """ Port for incoming traffic. """ return pulumi.get(self, "port") @port.setter def port(self, value: pulumi.Input[int]): pulumi.set(self, "port", value) @property @pulumi.getter def path(self) -> Optional[pulumi.Input[str]]: """ URL path to set for health checking requests for every target in the target group. For example `/ping`. The default path is `/`. """ return pulumi.get(self, "path") @path.setter def path(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "path", value) @pulumi.input_type class LbNetworkLoadBalancerAttachedTargetGroupHealthcheckTcpOptionsArgs: def __init__(__self__, *, port: pulumi.Input[int]): """ :param pulumi.Input[int] port: Port for incoming traffic. """ pulumi.set(__self__, "port", port) @property @pulumi.getter def port(self) -> pulumi.Input[int]: """ Port for incoming traffic. """ return pulumi.get(self, "port") @port.setter def port(self, value: pulumi.Input[int]): pulumi.set(self, "port", value) @pulumi.input_type class LbNetworkLoadBalancerListenerArgs: def __init__(__self__, *, name: pulumi.Input[str], port: pulumi.Input[int], external_address_spec: Optional[pulumi.Input['LbNetworkLoadBalancerListenerExternalAddressSpecArgs']] = None, internal_address_spec: Optional[pulumi.Input['LbNetworkLoadBalancerListenerInternalAddressSpecArgs']] = None, protocol: Optional[pulumi.Input[str]] = None, target_port: Optional[pulumi.Input[int]] = None): """ :param pulumi.Input[str] name: Name of the listener. The name must be unique for each listener on a single load balancer. :param pulumi.Input[int] port: Port for incoming traffic. :param pulumi.Input['LbNetworkLoadBalancerListenerExternalAddressSpecArgs'] external_address_spec: External IP address specification. The structure is documented below. :param pulumi.Input['LbNetworkLoadBalancerListenerInternalAddressSpecArgs'] internal_address_spec: Internal IP address specification. The structure is documented below. :param pulumi.Input[str] protocol: Protocol for incoming traffic. TCP or UDP and the default is TCP. :param pulumi.Input[int] target_port: Port of a target. The default is the same as listener's port. """ pulumi.set(__self__, "name", name) pulumi.set(__self__, "port", port) if external_address_spec is not None: pulumi.set(__self__, "external_address_spec", external_address_spec) if internal_address_spec is not None: pulumi.set(__self__, "internal_address_spec", internal_address_spec) if protocol is not None: pulumi.set(__self__, "protocol", protocol) if target_port is not None: pulumi.set(__self__, "target_port", target_port) @property @pulumi.getter def name(self) -> pulumi.Input[str]: """ Name of the listener. The name must be unique for each listener on a single load balancer. """ return pulumi.get(self, "name") @name.setter def name(self, value: pulumi.Input[str]): pulumi.set(self, "name", value) @property @pulumi.getter def port(self) -> pulumi.Input[int]: """ Port for incoming traffic. """ return pulumi.get(self, "port") @port.setter def port(self, value: pulumi.Input[int]): pulumi.set(self, "port", value) @property @pulumi.getter(name="externalAddressSpec") def external_address_spec(self) -> Optional[pulumi.Input['LbNetworkLoadBalancerListenerExternalAddressSpecArgs']]: """ External IP address specification. The structure is documented below. """ return pulumi.get(self, "external_address_spec") @external_address_spec.setter def external_address_spec(self, value: Optional[pulumi.Input['LbNetworkLoadBalancerListenerExternalAddressSpecArgs']]): pulumi.set(self, "external_address_spec", value) @property @pulumi.getter(name="internalAddressSpec") def internal_address_spec(self) -> Optional[pulumi.Input['LbNetworkLoadBalancerListenerInternalAddressSpecArgs']]: """ Internal IP address specification. The structure is documented below. """ return pulumi.get(self, "internal_address_spec") @internal_address_spec.setter def internal_address_spec(self, value: Optional[pulumi.Input['LbNetworkLoadBalancerListenerInternalAddressSpecArgs']]): pulumi.set(self, "internal_address_spec", value) @property @pulumi.getter def protocol(self) -> Optional[pulumi.Input[str]]: """ Protocol for incoming traffic. TCP or UDP and the default is TCP. """ return pulumi.get(self, "protocol") @protocol.setter def protocol(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "protocol", value) @property @pulumi.getter(name="targetPort") def target_port(self) -> Optional[pulumi.Input[int]]: """ Port of a target. The default is the same as listener's port. """ return pulumi.get(self, "target_port") @target_port.setter def target_port(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "target_port", value) @pulumi.input_type class LbNetworkLoadBalancerListenerExternalAddressSpecArgs: def __init__(__self__, *, address: Optional[pulumi.Input[str]] = None, ip_version: Optional[pulumi.Input[str]] = None): """ :param pulumi.Input[str] address: Internal IP address for a listener. Must belong to the subnet that is referenced in subnet_id. IP address will be allocated if it wasn't been set. :param pulumi.Input[str] ip_version: IP version of the internal addresses that the load balancer works with. Must be one of ipv4 or ipv6. The default is ipv4. """ if address is not None: pulumi.set(__self__, "address", address) if ip_version is not None: pulumi.set(__self__, "ip_version", ip_version) @property @pulumi.getter def address(self) -> Optional[pulumi.Input[str]]: """ Internal IP address for a listener. Must belong to the subnet that is referenced in subnet_id. IP address will be allocated if it wasn't been set. """ return pulumi.get(self, "address") @address.setter def address(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "address", value) @property @pulumi.getter(name="ipVersion") def ip_version(self) -> Optional[pulumi.Input[str]]: """ IP version of the internal addresses that the load balancer works with. Must be one of ipv4 or ipv6. The default is ipv4. """ return pulumi.get(self, "ip_version") @ip_version.setter def ip_version(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "ip_version", value) @pulumi.input_type class LbNetworkLoadBalancerListenerInternalAddressSpecArgs: def __init__(__self__, *, subnet_id: pulumi.Input[str], address: Optional[pulumi.Input[str]] = None, ip_version: Optional[pulumi.Input[str]] = None): """ :param pulumi.Input[str] subnet_id: ID of the subnet to which the internal IP address belongs. :param pulumi.Input[str] address: Internal IP address for a listener. Must belong to the subnet that is referenced in subnet_id. IP address will be allocated if it wasn't been set. :param pulumi.Input[str] ip_version: IP version of the internal addresses that the load balancer works with. Must be one of ipv4 or ipv6. The default is ipv4. """ pulumi.set(__self__, "subnet_id", subnet_id) if address is not None: pulumi.set(__self__, "address", address) if ip_version is not None: pulumi.set(__self__, "ip_version", ip_version) @property @pulumi.getter(name="subnetId") def subnet_id(self) -> pulumi.Input[str]: """ ID of the subnet to which the internal IP address belongs. """ return pulumi.get(self, "subnet_id") @subnet_id.setter def subnet_id(self, value: pulumi.Input[str]): pulumi.set(self, "subnet_id", value) @property @pulumi.getter def address(self) -> Optional[pulumi.Input[str]]: """ Internal IP address for a listener. Must belong to the subnet that is referenced in subnet_id. IP address will be allocated if it wasn't been set. """ return pulumi.get(self, "address") @address.setter def address(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "address", value) @property @pulumi.getter(name="ipVersion") def ip_version(self) -> Optional[pulumi.Input[str]]: """ IP version of the internal addresses that the load balancer works with. Must be one of ipv4 or ipv6. The default is ipv4. """ return pulumi.get(self, "ip_version") @ip_version.setter def ip_version(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "ip_version", value) @pulumi.input_type class LbTargetGroupTargetArgs: def __init__(__self__, *, address: pulumi.Input[str], subnet_id: pulumi.Input[str]): """ :param pulumi.Input[str] address: IP address of the target. :param pulumi.Input[str] subnet_id: ID of the subnet that targets are connected to. All targets in the target group must be connected to the same subnet within a single availability zone. """ pulumi.set(__self__, "address", address) pulumi.set(__self__, "subnet_id", subnet_id) @property @pulumi.getter def address(self) -> pulumi.Input[str]: """ IP address of the target. """ return pulumi.get(self, "address") @address.setter def address(self, value: pulumi.Input[str]): pulumi.set(self, "address", value) @property @pulumi.getter(name="subnetId") def subnet_id(self) -> pulumi.Input[str]: """ ID of the subnet that targets are connected to. All targets in the target group must be connected to the same subnet within a single availability zone. """ return pulumi.get(self, "subnet_id") @subnet_id.setter def subnet_id(self, value: pulumi.Input[str]): pulumi.set(self, "subnet_id", value) @pulumi.input_type class MdbClickhouseClusterAccessArgs: def __init__(__self__, *, data_lens: Optional[pulumi.Input[bool]] = None, metrika: Optional[pulumi.Input[bool]] = None, serverless: Optional[pulumi.Input[bool]] = None, web_sql: Optional[pulumi.Input[bool]] = None): """ :param pulumi.Input[bool] data_lens: Allow access for DataLens. Can be either `true` or `false`. :param pulumi.Input[bool] metrika: Allow access for Yandex.Metrika. Can be either `true` or `false`. :param pulumi.Input[bool] serverless: Allow access for Serverless. Can be either `true` or `false`. :param pulumi.Input[bool] web_sql: Allow access for Web SQL. Can be either `true` or `false`. """ if data_lens is not None: pulumi.set(__self__, "data_lens", data_lens) if metrika is not None: pulumi.set(__self__, "metrika", metrika) if serverless is not None: pulumi.set(__self__, "serverless", serverless) if web_sql is not None: pulumi.set(__self__, "web_sql", web_sql) @property @pulumi.getter(name="dataLens") def data_lens(self) -> Optional[pulumi.Input[bool]]: """ Allow access for DataLens. Can be either `true` or `false`. """ return pulumi.get(self, "data_lens") @data_lens.setter def data_lens(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "data_lens", value) @property @pulumi.getter def metrika(self) -> Optional[pulumi.Input[bool]]: """ Allow access for Yandex.Metrika. Can be either `true` or `false`. """ return pulumi.get(self, "metrika") @metrika.setter def metrika(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "metrika", value) @property @pulumi.getter def serverless(self) -> Optional[pulumi.Input[bool]]: """ Allow access for Serverless. Can be either `true` or `false`. """ return pulumi.get(self, "serverless") @serverless.setter def serverless(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "serverless", value) @property @pulumi.getter(name="webSql") def web_sql(self) -> Optional[pulumi.Input[bool]]: """ Allow access for Web SQL. Can be either `true` or `false`. """ return pulumi.get(self, "web_sql") @web_sql.setter def web_sql(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "web_sql", value) @pulumi.input_type class MdbClickhouseClusterBackupWindowStartArgs: def __init__(__self__, *, hours: Optional[pulumi.Input[int]] = None, minutes: Optional[pulumi.Input[int]] = None): """ :param pulumi.Input[int] hours: The hour at which backup will be started. :param pulumi.Input[int] minutes: The minute at which backup will be started. """ if hours is not None: pulumi.set(__self__, "hours", hours) if minutes is not None: pulumi.set(__self__, "minutes", minutes) @property @pulumi.getter def hours(self) -> Optional[pulumi.Input[int]]: """ The hour at which backup will be started. """ return pulumi.get(self, "hours") @hours.setter def hours(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "hours", value) @property @pulumi.getter def minutes(self) -> Optional[pulumi.Input[int]]: """ The minute at which backup will be started. """ return pulumi.get(self, "minutes") @minutes.setter def minutes(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "minutes", value) @pulumi.input_type class MdbClickhouseClusterClickhouseArgs: def __init__(__self__, *, resources: pulumi.Input['MdbClickhouseClusterClickhouseResourcesArgs'], config: Optional[pulumi.Input['MdbClickhouseClusterClickhouseConfigArgs']] = None): """ :param pulumi.Input['MdbClickhouseClusterClickhouseResourcesArgs'] resources: Resources allocated to hosts of the ZooKeeper subcluster. The structure is documented below. :param pulumi.Input['MdbClickhouseClusterClickhouseConfigArgs'] config: Main ClickHouse cluster configuration. """ pulumi.set(__self__, "resources", resources) if config is not None: pulumi.set(__self__, "config", config) @property @pulumi.getter def resources(self) -> pulumi.Input['MdbClickhouseClusterClickhouseResourcesArgs']: """ Resources allocated to hosts of the ZooKeeper subcluster. The structure is documented below. """ return pulumi.get(self, "resources") @resources.setter def resources(self, value: pulumi.Input['MdbClickhouseClusterClickhouseResourcesArgs']): pulumi.set(self, "resources", value) @property @pulumi.getter def config(self) -> Optional[pulumi.Input['MdbClickhouseClusterClickhouseConfigArgs']]: """ Main ClickHouse cluster configuration. """ return pulumi.get(self, "config") @config.setter def config(self, value: Optional[pulumi.Input['MdbClickhouseClusterClickhouseConfigArgs']]): pulumi.set(self, "config", value) @pulumi.input_type class MdbClickhouseClusterClickhouseConfigArgs: def __init__(__self__, *, background_pool_size: Optional[pulumi.Input[int]] = None, background_schedule_pool_size: Optional[pulumi.Input[int]] = None, compressions: Optional[pulumi.Input[Sequence[pulumi.Input['MdbClickhouseClusterClickhouseConfigCompressionArgs']]]] = None, geobase_uri: Optional[pulumi.Input[str]] = None, graphite_rollups: Optional[pulumi.Input[Sequence[pulumi.Input['MdbClickhouseClusterClickhouseConfigGraphiteRollupArgs']]]] = None, kafka: Optional[pulumi.Input['MdbClickhouseClusterClickhouseConfigKafkaArgs']] = None, kafka_topics: Optional[pulumi.Input[Sequence[pulumi.Input['MdbClickhouseClusterClickhouseConfigKafkaTopicArgs']]]] = None, keep_alive_timeout: Optional[pulumi.Input[int]] = None, log_level: Optional[pulumi.Input[str]] = None, mark_cache_size: Optional[pulumi.Input[int]] = None, max_concurrent_queries: Optional[pulumi.Input[int]] = None, max_connections: Optional[pulumi.Input[int]] = None, max_partition_size_to_drop: Optional[pulumi.Input[int]] = None, max_table_size_to_drop: Optional[pulumi.Input[int]] = None, merge_tree: Optional[pulumi.Input['MdbClickhouseClusterClickhouseConfigMergeTreeArgs']] = None, metric_log_enabled: Optional[pulumi.Input[bool]] = None, metric_log_retention_size: Optional[pulumi.Input[int]] = None, metric_log_retention_time: Optional[pulumi.Input[int]] = None, part_log_retention_size: Optional[pulumi.Input[int]] = None, part_log_retention_time: Optional[pulumi.Input[int]] = None, query_log_retention_size: Optional[pulumi.Input[int]] = None, query_log_retention_time: Optional[pulumi.Input[int]] = None, query_thread_log_enabled: Optional[pulumi.Input[bool]] = None, query_thread_log_retention_size: Optional[pulumi.Input[int]] = None, query_thread_log_retention_time: Optional[pulumi.Input[int]] = None, rabbitmq: Optional[pulumi.Input['MdbClickhouseClusterClickhouseConfigRabbitmqArgs']] = None, text_log_enabled: Optional[pulumi.Input[bool]] = None, text_log_level: Optional[pulumi.Input[str]] = None, text_log_retention_size: Optional[pulumi.Input[int]] = None, text_log_retention_time: Optional[pulumi.Input[int]] = None, timezone: Optional[pulumi.Input[str]] = None, trace_log_enabled: Optional[pulumi.Input[bool]] = None, trace_log_retention_size: Optional[pulumi.Input[int]] = None, trace_log_retention_time: Optional[pulumi.Input[int]] = None, uncompressed_cache_size: Optional[pulumi.Input[int]] = None): """ :param pulumi.Input[Sequence[pulumi.Input['MdbClickhouseClusterClickhouseConfigCompressionArgs']]] compressions: Data compression configuration. The structure is documented below. :param pulumi.Input[Sequence[pulumi.Input['MdbClickhouseClusterClickhouseConfigGraphiteRollupArgs']]] graphite_rollups: Graphite rollup configuration. The structure is documented below. :param pulumi.Input['MdbClickhouseClusterClickhouseConfigKafkaArgs'] kafka: Kafka connection configuration. The structure is documented below. :param pulumi.Input[Sequence[pulumi.Input['MdbClickhouseClusterClickhouseConfigKafkaTopicArgs']]] kafka_topics: Kafka topic connection configuration. The structure is documented below. :param pulumi.Input['MdbClickhouseClusterClickhouseConfigMergeTreeArgs'] merge_tree: MergeTree engine configuration. The structure is documented below. :param pulumi.Input['MdbClickhouseClusterClickhouseConfigRabbitmqArgs'] rabbitmq: RabbitMQ connection configuration. The structure is documented below. """ if background_pool_size is not None: pulumi.set(__self__, "background_pool_size", background_pool_size) if background_schedule_pool_size is not None: pulumi.set(__self__, "background_schedule_pool_size", background_schedule_pool_size) if compressions is not None: pulumi.set(__self__, "compressions", compressions) if geobase_uri is not None: pulumi.set(__self__, "geobase_uri", geobase_uri) if graphite_rollups is not None: pulumi.set(__self__, "graphite_rollups", graphite_rollups) if kafka is not None: pulumi.set(__self__, "kafka", kafka) if kafka_topics is not None: pulumi.set(__self__, "kafka_topics", kafka_topics) if keep_alive_timeout is not None: pulumi.set(__self__, "keep_alive_timeout", keep_alive_timeout) if log_level is not None: pulumi.set(__self__, "log_level", log_level) if mark_cache_size is not None: pulumi.set(__self__, "mark_cache_size", mark_cache_size) if max_concurrent_queries is not None: pulumi.set(__self__, "max_concurrent_queries", max_concurrent_queries) if max_connections is not None: pulumi.set(__self__, "max_connections", max_connections) if max_partition_size_to_drop is not None: pulumi.set(__self__, "max_partition_size_to_drop", max_partition_size_to_drop) if max_table_size_to_drop is not None: pulumi.set(__self__, "max_table_size_to_drop", max_table_size_to_drop) if merge_tree is not None: pulumi.set(__self__, "merge_tree", merge_tree) if metric_log_enabled is not None: pulumi.set(__self__, "metric_log_enabled", metric_log_enabled) if metric_log_retention_size is not None: pulumi.set(__self__, "metric_log_retention_size", metric_log_retention_size) if metric_log_retention_time is not None: pulumi.set(__self__, "metric_log_retention_time", metric_log_retention_time) if part_log_retention_size is not None: pulumi.set(__self__, "part_log_retention_size", part_log_retention_size) if part_log_retention_time is not None: pulumi.set(__self__, "part_log_retention_time", part_log_retention_time) if query_log_retention_size is not None: pulumi.set(__self__, "query_log_retention_size", query_log_retention_size) if query_log_retention_time is not None: pulumi.set(__self__, "query_log_retention_time", query_log_retention_time) if query_thread_log_enabled is not None: pulumi.set(__self__, "query_thread_log_enabled", query_thread_log_enabled) if query_thread_log_retention_size is not None: pulumi.set(__self__, "query_thread_log_retention_size", query_thread_log_retention_size) if query_thread_log_retention_time is not None: pulumi.set(__self__, "query_thread_log_retention_time", query_thread_log_retention_time) if rabbitmq is not None: pulumi.set(__self__, "rabbitmq", rabbitmq) if text_log_enabled is not None: pulumi.set(__self__, "text_log_enabled", text_log_enabled) if text_log_level is not None: pulumi.set(__self__, "text_log_level", text_log_level) if text_log_retention_size is not None: pulumi.set(__self__, "text_log_retention_size", text_log_retention_size) if text_log_retention_time is not None: pulumi.set(__self__, "text_log_retention_time", text_log_retention_time) if timezone is not None: pulumi.set(__self__, "timezone", timezone) if trace_log_enabled is not None: pulumi.set(__self__, "trace_log_enabled", trace_log_enabled) if trace_log_retention_size is not None: pulumi.set(__self__, "trace_log_retention_size", trace_log_retention_size) if trace_log_retention_time is not None: pulumi.set(__self__, "trace_log_retention_time", trace_log_retention_time) if uncompressed_cache_size is not None: pulumi.set(__self__, "uncompressed_cache_size", uncompressed_cache_size) @property @pulumi.getter(name="backgroundPoolSize") def background_pool_size(self) -> Optional[pulumi.Input[int]]: return pulumi.get(self, "background_pool_size") @background_pool_size.setter def background_pool_size(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "background_pool_size", value) @property @pulumi.getter(name="backgroundSchedulePoolSize") def background_schedule_pool_size(self) -> Optional[pulumi.Input[int]]: return pulumi.get(self, "background_schedule_pool_size") @background_schedule_pool_size.setter def background_schedule_pool_size(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "background_schedule_pool_size", value) @property @pulumi.getter def compressions(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['MdbClickhouseClusterClickhouseConfigCompressionArgs']]]]: """ Data compression configuration. The structure is documented below. """ return pulumi.get(self, "compressions") @compressions.setter def compressions(self, value: Optional[pulumi.Input[Sequence[pulumi.Input['MdbClickhouseClusterClickhouseConfigCompressionArgs']]]]): pulumi.set(self, "compressions", value) @property @pulumi.getter(name="geobaseUri") def geobase_uri(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "geobase_uri") @geobase_uri.setter def geobase_uri(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "geobase_uri", value) @property @pulumi.getter(name="graphiteRollups") def graphite_rollups(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['MdbClickhouseClusterClickhouseConfigGraphiteRollupArgs']]]]: """ Graphite rollup configuration. The structure is documented below. """ return pulumi.get(self, "graphite_rollups") @graphite_rollups.setter def graphite_rollups(self, value: Optional[pulumi.Input[Sequence[pulumi.Input['MdbClickhouseClusterClickhouseConfigGraphiteRollupArgs']]]]): pulumi.set(self, "graphite_rollups", value) @property @pulumi.getter def kafka(self) -> Optional[pulumi.Input['MdbClickhouseClusterClickhouseConfigKafkaArgs']]: """ Kafka connection configuration. The structure is documented below. """ return pulumi.get(self, "kafka") @kafka.setter def kafka(self, value: Optional[pulumi.Input['MdbClickhouseClusterClickhouseConfigKafkaArgs']]): pulumi.set(self, "kafka", value) @property @pulumi.getter(name="kafkaTopics") def kafka_topics(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['MdbClickhouseClusterClickhouseConfigKafkaTopicArgs']]]]: """ Kafka topic connection configuration. The structure is documented below. """ return pulumi.get(self, "kafka_topics") @kafka_topics.setter def kafka_topics(self, value: Optional[pulumi.Input[Sequence[pulumi.Input['MdbClickhouseClusterClickhouseConfigKafkaTopicArgs']]]]): pulumi.set(self, "kafka_topics", value) @property @pulumi.getter(name="keepAliveTimeout") def keep_alive_timeout(self) -> Optional[pulumi.Input[int]]: return pulumi.get(self, "keep_alive_timeout") @keep_alive_timeout.setter def keep_alive_timeout(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "keep_alive_timeout", value) @property @pulumi.getter(name="logLevel") def log_level(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "log_level") @log_level.setter def log_level(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "log_level", value) @property @pulumi.getter(name="markCacheSize") def mark_cache_size(self) -> Optional[pulumi.Input[int]]: return pulumi.get(self, "mark_cache_size") @mark_cache_size.setter def mark_cache_size(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "mark_cache_size", value) @property @pulumi.getter(name="maxConcurrentQueries") def max_concurrent_queries(self) -> Optional[pulumi.Input[int]]: return pulumi.get(self, "max_concurrent_queries") @max_concurrent_queries.setter def max_concurrent_queries(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "max_concurrent_queries", value) @property @pulumi.getter(name="maxConnections") def max_connections(self) -> Optional[pulumi.Input[int]]: return pulumi.get(self, "max_connections") @max_connections.setter def max_connections(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "max_connections", value) @property @pulumi.getter(name="maxPartitionSizeToDrop") def max_partition_size_to_drop(self) -> Optional[pulumi.Input[int]]: return pulumi.get(self, "max_partition_size_to_drop") @max_partition_size_to_drop.setter def max_partition_size_to_drop(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "max_partition_size_to_drop", value) @property @pulumi.getter(name="maxTableSizeToDrop") def max_table_size_to_drop(self) -> Optional[pulumi.Input[int]]: return pulumi.get(self, "max_table_size_to_drop") @max_table_size_to_drop.setter def max_table_size_to_drop(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "max_table_size_to_drop", value) @property @pulumi.getter(name="mergeTree") def merge_tree(self) -> Optional[pulumi.Input['MdbClickhouseClusterClickhouseConfigMergeTreeArgs']]: """ MergeTree engine configuration. The structure is documented below. """ return pulumi.get(self, "merge_tree") @merge_tree.setter def merge_tree(self, value: Optional[pulumi.Input['MdbClickhouseClusterClickhouseConfigMergeTreeArgs']]): pulumi.set(self, "merge_tree", value) @property @pulumi.getter(name="metricLogEnabled") def metric_log_enabled(self) -> Optional[pulumi.Input[bool]]: return pulumi.get(self, "metric_log_enabled") @metric_log_enabled.setter def metric_log_enabled(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "metric_log_enabled", value) @property @pulumi.getter(name="metricLogRetentionSize") def metric_log_retention_size(self) -> Optional[pulumi.Input[int]]: return pulumi.get(self, "metric_log_retention_size") @metric_log_retention_size.setter def metric_log_retention_size(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "metric_log_retention_size", value) @property @pulumi.getter(name="metricLogRetentionTime") def metric_log_retention_time(self) -> Optional[pulumi.Input[int]]: return pulumi.get(self, "metric_log_retention_time") @metric_log_retention_time.setter def metric_log_retention_time(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "metric_log_retention_time", value) @property @pulumi.getter(name="partLogRetentionSize") def part_log_retention_size(self) -> Optional[pulumi.Input[int]]: return pulumi.get(self, "part_log_retention_size") @part_log_retention_size.setter def part_log_retention_size(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "part_log_retention_size", value) @property @pulumi.getter(name="partLogRetentionTime") def part_log_retention_time(self) -> Optional[pulumi.Input[int]]: return pulumi.get(self, "part_log_retention_time") @part_log_retention_time.setter def part_log_retention_time(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "part_log_retention_time", value) @property @pulumi.getter(name="queryLogRetentionSize") def query_log_retention_size(self) -> Optional[pulumi.Input[int]]: return pulumi.get(self, "query_log_retention_size") @query_log_retention_size.setter def query_log_retention_size(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "query_log_retention_size", value) @property @pulumi.getter(name="queryLogRetentionTime") def query_log_retention_time(self) -> Optional[pulumi.Input[int]]: return pulumi.get(self, "query_log_retention_time") @query_log_retention_time.setter def query_log_retention_time(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "query_log_retention_time", value) @property @pulumi.getter(name="queryThreadLogEnabled") def query_thread_log_enabled(self) -> Optional[pulumi.Input[bool]]: return pulumi.get(self, "query_thread_log_enabled") @query_thread_log_enabled.setter def query_thread_log_enabled(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "query_thread_log_enabled", value) @property @pulumi.getter(name="queryThreadLogRetentionSize") def query_thread_log_retention_size(self) -> Optional[pulumi.Input[int]]: return pulumi.get(self, "query_thread_log_retention_size") @query_thread_log_retention_size.setter def query_thread_log_retention_size(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "query_thread_log_retention_size", value) @property @pulumi.getter(name="queryThreadLogRetentionTime") def query_thread_log_retention_time(self) -> Optional[pulumi.Input[int]]: return pulumi.get(self, "query_thread_log_retention_time") @query_thread_log_retention_time.setter def query_thread_log_retention_time(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "query_thread_log_retention_time", value) @property @pulumi.getter def rabbitmq(self) -> Optional[pulumi.Input['MdbClickhouseClusterClickhouseConfigRabbitmqArgs']]: """ RabbitMQ connection configuration. The structure is documented below. """ return pulumi.get(self, "rabbitmq") @rabbitmq.setter def rabbitmq(self, value: Optional[pulumi.Input['MdbClickhouseClusterClickhouseConfigRabbitmqArgs']]): pulumi.set(self, "rabbitmq", value) @property @pulumi.getter(name="textLogEnabled") def text_log_enabled(self) -> Optional[pulumi.Input[bool]]: return pulumi.get(self, "text_log_enabled") @text_log_enabled.setter def text_log_enabled(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "text_log_enabled", value) @property @pulumi.getter(name="textLogLevel") def text_log_level(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "text_log_level") @text_log_level.setter def text_log_level(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "text_log_level", value) @property @pulumi.getter(name="textLogRetentionSize") def text_log_retention_size(self) -> Optional[pulumi.Input[int]]: return pulumi.get(self, "text_log_retention_size") @text_log_retention_size.setter def text_log_retention_size(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "text_log_retention_size", value) @property @pulumi.getter(name="textLogRetentionTime") def text_log_retention_time(self) -> Optional[pulumi.Input[int]]: return pulumi.get(self, "text_log_retention_time") @text_log_retention_time.setter def text_log_retention_time(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "text_log_retention_time", value) @property @pulumi.getter def timezone(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "timezone") @timezone.setter def timezone(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "timezone", value) @property @pulumi.getter(name="traceLogEnabled") def trace_log_enabled(self) -> Optional[pulumi.Input[bool]]: return pulumi.get(self, "trace_log_enabled") @trace_log_enabled.setter def trace_log_enabled(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "trace_log_enabled", value) @property @pulumi.getter(name="traceLogRetentionSize") def trace_log_retention_size(self) -> Optional[pulumi.Input[int]]: return pulumi.get(self, "trace_log_retention_size") @trace_log_retention_size.setter def trace_log_retention_size(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "trace_log_retention_size", value) @property @pulumi.getter(name="traceLogRetentionTime") def trace_log_retention_time(self) -> Optional[pulumi.Input[int]]: return pulumi.get(self, "trace_log_retention_time") @trace_log_retention_time.setter def trace_log_retention_time(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "trace_log_retention_time", value) @property @pulumi.getter(name="uncompressedCacheSize") def uncompressed_cache_size(self) -> Optional[pulumi.Input[int]]: return pulumi.get(self, "uncompressed_cache_size") @uncompressed_cache_size.setter def uncompressed_cache_size(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "uncompressed_cache_size", value) @pulumi.input_type class MdbClickhouseClusterClickhouseConfigCompressionArgs: def __init__(__self__, *, method: pulumi.Input[str], min_part_size: pulumi.Input[int], min_part_size_ratio: pulumi.Input[float]): """ :param pulumi.Input[str] method: Method: Compression method. Two methods are available: LZ4 and zstd. :param pulumi.Input[int] min_part_size: Min part size: Minimum size (in bytes) of a data part in a table. ClickHouse only applies the rule to tables with data parts greater than or equal to the Min part size value. :param pulumi.Input[float] min_part_size_ratio: Min part size ratio: Minimum table part size to total table size ratio. ClickHouse only applies the rule to tables in which this ratio is greater than or equal to the Min part size ratio value. """ pulumi.set(__self__, "method", method) pulumi.set(__self__, "min_part_size", min_part_size) pulumi.set(__self__, "min_part_size_ratio", min_part_size_ratio) @property @pulumi.getter def method(self) -> pulumi.Input[str]: """ Method: Compression method. Two methods are available: LZ4 and zstd. """ return pulumi.get(self, "method") @method.setter def method(self, value: pulumi.Input[str]): pulumi.set(self, "method", value) @property @pulumi.getter(name="minPartSize") def min_part_size(self) -> pulumi.Input[int]: """ Min part size: Minimum size (in bytes) of a data part in a table. ClickHouse only applies the rule to tables with data parts greater than or equal to the Min part size value. """ return pulumi.get(self, "min_part_size") @min_part_size.setter def min_part_size(self, value: pulumi.Input[int]): pulumi.set(self, "min_part_size", value) @property @pulumi.getter(name="minPartSizeRatio") def min_part_size_ratio(self) -> pulumi.Input[float]: """ Min part size ratio: Minimum table part size to total table size ratio. ClickHouse only applies the rule to tables in which this ratio is greater than or equal to the Min part size ratio value. """ return pulumi.get(self, "min_part_size_ratio") @min_part_size_ratio.setter def min_part_size_ratio(self, value: pulumi.Input[float]): pulumi.set(self, "min_part_size_ratio", value) @pulumi.input_type class MdbClickhouseClusterClickhouseConfigGraphiteRollupArgs: def __init__(__self__, *, name: pulumi.Input[str], patterns: Optional[pulumi.Input[Sequence[pulumi.Input['MdbClickhouseClusterClickhouseConfigGraphiteRollupPatternArgs']]]] = None): """ :param pulumi.Input[str] name: Graphite rollup configuration name. :param pulumi.Input[Sequence[pulumi.Input['MdbClickhouseClusterClickhouseConfigGraphiteRollupPatternArgs']]] patterns: Set of thinning rules. """ pulumi.set(__self__, "name", name) if patterns is not None: pulumi.set(__self__, "patterns", patterns) @property @pulumi.getter def name(self) -> pulumi.Input[str]: """ Graphite rollup configuration name. """ return pulumi.get(self, "name") @name.setter def name(self, value: pulumi.Input[str]): pulumi.set(self, "name", value) @property @pulumi.getter def patterns(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['MdbClickhouseClusterClickhouseConfigGraphiteRollupPatternArgs']]]]: """ Set of thinning rules. """ return pulumi.get(self, "patterns") @patterns.setter def patterns(self, value: Optional[pulumi.Input[Sequence[pulumi.Input['MdbClickhouseClusterClickhouseConfigGraphiteRollupPatternArgs']]]]): pulumi.set(self, "patterns", value) @pulumi.input_type class MdbClickhouseClusterClickhouseConfigGraphiteRollupPatternArgs: def __init__(__self__, *, function: pulumi.Input[str], regexp: Optional[pulumi.Input[str]] = None, retentions: Optional[pulumi.Input[Sequence[pulumi.Input['MdbClickhouseClusterClickhouseConfigGraphiteRollupPatternRetentionArgs']]]] = None): """ :param pulumi.Input[str] function: Aggregation function name. :param pulumi.Input[str] regexp: Regular expression that the metric name must match. :param pulumi.Input[Sequence[pulumi.Input['MdbClickhouseClusterClickhouseConfigGraphiteRollupPatternRetentionArgs']]] retentions: Retain parameters. """ pulumi.set(__self__, "function", function) if regexp is not None: pulumi.set(__self__, "regexp", regexp) if retentions is not None: pulumi.set(__self__, "retentions", retentions) @property @pulumi.getter def function(self) -> pulumi.Input[str]: """ Aggregation function name. """ return pulumi.get(self, "function") @function.setter def function(self, value: pulumi.Input[str]): pulumi.set(self, "function", value) @property @pulumi.getter def regexp(self) -> Optional[pulumi.Input[str]]: """ Regular expression that the metric name must match. """ return pulumi.get(self, "regexp") @regexp.setter def regexp(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "regexp", value) @property @pulumi.getter def retentions(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['MdbClickhouseClusterClickhouseConfigGraphiteRollupPatternRetentionArgs']]]]: """ Retain parameters. """ return pulumi.get(self, "retentions") @retentions.setter def retentions(self, value: Optional[pulumi.Input[Sequence[pulumi.Input['MdbClickhouseClusterClickhouseConfigGraphiteRollupPatternRetentionArgs']]]]): pulumi.set(self, "retentions", value) @pulumi.input_type class MdbClickhouseClusterClickhouseConfigGraphiteRollupPatternRetentionArgs: def __init__(__self__, *, age: pulumi.Input[int], precision: pulumi.Input[int]): """ :param pulumi.Input[int] age: Minimum data age in seconds. :param pulumi.Input[int] precision: Accuracy of determining the age of the data in seconds. """ pulumi.set(__self__, "age", age) pulumi.set(__self__, "precision", precision) @property @pulumi.getter def age(self) -> pulumi.Input[int]: """ Minimum data age in seconds. """ return pulumi.get(self, "age") @age.setter def age(self, value: pulumi.Input[int]): pulumi.set(self, "age", value) @property @pulumi.getter def precision(self) -> pulumi.Input[int]: """ Accuracy of determining the age of the data in seconds. """ return pulumi.get(self, "precision") @precision.setter def precision(self, value: pulumi.Input[int]): pulumi.set(self, "precision", value) @pulumi.input_type class MdbClickhouseClusterClickhouseConfigKafkaArgs: def __init__(__self__, *, sasl_mechanism: Optional[pulumi.Input[str]] = None, sasl_password: Optional[pulumi.Input[str]] = None, sasl_username: Optional[pulumi.Input[str]] = None, security_protocol: Optional[pulumi.Input[str]] = None): """ :param pulumi.Input[str] sasl_mechanism: SASL mechanism used in kafka authentication. :param pulumi.Input[str] sasl_password: <PASSWORD> on kafka server. :param pulumi.Input[str] sasl_username: Username on kafka server. :param pulumi.Input[str] security_protocol: Security protocol used to connect to kafka server. """ if sasl_mechanism is not None: pulumi.set(__self__, "sasl_mechanism", sasl_mechanism) if sasl_password is not None: pulumi.set(__self__, "sasl_password", sasl_password) if sasl_username is not None: pulumi.set(__self__, "sasl_username", sasl_username) if security_protocol is not None: pulumi.set(__self__, "security_protocol", security_protocol) @property @pulumi.getter(name="saslMechanism") def sasl_mechanism(self) -> Optional[pulumi.Input[str]]: """ SASL mechanism used in kafka authentication. """ return pulumi.get(self, "sasl_mechanism") @sasl_mechanism.setter def sasl_mechanism(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "sasl_mechanism", value) @property @pulumi.getter(name="saslPassword") def sasl_password(self) -> Optional[pulumi.Input[str]]: """ User password on kafka server. """ return pulumi.get(self, "sasl_password") @sasl_password.setter def sasl_password(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "sasl_password", value) @property @pulumi.getter(name="saslUsername") def sasl_username(self) -> Optional[pulumi.Input[str]]: """ Username on kafka server. """ return pulumi.get(self, "sasl_username") @sasl_username.setter def sasl_username(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "sasl_username", value) @property @pulumi.getter(name="securityProtocol") def security_protocol(self) -> Optional[pulumi.Input[str]]: """ Security protocol used to connect to kafka server. """ return pulumi.get(self, "security_protocol") @security_protocol.setter def security_protocol(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "security_protocol", value) @pulumi.input_type class MdbClickhouseClusterClickhouseConfigKafkaTopicArgs: def __init__(__self__, *, name: pulumi.Input[str], settings: Optional[pulumi.Input['MdbClickhouseClusterClickhouseConfigKafkaTopicSettingsArgs']] = None): """ :param pulumi.Input[str] name: Graphite rollup configuration name. :param pulumi.Input['MdbClickhouseClusterClickhouseConfigKafkaTopicSettingsArgs'] settings: Kafka connection settngs sanem as `kafka` block. """ pulumi.set(__self__, "name", name) if settings is not None: pulumi.set(__self__, "settings", settings) @property @pulumi.getter def name(self) -> pulumi.Input[str]: """ Graphite rollup configuration name. """ return pulumi.get(self, "name") @name.setter def name(self, value: pulumi.Input[str]): pulumi.set(self, "name", value) @property @pulumi.getter def settings(self) -> Optional[pulumi.Input['MdbClickhouseClusterClickhouseConfigKafkaTopicSettingsArgs']]: """ Kafka connection settngs sanem as `kafka` block. """ return pulumi.get(self, "settings") @settings.setter def settings(self, value: Optional[pulumi.Input['MdbClickhouseClusterClickhouseConfigKafkaTopicSettingsArgs']]): pulumi.set(self, "settings", value) @pulumi.input_type class MdbClickhouseClusterClickhouseConfigKafkaTopicSettingsArgs: def __init__(__self__, *, sasl_mechanism: Optional[pulumi.Input[str]] = None, sasl_password: Optional[pulumi.Input[str]] = None, sasl_username: Optional[pulumi.Input[str]] = None, security_protocol: Optional[pulumi.Input[str]] = None): """ :param pulumi.Input[str] sasl_mechanism: SASL mechanism used in kafka authentication. :param pulumi.Input[str] sasl_password: <PASSWORD> on kafka server. :param pulumi.Input[str] sasl_username: Username on kafka server. :param pulumi.Input[str] security_protocol: Security protocol used to connect to kafka server. """ if sasl_mechanism is not None: pulumi.set(__self__, "sasl_mechanism", sasl_mechanism) if sasl_password is not None: pulumi.set(__self__, "sasl_password", sasl_password) if sasl_username is not None: pulumi.set(__self__, "sasl_username", sasl_username) if security_protocol is not None: pulumi.set(__self__, "security_protocol", security_protocol) @property @pulumi.getter(name="saslMechanism") def sasl_mechanism(self) -> Optional[pulumi.Input[str]]: """ SASL mechanism used in kafka authentication. """ return pulumi.get(self, "sasl_mechanism") @sasl_mechanism.setter def sasl_mechanism(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "sasl_mechanism", value) @property @pulumi.getter(name="saslPassword") def sasl_password(self) -> Optional[pulumi.Input[str]]: """ User password on kafka server. """ return pulumi.get(self, "sasl_password") @sasl_password.setter def sasl_password(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "sasl_password", value) @property @pulumi.getter(name="saslUsername") def sasl_username(self) -> Optional[pulumi.Input[str]]: """ Username on kafka server. """ return pulumi.get(self, "sasl_username") @sasl_username.setter def sasl_username(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "sasl_username", value) @property @pulumi.getter(name="securityProtocol") def security_protocol(self) -> Optional[pulumi.Input[str]]: """ Security protocol used to connect to kafka server. """ return pulumi.get(self, "security_protocol") @security_protocol.setter def security_protocol(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "security_protocol", value) @pulumi.input_type class MdbClickhouseClusterClickhouseConfigMergeTreeArgs: def __init__(__self__, *, max_bytes_to_merge_at_min_space_in_pool: Optional[pulumi.Input[int]] = None, max_replicated_merges_in_queue: Optional[pulumi.Input[int]] = None, number_of_free_entries_in_pool_to_lower_max_size_of_merge: Optional[pulumi.Input[int]] = None, parts_to_delay_insert: Optional[pulumi.Input[int]] = None, parts_to_throw_insert: Optional[pulumi.Input[int]] = None, replicated_deduplication_window: Optional[pulumi.Input[int]] = None, replicated_deduplication_window_seconds: Optional[pulumi.Input[int]] = None): """ :param pulumi.Input[int] max_bytes_to_merge_at_min_space_in_pool: Max bytes to merge at min space in pool: Maximum total size of a data part to merge when the number of free threads in the background pool is minimum. :param pulumi.Input[int] max_replicated_merges_in_queue: Max replicated merges in queue: Maximum number of merge tasks that can be in the ReplicatedMergeTree queue at the same time. :param pulumi.Input[int] number_of_free_entries_in_pool_to_lower_max_size_of_merge: Number of free entries in pool to lower max size of merge: Threshold value of free entries in the pool. If the number of entries in the pool falls below this value, ClickHouse reduces the maximum size of a data part to merge. This helps handle small merges faster, rather than filling the pool with lengthy merges. :param pulumi.Input[int] parts_to_delay_insert: Parts to delay insert: Number of active data parts in a table, on exceeding which ClickHouse starts artificially reduce the rate of inserting data into the table. :param pulumi.Input[int] parts_to_throw_insert: Parts to throw insert: Threshold value of active data parts in a table, on exceeding which ClickHouse throws the 'Too many parts ...' exception. :param pulumi.Input[int] replicated_deduplication_window: Replicated deduplication window: Number of recent hash blocks that ZooKeeper will store (the old ones will be deleted). :param pulumi.Input[int] replicated_deduplication_window_seconds: Replicated deduplication window seconds: Time during which ZooKeeper stores the hash blocks (the old ones wil be deleted). """ if max_bytes_to_merge_at_min_space_in_pool is not None: pulumi.set(__self__, "max_bytes_to_merge_at_min_space_in_pool", max_bytes_to_merge_at_min_space_in_pool) if max_replicated_merges_in_queue is not None: pulumi.set(__self__, "max_replicated_merges_in_queue", max_replicated_merges_in_queue) if number_of_free_entries_in_pool_to_lower_max_size_of_merge is not None: pulumi.set(__self__, "number_of_free_entries_in_pool_to_lower_max_size_of_merge", number_of_free_entries_in_pool_to_lower_max_size_of_merge) if parts_to_delay_insert is not None: pulumi.set(__self__, "parts_to_delay_insert", parts_to_delay_insert) if parts_to_throw_insert is not None: pulumi.set(__self__, "parts_to_throw_insert", parts_to_throw_insert) if replicated_deduplication_window is not None: pulumi.set(__self__, "replicated_deduplication_window", replicated_deduplication_window) if replicated_deduplication_window_seconds is not None: pulumi.set(__self__, "replicated_deduplication_window_seconds", replicated_deduplication_window_seconds) @property @pulumi.getter(name="maxBytesToMergeAtMinSpaceInPool") def max_bytes_to_merge_at_min_space_in_pool(self) -> Optional[pulumi.Input[int]]: """ Max bytes to merge at min space in pool: Maximum total size of a data part to merge when the number of free threads in the background pool is minimum. """ return pulumi.get(self, "max_bytes_to_merge_at_min_space_in_pool") @max_bytes_to_merge_at_min_space_in_pool.setter def max_bytes_to_merge_at_min_space_in_pool(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "max_bytes_to_merge_at_min_space_in_pool", value) @property @pulumi.getter(name="maxReplicatedMergesInQueue") def max_replicated_merges_in_queue(self) -> Optional[pulumi.Input[int]]: """ Max replicated merges in queue: Maximum number of merge tasks that can be in the ReplicatedMergeTree queue at the same time. """ return pulumi.get(self, "max_replicated_merges_in_queue") @max_replicated_merges_in_queue.setter def max_replicated_merges_in_queue(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "max_replicated_merges_in_queue", value) @property @pulumi.getter(name="numberOfFreeEntriesInPoolToLowerMaxSizeOfMerge") def number_of_free_entries_in_pool_to_lower_max_size_of_merge(self) -> Optional[pulumi.Input[int]]: """ Number of free entries in pool to lower max size of merge: Threshold value of free entries in the pool. If the number of entries in the pool falls below this value, ClickHouse reduces the maximum size of a data part to merge. This helps handle small merges faster, rather than filling the pool with lengthy merges. """ return pulumi.get(self, "number_of_free_entries_in_pool_to_lower_max_size_of_merge") @number_of_free_entries_in_pool_to_lower_max_size_of_merge.setter def number_of_free_entries_in_pool_to_lower_max_size_of_merge(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "number_of_free_entries_in_pool_to_lower_max_size_of_merge", value) @property @pulumi.getter(name="partsToDelayInsert") def parts_to_delay_insert(self) -> Optional[pulumi.Input[int]]: """ Parts to delay insert: Number of active data parts in a table, on exceeding which ClickHouse starts artificially reduce the rate of inserting data into the table. """ return pulumi.get(self, "parts_to_delay_insert") @parts_to_delay_insert.setter def parts_to_delay_insert(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "parts_to_delay_insert", value) @property @pulumi.getter(name="partsToThrowInsert") def parts_to_throw_insert(self) -> Optional[pulumi.Input[int]]: """ Parts to throw insert: Threshold value of active data parts in a table, on exceeding which ClickHouse throws the 'Too many parts ...' exception. """ return pulumi.get(self, "parts_to_throw_insert") @parts_to_throw_insert.setter def parts_to_throw_insert(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "parts_to_throw_insert", value) @property @pulumi.getter(name="replicatedDeduplicationWindow") def replicated_deduplication_window(self) -> Optional[pulumi.Input[int]]: """ Replicated deduplication window: Number of recent hash blocks that ZooKeeper will store (the old ones will be deleted). """ return pulumi.get(self, "replicated_deduplication_window") @replicated_deduplication_window.setter def replicated_deduplication_window(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "replicated_deduplication_window", value) @property @pulumi.getter(name="replicatedDeduplicationWindowSeconds") def replicated_deduplication_window_seconds(self) -> Optional[pulumi.Input[int]]: """ Replicated deduplication window seconds: Time during which ZooKeeper stores the hash blocks (the old ones wil be deleted). """ return pulumi.get(self, "replicated_deduplication_window_seconds") @replicated_deduplication_window_seconds.setter def replicated_deduplication_window_seconds(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "replicated_deduplication_window_seconds", value) @pulumi.input_type class MdbClickhouseClusterClickhouseConfigRabbitmqArgs: def __init__(__self__, *, password: Optional[pulumi.Input[str]] = None, username: Optional[pulumi.Input[str]] = None): """ :param pulumi.Input[str] password: <PASSWORD>. :param pulumi.Input[str] username: RabbitMQ username. """ if password is not None: pulumi.set(__self__, "password", password) if username is not None: pulumi.set(__self__, "username", username) @property @pulumi.getter def password(self) -> Optional[pulumi.Input[str]]: """ RabbitMQ user password. """ return pulumi.get(self, "password") @password.setter def password(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "password", value) @property @pulumi.getter def username(self) -> Optional[pulumi.Input[str]]: """ RabbitMQ username. """ return pulumi.get(self, "username") @username.setter def username(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "username", value) @pulumi.input_type class MdbClickhouseClusterClickhouseResourcesArgs: def __init__(__self__, *, disk_size: pulumi.Input[int], disk_type_id: pulumi.Input[str], resource_preset_id: pulumi.Input[str]): """ :param pulumi.Input[int] disk_size: Volume of the storage available to a ZooKeeper host, in gigabytes. :param pulumi.Input[str] disk_type_id: Type of the storage of ZooKeeper hosts. For more information see [the official documentation](https://cloud.yandex.com/docs/managed-clickhouse/concepts/storage). """ pulumi.set(__self__, "disk_size", disk_size) pulumi.set(__self__, "disk_type_id", disk_type_id) pulumi.set(__self__, "resource_preset_id", resource_preset_id) @property @pulumi.getter(name="diskSize") def disk_size(self) -> pulumi.Input[int]: """ Volume of the storage available to a ZooKeeper host, in gigabytes. """ return pulumi.get(self, "disk_size") @disk_size.setter def disk_size(self, value: pulumi.Input[int]): pulumi.set(self, "disk_size", value) @property @pulumi.getter(name="diskTypeId") def disk_type_id(self) -> pulumi.Input[str]: """ Type of the storage of ZooKeeper hosts. For more information see [the official documentation](https://cloud.yandex.com/docs/managed-clickhouse/concepts/storage). """ return pulumi.get(self, "disk_type_id") @disk_type_id.setter def disk_type_id(self, value: pulumi.Input[str]): pulumi.set(self, "disk_type_id", value) @property @pulumi.getter(name="resourcePresetId") def resource_preset_id(self) -> pulumi.Input[str]: return pulumi.get(self, "resource_preset_id") @resource_preset_id.setter def resource_preset_id(self, value: pulumi.Input[str]): pulumi.set(self, "resource_preset_id", value) @pulumi.input_type class MdbClickhouseClusterCloudStorageArgs: def __init__(__self__, *, enabled: pulumi.Input[bool]): """ :param pulumi.Input[bool] enabled: Whether to use Yandex Object Storage for storing ClickHouse data. Can be either `true` or `false`. """ pulumi.set(__self__, "enabled", enabled) @property @pulumi.getter def enabled(self) -> pulumi.Input[bool]: """ Whether to use Yandex Object Storage for storing ClickHouse data. Can be either `true` or `false`. """ return pulumi.get(self, "enabled") @enabled.setter def enabled(self, value: pulumi.Input[bool]): pulumi.set(self, "enabled", value) @pulumi.input_type class MdbClickhouseClusterDatabaseArgs: def __init__(__self__, *, name: pulumi.Input[str]): """ :param pulumi.Input[str] name: Graphite rollup configuration name. """ pulumi.set(__self__, "name", name) @property @pulumi.getter def name(self) -> pulumi.Input[str]: """ Graphite rollup configuration name. """ return pulumi.get(self, "name") @name.setter def name(self, value: pulumi.Input[str]): pulumi.set(self, "name", value) @pulumi.input_type class MdbClickhouseClusterFormatSchemaArgs: def __init__(__self__, *, name: pulumi.Input[str], type: pulumi.Input[str], uri: pulumi.Input[str]): """ :param pulumi.Input[str] name: Graphite rollup configuration name. :param pulumi.Input[str] type: Type of maintenance window. Can be either `ANYTIME` or `WEEKLY`. A day and hour of window need to be specified with weekly window. :param pulumi.Input[str] uri: Model file URL. You can only use models stored in Yandex Object Storage. """ pulumi.set(__self__, "name", name) pulumi.set(__self__, "type", type) pulumi.set(__self__, "uri", uri) @property @pulumi.getter def name(self) -> pulumi.Input[str]: """ Graphite rollup configuration name. """ return pulumi.get(self, "name") @name.setter def name(self, value: pulumi.Input[str]): pulumi.set(self, "name", value) @property @pulumi.getter def type(self) -> pulumi.Input[str]: """ Type of maintenance window. Can be either `ANYTIME` or `WEEKLY`. A day and hour of window need to be specified with weekly window. """ return pulumi.get(self, "type") @type.setter def type(self, value: pulumi.Input[str]): pulumi.set(self, "type", value) @property @pulumi.getter def uri(self) -> pulumi.Input[str]: """ Model file URL. You can only use models stored in Yandex Object Storage. """ return pulumi.get(self, "uri") @uri.setter def uri(self, value: pulumi.Input[str]): pulumi.set(self, "uri", value) @pulumi.input_type class MdbClickhouseClusterHostArgs: def __init__(__self__, *, type: pulumi.Input[str], zone: pulumi.Input[str], assign_public_ip: Optional[pulumi.Input[bool]] = None, fqdn: Optional[pulumi.Input[str]] = None, shard_name: Optional[pulumi.Input[str]] = None, subnet_id: Optional[pulumi.Input[str]] = None): """ :param pulumi.Input[str] type: Type of maintenance window. Can be either `ANYTIME` or `WEEKLY`. A day and hour of window need to be specified with weekly window. :param pulumi.Input[str] zone: The availability zone where the ClickHouse host will be created. For more information see [the official documentation](https://cloud.yandex.com/docs/overview/concepts/geo-scope). :param pulumi.Input[bool] assign_public_ip: Sets whether the host should get a public IP address on creation. Can be either `true` or `false`. :param pulumi.Input[str] fqdn: The fully qualified domain name of the host. :param pulumi.Input[str] shard_name: The name of the shard to which the host belongs. :param pulumi.Input[str] subnet_id: The ID of the subnet, to which the host belongs. The subnet must be a part of the network to which the cluster belongs. """ pulumi.set(__self__, "type", type) pulumi.set(__self__, "zone", zone) if assign_public_ip is not None: pulumi.set(__self__, "assign_public_ip", assign_public_ip) if fqdn is not None: pulumi.set(__self__, "fqdn", fqdn) if shard_name is not None: pulumi.set(__self__, "shard_name", shard_name) if subnet_id is not None: pulumi.set(__self__, "subnet_id", subnet_id) @property @pulumi.getter def type(self) -> pulumi.Input[str]: """ Type of maintenance window. Can be either `ANYTIME` or `WEEKLY`. A day and hour of window need to be specified with weekly window. """ return pulumi.get(self, "type") @type.setter def type(self, value: pulumi.Input[str]): pulumi.set(self, "type", value) @property @pulumi.getter def zone(self) -> pulumi.Input[str]: """ The availability zone where the ClickHouse host will be created. For more information see [the official documentation](https://cloud.yandex.com/docs/overview/concepts/geo-scope). """ return pulumi.get(self, "zone") @zone.setter def zone(self, value: pulumi.Input[str]): pulumi.set(self, "zone", value) @property @pulumi.getter(name="assignPublicIp") def assign_public_ip(self) -> Optional[pulumi.Input[bool]]: """ Sets whether the host should get a public IP address on creation. Can be either `true` or `false`. """ return pulumi.get(self, "assign_public_ip") @assign_public_ip.setter def assign_public_ip(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "assign_public_ip", value) @property @pulumi.getter def fqdn(self) -> Optional[pulumi.Input[str]]: """ The fully qualified domain name of the host. """ return pulumi.get(self, "fqdn") @fqdn.setter def fqdn(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "fqdn", value) @property @pulumi.getter(name="shardName") def shard_name(self) -> Optional[pulumi.Input[str]]: """ The name of the shard to which the host belongs. """ return pulumi.get(self, "shard_name") @shard_name.setter def shard_name(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "shard_name", value) @property @pulumi.getter(name="subnetId") def subnet_id(self) -> Optional[pulumi.Input[str]]: """ The ID of the subnet, to which the host belongs. The subnet must be a part of the network to which the cluster belongs. """ return pulumi.get(self, "subnet_id") @subnet_id.setter def subnet_id(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "subnet_id", value) @pulumi.input_type class MdbClickhouseClusterMaintenanceWindowArgs: def __init__(__self__, *, type: pulumi.Input[str], day: Optional[pulumi.Input[str]] = None, hour: Optional[pulumi.Input[int]] = None): """ :param pulumi.Input[str] type: Type of maintenance window. Can be either `ANYTIME` or `WEEKLY`. A day and hour of window need to be specified with weekly window. :param pulumi.Input[str] day: Day of week for maintenance window if window type is weekly. Possible values: `MON`, `TUE`, `WED`, `THU`, `FRI`, `SAT`, `SUN`. :param pulumi.Input[int] hour: Hour of day in UTC time zone (1-24) for maintenance window if window type is weekly. """ pulumi.set(__self__, "type", type) if day is not None: pulumi.set(__self__, "day", day) if hour is not None: pulumi.set(__self__, "hour", hour) @property @pulumi.getter def type(self) -> pulumi.Input[str]: """ Type of maintenance window. Can be either `ANYTIME` or `WEEKLY`. A day and hour of window need to be specified with weekly window. """ return pulumi.get(self, "type") @type.setter def type(self, value: pulumi.Input[str]): pulumi.set(self, "type", value) @property @pulumi.getter def day(self) -> Optional[pulumi.Input[str]]: """ Day of week for maintenance window if window type is weekly. Possible values: `MON`, `TUE`, `WED`, `THU`, `FRI`, `SAT`, `SUN`. """ return pulumi.get(self, "day") @day.setter def day(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "day", value) @property @pulumi.getter def hour(self) -> Optional[pulumi.Input[int]]: """ Hour of day in UTC time zone (1-24) for maintenance window if window type is weekly. """ return pulumi.get(self, "hour") @hour.setter def hour(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "hour", value) @pulumi.input_type class MdbClickhouseClusterMlModelArgs: def __init__(__self__, *, name: pulumi.Input[str], type: pulumi.Input[str], uri: pulumi.Input[str]): """ :param pulumi.Input[str] name: Graphite rollup configuration name. :param pulumi.Input[str] type: Type of maintenance window. Can be either `ANYTIME` or `WEEKLY`. A day and hour of window need to be specified with weekly window. :param pulumi.Input[str] uri: Model file URL. You can only use models stored in Yandex Object Storage. """ pulumi.set(__self__, "name", name) pulumi.set(__self__, "type", type) pulumi.set(__self__, "uri", uri) @property @pulumi.getter def name(self) -> pulumi.Input[str]: """ Graphite rollup configuration name. """ return pulumi.get(self, "name") @name.setter def name(self, value: pulumi.Input[str]): pulumi.set(self, "name", value) @property @pulumi.getter def type(self) -> pulumi.Input[str]: """ Type of maintenance window. Can be either `ANYTIME` or `WEEKLY`. A day and hour of window need to be specified with weekly window. """ return pulumi.get(self, "type") @type.setter def type(self, value: pulumi.Input[str]): pulumi.set(self, "type", value) @property @pulumi.getter def uri(self) -> pulumi.Input[str]: """ Model file URL. You can only use models stored in Yandex Object Storage. """ return pulumi.get(self, "uri") @uri.setter def uri(self, value: pulumi.Input[str]): pulumi.set(self, "uri", value) @pulumi.input_type class MdbClickhouseClusterShardGroupArgs: def __init__(__self__, *, name: pulumi.Input[str], shard_names: pulumi.Input[Sequence[pulumi.Input[str]]], description: Optional[pulumi.Input[str]] = None): """ :param pulumi.Input[str] name: Graphite rollup configuration name. :param pulumi.Input[Sequence[pulumi.Input[str]]] shard_names: List of shards names that belong to the shard group. :param pulumi.Input[str] description: Description of the shard group. """ pulumi.set(__self__, "name", name) pulumi.set(__self__, "shard_names", shard_names) if description is not None: pulumi.set(__self__, "description", description) @property @pulumi.getter def name(self) -> pulumi.Input[str]: """ Graphite rollup configuration name. """ return pulumi.get(self, "name") @name.setter def name(self, value: pulumi.Input[str]): pulumi.set(self, "name", value) @property @pulumi.getter(name="shardNames") def shard_names(self) -> pulumi.Input[Sequence[pulumi.Input[str]]]: """ List of shards names that belong to the shard group. """ return pulumi.get(self, "shard_names") @shard_names.setter def shard_names(self, value: pulumi.Input[Sequence[pulumi.Input[str]]]): pulumi.set(self, "shard_names", value) @property @pulumi.getter def description(self) -> Optional[pulumi.Input[str]]: """ Description of the shard group. """ return pulumi.get(self, "description") @description.setter def description(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "description", value) @pulumi.input_type class MdbClickhouseClusterUserArgs: def __init__(__self__, *, name: pulumi.Input[str], password: pulumi.Input[str], permissions: Optional[pulumi.Input[Sequence[pulumi.Input['MdbClickhouseClusterUserPermissionArgs']]]] = None, quotas: Optional[pulumi.Input[Sequence[pulumi.Input['MdbClickhouseClusterUserQuotaArgs']]]] = None, settings: Optional[pulumi.Input['MdbClickhouseClusterUserSettingsArgs']] = None): """ :param pulumi.Input[str] name: Graphite rollup configuration name. :param pulumi.Input[str] password: <PASSWORD>. :param pulumi.Input[Sequence[pulumi.Input['MdbClickhouseClusterUserPermissionArgs']]] permissions: Set of permissions granted to the user. The structure is documented below. :param pulumi.Input[Sequence[pulumi.Input['MdbClickhouseClusterUserQuotaArgs']]] quotas: Set of user quotas. The structure is documented below. :param pulumi.Input['MdbClickhouseClusterUserSettingsArgs'] settings: Kafka connection settngs sanem as `kafka` block. """ pulumi.set(__self__, "name", name) pulumi.set(__self__, "password", password) if permissions is not None: pulumi.set(__self__, "permissions", permissions) if quotas is not None: pulumi.set(__self__, "quotas", quotas) if settings is not None: pulumi.set(__self__, "settings", settings) @property @pulumi.getter def name(self) -> pulumi.Input[str]: """ Graphite rollup configuration name. """ return pulumi.get(self, "name") @name.setter def name(self, value: pulumi.Input[str]): pulumi.set(self, "name", value) @property @pulumi.getter def password(self) -> pulumi.Input[str]: """ RabbitMQ user password. """ return pulumi.get(self, "password") @password.setter def password(self, value: pulumi.Input[str]): pulumi.set(self, "password", value) @property @pulumi.getter def permissions(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['MdbClickhouseClusterUserPermissionArgs']]]]: """ Set of permissions granted to the user. The structure is documented below. """ return pulumi.get(self, "permissions") @permissions.setter def permissions(self, value: Optional[pulumi.Input[Sequence[pulumi.Input['MdbClickhouseClusterUserPermissionArgs']]]]): pulumi.set(self, "permissions", value) @property @pulumi.getter def quotas(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['MdbClickhouseClusterUserQuotaArgs']]]]: """ Set of user quotas. The structure is documented below. """ return pulumi.get(self, "quotas") @quotas.setter def quotas(self, value: Optional[pulumi.Input[Sequence[pulumi.Input['MdbClickhouseClusterUserQuotaArgs']]]]): pulumi.set(self, "quotas", value) @property @pulumi.getter def settings(self) -> Optional[pulumi.Input['MdbClickhouseClusterUserSettingsArgs']]: """ Kafka connection settngs sanem as `kafka` block. """ return pulumi.get(self, "settings") @settings.setter def settings(self, value: Optional[pulumi.Input['MdbClickhouseClusterUserSettingsArgs']]): pulumi.set(self, "settings", value) @pulumi.input_type class MdbClickhouseClusterUserPermissionArgs: def __init__(__self__, *, database_name: pulumi.Input[str]): """ :param pulumi.Input[str] database_name: The name of the database that the permission grants access to. """ pulumi.set(__self__, "database_name", database_name) @property @pulumi.getter(name="databaseName") def database_name(self) -> pulumi.Input[str]: """ The name of the database that the permission grants access to. """ return pulumi.get(self, "database_name") @database_name.setter def database_name(self, value: pulumi.Input[str]): pulumi.set(self, "database_name", value) @pulumi.input_type class MdbClickhouseClusterUserQuotaArgs: def __init__(__self__, *, interval_duration: pulumi.Input[int], errors: Optional[pulumi.Input[int]] = None, execution_time: Optional[pulumi.Input[int]] = None, queries: Optional[pulumi.Input[int]] = None, read_rows: Optional[pulumi.Input[int]] = None, result_rows: Optional[pulumi.Input[int]] = None): """ :param pulumi.Input[int] interval_duration: Duration of interval for quota in milliseconds. :param pulumi.Input[int] errors: The number of queries that threw exception. :param pulumi.Input[int] execution_time: The total query execution time, in milliseconds (wall time). :param pulumi.Input[int] queries: The total number of queries. :param pulumi.Input[int] read_rows: The total number of source rows read from tables for running the query, on all remote servers. :param pulumi.Input[int] result_rows: The total number of rows given as the result. """ pulumi.set(__self__, "interval_duration", interval_duration) if errors is not None: pulumi.set(__self__, "errors", errors) if execution_time is not None: pulumi.set(__self__, "execution_time", execution_time) if queries is not None: pulumi.set(__self__, "queries", queries) if read_rows is not None: pulumi.set(__self__, "read_rows", read_rows) if result_rows is not None: pulumi.set(__self__, "result_rows", result_rows) @property @pulumi.getter(name="intervalDuration") def interval_duration(self) -> pulumi.Input[int]: """ Duration of interval for quota in milliseconds. """ return pulumi.get(self, "interval_duration") @interval_duration.setter def interval_duration(self, value: pulumi.Input[int]): pulumi.set(self, "interval_duration", value) @property @pulumi.getter def errors(self) -> Optional[pulumi.Input[int]]: """ The number of queries that threw exception. """ return pulumi.get(self, "errors") @errors.setter def errors(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "errors", value) @property @pulumi.getter(name="executionTime") def execution_time(self) -> Optional[pulumi.Input[int]]: """ The total query execution time, in milliseconds (wall time). """ return pulumi.get(self, "execution_time") @execution_time.setter def execution_time(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "execution_time", value) @property @pulumi.getter def queries(self) -> Optional[pulumi.Input[int]]: """ The total number of queries. """ return pulumi.get(self, "queries") @queries.setter def queries(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "queries", value) @property @pulumi.getter(name="readRows") def read_rows(self) -> Optional[pulumi.Input[int]]: """ The total number of source rows read from tables for running the query, on all remote servers. """ return pulumi.get(self, "read_rows") @read_rows.setter def read_rows(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "read_rows", value) @property @pulumi.getter(name="resultRows") def result_rows(self) -> Optional[pulumi.Input[int]]: """ The total number of rows given as the result. """ return pulumi.get(self, "result_rows") @result_rows.setter def result_rows(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "result_rows", value) @pulumi.input_type class MdbClickhouseClusterUserSettingsArgs: def __init__(__self__, *, add_http_cors_header: Optional[pulumi.Input[bool]] = None, allow_ddl: Optional[pulumi.Input[bool]] = None, compile: Optional[pulumi.Input[bool]] = None, compile_expressions: Optional[pulumi.Input[bool]] = None, connect_timeout: Optional[pulumi.Input[int]] = None, count_distinct_implementation: Optional[pulumi.Input[str]] = None, distinct_overflow_mode: Optional[pulumi.Input[str]] = None, distributed_aggregation_memory_efficient: Optional[pulumi.Input[bool]] = None, distributed_ddl_task_timeout: Optional[pulumi.Input[int]] = None, distributed_product_mode: Optional[pulumi.Input[str]] = None, empty_result_for_aggregation_by_empty_set: Optional[pulumi.Input[bool]] = None, enable_http_compression: Optional[pulumi.Input[bool]] = None, fallback_to_stale_replicas_for_distributed_queries: Optional[pulumi.Input[bool]] = None, force_index_by_date: Optional[pulumi.Input[bool]] = None, force_primary_key: Optional[pulumi.Input[bool]] = None, group_by_overflow_mode: Optional[pulumi.Input[str]] = None, group_by_two_level_threshold: Optional[pulumi.Input[int]] = None, group_by_two_level_threshold_bytes: Optional[pulumi.Input[int]] = None, http_connection_timeout: Optional[pulumi.Input[int]] = None, http_headers_progress_interval: Optional[pulumi.Input[int]] = None, http_receive_timeout: Optional[pulumi.Input[int]] = None, http_send_timeout: Optional[pulumi.Input[int]] = None, input_format_defaults_for_omitted_fields: Optional[pulumi.Input[bool]] = None, input_format_values_interpret_expressions: Optional[pulumi.Input[bool]] = None, insert_quorum: Optional[pulumi.Input[int]] = None, insert_quorum_timeout: Optional[pulumi.Input[int]] = None, join_overflow_mode: Optional[pulumi.Input[str]] = None, join_use_nulls: Optional[pulumi.Input[bool]] = None, joined_subquery_requires_alias: Optional[pulumi.Input[bool]] = None, low_cardinality_allow_in_native_format: Optional[pulumi.Input[bool]] = None, max_ast_depth: Optional[pulumi.Input[int]] = None, max_ast_elements: Optional[pulumi.Input[int]] = None, max_block_size: Optional[pulumi.Input[int]] = None, max_bytes_before_external_group_by: Optional[pulumi.Input[int]] = None, max_bytes_before_external_sort: Optional[pulumi.Input[int]] = None, max_bytes_in_distinct: Optional[pulumi.Input[int]] = None, max_bytes_in_join: Optional[pulumi.Input[int]] = None, max_bytes_in_set: Optional[pulumi.Input[int]] = None, max_bytes_to_read: Optional[pulumi.Input[int]] = None, max_bytes_to_sort: Optional[pulumi.Input[int]] = None, max_bytes_to_transfer: Optional[pulumi.Input[int]] = None, max_columns_to_read: Optional[pulumi.Input[int]] = None, max_execution_time: Optional[pulumi.Input[int]] = None, max_expanded_ast_elements: Optional[pulumi.Input[int]] = None, max_insert_block_size: Optional[pulumi.Input[int]] = None, max_memory_usage: Optional[pulumi.Input[int]] = None, max_memory_usage_for_user: Optional[pulumi.Input[int]] = None, max_network_bandwidth: Optional[pulumi.Input[int]] = None, max_network_bandwidth_for_user: Optional[pulumi.Input[int]] = None, max_query_size: Optional[pulumi.Input[int]] = None, max_replica_delay_for_distributed_queries: Optional[pulumi.Input[int]] = None, max_result_bytes: Optional[pulumi.Input[int]] = None, max_result_rows: Optional[pulumi.Input[int]] = None, max_rows_in_distinct: Optional[pulumi.Input[int]] = None, max_rows_in_join: Optional[pulumi.Input[int]] = None, max_rows_in_set: Optional[pulumi.Input[int]] = None, max_rows_to_group_by: Optional[pulumi.Input[int]] = None, max_rows_to_read: Optional[pulumi.Input[int]] = None, max_rows_to_sort: Optional[pulumi.Input[int]] = None, max_rows_to_transfer: Optional[pulumi.Input[int]] = None, max_temporary_columns: Optional[pulumi.Input[int]] = None, max_temporary_non_const_columns: Optional[pulumi.Input[int]] = None, max_threads: Optional[pulumi.Input[int]] = None, merge_tree_max_bytes_to_use_cache: Optional[pulumi.Input[int]] = None, merge_tree_max_rows_to_use_cache: Optional[pulumi.Input[int]] = None, merge_tree_min_bytes_for_concurrent_read: Optional[pulumi.Input[int]] = None, merge_tree_min_rows_for_concurrent_read: Optional[pulumi.Input[int]] = None, min_bytes_to_use_direct_io: Optional[pulumi.Input[int]] = None, min_count_to_compile: Optional[pulumi.Input[int]] = None, min_count_to_compile_expression: Optional[pulumi.Input[int]] = None, min_execution_speed: Optional[pulumi.Input[int]] = None, min_execution_speed_bytes: Optional[pulumi.Input[int]] = None, min_insert_block_size_bytes: Optional[pulumi.Input[int]] = None, min_insert_block_size_rows: Optional[pulumi.Input[int]] = None, output_format_json_quote64bit_integers: Optional[pulumi.Input[bool]] = None, output_format_json_quote_denormals: Optional[pulumi.Input[bool]] = None, priority: Optional[pulumi.Input[int]] = None, quota_mode: Optional[pulumi.Input[str]] = None, read_overflow_mode: Optional[pulumi.Input[str]] = None, readonly: Optional[pulumi.Input[int]] = None, receive_timeout: Optional[pulumi.Input[int]] = None, replication_alter_partitions_sync: Optional[pulumi.Input[int]] = None, result_overflow_mode: Optional[pulumi.Input[str]] = None, select_sequential_consistency: Optional[pulumi.Input[bool]] = None, send_progress_in_http_headers: Optional[pulumi.Input[bool]] = None, send_timeout: Optional[pulumi.Input[int]] = None, set_overflow_mode: Optional[pulumi.Input[str]] = None, skip_unavailable_shards: Optional[pulumi.Input[bool]] = None, sort_overflow_mode: Optional[pulumi.Input[str]] = None, timeout_overflow_mode: Optional[pulumi.Input[str]] = None, transfer_overflow_mode: Optional[pulumi.Input[str]] = None, transform_null_in: Optional[pulumi.Input[bool]] = None, use_uncompressed_cache: Optional[pulumi.Input[bool]] = None): """ :param pulumi.Input[bool] add_http_cors_header: Include CORS headers in HTTP responces. :param pulumi.Input[bool] allow_ddl: Allows or denies DDL queries. :param pulumi.Input[bool] compile: Enable compilation of queries. :param pulumi.Input[bool] compile_expressions: Turn on expression compilation. :param pulumi.Input[int] connect_timeout: Connect timeout in milliseconds on the socket used for communicating with the client. :param pulumi.Input[str] count_distinct_implementation: Specifies which of the uniq* functions should be used to perform the COUNT(DISTINCT …) construction. :param pulumi.Input[str] distinct_overflow_mode: Sets behaviour on overflow when using DISTINCT. Possible values: :param pulumi.Input[bool] distributed_aggregation_memory_efficient: Determine the behavior of distributed subqueries. :param pulumi.Input[int] distributed_ddl_task_timeout: Timeout for DDL queries, in milliseconds. :param pulumi.Input[str] distributed_product_mode: Changes the behaviour of distributed subqueries. :param pulumi.Input[bool] empty_result_for_aggregation_by_empty_set: Allows to retunr empty result. :param pulumi.Input[bool] enable_http_compression: Enables or disables data compression in the response to an HTTP request. :param pulumi.Input[bool] fallback_to_stale_replicas_for_distributed_queries: Forces a query to an out-of-date replica if updated data is not available. :param pulumi.Input[bool] force_index_by_date: Disables query execution if the index can’t be used by date. :param pulumi.Input[bool] force_primary_key: Disables query execution if indexing by the primary key is not possible. :param pulumi.Input[str] group_by_overflow_mode: Sets behaviour on overflow while GROUP BY operation. Possible values: :param pulumi.Input[int] group_by_two_level_threshold: Sets the threshold of the number of keys, after that the two-level aggregation should be used. :param pulumi.Input[int] group_by_two_level_threshold_bytes: Sets the threshold of the number of bytes, after that the two-level aggregation should be used. :param pulumi.Input[int] http_connection_timeout: Timeout for HTTP connection in milliseconds. :param pulumi.Input[int] http_headers_progress_interval: Sets minimal interval between notifications about request process in HTTP header X-ClickHouse-Progress. :param pulumi.Input[int] http_receive_timeout: Timeout for HTTP connection in milliseconds. :param pulumi.Input[int] http_send_timeout: Timeout for HTTP connection in milliseconds. :param pulumi.Input[bool] input_format_defaults_for_omitted_fields: When performing INSERT queries, replace omitted input column values with default values of the respective columns. :param pulumi.Input[bool] input_format_values_interpret_expressions: Enables or disables the full SQL parser if the fast stream parser can’t parse the data. :param pulumi.Input[int] insert_quorum: Enables the quorum writes. :param pulumi.Input[int] insert_quorum_timeout: Write to a quorum timeout in milliseconds. :param pulumi.Input[str] join_overflow_mode: Sets behaviour on overflow in JOIN. Possible values: :param pulumi.Input[bool] join_use_nulls: Sets the type of JOIN behaviour. When merging tables, empty cells may appear. ClickHouse fills them differently based on this setting. :param pulumi.Input[bool] joined_subquery_requires_alias: Require aliases for subselects and table functions in FROM that more than one table is present. :param pulumi.Input[bool] low_cardinality_allow_in_native_format: Allows or restricts using the LowCardinality data type with the Native format. :param pulumi.Input[int] max_ast_depth: Maximum abstract syntax tree depth. :param pulumi.Input[int] max_ast_elements: Maximum abstract syntax tree elements. :param pulumi.Input[int] max_block_size: A recommendation for what size of the block (in a count of rows) to load from tables. :param pulumi.Input[int] max_bytes_before_external_group_by: Limit in bytes for using memoru for GROUP BY before using swap on disk. :param pulumi.Input[int] max_bytes_before_external_sort: This setting is equivalent of the max_bytes_before_external_group_by setting, except for it is for sort operation (ORDER BY), not aggregation. :param pulumi.Input[int] max_bytes_in_distinct: Limits the maximum size of a hash table in bytes (uncompressed data) when using DISTINCT. :param pulumi.Input[int] max_bytes_in_join: Limit on maximum size of the hash table for JOIN, in bytes. :param pulumi.Input[int] max_bytes_in_set: Limit on the number of bytes in the set resulting from the execution of the IN section. :param pulumi.Input[int] max_bytes_to_read: Limits the maximum number of bytes (uncompressed data) that can be read from a table when running a query. :param pulumi.Input[int] max_bytes_to_sort: Limits the maximum number of bytes (uncompressed data) that can be read from a table for sorting. :param pulumi.Input[int] max_bytes_to_transfer: Limits the maximum number of bytes (uncompressed data) that can be passed to a remote server or saved in a temporary table when using GLOBAL IN. :param pulumi.Input[int] max_columns_to_read: Limits the maximum number of columns that can be read from a table in a single query. :param pulumi.Input[int] max_execution_time: Limits the maximum query execution time in milliseconds. :param pulumi.Input[int] max_expanded_ast_elements: Maximum abstract syntax tree depth after after expansion of aliases. :param pulumi.Input[int] max_insert_block_size: The size of blocks (in a count of rows) to form for insertion into a table. :param pulumi.Input[int] max_memory_usage: Limits the maximum memory usage (in bytes) for processing queries on a single server. :param pulumi.Input[int] max_memory_usage_for_user: Limits the maximum memory usage (in bytes) for processing of user's queries on a single server. :param pulumi.Input[int] max_network_bandwidth: Limits the speed of the data exchange over the network in bytes per second. :param pulumi.Input[int] max_network_bandwidth_for_user: Limits the speed of the data exchange over the network in bytes per second. :param pulumi.Input[int] max_query_size: The maximum part of a query that can be taken to RAM for parsing with the SQL parser. :param pulumi.Input[int] max_replica_delay_for_distributed_queries: Disables lagging replicas for distributed queries. :param pulumi.Input[int] max_result_bytes: Limits the number of bytes in the result. :param pulumi.Input[int] max_result_rows: Limits the number of rows in the result. :param pulumi.Input[int] max_rows_in_distinct: Limits the maximum number of different rows when using DISTINCT. :param pulumi.Input[int] max_rows_in_join: Limit on maximum size of the hash table for JOIN, in rows. :param pulumi.Input[int] max_rows_in_set: Limit on the number of rows in the set resulting from the execution of the IN section. :param pulumi.Input[int] max_rows_to_group_by: Limits the maximum number of unique keys received from aggregation function. :param pulumi.Input[int] max_rows_to_read: Limits the maximum number of rows that can be read from a table when running a query. :param pulumi.Input[int] max_rows_to_sort: Limits the maximum number of rows that can be read from a table for sorting. :param pulumi.Input[int] max_rows_to_transfer: Limits the maximum number of rows that can be passed to a remote server or saved in a temporary table when using GLOBAL IN. :param pulumi.Input[int] max_temporary_columns: Limits the maximum number of temporary columns that must be kept in RAM at the same time when running a query, including constant columns. :param pulumi.Input[int] max_temporary_non_const_columns: Limits the maximum number of temporary columns that must be kept in RAM at the same time when running a query, excluding constant columns. :param pulumi.Input[int] max_threads: The maximum number of query processing threads, excluding threads for retrieving data from remote servers. :param pulumi.Input[int] merge_tree_max_bytes_to_use_cache: If ClickHouse should read more than merge_tree_max_bytes_to_use_cache bytes in one query, it doesn’t use the cache of uncompressed blocks. :param pulumi.Input[int] merge_tree_max_rows_to_use_cache: If ClickHouse should read more than merge_tree_max_rows_to_use_cache rows in one query, it doesn’t use the cache of uncompressed blocks. :param pulumi.Input[int] merge_tree_min_bytes_for_concurrent_read: If the number of bytes to read from one file of a MergeTree-engine table exceeds merge_tree_min_bytes_for_concurrent_read, then ClickHouse tries to concurrently read from this file in several threads. :param pulumi.Input[int] merge_tree_min_rows_for_concurrent_read: If the number of rows to be read from a file of a MergeTree table exceeds merge_tree_min_rows_for_concurrent_read then ClickHouse tries to perform a concurrent reading from this file on several threads. :param pulumi.Input[int] min_bytes_to_use_direct_io: The minimum data volume required for using direct I/O access to the storage disk. :param pulumi.Input[int] min_count_to_compile: How many times to potentially use a compiled chunk of code before running compilation. :param pulumi.Input[int] min_count_to_compile_expression: A query waits for expression compilation process to complete prior to continuing execution. :param pulumi.Input[int] min_execution_speed: Minimal execution speed in rows per second. :param pulumi.Input[int] min_execution_speed_bytes: Minimal execution speed in bytes per second. :param pulumi.Input[int] min_insert_block_size_bytes: Sets the minimum number of bytes in the block which can be inserted into a table by an INSERT query. :param pulumi.Input[int] min_insert_block_size_rows: Sets the minimum number of rows in the block which can be inserted into a table by an INSERT query. :param pulumi.Input[bool] output_format_json_quote64bit_integers: If the value is true, integers appear in quotes when using JSON* Int64 and UInt64 formats (for compatibility with most JavaScript implementations); otherwise, integers are output without the quotes. :param pulumi.Input[bool] output_format_json_quote_denormals: Enables +nan, -nan, +inf, -inf outputs in JSON output format. :param pulumi.Input[int] priority: Query priority. :param pulumi.Input[str] quota_mode: Quota accounting mode. :param pulumi.Input[str] read_overflow_mode: Sets behaviour on overflow while read. Possible values: :param pulumi.Input[int] readonly: Restricts permissions for reading data, write data and change settings queries. :param pulumi.Input[int] receive_timeout: Receive timeout in milliseconds on the socket used for communicating with the client. :param pulumi.Input[int] replication_alter_partitions_sync: For ALTER ... ATTACH|DETACH|DROP queries, you can use the replication_alter_partitions_sync setting to set up waiting. :param pulumi.Input[str] result_overflow_mode: Sets behaviour on overflow in result. Possible values: :param pulumi.Input[bool] select_sequential_consistency: Enables or disables sequential consistency for SELECT queries. :param pulumi.Input[bool] send_progress_in_http_headers: Enables or disables X-ClickHouse-Progress HTTP response headers in clickhouse-server responses. :param pulumi.Input[int] send_timeout: Send timeout in milliseconds on the socket used for communicating with the client. :param pulumi.Input[str] set_overflow_mode: Sets behaviour on overflow in the set resulting. Possible values: :param pulumi.Input[bool] skip_unavailable_shards: Enables or disables silently skipping of unavailable shards. :param pulumi.Input[str] sort_overflow_mode: Sets behaviour on overflow while sort. Possible values: :param pulumi.Input[str] timeout_overflow_mode: Sets behaviour on overflow. Possible values: :param pulumi.Input[str] transfer_overflow_mode: Sets behaviour on overflow. Possible values: :param pulumi.Input[bool] transform_null_in: Enables equality of NULL values for IN operator. :param pulumi.Input[bool] use_uncompressed_cache: Whether to use a cache of uncompressed blocks. """ if add_http_cors_header is not None: pulumi.set(__self__, "add_http_cors_header", add_http_cors_header) if allow_ddl is not None: pulumi.set(__self__, "allow_ddl", allow_ddl) if compile is not None: pulumi.set(__self__, "compile", compile) if compile_expressions is not None: pulumi.set(__self__, "compile_expressions", compile_expressions) if connect_timeout is not None: pulumi.set(__self__, "connect_timeout", connect_timeout) if count_distinct_implementation is not None: pulumi.set(__self__, "count_distinct_implementation", count_distinct_implementation) if distinct_overflow_mode is not None: pulumi.set(__self__, "distinct_overflow_mode", distinct_overflow_mode) if distributed_aggregation_memory_efficient is not None: pulumi.set(__self__, "distributed_aggregation_memory_efficient", distributed_aggregation_memory_efficient) if distributed_ddl_task_timeout is not None: pulumi.set(__self__, "distributed_ddl_task_timeout", distributed_ddl_task_timeout) if distributed_product_mode is not None: pulumi.set(__self__, "distributed_product_mode", distributed_product_mode) if empty_result_for_aggregation_by_empty_set is not None: pulumi.set(__self__, "empty_result_for_aggregation_by_empty_set", empty_result_for_aggregation_by_empty_set) if enable_http_compression is not None: pulumi.set(__self__, "enable_http_compression", enable_http_compression) if fallback_to_stale_replicas_for_distributed_queries is not None: pulumi.set(__self__, "fallback_to_stale_replicas_for_distributed_queries", fallback_to_stale_replicas_for_distributed_queries) if force_index_by_date is not None: pulumi.set(__self__, "force_index_by_date", force_index_by_date) if force_primary_key is not None: pulumi.set(__self__, "force_primary_key", force_primary_key) if group_by_overflow_mode is not None: pulumi.set(__self__, "group_by_overflow_mode", group_by_overflow_mode) if group_by_two_level_threshold is not None: pulumi.set(__self__, "group_by_two_level_threshold", group_by_two_level_threshold) if group_by_two_level_threshold_bytes is not None: pulumi.set(__self__, "group_by_two_level_threshold_bytes", group_by_two_level_threshold_bytes) if http_connection_timeout is not None: pulumi.set(__self__, "http_connection_timeout", http_connection_timeout) if http_headers_progress_interval is not None: pulumi.set(__self__, "http_headers_progress_interval", http_headers_progress_interval) if http_receive_timeout is not None: pulumi.set(__self__, "http_receive_timeout", http_receive_timeout) if http_send_timeout is not None: pulumi.set(__self__, "http_send_timeout", http_send_timeout) if input_format_defaults_for_omitted_fields is not None: pulumi.set(__self__, "input_format_defaults_for_omitted_fields", input_format_defaults_for_omitted_fields) if input_format_values_interpret_expressions is not None: pulumi.set(__self__, "input_format_values_interpret_expressions", input_format_values_interpret_expressions) if insert_quorum is not None: pulumi.set(__self__, "insert_quorum", insert_quorum) if insert_quorum_timeout is not None: pulumi.set(__self__, "insert_quorum_timeout", insert_quorum_timeout) if join_overflow_mode is not None: pulumi.set(__self__, "join_overflow_mode", join_overflow_mode) if join_use_nulls is not None: pulumi.set(__self__, "join_use_nulls", join_use_nulls) if joined_subquery_requires_alias is not None: pulumi.set(__self__, "joined_subquery_requires_alias", joined_subquery_requires_alias) if low_cardinality_allow_in_native_format is not None: pulumi.set(__self__, "low_cardinality_allow_in_native_format", low_cardinality_allow_in_native_format) if max_ast_depth is not None: pulumi.set(__self__, "max_ast_depth", max_ast_depth) if max_ast_elements is not None: pulumi.set(__self__, "max_ast_elements", max_ast_elements) if max_block_size is not None: pulumi.set(__self__, "max_block_size", max_block_size) if max_bytes_before_external_group_by is not None: pulumi.set(__self__, "max_bytes_before_external_group_by", max_bytes_before_external_group_by) if max_bytes_before_external_sort is not None: pulumi.set(__self__, "max_bytes_before_external_sort", max_bytes_before_external_sort) if max_bytes_in_distinct is not None: pulumi.set(__self__, "max_bytes_in_distinct", max_bytes_in_distinct) if max_bytes_in_join is not None: pulumi.set(__self__, "max_bytes_in_join", max_bytes_in_join) if max_bytes_in_set is not None: pulumi.set(__self__, "max_bytes_in_set", max_bytes_in_set) if max_bytes_to_read is not None: pulumi.set(__self__, "max_bytes_to_read", max_bytes_to_read) if max_bytes_to_sort is not None: pulumi.set(__self__, "max_bytes_to_sort", max_bytes_to_sort) if max_bytes_to_transfer is not None: pulumi.set(__self__, "max_bytes_to_transfer", max_bytes_to_transfer) if max_columns_to_read is not None: pulumi.set(__self__, "max_columns_to_read", max_columns_to_read) if max_execution_time is not None: pulumi.set(__self__, "max_execution_time", max_execution_time) if max_expanded_ast_elements is not None: pulumi.set(__self__, "max_expanded_ast_elements", max_expanded_ast_elements) if max_insert_block_size is not None: pulumi.set(__self__, "max_insert_block_size", max_insert_block_size) if max_memory_usage is not None: pulumi.set(__self__, "max_memory_usage", max_memory_usage) if max_memory_usage_for_user is not None: pulumi.set(__self__, "max_memory_usage_for_user", max_memory_usage_for_user) if max_network_bandwidth is not None: pulumi.set(__self__, "max_network_bandwidth", max_network_bandwidth) if max_network_bandwidth_for_user is not None: pulumi.set(__self__, "max_network_bandwidth_for_user", max_network_bandwidth_for_user) if max_query_size is not None: pulumi.set(__self__, "max_query_size", max_query_size) if max_replica_delay_for_distributed_queries is not None: pulumi.set(__self__, "max_replica_delay_for_distributed_queries", max_replica_delay_for_distributed_queries) if max_result_bytes is not None: pulumi.set(__self__, "max_result_bytes", max_result_bytes) if max_result_rows is not None: pulumi.set(__self__, "max_result_rows", max_result_rows) if max_rows_in_distinct is not None: pulumi.set(__self__, "max_rows_in_distinct", max_rows_in_distinct) if max_rows_in_join is not None: pulumi.set(__self__, "max_rows_in_join", max_rows_in_join) if max_rows_in_set is not None: pulumi.set(__self__, "max_rows_in_set", max_rows_in_set) if max_rows_to_group_by is not None: pulumi.set(__self__, "max_rows_to_group_by", max_rows_to_group_by) if max_rows_to_read is not None: pulumi.set(__self__, "max_rows_to_read", max_rows_to_read) if max_rows_to_sort is not None: pulumi.set(__self__, "max_rows_to_sort", max_rows_to_sort) if max_rows_to_transfer is not None: pulumi.set(__self__, "max_rows_to_transfer", max_rows_to_transfer) if max_temporary_columns is not None: pulumi.set(__self__, "max_temporary_columns", max_temporary_columns) if max_temporary_non_const_columns is not None: pulumi.set(__self__, "max_temporary_non_const_columns", max_temporary_non_const_columns) if max_threads is not None: pulumi.set(__self__, "max_threads", max_threads) if merge_tree_max_bytes_to_use_cache is not None: pulumi.set(__self__, "merge_tree_max_bytes_to_use_cache", merge_tree_max_bytes_to_use_cache) if merge_tree_max_rows_to_use_cache is not None: pulumi.set(__self__, "merge_tree_max_rows_to_use_cache", merge_tree_max_rows_to_use_cache) if merge_tree_min_bytes_for_concurrent_read is not None: pulumi.set(__self__, "merge_tree_min_bytes_for_concurrent_read", merge_tree_min_bytes_for_concurrent_read) if merge_tree_min_rows_for_concurrent_read is not None: pulumi.set(__self__, "merge_tree_min_rows_for_concurrent_read", merge_tree_min_rows_for_concurrent_read) if min_bytes_to_use_direct_io is not None: pulumi.set(__self__, "min_bytes_to_use_direct_io", min_bytes_to_use_direct_io) if min_count_to_compile is not None: pulumi.set(__self__, "min_count_to_compile", min_count_to_compile) if min_count_to_compile_expression is not None: pulumi.set(__self__, "min_count_to_compile_expression", min_count_to_compile_expression) if min_execution_speed is not None: pulumi.set(__self__, "min_execution_speed", min_execution_speed) if min_execution_speed_bytes is not None: pulumi.set(__self__, "min_execution_speed_bytes", min_execution_speed_bytes) if min_insert_block_size_bytes is not None: pulumi.set(__self__, "min_insert_block_size_bytes", min_insert_block_size_bytes) if min_insert_block_size_rows is not None: pulumi.set(__self__, "min_insert_block_size_rows", min_insert_block_size_rows) if output_format_json_quote64bit_integers is not None: pulumi.set(__self__, "output_format_json_quote64bit_integers", output_format_json_quote64bit_integers) if output_format_json_quote_denormals is not None: pulumi.set(__self__, "output_format_json_quote_denormals", output_format_json_quote_denormals) if priority is not None: pulumi.set(__self__, "priority", priority) if quota_mode is not None: pulumi.set(__self__, "quota_mode", quota_mode) if read_overflow_mode is not None: pulumi.set(__self__, "read_overflow_mode", read_overflow_mode) if readonly is not None: pulumi.set(__self__, "readonly", readonly) if receive_timeout is not None: pulumi.set(__self__, "receive_timeout", receive_timeout) if replication_alter_partitions_sync is not None: pulumi.set(__self__, "replication_alter_partitions_sync", replication_alter_partitions_sync) if result_overflow_mode is not None: pulumi.set(__self__, "result_overflow_mode", result_overflow_mode) if select_sequential_consistency is not None: pulumi.set(__self__, "select_sequential_consistency", select_sequential_consistency) if send_progress_in_http_headers is not None: pulumi.set(__self__, "send_progress_in_http_headers", send_progress_in_http_headers) if send_timeout is not None: pulumi.set(__self__, "send_timeout", send_timeout) if set_overflow_mode is not None: pulumi.set(__self__, "set_overflow_mode", set_overflow_mode) if skip_unavailable_shards is not None: pulumi.set(__self__, "skip_unavailable_shards", skip_unavailable_shards) if sort_overflow_mode is not None: pulumi.set(__self__, "sort_overflow_mode", sort_overflow_mode) if timeout_overflow_mode is not None: pulumi.set(__self__, "timeout_overflow_mode", timeout_overflow_mode) if transfer_overflow_mode is not None: pulumi.set(__self__, "transfer_overflow_mode", transfer_overflow_mode) if transform_null_in is not None: pulumi.set(__self__, "transform_null_in", transform_null_in) if use_uncompressed_cache is not None: pulumi.set(__self__, "use_uncompressed_cache", use_uncompressed_cache) @property @pulumi.getter(name="addHttpCorsHeader") def add_http_cors_header(self) -> Optional[pulumi.Input[bool]]: """ Include CORS headers in HTTP responces. """ return pulumi.get(self, "add_http_cors_header") @add_http_cors_header.setter def add_http_cors_header(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "add_http_cors_header", value) @property @pulumi.getter(name="allowDdl") def allow_ddl(self) -> Optional[pulumi.Input[bool]]: """ Allows or denies DDL queries. """ return pulumi.get(self, "allow_ddl") @allow_ddl.setter def allow_ddl(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "allow_ddl", value) @property @pulumi.getter def compile(self) -> Optional[pulumi.Input[bool]]: """ Enable compilation of queries. """ return pulumi.get(self, "compile") @compile.setter def compile(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "compile", value) @property @pulumi.getter(name="compileExpressions") def compile_expressions(self) -> Optional[pulumi.Input[bool]]: """ Turn on expression compilation. """ return pulumi.get(self, "compile_expressions") @compile_expressions.setter def compile_expressions(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "compile_expressions", value) @property @pulumi.getter(name="connectTimeout") def connect_timeout(self) -> Optional[pulumi.Input[int]]: """ Connect timeout in milliseconds on the socket used for communicating with the client. """ return pulumi.get(self, "connect_timeout") @connect_timeout.setter def connect_timeout(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "connect_timeout", value) @property @pulumi.getter(name="countDistinctImplementation") def count_distinct_implementation(self) -> Optional[pulumi.Input[str]]: """ Specifies which of the uniq* functions should be used to perform the COUNT(DISTINCT …) construction. """ return pulumi.get(self, "count_distinct_implementation") @count_distinct_implementation.setter def count_distinct_implementation(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "count_distinct_implementation", value) @property @pulumi.getter(name="distinctOverflowMode") def distinct_overflow_mode(self) -> Optional[pulumi.Input[str]]: """ Sets behaviour on overflow when using DISTINCT. Possible values: """ return pulumi.get(self, "distinct_overflow_mode") @distinct_overflow_mode.setter def distinct_overflow_mode(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "distinct_overflow_mode", value) @property @pulumi.getter(name="distributedAggregationMemoryEfficient") def distributed_aggregation_memory_efficient(self) -> Optional[pulumi.Input[bool]]: """ Determine the behavior of distributed subqueries. """ return pulumi.get(self, "distributed_aggregation_memory_efficient") @distributed_aggregation_memory_efficient.setter def distributed_aggregation_memory_efficient(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "distributed_aggregation_memory_efficient", value) @property @pulumi.getter(name="distributedDdlTaskTimeout") def distributed_ddl_task_timeout(self) -> Optional[pulumi.Input[int]]: """ Timeout for DDL queries, in milliseconds. """ return pulumi.get(self, "distributed_ddl_task_timeout") @distributed_ddl_task_timeout.setter def distributed_ddl_task_timeout(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "distributed_ddl_task_timeout", value) @property @pulumi.getter(name="distributedProductMode") def distributed_product_mode(self) -> Optional[pulumi.Input[str]]: """ Changes the behaviour of distributed subqueries. """ return pulumi.get(self, "distributed_product_mode") @distributed_product_mode.setter def distributed_product_mode(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "distributed_product_mode", value) @property @pulumi.getter(name="emptyResultForAggregationByEmptySet") def empty_result_for_aggregation_by_empty_set(self) -> Optional[pulumi.Input[bool]]: """ Allows to retunr empty result. """ return pulumi.get(self, "empty_result_for_aggregation_by_empty_set") @empty_result_for_aggregation_by_empty_set.setter def empty_result_for_aggregation_by_empty_set(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "empty_result_for_aggregation_by_empty_set", value) @property @pulumi.getter(name="enableHttpCompression") def enable_http_compression(self) -> Optional[pulumi.Input[bool]]: """ Enables or disables data compression in the response to an HTTP request. """ return pulumi.get(self, "enable_http_compression") @enable_http_compression.setter def enable_http_compression(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "enable_http_compression", value) @property @pulumi.getter(name="fallbackToStaleReplicasForDistributedQueries") def fallback_to_stale_replicas_for_distributed_queries(self) -> Optional[pulumi.Input[bool]]: """ Forces a query to an out-of-date replica if updated data is not available. """ return pulumi.get(self, "fallback_to_stale_replicas_for_distributed_queries") @fallback_to_stale_replicas_for_distributed_queries.setter def fallback_to_stale_replicas_for_distributed_queries(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "fallback_to_stale_replicas_for_distributed_queries", value) @property @pulumi.getter(name="forceIndexByDate") def force_index_by_date(self) -> Optional[pulumi.Input[bool]]: """ Disables query execution if the index can’t be used by date. """ return pulumi.get(self, "force_index_by_date") @force_index_by_date.setter def force_index_by_date(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "force_index_by_date", value) @property @pulumi.getter(name="forcePrimaryKey") def force_primary_key(self) -> Optional[pulumi.Input[bool]]: """ Disables query execution if indexing by the primary key is not possible. """ return pulumi.get(self, "force_primary_key") @force_primary_key.setter def force_primary_key(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "force_primary_key", value) @property @pulumi.getter(name="groupByOverflowMode") def group_by_overflow_mode(self) -> Optional[pulumi.Input[str]]: """ Sets behaviour on overflow while GROUP BY operation. Possible values: """ return pulumi.get(self, "group_by_overflow_mode") @group_by_overflow_mode.setter def group_by_overflow_mode(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "group_by_overflow_mode", value) @property @pulumi.getter(name="groupByTwoLevelThreshold") def group_by_two_level_threshold(self) -> Optional[pulumi.Input[int]]: """ Sets the threshold of the number of keys, after that the two-level aggregation should be used. """ return pulumi.get(self, "group_by_two_level_threshold") @group_by_two_level_threshold.setter def group_by_two_level_threshold(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "group_by_two_level_threshold", value) @property @pulumi.getter(name="groupByTwoLevelThresholdBytes") def group_by_two_level_threshold_bytes(self) -> Optional[pulumi.Input[int]]: """ Sets the threshold of the number of bytes, after that the two-level aggregation should be used. """ return pulumi.get(self, "group_by_two_level_threshold_bytes") @group_by_two_level_threshold_bytes.setter def group_by_two_level_threshold_bytes(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "group_by_two_level_threshold_bytes", value) @property @pulumi.getter(name="httpConnectionTimeout") def http_connection_timeout(self) -> Optional[pulumi.Input[int]]: """ Timeout for HTTP connection in milliseconds. """ return pulumi.get(self, "http_connection_timeout") @http_connection_timeout.setter def http_connection_timeout(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "http_connection_timeout", value) @property @pulumi.getter(name="httpHeadersProgressInterval") def http_headers_progress_interval(self) -> Optional[pulumi.Input[int]]: """ Sets minimal interval between notifications about request process in HTTP header X-ClickHouse-Progress. """ return pulumi.get(self, "http_headers_progress_interval") @http_headers_progress_interval.setter def http_headers_progress_interval(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "http_headers_progress_interval", value) @property @pulumi.getter(name="httpReceiveTimeout") def http_receive_timeout(self) -> Optional[pulumi.Input[int]]: """ Timeout for HTTP connection in milliseconds. """ return pulumi.get(self, "http_receive_timeout") @http_receive_timeout.setter def http_receive_timeout(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "http_receive_timeout", value) @property @pulumi.getter(name="httpSendTimeout") def http_send_timeout(self) -> Optional[pulumi.Input[int]]: """ Timeout for HTTP connection in milliseconds. """ return pulumi.get(self, "http_send_timeout") @http_send_timeout.setter def http_send_timeout(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "http_send_timeout", value) @property @pulumi.getter(name="inputFormatDefaultsForOmittedFields") def input_format_defaults_for_omitted_fields(self) -> Optional[pulumi.Input[bool]]: """ When performing INSERT queries, replace omitted input column values with default values of the respective columns. """ return pulumi.get(self, "input_format_defaults_for_omitted_fields") @input_format_defaults_for_omitted_fields.setter def input_format_defaults_for_omitted_fields(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "input_format_defaults_for_omitted_fields", value) @property @pulumi.getter(name="inputFormatValuesInterpretExpressions") def input_format_values_interpret_expressions(self) -> Optional[pulumi.Input[bool]]: """ Enables or disables the full SQL parser if the fast stream parser can’t parse the data. """ return pulumi.get(self, "input_format_values_interpret_expressions") @input_format_values_interpret_expressions.setter def input_format_values_interpret_expressions(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "input_format_values_interpret_expressions", value) @property @pulumi.getter(name="insertQuorum") def insert_quorum(self) -> Optional[pulumi.Input[int]]: """ Enables the quorum writes. """ return pulumi.get(self, "insert_quorum") @insert_quorum.setter def insert_quorum(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "insert_quorum", value) @property @pulumi.getter(name="insertQuorumTimeout") def insert_quorum_timeout(self) -> Optional[pulumi.Input[int]]: """ Write to a quorum timeout in milliseconds. """ return pulumi.get(self, "insert_quorum_timeout") @insert_quorum_timeout.setter def insert_quorum_timeout(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "insert_quorum_timeout", value) @property @pulumi.getter(name="joinOverflowMode") def join_overflow_mode(self) -> Optional[pulumi.Input[str]]: """ Sets behaviour on overflow in JOIN. Possible values: """ return pulumi.get(self, "join_overflow_mode") @join_overflow_mode.setter def join_overflow_mode(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "join_overflow_mode", value) @property @pulumi.getter(name="joinUseNulls") def join_use_nulls(self) -> Optional[pulumi.Input[bool]]: """ Sets the type of JOIN behaviour. When merging tables, empty cells may appear. ClickHouse fills them differently based on this setting. """ return pulumi.get(self, "join_use_nulls") @join_use_nulls.setter def join_use_nulls(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "join_use_nulls", value) @property @pulumi.getter(name="joinedSubqueryRequiresAlias") def joined_subquery_requires_alias(self) -> Optional[pulumi.Input[bool]]: """ Require aliases for subselects and table functions in FROM that more than one table is present. """ return pulumi.get(self, "joined_subquery_requires_alias") @joined_subquery_requires_alias.setter def joined_subquery_requires_alias(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "joined_subquery_requires_alias", value) @property @pulumi.getter(name="lowCardinalityAllowInNativeFormat") def low_cardinality_allow_in_native_format(self) -> Optional[pulumi.Input[bool]]: """ Allows or restricts using the LowCardinality data type with the Native format. """ return pulumi.get(self, "low_cardinality_allow_in_native_format") @low_cardinality_allow_in_native_format.setter def low_cardinality_allow_in_native_format(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "low_cardinality_allow_in_native_format", value) @property @pulumi.getter(name="maxAstDepth") def max_ast_depth(self) -> Optional[pulumi.Input[int]]: """ Maximum abstract syntax tree depth. """ return pulumi.get(self, "max_ast_depth") @max_ast_depth.setter def max_ast_depth(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "max_ast_depth", value) @property @pulumi.getter(name="maxAstElements") def max_ast_elements(self) -> Optional[pulumi.Input[int]]: """ Maximum abstract syntax tree elements. """ return pulumi.get(self, "max_ast_elements") @max_ast_elements.setter def max_ast_elements(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "max_ast_elements", value) @property @pulumi.getter(name="maxBlockSize") def max_block_size(self) -> Optional[pulumi.Input[int]]: """ A recommendation for what size of the block (in a count of rows) to load from tables. """ return pulumi.get(self, "max_block_size") @max_block_size.setter def max_block_size(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "max_block_size", value) @property @pulumi.getter(name="maxBytesBeforeExternalGroupBy") def max_bytes_before_external_group_by(self) -> Optional[pulumi.Input[int]]: """ Limit in bytes for using memoru for GROUP BY before using swap on disk. """ return pulumi.get(self, "max_bytes_before_external_group_by") @max_bytes_before_external_group_by.setter def max_bytes_before_external_group_by(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "max_bytes_before_external_group_by", value) @property @pulumi.getter(name="maxBytesBeforeExternalSort") def max_bytes_before_external_sort(self) -> Optional[pulumi.Input[int]]: """ This setting is equivalent of the max_bytes_before_external_group_by setting, except for it is for sort operation (ORDER BY), not aggregation. """ return pulumi.get(self, "max_bytes_before_external_sort") @max_bytes_before_external_sort.setter def max_bytes_before_external_sort(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "max_bytes_before_external_sort", value) @property @pulumi.getter(name="maxBytesInDistinct") def max_bytes_in_distinct(self) -> Optional[pulumi.Input[int]]: """ Limits the maximum size of a hash table in bytes (uncompressed data) when using DISTINCT. """ return pulumi.get(self, "max_bytes_in_distinct") @max_bytes_in_distinct.setter def max_bytes_in_distinct(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "max_bytes_in_distinct", value) @property @pulumi.getter(name="maxBytesInJoin") def max_bytes_in_join(self) -> Optional[pulumi.Input[int]]: """ Limit on maximum size of the hash table for JOIN, in bytes. """ return pulumi.get(self, "max_bytes_in_join") @max_bytes_in_join.setter def max_bytes_in_join(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "max_bytes_in_join", value) @property @pulumi.getter(name="maxBytesInSet") def max_bytes_in_set(self) -> Optional[pulumi.Input[int]]: """ Limit on the number of bytes in the set resulting from the execution of the IN section. """ return pulumi.get(self, "max_bytes_in_set") @max_bytes_in_set.setter def max_bytes_in_set(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "max_bytes_in_set", value) @property @pulumi.getter(name="maxBytesToRead") def max_bytes_to_read(self) -> Optional[pulumi.Input[int]]: """ Limits the maximum number of bytes (uncompressed data) that can be read from a table when running a query. """ return pulumi.get(self, "max_bytes_to_read") @max_bytes_to_read.setter def max_bytes_to_read(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "max_bytes_to_read", value) @property @pulumi.getter(name="maxBytesToSort") def max_bytes_to_sort(self) -> Optional[pulumi.Input[int]]: """ Limits the maximum number of bytes (uncompressed data) that can be read from a table for sorting. """ return pulumi.get(self, "max_bytes_to_sort") @max_bytes_to_sort.setter def max_bytes_to_sort(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "max_bytes_to_sort", value) @property @pulumi.getter(name="maxBytesToTransfer") def max_bytes_to_transfer(self) -> Optional[pulumi.Input[int]]: """ Limits the maximum number of bytes (uncompressed data) that can be passed to a remote server or saved in a temporary table when using GLOBAL IN. """ return pulumi.get(self, "max_bytes_to_transfer") @max_bytes_to_transfer.setter def max_bytes_to_transfer(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "max_bytes_to_transfer", value) @property @pulumi.getter(name="maxColumnsToRead") def max_columns_to_read(self) -> Optional[pulumi.Input[int]]: """ Limits the maximum number of columns that can be read from a table in a single query. """ return pulumi.get(self, "max_columns_to_read") @max_columns_to_read.setter def max_columns_to_read(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "max_columns_to_read", value) @property @pulumi.getter(name="maxExecutionTime") def max_execution_time(self) -> Optional[pulumi.Input[int]]: """ Limits the maximum query execution time in milliseconds. """ return pulumi.get(self, "max_execution_time") @max_execution_time.setter def max_execution_time(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "max_execution_time", value) @property @pulumi.getter(name="maxExpandedAstElements") def max_expanded_ast_elements(self) -> Optional[pulumi.Input[int]]: """ Maximum abstract syntax tree depth after after expansion of aliases. """ return pulumi.get(self, "max_expanded_ast_elements") @max_expanded_ast_elements.setter def max_expanded_ast_elements(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "max_expanded_ast_elements", value) @property @pulumi.getter(name="maxInsertBlockSize") def max_insert_block_size(self) -> Optional[pulumi.Input[int]]: """ The size of blocks (in a count of rows) to form for insertion into a table. """ return pulumi.get(self, "max_insert_block_size") @max_insert_block_size.setter def max_insert_block_size(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "max_insert_block_size", value) @property @pulumi.getter(name="maxMemoryUsage") def max_memory_usage(self) -> Optional[pulumi.Input[int]]: """ Limits the maximum memory usage (in bytes) for processing queries on a single server. """ return pulumi.get(self, "max_memory_usage") @max_memory_usage.setter def max_memory_usage(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "max_memory_usage", value) @property @pulumi.getter(name="maxMemoryUsageForUser") def max_memory_usage_for_user(self) -> Optional[pulumi.Input[int]]: """ Limits the maximum memory usage (in bytes) for processing of user's queries on a single server. """ return pulumi.get(self, "max_memory_usage_for_user") @max_memory_usage_for_user.setter def max_memory_usage_for_user(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "max_memory_usage_for_user", value) @property @pulumi.getter(name="maxNetworkBandwidth") def max_network_bandwidth(self) -> Optional[pulumi.Input[int]]: """ Limits the speed of the data exchange over the network in bytes per second. """ return pulumi.get(self, "max_network_bandwidth") @max_network_bandwidth.setter def max_network_bandwidth(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "max_network_bandwidth", value) @property @pulumi.getter(name="maxNetworkBandwidthForUser") def max_network_bandwidth_for_user(self) -> Optional[pulumi.Input[int]]: """ Limits the speed of the data exchange over the network in bytes per second. """ return pulumi.get(self, "max_network_bandwidth_for_user") @max_network_bandwidth_for_user.setter def max_network_bandwidth_for_user(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "max_network_bandwidth_for_user", value) @property @pulumi.getter(name="maxQuerySize") def max_query_size(self) -> Optional[pulumi.Input[int]]: """ The maximum part of a query that can be taken to RAM for parsing with the SQL parser. """ return pulumi.get(self, "max_query_size") @max_query_size.setter def max_query_size(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "max_query_size", value) @property @pulumi.getter(name="maxReplicaDelayForDistributedQueries") def max_replica_delay_for_distributed_queries(self) -> Optional[pulumi.Input[int]]: """ Disables lagging replicas for distributed queries. """ return pulumi.get(self, "max_replica_delay_for_distributed_queries") @max_replica_delay_for_distributed_queries.setter def max_replica_delay_for_distributed_queries(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "max_replica_delay_for_distributed_queries", value) @property @pulumi.getter(name="maxResultBytes") def max_result_bytes(self) -> Optional[pulumi.Input[int]]: """ Limits the number of bytes in the result. """ return pulumi.get(self, "max_result_bytes") @max_result_bytes.setter def max_result_bytes(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "max_result_bytes", value) @property @pulumi.getter(name="maxResultRows") def max_result_rows(self) -> Optional[pulumi.Input[int]]: """ Limits the number of rows in the result. """ return pulumi.get(self, "max_result_rows") @max_result_rows.setter def max_result_rows(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "max_result_rows", value) @property @pulumi.getter(name="maxRowsInDistinct") def max_rows_in_distinct(self) -> Optional[pulumi.Input[int]]: """ Limits the maximum number of different rows when using DISTINCT. """ return pulumi.get(self, "max_rows_in_distinct") @max_rows_in_distinct.setter def max_rows_in_distinct(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "max_rows_in_distinct", value) @property @pulumi.getter(name="maxRowsInJoin") def max_rows_in_join(self) -> Optional[pulumi.Input[int]]: """ Limit on maximum size of the hash table for JOIN, in rows. """ return pulumi.get(self, "max_rows_in_join") @max_rows_in_join.setter def max_rows_in_join(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "max_rows_in_join", value) @property @pulumi.getter(name="maxRowsInSet") def max_rows_in_set(self) -> Optional[pulumi.Input[int]]: """ Limit on the number of rows in the set resulting from the execution of the IN section. """ return pulumi.get(self, "max_rows_in_set") @max_rows_in_set.setter def max_rows_in_set(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "max_rows_in_set", value) @property @pulumi.getter(name="maxRowsToGroupBy") def max_rows_to_group_by(self) -> Optional[pulumi.Input[int]]: """ Limits the maximum number of unique keys received from aggregation function. """ return pulumi.get(self, "max_rows_to_group_by") @max_rows_to_group_by.setter def max_rows_to_group_by(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "max_rows_to_group_by", value) @property @pulumi.getter(name="maxRowsToRead") def max_rows_to_read(self) -> Optional[pulumi.Input[int]]: """ Limits the maximum number of rows that can be read from a table when running a query. """ return pulumi.get(self, "max_rows_to_read") @max_rows_to_read.setter def max_rows_to_read(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "max_rows_to_read", value) @property @pulumi.getter(name="maxRowsToSort") def max_rows_to_sort(self) -> Optional[pulumi.Input[int]]: """ Limits the maximum number of rows that can be read from a table for sorting. """ return pulumi.get(self, "max_rows_to_sort") @max_rows_to_sort.setter def max_rows_to_sort(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "max_rows_to_sort", value) @property @pulumi.getter(name="maxRowsToTransfer") def max_rows_to_transfer(self) -> Optional[pulumi.Input[int]]: """ Limits the maximum number of rows that can be passed to a remote server or saved in a temporary table when using GLOBAL IN. """ return pulumi.get(self, "max_rows_to_transfer") @max_rows_to_transfer.setter def max_rows_to_transfer(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "max_rows_to_transfer", value) @property @pulumi.getter(name="maxTemporaryColumns") def max_temporary_columns(self) -> Optional[pulumi.Input[int]]: """ Limits the maximum number of temporary columns that must be kept in RAM at the same time when running a query, including constant columns. """ return pulumi.get(self, "max_temporary_columns") @max_temporary_columns.setter def max_temporary_columns(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "max_temporary_columns", value) @property @pulumi.getter(name="maxTemporaryNonConstColumns") def max_temporary_non_const_columns(self) -> Optional[pulumi.Input[int]]: """ Limits the maximum number of temporary columns that must be kept in RAM at the same time when running a query, excluding constant columns. """ return pulumi.get(self, "max_temporary_non_const_columns") @max_temporary_non_const_columns.setter def max_temporary_non_const_columns(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "max_temporary_non_const_columns", value) @property @pulumi.getter(name="maxThreads") def max_threads(self) -> Optional[pulumi.Input[int]]: """ The maximum number of query processing threads, excluding threads for retrieving data from remote servers. """ return pulumi.get(self, "max_threads") @max_threads.setter def max_threads(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "max_threads", value) @property @pulumi.getter(name="mergeTreeMaxBytesToUseCache") def merge_tree_max_bytes_to_use_cache(self) -> Optional[pulumi.Input[int]]: """ If ClickHouse should read more than merge_tree_max_bytes_to_use_cache bytes in one query, it doesn’t use the cache of uncompressed blocks. """ return pulumi.get(self, "merge_tree_max_bytes_to_use_cache") @merge_tree_max_bytes_to_use_cache.setter def merge_tree_max_bytes_to_use_cache(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "merge_tree_max_bytes_to_use_cache", value) @property @pulumi.getter(name="mergeTreeMaxRowsToUseCache") def merge_tree_max_rows_to_use_cache(self) -> Optional[pulumi.Input[int]]: """ If ClickHouse should read more than merge_tree_max_rows_to_use_cache rows in one query, it doesn’t use the cache of uncompressed blocks. """ return pulumi.get(self, "merge_tree_max_rows_to_use_cache") @merge_tree_max_rows_to_use_cache.setter def merge_tree_max_rows_to_use_cache(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "merge_tree_max_rows_to_use_cache", value) @property @pulumi.getter(name="mergeTreeMinBytesForConcurrentRead") def merge_tree_min_bytes_for_concurrent_read(self) -> Optional[pulumi.Input[int]]: """ If the number of bytes to read from one file of a MergeTree-engine table exceeds merge_tree_min_bytes_for_concurrent_read, then ClickHouse tries to concurrently read from this file in several threads. """ return pulumi.get(self, "merge_tree_min_bytes_for_concurrent_read") @merge_tree_min_bytes_for_concurrent_read.setter def merge_tree_min_bytes_for_concurrent_read(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "merge_tree_min_bytes_for_concurrent_read", value) @property @pulumi.getter(name="mergeTreeMinRowsForConcurrentRead") def merge_tree_min_rows_for_concurrent_read(self) -> Optional[pulumi.Input[int]]: """ If the number of rows to be read from a file of a MergeTree table exceeds merge_tree_min_rows_for_concurrent_read then ClickHouse tries to perform a concurrent reading from this file on several threads. """ return pulumi.get(self, "merge_tree_min_rows_for_concurrent_read") @merge_tree_min_rows_for_concurrent_read.setter def merge_tree_min_rows_for_concurrent_read(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "merge_tree_min_rows_for_concurrent_read", value) @property @pulumi.getter(name="minBytesToUseDirectIo") def min_bytes_to_use_direct_io(self) -> Optional[pulumi.Input[int]]: """ The minimum data volume required for using direct I/O access to the storage disk. """ return pulumi.get(self, "min_bytes_to_use_direct_io") @min_bytes_to_use_direct_io.setter def min_bytes_to_use_direct_io(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "min_bytes_to_use_direct_io", value) @property @pulumi.getter(name="minCountToCompile") def min_count_to_compile(self) -> Optional[pulumi.Input[int]]: """ How many times to potentially use a compiled chunk of code before running compilation. """ return pulumi.get(self, "min_count_to_compile") @min_count_to_compile.setter def min_count_to_compile(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "min_count_to_compile", value) @property @pulumi.getter(name="minCountToCompileExpression") def min_count_to_compile_expression(self) -> Optional[pulumi.Input[int]]: """ A query waits for expression compilation process to complete prior to continuing execution. """ return pulumi.get(self, "min_count_to_compile_expression") @min_count_to_compile_expression.setter def min_count_to_compile_expression(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "min_count_to_compile_expression", value) @property @pulumi.getter(name="minExecutionSpeed") def min_execution_speed(self) -> Optional[pulumi.Input[int]]: """ Minimal execution speed in rows per second. """ return pulumi.get(self, "min_execution_speed") @min_execution_speed.setter def min_execution_speed(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "min_execution_speed", value) @property @pulumi.getter(name="minExecutionSpeedBytes") def min_execution_speed_bytes(self) -> Optional[pulumi.Input[int]]: """ Minimal execution speed in bytes per second. """ return pulumi.get(self, "min_execution_speed_bytes") @min_execution_speed_bytes.setter def min_execution_speed_bytes(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "min_execution_speed_bytes", value) @property @pulumi.getter(name="minInsertBlockSizeBytes") def min_insert_block_size_bytes(self) -> Optional[pulumi.Input[int]]: """ Sets the minimum number of bytes in the block which can be inserted into a table by an INSERT query. """ return pulumi.get(self, "min_insert_block_size_bytes") @min_insert_block_size_bytes.setter def min_insert_block_size_bytes(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "min_insert_block_size_bytes", value) @property @pulumi.getter(name="minInsertBlockSizeRows") def min_insert_block_size_rows(self) -> Optional[pulumi.Input[int]]: """ Sets the minimum number of rows in the block which can be inserted into a table by an INSERT query. """ return pulumi.get(self, "min_insert_block_size_rows") @min_insert_block_size_rows.setter def min_insert_block_size_rows(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "min_insert_block_size_rows", value) @property @pulumi.getter(name="outputFormatJsonQuote64bitIntegers") def output_format_json_quote64bit_integers(self) -> Optional[pulumi.Input[bool]]: """ If the value is true, integers appear in quotes when using JSON* Int64 and UInt64 formats (for compatibility with most JavaScript implementations); otherwise, integers are output without the quotes. """ return pulumi.get(self, "output_format_json_quote64bit_integers") @output_format_json_quote64bit_integers.setter def output_format_json_quote64bit_integers(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "output_format_json_quote64bit_integers", value) @property @pulumi.getter(name="outputFormatJsonQuoteDenormals") def output_format_json_quote_denormals(self) -> Optional[pulumi.Input[bool]]: """ Enables +nan, -nan, +inf, -inf outputs in JSON output format. """ return pulumi.get(self, "output_format_json_quote_denormals") @output_format_json_quote_denormals.setter def output_format_json_quote_denormals(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "output_format_json_quote_denormals", value) @property @pulumi.getter def priority(self) -> Optional[pulumi.Input[int]]: """ Query priority. """ return pulumi.get(self, "priority") @priority.setter def priority(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "priority", value) @property @pulumi.getter(name="quotaMode") def quota_mode(self) -> Optional[pulumi.Input[str]]: """ Quota accounting mode. """ return pulumi.get(self, "quota_mode") @quota_mode.setter def quota_mode(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "quota_mode", value) @property @pulumi.getter(name="readOverflowMode") def read_overflow_mode(self) -> Optional[pulumi.Input[str]]: """ Sets behaviour on overflow while read. Possible values: """ return pulumi.get(self, "read_overflow_mode") @read_overflow_mode.setter def read_overflow_mode(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "read_overflow_mode", value) @property @pulumi.getter def readonly(self) -> Optional[pulumi.Input[int]]: """ Restricts permissions for reading data, write data and change settings queries. """ return pulumi.get(self, "readonly") @readonly.setter def readonly(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "readonly", value) @property @pulumi.getter(name="receiveTimeout") def receive_timeout(self) -> Optional[pulumi.Input[int]]: """ Receive timeout in milliseconds on the socket used for communicating with the client. """ return pulumi.get(self, "receive_timeout") @receive_timeout.setter def receive_timeout(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "receive_timeout", value) @property @pulumi.getter(name="replicationAlterPartitionsSync") def replication_alter_partitions_sync(self) -> Optional[pulumi.Input[int]]: """ For ALTER ... ATTACH|DETACH|DROP queries, you can use the replication_alter_partitions_sync setting to set up waiting. """ return pulumi.get(self, "replication_alter_partitions_sync") @replication_alter_partitions_sync.setter def replication_alter_partitions_sync(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "replication_alter_partitions_sync", value) @property @pulumi.getter(name="resultOverflowMode") def result_overflow_mode(self) -> Optional[pulumi.Input[str]]: """ Sets behaviour on overflow in result. Possible values: """ return pulumi.get(self, "result_overflow_mode") @result_overflow_mode.setter def result_overflow_mode(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "result_overflow_mode", value) @property @pulumi.getter(name="selectSequentialConsistency") def select_sequential_consistency(self) -> Optional[pulumi.Input[bool]]: """ Enables or disables sequential consistency for SELECT queries. """ return pulumi.get(self, "select_sequential_consistency") @select_sequential_consistency.setter def select_sequential_consistency(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "select_sequential_consistency", value) @property @pulumi.getter(name="sendProgressInHttpHeaders") def send_progress_in_http_headers(self) -> Optional[pulumi.Input[bool]]: """ Enables or disables X-ClickHouse-Progress HTTP response headers in clickhouse-server responses. """ return pulumi.get(self, "send_progress_in_http_headers") @send_progress_in_http_headers.setter def send_progress_in_http_headers(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "send_progress_in_http_headers", value) @property @pulumi.getter(name="sendTimeout") def send_timeout(self) -> Optional[pulumi.Input[int]]: """ Send timeout in milliseconds on the socket used for communicating with the client. """ return pulumi.get(self, "send_timeout") @send_timeout.setter def send_timeout(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "send_timeout", value) @property @pulumi.getter(name="setOverflowMode") def set_overflow_mode(self) -> Optional[pulumi.Input[str]]: """ Sets behaviour on overflow in the set resulting. Possible values: """ return pulumi.get(self, "set_overflow_mode") @set_overflow_mode.setter def set_overflow_mode(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "set_overflow_mode", value) @property @pulumi.getter(name="skipUnavailableShards") def skip_unavailable_shards(self) -> Optional[pulumi.Input[bool]]: """ Enables or disables silently skipping of unavailable shards. """ return pulumi.get(self, "skip_unavailable_shards") @skip_unavailable_shards.setter def skip_unavailable_shards(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "skip_unavailable_shards", value) @property @pulumi.getter(name="sortOverflowMode") def sort_overflow_mode(self) -> Optional[pulumi.Input[str]]: """ Sets behaviour on overflow while sort. Possible values: """ return pulumi.get(self, "sort_overflow_mode") @sort_overflow_mode.setter def sort_overflow_mode(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "sort_overflow_mode", value) @property @pulumi.getter(name="timeoutOverflowMode") def timeout_overflow_mode(self) -> Optional[pulumi.Input[str]]: """ Sets behaviour on overflow. Possible values: """ return pulumi.get(self, "timeout_overflow_mode") @timeout_overflow_mode.setter def timeout_overflow_mode(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "timeout_overflow_mode", value) @property @pulumi.getter(name="transferOverflowMode") def transfer_overflow_mode(self) -> Optional[pulumi.Input[str]]: """ Sets behaviour on overflow. Possible values: """ return pulumi.get(self, "transfer_overflow_mode") @transfer_overflow_mode.setter def transfer_overflow_mode(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "transfer_overflow_mode", value) @property @pulumi.getter(name="transformNullIn") def transform_null_in(self) -> Optional[pulumi.Input[bool]]: """ Enables equality of NULL values for IN operator. """ return pulumi.get(self, "transform_null_in") @transform_null_in.setter def transform_null_in(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "transform_null_in", value) @property @pulumi.getter(name="useUncompressedCache") def use_uncompressed_cache(self) -> Optional[pulumi.Input[bool]]: """ Whether to use a cache of uncompressed blocks. """ return pulumi.get(self, "use_uncompressed_cache") @use_uncompressed_cache.setter def use_uncompressed_cache(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "use_uncompressed_cache", value) @pulumi.input_type class MdbClickhouseClusterZookeeperArgs: def __init__(__self__, *, resources: Optional[pulumi.Input['MdbClickhouseClusterZookeeperResourcesArgs']] = None): """ :param pulumi.Input['MdbClickhouseClusterZookeeperResourcesArgs'] resources: Resources allocated to hosts of the ZooKeeper subcluster. The structure is documented below. """ if resources is not None: pulumi.set(__self__, "resources", resources) @property @pulumi.getter def resources(self) -> Optional[pulumi.Input['MdbClickhouseClusterZookeeperResourcesArgs']]: """ Resources allocated to hosts of the ZooKeeper subcluster. The structure is documented below. """ return pulumi.get(self, "resources") @resources.setter def resources(self, value: Optional[pulumi.Input['MdbClickhouseClusterZookeeperResourcesArgs']]): pulumi.set(self, "resources", value) @pulumi.input_type class MdbClickhouseClusterZookeeperResourcesArgs: def __init__(__self__, *, disk_size: Optional[pulumi.Input[int]] = None, disk_type_id: Optional[pulumi.Input[str]] = None, resource_preset_id: Optional[pulumi.Input[str]] = None): """ :param pulumi.Input[int] disk_size: Volume of the storage available to a ZooKeeper host, in gigabytes. :param pulumi.Input[str] disk_type_id: Type of the storage of ZooKeeper hosts. For more information see [the official documentation](https://cloud.yandex.com/docs/managed-clickhouse/concepts/storage). """ if disk_size is not None: pulumi.set(__self__, "disk_size", disk_size) if disk_type_id is not None: pulumi.set(__self__, "disk_type_id", disk_type_id) if resource_preset_id is not None: pulumi.set(__self__, "resource_preset_id", resource_preset_id) @property @pulumi.getter(name="diskSize") def disk_size(self) -> Optional[pulumi.Input[int]]: """ Volume of the storage available to a ZooKeeper host, in gigabytes. """ return pulumi.get(self, "disk_size") @disk_size.setter def disk_size(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "disk_size", value) @property @pulumi.getter(name="diskTypeId") def disk_type_id(self) -> Optional[pulumi.Input[str]]: """ Type of the storage of ZooKeeper hosts. For more information see [the official documentation](https://cloud.yandex.com/docs/managed-clickhouse/concepts/storage). """ return pulumi.get(self, "disk_type_id") @disk_type_id.setter def disk_type_id(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "disk_type_id", value) @property @pulumi.getter(name="resourcePresetId") def resource_preset_id(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "resource_preset_id") @resource_preset_id.setter def resource_preset_id(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "resource_preset_id", value) @pulumi.input_type class MdbElasticSearchClusterConfigArgs: def __init__(__self__, *, admin_password: pulumi.Input[str], data_node: pulumi.Input['MdbElasticSearchClusterConfigDataNodeArgs'], edition: Optional[pulumi.Input[str]] = None, master_node: Optional[pulumi.Input['MdbElasticSearchClusterConfigMasterNodeArgs']] = None, plugins: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, version: Optional[pulumi.Input[str]] = None): """ :param pulumi.Input[str] admin_password: Password for admin user of Elasticsearch. :param pulumi.Input['MdbElasticSearchClusterConfigDataNodeArgs'] data_node: Configuration for Elasticsearch data nodes subcluster. The structure is documented below. :param pulumi.Input[str] edition: Edition of Elasticsearch. For more information, see [the official documentation](https://cloud.yandex.com/en-ru/docs/managed-elasticsearch/concepts/es-editions). :param pulumi.Input['MdbElasticSearchClusterConfigMasterNodeArgs'] master_node: Configuration for Elasticsearch master nodes subcluster. The structure is documented below. :param pulumi.Input[Sequence[pulumi.Input[str]]] plugins: A set of Elasticsearch plugins to install. :param pulumi.Input[str] version: Version of Elasticsearch. """ pulumi.set(__self__, "admin_password", admin_password) pulumi.set(__self__, "data_node", data_node) if edition is not None: pulumi.set(__self__, "edition", edition) if master_node is not None: pulumi.set(__self__, "master_node", master_node) if plugins is not None: pulumi.set(__self__, "plugins", plugins) if version is not None: pulumi.set(__self__, "version", version) @property @pulumi.getter(name="adminPassword") def admin_password(self) -> pulumi.Input[str]: """ Password for admin user of Elasticsearch. """ return pulumi.get(self, "admin_password") @admin_password.setter def admin_password(self, value: pulumi.Input[str]): pulumi.set(self, "admin_password", value) @property @pulumi.getter(name="dataNode") def data_node(self) -> pulumi.Input['MdbElasticSearchClusterConfigDataNodeArgs']: """ Configuration for Elasticsearch data nodes subcluster. The structure is documented below. """ return pulumi.get(self, "data_node") @data_node.setter def data_node(self, value: pulumi.Input['MdbElasticSearchClusterConfigDataNodeArgs']): pulumi.set(self, "data_node", value) @property @pulumi.getter def edition(self) -> Optional[pulumi.Input[str]]: """ Edition of Elasticsearch. For more information, see [the official documentation](https://cloud.yandex.com/en-ru/docs/managed-elasticsearch/concepts/es-editions). """ return pulumi.get(self, "edition") @edition.setter def edition(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "edition", value) @property @pulumi.getter(name="masterNode") def master_node(self) -> Optional[pulumi.Input['MdbElasticSearchClusterConfigMasterNodeArgs']]: """ Configuration for Elasticsearch master nodes subcluster. The structure is documented below. """ return pulumi.get(self, "master_node") @master_node.setter def master_node(self, value: Optional[pulumi.Input['MdbElasticSearchClusterConfigMasterNodeArgs']]): pulumi.set(self, "master_node", value) @property @pulumi.getter def plugins(self) -> Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]: """ A set of Elasticsearch plugins to install. """ return pulumi.get(self, "plugins") @plugins.setter def plugins(self, value: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]): pulumi.set(self, "plugins", value) @property @pulumi.getter def version(self) -> Optional[pulumi.Input[str]]: """ Version of Elasticsearch. """ return pulumi.get(self, "version") @version.setter def version(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "version", value) @pulumi.input_type class MdbElasticSearchClusterConfigDataNodeArgs: def __init__(__self__, *, resources: pulumi.Input['MdbElasticSearchClusterConfigDataNodeResourcesArgs']): """ :param pulumi.Input['MdbElasticSearchClusterConfigDataNodeResourcesArgs'] resources: Resources allocated to hosts of the Elasticsearch master nodes subcluster. The structure is documented below. """ pulumi.set(__self__, "resources", resources) @property @pulumi.getter def resources(self) -> pulumi.Input['MdbElasticSearchClusterConfigDataNodeResourcesArgs']: """ Resources allocated to hosts of the Elasticsearch master nodes subcluster. The structure is documented below. """ return pulumi.get(self, "resources") @resources.setter def resources(self, value: pulumi.Input['MdbElasticSearchClusterConfigDataNodeResourcesArgs']): pulumi.set(self, "resources", value) @pulumi.input_type class MdbElasticSearchClusterConfigDataNodeResourcesArgs: def __init__(__self__, *, disk_size: pulumi.Input[int], disk_type_id: pulumi.Input[str], resource_preset_id: pulumi.Input[str]): """ :param pulumi.Input[int] disk_size: Volume of the storage available to a host, in gigabytes. :param pulumi.Input[str] disk_type_id: Type of the storage of Elasticsearch hosts. """ pulumi.set(__self__, "disk_size", disk_size) pulumi.set(__self__, "disk_type_id", disk_type_id) pulumi.set(__self__, "resource_preset_id", resource_preset_id) @property @pulumi.getter(name="diskSize") def disk_size(self) -> pulumi.Input[int]: """ Volume of the storage available to a host, in gigabytes. """ return pulumi.get(self, "disk_size") @disk_size.setter def disk_size(self, value: pulumi.Input[int]): pulumi.set(self, "disk_size", value) @property @pulumi.getter(name="diskTypeId") def disk_type_id(self) -> pulumi.Input[str]: """ Type of the storage of Elasticsearch hosts. """ return pulumi.get(self, "disk_type_id") @disk_type_id.setter def disk_type_id(self, value: pulumi.Input[str]): pulumi.set(self, "disk_type_id", value) @property @pulumi.getter(name="resourcePresetId") def resource_preset_id(self) -> pulumi.Input[str]: return pulumi.get(self, "resource_preset_id") @resource_preset_id.setter def resource_preset_id(self, value: pulumi.Input[str]): pulumi.set(self, "resource_preset_id", value) @pulumi.input_type class MdbElasticSearchClusterConfigMasterNodeArgs: def __init__(__self__, *, resources: pulumi.Input['MdbElasticSearchClusterConfigMasterNodeResourcesArgs']): """ :param pulumi.Input['MdbElasticSearchClusterConfigMasterNodeResourcesArgs'] resources: Resources allocated to hosts of the Elasticsearch master nodes subcluster. The structure is documented below. """ pulumi.set(__self__, "resources", resources) @property @pulumi.getter def resources(self) -> pulumi.Input['MdbElasticSearchClusterConfigMasterNodeResourcesArgs']: """ Resources allocated to hosts of the Elasticsearch master nodes subcluster. The structure is documented below. """ return pulumi.get(self, "resources") @resources.setter def resources(self, value: pulumi.Input['MdbElasticSearchClusterConfigMasterNodeResourcesArgs']): pulumi.set(self, "resources", value) @pulumi.input_type class MdbElasticSearchClusterConfigMasterNodeResourcesArgs: def __init__(__self__, *, disk_size: pulumi.Input[int], disk_type_id: pulumi.Input[str], resource_preset_id: pulumi.Input[str]): """ :param pulumi.Input[int] disk_size: Volume of the storage available to a host, in gigabytes. :param pulumi.Input[str] disk_type_id: Type of the storage of Elasticsearch hosts. """ pulumi.set(__self__, "disk_size", disk_size) pulumi.set(__self__, "disk_type_id", disk_type_id) pulumi.set(__self__, "resource_preset_id", resource_preset_id) @property @pulumi.getter(name="diskSize") def disk_size(self) -> pulumi.Input[int]: """ Volume of the storage available to a host, in gigabytes. """ return pulumi.get(self, "disk_size") @disk_size.setter def disk_size(self, value: pulumi.Input[int]): pulumi.set(self, "disk_size", value) @property @pulumi.getter(name="diskTypeId") def disk_type_id(self) -> pulumi.Input[str]: """ Type of the storage of Elasticsearch hosts. """ return pulumi.get(self, "disk_type_id") @disk_type_id.setter def disk_type_id(self, value: pulumi.Input[str]): pulumi.set(self, "disk_type_id", value) @property @pulumi.getter(name="resourcePresetId") def resource_preset_id(self) -> pulumi.Input[str]: return pulumi.get(self, "resource_preset_id") @resource_preset_id.setter def resource_preset_id(self, value: pulumi.Input[str]): pulumi.set(self, "resource_preset_id", value) @pulumi.input_type class MdbElasticSearchClusterHostArgs: def __init__(__self__, *, name: pulumi.Input[str], type: pulumi.Input[str], zone: pulumi.Input[str], assign_public_ip: Optional[pulumi.Input[bool]] = None, fqdn: Optional[pulumi.Input[str]] = None, subnet_id: Optional[pulumi.Input[str]] = None): """ :param pulumi.Input[str] name: User defined host name. :param pulumi.Input[str] type: The type of the host to be deployed. Can be either `DATA_NODE` or `MASTER_NODE`. :param pulumi.Input[str] zone: The availability zone where the Elasticsearch host will be created. For more information see [the official documentation](https://cloud.yandex.com/docs/overview/concepts/geo-scope). :param pulumi.Input[bool] assign_public_ip: Sets whether the host should get a public IP address on creation. Can be either `true` or `false`. :param pulumi.Input[str] fqdn: The fully qualified domain name of the host. :param pulumi.Input[str] subnet_id: The ID of the subnet, to which the host belongs. The subnet must be a part of the network to which the cluster belongs. """ pulumi.set(__self__, "name", name) pulumi.set(__self__, "type", type) pulumi.set(__self__, "zone", zone) if assign_public_ip is not None: pulumi.set(__self__, "assign_public_ip", assign_public_ip) if fqdn is not None: pulumi.set(__self__, "fqdn", fqdn) if subnet_id is not None: pulumi.set(__self__, "subnet_id", subnet_id) @property @pulumi.getter def name(self) -> pulumi.Input[str]: """ User defined host name. """ return pulumi.get(self, "name") @name.setter def name(self, value: pulumi.Input[str]): pulumi.set(self, "name", value) @property @pulumi.getter def type(self) -> pulumi.Input[str]: """ The type of the host to be deployed. Can be either `DATA_NODE` or `MASTER_NODE`. """ return pulumi.get(self, "type") @type.setter def type(self, value: pulumi.Input[str]): pulumi.set(self, "type", value) @property @pulumi.getter def zone(self) -> pulumi.Input[str]: """ The availability zone where the Elasticsearch host will be created. For more information see [the official documentation](https://cloud.yandex.com/docs/overview/concepts/geo-scope). """ return pulumi.get(self, "zone") @zone.setter def zone(self, value: pulumi.Input[str]): pulumi.set(self, "zone", value) @property @pulumi.getter(name="assignPublicIp") def assign_public_ip(self) -> Optional[pulumi.Input[bool]]: """ Sets whether the host should get a public IP address on creation. Can be either `true` or `false`. """ return pulumi.get(self, "assign_public_ip") @assign_public_ip.setter def assign_public_ip(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "assign_public_ip", value) @property @pulumi.getter def fqdn(self) -> Optional[pulumi.Input[str]]: """ The fully qualified domain name of the host. """ return pulumi.get(self, "fqdn") @fqdn.setter def fqdn(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "fqdn", value) @property @pulumi.getter(name="subnetId") def subnet_id(self) -> Optional[pulumi.Input[str]]: """ The ID of the subnet, to which the host belongs. The subnet must be a part of the network to which the cluster belongs. """ return pulumi.get(self, "subnet_id") @subnet_id.setter def subnet_id(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "subnet_id", value) @pulumi.input_type class MdbGreenplumClusterAccessArgs: def __init__(__self__, *, data_lens: Optional[pulumi.Input[bool]] = None, web_sql: Optional[pulumi.Input[bool]] = None): """ :param pulumi.Input[bool] data_lens: Allow access for [Yandex DataLens](https://cloud.yandex.com/services/datalens). :param pulumi.Input[bool] web_sql: Allows access for SQL queries in the management console """ if data_lens is not None: pulumi.set(__self__, "data_lens", data_lens) if web_sql is not None: pulumi.set(__self__, "web_sql", web_sql) @property @pulumi.getter(name="dataLens") def data_lens(self) -> Optional[pulumi.Input[bool]]: """ Allow access for [Yandex DataLens](https://cloud.yandex.com/services/datalens). """ return pulumi.get(self, "data_lens") @data_lens.setter def data_lens(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "data_lens", value) @property @pulumi.getter(name="webSql") def web_sql(self) -> Optional[pulumi.Input[bool]]: """ Allows access for SQL queries in the management console """ return pulumi.get(self, "web_sql") @web_sql.setter def web_sql(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "web_sql", value) @pulumi.input_type class MdbGreenplumClusterBackupWindowStartArgs: def __init__(__self__, *, hours: Optional[pulumi.Input[int]] = None, minutes: Optional[pulumi.Input[int]] = None): """ :param pulumi.Input[int] hours: The hour at which backup will be started (UTC). :param pulumi.Input[int] minutes: The minute at which backup will be started (UTC). """ if hours is not None: pulumi.set(__self__, "hours", hours) if minutes is not None: pulumi.set(__self__, "minutes", minutes) @property @pulumi.getter def hours(self) -> Optional[pulumi.Input[int]]: """ The hour at which backup will be started (UTC). """ return pulumi.get(self, "hours") @hours.setter def hours(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "hours", value) @property @pulumi.getter def minutes(self) -> Optional[pulumi.Input[int]]: """ The minute at which backup will be started (UTC). """ return pulumi.get(self, "minutes") @minutes.setter def minutes(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "minutes", value) @pulumi.input_type class MdbGreenplumClusterMasterHostArgs: def __init__(__self__, *, assign_public_ip: Optional[pulumi.Input[bool]] = None, fqdn: Optional[pulumi.Input[str]] = None): """ :param pulumi.Input[bool] assign_public_ip: Sets whether the master hosts should get a public IP address on creation. Changing this parameter for an existing host is not supported at the moment. :param pulumi.Input[str] fqdn: (Computed) The fully qualified domain name of the host. """ if assign_public_ip is not None: pulumi.set(__self__, "assign_public_ip", assign_public_ip) if fqdn is not None: pulumi.set(__self__, "fqdn", fqdn) @property @pulumi.getter(name="assignPublicIp") def assign_public_ip(self) -> Optional[pulumi.Input[bool]]: """ Sets whether the master hosts should get a public IP address on creation. Changing this parameter for an existing host is not supported at the moment. """ return pulumi.get(self, "assign_public_ip") @assign_public_ip.setter def assign_public_ip(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "assign_public_ip", value) @property @pulumi.getter def fqdn(self) -> Optional[pulumi.Input[str]]: """ (Computed) The fully qualified domain name of the host. """ return pulumi.get(self, "fqdn") @fqdn.setter def fqdn(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "fqdn", value) @pulumi.input_type class MdbGreenplumClusterMasterSubclusterArgs: def __init__(__self__, *, resources: pulumi.Input['MdbGreenplumClusterMasterSubclusterResourcesArgs']): """ :param pulumi.Input['MdbGreenplumClusterMasterSubclusterResourcesArgs'] resources: Resources allocated to hosts for segment subcluster of the Greenplum cluster. The structure is documented below. """ pulumi.set(__self__, "resources", resources) @property @pulumi.getter def resources(self) -> pulumi.Input['MdbGreenplumClusterMasterSubclusterResourcesArgs']: """ Resources allocated to hosts for segment subcluster of the Greenplum cluster. The structure is documented below. """ return pulumi.get(self, "resources") @resources.setter def resources(self, value: pulumi.Input['MdbGreenplumClusterMasterSubclusterResourcesArgs']): pulumi.set(self, "resources", value) @pulumi.input_type class MdbGreenplumClusterMasterSubclusterResourcesArgs: def __init__(__self__, *, disk_size: pulumi.Input[int], disk_type_id: pulumi.Input[str], resource_preset_id: pulumi.Input[str]): pulumi.set(__self__, "disk_size", disk_size) pulumi.set(__self__, "disk_type_id", disk_type_id) pulumi.set(__self__, "resource_preset_id", resource_preset_id) @property @pulumi.getter(name="diskSize") def disk_size(self) -> pulumi.Input[int]: return pulumi.get(self, "disk_size") @disk_size.setter def disk_size(self, value: pulumi.Input[int]): pulumi.set(self, "disk_size", value) @property @pulumi.getter(name="diskTypeId") def disk_type_id(self) -> pulumi.Input[str]: return pulumi.get(self, "disk_type_id") @disk_type_id.setter def disk_type_id(self, value: pulumi.Input[str]): pulumi.set(self, "disk_type_id", value) @property @pulumi.getter(name="resourcePresetId") def resource_preset_id(self) -> pulumi.Input[str]: return pulumi.get(self, "resource_preset_id") @resource_preset_id.setter def resource_preset_id(self, value: pulumi.Input[str]): pulumi.set(self, "resource_preset_id", value) @pulumi.input_type class MdbGreenplumClusterSegmentHostArgs: def __init__(__self__, *, fqdn: Optional[pulumi.Input[str]] = None): """ :param pulumi.Input[str] fqdn: (Computed) The fully qualified domain name of the host. """ if fqdn is not None: pulumi.set(__self__, "fqdn", fqdn) @property @pulumi.getter def fqdn(self) -> Optional[pulumi.Input[str]]: """ (Computed) The fully qualified domain name of the host. """ return pulumi.get(self, "fqdn") @fqdn.setter def fqdn(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "fqdn", value) @pulumi.input_type class MdbGreenplumClusterSegmentSubclusterArgs: def __init__(__self__, *, resources: pulumi.Input['MdbGreenplumClusterSegmentSubclusterResourcesArgs']): """ :param pulumi.Input['MdbGreenplumClusterSegmentSubclusterResourcesArgs'] resources: Resources allocated to hosts for segment subcluster of the Greenplum cluster. The structure is documented below. """ pulumi.set(__self__, "resources", resources) @property @pulumi.getter def resources(self) -> pulumi.Input['MdbGreenplumClusterSegmentSubclusterResourcesArgs']: """ Resources allocated to hosts for segment subcluster of the Greenplum cluster. The structure is documented below. """ return pulumi.get(self, "resources") @resources.setter def resources(self, value: pulumi.Input['MdbGreenplumClusterSegmentSubclusterResourcesArgs']): pulumi.set(self, "resources", value) @pulumi.input_type class MdbGreenplumClusterSegmentSubclusterResourcesArgs: def __init__(__self__, *, disk_size: pulumi.Input[int], disk_type_id: pulumi.Input[str], resource_preset_id: pulumi.Input[str]): pulumi.set(__self__, "disk_size", disk_size) pulumi.set(__self__, "disk_type_id", disk_type_id) pulumi.set(__self__, "resource_preset_id", resource_preset_id) @property @pulumi.getter(name="diskSize") def disk_size(self) -> pulumi.Input[int]: return pulumi.get(self, "disk_size") @disk_size.setter def disk_size(self, value: pulumi.Input[int]): pulumi.set(self, "disk_size", value) @property @pulumi.getter(name="diskTypeId") def disk_type_id(self) -> pulumi.Input[str]: return pulumi.get(self, "disk_type_id") @disk_type_id.setter def disk_type_id(self, value: pulumi.Input[str]): pulumi.set(self, "disk_type_id", value) @property @pulumi.getter(name="resourcePresetId") def resource_preset_id(self) -> pulumi.Input[str]: return pulumi.get(self, "resource_preset_id") @resource_preset_id.setter def resource_preset_id(self, value: pulumi.Input[str]): pulumi.set(self, "resource_preset_id", value) @pulumi.input_type class MdbKafkaClusterConfigArgs: def __init__(__self__, *, kafka: pulumi.Input['MdbKafkaClusterConfigKafkaArgs'], version: pulumi.Input[str], zones: pulumi.Input[Sequence[pulumi.Input[str]]], assign_public_ip: Optional[pulumi.Input[bool]] = None, brokers_count: Optional[pulumi.Input[int]] = None, schema_registry: Optional[pulumi.Input[bool]] = None, unmanaged_topics: Optional[pulumi.Input[bool]] = None, zookeeper: Optional[pulumi.Input['MdbKafkaClusterConfigZookeeperArgs']] = None): """ :param pulumi.Input['MdbKafkaClusterConfigKafkaArgs'] kafka: Configuration of the Kafka subcluster. The structure is documented below. :param pulumi.Input[str] version: Version of the Kafka server software. :param pulumi.Input[Sequence[pulumi.Input[str]]] zones: List of availability zones. :param pulumi.Input[bool] assign_public_ip: Determines whether each broker will be assigned a public IP address. The default is `false`. :param pulumi.Input[int] brokers_count: Count of brokers per availability zone. The default is `1`. :param pulumi.Input[bool] schema_registry: Enables managed schema registry on cluster. The default is `false`. :param pulumi.Input[bool] unmanaged_topics: Allows to use Kafka AdminAPI to manage topics. The default is `false`. :param pulumi.Input['MdbKafkaClusterConfigZookeeperArgs'] zookeeper: Configuration of the ZooKeeper subcluster. The structure is documented below. """ pulumi.set(__self__, "kafka", kafka) pulumi.set(__self__, "version", version) pulumi.set(__self__, "zones", zones) if assign_public_ip is not None: pulumi.set(__self__, "assign_public_ip", assign_public_ip) if brokers_count is not None: pulumi.set(__self__, "brokers_count", brokers_count) if schema_registry is not None: pulumi.set(__self__, "schema_registry", schema_registry) if unmanaged_topics is not None: pulumi.set(__self__, "unmanaged_topics", unmanaged_topics) if zookeeper is not None: pulumi.set(__self__, "zookeeper", zookeeper) @property @pulumi.getter def kafka(self) -> pulumi.Input['MdbKafkaClusterConfigKafkaArgs']: """ Configuration of the Kafka subcluster. The structure is documented below. """ return pulumi.get(self, "kafka") @kafka.setter def kafka(self, value: pulumi.Input['MdbKafkaClusterConfigKafkaArgs']): pulumi.set(self, "kafka", value) @property @pulumi.getter def version(self) -> pulumi.Input[str]: """ Version of the Kafka server software. """ return pulumi.get(self, "version") @version.setter def version(self, value: pulumi.Input[str]): pulumi.set(self, "version", value) @property @pulumi.getter def zones(self) -> pulumi.Input[Sequence[pulumi.Input[str]]]: """ List of availability zones. """ return pulumi.get(self, "zones") @zones.setter def zones(self, value: pulumi.Input[Sequence[pulumi.Input[str]]]): pulumi.set(self, "zones", value) @property @pulumi.getter(name="assignPublicIp") def assign_public_ip(self) -> Optional[pulumi.Input[bool]]: """ Determines whether each broker will be assigned a public IP address. The default is `false`. """ return pulumi.get(self, "assign_public_ip") @assign_public_ip.setter def assign_public_ip(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "assign_public_ip", value) @property @pulumi.getter(name="brokersCount") def brokers_count(self) -> Optional[pulumi.Input[int]]: """ Count of brokers per availability zone. The default is `1`. """ return pulumi.get(self, "brokers_count") @brokers_count.setter def brokers_count(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "brokers_count", value) @property @pulumi.getter(name="schemaRegistry") def schema_registry(self) -> Optional[pulumi.Input[bool]]: """ Enables managed schema registry on cluster. The default is `false`. """ return pulumi.get(self, "schema_registry") @schema_registry.setter def schema_registry(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "schema_registry", value) @property @pulumi.getter(name="unmanagedTopics") def unmanaged_topics(self) -> Optional[pulumi.Input[bool]]: """ Allows to use Kafka AdminAPI to manage topics. The default is `false`. """ return pulumi.get(self, "unmanaged_topics") @unmanaged_topics.setter def unmanaged_topics(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "unmanaged_topics", value) @property @pulumi.getter def zookeeper(self) -> Optional[pulumi.Input['MdbKafkaClusterConfigZookeeperArgs']]: """ Configuration of the ZooKeeper subcluster. The structure is documented below. """ return pulumi.get(self, "zookeeper") @zookeeper.setter def zookeeper(self, value: Optional[pulumi.Input['MdbKafkaClusterConfigZookeeperArgs']]): pulumi.set(self, "zookeeper", value) @pulumi.input_type class MdbKafkaClusterConfigKafkaArgs: def __init__(__self__, *, resources: pulumi.Input['MdbKafkaClusterConfigKafkaResourcesArgs'], kafka_config: Optional[pulumi.Input['MdbKafkaClusterConfigKafkaKafkaConfigArgs']] = None): """ :param pulumi.Input['MdbKafkaClusterConfigKafkaResourcesArgs'] resources: Resources allocated to hosts of the ZooKeeper subcluster. The structure is documented below. :param pulumi.Input['MdbKafkaClusterConfigKafkaKafkaConfigArgs'] kafka_config: User-defined settings for the Kafka cluster. The structure is documented below. """ pulumi.set(__self__, "resources", resources) if kafka_config is not None: pulumi.set(__self__, "kafka_config", kafka_config) @property @pulumi.getter def resources(self) -> pulumi.Input['MdbKafkaClusterConfigKafkaResourcesArgs']: """ Resources allocated to hosts of the ZooKeeper subcluster. The structure is documented below. """ return pulumi.get(self, "resources") @resources.setter def resources(self, value: pulumi.Input['MdbKafkaClusterConfigKafkaResourcesArgs']): pulumi.set(self, "resources", value) @property @pulumi.getter(name="kafkaConfig") def kafka_config(self) -> Optional[pulumi.Input['MdbKafkaClusterConfigKafkaKafkaConfigArgs']]: """ User-defined settings for the Kafka cluster. The structure is documented below. """ return pulumi.get(self, "kafka_config") @kafka_config.setter def kafka_config(self, value: Optional[pulumi.Input['MdbKafkaClusterConfigKafkaKafkaConfigArgs']]): pulumi.set(self, "kafka_config", value) @pulumi.input_type class MdbKafkaClusterConfigKafkaKafkaConfigArgs: def __init__(__self__, *, auto_create_topics_enable: Optional[pulumi.Input[bool]] = None, compression_type: Optional[pulumi.Input[str]] = None, default_replication_factor: Optional[pulumi.Input[str]] = None, log_flush_interval_messages: Optional[pulumi.Input[str]] = None, log_flush_interval_ms: Optional[pulumi.Input[str]] = None, log_flush_scheduler_interval_ms: Optional[pulumi.Input[str]] = None, log_preallocate: Optional[pulumi.Input[bool]] = None, log_retention_bytes: Optional[pulumi.Input[str]] = None, log_retention_hours: Optional[pulumi.Input[str]] = None, log_retention_minutes: Optional[pulumi.Input[str]] = None, log_retention_ms: Optional[pulumi.Input[str]] = None, log_segment_bytes: Optional[pulumi.Input[str]] = None, num_partitions: Optional[pulumi.Input[str]] = None, socket_receive_buffer_bytes: Optional[pulumi.Input[str]] = None, socket_send_buffer_bytes: Optional[pulumi.Input[str]] = None): if auto_create_topics_enable is not None: pulumi.set(__self__, "auto_create_topics_enable", auto_create_topics_enable) if compression_type is not None: pulumi.set(__self__, "compression_type", compression_type) if default_replication_factor is not None: pulumi.set(__self__, "default_replication_factor", default_replication_factor) if log_flush_interval_messages is not None: pulumi.set(__self__, "log_flush_interval_messages", log_flush_interval_messages) if log_flush_interval_ms is not None: pulumi.set(__self__, "log_flush_interval_ms", log_flush_interval_ms) if log_flush_scheduler_interval_ms is not None: pulumi.set(__self__, "log_flush_scheduler_interval_ms", log_flush_scheduler_interval_ms) if log_preallocate is not None: pulumi.set(__self__, "log_preallocate", log_preallocate) if log_retention_bytes is not None: pulumi.set(__self__, "log_retention_bytes", log_retention_bytes) if log_retention_hours is not None: pulumi.set(__self__, "log_retention_hours", log_retention_hours) if log_retention_minutes is not None: pulumi.set(__self__, "log_retention_minutes", log_retention_minutes) if log_retention_ms is not None: pulumi.set(__self__, "log_retention_ms", log_retention_ms) if log_segment_bytes is not None: pulumi.set(__self__, "log_segment_bytes", log_segment_bytes) if num_partitions is not None: pulumi.set(__self__, "num_partitions", num_partitions) if socket_receive_buffer_bytes is not None: pulumi.set(__self__, "socket_receive_buffer_bytes", socket_receive_buffer_bytes) if socket_send_buffer_bytes is not None: pulumi.set(__self__, "socket_send_buffer_bytes", socket_send_buffer_bytes) @property @pulumi.getter(name="autoCreateTopicsEnable") def auto_create_topics_enable(self) -> Optional[pulumi.Input[bool]]: return pulumi.get(self, "auto_create_topics_enable") @auto_create_topics_enable.setter def auto_create_topics_enable(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "auto_create_topics_enable", value) @property @pulumi.getter(name="compressionType") def compression_type(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "compression_type") @compression_type.setter def compression_type(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "compression_type", value) @property @pulumi.getter(name="defaultReplicationFactor") def default_replication_factor(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "default_replication_factor") @default_replication_factor.setter def default_replication_factor(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "default_replication_factor", value) @property @pulumi.getter(name="logFlushIntervalMessages") def log_flush_interval_messages(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "log_flush_interval_messages") @log_flush_interval_messages.setter def log_flush_interval_messages(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "log_flush_interval_messages", value) @property @pulumi.getter(name="logFlushIntervalMs") def log_flush_interval_ms(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "log_flush_interval_ms") @log_flush_interval_ms.setter def log_flush_interval_ms(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "log_flush_interval_ms", value) @property @pulumi.getter(name="logFlushSchedulerIntervalMs") def log_flush_scheduler_interval_ms(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "log_flush_scheduler_interval_ms") @log_flush_scheduler_interval_ms.setter def log_flush_scheduler_interval_ms(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "log_flush_scheduler_interval_ms", value) @property @pulumi.getter(name="logPreallocate") def log_preallocate(self) -> Optional[pulumi.Input[bool]]: return pulumi.get(self, "log_preallocate") @log_preallocate.setter def log_preallocate(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "log_preallocate", value) @property @pulumi.getter(name="logRetentionBytes") def log_retention_bytes(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "log_retention_bytes") @log_retention_bytes.setter def log_retention_bytes(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "log_retention_bytes", value) @property @pulumi.getter(name="logRetentionHours") def log_retention_hours(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "log_retention_hours") @log_retention_hours.setter def log_retention_hours(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "log_retention_hours", value) @property @pulumi.getter(name="logRetentionMinutes") def log_retention_minutes(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "log_retention_minutes") @log_retention_minutes.setter def log_retention_minutes(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "log_retention_minutes", value) @property @pulumi.getter(name="logRetentionMs") def log_retention_ms(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "log_retention_ms") @log_retention_ms.setter def log_retention_ms(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "log_retention_ms", value) @property @pulumi.getter(name="logSegmentBytes") def log_segment_bytes(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "log_segment_bytes") @log_segment_bytes.setter def log_segment_bytes(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "log_segment_bytes", value) @property @pulumi.getter(name="numPartitions") def num_partitions(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "num_partitions") @num_partitions.setter def num_partitions(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "num_partitions", value) @property @pulumi.getter(name="socketReceiveBufferBytes") def socket_receive_buffer_bytes(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "socket_receive_buffer_bytes") @socket_receive_buffer_bytes.setter def socket_receive_buffer_bytes(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "socket_receive_buffer_bytes", value) @property @pulumi.getter(name="socketSendBufferBytes") def socket_send_buffer_bytes(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "socket_send_buffer_bytes") @socket_send_buffer_bytes.setter def socket_send_buffer_bytes(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "socket_send_buffer_bytes", value) @pulumi.input_type class MdbKafkaClusterConfigKafkaResourcesArgs: def __init__(__self__, *, disk_size: pulumi.Input[int], disk_type_id: pulumi.Input[str], resource_preset_id: pulumi.Input[str]): """ :param pulumi.Input[int] disk_size: Volume of the storage available to a ZooKeeper host, in gigabytes. :param pulumi.Input[str] disk_type_id: Type of the storage of ZooKeeper hosts. For more information see [the official documentation](https://cloud.yandex.com/docs/managed-kafka/concepts/storage). """ pulumi.set(__self__, "disk_size", disk_size) pulumi.set(__self__, "disk_type_id", disk_type_id) pulumi.set(__self__, "resource_preset_id", resource_preset_id) @property @pulumi.getter(name="diskSize") def disk_size(self) -> pulumi.Input[int]: """ Volume of the storage available to a ZooKeeper host, in gigabytes. """ return pulumi.get(self, "disk_size") @disk_size.setter def disk_size(self, value: pulumi.Input[int]): pulumi.set(self, "disk_size", value) @property @pulumi.getter(name="diskTypeId") def disk_type_id(self) -> pulumi.Input[str]: """ Type of the storage of ZooKeeper hosts. For more information see [the official documentation](https://cloud.yandex.com/docs/managed-kafka/concepts/storage). """ return pulumi.get(self, "disk_type_id") @disk_type_id.setter def disk_type_id(self, value: pulumi.Input[str]): pulumi.set(self, "disk_type_id", value) @property @pulumi.getter(name="resourcePresetId") def resource_preset_id(self) -> pulumi.Input[str]: return pulumi.get(self, "resource_preset_id") @resource_preset_id.setter def resource_preset_id(self, value: pulumi.Input[str]): pulumi.set(self, "resource_preset_id", value) @pulumi.input_type class MdbKafkaClusterConfigZookeeperArgs: def __init__(__self__, *, resources: Optional[pulumi.Input['MdbKafkaClusterConfigZookeeperResourcesArgs']] = None): """ :param pulumi.Input['MdbKafkaClusterConfigZookeeperResourcesArgs'] resources: Resources allocated to hosts of the ZooKeeper subcluster. The structure is documented below. """ if resources is not None: pulumi.set(__self__, "resources", resources) @property @pulumi.getter def resources(self) -> Optional[pulumi.Input['MdbKafkaClusterConfigZookeeperResourcesArgs']]: """ Resources allocated to hosts of the ZooKeeper subcluster. The structure is documented below. """ return pulumi.get(self, "resources") @resources.setter def resources(self, value: Optional[pulumi.Input['MdbKafkaClusterConfigZookeeperResourcesArgs']]): pulumi.set(self, "resources", value) @pulumi.input_type class MdbKafkaClusterConfigZookeeperResourcesArgs: def __init__(__self__, *, disk_size: Optional[pulumi.Input[int]] = None, disk_type_id: Optional[pulumi.Input[str]] = None, resource_preset_id: Optional[pulumi.Input[str]] = None): """ :param pulumi.Input[int] disk_size: Volume of the storage available to a ZooKeeper host, in gigabytes. :param pulumi.Input[str] disk_type_id: Type of the storage of ZooKeeper hosts. For more information see [the official documentation](https://cloud.yandex.com/docs/managed-kafka/concepts/storage). """ if disk_size is not None: pulumi.set(__self__, "disk_size", disk_size) if disk_type_id is not None: pulumi.set(__self__, "disk_type_id", disk_type_id) if resource_preset_id is not None: pulumi.set(__self__, "resource_preset_id", resource_preset_id) @property @pulumi.getter(name="diskSize") def disk_size(self) -> Optional[pulumi.Input[int]]: """ Volume of the storage available to a ZooKeeper host, in gigabytes. """ return pulumi.get(self, "disk_size") @disk_size.setter def disk_size(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "disk_size", value) @property @pulumi.getter(name="diskTypeId") def disk_type_id(self) -> Optional[pulumi.Input[str]]: """ Type of the storage of ZooKeeper hosts. For more information see [the official documentation](https://cloud.yandex.com/docs/managed-kafka/concepts/storage). """ return pulumi.get(self, "disk_type_id") @disk_type_id.setter def disk_type_id(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "disk_type_id", value) @property @pulumi.getter(name="resourcePresetId") def resource_preset_id(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "resource_preset_id") @resource_preset_id.setter def resource_preset_id(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "resource_preset_id", value) @pulumi.input_type class MdbKafkaClusterHostArgs: def __init__(__self__, *, assign_public_ip: Optional[pulumi.Input[bool]] = None, health: Optional[pulumi.Input[str]] = None, name: Optional[pulumi.Input[str]] = None, role: Optional[pulumi.Input[str]] = None, subnet_id: Optional[pulumi.Input[str]] = None, zone_id: Optional[pulumi.Input[str]] = None): """ :param pulumi.Input[bool] assign_public_ip: Determines whether each broker will be assigned a public IP address. The default is `false`. :param pulumi.Input[str] health: Health of the host. :param pulumi.Input[str] name: The name of the topic. :param pulumi.Input[str] role: The role type to grant to the topic. :param pulumi.Input[str] subnet_id: The ID of the subnet, to which the host belongs. :param pulumi.Input[str] zone_id: The availability zone where the Kafka host was created. """ if assign_public_ip is not None: pulumi.set(__self__, "assign_public_ip", assign_public_ip) if health is not None: pulumi.set(__self__, "health", health) if name is not None: pulumi.set(__self__, "name", name) if role is not None: pulumi.set(__self__, "role", role) if subnet_id is not None: pulumi.set(__self__, "subnet_id", subnet_id) if zone_id is not None: pulumi.set(__self__, "zone_id", zone_id) @property @pulumi.getter(name="assignPublicIp") def assign_public_ip(self) -> Optional[pulumi.Input[bool]]: """ Determines whether each broker will be assigned a public IP address. The default is `false`. """ return pulumi.get(self, "assign_public_ip") @assign_public_ip.setter def assign_public_ip(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "assign_public_ip", value) @property @pulumi.getter def health(self) -> Optional[pulumi.Input[str]]: """ Health of the host. """ return pulumi.get(self, "health") @health.setter def health(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "health", value) @property @pulumi.getter def name(self) -> Optional[pulumi.Input[str]]: """ The name of the topic. """ return pulumi.get(self, "name") @name.setter def name(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "name", value) @property @pulumi.getter def role(self) -> Optional[pulumi.Input[str]]: """ The role type to grant to the topic. """ return pulumi.get(self, "role") @role.setter def role(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "role", value) @property @pulumi.getter(name="subnetId") def subnet_id(self) -> Optional[pulumi.Input[str]]: """ The ID of the subnet, to which the host belongs. """ return pulumi.get(self, "subnet_id") @subnet_id.setter def subnet_id(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "subnet_id", value) @property @pulumi.getter(name="zoneId") def zone_id(self) -> Optional[pulumi.Input[str]]: """ The availability zone where the Kafka host was created. """ return pulumi.get(self, "zone_id") @zone_id.setter def zone_id(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "zone_id", value) @pulumi.input_type class MdbKafkaClusterTopicArgs: def __init__(__self__, *, name: pulumi.Input[str], partitions: pulumi.Input[int], replication_factor: pulumi.Input[int], topic_config: Optional[pulumi.Input['MdbKafkaClusterTopicTopicConfigArgs']] = None): """ :param pulumi.Input[str] name: The name of the topic. :param pulumi.Input[int] partitions: The number of the topic's partitions. :param pulumi.Input[int] replication_factor: Amount of data copies (replicas) for the topic in the cluster. :param pulumi.Input['MdbKafkaClusterTopicTopicConfigArgs'] topic_config: User-defined settings for the topic. The structure is documented below. """ pulumi.set(__self__, "name", name) pulumi.set(__self__, "partitions", partitions) pulumi.set(__self__, "replication_factor", replication_factor) if topic_config is not None: pulumi.set(__self__, "topic_config", topic_config) @property @pulumi.getter def name(self) -> pulumi.Input[str]: """ The name of the topic. """ return pulumi.get(self, "name") @name.setter def name(self, value: pulumi.Input[str]): pulumi.set(self, "name", value) @property @pulumi.getter def partitions(self) -> pulumi.Input[int]: """ The number of the topic's partitions. """ return pulumi.get(self, "partitions") @partitions.setter def partitions(self, value: pulumi.Input[int]): pulumi.set(self, "partitions", value) @property @pulumi.getter(name="replicationFactor") def replication_factor(self) -> pulumi.Input[int]: """ Amount of data copies (replicas) for the topic in the cluster. """ return pulumi.get(self, "replication_factor") @replication_factor.setter def replication_factor(self, value: pulumi.Input[int]): pulumi.set(self, "replication_factor", value) @property @pulumi.getter(name="topicConfig") def topic_config(self) -> Optional[pulumi.Input['MdbKafkaClusterTopicTopicConfigArgs']]: """ User-defined settings for the topic. The structure is documented below. """ return pulumi.get(self, "topic_config") @topic_config.setter def topic_config(self, value: Optional[pulumi.Input['MdbKafkaClusterTopicTopicConfigArgs']]): pulumi.set(self, "topic_config", value) @pulumi.input_type class MdbKafkaClusterTopicTopicConfigArgs: def __init__(__self__, *, cleanup_policy: Optional[pulumi.Input[str]] = None, compression_type: Optional[pulumi.Input[str]] = None, delete_retention_ms: Optional[pulumi.Input[str]] = None, file_delete_delay_ms: Optional[pulumi.Input[str]] = None, flush_messages: Optional[pulumi.Input[str]] = None, flush_ms: Optional[pulumi.Input[str]] = None, max_message_bytes: Optional[pulumi.Input[str]] = None, min_compaction_lag_ms: Optional[pulumi.Input[str]] = None, min_insync_replicas: Optional[pulumi.Input[str]] = None, preallocate: Optional[pulumi.Input[bool]] = None, retention_bytes: Optional[pulumi.Input[str]] = None, retention_ms: Optional[pulumi.Input[str]] = None, segment_bytes: Optional[pulumi.Input[str]] = None): if cleanup_policy is not None: pulumi.set(__self__, "cleanup_policy", cleanup_policy) if compression_type is not None: pulumi.set(__self__, "compression_type", compression_type) if delete_retention_ms is not None: pulumi.set(__self__, "delete_retention_ms", delete_retention_ms) if file_delete_delay_ms is not None: pulumi.set(__self__, "file_delete_delay_ms", file_delete_delay_ms) if flush_messages is not None: pulumi.set(__self__, "flush_messages", flush_messages) if flush_ms is not None: pulumi.set(__self__, "flush_ms", flush_ms) if max_message_bytes is not None: pulumi.set(__self__, "max_message_bytes", max_message_bytes) if min_compaction_lag_ms is not None: pulumi.set(__self__, "min_compaction_lag_ms", min_compaction_lag_ms) if min_insync_replicas is not None: pulumi.set(__self__, "min_insync_replicas", min_insync_replicas) if preallocate is not None: pulumi.set(__self__, "preallocate", preallocate) if retention_bytes is not None: pulumi.set(__self__, "retention_bytes", retention_bytes) if retention_ms is not None: pulumi.set(__self__, "retention_ms", retention_ms) if segment_bytes is not None: pulumi.set(__self__, "segment_bytes", segment_bytes) @property @pulumi.getter(name="cleanupPolicy") def cleanup_policy(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "cleanup_policy") @cleanup_policy.setter def cleanup_policy(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "cleanup_policy", value) @property @pulumi.getter(name="compressionType") def compression_type(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "compression_type") @compression_type.setter def compression_type(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "compression_type", value) @property @pulumi.getter(name="deleteRetentionMs") def delete_retention_ms(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "delete_retention_ms") @delete_retention_ms.setter def delete_retention_ms(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "delete_retention_ms", value) @property @pulumi.getter(name="fileDeleteDelayMs") def file_delete_delay_ms(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "file_delete_delay_ms") @file_delete_delay_ms.setter def file_delete_delay_ms(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "file_delete_delay_ms", value) @property @pulumi.getter(name="flushMessages") def flush_messages(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "flush_messages") @flush_messages.setter def flush_messages(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "flush_messages", value) @property @pulumi.getter(name="flushMs") def flush_ms(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "flush_ms") @flush_ms.setter def flush_ms(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "flush_ms", value) @property @pulumi.getter(name="maxMessageBytes") def max_message_bytes(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "max_message_bytes") @max_message_bytes.setter def max_message_bytes(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "max_message_bytes", value) @property @pulumi.getter(name="minCompactionLagMs") def min_compaction_lag_ms(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "min_compaction_lag_ms") @min_compaction_lag_ms.setter def min_compaction_lag_ms(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "min_compaction_lag_ms", value) @property @pulumi.getter(name="minInsyncReplicas") def min_insync_replicas(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "min_insync_replicas") @min_insync_replicas.setter def min_insync_replicas(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "min_insync_replicas", value) @property @pulumi.getter def preallocate(self) -> Optional[pulumi.Input[bool]]: return pulumi.get(self, "preallocate") @preallocate.setter def preallocate(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "preallocate", value) @property @pulumi.getter(name="retentionBytes") def retention_bytes(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "retention_bytes") @retention_bytes.setter def retention_bytes(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "retention_bytes", value) @property @pulumi.getter(name="retentionMs") def retention_ms(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "retention_ms") @retention_ms.setter def retention_ms(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "retention_ms", value) @property @pulumi.getter(name="segmentBytes") def segment_bytes(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "segment_bytes") @segment_bytes.setter def segment_bytes(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "segment_bytes", value) @pulumi.input_type class MdbKafkaClusterUserArgs: def __init__(__self__, *, name: pulumi.Input[str], password: pulumi.Input[str], permissions: Optional[pulumi.Input[Sequence[pulumi.Input['MdbKafkaClusterUserPermissionArgs']]]] = None): """ :param pulumi.Input[str] name: The name of the topic. :param pulumi.Input[str] password: The <PASSWORD> the <PASSWORD>. :param pulumi.Input[Sequence[pulumi.Input['MdbKafkaClusterUserPermissionArgs']]] permissions: Set of permissions granted to the user. The structure is documented below. """ pulumi.set(__self__, "name", name) pulumi.set(__self__, "password", password) if permissions is not None: pulumi.set(__self__, "permissions", permissions) @property @pulumi.getter def name(self) -> pulumi.Input[str]: """ The name of the topic. """ return pulumi.get(self, "name") @name.setter def name(self, value: pulumi.Input[str]): pulumi.set(self, "name", value) @property @pulumi.getter def password(self) -> pulumi.Input[str]: """ The password of the user. """ return pulumi.get(self, "password") @password.setter def password(self, value: pulumi.Input[str]): pulumi.set(self, "password", value) @property @pulumi.getter def permissions(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['MdbKafkaClusterUserPermissionArgs']]]]: """ Set of permissions granted to the user. The structure is documented below. """ return pulumi.get(self, "permissions") @permissions.setter def permissions(self, value: Optional[pulumi.Input[Sequence[pulumi.Input['MdbKafkaClusterUserPermissionArgs']]]]): pulumi.set(self, "permissions", value) @pulumi.input_type class MdbKafkaClusterUserPermissionArgs: def __init__(__self__, *, role: pulumi.Input[str], topic_name: pulumi.Input[str]): """ :param pulumi.Input[str] role: The role type to grant to the topic. :param pulumi.Input[str] topic_name: The name of the topic that the permission grants access to. """ pulumi.set(__self__, "role", role) pulumi.set(__self__, "topic_name", topic_name) @property @pulumi.getter def role(self) -> pulumi.Input[str]: """ The role type to grant to the topic. """ return pulumi.get(self, "role") @role.setter def role(self, value: pulumi.Input[str]): pulumi.set(self, "role", value) @property @pulumi.getter(name="topicName") def topic_name(self) -> pulumi.Input[str]: """ The name of the topic that the permission grants access to. """ return pulumi.get(self, "topic_name") @topic_name.setter def topic_name(self, value: pulumi.Input[str]): pulumi.set(self, "topic_name", value) @pulumi.input_type class MdbKafkaTopicTopicConfigArgs: def __init__(__self__, *, cleanup_policy: Optional[pulumi.Input[str]] = None, compression_type: Optional[pulumi.Input[str]] = None, delete_retention_ms: Optional[pulumi.Input[str]] = None, file_delete_delay_ms: Optional[pulumi.Input[str]] = None, flush_messages: Optional[pulumi.Input[str]] = None, flush_ms: Optional[pulumi.Input[str]] = None, max_message_bytes: Optional[pulumi.Input[str]] = None, min_compaction_lag_ms: Optional[pulumi.Input[str]] = None, min_insync_replicas: Optional[pulumi.Input[str]] = None, preallocate: Optional[pulumi.Input[bool]] = None, retention_bytes: Optional[pulumi.Input[str]] = None, retention_ms: Optional[pulumi.Input[str]] = None, segment_bytes: Optional[pulumi.Input[str]] = None): if cleanup_policy is not None: pulumi.set(__self__, "cleanup_policy", cleanup_policy) if compression_type is not None: pulumi.set(__self__, "compression_type", compression_type) if delete_retention_ms is not None: pulumi.set(__self__, "delete_retention_ms", delete_retention_ms) if file_delete_delay_ms is not None: pulumi.set(__self__, "file_delete_delay_ms", file_delete_delay_ms) if flush_messages is not None: pulumi.set(__self__, "flush_messages", flush_messages) if flush_ms is not None: pulumi.set(__self__, "flush_ms", flush_ms) if max_message_bytes is not None: pulumi.set(__self__, "max_message_bytes", max_message_bytes) if min_compaction_lag_ms is not None: pulumi.set(__self__, "min_compaction_lag_ms", min_compaction_lag_ms) if min_insync_replicas is not None: pulumi.set(__self__, "min_insync_replicas", min_insync_replicas) if preallocate is not None: pulumi.set(__self__, "preallocate", preallocate) if retention_bytes is not None: pulumi.set(__self__, "retention_bytes", retention_bytes) if retention_ms is not None: pulumi.set(__self__, "retention_ms", retention_ms) if segment_bytes is not None: pulumi.set(__self__, "segment_bytes", segment_bytes) @property @pulumi.getter(name="cleanupPolicy") def cleanup_policy(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "cleanup_policy") @cleanup_policy.setter def cleanup_policy(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "cleanup_policy", value) @property @pulumi.getter(name="compressionType") def compression_type(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "compression_type") @compression_type.setter def compression_type(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "compression_type", value) @property @pulumi.getter(name="deleteRetentionMs") def delete_retention_ms(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "delete_retention_ms") @delete_retention_ms.setter def delete_retention_ms(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "delete_retention_ms", value) @property @pulumi.getter(name="fileDeleteDelayMs") def file_delete_delay_ms(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "file_delete_delay_ms") @file_delete_delay_ms.setter def file_delete_delay_ms(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "file_delete_delay_ms", value) @property @pulumi.getter(name="flushMessages") def flush_messages(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "flush_messages") @flush_messages.setter def flush_messages(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "flush_messages", value) @property @pulumi.getter(name="flushMs") def flush_ms(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "flush_ms") @flush_ms.setter def flush_ms(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "flush_ms", value) @property @pulumi.getter(name="maxMessageBytes") def max_message_bytes(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "max_message_bytes") @max_message_bytes.setter def max_message_bytes(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "max_message_bytes", value) @property @pulumi.getter(name="minCompactionLagMs") def min_compaction_lag_ms(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "min_compaction_lag_ms") @min_compaction_lag_ms.setter def min_compaction_lag_ms(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "min_compaction_lag_ms", value) @property @pulumi.getter(name="minInsyncReplicas") def min_insync_replicas(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "min_insync_replicas") @min_insync_replicas.setter def min_insync_replicas(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "min_insync_replicas", value) @property @pulumi.getter def preallocate(self) -> Optional[pulumi.Input[bool]]: return pulumi.get(self, "preallocate") @preallocate.setter def preallocate(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "preallocate", value) @property @pulumi.getter(name="retentionBytes") def retention_bytes(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "retention_bytes") @retention_bytes.setter def retention_bytes(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "retention_bytes", value) @property @pulumi.getter(name="retentionMs") def retention_ms(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "retention_ms") @retention_ms.setter def retention_ms(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "retention_ms", value) @property @pulumi.getter(name="segmentBytes") def segment_bytes(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "segment_bytes") @segment_bytes.setter def segment_bytes(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "segment_bytes", value) @pulumi.input_type class MdbMongodbClusterClusterConfigArgs: def __init__(__self__, *, version: pulumi.Input[str], access: Optional[pulumi.Input['MdbMongodbClusterClusterConfigAccessArgs']] = None, backup_window_start: Optional[pulumi.Input['MdbMongodbClusterClusterConfigBackupWindowStartArgs']] = None, feature_compatibility_version: Optional[pulumi.Input[str]] = None): """ :param pulumi.Input[str] version: Version of MongoDB (either 5.0, 4.4, 4.2 or 4.0). :param pulumi.Input['MdbMongodbClusterClusterConfigAccessArgs'] access: Shows whether cluster has access to data lens. The structure is documented below. :param pulumi.Input['MdbMongodbClusterClusterConfigBackupWindowStartArgs'] backup_window_start: Time to start the daily backup, in the UTC timezone. The structure is documented below. :param pulumi.Input[str] feature_compatibility_version: Feature compatibility version of MongoDB. If not provided version is taken. Can be either `5.0`, `4.4`, `4.2` and `4.0`. """ pulumi.set(__self__, "version", version) if access is not None: pulumi.set(__self__, "access", access) if backup_window_start is not None: pulumi.set(__self__, "backup_window_start", backup_window_start) if feature_compatibility_version is not None: pulumi.set(__self__, "feature_compatibility_version", feature_compatibility_version) @property @pulumi.getter def version(self) -> pulumi.Input[str]: """ Version of MongoDB (either 5.0, 4.4, 4.2 or 4.0). """ return pulumi.get(self, "version") @version.setter def version(self, value: pulumi.Input[str]): pulumi.set(self, "version", value) @property @pulumi.getter def access(self) -> Optional[pulumi.Input['MdbMongodbClusterClusterConfigAccessArgs']]: """ Shows whether cluster has access to data lens. The structure is documented below. """ return pulumi.get(self, "access") @access.setter def access(self, value: Optional[pulumi.Input['MdbMongodbClusterClusterConfigAccessArgs']]): pulumi.set(self, "access", value) @property @pulumi.getter(name="backupWindowStart") def backup_window_start(self) -> Optional[pulumi.Input['MdbMongodbClusterClusterConfigBackupWindowStartArgs']]: """ Time to start the daily backup, in the UTC timezone. The structure is documented below. """ return pulumi.get(self, "backup_window_start") @backup_window_start.setter def backup_window_start(self, value: Optional[pulumi.Input['MdbMongodbClusterClusterConfigBackupWindowStartArgs']]): pulumi.set(self, "backup_window_start", value) @property @pulumi.getter(name="featureCompatibilityVersion") def feature_compatibility_version(self) -> Optional[pulumi.Input[str]]: """ Feature compatibility version of MongoDB. If not provided version is taken. Can be either `5.0`, `4.4`, `4.2` and `4.0`. """ return pulumi.get(self, "feature_compatibility_version") @feature_compatibility_version.setter def feature_compatibility_version(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "feature_compatibility_version", value) @pulumi.input_type class MdbMongodbClusterClusterConfigAccessArgs: def __init__(__self__, *, data_lens: Optional[pulumi.Input[bool]] = None): """ :param pulumi.Input[bool] data_lens: Allow access for DataLens. """ if data_lens is not None: pulumi.set(__self__, "data_lens", data_lens) @property @pulumi.getter(name="dataLens") def data_lens(self) -> Optional[pulumi.Input[bool]]: """ Allow access for DataLens. """ return pulumi.get(self, "data_lens") @data_lens.setter def data_lens(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "data_lens", value) @pulumi.input_type class MdbMongodbClusterClusterConfigBackupWindowStartArgs: def __init__(__self__, *, hours: Optional[pulumi.Input[int]] = None, minutes: Optional[pulumi.Input[int]] = None): """ :param pulumi.Input[int] hours: The hour at which backup will be started. :param pulumi.Input[int] minutes: The minute at which backup will be started. """ if hours is not None: pulumi.set(__self__, "hours", hours) if minutes is not None: pulumi.set(__self__, "minutes", minutes) @property @pulumi.getter def hours(self) -> Optional[pulumi.Input[int]]: """ The hour at which backup will be started. """ return pulumi.get(self, "hours") @hours.setter def hours(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "hours", value) @property @pulumi.getter def minutes(self) -> Optional[pulumi.Input[int]]: """ The minute at which backup will be started. """ return pulumi.get(self, "minutes") @minutes.setter def minutes(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "minutes", value) @pulumi.input_type class MdbMongodbClusterDatabaseArgs: def __init__(__self__, *, name: pulumi.Input[str]): """ :param pulumi.Input[str] name: The fully qualified domain name of the host. Computed on server side. """ pulumi.set(__self__, "name", name) @property @pulumi.getter def name(self) -> pulumi.Input[str]: """ The fully qualified domain name of the host. Computed on server side. """ return pulumi.get(self, "name") @name.setter def name(self, value: pulumi.Input[str]): pulumi.set(self, "name", value) @pulumi.input_type class MdbMongodbClusterHostArgs: def __init__(__self__, *, subnet_id: pulumi.Input[str], zone_id: pulumi.Input[str], assign_public_ip: Optional[pulumi.Input[bool]] = None, health: Optional[pulumi.Input[str]] = None, name: Optional[pulumi.Input[str]] = None, role: Optional[pulumi.Input[str]] = None, shard_name: Optional[pulumi.Input[str]] = None, type: Optional[pulumi.Input[str]] = None): """ :param pulumi.Input[str] subnet_id: The ID of the subnet, to which the host belongs. The subnet must be a part of the network to which the cluster belongs. :param pulumi.Input[str] zone_id: The availability zone where the MongoDB host will be created. For more information see [the official documentation](https://cloud.yandex.com/docs/overview/concepts/geo-scope). :param pulumi.Input[bool] assign_public_ip: -(Optional) Should this host have assigned public IP assigned. Can be either `true` or `false`. :param pulumi.Input[str] health: The health of the host. :param pulumi.Input[str] name: The fully qualified domain name of the host. Computed on server side. :param pulumi.Input[str] role: The role of the cluster (either PRIMARY or SECONDARY). :param pulumi.Input[str] shard_name: The name of the shard to which the host belongs. :param pulumi.Input[str] type: Type of maintenance window. Can be either `ANYTIME` or `WEEKLY`. A day and hour of window need to be specified with weekly window. """ pulumi.set(__self__, "subnet_id", subnet_id) pulumi.set(__self__, "zone_id", zone_id) if assign_public_ip is not None: pulumi.set(__self__, "assign_public_ip", assign_public_ip) if health is not None: pulumi.set(__self__, "health", health) if name is not None: pulumi.set(__self__, "name", name) if role is not None: pulumi.set(__self__, "role", role) if shard_name is not None: pulumi.set(__self__, "shard_name", shard_name) if type is not None: pulumi.set(__self__, "type", type) @property @pulumi.getter(name="subnetId") def subnet_id(self) -> pulumi.Input[str]: """ The ID of the subnet, to which the host belongs. The subnet must be a part of the network to which the cluster belongs. """ return pulumi.get(self, "subnet_id") @subnet_id.setter def subnet_id(self, value: pulumi.Input[str]): pulumi.set(self, "subnet_id", value) @property @pulumi.getter(name="zoneId") def zone_id(self) -> pulumi.Input[str]: """ The availability zone where the MongoDB host will be created. For more information see [the official documentation](https://cloud.yandex.com/docs/overview/concepts/geo-scope). """ return pulumi.get(self, "zone_id") @zone_id.setter def zone_id(self, value: pulumi.Input[str]): pulumi.set(self, "zone_id", value) @property @pulumi.getter(name="assignPublicIp") def assign_public_ip(self) -> Optional[pulumi.Input[bool]]: """ -(Optional) Should this host have assigned public IP assigned. Can be either `true` or `false`. """ return pulumi.get(self, "assign_public_ip") @assign_public_ip.setter def assign_public_ip(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "assign_public_ip", value) @property @pulumi.getter def health(self) -> Optional[pulumi.Input[str]]: """ The health of the host. """ return pulumi.get(self, "health") @health.setter def health(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "health", value) @property @pulumi.getter def name(self) -> Optional[pulumi.Input[str]]: """ The fully qualified domain name of the host. Computed on server side. """ return pulumi.get(self, "name") @name.setter def name(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "name", value) @property @pulumi.getter def role(self) -> Optional[pulumi.Input[str]]: """ The role of the cluster (either PRIMARY or SECONDARY). """ return pulumi.get(self, "role") @role.setter def role(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "role", value) @property @pulumi.getter(name="shardName") def shard_name(self) -> Optional[pulumi.Input[str]]: """ The name of the shard to which the host belongs. """ return pulumi.get(self, "shard_name") @shard_name.setter def shard_name(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "shard_name", value) @property @pulumi.getter def type(self) -> Optional[pulumi.Input[str]]: """ Type of maintenance window. Can be either `ANYTIME` or `WEEKLY`. A day and hour of window need to be specified with weekly window. """ return pulumi.get(self, "type") @type.setter def type(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "type", value) @pulumi.input_type class MdbMongodbClusterMaintenanceWindowArgs: def __init__(__self__, *, type: pulumi.Input[str], day: Optional[pulumi.Input[str]] = None, hour: Optional[pulumi.Input[int]] = None): """ :param pulumi.Input[str] type: Type of maintenance window. Can be either `ANYTIME` or `WEEKLY`. A day and hour of window need to be specified with weekly window. :param pulumi.Input[str] day: Day of week for maintenance window if window type is weekly. Possible values: `MON`, `TUE`, `WED`, `THU`, `FRI`, `SAT`, `SUN`. :param pulumi.Input[int] hour: Hour of day in UTC time zone (1-24) for maintenance window if window type is weekly. """ pulumi.set(__self__, "type", type) if day is not None: pulumi.set(__self__, "day", day) if hour is not None: pulumi.set(__self__, "hour", hour) @property @pulumi.getter def type(self) -> pulumi.Input[str]: """ Type of maintenance window. Can be either `ANYTIME` or `WEEKLY`. A day and hour of window need to be specified with weekly window. """ return pulumi.get(self, "type") @type.setter def type(self, value: pulumi.Input[str]): pulumi.set(self, "type", value) @property @pulumi.getter def day(self) -> Optional[pulumi.Input[str]]: """ Day of week for maintenance window if window type is weekly. Possible values: `MON`, `TUE`, `WED`, `THU`, `FRI`, `SAT`, `SUN`. """ return pulumi.get(self, "day") @day.setter def day(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "day", value) @property @pulumi.getter def hour(self) -> Optional[pulumi.Input[int]]: """ Hour of day in UTC time zone (1-24) for maintenance window if window type is weekly. """ return pulumi.get(self, "hour") @hour.setter def hour(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "hour", value) @pulumi.input_type class MdbMongodbClusterResourcesArgs: def __init__(__self__, *, disk_size: pulumi.Input[int], disk_type_id: pulumi.Input[str], resource_preset_id: pulumi.Input[str]): """ :param pulumi.Input[int] disk_size: Volume of the storage available to a MongoDB host, in gigabytes. :param pulumi.Input[str] disk_type_id: Type of the storage of MongoDB hosts. For more information see [the official documentation](https://cloud.yandex.com/docs/managed-clickhouse/concepts/storage). """ pulumi.set(__self__, "disk_size", disk_size) pulumi.set(__self__, "disk_type_id", disk_type_id) pulumi.set(__self__, "resource_preset_id", resource_preset_id) @property @pulumi.getter(name="diskSize") def disk_size(self) -> pulumi.Input[int]: """ Volume of the storage available to a MongoDB host, in gigabytes. """ return pulumi.get(self, "disk_size") @disk_size.setter def disk_size(self, value: pulumi.Input[int]): pulumi.set(self, "disk_size", value) @property @pulumi.getter(name="diskTypeId") def disk_type_id(self) -> pulumi.Input[str]: """ Type of the storage of MongoDB hosts. For more information see [the official documentation](https://cloud.yandex.com/docs/managed-clickhouse/concepts/storage). """ return pulumi.get(self, "disk_type_id") @disk_type_id.setter def disk_type_id(self, value: pulumi.Input[str]): pulumi.set(self, "disk_type_id", value) @property @pulumi.getter(name="resourcePresetId") def resource_preset_id(self) -> pulumi.Input[str]: return pulumi.get(self, "resource_preset_id") @resource_preset_id.setter def resource_preset_id(self, value: pulumi.Input[str]): pulumi.set(self, "resource_preset_id", value) @pulumi.input_type class MdbMongodbClusterUserArgs: def __init__(__self__, *, name: pulumi.Input[str], password: pulumi.Input[str], permissions: Optional[pulumi.Input[Sequence[pulumi.Input['MdbMongodbClusterUserPermissionArgs']]]] = None): """ :param pulumi.Input[str] name: The fully qualified domain name of the host. Computed on server side. :param pulumi.Input[str] password: <PASSWORD>. :param pulumi.Input[Sequence[pulumi.Input['MdbMongodbClusterUserPermissionArgs']]] permissions: Set of permissions granted to the user. The structure is documented below. """ pulumi.set(__self__, "name", name) pulumi.set(__self__, "password", password) if permissions is not None: pulumi.set(__self__, "permissions", permissions) @property @pulumi.getter def name(self) -> pulumi.Input[str]: """ The fully qualified domain name of the host. Computed on server side. """ return pulumi.get(self, "name") @name.setter def name(self, value: pulumi.Input[str]): pulumi.set(self, "name", value) @property @pulumi.getter def password(self) -> pulumi.Input[str]: """ The password of the user. """ return pulumi.get(self, "password") @password.setter def password(self, value: pulumi.Input[str]): pulumi.set(self, "password", value) @property @pulumi.getter def permissions(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['MdbMongodbClusterUserPermissionArgs']]]]: """ Set of permissions granted to the user. The structure is documented below. """ return pulumi.get(self, "permissions") @permissions.setter def permissions(self, value: Optional[pulumi.Input[Sequence[pulumi.Input['MdbMongodbClusterUserPermissionArgs']]]]): pulumi.set(self, "permissions", value) @pulumi.input_type class MdbMongodbClusterUserPermissionArgs: def __init__(__self__, *, database_name: pulumi.Input[str], roles: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None): """ :param pulumi.Input[str] database_name: The name of the database that the permission grants access to. """ pulumi.set(__self__, "database_name", database_name) if roles is not None: pulumi.set(__self__, "roles", roles) @property @pulumi.getter(name="databaseName") def database_name(self) -> pulumi.Input[str]: """ The name of the database that the permission grants access to. """ return pulumi.get(self, "database_name") @database_name.setter def database_name(self, value: pulumi.Input[str]): pulumi.set(self, "database_name", value) @property @pulumi.getter def roles(self) -> Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]: return pulumi.get(self, "roles") @roles.setter def roles(self, value: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]): pulumi.set(self, "roles", value) @pulumi.input_type class MdbMysqlClusterAccessArgs: def __init__(__self__, *, data_lens: Optional[pulumi.Input[bool]] = None, web_sql: Optional[pulumi.Input[bool]] = None): """ :param pulumi.Input[bool] data_lens: Allow access for [Yandex DataLens](https://cloud.yandex.com/services/datalens). :param pulumi.Input[bool] web_sql: Allows access for [SQL queries in the management console](https://cloud.yandex.com/docs/managed-mysql/operations/web-sql-query). """ if data_lens is not None: pulumi.set(__self__, "data_lens", data_lens) if web_sql is not None: pulumi.set(__self__, "web_sql", web_sql) @property @pulumi.getter(name="dataLens") def data_lens(self) -> Optional[pulumi.Input[bool]]: """ Allow access for [Yandex DataLens](https://cloud.yandex.com/services/datalens). """ return pulumi.get(self, "data_lens") @data_lens.setter def data_lens(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "data_lens", value) @property @pulumi.getter(name="webSql") def web_sql(self) -> Optional[pulumi.Input[bool]]: """ Allows access for [SQL queries in the management console](https://cloud.yandex.com/docs/managed-mysql/operations/web-sql-query). """ return pulumi.get(self, "web_sql") @web_sql.setter def web_sql(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "web_sql", value) @pulumi.input_type class MdbMysqlClusterBackupWindowStartArgs: def __init__(__self__, *, hours: Optional[pulumi.Input[int]] = None, minutes: Optional[pulumi.Input[int]] = None): """ :param pulumi.Input[int] hours: The hour at which backup will be started. :param pulumi.Input[int] minutes: The minute at which backup will be started. """ if hours is not None: pulumi.set(__self__, "hours", hours) if minutes is not None: pulumi.set(__self__, "minutes", minutes) @property @pulumi.getter def hours(self) -> Optional[pulumi.Input[int]]: """ The hour at which backup will be started. """ return pulumi.get(self, "hours") @hours.setter def hours(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "hours", value) @property @pulumi.getter def minutes(self) -> Optional[pulumi.Input[int]]: """ The minute at which backup will be started. """ return pulumi.get(self, "minutes") @minutes.setter def minutes(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "minutes", value) @pulumi.input_type class MdbMysqlClusterDatabaseArgs: def __init__(__self__, *, name: pulumi.Input[str]): """ :param pulumi.Input[str] name: Host state name. It should be set for all hosts or unset for all hosts. This field can be used by another host, to select which host will be its replication source. Please refer to `replication_source_name` parameter. """ pulumi.set(__self__, "name", name) @property @pulumi.getter def name(self) -> pulumi.Input[str]: """ Host state name. It should be set for all hosts or unset for all hosts. This field can be used by another host, to select which host will be its replication source. Please refer to `replication_source_name` parameter. """ return pulumi.get(self, "name") @name.setter def name(self, value: pulumi.Input[str]): pulumi.set(self, "name", value) @pulumi.input_type class MdbMysqlClusterHostArgs: def __init__(__self__, *, zone: pulumi.Input[str], assign_public_ip: Optional[pulumi.Input[bool]] = None, fqdn: Optional[pulumi.Input[str]] = None, name: Optional[pulumi.Input[str]] = None, replication_source: Optional[pulumi.Input[str]] = None, replication_source_name: Optional[pulumi.Input[str]] = None, subnet_id: Optional[pulumi.Input[str]] = None): """ :param pulumi.Input[str] zone: The availability zone where the MySQL host will be created. :param pulumi.Input[bool] assign_public_ip: Sets whether the host should get a public IP address. It can be changed on the fly only when `name` is set. :param pulumi.Input[str] fqdn: The fully qualified domain name of the host. :param pulumi.Input[str] name: Host state name. It should be set for all hosts or unset for all hosts. This field can be used by another host, to select which host will be its replication source. Please refer to `replication_source_name` parameter. :param pulumi.Input[str] replication_source: Host replication source (fqdn), when replication_source is empty then host is in HA group. :param pulumi.Input[str] replication_source_name: Host replication source name points to host's `name` from which this host should replicate. When not set then host in HA group. It works only when `name` is set. :param pulumi.Input[str] subnet_id: The ID of the subnet, to which the host belongs. The subnet must be a part of the network to which the cluster belongs. """ pulumi.set(__self__, "zone", zone) if assign_public_ip is not None: pulumi.set(__self__, "assign_public_ip", assign_public_ip) if fqdn is not None: pulumi.set(__self__, "fqdn", fqdn) if name is not None: pulumi.set(__self__, "name", name) if replication_source is not None: pulumi.set(__self__, "replication_source", replication_source) if replication_source_name is not None: pulumi.set(__self__, "replication_source_name", replication_source_name) if subnet_id is not None: pulumi.set(__self__, "subnet_id", subnet_id) @property @pulumi.getter def zone(self) -> pulumi.Input[str]: """ The availability zone where the MySQL host will be created. """ return pulumi.get(self, "zone") @zone.setter def zone(self, value: pulumi.Input[str]): pulumi.set(self, "zone", value) @property @pulumi.getter(name="assignPublicIp") def assign_public_ip(self) -> Optional[pulumi.Input[bool]]: """ Sets whether the host should get a public IP address. It can be changed on the fly only when `name` is set. """ return pulumi.get(self, "assign_public_ip") @assign_public_ip.setter def assign_public_ip(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "assign_public_ip", value) @property @pulumi.getter def fqdn(self) -> Optional[pulumi.Input[str]]: """ The fully qualified domain name of the host. """ return pulumi.get(self, "fqdn") @fqdn.setter def fqdn(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "fqdn", value) @property @pulumi.getter def name(self) -> Optional[pulumi.Input[str]]: """ Host state name. It should be set for all hosts or unset for all hosts. This field can be used by another host, to select which host will be its replication source. Please refer to `replication_source_name` parameter. """ return pulumi.get(self, "name") @name.setter def name(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "name", value) @property @pulumi.getter(name="replicationSource") def replication_source(self) -> Optional[pulumi.Input[str]]: """ Host replication source (fqdn), when replication_source is empty then host is in HA group. """ return pulumi.get(self, "replication_source") @replication_source.setter def replication_source(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "replication_source", value) @property @pulumi.getter(name="replicationSourceName") def replication_source_name(self) -> Optional[pulumi.Input[str]]: """ Host replication source name points to host's `name` from which this host should replicate. When not set then host in HA group. It works only when `name` is set. """ return pulumi.get(self, "replication_source_name") @replication_source_name.setter def replication_source_name(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "replication_source_name", value) @property @pulumi.getter(name="subnetId") def subnet_id(self) -> Optional[pulumi.Input[str]]: """ The ID of the subnet, to which the host belongs. The subnet must be a part of the network to which the cluster belongs. """ return pulumi.get(self, "subnet_id") @subnet_id.setter def subnet_id(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "subnet_id", value) @pulumi.input_type class MdbMysqlClusterMaintenanceWindowArgs: def __init__(__self__, *, type: pulumi.Input[str], day: Optional[pulumi.Input[str]] = None, hour: Optional[pulumi.Input[int]] = None): """ :param pulumi.Input[str] type: Type of maintenance window. Can be either `ANYTIME` or `WEEKLY`. A day and hour of window need to be specified with weekly window. :param pulumi.Input[str] day: Day of the week (in `DDD` format). Allowed values: "MON", "TUE", "WED", "THU", "FRI", "SAT", "SUN" :param pulumi.Input[int] hour: Hour of the day in UTC (in `HH` format). Allowed value is between 0 and 23. """ pulumi.set(__self__, "type", type) if day is not None: pulumi.set(__self__, "day", day) if hour is not None: pulumi.set(__self__, "hour", hour) @property @pulumi.getter def type(self) -> pulumi.Input[str]: """ Type of maintenance window. Can be either `ANYTIME` or `WEEKLY`. A day and hour of window need to be specified with weekly window. """ return pulumi.get(self, "type") @type.setter def type(self, value: pulumi.Input[str]): pulumi.set(self, "type", value) @property @pulumi.getter def day(self) -> Optional[pulumi.Input[str]]: """ Day of the week (in `DDD` format). Allowed values: "MON", "TUE", "WED", "THU", "FRI", "SAT", "SUN" """ return pulumi.get(self, "day") @day.setter def day(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "day", value) @property @pulumi.getter def hour(self) -> Optional[pulumi.Input[int]]: """ Hour of the day in UTC (in `HH` format). Allowed value is between 0 and 23. """ return pulumi.get(self, "hour") @hour.setter def hour(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "hour", value) @pulumi.input_type class MdbMysqlClusterResourcesArgs: def __init__(__self__, *, disk_size: pulumi.Input[int], disk_type_id: pulumi.Input[str], resource_preset_id: pulumi.Input[str]): """ :param pulumi.Input[int] disk_size: Volume of the storage available to a MySQL host, in gigabytes. :param pulumi.Input[str] disk_type_id: Type of the storage of MySQL hosts. """ pulumi.set(__self__, "disk_size", disk_size) pulumi.set(__self__, "disk_type_id", disk_type_id) pulumi.set(__self__, "resource_preset_id", resource_preset_id) @property @pulumi.getter(name="diskSize") def disk_size(self) -> pulumi.Input[int]: """ Volume of the storage available to a MySQL host, in gigabytes. """ return pulumi.get(self, "disk_size") @disk_size.setter def disk_size(self, value: pulumi.Input[int]): pulumi.set(self, "disk_size", value) @property @pulumi.getter(name="diskTypeId") def disk_type_id(self) -> pulumi.Input[str]: """ Type of the storage of MySQL hosts. """ return pulumi.get(self, "disk_type_id") @disk_type_id.setter def disk_type_id(self, value: pulumi.Input[str]): pulumi.set(self, "disk_type_id", value) @property @pulumi.getter(name="resourcePresetId") def resource_preset_id(self) -> pulumi.Input[str]: return pulumi.get(self, "resource_preset_id") @resource_preset_id.setter def resource_preset_id(self, value: pulumi.Input[str]): pulumi.set(self, "resource_preset_id", value) @pulumi.input_type class MdbMysqlClusterRestoreArgs: def __init__(__self__, *, backup_id: pulumi.Input[str], time: Optional[pulumi.Input[str]] = None): """ :param pulumi.Input[str] backup_id: Backup ID. The cluster will be created from the specified backup. [How to get a list of MySQL backups](https://cloud.yandex.com/docs/managed-mysql/operations/cluster-backups). :param pulumi.Input[str] time: Timestamp of the moment to which the MySQL cluster should be restored. (Format: "2006-01-02T15:04:05" - UTC). When not set, current time is used. """ pulumi.set(__self__, "backup_id", backup_id) if time is not None: pulumi.set(__self__, "time", time) @property @pulumi.getter(name="backupId") def backup_id(self) -> pulumi.Input[str]: """ Backup ID. The cluster will be created from the specified backup. [How to get a list of MySQL backups](https://cloud.yandex.com/docs/managed-mysql/operations/cluster-backups). """ return pulumi.get(self, "backup_id") @backup_id.setter def backup_id(self, value: pulumi.Input[str]): pulumi.set(self, "backup_id", value) @property @pulumi.getter def time(self) -> Optional[pulumi.Input[str]]: """ Timestamp of the moment to which the MySQL cluster should be restored. (Format: "2006-01-02T15:04:05" - UTC). When not set, current time is used. """ return pulumi.get(self, "time") @time.setter def time(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "time", value) @pulumi.input_type class MdbMysqlClusterUserArgs: def __init__(__self__, *, name: pulumi.Input[str], password: pulumi.Input[str], authentication_plugin: Optional[pulumi.Input[str]] = None, connection_limits: Optional[pulumi.Input['MdbMysqlClusterUserConnectionLimitsArgs']] = None, global_permissions: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, permissions: Optional[pulumi.Input[Sequence[pulumi.Input['MdbMysqlClusterUserPermissionArgs']]]] = None): """ :param pulumi.Input[str] name: Host state name. It should be set for all hosts or unset for all hosts. This field can be used by another host, to select which host will be its replication source. Please refer to `replication_source_name` parameter. :param pulumi.Input[str] password: <PASSWORD>. :param pulumi.Input[str] authentication_plugin: Authentication plugin. Allowed values: `MYSQL_NATIVE_PASSWORD`, `CACHING_SHA2_PASSWORD`, `<PASSWORD>` (for version 5.7 `MYSQL_NATIVE_PASSWORD`, `<PASSWORD>`) :param pulumi.Input['MdbMysqlClusterUserConnectionLimitsArgs'] connection_limits: User's connection limits. The structure is documented below. If the attribute is not specified there will be no changes. :param pulumi.Input[Sequence[pulumi.Input[str]]] global_permissions: List user's global permissions Allowed permissions: `REPLICATION_CLIENT`, `REPLICATION_SLAVE`, `PROCESS` for clear list use empty list. If the attribute is not specified there will be no changes. :param pulumi.Input[Sequence[pulumi.Input['MdbMysqlClusterUserPermissionArgs']]] permissions: Set of permissions granted to the user. The structure is documented below. """ pulumi.set(__self__, "name", name) pulumi.set(__self__, "password", password) if authentication_plugin is not None: pulumi.set(__self__, "authentication_plugin", authentication_plugin) if connection_limits is not None: pulumi.set(__self__, "connection_limits", connection_limits) if global_permissions is not None: pulumi.set(__self__, "global_permissions", global_permissions) if permissions is not None: pulumi.set(__self__, "permissions", permissions) @property @pulumi.getter def name(self) -> pulumi.Input[str]: """ Host state name. It should be set for all hosts or unset for all hosts. This field can be used by another host, to select which host will be its replication source. Please refer to `replication_source_name` parameter. """ return pulumi.get(self, "name") @name.setter def name(self, value: pulumi.Input[str]): pulumi.set(self, "name", value) @property @pulumi.getter def password(self) -> pulumi.Input[str]: """ The password of the user. """ return pulumi.get(self, "password") @password.setter def password(self, value: pulumi.Input[str]): pulumi.set(self, "password", value) @property @pulumi.getter(name="authenticationPlugin") def authentication_plugin(self) -> Optional[pulumi.Input[str]]: """ Authentication plugin. Allowed values: `MYSQL_NATIVE_PASSWORD`, `CACHING_SHA2_PASSWORD`, `SHA256_PASSWORD` (for version 5.7 `MYSQL_NATIVE_PASSWORD`, `SHA256_PASSWORD`) """ return pulumi.get(self, "authentication_plugin") @authentication_plugin.setter def authentication_plugin(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "authentication_plugin", value) @property @pulumi.getter(name="connectionLimits") def connection_limits(self) -> Optional[pulumi.Input['MdbMysqlClusterUserConnectionLimitsArgs']]: """ User's connection limits. The structure is documented below. If the attribute is not specified there will be no changes. """ return pulumi.get(self, "connection_limits") @connection_limits.setter def connection_limits(self, value: Optional[pulumi.Input['MdbMysqlClusterUserConnectionLimitsArgs']]): pulumi.set(self, "connection_limits", value) @property @pulumi.getter(name="globalPermissions") def global_permissions(self) -> Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]: """ List user's global permissions Allowed permissions: `REPLICATION_CLIENT`, `REPLICATION_SLAVE`, `PROCESS` for clear list use empty list. If the attribute is not specified there will be no changes. """ return pulumi.get(self, "global_permissions") @global_permissions.setter def global_permissions(self, value: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]): pulumi.set(self, "global_permissions", value) @property @pulumi.getter def permissions(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['MdbMysqlClusterUserPermissionArgs']]]]: """ Set of permissions granted to the user. The structure is documented below. """ return pulumi.get(self, "permissions") @permissions.setter def permissions(self, value: Optional[pulumi.Input[Sequence[pulumi.Input['MdbMysqlClusterUserPermissionArgs']]]]): pulumi.set(self, "permissions", value) @pulumi.input_type class MdbMysqlClusterUserConnectionLimitsArgs: def __init__(__self__, *, max_connections_per_hour: Optional[pulumi.Input[int]] = None, max_questions_per_hour: Optional[pulumi.Input[int]] = None, max_updates_per_hour: Optional[pulumi.Input[int]] = None, max_user_connections: Optional[pulumi.Input[int]] = None): """ :param pulumi.Input[int] max_connections_per_hour: Max connections per hour. :param pulumi.Input[int] max_questions_per_hour: Max questions per hour. :param pulumi.Input[int] max_updates_per_hour: Max updates per hour. :param pulumi.Input[int] max_user_connections: Max user connections. """ if max_connections_per_hour is not None: pulumi.set(__self__, "max_connections_per_hour", max_connections_per_hour) if max_questions_per_hour is not None: pulumi.set(__self__, "max_questions_per_hour", max_questions_per_hour) if max_updates_per_hour is not None: pulumi.set(__self__, "max_updates_per_hour", max_updates_per_hour) if max_user_connections is not None: pulumi.set(__self__, "max_user_connections", max_user_connections) @property @pulumi.getter(name="maxConnectionsPerHour") def max_connections_per_hour(self) -> Optional[pulumi.Input[int]]: """ Max connections per hour. """ return pulumi.get(self, "max_connections_per_hour") @max_connections_per_hour.setter def max_connections_per_hour(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "max_connections_per_hour", value) @property @pulumi.getter(name="maxQuestionsPerHour") def max_questions_per_hour(self) -> Optional[pulumi.Input[int]]: """ Max questions per hour. """ return pulumi.get(self, "max_questions_per_hour") @max_questions_per_hour.setter def max_questions_per_hour(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "max_questions_per_hour", value) @property @pulumi.getter(name="maxUpdatesPerHour") def max_updates_per_hour(self) -> Optional[pulumi.Input[int]]: """ Max updates per hour. """ return pulumi.get(self, "max_updates_per_hour") @max_updates_per_hour.setter def max_updates_per_hour(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "max_updates_per_hour", value) @property @pulumi.getter(name="maxUserConnections") def max_user_connections(self) -> Optional[pulumi.Input[int]]: """ Max user connections. """ return pulumi.get(self, "max_user_connections") @max_user_connections.setter def max_user_connections(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "max_user_connections", value) @pulumi.input_type class MdbMysqlClusterUserPermissionArgs: def __init__(__self__, *, database_name: pulumi.Input[str], roles: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None): """ :param pulumi.Input[str] database_name: The name of the database that the permission grants access to. :param pulumi.Input[Sequence[pulumi.Input[str]]] roles: List user's roles in the database. Allowed roles: `ALL`,`ALTER`,`ALTER_ROUTINE`,`CREATE`,`CREATE_ROUTINE`,`CREATE_TEMPORARY_TABLES`, `CREATE_VIEW`,`DELETE`,`DROP`,`EVENT`,`EXECUTE`,`INDEX`,`INSERT`,`LOCK_TABLES`,`SELECT`,`SHOW_VIEW`,`TRIGGER`,`UPDATE`. """ pulumi.set(__self__, "database_name", database_name) if roles is not None: pulumi.set(__self__, "roles", roles) @property @pulumi.getter(name="databaseName") def database_name(self) -> pulumi.Input[str]: """ The name of the database that the permission grants access to. """ return pulumi.get(self, "database_name") @database_name.setter def database_name(self, value: pulumi.Input[str]): pulumi.set(self, "database_name", value) @property @pulumi.getter def roles(self) -> Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]: """ List user's roles in the database. Allowed roles: `ALL`,`ALTER`,`ALTER_ROUTINE`,`CREATE`,`CREATE_ROUTINE`,`CREATE_TEMPORARY_TABLES`, `CREATE_VIEW`,`DELETE`,`DROP`,`EVENT`,`EXECUTE`,`INDEX`,`INSERT`,`LOCK_TABLES`,`SELECT`,`SHOW_VIEW`,`TRIGGER`,`UPDATE`. """ return pulumi.get(self, "roles") @roles.setter def roles(self, value: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]): pulumi.set(self, "roles", value) @pulumi.input_type class MdbRedisClusterConfigArgs: def __init__(__self__, *, password: pulumi.Input[str], version: pulumi.Input[str], databases: Optional[pulumi.Input[int]] = None, maxmemory_policy: Optional[pulumi.Input[str]] = None, notify_keyspace_events: Optional[pulumi.Input[str]] = None, slowlog_log_slower_than: Optional[pulumi.Input[int]] = None, slowlog_max_len: Optional[pulumi.Input[int]] = None, timeout: Optional[pulumi.Input[int]] = None): """ :param pulumi.Input[str] password: Password for the Redis cluster. :param pulumi.Input[str] version: Version of Redis (5.0, 6.0 or 6.2). :param pulumi.Input[int] databases: Number of databases (changing requires redis-server restart). :param pulumi.Input[str] maxmemory_policy: Redis key eviction policy for a dataset that reaches maximum memory. Can be any of the listed in [the official RedisDB documentation](https://docs.redislabs.com/latest/rs/administering/database-operations/eviction-policy/). :param pulumi.Input[str] notify_keyspace_events: Select the events that Redis will notify among a set of classes. :param pulumi.Input[int] slowlog_log_slower_than: Log slow queries below this number in microseconds. :param pulumi.Input[int] slowlog_max_len: Slow queries log length. :param pulumi.Input[int] timeout: Close the connection after a client is idle for N seconds. """ pulumi.set(__self__, "password", password) pulumi.set(__self__, "version", version) if databases is not None: pulumi.set(__self__, "databases", databases) if maxmemory_policy is not None: pulumi.set(__self__, "maxmemory_policy", maxmemory_policy) if notify_keyspace_events is not None: pulumi.set(__self__, "notify_keyspace_events", notify_keyspace_events) if slowlog_log_slower_than is not None: pulumi.set(__self__, "slowlog_log_slower_than", slowlog_log_slower_than) if slowlog_max_len is not None: pulumi.set(__self__, "slowlog_max_len", slowlog_max_len) if timeout is not None: pulumi.set(__self__, "timeout", timeout) @property @pulumi.getter def password(self) -> pulumi.Input[str]: """ Password for the Redis cluster. """ return pulumi.get(self, "password") @password.setter def password(self, value: pulumi.Input[str]): pulumi.set(self, "password", value) @property @pulumi.getter def version(self) -> pulumi.Input[str]: """ Version of Redis (5.0, 6.0 or 6.2). """ return pulumi.get(self, "version") @version.setter def version(self, value: pulumi.Input[str]): pulumi.set(self, "version", value) @property @pulumi.getter def databases(self) -> Optional[pulumi.Input[int]]: """ Number of databases (changing requires redis-server restart). """ return pulumi.get(self, "databases") @databases.setter def databases(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "databases", value) @property @pulumi.getter(name="maxmemoryPolicy") def maxmemory_policy(self) -> Optional[pulumi.Input[str]]: """ Redis key eviction policy for a dataset that reaches maximum memory. Can be any of the listed in [the official RedisDB documentation](https://docs.redislabs.com/latest/rs/administering/database-operations/eviction-policy/). """ return pulumi.get(self, "maxmemory_policy") @maxmemory_policy.setter def maxmemory_policy(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "maxmemory_policy", value) @property @pulumi.getter(name="notifyKeyspaceEvents") def notify_keyspace_events(self) -> Optional[pulumi.Input[str]]: """ Select the events that Redis will notify among a set of classes. """ return pulumi.get(self, "notify_keyspace_events") @notify_keyspace_events.setter def notify_keyspace_events(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "notify_keyspace_events", value) @property @pulumi.getter(name="slowlogLogSlowerThan") def slowlog_log_slower_than(self) -> Optional[pulumi.Input[int]]: """ Log slow queries below this number in microseconds. """ return pulumi.get(self, "slowlog_log_slower_than") @slowlog_log_slower_than.setter def slowlog_log_slower_than(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "slowlog_log_slower_than", value) @property @pulumi.getter(name="slowlogMaxLen") def slowlog_max_len(self) -> Optional[pulumi.Input[int]]: """ Slow queries log length. """ return pulumi.get(self, "slowlog_max_len") @slowlog_max_len.setter def slowlog_max_len(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "slowlog_max_len", value) @property @pulumi.getter def timeout(self) -> Optional[pulumi.Input[int]]: """ Close the connection after a client is idle for N seconds. """ return pulumi.get(self, "timeout") @timeout.setter def timeout(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "timeout", value) @pulumi.input_type class MdbRedisClusterHostArgs: def __init__(__self__, *, zone: pulumi.Input[str], fqdn: Optional[pulumi.Input[str]] = None, shard_name: Optional[pulumi.Input[str]] = None, subnet_id: Optional[pulumi.Input[str]] = None): """ :param pulumi.Input[str] zone: The availability zone where the Redis host will be created. For more information see [the official documentation](https://cloud.yandex.com/docs/overview/concepts/geo-scope). :param pulumi.Input[str] fqdn: The fully qualified domain name of the host. :param pulumi.Input[str] shard_name: The name of the shard to which the host belongs. :param pulumi.Input[str] subnet_id: The ID of the subnet, to which the host belongs. The subnet must be a part of the network to which the cluster belongs. """ pulumi.set(__self__, "zone", zone) if fqdn is not None: pulumi.set(__self__, "fqdn", fqdn) if shard_name is not None: pulumi.set(__self__, "shard_name", shard_name) if subnet_id is not None: pulumi.set(__self__, "subnet_id", subnet_id) @property @pulumi.getter def zone(self) -> pulumi.Input[str]: """ The availability zone where the Redis host will be created. For more information see [the official documentation](https://cloud.yandex.com/docs/overview/concepts/geo-scope). """ return pulumi.get(self, "zone") @zone.setter def zone(self, value: pulumi.Input[str]): pulumi.set(self, "zone", value) @property @pulumi.getter def fqdn(self) -> Optional[pulumi.Input[str]]: """ The fully qualified domain name of the host. """ return pulumi.get(self, "fqdn") @fqdn.setter def fqdn(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "fqdn", value) @property @pulumi.getter(name="shardName") def shard_name(self) -> Optional[pulumi.Input[str]]: """ The name of the shard to which the host belongs. """ return pulumi.get(self, "shard_name") @shard_name.setter def shard_name(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "shard_name", value) @property @pulumi.getter(name="subnetId") def subnet_id(self) -> Optional[pulumi.Input[str]]: """ The ID of the subnet, to which the host belongs. The subnet must be a part of the network to which the cluster belongs. """ return pulumi.get(self, "subnet_id") @subnet_id.setter def subnet_id(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "subnet_id", value) @pulumi.input_type class MdbRedisClusterMaintenanceWindowArgs: def __init__(__self__, *, type: pulumi.Input[str], day: Optional[pulumi.Input[str]] = None, hour: Optional[pulumi.Input[int]] = None): """ :param pulumi.Input[str] type: Type of maintenance window. Can be either `ANYTIME` or `WEEKLY`. A day and hour of window need to be specified with weekly window. :param pulumi.Input[str] day: Day of week for maintenance window if window type is weekly. Possible values: `MON`, `TUE`, `WED`, `THU`, `FRI`, `SAT`, `SUN`. :param pulumi.Input[int] hour: Hour of day in UTC time zone (1-24) for maintenance window if window type is weekly. """ pulumi.set(__self__, "type", type) if day is not None: pulumi.set(__self__, "day", day) if hour is not None: pulumi.set(__self__, "hour", hour) @property @pulumi.getter def type(self) -> pulumi.Input[str]: """ Type of maintenance window. Can be either `ANYTIME` or `WEEKLY`. A day and hour of window need to be specified with weekly window. """ return pulumi.get(self, "type") @type.setter def type(self, value: pulumi.Input[str]): pulumi.set(self, "type", value) @property @pulumi.getter def day(self) -> Optional[pulumi.Input[str]]: """ Day of week for maintenance window if window type is weekly. Possible values: `MON`, `TUE`, `WED`, `THU`, `FRI`, `SAT`, `SUN`. """ return pulumi.get(self, "day") @day.setter def day(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "day", value) @property @pulumi.getter def hour(self) -> Optional[pulumi.Input[int]]: """ Hour of day in UTC time zone (1-24) for maintenance window if window type is weekly. """ return pulumi.get(self, "hour") @hour.setter def hour(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "hour", value) @pulumi.input_type class MdbRedisClusterResourcesArgs: def __init__(__self__, *, disk_size: pulumi.Input[int], resource_preset_id: pulumi.Input[str], disk_type_id: Optional[pulumi.Input[str]] = None): """ :param pulumi.Input[int] disk_size: Volume of the storage available to a host, in gigabytes. :param pulumi.Input[str] disk_type_id: Type of the storage of Redis hosts - environment default is used if missing. """ pulumi.set(__self__, "disk_size", disk_size) pulumi.set(__self__, "resource_preset_id", resource_preset_id) if disk_type_id is not None: pulumi.set(__self__, "disk_type_id", disk_type_id) @property @pulumi.getter(name="diskSize") def disk_size(self) -> pulumi.Input[int]: """ Volume of the storage available to a host, in gigabytes. """ return pulumi.get(self, "disk_size") @disk_size.setter def disk_size(self, value: pulumi.Input[int]): pulumi.set(self, "disk_size", value) @property @pulumi.getter(name="resourcePresetId") def resource_preset_id(self) -> pulumi.Input[str]: return pulumi.get(self, "resource_preset_id") @resource_preset_id.setter def resource_preset_id(self, value: pulumi.Input[str]): pulumi.set(self, "resource_preset_id", value) @property @pulumi.getter(name="diskTypeId") def disk_type_id(self) -> Optional[pulumi.Input[str]]: """ Type of the storage of Redis hosts - environment default is used if missing. """ return pulumi.get(self, "disk_type_id") @disk_type_id.setter def disk_type_id(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "disk_type_id", value) @pulumi.input_type class MdbSqlServerClusterBackupWindowStartArgs: def __init__(__self__, *, hours: Optional[pulumi.Input[int]] = None, minutes: Optional[pulumi.Input[int]] = None): """ :param pulumi.Input[int] hours: The hour at which backup will be started. :param pulumi.Input[int] minutes: The minute at which backup will be started. """ if hours is not None: pulumi.set(__self__, "hours", hours) if minutes is not None: pulumi.set(__self__, "minutes", minutes) @property @pulumi.getter def hours(self) -> Optional[pulumi.Input[int]]: """ The hour at which backup will be started. """ return pulumi.get(self, "hours") @hours.setter def hours(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "hours", value) @property @pulumi.getter def minutes(self) -> Optional[pulumi.Input[int]]: """ The minute at which backup will be started. """ return pulumi.get(self, "minutes") @minutes.setter def minutes(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "minutes", value) @pulumi.input_type class MdbSqlServerClusterDatabaseArgs: def __init__(__self__, *, name: pulumi.Input[str]): """ :param pulumi.Input[str] name: The name of the database. """ pulumi.set(__self__, "name", name) @property @pulumi.getter def name(self) -> pulumi.Input[str]: """ The name of the database. """ return pulumi.get(self, "name") @name.setter def name(self, value: pulumi.Input[str]): pulumi.set(self, "name", value) @pulumi.input_type class MdbSqlServerClusterHostArgs: def __init__(__self__, *, zone: pulumi.Input[str], assign_public_ip: Optional[pulumi.Input[bool]] = None, fqdn: Optional[pulumi.Input[str]] = None, subnet_id: Optional[pulumi.Input[str]] = None): """ :param pulumi.Input[str] zone: The availability zone where the SQLServer host will be created. :param pulumi.Input[bool] assign_public_ip: Sets whether the host should get a public IP address on creation. Changing this parameter for an existing host is not supported at the moment :param pulumi.Input[str] fqdn: The fully qualified domain name of the host. :param pulumi.Input[str] subnet_id: The ID of the subnet, to which the host belongs. The subnet must be a part of the network to which the cluster belongs. """ pulumi.set(__self__, "zone", zone) if assign_public_ip is not None: pulumi.set(__self__, "assign_public_ip", assign_public_ip) if fqdn is not None: pulumi.set(__self__, "fqdn", fqdn) if subnet_id is not None: pulumi.set(__self__, "subnet_id", subnet_id) @property @pulumi.getter def zone(self) -> pulumi.Input[str]: """ The availability zone where the SQLServer host will be created. """ return pulumi.get(self, "zone") @zone.setter def zone(self, value: pulumi.Input[str]): pulumi.set(self, "zone", value) @property @pulumi.getter(name="assignPublicIp") def assign_public_ip(self) -> Optional[pulumi.Input[bool]]: """ Sets whether the host should get a public IP address on creation. Changing this parameter for an existing host is not supported at the moment """ return pulumi.get(self, "assign_public_ip") @assign_public_ip.setter def assign_public_ip(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "assign_public_ip", value) @property @pulumi.getter def fqdn(self) -> Optional[pulumi.Input[str]]: """ The fully qualified domain name of the host. """ return pulumi.get(self, "fqdn") @fqdn.setter def fqdn(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "fqdn", value) @property @pulumi.getter(name="subnetId") def subnet_id(self) -> Optional[pulumi.Input[str]]: """ The ID of the subnet, to which the host belongs. The subnet must be a part of the network to which the cluster belongs. """ return pulumi.get(self, "subnet_id") @subnet_id.setter def subnet_id(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "subnet_id", value) @pulumi.input_type class MdbSqlServerClusterResourcesArgs: def __init__(__self__, *, disk_size: pulumi.Input[int], disk_type_id: pulumi.Input[str], resource_preset_id: pulumi.Input[str]): """ :param pulumi.Input[int] disk_size: Volume of the storage available to a SQLServer host, in gigabytes. :param pulumi.Input[str] disk_type_id: Type of the storage of SQLServer hosts. """ pulumi.set(__self__, "disk_size", disk_size) pulumi.set(__self__, "disk_type_id", disk_type_id) pulumi.set(__self__, "resource_preset_id", resource_preset_id) @property @pulumi.getter(name="diskSize") def disk_size(self) -> pulumi.Input[int]: """ Volume of the storage available to a SQLServer host, in gigabytes. """ return pulumi.get(self, "disk_size") @disk_size.setter def disk_size(self, value: pulumi.Input[int]): pulumi.set(self, "disk_size", value) @property @pulumi.getter(name="diskTypeId") def disk_type_id(self) -> pulumi.Input[str]: """ Type of the storage of SQLServer hosts. """ return pulumi.get(self, "disk_type_id") @disk_type_id.setter def disk_type_id(self, value: pulumi.Input[str]): pulumi.set(self, "disk_type_id", value) @property @pulumi.getter(name="resourcePresetId") def resource_preset_id(self) -> pulumi.Input[str]: return pulumi.get(self, "resource_preset_id") @resource_preset_id.setter def resource_preset_id(self, value: pulumi.Input[str]): pulumi.set(self, "resource_preset_id", value) @pulumi.input_type class MdbSqlServerClusterUserArgs: def __init__(__self__, *, name: pulumi.Input[str], password: pulumi.Input[str], permissions: Optional[pulumi.Input[Sequence[pulumi.Input['MdbSqlServerClusterUserPermissionArgs']]]] = None): """ :param pulumi.Input[str] name: The name of the database. :param pulumi.Input[str] password: <PASSWORD>. :param pulumi.Input[Sequence[pulumi.Input['MdbSqlServerClusterUserPermissionArgs']]] permissions: Set of permissions granted to the user. The structure is documented below. """ pulumi.set(__self__, "name", name) pulumi.set(__self__, "password", password) if permissions is not None: pulumi.set(__self__, "permissions", permissions) @property @pulumi.getter def name(self) -> pulumi.Input[str]: """ The name of the database. """ return pulumi.get(self, "name") @name.setter def name(self, value: pulumi.Input[str]): pulumi.set(self, "name", value) @property @pulumi.getter def password(self) -> pulumi.Input[str]: """ The password of the user. """ return pulumi.get(self, "password") @password.setter def password(self, value: pulumi.Input[str]): pulumi.set(self, "password", value) @property @pulumi.getter def permissions(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['MdbSqlServerClusterUserPermissionArgs']]]]: """ Set of permissions granted to the user. The structure is documented below. """ return pulumi.get(self, "permissions") @permissions.setter def permissions(self, value: Optional[pulumi.Input[Sequence[pulumi.Input['MdbSqlServerClusterUserPermissionArgs']]]]): pulumi.set(self, "permissions", value) @pulumi.input_type class MdbSqlServerClusterUserPermissionArgs: def __init__(__self__, *, database_name: pulumi.Input[str], roles: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None): """ :param pulumi.Input[str] database_name: The name of the database that the permission grants access to. :param pulumi.Input[Sequence[pulumi.Input[str]]] roles: List user's roles in the database. Allowed roles: `OWNER`, `SECURITYADMIN`, `ACCESSADMIN`, `BACKUPOPERATOR`, `DDLADMIN`, `DATAWRITER`, `DATAREADER`, `DENYDATAWRITER`, `DENYDATAREADER`. """ pulumi.set(__self__, "database_name", database_name) if roles is not None: pulumi.set(__self__, "roles", roles) @property @pulumi.getter(name="databaseName") def database_name(self) -> pulumi.Input[str]: """ The name of the database that the permission grants access to. """ return pulumi.get(self, "database_name") @database_name.setter def database_name(self, value: pulumi.Input[str]): pulumi.set(self, "database_name", value) @property @pulumi.getter def roles(self) -> Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]: """ List user's roles in the database. Allowed roles: `OWNER`, `SECURITYADMIN`, `ACCESSADMIN`, `BACKUPOPERATOR`, `DDLADMIN`, `DATAWRITER`, `DATAREADER`, `DENYDATAWRITER`, `DENYDATAREADER`. """ return pulumi.get(self, "roles") @roles.setter def roles(self, value: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]): pulumi.set(self, "roles", value) @pulumi.input_type class StorageBucketCorsRuleArgs: def __init__(__self__, *, allowed_methods: pulumi.Input[Sequence[pulumi.Input[str]]], allowed_origins: pulumi.Input[Sequence[pulumi.Input[str]]], allowed_headers: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, expose_headers: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, max_age_seconds: Optional[pulumi.Input[int]] = None): """ :param pulumi.Input[Sequence[pulumi.Input[str]]] allowed_methods: Specifies which methods are allowed. Can be `GET`, `PUT`, `POST`, `DELETE` or `HEAD`. :param pulumi.Input[Sequence[pulumi.Input[str]]] allowed_origins: Specifies which origins are allowed. :param pulumi.Input[Sequence[pulumi.Input[str]]] allowed_headers: Specifies which headers are allowed. :param pulumi.Input[Sequence[pulumi.Input[str]]] expose_headers: Specifies expose header in the response. :param pulumi.Input[int] max_age_seconds: Specifies time in seconds that browser can cache the response for a preflight request. """ pulumi.set(__self__, "allowed_methods", allowed_methods) pulumi.set(__self__, "allowed_origins", allowed_origins) if allowed_headers is not None: pulumi.set(__self__, "allowed_headers", allowed_headers) if expose_headers is not None: pulumi.set(__self__, "expose_headers", expose_headers) if max_age_seconds is not None: pulumi.set(__self__, "max_age_seconds", max_age_seconds) @property @pulumi.getter(name="allowedMethods") def allowed_methods(self) -> pulumi.Input[Sequence[pulumi.Input[str]]]: """ Specifies which methods are allowed. Can be `GET`, `PUT`, `POST`, `DELETE` or `HEAD`. """ return pulumi.get(self, "allowed_methods") @allowed_methods.setter def allowed_methods(self, value: pulumi.Input[Sequence[pulumi.Input[str]]]): pulumi.set(self, "allowed_methods", value) @property @pulumi.getter(name="allowedOrigins") def allowed_origins(self) -> pulumi.Input[Sequence[pulumi.Input[str]]]: """ Specifies which origins are allowed. """ return pulumi.get(self, "allowed_origins") @allowed_origins.setter def allowed_origins(self, value: pulumi.Input[Sequence[pulumi.Input[str]]]): pulumi.set(self, "allowed_origins", value) @property @pulumi.getter(name="allowedHeaders") def allowed_headers(self) -> Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]: """ Specifies which headers are allowed. """ return pulumi.get(self, "allowed_headers") @allowed_headers.setter def allowed_headers(self, value: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]): pulumi.set(self, "allowed_headers", value) @property @pulumi.getter(name="exposeHeaders") def expose_headers(self) -> Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]: """ Specifies expose header in the response. """ return pulumi.get(self, "expose_headers") @expose_headers.setter def expose_headers(self, value: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]): pulumi.set(self, "expose_headers", value) @property @pulumi.getter(name="maxAgeSeconds") def max_age_seconds(self) -> Optional[pulumi.Input[int]]: """ Specifies time in seconds that browser can cache the response for a preflight request. """ return pulumi.get(self, "max_age_seconds") @max_age_seconds.setter def max_age_seconds(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "max_age_seconds", value) @pulumi.input_type class StorageBucketGrantArgs: def __init__(__self__, *, permissions: pulumi.Input[Sequence[pulumi.Input[str]]], type: pulumi.Input[str], id: Optional[pulumi.Input[str]] = None, uri: Optional[pulumi.Input[str]] = None): """ :param pulumi.Input[str] id: Unique identifier for the rule. Must be less than or equal to 255 characters in length. """ pulumi.set(__self__, "permissions", permissions) pulumi.set(__self__, "type", type) if id is not None: pulumi.set(__self__, "id", id) if uri is not None: pulumi.set(__self__, "uri", uri) @property @pulumi.getter def permissions(self) -> pulumi.Input[Sequence[pulumi.Input[str]]]: return pulumi.get(self, "permissions") @permissions.setter def permissions(self, value: pulumi.Input[Sequence[pulumi.Input[str]]]): pulumi.set(self, "permissions", value) @property @pulumi.getter def type(self) -> pulumi.Input[str]: return pulumi.get(self, "type") @type.setter def type(self, value: pulumi.Input[str]): pulumi.set(self, "type", value) @property @pulumi.getter def id(self) -> Optional[pulumi.Input[str]]: """ Unique identifier for the rule. Must be less than or equal to 255 characters in length. """ return pulumi.get(self, "id") @id.setter def id(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "id", value) @property @pulumi.getter def uri(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "uri") @uri.setter def uri(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "uri", value) @pulumi.input_type class StorageBucketLifecycleRuleArgs: def __init__(__self__, *, enabled: pulumi.Input[bool], abort_incomplete_multipart_upload_days: Optional[pulumi.Input[int]] = None, expiration: Optional[pulumi.Input['StorageBucketLifecycleRuleExpirationArgs']] = None, id: Optional[pulumi.Input[str]] = None, noncurrent_version_expiration: Optional[pulumi.Input['StorageBucketLifecycleRuleNoncurrentVersionExpirationArgs']] = None, noncurrent_version_transitions: Optional[pulumi.Input[Sequence[pulumi.Input['StorageBucketLifecycleRuleNoncurrentVersionTransitionArgs']]]] = None, prefix: Optional[pulumi.Input[str]] = None, transitions: Optional[pulumi.Input[Sequence[pulumi.Input['StorageBucketLifecycleRuleTransitionArgs']]]] = None): """ :param pulumi.Input[bool] enabled: Specifies lifecycle rule status. :param pulumi.Input[int] abort_incomplete_multipart_upload_days: Specifies the number of days after initiating a multipart upload when the multipart upload must be completed. :param pulumi.Input['StorageBucketLifecycleRuleExpirationArgs'] expiration: Specifies a period in the object's expire (documented below). :param pulumi.Input[str] id: Unique identifier for the rule. Must be less than or equal to 255 characters in length. :param pulumi.Input['StorageBucketLifecycleRuleNoncurrentVersionExpirationArgs'] noncurrent_version_expiration: Specifies when noncurrent object versions expire (documented below). :param pulumi.Input[Sequence[pulumi.Input['StorageBucketLifecycleRuleNoncurrentVersionTransitionArgs']]] noncurrent_version_transitions: Specifies when noncurrent object versions transitions (documented below). :param pulumi.Input[str] prefix: Object key prefix identifying one or more objects to which the rule applies. :param pulumi.Input[Sequence[pulumi.Input['StorageBucketLifecycleRuleTransitionArgs']]] transitions: Specifies a period in the object's transitions (documented below). """ pulumi.set(__self__, "enabled", enabled) if abort_incomplete_multipart_upload_days is not None: pulumi.set(__self__, "abort_incomplete_multipart_upload_days", abort_incomplete_multipart_upload_days) if expiration is not None: pulumi.set(__self__, "expiration", expiration) if id is not None: pulumi.set(__self__, "id", id) if noncurrent_version_expiration is not None: pulumi.set(__self__, "noncurrent_version_expiration", noncurrent_version_expiration) if noncurrent_version_transitions is not None: pulumi.set(__self__, "noncurrent_version_transitions", noncurrent_version_transitions) if prefix is not None: pulumi.set(__self__, "prefix", prefix) if transitions is not None: pulumi.set(__self__, "transitions", transitions) @property @pulumi.getter def enabled(self) -> pulumi.Input[bool]: """ Specifies lifecycle rule status. """ return pulumi.get(self, "enabled") @enabled.setter def enabled(self, value: pulumi.Input[bool]): pulumi.set(self, "enabled", value) @property @pulumi.getter(name="abortIncompleteMultipartUploadDays") def abort_incomplete_multipart_upload_days(self) -> Optional[pulumi.Input[int]]: """ Specifies the number of days after initiating a multipart upload when the multipart upload must be completed. """ return pulumi.get(self, "abort_incomplete_multipart_upload_days") @abort_incomplete_multipart_upload_days.setter def abort_incomplete_multipart_upload_days(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "abort_incomplete_multipart_upload_days", value) @property @pulumi.getter def expiration(self) -> Optional[pulumi.Input['StorageBucketLifecycleRuleExpirationArgs']]: """ Specifies a period in the object's expire (documented below). """ return pulumi.get(self, "expiration") @expiration.setter def expiration(self, value: Optional[pulumi.Input['StorageBucketLifecycleRuleExpirationArgs']]): pulumi.set(self, "expiration", value) @property @pulumi.getter def id(self) -> Optional[pulumi.Input[str]]: """ Unique identifier for the rule. Must be less than or equal to 255 characters in length. """ return pulumi.get(self, "id") @id.setter def id(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "id", value) @property @pulumi.getter(name="noncurrentVersionExpiration") def noncurrent_version_expiration(self) -> Optional[pulumi.Input['StorageBucketLifecycleRuleNoncurrentVersionExpirationArgs']]: """ Specifies when noncurrent object versions expire (documented below). """ return pulumi.get(self, "noncurrent_version_expiration") @noncurrent_version_expiration.setter def noncurrent_version_expiration(self, value: Optional[pulumi.Input['StorageBucketLifecycleRuleNoncurrentVersionExpirationArgs']]): pulumi.set(self, "noncurrent_version_expiration", value) @property @pulumi.getter(name="noncurrentVersionTransitions") def noncurrent_version_transitions(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['StorageBucketLifecycleRuleNoncurrentVersionTransitionArgs']]]]: """ Specifies when noncurrent object versions transitions (documented below). """ return pulumi.get(self, "noncurrent_version_transitions") @noncurrent_version_transitions.setter def noncurrent_version_transitions(self, value: Optional[pulumi.Input[Sequence[pulumi.Input['StorageBucketLifecycleRuleNoncurrentVersionTransitionArgs']]]]): pulumi.set(self, "noncurrent_version_transitions", value) @property @pulumi.getter def prefix(self) -> Optional[pulumi.Input[str]]: """ Object key prefix identifying one or more objects to which the rule applies. """ return pulumi.get(self, "prefix") @prefix.setter def prefix(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "prefix", value) @property @pulumi.getter def transitions(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['StorageBucketLifecycleRuleTransitionArgs']]]]: """ Specifies a period in the object's transitions (documented below). """ return pulumi.get(self, "transitions") @transitions.setter def transitions(self, value: Optional[pulumi.Input[Sequence[pulumi.Input['StorageBucketLifecycleRuleTransitionArgs']]]]): pulumi.set(self, "transitions", value) @pulumi.input_type class StorageBucketLifecycleRuleExpirationArgs: def __init__(__self__, *, date: Optional[pulumi.Input[str]] = None, days: Optional[pulumi.Input[int]] = None, expired_object_delete_marker: Optional[pulumi.Input[bool]] = None): """ :param pulumi.Input[str] date: Specifies the date after which you want the corresponding action to take effect. :param pulumi.Input[int] days: Specifies the number of days after object creation when the specific rule action takes effect. :param pulumi.Input[bool] expired_object_delete_marker: On a versioned bucket (versioning-enabled or versioning-suspended bucket), you can add this element in the lifecycle configuration to direct Object Storage to delete expired object delete markers. """ if date is not None: pulumi.set(__self__, "date", date) if days is not None: pulumi.set(__self__, "days", days) if expired_object_delete_marker is not None: pulumi.set(__self__, "expired_object_delete_marker", expired_object_delete_marker) @property @pulumi.getter def date(self) -> Optional[pulumi.Input[str]]: """ Specifies the date after which you want the corresponding action to take effect. """ return pulumi.get(self, "date") @date.setter def date(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "date", value) @property @pulumi.getter def days(self) -> Optional[pulumi.Input[int]]: """ Specifies the number of days after object creation when the specific rule action takes effect. """ return pulumi.get(self, "days") @days.setter def days(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "days", value) @property @pulumi.getter(name="expiredObjectDeleteMarker") def expired_object_delete_marker(self) -> Optional[pulumi.Input[bool]]: """ On a versioned bucket (versioning-enabled or versioning-suspended bucket), you can add this element in the lifecycle configuration to direct Object Storage to delete expired object delete markers. """ return pulumi.get(self, "expired_object_delete_marker") @expired_object_delete_marker.setter def expired_object_delete_marker(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "expired_object_delete_marker", value) @pulumi.input_type class StorageBucketLifecycleRuleNoncurrentVersionExpirationArgs: def __init__(__self__, *, days: Optional[pulumi.Input[int]] = None): """ :param pulumi.Input[int] days: Specifies the number of days noncurrent object versions expire. """ if days is not None: pulumi.set(__self__, "days", days) @property @pulumi.getter def days(self) -> Optional[pulumi.Input[int]]: """ Specifies the number of days noncurrent object versions expire. """ return pulumi.get(self, "days") @days.setter def days(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "days", value) @pulumi.input_type class StorageBucketLifecycleRuleNoncurrentVersionTransitionArgs: def __init__(__self__, *, storage_class: pulumi.Input[str], days: Optional[pulumi.Input[int]] = None): """ :param pulumi.Input[str] storage_class: Specifies the storage class to which you want the noncurrent object versions to transition. Can only be `COLD` or `STANDARD_IA`. :param pulumi.Input[int] days: Specifies the number of days noncurrent object versions transition. """ pulumi.set(__self__, "storage_class", storage_class) if days is not None: pulumi.set(__self__, "days", days) @property @pulumi.getter(name="storageClass") def storage_class(self) -> pulumi.Input[str]: """ Specifies the storage class to which you want the noncurrent object versions to transition. Can only be `COLD` or `STANDARD_IA`. """ return pulumi.get(self, "storage_class") @storage_class.setter def storage_class(self, value: pulumi.Input[str]): pulumi.set(self, "storage_class", value) @property @pulumi.getter def days(self) -> Optional[pulumi.Input[int]]: """ Specifies the number of days noncurrent object versions transition. """ return pulumi.get(self, "days") @days.setter def days(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "days", value) @pulumi.input_type class StorageBucketLifecycleRuleTransitionArgs: def __init__(__self__, *, storage_class: pulumi.Input[str], date: Optional[pulumi.Input[str]] = None, days: Optional[pulumi.Input[int]] = None): """ :param pulumi.Input[str] storage_class: Specifies the storage class to which you want the object to transition. Can only be `COLD` or `STANDARD_IA`. :param pulumi.Input[str] date: Specifies the date after which you want the corresponding action to take effect. :param pulumi.Input[int] days: Specifies the number of days after object creation when the specific rule action takes effect. """ pulumi.set(__self__, "storage_class", storage_class) if date is not None: pulumi.set(__self__, "date", date) if days is not None: pulumi.set(__self__, "days", days) @property @pulumi.getter(name="storageClass") def storage_class(self) -> pulumi.Input[str]: """ Specifies the storage class to which you want the object to transition. Can only be `COLD` or `STANDARD_IA`. """ return pulumi.get(self, "storage_class") @storage_class.setter def storage_class(self, value: pulumi.Input[str]): pulumi.set(self, "storage_class", value) @property @pulumi.getter def date(self) -> Optional[pulumi.Input[str]]: """ Specifies the date after which you want the corresponding action to take effect. """ return pulumi.get(self, "date") @date.setter def date(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "date", value) @property @pulumi.getter def days(self) -> Optional[pulumi.Input[int]]: """ Specifies the number of days after object creation when the specific rule action takes effect. """ return pulumi.get(self, "days") @days.setter def days(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "days", value) @pulumi.input_type class StorageBucketLoggingArgs: def __init__(__self__, *, target_bucket: pulumi.Input[str], target_prefix: Optional[pulumi.Input[str]] = None): """ :param pulumi.Input[str] target_bucket: The name of the bucket that will receive the log objects. :param pulumi.Input[str] target_prefix: To specify a key prefix for log objects. """ pulumi.set(__self__, "target_bucket", target_bucket) if target_prefix is not None: pulumi.set(__self__, "target_prefix", target_prefix) @property @pulumi.getter(name="targetBucket") def target_bucket(self) -> pulumi.Input[str]: """ The name of the bucket that will receive the log objects. """ return pulumi.get(self, "target_bucket") @target_bucket.setter def target_bucket(self, value: pulumi.Input[str]): pulumi.set(self, "target_bucket", value) @property @pulumi.getter(name="targetPrefix") def target_prefix(self) -> Optional[pulumi.Input[str]]: """ To specify a key prefix for log objects. """ return pulumi.get(self, "target_prefix") @target_prefix.setter def target_prefix(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "target_prefix", value) @pulumi.input_type class StorageBucketServerSideEncryptionConfigurationArgs: def __init__(__self__, *, rule: pulumi.Input['StorageBucketServerSideEncryptionConfigurationRuleArgs']): """ :param pulumi.Input['StorageBucketServerSideEncryptionConfigurationRuleArgs'] rule: A single object for server-side encryption by default configuration. (documented below) """ pulumi.set(__self__, "rule", rule) @property @pulumi.getter def rule(self) -> pulumi.Input['StorageBucketServerSideEncryptionConfigurationRuleArgs']: """ A single object for server-side encryption by default configuration. (documented below) """ return pulumi.get(self, "rule") @rule.setter def rule(self, value: pulumi.Input['StorageBucketServerSideEncryptionConfigurationRuleArgs']): pulumi.set(self, "rule", value) @pulumi.input_type class StorageBucketServerSideEncryptionConfigurationRuleArgs: def __init__(__self__, *, apply_server_side_encryption_by_default: pulumi.Input['StorageBucketServerSideEncryptionConfigurationRuleApplyServerSideEncryptionByDefaultArgs']): """ :param pulumi.Input['StorageBucketServerSideEncryptionConfigurationRuleApplyServerSideEncryptionByDefaultArgs'] apply_server_side_encryption_by_default: A single object for setting server-side encryption by default. (documented below) """ pulumi.set(__self__, "apply_server_side_encryption_by_default", apply_server_side_encryption_by_default) @property @pulumi.getter(name="applyServerSideEncryptionByDefault") def apply_server_side_encryption_by_default(self) -> pulumi.Input['StorageBucketServerSideEncryptionConfigurationRuleApplyServerSideEncryptionByDefaultArgs']: """ A single object for setting server-side encryption by default. (documented below) """ return pulumi.get(self, "apply_server_side_encryption_by_default") @apply_server_side_encryption_by_default.setter def apply_server_side_encryption_by_default(self, value: pulumi.Input['StorageBucketServerSideEncryptionConfigurationRuleApplyServerSideEncryptionByDefaultArgs']): pulumi.set(self, "apply_server_side_encryption_by_default", value) @pulumi.input_type class StorageBucketServerSideEncryptionConfigurationRuleApplyServerSideEncryptionByDefaultArgs: def __init__(__self__, *, kms_master_key_id: pulumi.Input[str], sse_algorithm: pulumi.Input[str]): """ :param pulumi.Input[str] kms_master_key_id: The KMS master key ID used for the SSE-KMS encryption. :param pulumi.Input[str] sse_algorithm: The server-side encryption algorithm to use. Single valid value is `aws:kms` """ pulumi.set(__self__, "kms_master_key_id", kms_master_key_id) pulumi.set(__self__, "sse_algorithm", sse_algorithm) @property @pulumi.getter(name="kmsMasterKeyId") def kms_master_key_id(self) -> pulumi.Input[str]: """ The KMS master key ID used for the SSE-KMS encryption. """ return pulumi.get(self, "kms_master_key_id") @kms_master_key_id.setter def kms_master_key_id(self, value: pulumi.Input[str]): pulumi.set(self, "kms_master_key_id", value) @property @pulumi.getter(name="sseAlgorithm") def sse_algorithm(self) -> pulumi.Input[str]: """ The server-side encryption algorithm to use. Single valid value is `aws:kms` """ return pulumi.get(self, "sse_algorithm") @sse_algorithm.setter def sse_algorithm(self, value: pulumi.Input[str]): pulumi.set(self, "sse_algorithm", value) @pulumi.input_type class StorageBucketVersioningArgs: def __init__(__self__, *, enabled: Optional[pulumi.Input[bool]] = None): """ :param pulumi.Input[bool] enabled: Enable versioning. Once you version-enable a bucket, it can never return to an unversioned state. You can, however, suspend versioning on that bucket. """ if enabled is not None: pulumi.set(__self__, "enabled", enabled) @property @pulumi.getter def enabled(self) -> Optional[pulumi.Input[bool]]: """ Enable versioning. Once you version-enable a bucket, it can never return to an unversioned state. You can, however, suspend versioning on that bucket. """ return pulumi.get(self, "enabled") @enabled.setter def enabled(self, value: Optional[pulumi.Input[bool]]): pulumi.set(self, "enabled", value) @pulumi.input_type class StorageBucketWebsiteArgs: def __init__(__self__, *, error_document: Optional[pulumi.Input[str]] = None, index_document: Optional[pulumi.Input[str]] = None, redirect_all_requests_to: Optional[pulumi.Input[str]] = None, routing_rules: Optional[pulumi.Input[str]] = None): """ :param pulumi.Input[str] error_document: An absolute path to the document to return in case of a 4XX error. :param pulumi.Input[str] index_document: Storage returns this index document when requests are made to the root domain or any of the subfolders. :param pulumi.Input[str] redirect_all_requests_to: A hostname to redirect all website requests for this bucket to. Hostname can optionally be prefixed with a protocol (`http://` or `https://`) to use when redirecting requests. The default is the protocol that is used in the original request. :param pulumi.Input[str] routing_rules: A json array containing [routing rules](https://cloud.yandex.com/docs/storage/s3/api-ref/hosting/upload#request-scheme) describing redirect behavior and when redirects are applied. """ if error_document is not None: pulumi.set(__self__, "error_document", error_document) if index_document is not None: pulumi.set(__self__, "index_document", index_document) if redirect_all_requests_to is not None: pulumi.set(__self__, "redirect_all_requests_to", redirect_all_requests_to) if routing_rules is not None: pulumi.set(__self__, "routing_rules", routing_rules) @property @pulumi.getter(name="errorDocument") def error_document(self) -> Optional[pulumi.Input[str]]: """ An absolute path to the document to return in case of a 4XX error. """ return pulumi.get(self, "error_document") @error_document.setter def error_document(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "error_document", value) @property @pulumi.getter(name="indexDocument") def index_document(self) -> Optional[pulumi.Input[str]]: """ Storage returns this index document when requests are made to the root domain or any of the subfolders. """ return pulumi.get(self, "index_document") @index_document.setter def index_document(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "index_document", value) @property @pulumi.getter(name="redirectAllRequestsTo") def redirect_all_requests_to(self) -> Optional[pulumi.Input[str]]: """ A hostname to redirect all website requests for this bucket to. Hostname can optionally be prefixed with a protocol (`http://` or `https://`) to use when redirecting requests. The default is the protocol that is used in the original request. """ return pulumi.get(self, "redirect_all_requests_to") @redirect_all_requests_to.setter def redirect_all_requests_to(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "redirect_all_requests_to", value) @property @pulumi.getter(name="routingRules") def routing_rules(self) -> Optional[pulumi.Input[str]]: """ A json array containing [routing rules](https://cloud.yandex.com/docs/storage/s3/api-ref/hosting/upload#request-scheme) describing redirect behavior and when redirects are applied. """ return pulumi.get(self, "routing_rules") @routing_rules.setter def routing_rules(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "routing_rules", value) @pulumi.input_type class VpcAddressExternalIpv4AddressArgs: def __init__(__self__, *, address: Optional[pulumi.Input[str]] = None, ddos_protection_provider: Optional[pulumi.Input[str]] = None, outgoing_smtp_capability: Optional[pulumi.Input[str]] = None, zone_id: Optional[pulumi.Input[str]] = None): """ :param pulumi.Input[str] address: - Desired IP. :param pulumi.Input[str] ddos_protection_provider: Enable DDOS protection. Possible values are: "qrator" :param pulumi.Input[str] outgoing_smtp_capability: Wanted outgoing smtp capability. :param pulumi.Input[str] zone_id: - Zone for allocating address. """ if address is not None: pulumi.set(__self__, "address", address) if ddos_protection_provider is not None: pulumi.set(__self__, "ddos_protection_provider", ddos_protection_provider) if outgoing_smtp_capability is not None: pulumi.set(__self__, "outgoing_smtp_capability", outgoing_smtp_capability) if zone_id is not None: pulumi.set(__self__, "zone_id", zone_id) @property @pulumi.getter def address(self) -> Optional[pulumi.Input[str]]: """ - Desired IP. """ return pulumi.get(self, "address") @address.setter def address(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "address", value) @property @pulumi.getter(name="ddosProtectionProvider") def ddos_protection_provider(self) -> Optional[pulumi.Input[str]]: """ Enable DDOS protection. Possible values are: "qrator" """ return pulumi.get(self, "ddos_protection_provider") @ddos_protection_provider.setter def ddos_protection_provider(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "ddos_protection_provider", value) @property @pulumi.getter(name="outgoingSmtpCapability") def outgoing_smtp_capability(self) -> Optional[pulumi.Input[str]]: """ Wanted outgoing smtp capability. """ return pulumi.get(self, "outgoing_smtp_capability") @outgoing_smtp_capability.setter def outgoing_smtp_capability(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "outgoing_smtp_capability", value) @property @pulumi.getter(name="zoneId") def zone_id(self) -> Optional[pulumi.Input[str]]: """ - Zone for allocating address. """ return pulumi.get(self, "zone_id") @zone_id.setter def zone_id(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "zone_id", value) @pulumi.input_type class VpcDefaultSecurityGroupEgressArgs: def __init__(__self__, *, protocol: pulumi.Input[str], description: Optional[pulumi.Input[str]] = None, from_port: Optional[pulumi.Input[int]] = None, id: Optional[pulumi.Input[str]] = None, labels: Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]] = None, port: Optional[pulumi.Input[int]] = None, predefined_target: Optional[pulumi.Input[str]] = None, security_group_id: Optional[pulumi.Input[str]] = None, to_port: Optional[pulumi.Input[int]] = None, v4_cidr_blocks: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, v6_cidr_blocks: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None): """ :param pulumi.Input[str] description: Description of the security group. :param pulumi.Input[str] id: Id of the security group. :param pulumi.Input[Mapping[str, pulumi.Input[str]]] labels: Labels to assign to this security group. """ pulumi.set(__self__, "protocol", protocol) if description is not None: pulumi.set(__self__, "description", description) if from_port is not None: pulumi.set(__self__, "from_port", from_port) if id is not None: pulumi.set(__self__, "id", id) if labels is not None: pulumi.set(__self__, "labels", labels) if port is not None: pulumi.set(__self__, "port", port) if predefined_target is not None: pulumi.set(__self__, "predefined_target", predefined_target) if security_group_id is not None: pulumi.set(__self__, "security_group_id", security_group_id) if to_port is not None: pulumi.set(__self__, "to_port", to_port) if v4_cidr_blocks is not None: pulumi.set(__self__, "v4_cidr_blocks", v4_cidr_blocks) if v6_cidr_blocks is not None: pulumi.set(__self__, "v6_cidr_blocks", v6_cidr_blocks) @property @pulumi.getter def protocol(self) -> pulumi.Input[str]: return pulumi.get(self, "protocol") @protocol.setter def protocol(self, value: pulumi.Input[str]): pulumi.set(self, "protocol", value) @property @pulumi.getter def description(self) -> Optional[pulumi.Input[str]]: """ Description of the security group. """ return pulumi.get(self, "description") @description.setter def description(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "description", value) @property @pulumi.getter(name="fromPort") def from_port(self) -> Optional[pulumi.Input[int]]: return pulumi.get(self, "from_port") @from_port.setter def from_port(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "from_port", value) @property @pulumi.getter def id(self) -> Optional[pulumi.Input[str]]: """ Id of the security group. """ return pulumi.get(self, "id") @id.setter def id(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "id", value) @property @pulumi.getter def labels(self) -> Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]]: """ Labels to assign to this security group. """ return pulumi.get(self, "labels") @labels.setter def labels(self, value: Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]]): pulumi.set(self, "labels", value) @property @pulumi.getter def port(self) -> Optional[pulumi.Input[int]]: return pulumi.get(self, "port") @port.setter def port(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "port", value) @property @pulumi.getter(name="predefinedTarget") def predefined_target(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "predefined_target") @predefined_target.setter def predefined_target(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "predefined_target", value) @property @pulumi.getter(name="securityGroupId") def security_group_id(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "security_group_id") @security_group_id.setter def security_group_id(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "security_group_id", value) @property @pulumi.getter(name="toPort") def to_port(self) -> Optional[pulumi.Input[int]]: return pulumi.get(self, "to_port") @to_port.setter def to_port(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "to_port", value) @property @pulumi.getter(name="v4CidrBlocks") def v4_cidr_blocks(self) -> Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]: return pulumi.get(self, "v4_cidr_blocks") @v4_cidr_blocks.setter def v4_cidr_blocks(self, value: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]): pulumi.set(self, "v4_cidr_blocks", value) @property @pulumi.getter(name="v6CidrBlocks") def v6_cidr_blocks(self) -> Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]: return pulumi.get(self, "v6_cidr_blocks") @v6_cidr_blocks.setter def v6_cidr_blocks(self, value: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]): pulumi.set(self, "v6_cidr_blocks", value) @pulumi.input_type class VpcDefaultSecurityGroupIngressArgs: def __init__(__self__, *, protocol: pulumi.Input[str], description: Optional[pulumi.Input[str]] = None, from_port: Optional[pulumi.Input[int]] = None, id: Optional[pulumi.Input[str]] = None, labels: Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]] = None, port: Optional[pulumi.Input[int]] = None, predefined_target: Optional[pulumi.Input[str]] = None, security_group_id: Optional[pulumi.Input[str]] = None, to_port: Optional[pulumi.Input[int]] = None, v4_cidr_blocks: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, v6_cidr_blocks: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None): """ :param pulumi.Input[str] description: Description of the security group. :param pulumi.Input[str] id: Id of the security group. :param pulumi.Input[Mapping[str, pulumi.Input[str]]] labels: Labels to assign to this security group. """ pulumi.set(__self__, "protocol", protocol) if description is not None: pulumi.set(__self__, "description", description) if from_port is not None: pulumi.set(__self__, "from_port", from_port) if id is not None: pulumi.set(__self__, "id", id) if labels is not None: pulumi.set(__self__, "labels", labels) if port is not None: pulumi.set(__self__, "port", port) if predefined_target is not None: pulumi.set(__self__, "predefined_target", predefined_target) if security_group_id is not None: pulumi.set(__self__, "security_group_id", security_group_id) if to_port is not None: pulumi.set(__self__, "to_port", to_port) if v4_cidr_blocks is not None: pulumi.set(__self__, "v4_cidr_blocks", v4_cidr_blocks) if v6_cidr_blocks is not None: pulumi.set(__self__, "v6_cidr_blocks", v6_cidr_blocks) @property @pulumi.getter def protocol(self) -> pulumi.Input[str]: return pulumi.get(self, "protocol") @protocol.setter def protocol(self, value: pulumi.Input[str]): pulumi.set(self, "protocol", value) @property @pulumi.getter def description(self) -> Optional[pulumi.Input[str]]: """ Description of the security group. """ return pulumi.get(self, "description") @description.setter def description(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "description", value) @property @pulumi.getter(name="fromPort") def from_port(self) -> Optional[pulumi.Input[int]]: return pulumi.get(self, "from_port") @from_port.setter def from_port(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "from_port", value) @property @pulumi.getter def id(self) -> Optional[pulumi.Input[str]]: """ Id of the security group. """ return pulumi.get(self, "id") @id.setter def id(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "id", value) @property @pulumi.getter def labels(self) -> Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]]: """ Labels to assign to this security group. """ return pulumi.get(self, "labels") @labels.setter def labels(self, value: Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]]): pulumi.set(self, "labels", value) @property @pulumi.getter def port(self) -> Optional[pulumi.Input[int]]: return pulumi.get(self, "port") @port.setter def port(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "port", value) @property @pulumi.getter(name="predefinedTarget") def predefined_target(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "predefined_target") @predefined_target.setter def predefined_target(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "predefined_target", value) @property @pulumi.getter(name="securityGroupId") def security_group_id(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "security_group_id") @security_group_id.setter def security_group_id(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "security_group_id", value) @property @pulumi.getter(name="toPort") def to_port(self) -> Optional[pulumi.Input[int]]: return pulumi.get(self, "to_port") @to_port.setter def to_port(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "to_port", value) @property @pulumi.getter(name="v4CidrBlocks") def v4_cidr_blocks(self) -> Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]: return pulumi.get(self, "v4_cidr_blocks") @v4_cidr_blocks.setter def v4_cidr_blocks(self, value: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]): pulumi.set(self, "v4_cidr_blocks", value) @property @pulumi.getter(name="v6CidrBlocks") def v6_cidr_blocks(self) -> Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]: return pulumi.get(self, "v6_cidr_blocks") @v6_cidr_blocks.setter def v6_cidr_blocks(self, value: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]): pulumi.set(self, "v6_cidr_blocks", value) @pulumi.input_type class VpcRouteTableStaticRouteArgs: def __init__(__self__, *, destination_prefix: Optional[pulumi.Input[str]] = None, next_hop_address: Optional[pulumi.Input[str]] = None): """ :param pulumi.Input[str] destination_prefix: Route prefix in CIDR notation. :param pulumi.Input[str] next_hop_address: Address of the next hop. """ if destination_prefix is not None: pulumi.set(__self__, "destination_prefix", destination_prefix) if next_hop_address is not None: pulumi.set(__self__, "next_hop_address", next_hop_address) @property @pulumi.getter(name="destinationPrefix") def destination_prefix(self) -> Optional[pulumi.Input[str]]: """ Route prefix in CIDR notation. """ return pulumi.get(self, "destination_prefix") @destination_prefix.setter def destination_prefix(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "destination_prefix", value) @property @pulumi.getter(name="nextHopAddress") def next_hop_address(self) -> Optional[pulumi.Input[str]]: """ Address of the next hop. """ return pulumi.get(self, "next_hop_address") @next_hop_address.setter def next_hop_address(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "next_hop_address", value) @pulumi.input_type class VpcSecurityGroupEgressArgs: def __init__(__self__, *, protocol: pulumi.Input[str], description: Optional[pulumi.Input[str]] = None, from_port: Optional[pulumi.Input[int]] = None, id: Optional[pulumi.Input[str]] = None, labels: Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]] = None, port: Optional[pulumi.Input[int]] = None, predefined_target: Optional[pulumi.Input[str]] = None, security_group_id: Optional[pulumi.Input[str]] = None, to_port: Optional[pulumi.Input[int]] = None, v4_cidr_blocks: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, v6_cidr_blocks: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None): """ :param pulumi.Input[str] description: Description of the security group. :param pulumi.Input[str] id: Id of the rule. :param pulumi.Input[Mapping[str, pulumi.Input[str]]] labels: Labels to assign to this security group. """ pulumi.set(__self__, "protocol", protocol) if description is not None: pulumi.set(__self__, "description", description) if from_port is not None: pulumi.set(__self__, "from_port", from_port) if id is not None: pulumi.set(__self__, "id", id) if labels is not None: pulumi.set(__self__, "labels", labels) if port is not None: pulumi.set(__self__, "port", port) if predefined_target is not None: pulumi.set(__self__, "predefined_target", predefined_target) if security_group_id is not None: pulumi.set(__self__, "security_group_id", security_group_id) if to_port is not None: pulumi.set(__self__, "to_port", to_port) if v4_cidr_blocks is not None: pulumi.set(__self__, "v4_cidr_blocks", v4_cidr_blocks) if v6_cidr_blocks is not None: pulumi.set(__self__, "v6_cidr_blocks", v6_cidr_blocks) @property @pulumi.getter def protocol(self) -> pulumi.Input[str]: return pulumi.get(self, "protocol") @protocol.setter def protocol(self, value: pulumi.Input[str]): pulumi.set(self, "protocol", value) @property @pulumi.getter def description(self) -> Optional[pulumi.Input[str]]: """ Description of the security group. """ return pulumi.get(self, "description") @description.setter def description(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "description", value) @property @pulumi.getter(name="fromPort") def from_port(self) -> Optional[pulumi.Input[int]]: return pulumi.get(self, "from_port") @from_port.setter def from_port(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "from_port", value) @property @pulumi.getter def id(self) -> Optional[pulumi.Input[str]]: """ Id of the rule. """ return pulumi.get(self, "id") @id.setter def id(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "id", value) @property @pulumi.getter def labels(self) -> Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]]: """ Labels to assign to this security group. """ return pulumi.get(self, "labels") @labels.setter def labels(self, value: Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]]): pulumi.set(self, "labels", value) @property @pulumi.getter def port(self) -> Optional[pulumi.Input[int]]: return pulumi.get(self, "port") @port.setter def port(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "port", value) @property @pulumi.getter(name="predefinedTarget") def predefined_target(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "predefined_target") @predefined_target.setter def predefined_target(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "predefined_target", value) @property @pulumi.getter(name="securityGroupId") def security_group_id(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "security_group_id") @security_group_id.setter def security_group_id(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "security_group_id", value) @property @pulumi.getter(name="toPort") def to_port(self) -> Optional[pulumi.Input[int]]: return pulumi.get(self, "to_port") @to_port.setter def to_port(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "to_port", value) @property @pulumi.getter(name="v4CidrBlocks") def v4_cidr_blocks(self) -> Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]: return pulumi.get(self, "v4_cidr_blocks") @v4_cidr_blocks.setter def v4_cidr_blocks(self, value: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]): pulumi.set(self, "v4_cidr_blocks", value) @property @pulumi.getter(name="v6CidrBlocks") def v6_cidr_blocks(self) -> Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]: return pulumi.get(self, "v6_cidr_blocks") @v6_cidr_blocks.setter def v6_cidr_blocks(self, value: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]): pulumi.set(self, "v6_cidr_blocks", value) @pulumi.input_type class VpcSecurityGroupIngressArgs: def __init__(__self__, *, protocol: pulumi.Input[str], description: Optional[pulumi.Input[str]] = None, from_port: Optional[pulumi.Input[int]] = None, id: Optional[pulumi.Input[str]] = None, labels: Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]] = None, port: Optional[pulumi.Input[int]] = None, predefined_target: Optional[pulumi.Input[str]] = None, security_group_id: Optional[pulumi.Input[str]] = None, to_port: Optional[pulumi.Input[int]] = None, v4_cidr_blocks: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, v6_cidr_blocks: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None): """ :param pulumi.Input[str] description: Description of the security group. :param pulumi.Input[str] id: Id of the rule. :param pulumi.Input[Mapping[str, pulumi.Input[str]]] labels: Labels to assign to this security group. """ pulumi.set(__self__, "protocol", protocol) if description is not None: pulumi.set(__self__, "description", description) if from_port is not None: pulumi.set(__self__, "from_port", from_port) if id is not None: pulumi.set(__self__, "id", id) if labels is not None: pulumi.set(__self__, "labels", labels) if port is not None: pulumi.set(__self__, "port", port) if predefined_target is not None: pulumi.set(__self__, "predefined_target", predefined_target) if security_group_id is not None: pulumi.set(__self__, "security_group_id", security_group_id) if to_port is not None: pulumi.set(__self__, "to_port", to_port) if v4_cidr_blocks is not None: pulumi.set(__self__, "v4_cidr_blocks", v4_cidr_blocks) if v6_cidr_blocks is not None: pulumi.set(__self__, "v6_cidr_blocks", v6_cidr_blocks) @property @pulumi.getter def protocol(self) -> pulumi.Input[str]: return pulumi.get(self, "protocol") @protocol.setter def protocol(self, value: pulumi.Input[str]): pulumi.set(self, "protocol", value) @property @pulumi.getter def description(self) -> Optional[pulumi.Input[str]]: """ Description of the security group. """ return pulumi.get(self, "description") @description.setter def description(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "description", value) @property @pulumi.getter(name="fromPort") def from_port(self) -> Optional[pulumi.Input[int]]: return pulumi.get(self, "from_port") @from_port.setter def from_port(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "from_port", value) @property @pulumi.getter def id(self) -> Optional[pulumi.Input[str]]: """ Id of the rule. """ return pulumi.get(self, "id") @id.setter def id(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "id", value) @property @pulumi.getter def labels(self) -> Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]]: """ Labels to assign to this security group. """ return pulumi.get(self, "labels") @labels.setter def labels(self, value: Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]]): pulumi.set(self, "labels", value) @property @pulumi.getter def port(self) -> Optional[pulumi.Input[int]]: return pulumi.get(self, "port") @port.setter def port(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "port", value) @property @pulumi.getter(name="predefinedTarget") def predefined_target(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "predefined_target") @predefined_target.setter def predefined_target(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "predefined_target", value) @property @pulumi.getter(name="securityGroupId") def security_group_id(self) -> Optional[pulumi.Input[str]]: return pulumi.get(self, "security_group_id") @security_group_id.setter def security_group_id(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "security_group_id", value) @property @pulumi.getter(name="toPort") def to_port(self) -> Optional[pulumi.Input[int]]: return pulumi.get(self, "to_port") @to_port.setter def to_port(self, value: Optional[pulumi.Input[int]]): pulumi.set(self, "to_port", value) @property @pulumi.getter(name="v4CidrBlocks") def v4_cidr_blocks(self) -> Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]: return pulumi.get(self, "v4_cidr_blocks") @v4_cidr_blocks.setter def v4_cidr_blocks(self, value: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]): pulumi.set(self, "v4_cidr_blocks", value) @property @pulumi.getter(name="v6CidrBlocks") def v6_cidr_blocks(self) -> Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]: return pulumi.get(self, "v6_cidr_blocks") @v6_cidr_blocks.setter def v6_cidr_blocks(self, value: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]): pulumi.set(self, "v6_cidr_blocks", value) @pulumi.input_type class VpcSubnetDhcpOptionsArgs: def __init__(__self__, *, domain_name: Optional[pulumi.Input[str]] = None, domain_name_servers: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, ntp_servers: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None): """ :param pulumi.Input[str] domain_name: Domain name. :param pulumi.Input[Sequence[pulumi.Input[str]]] domain_name_servers: Domain name server IP addresses. :param pulumi.Input[Sequence[pulumi.Input[str]]] ntp_servers: NTP server IP addresses. """ if domain_name is not None: pulumi.set(__self__, "domain_name", domain_name) if domain_name_servers is not None: pulumi.set(__self__, "domain_name_servers", domain_name_servers) if ntp_servers is not None: pulumi.set(__self__, "ntp_servers", ntp_servers) @property @pulumi.getter(name="domainName") def domain_name(self) -> Optional[pulumi.Input[str]]: """ Domain name. """ return pulumi.get(self, "domain_name") @domain_name.setter def domain_name(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "domain_name", value) @property @pulumi.getter(name="domainNameServers") def domain_name_servers(self) -> Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]: """ Domain name server IP addresses. """ return pulumi.get(self, "domain_name_servers") @domain_name_servers.setter def domain_name_servers(self, value: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]): pulumi.set(self, "domain_name_servers", value) @property @pulumi.getter(name="ntpServers") def ntp_servers(self) -> Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]: """ NTP server IP addresses. """ return pulumi.get(self, "ntp_servers") @ntp_servers.setter def ntp_servers(self, value: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]): pulumi.set(self, "ntp_servers", value) @pulumi.input_type class YdbDatabaseDedicatedLocationArgs: def __init__(__self__, *, region: Optional[pulumi.Input['YdbDatabaseDedicatedLocationRegionArgs']] = None): """ :param pulumi.Input['YdbDatabaseDedicatedLocationRegionArgs'] region: Region for the Yandex Database cluster. The structure is documented below. """ if region is not None: pulumi.set(__self__, "region", region) @property @pulumi.getter def region(self) -> Optional[pulumi.Input['YdbDatabaseDedicatedLocationRegionArgs']]: """ Region for the Yandex Database cluster. The structure is documented below. """ return pulumi.get(self, "region") @region.setter def region(self, value: Optional[pulumi.Input['YdbDatabaseDedicatedLocationRegionArgs']]): pulumi.set(self, "region", value) @pulumi.input_type class YdbDatabaseDedicatedLocationRegionArgs: def __init__(__self__, *, id: pulumi.Input[str]): """ :param pulumi.Input[str] id: Region ID for the Yandex Database cluster. """ pulumi.set(__self__, "id", id) @property @pulumi.getter def id(self) -> pulumi.Input[str]: """ Region ID for the Yandex Database cluster. """ return pulumi.get(self, "id") @id.setter def id(self, value: pulumi.Input[str]): pulumi.set(self, "id", value) @pulumi.input_type class YdbDatabaseDedicatedScalePolicyArgs: def __init__(__self__, *, fixed_scale: pulumi.Input['YdbDatabaseDedicatedScalePolicyFixedScaleArgs']): """ :param pulumi.Input['YdbDatabaseDedicatedScalePolicyFixedScaleArgs'] fixed_scale: Fixed scaling policy for the Yandex Database cluster. The structure is documented below. """ pulumi.set(__self__, "fixed_scale", fixed_scale) @property @pulumi.getter(name="fixedScale") def fixed_scale(self) -> pulumi.Input['YdbDatabaseDedicatedScalePolicyFixedScaleArgs']: """ Fixed scaling policy for the Yandex Database cluster. The structure is documented below. """ return pulumi.get(self, "fixed_scale") @fixed_scale.setter def fixed_scale(self, value: pulumi.Input['YdbDatabaseDedicatedScalePolicyFixedScaleArgs']): pulumi.set(self, "fixed_scale", value) @pulumi.input_type class YdbDatabaseDedicatedScalePolicyFixedScaleArgs: def __init__(__self__, *, size: pulumi.Input[int]): """ :param pulumi.Input[int] size: Number of instances for the Yandex Database cluster. """ pulumi.set(__self__, "size", size) @property @pulumi.getter def size(self) -> pulumi.Input[int]: """ Number of instances for the Yandex Database cluster. """ return pulumi.get(self, "size") @size.setter def size(self, value: pulumi.Input[int]): pulumi.set(self, "size", value) @pulumi.input_type class YdbDatabaseDedicatedStorageConfigArgs: def __init__(__self__, *, group_count: pulumi.Input[int], storage_type_id: pulumi.Input[str]): """ :param pulumi.Input[int] group_count: Amount of storage groups of selected type for the Yandex Database cluster. :param pulumi.Input[str] storage_type_id: Storage type ID for the Yandex Database cluster. Available presets can be obtained via `yc ydb storage-type list` command. """ pulumi.set(__self__, "group_count", group_count) pulumi.set(__self__, "storage_type_id", storage_type_id) @property @pulumi.getter(name="groupCount") def group_count(self) -> pulumi.Input[int]: """ Amount of storage groups of selected type for the Yandex Database cluster. """ return pulumi.get(self, "group_count") @group_count.setter def group_count(self, value: pulumi.Input[int]): pulumi.set(self, "group_count", value) @property @pulumi.getter(name="storageTypeId") def storage_type_id(self) -> pulumi.Input[str]: """ Storage type ID for the Yandex Database cluster. Available presets can be obtained via `yc ydb storage-type list` command. """ return pulumi.get(self, "storage_type_id") @storage_type_id.setter def storage_type_id(self, value: pulumi.Input[str]): pulumi.set(self, "storage_type_id", value) @pulumi.input_type class GetAlbBackendGroupGrpcBackendArgs: def __init__(__self__, *, healthcheck: 'GetAlbBackendGroupGrpcBackendHealthcheckArgs', load_balancing_config: 'GetAlbBackendGroupGrpcBackendLoadBalancingConfigArgs', name: str, port: int, target_group_ids: Sequence[str], tls: 'GetAlbBackendGroupGrpcBackendTlsArgs', weight: int): """ :param 'GetAlbBackendGroupGrpcBackendHealthcheckArgs' healthcheck: Healthcheck specification that will be used by this backend. Structure is documented below. :param 'GetAlbBackendGroupGrpcBackendLoadBalancingConfigArgs' load_balancing_config: Load Balancing Config specification that will be used by this backend. Structure is documented below. :param str name: - Name of the Backend Group. :param int port: Port for incoming traffic. :param Sequence[str] target_group_ids: References target groups for the backend. :param 'GetAlbBackendGroupGrpcBackendTlsArgs' tls: Tls specification that will be used by this backend. Structure is documented below. :param int weight: Weight of the backend. Traffic will be split between backends of the same BackendGroup according to their weights. """ pulumi.set(__self__, "healthcheck", healthcheck) pulumi.set(__self__, "load_balancing_config", load_balancing_config) pulumi.set(__self__, "name", name) pulumi.set(__self__, "port", port) pulumi.set(__self__, "target_group_ids", target_group_ids) pulumi.set(__self__, "tls", tls) pulumi.set(__self__, "weight", weight) @property @pulumi.getter def healthcheck(self) -> 'GetAlbBackendGroupGrpcBackendHealthcheckArgs': """ Healthcheck specification that will be used by this backend. Structure is documented below. """ return pulumi.get(self, "healthcheck") @healthcheck.setter def healthcheck(self, value: 'GetAlbBackendGroupGrpcBackendHealthcheckArgs'): pulumi.set(self, "healthcheck", value) @property @pulumi.getter(name="loadBalancingConfig") def load_balancing_config(self) -> 'GetAlbBackendGroupGrpcBackendLoadBalancingConfigArgs': """ Load Balancing Config specification that will be used by this backend. Structure is documented below. """ return pulumi.get(self, "load_balancing_config") @load_balancing_config.setter def load_balancing_config(self, value: 'GetAlbBackendGroupGrpcBackendLoadBalancingConfigArgs'): pulumi.set(self, "load_balancing_config", value) @property @pulumi.getter def name(self) -> str: """ - Name of the Backend Group. """ return pulumi.get(self, "name") @name.setter def name(self, value: str): pulumi.set(self, "name", value) @property @pulumi.getter def port(self) -> int: """ Port for incoming traffic. """ return pulumi.get(self, "port") @port.setter def port(self, value: int): pulumi.set(self, "port", value) @property @pulumi.getter(name="targetGroupIds") def target_group_ids(self) -> Sequence[str]: """ References target groups for the backend. """ return pulumi.get(self, "target_group_ids") @target_group_ids.setter def target_group_ids(self, value: Sequence[str]): pulumi.set(self, "target_group_ids", value) @property @pulumi.getter def tls(self) -> 'GetAlbBackendGroupGrpcBackendTlsArgs': """ Tls specification that will be used by this backend. Structure is documented below. """ return pulumi.get(self, "tls") @tls.setter def tls(self, value: 'GetAlbBackendGroupGrpcBackendTlsArgs'): pulumi.set(self, "tls", value) @property @pulumi.getter def weight(self) -> int: """ Weight of the backend. Traffic will be split between backends of the same BackendGroup according to their weights. """ return pulumi.get(self, "weight") @weight.setter def weight(self, value: int): pulumi.set(self, "weight", value) @pulumi.input_type class GetAlbBackendGroupGrpcBackendHealthcheckArgs: def __init__(__self__, *, grpc_healthcheck: 'GetAlbBackendGroupGrpcBackendHealthcheckGrpcHealthcheckArgs', healthcheck_port: int, healthy_threshold: int, http_healthcheck: 'GetAlbBackendGroupGrpcBackendHealthcheckHttpHealthcheckArgs', interval: str, interval_jitter_percent: float, stream_healthcheck: 'GetAlbBackendGroupGrpcBackendHealthcheckStreamHealthcheckArgs', timeout: str, unhealthy_threshold: int): """ :param 'GetAlbBackendGroupGrpcBackendHealthcheckGrpcHealthcheckArgs' grpc_healthcheck: Grpc Healthcheck specification that will be used by this healthcheck. Structure is documented below. :param int healthcheck_port: Optional alternative port for health checking. :param int healthy_threshold: Number of consecutive successful health checks required to promote endpoint into the healthy state. 0 means 1. Note that during startup, only a single successful health check is required to mark a host healthy. :param 'GetAlbBackendGroupGrpcBackendHealthcheckHttpHealthcheckArgs' http_healthcheck: Http Healthcheck specification that will be used by this healthcheck. Structure is documented below. :param str interval: Interval between health checks. :param float interval_jitter_percent: An optional jitter amount as a percentage of interval. If specified, during every interval value of (interval_ms * interval_jitter_percent / 100) will be added to the wait time. :param 'GetAlbBackendGroupGrpcBackendHealthcheckStreamHealthcheckArgs' stream_healthcheck: Stream Healthcheck specification that will be used by this healthcheck. Structure is documented below. :param str timeout: Time to wait for a health check response. :param int unhealthy_threshold: Number of consecutive failed health checks required to demote endpoint into the unhealthy state. 0 means 1. Note that for HTTP health checks, a single 503 immediately makes endpoint unhealthy. """ pulumi.set(__self__, "grpc_healthcheck", grpc_healthcheck) pulumi.set(__self__, "healthcheck_port", healthcheck_port) pulumi.set(__self__, "healthy_threshold", healthy_threshold) pulumi.set(__self__, "http_healthcheck", http_healthcheck) pulumi.set(__self__, "interval", interval) pulumi.set(__self__, "interval_jitter_percent", interval_jitter_percent) pulumi.set(__self__, "stream_healthcheck", stream_healthcheck) pulumi.set(__self__, "timeout", timeout) pulumi.set(__self__, "unhealthy_threshold", unhealthy_threshold) @property @pulumi.getter(name="grpcHealthcheck") def grpc_healthcheck(self) -> 'GetAlbBackendGroupGrpcBackendHealthcheckGrpcHealthcheckArgs': """ Grpc Healthcheck specification that will be used by this healthcheck. Structure is documented below. """ return pulumi.get(self, "grpc_healthcheck") @grpc_healthcheck.setter def grpc_healthcheck(self, value: 'GetAlbBackendGroupGrpcBackendHealthcheckGrpcHealthcheckArgs'): pulumi.set(self, "grpc_healthcheck", value) @property @pulumi.getter(name="healthcheckPort") def healthcheck_port(self) -> int: """ Optional alternative port for health checking. """ return pulumi.get(self, "healthcheck_port") @healthcheck_port.setter def healthcheck_port(self, value: int): pulumi.set(self, "healthcheck_port", value) @property @pulumi.getter(name="healthyThreshold") def healthy_threshold(self) -> int: """ Number of consecutive successful health checks required to promote endpoint into the healthy state. 0 means 1. Note that during startup, only a single successful health check is required to mark a host healthy. """ return pulumi.get(self, "healthy_threshold") @healthy_threshold.setter def healthy_threshold(self, value: int): pulumi.set(self, "healthy_threshold", value) @property @pulumi.getter(name="httpHealthcheck") def http_healthcheck(self) -> 'GetAlbBackendGroupGrpcBackendHealthcheckHttpHealthcheckArgs': """ Http Healthcheck specification that will be used by this healthcheck. Structure is documented below. """ return pulumi.get(self, "http_healthcheck") @http_healthcheck.setter def http_healthcheck(self, value: 'GetAlbBackendGroupGrpcBackendHealthcheckHttpHealthcheckArgs'): pulumi.set(self, "http_healthcheck", value) @property @pulumi.getter def interval(self) -> str: """ Interval between health checks. """ return pulumi.get(self, "interval") @interval.setter def interval(self, value: str): pulumi.set(self, "interval", value) @property @pulumi.getter(name="intervalJitterPercent") def interval_jitter_percent(self) -> float: """ An optional jitter amount as a percentage of interval. If specified, during every interval value of (interval_ms * interval_jitter_percent / 100) will be added to the wait time. """ return pulumi.get(self, "interval_jitter_percent") @interval_jitter_percent.setter def interval_jitter_percent(self, value: float): pulumi.set(self, "interval_jitter_percent", value) @property @pulumi.getter(name="streamHealthcheck") def stream_healthcheck(self) -> 'GetAlbBackendGroupGrpcBackendHealthcheckStreamHealthcheckArgs': """ Stream Healthcheck specification that will be used by this healthcheck. Structure is documented below. """ return pulumi.get(self, "stream_healthcheck") @stream_healthcheck.setter def stream_healthcheck(self, value: 'GetAlbBackendGroupGrpcBackendHealthcheckStreamHealthcheckArgs'): pulumi.set(self, "stream_healthcheck", value) @property @pulumi.getter def timeout(self) -> str: """ Time to wait for a health check response. """ return pulumi.get(self, "timeout") @timeout.setter def timeout(self, value: str): pulumi.set(self, "timeout", value) @property @pulumi.getter(name="unhealthyThreshold") def unhealthy_threshold(self) -> int: """ Number of consecutive failed health checks required to demote endpoint into the unhealthy state. 0 means 1. Note that for HTTP health checks, a single 503 immediately makes endpoint unhealthy. """ return pulumi.get(self, "unhealthy_threshold") @unhealthy_threshold.setter def unhealthy_threshold(self, value: int): pulumi.set(self, "unhealthy_threshold", value) @pulumi.input_type class GetAlbBackendGroupGrpcBackendHealthcheckGrpcHealthcheckArgs: def __init__(__self__, *, service_name: str): """ :param str service_name: Optional service name for grpc.health.v1.HealthCheckRequest message. """ pulumi.set(__self__, "service_name", service_name) @property @pulumi.getter(name="serviceName") def service_name(self) -> str: """ Optional service name for grpc.health.v1.HealthCheckRequest message. """ return pulumi.get(self, "service_name") @service_name.setter def service_name(self, value: str): pulumi.set(self, "service_name", value) @pulumi.input_type class GetAlbBackendGroupGrpcBackendHealthcheckHttpHealthcheckArgs: def __init__(__self__, *, host: str, http2: bool, path: str): """ :param str host: Optional "Host" HTTP header value. :param bool http2: If set, health checks will use HTTP2. :param str path: HTTP path. """ pulumi.set(__self__, "host", host) pulumi.set(__self__, "http2", http2) pulumi.set(__self__, "path", path) @property @pulumi.getter def host(self) -> str: """ Optional "Host" HTTP header value. """ return pulumi.get(self, "host") @host.setter def host(self, value: str): pulumi.set(self, "host", value) @property @pulumi.getter def http2(self) -> bool: """ If set, health checks will use HTTP2. """ return pulumi.get(self, "http2") @http2.setter def http2(self, value: bool): pulumi.set(self, "http2", value) @property @pulumi.getter def path(self) -> str: """ HTTP path. """ return pulumi.get(self, "path") @path.setter def path(self, value: str): pulumi.set(self, "path", value) @pulumi.input_type class GetAlbBackendGroupGrpcBackendHealthcheckStreamHealthcheckArgs: def __init__(__self__, *, receive: str, send: str): """ :param str receive: Optional text to search in reply. :param str send: Optional message to send. If empty, it's a connect-only health check. """ pulumi.set(__self__, "receive", receive) pulumi.set(__self__, "send", send) @property @pulumi.getter def receive(self) -> str: """ Optional text to search in reply. """ return pulumi.get(self, "receive") @receive.setter def receive(self, value: str): pulumi.set(self, "receive", value) @property @pulumi.getter def send(self) -> str: """ Optional message to send. If empty, it's a connect-only health check. """ return pulumi.get(self, "send") @send.setter def send(self, value: str): pulumi.set(self, "send", value) @pulumi.input_type class GetAlbBackendGroupGrpcBackendLoadBalancingConfigArgs: def __init__(__self__, *, locality_aware_routing_percent: int, panic_threshold: int, strict_locality: bool): """ :param int locality_aware_routing_percent: Percent of traffic to be sent to the same availability zone. The rest will be equally divided between other zones. :param int panic_threshold: If percentage of healthy hosts in the backend is lower than panic_threshold, traffic will be routed to all backends no matter what the health status is. This helps to avoid healthy backends overloading when everything is bad. Zero means no panic threshold. :param bool strict_locality: If set, will route requests only to the same availability zone. Balancer won't know about endpoints in other zones. """ pulumi.set(__self__, "locality_aware_routing_percent", locality_aware_routing_percent) pulumi.set(__self__, "panic_threshold", panic_threshold) pulumi.set(__self__, "strict_locality", strict_locality) @property @pulumi.getter(name="localityAwareRoutingPercent") def locality_aware_routing_percent(self) -> int: """ Percent of traffic to be sent to the same availability zone. The rest will be equally divided between other zones. """ return pulumi.get(self, "locality_aware_routing_percent") @locality_aware_routing_percent.setter def locality_aware_routing_percent(self, value: int): pulumi.set(self, "locality_aware_routing_percent", value) @property @pulumi.getter(name="panicThreshold") def panic_threshold(self) -> int: """ If percentage of healthy hosts in the backend is lower than panic_threshold, traffic will be routed to all backends no matter what the health status is. This helps to avoid healthy backends overloading when everything is bad. Zero means no panic threshold. """ return pulumi.get(self, "panic_threshold") @panic_threshold.setter def panic_threshold(self, value: int): pulumi.set(self, "panic_threshold", value) @property @pulumi.getter(name="strictLocality") def strict_locality(self) -> bool: """ If set, will route requests only to the same availability zone. Balancer won't know about endpoints in other zones. """ return pulumi.get(self, "strict_locality") @strict_locality.setter def strict_locality(self, value: bool): pulumi.set(self, "strict_locality", value) @pulumi.input_type class GetAlbBackendGroupGrpcBackendTlsArgs: def __init__(__self__, *, sni: str, validation_context: 'GetAlbBackendGroupGrpcBackendTlsValidationContextArgs'): """ :param str sni: [SNI](https://en.wikipedia.org/wiki/Server_Name_Indication) string for TLS connections. * `validation_context.0.trusted_ca_id` - Trusted CA certificate ID in the Certificate Manager. * `validation_context.0.trusted_ca_bytes` - PEM-encoded trusted CA certificate chain. """ pulumi.set(__self__, "sni", sni) pulumi.set(__self__, "validation_context", validation_context) @property @pulumi.getter def sni(self) -> str: """ [SNI](https://en.wikipedia.org/wiki/Server_Name_Indication) string for TLS connections. * `validation_context.0.trusted_ca_id` - Trusted CA certificate ID in the Certificate Manager. * `validation_context.0.trusted_ca_bytes` - PEM-encoded trusted CA certificate chain. """ return pulumi.get(self, "sni") @sni.setter def sni(self, value: str): pulumi.set(self, "sni", value) @property @pulumi.getter(name="validationContext") def validation_context(self) -> 'GetAlbBackendGroupGrpcBackendTlsValidationContextArgs': return pulumi.get(self, "validation_context") @validation_context.setter def validation_context(self, value: 'GetAlbBackendGroupGrpcBackendTlsValidationContextArgs'): pulumi.set(self, "validation_context", value) @pulumi.input_type class GetAlbBackendGroupGrpcBackendTlsValidationContextArgs: def __init__(__self__, *, trusted_ca_bytes: str, trusted_ca_id: str): pulumi.set(__self__, "trusted_ca_bytes", trusted_ca_bytes) pulumi.set(__self__, "trusted_ca_id", trusted_ca_id) @property @pulumi.getter(name="trustedCaBytes") def trusted_ca_bytes(self) -> str: return pulumi.get(self, "trusted_ca_bytes") @trusted_ca_bytes.setter def trusted_ca_bytes(self, value: str): pulumi.set(self, "trusted_ca_bytes", value) @property @pulumi.getter(name="trustedCaId") def trusted_ca_id(self) -> str: return pulumi.get(self, "trusted_ca_id") @trusted_ca_id.setter def trusted_ca_id(self, value: str): pulumi.set(self, "trusted_ca_id", value) @pulumi.input_type class GetAlbBackendGroupHttpBackendArgs: def __init__(__self__, *, healthcheck: 'GetAlbBackendGroupHttpBackendHealthcheckArgs', http2: bool, load_balancing_config: 'GetAlbBackendGroupHttpBackendLoadBalancingConfigArgs', name: str, port: int, target_group_ids: Sequence[str], tls: 'GetAlbBackendGroupHttpBackendTlsArgs', weight: int): """ :param 'GetAlbBackendGroupHttpBackendHealthcheckArgs' healthcheck: Healthcheck specification that will be used by this backend. Structure is documented below. :param bool http2: If set, health checks will use HTTP2. :param 'GetAlbBackendGroupHttpBackendLoadBalancingConfigArgs' load_balancing_config: Load Balancing Config specification that will be used by this backend. Structure is documented below. :param str name: - Name of the Backend Group. :param int port: Port for incoming traffic. :param Sequence[str] target_group_ids: References target groups for the backend. :param 'GetAlbBackendGroupHttpBackendTlsArgs' tls: Tls specification that will be used by this backend. Structure is documented below. :param int weight: Weight of the backend. Traffic will be split between backends of the same BackendGroup according to their weights. """ pulumi.set(__self__, "healthcheck", healthcheck) pulumi.set(__self__, "http2", http2) pulumi.set(__self__, "load_balancing_config", load_balancing_config) pulumi.set(__self__, "name", name) pulumi.set(__self__, "port", port) pulumi.set(__self__, "target_group_ids", target_group_ids) pulumi.set(__self__, "tls", tls) pulumi.set(__self__, "weight", weight) @property @pulumi.getter def healthcheck(self) -> 'GetAlbBackendGroupHttpBackendHealthcheckArgs': """ Healthcheck specification that will be used by this backend. Structure is documented below. """ return pulumi.get(self, "healthcheck") @healthcheck.setter def healthcheck(self, value: 'GetAlbBackendGroupHttpBackendHealthcheckArgs'): pulumi.set(self, "healthcheck", value) @property @pulumi.getter def http2(self) -> bool: """ If set, health checks will use HTTP2. """ return pulumi.get(self, "http2") @http2.setter def http2(self, value: bool): pulumi.set(self, "http2", value) @property @pulumi.getter(name="loadBalancingConfig") def load_balancing_config(self) -> 'GetAlbBackendGroupHttpBackendLoadBalancingConfigArgs': """ Load Balancing Config specification that will be used by this backend. Structure is documented below. """ return pulumi.get(self, "load_balancing_config") @load_balancing_config.setter def load_balancing_config(self, value: 'GetAlbBackendGroupHttpBackendLoadBalancingConfigArgs'): pulumi.set(self, "load_balancing_config", value) @property @pulumi.getter def name(self) -> str: """ - Name of the Backend Group. """ return pulumi.get(self, "name") @name.setter def name(self, value: str): pulumi.set(self, "name", value) @property @pulumi.getter def port(self) -> int: """ Port for incoming traffic. """ return pulumi.get(self, "port") @port.setter def port(self, value: int): pulumi.set(self, "port", value) @property @pulumi.getter(name="targetGroupIds") def target_group_ids(self) -> Sequence[str]: """ References target groups for the backend. """ return pulumi.get(self, "target_group_ids") @target_group_ids.setter def target_group_ids(self, value: Sequence[str]): pulumi.set(self, "target_group_ids", value) @property @pulumi.getter def tls(self) -> 'GetAlbBackendGroupHttpBackendTlsArgs': """ Tls specification that will be used by this backend. Structure is documented below. """ return pulumi.get(self, "tls") @tls.setter def tls(self, value: 'GetAlbBackendGroupHttpBackendTlsArgs'): pulumi.set(self, "tls", value) @property @pulumi.getter def weight(self) -> int: """ Weight of the backend. Traffic will be split between backends of the same BackendGroup according to their weights. """ return pulumi.get(self, "weight") @weight.setter def weight(self, value: int): pulumi.set(self, "weight", value) @pulumi.input_type class GetAlbBackendGroupHttpBackendHealthcheckArgs: def __init__(__self__, *, grpc_healthcheck: 'GetAlbBackendGroupHttpBackendHealthcheckGrpcHealthcheckArgs', healthcheck_port: int, healthy_threshold: int, http_healthcheck: 'GetAlbBackendGroupHttpBackendHealthcheckHttpHealthcheckArgs', interval: str, interval_jitter_percent: float, stream_healthcheck: 'GetAlbBackendGroupHttpBackendHealthcheckStreamHealthcheckArgs', timeout: str, unhealthy_threshold: int): """ :param 'GetAlbBackendGroupHttpBackendHealthcheckGrpcHealthcheckArgs' grpc_healthcheck: Grpc Healthcheck specification that will be used by this healthcheck. Structure is documented below. :param int healthcheck_port: Optional alternative port for health checking. :param int healthy_threshold: Number of consecutive successful health checks required to promote endpoint into the healthy state. 0 means 1. Note that during startup, only a single successful health check is required to mark a host healthy. :param 'GetAlbBackendGroupHttpBackendHealthcheckHttpHealthcheckArgs' http_healthcheck: Http Healthcheck specification that will be used by this healthcheck. Structure is documented below. :param str interval: Interval between health checks. :param float interval_jitter_percent: An optional jitter amount as a percentage of interval. If specified, during every interval value of (interval_ms * interval_jitter_percent / 100) will be added to the wait time. :param 'GetAlbBackendGroupHttpBackendHealthcheckStreamHealthcheckArgs' stream_healthcheck: Stream Healthcheck specification that will be used by this healthcheck. Structure is documented below. :param str timeout: Time to wait for a health check response. :param int unhealthy_threshold: Number of consecutive failed health checks required to demote endpoint into the unhealthy state. 0 means 1. Note that for HTTP health checks, a single 503 immediately makes endpoint unhealthy. """ pulumi.set(__self__, "grpc_healthcheck", grpc_healthcheck) pulumi.set(__self__, "healthcheck_port", healthcheck_port) pulumi.set(__self__, "healthy_threshold", healthy_threshold) pulumi.set(__self__, "http_healthcheck", http_healthcheck) pulumi.set(__self__, "interval", interval) pulumi.set(__self__, "interval_jitter_percent", interval_jitter_percent) pulumi.set(__self__, "stream_healthcheck", stream_healthcheck) pulumi.set(__self__, "timeout", timeout) pulumi.set(__self__, "unhealthy_threshold", unhealthy_threshold) @property @pulumi.getter(name="grpcHealthcheck") def grpc_healthcheck(self) -> 'GetAlbBackendGroupHttpBackendHealthcheckGrpcHealthcheckArgs': """ Grpc Healthcheck specification that will be used by this healthcheck. Structure is documented below. """ return pulumi.get(self, "grpc_healthcheck") @grpc_healthcheck.setter def grpc_healthcheck(self, value: 'GetAlbBackendGroupHttpBackendHealthcheckGrpcHealthcheckArgs'): pulumi.set(self, "grpc_healthcheck", value) @property @pulumi.getter(name="healthcheckPort") def healthcheck_port(self) -> int: """ Optional alternative port for health checking. """ return pulumi.get(self, "healthcheck_port") @healthcheck_port.setter def healthcheck_port(self, value: int): pulumi.set(self, "healthcheck_port", value) @property @pulumi.getter(name="healthyThreshold") def healthy_threshold(self) -> int: """ Number of consecutive successful health checks required to promote endpoint into the healthy state. 0 means 1. Note that during startup, only a single successful health check is required to mark a host healthy. """ return pulumi.get(self, "healthy_threshold") @healthy_threshold.setter def healthy_threshold(self, value: int): pulumi.set(self, "healthy_threshold", value) @property @pulumi.getter(name="httpHealthcheck") def http_healthcheck(self) -> 'GetAlbBackendGroupHttpBackendHealthcheckHttpHealthcheckArgs': """ Http Healthcheck specification that will be used by this healthcheck. Structure is documented below. """ return pulumi.get(self, "http_healthcheck") @http_healthcheck.setter def http_healthcheck(self, value: 'GetAlbBackendGroupHttpBackendHealthcheckHttpHealthcheckArgs'): pulumi.set(self, "http_healthcheck", value) @property @pulumi.getter def interval(self) -> str: """ Interval between health checks. """ return pulumi.get(self, "interval") @interval.setter def interval(self, value: str): pulumi.set(self, "interval", value) @property @pulumi.getter(name="intervalJitterPercent") def interval_jitter_percent(self) -> float: """ An optional jitter amount as a percentage of interval. If specified, during every interval value of (interval_ms * interval_jitter_percent / 100) will be added to the wait time. """ return pulumi.get(self, "interval_jitter_percent") @interval_jitter_percent.setter def interval_jitter_percent(self, value: float): pulumi.set(self, "interval_jitter_percent", value) @property @pulumi.getter(name="streamHealthcheck") def stream_healthcheck(self) -> 'GetAlbBackendGroupHttpBackendHealthcheckStreamHealthcheckArgs': """ Stream Healthcheck specification that will be used by this healthcheck. Structure is documented below. """ return pulumi.get(self, "stream_healthcheck") @stream_healthcheck.setter def stream_healthcheck(self, value: 'GetAlbBackendGroupHttpBackendHealthcheckStreamHealthcheckArgs'): pulumi.set(self, "stream_healthcheck", value) @property @pulumi.getter def timeout(self) -> str: """ Time to wait for a health check response. """ return pulumi.get(self, "timeout") @timeout.setter def timeout(self, value: str): pulumi.set(self, "timeout", value) @property @pulumi.getter(name="unhealthyThreshold") def unhealthy_threshold(self) -> int: """ Number of consecutive failed health checks required to demote endpoint into the unhealthy state. 0 means 1. Note that for HTTP health checks, a single 503 immediately makes endpoint unhealthy. """ return pulumi.get(self, "unhealthy_threshold") @unhealthy_threshold.setter def unhealthy_threshold(self, value: int): pulumi.set(self, "unhealthy_threshold", value) @pulumi.input_type class GetAlbBackendGroupHttpBackendHealthcheckGrpcHealthcheckArgs: def __init__(__self__, *, service_name: str): """ :param str service_name: Optional service name for grpc.health.v1.HealthCheckRequest message. """ pulumi.set(__self__, "service_name", service_name) @property @pulumi.getter(name="serviceName") def service_name(self) -> str: """ Optional service name for grpc.health.v1.HealthCheckRequest message. """ return pulumi.get(self, "service_name") @service_name.setter def service_name(self, value: str): pulumi.set(self, "service_name", value) @pulumi.input_type class GetAlbBackendGroupHttpBackendHealthcheckHttpHealthcheckArgs: def __init__(__self__, *, host: str, http2: bool, path: str): """ :param str host: Optional "Host" HTTP header value. :param bool http2: If set, health checks will use HTTP2. :param str path: HTTP path. """ pulumi.set(__self__, "host", host) pulumi.set(__self__, "http2", http2) pulumi.set(__self__, "path", path) @property @pulumi.getter def host(self) -> str: """ Optional "Host" HTTP header value. """ return pulumi.get(self, "host") @host.setter def host(self, value: str): pulumi.set(self, "host", value) @property @pulumi.getter def http2(self) -> bool: """ If set, health checks will use HTTP2. """ return pulumi.get(self, "http2") @http2.setter def http2(self, value: bool): pulumi.set(self, "http2", value) @property @pulumi.getter def path(self) -> str: """ HTTP path. """ return pulumi.get(self, "path") @path.setter def path(self, value: str): pulumi.set(self, "path", value) @pulumi.input_type class GetAlbBackendGroupHttpBackendHealthcheckStreamHealthcheckArgs: def __init__(__self__, *, receive: str, send: str): """ :param str receive: Optional text to search in reply. :param str send: Optional message to send. If empty, it's a connect-only health check. """ pulumi.set(__self__, "receive", receive) pulumi.set(__self__, "send", send) @property @pulumi.getter def receive(self) -> str: """ Optional text to search in reply. """ return pulumi.get(self, "receive") @receive.setter def receive(self, value: str): pulumi.set(self, "receive", value) @property @pulumi.getter def send(self) -> str: """ Optional message to send. If empty, it's a connect-only health check. """ return pulumi.get(self, "send") @send.setter def send(self, value: str): pulumi.set(self, "send", value) @pulumi.input_type class GetAlbBackendGroupHttpBackendLoadBalancingConfigArgs: def __init__(__self__, *, locality_aware_routing_percent: int, panic_threshold: int, strict_locality: bool): """ :param int locality_aware_routing_percent: Percent of traffic to be sent to the same availability zone. The rest will be equally divided between other zones. :param int panic_threshold: If percentage of healthy hosts in the backend is lower than panic_threshold, traffic will be routed to all backends no matter what the health status is. This helps to avoid healthy backends overloading when everything is bad. Zero means no panic threshold. :param bool strict_locality: If set, will route requests only to the same availability zone. Balancer won't know about endpoints in other zones. """ pulumi.set(__self__, "locality_aware_routing_percent", locality_aware_routing_percent) pulumi.set(__self__, "panic_threshold", panic_threshold) pulumi.set(__self__, "strict_locality", strict_locality) @property @pulumi.getter(name="localityAwareRoutingPercent") def locality_aware_routing_percent(self) -> int: """ Percent of traffic to be sent to the same availability zone. The rest will be equally divided between other zones. """ return pulumi.get(self, "locality_aware_routing_percent") @locality_aware_routing_percent.setter def locality_aware_routing_percent(self, value: int): pulumi.set(self, "locality_aware_routing_percent", value) @property @pulumi.getter(name="panicThreshold") def panic_threshold(self) -> int: """ If percentage of healthy hosts in the backend is lower than panic_threshold, traffic will be routed to all backends no matter what the health status is. This helps to avoid healthy backends overloading when everything is bad. Zero means no panic threshold. """ return pulumi.get(self, "panic_threshold") @panic_threshold.setter def panic_threshold(self, value: int): pulumi.set(self, "panic_threshold", value) @property @pulumi.getter(name="strictLocality") def strict_locality(self) -> bool: """ If set, will route requests only to the same availability zone. Balancer won't know about endpoints in other zones. """ return pulumi.get(self, "strict_locality") @strict_locality.setter def strict_locality(self, value: bool): pulumi.set(self, "strict_locality", value) @pulumi.input_type class GetAlbBackendGroupHttpBackendTlsArgs: def __init__(__self__, *, sni: str, validation_context: 'GetAlbBackendGroupHttpBackendTlsValidationContextArgs'): """ :param str sni: [SNI](https://en.wikipedia.org/wiki/Server_Name_Indication) string for TLS connections. * `validation_context.0.trusted_ca_id` - Trusted CA certificate ID in the Certificate Manager. * `validation_context.0.trusted_ca_bytes` - PEM-encoded trusted CA certificate chain. """ pulumi.set(__self__, "sni", sni) pulumi.set(__self__, "validation_context", validation_context) @property @pulumi.getter def sni(self) -> str: """ [SNI](https://en.wikipedia.org/wiki/Server_Name_Indication) string for TLS connections. * `validation_context.0.trusted_ca_id` - Trusted CA certificate ID in the Certificate Manager. * `validation_context.0.trusted_ca_bytes` - PEM-encoded trusted CA certificate chain. """ return pulumi.get(self, "sni") @sni.setter def sni(self, value: str): pulumi.set(self, "sni", value) @property @pulumi.getter(name="validationContext") def validation_context(self) -> 'GetAlbBackendGroupHttpBackendTlsValidationContextArgs': return pulumi.get(self, "validation_context") @validation_context.setter def validation_context(self, value: 'GetAlbBackendGroupHttpBackendTlsValidationContextArgs'): pulumi.set(self, "validation_context", value) @pulumi.input_type class GetAlbBackendGroupHttpBackendTlsValidationContextArgs: def __init__(__self__, *, trusted_ca_bytes: str, trusted_ca_id: str): pulumi.set(__self__, "trusted_ca_bytes", trusted_ca_bytes) pulumi.set(__self__, "trusted_ca_id", trusted_ca_id) @property @pulumi.getter(name="trustedCaBytes") def trusted_ca_bytes(self) -> str: return pulumi.get(self, "trusted_ca_bytes") @trusted_ca_bytes.setter def trusted_ca_bytes(self, value: str): pulumi.set(self, "trusted_ca_bytes", value) @property @pulumi.getter(name="trustedCaId") def trusted_ca_id(self) -> str: return pulumi.get(self, "trusted_ca_id") @trusted_ca_id.setter def trusted_ca_id(self, value: str): pulumi.set(self, "trusted_ca_id", value) @pulumi.input_type class GetComputeDiskDiskPlacementPolicyArgs: def __init__(__self__, *, disk_placement_group_id: str): pulumi.set(__self__, "disk_placement_group_id", disk_placement_group_id) @property @pulumi.getter(name="diskPlacementGroupId") def disk_placement_group_id(self) -> str: return pulumi.get(self, "disk_placement_group_id") @disk_placement_group_id.setter def disk_placement_group_id(self, value: str): pulumi.set(self, "disk_placement_group_id", value) @pulumi.input_type class GetComputeInstancePlacementPolicyArgs: def __init__(__self__, *, placement_group_id: str): pulumi.set(__self__, "placement_group_id", placement_group_id) @property @pulumi.getter(name="placementGroupId") def placement_group_id(self) -> str: return pulumi.get(self, "placement_group_id") @placement_group_id.setter def placement_group_id(self, value: str): pulumi.set(self, "placement_group_id", value) @pulumi.input_type class GetComputeInstanceSchedulingPolicyArgs: def __init__(__self__, *, preemptible: Optional[bool] = None): """ :param bool preemptible: (Optional) Specifies if the instance is preemptible. Defaults to false. """ if preemptible is not None: pulumi.set(__self__, "preemptible", preemptible) @property @pulumi.getter def preemptible(self) -> Optional[bool]: """ (Optional) Specifies if the instance is preemptible. Defaults to false. """ return pulumi.get(self, "preemptible") @preemptible.setter def preemptible(self, value: Optional[bool]): pulumi.set(self, "preemptible", value) @pulumi.input_type class GetFunctionScalingPolicyPolicyArgs: def __init__(__self__, *, tag: str, zone_instances_limit: Optional[int] = None, zone_requests_limit: Optional[int] = None): pulumi.set(__self__, "tag", tag) if zone_instances_limit is not None: pulumi.set(__self__, "zone_instances_limit", zone_instances_limit) if zone_requests_limit is not None: pulumi.set(__self__, "zone_requests_limit", zone_requests_limit) @property @pulumi.getter def tag(self) -> str: return pulumi.get(self, "tag") @tag.setter def tag(self, value: str): pulumi.set(self, "tag", value) @property @pulumi.getter(name="zoneInstancesLimit") def zone_instances_limit(self) -> Optional[int]: return pulumi.get(self, "zone_instances_limit") @zone_instances_limit.setter def zone_instances_limit(self, value: Optional[int]): pulumi.set(self, "zone_instances_limit", value) @property @pulumi.getter(name="zoneRequestsLimit") def zone_requests_limit(self) -> Optional[int]: return pulumi.get(self, "zone_requests_limit") @zone_requests_limit.setter def zone_requests_limit(self, value: Optional[int]): pulumi.set(self, "zone_requests_limit", value) @pulumi.input_type class GetIamPolicyBindingArgs: def __init__(__self__, *, members: Sequence[str], role: str): """ :param Sequence[str] members: An array of identities that will be granted the privilege in the `role`. Each entry can have one of the following values: * **userAccount:{user_id}**: A unique user ID that represents a specific Yandex account. * **serviceAccount:{service_account_id}**: A unique service account ID. :param str role: The role/permission that will be granted to the members. See the [IAM Roles] documentation for a complete list of roles. """ pulumi.set(__self__, "members", members) pulumi.set(__self__, "role", role) @property @pulumi.getter def members(self) -> Sequence[str]: """ An array of identities that will be granted the privilege in the `role`. Each entry can have one of the following values: * **userAccount:{user_id}**: A unique user ID that represents a specific Yandex account. * **serviceAccount:{service_account_id}**: A unique service account ID. """ return pulumi.get(self, "members") @members.setter def members(self, value: Sequence[str]): pulumi.set(self, "members", value) @property @pulumi.getter def role(self) -> str: """ The role/permission that will be granted to the members. See the [IAM Roles] documentation for a complete list of roles. """ return pulumi.get(self, "role") @role.setter def role(self, value: str): pulumi.set(self, "role", value) @pulumi.input_type class GetMdbClickhouseClusterCloudStorageArgs: def __init__(__self__, *, enabled: bool): """ :param bool enabled: (Required) Whether to use Yandex Object Storage for storing ClickHouse data. Can be either `true` or `false`. """ pulumi.set(__self__, "enabled", enabled) @property @pulumi.getter def enabled(self) -> bool: """ (Required) Whether to use Yandex Object Storage for storing ClickHouse data. Can be either `true` or `false`. """ return pulumi.get(self, "enabled") @enabled.setter def enabled(self, value: bool): pulumi.set(self, "enabled", value) @pulumi.input_type class GetMdbKafkaClusterConfigArgs: def __init__(__self__, *, kafka: 'GetMdbKafkaClusterConfigKafkaArgs', version: str, zones: Sequence[str], zookeeper: 'GetMdbKafkaClusterConfigZookeeperArgs', assign_public_ip: Optional[bool] = None, brokers_count: Optional[int] = None, schema_registry: Optional[bool] = None, unmanaged_topics: Optional[bool] = None): """ :param 'GetMdbKafkaClusterConfigKafkaArgs' kafka: (Optional) Configuration of the Kafka subcluster. The structure is documented below. :param str version: (Required) Version of the Kafka server software. :param Sequence[str] zones: (Optional) List of availability zones. :param 'GetMdbKafkaClusterConfigZookeeperArgs' zookeeper: (Optional) Configuration of the ZooKeeper subcluster. The structure is documented below. :param bool assign_public_ip: The flag that defines whether a public IP address is assigned to the node. :param int brokers_count: (Optional) Count of brokers per availability zone. :param bool schema_registry: (Optional) Enables managed schema registry on cluster. Can be either `true` or `false`. :param bool unmanaged_topics: (Optional) Allows to use Kafka AdminAPI to manage topics. Can be either `true` or `false`. """ pulumi.set(__self__, "kafka", kafka) pulumi.set(__self__, "version", version) pulumi.set(__self__, "zones", zones) pulumi.set(__self__, "zookeeper", zookeeper) if assign_public_ip is not None: pulumi.set(__self__, "assign_public_ip", assign_public_ip) if brokers_count is not None: pulumi.set(__self__, "brokers_count", brokers_count) if schema_registry is not None: pulumi.set(__self__, "schema_registry", schema_registry) if unmanaged_topics is not None: pulumi.set(__self__, "unmanaged_topics", unmanaged_topics) @property @pulumi.getter def kafka(self) -> 'GetMdbKafkaClusterConfigKafkaArgs': """ (Optional) Configuration of the Kafka subcluster. The structure is documented below. """ return pulumi.get(self, "kafka") @kafka.setter def kafka(self, value: 'GetMdbKafkaClusterConfigKafkaArgs'): pulumi.set(self, "kafka", value) @property @pulumi.getter def version(self) -> str: """ (Required) Version of the Kafka server software. """ return pulumi.get(self, "version") @version.setter def version(self, value: str): pulumi.set(self, "version", value) @property @pulumi.getter def zones(self) -> Sequence[str]: """ (Optional) List of availability zones. """ return pulumi.get(self, "zones") @zones.setter def zones(self, value: Sequence[str]): pulumi.set(self, "zones", value) @property @pulumi.getter def zookeeper(self) -> 'GetMdbKafkaClusterConfigZookeeperArgs': """ (Optional) Configuration of the ZooKeeper subcluster. The structure is documented below. """ return pulumi.get(self, "zookeeper") @zookeeper.setter def zookeeper(self, value: 'GetMdbKafkaClusterConfigZookeeperArgs'): pulumi.set(self, "zookeeper", value) @property @pulumi.getter(name="assignPublicIp") def assign_public_ip(self) -> Optional[bool]: """ The flag that defines whether a public IP address is assigned to the node. """ return pulumi.get(self, "assign_public_ip") @assign_public_ip.setter def assign_public_ip(self, value: Optional[bool]): pulumi.set(self, "assign_public_ip", value) @property @pulumi.getter(name="brokersCount") def brokers_count(self) -> Optional[int]: """ (Optional) Count of brokers per availability zone. """ return pulumi.get(self, "brokers_count") @brokers_count.setter def brokers_count(self, value: Optional[int]): pulumi.set(self, "brokers_count", value) @property @pulumi.getter(name="schemaRegistry") def schema_registry(self) -> Optional[bool]: """ (Optional) Enables managed schema registry on cluster. Can be either `true` or `false`. """ return pulumi.get(self, "schema_registry") @schema_registry.setter def schema_registry(self, value: Optional[bool]): pulumi.set(self, "schema_registry", value) @property @pulumi.getter(name="unmanagedTopics") def unmanaged_topics(self) -> Optional[bool]: """ (Optional) Allows to use Kafka AdminAPI to manage topics. Can be either `true` or `false`. """ return pulumi.get(self, "unmanaged_topics") @unmanaged_topics.setter def unmanaged_topics(self, value: Optional[bool]): pulumi.set(self, "unmanaged_topics", value) @pulumi.input_type class GetMdbKafkaClusterConfigKafkaArgs: def __init__(__self__, *, resources: 'GetMdbKafkaClusterConfigKafkaResourcesArgs', kafka_config: Optional['GetMdbKafkaClusterConfigKafkaKafkaConfigArgs'] = None): """ :param 'GetMdbKafkaClusterConfigKafkaResourcesArgs' resources: (Optional) Resources allocated to hosts of the ZooKeeper subcluster. The structure is documented below. :param 'GetMdbKafkaClusterConfigKafkaKafkaConfigArgs' kafka_config: (Optional) User-defined settings for the Kafka cluster. The structure is documented below. """ pulumi.set(__self__, "resources", resources) if kafka_config is not None: pulumi.set(__self__, "kafka_config", kafka_config) @property @pulumi.getter def resources(self) -> 'GetMdbKafkaClusterConfigKafkaResourcesArgs': """ (Optional) Resources allocated to hosts of the ZooKeeper subcluster. The structure is documented below. """ return pulumi.get(self, "resources") @resources.setter def resources(self, value: 'GetMdbKafkaClusterConfigKafkaResourcesArgs'): pulumi.set(self, "resources", value) @property @pulumi.getter(name="kafkaConfig") def kafka_config(self) -> Optional['GetMdbKafkaClusterConfigKafkaKafkaConfigArgs']: """ (Optional) User-defined settings for the Kafka cluster. The structure is documented below. """ return pulumi.get(self, "kafka_config") @kafka_config.setter def kafka_config(self, value: Optional['GetMdbKafkaClusterConfigKafkaKafkaConfigArgs']): pulumi.set(self, "kafka_config", value) @pulumi.input_type class GetMdbKafkaClusterConfigKafkaKafkaConfigArgs: def __init__(__self__, *, auto_create_topics_enable: Optional[bool] = None, compression_type: Optional[str] = None, default_replication_factor: Optional[str] = None, log_flush_interval_messages: Optional[str] = None, log_flush_interval_ms: Optional[str] = None, log_flush_scheduler_interval_ms: Optional[str] = None, log_preallocate: Optional[bool] = None, log_retention_bytes: Optional[str] = None, log_retention_hours: Optional[str] = None, log_retention_minutes: Optional[str] = None, log_retention_ms: Optional[str] = None, log_segment_bytes: Optional[str] = None, num_partitions: Optional[str] = None, socket_receive_buffer_bytes: Optional[str] = None, socket_send_buffer_bytes: Optional[str] = None): if auto_create_topics_enable is not None: pulumi.set(__self__, "auto_create_topics_enable", auto_create_topics_enable) if compression_type is not None: pulumi.set(__self__, "compression_type", compression_type) if default_replication_factor is not None: pulumi.set(__self__, "default_replication_factor", default_replication_factor) if log_flush_interval_messages is not None: pulumi.set(__self__, "log_flush_interval_messages", log_flush_interval_messages) if log_flush_interval_ms is not None: pulumi.set(__self__, "log_flush_interval_ms", log_flush_interval_ms) if log_flush_scheduler_interval_ms is not None: pulumi.set(__self__, "log_flush_scheduler_interval_ms", log_flush_scheduler_interval_ms) if log_preallocate is not None: pulumi.set(__self__, "log_preallocate", log_preallocate) if log_retention_bytes is not None: pulumi.set(__self__, "log_retention_bytes", log_retention_bytes) if log_retention_hours is not None: pulumi.set(__self__, "log_retention_hours", log_retention_hours) if log_retention_minutes is not None: pulumi.set(__self__, "log_retention_minutes", log_retention_minutes) if log_retention_ms is not None: pulumi.set(__self__, "log_retention_ms", log_retention_ms) if log_segment_bytes is not None: pulumi.set(__self__, "log_segment_bytes", log_segment_bytes) if num_partitions is not None: pulumi.set(__self__, "num_partitions", num_partitions) if socket_receive_buffer_bytes is not None: pulumi.set(__self__, "socket_receive_buffer_bytes", socket_receive_buffer_bytes) if socket_send_buffer_bytes is not None: pulumi.set(__self__, "socket_send_buffer_bytes", socket_send_buffer_bytes) @property @pulumi.getter(name="autoCreateTopicsEnable") def auto_create_topics_enable(self) -> Optional[bool]: return pulumi.get(self, "auto_create_topics_enable") @auto_create_topics_enable.setter def auto_create_topics_enable(self, value: Optional[bool]): pulumi.set(self, "auto_create_topics_enable", value) @property @pulumi.getter(name="compressionType") def compression_type(self) -> Optional[str]: return pulumi.get(self, "compression_type") @compression_type.setter def compression_type(self, value: Optional[str]): pulumi.set(self, "compression_type", value) @property @pulumi.getter(name="defaultReplicationFactor") def default_replication_factor(self) -> Optional[str]: return pulumi.get(self, "default_replication_factor") @default_replication_factor.setter def default_replication_factor(self, value: Optional[str]): pulumi.set(self, "default_replication_factor", value) @property @pulumi.getter(name="logFlushIntervalMessages") def log_flush_interval_messages(self) -> Optional[str]: return pulumi.get(self, "log_flush_interval_messages") @log_flush_interval_messages.setter def log_flush_interval_messages(self, value: Optional[str]): pulumi.set(self, "log_flush_interval_messages", value) @property @pulumi.getter(name="logFlushIntervalMs") def log_flush_interval_ms(self) -> Optional[str]: return pulumi.get(self, "log_flush_interval_ms") @log_flush_interval_ms.setter def log_flush_interval_ms(self, value: Optional[str]): pulumi.set(self, "log_flush_interval_ms", value) @property @pulumi.getter(name="logFlushSchedulerIntervalMs") def log_flush_scheduler_interval_ms(self) -> Optional[str]: return pulumi.get(self, "log_flush_scheduler_interval_ms") @log_flush_scheduler_interval_ms.setter def log_flush_scheduler_interval_ms(self, value: Optional[str]): pulumi.set(self, "log_flush_scheduler_interval_ms", value) @property @pulumi.getter(name="logPreallocate") def log_preallocate(self) -> Optional[bool]: return pulumi.get(self, "log_preallocate") @log_preallocate.setter def log_preallocate(self, value: Optional[bool]): pulumi.set(self, "log_preallocate", value) @property @pulumi.getter(name="logRetentionBytes") def log_retention_bytes(self) -> Optional[str]: return pulumi.get(self, "log_retention_bytes") @log_retention_bytes.setter def log_retention_bytes(self, value: Optional[str]): pulumi.set(self, "log_retention_bytes", value) @property @pulumi.getter(name="logRetentionHours") def log_retention_hours(self) -> Optional[str]: return pulumi.get(self, "log_retention_hours") @log_retention_hours.setter def log_retention_hours(self, value: Optional[str]): pulumi.set(self, "log_retention_hours", value) @property @pulumi.getter(name="logRetentionMinutes") def log_retention_minutes(self) -> Optional[str]: return pulumi.get(self, "log_retention_minutes") @log_retention_minutes.setter def log_retention_minutes(self, value: Optional[str]): pulumi.set(self, "log_retention_minutes", value) @property @pulumi.getter(name="logRetentionMs") def log_retention_ms(self) -> Optional[str]: return pulumi.get(self, "log_retention_ms") @log_retention_ms.setter def log_retention_ms(self, value: Optional[str]): pulumi.set(self, "log_retention_ms", value) @property @pulumi.getter(name="logSegmentBytes") def log_segment_bytes(self) -> Optional[str]: return pulumi.get(self, "log_segment_bytes") @log_segment_bytes.setter def log_segment_bytes(self, value: Optional[str]): pulumi.set(self, "log_segment_bytes", value) @property @pulumi.getter(name="numPartitions") def num_partitions(self) -> Optional[str]: return pulumi.get(self, "num_partitions") @num_partitions.setter def num_partitions(self, value: Optional[str]): pulumi.set(self, "num_partitions", value) @property @pulumi.getter(name="socketReceiveBufferBytes") def socket_receive_buffer_bytes(self) -> Optional[str]: return pulumi.get(self, "socket_receive_buffer_bytes") @socket_receive_buffer_bytes.setter def socket_receive_buffer_bytes(self, value: Optional[str]): pulumi.set(self, "socket_receive_buffer_bytes", value) @property @pulumi.getter(name="socketSendBufferBytes") def socket_send_buffer_bytes(self) -> Optional[str]: return pulumi.get(self, "socket_send_buffer_bytes") @socket_send_buffer_bytes.setter def socket_send_buffer_bytes(self, value: Optional[str]): pulumi.set(self, "socket_send_buffer_bytes", value) @pulumi.input_type class GetMdbKafkaClusterConfigKafkaResourcesArgs: def __init__(__self__, *, disk_size: int, disk_type_id: str, resource_preset_id: str): """ :param int disk_size: (Optional) Volume of the storage available to a ZooKeeper host, in gigabytes. :param str disk_type_id: (Optional) Type of the storage of ZooKeeper hosts. For more information see [the official documentation](https://cloud.yandex.com/docs/managed-kafka/concepts/storage). """ pulumi.set(__self__, "disk_size", disk_size) pulumi.set(__self__, "disk_type_id", disk_type_id) pulumi.set(__self__, "resource_preset_id", resource_preset_id) @property @pulumi.getter(name="diskSize") def disk_size(self) -> int: """ (Optional) Volume of the storage available to a ZooKeeper host, in gigabytes. """ return pulumi.get(self, "disk_size") @disk_size.setter def disk_size(self, value: int): pulumi.set(self, "disk_size", value) @property @pulumi.getter(name="diskTypeId") def disk_type_id(self) -> str: """ (Optional) Type of the storage of ZooKeeper hosts. For more information see [the official documentation](https://cloud.yandex.com/docs/managed-kafka/concepts/storage). """ return pulumi.get(self, "disk_type_id") @disk_type_id.setter def disk_type_id(self, value: str): pulumi.set(self, "disk_type_id", value) @property @pulumi.getter(name="resourcePresetId") def resource_preset_id(self) -> str: return pulumi.get(self, "resource_preset_id") @resource_preset_id.setter def resource_preset_id(self, value: str): pulumi.set(self, "resource_preset_id", value) @pulumi.input_type class GetMdbKafkaClusterConfigZookeeperArgs: def __init__(__self__, *, resources: 'GetMdbKafkaClusterConfigZookeeperResourcesArgs'): """ :param 'GetMdbKafkaClusterConfigZookeeperResourcesArgs' resources: (Optional) Resources allocated to hosts of the ZooKeeper subcluster. The structure is documented below. """ pulumi.set(__self__, "resources", resources) @property @pulumi.getter def resources(self) -> 'GetMdbKafkaClusterConfigZookeeperResourcesArgs': """ (Optional) Resources allocated to hosts of the ZooKeeper subcluster. The structure is documented below. """ return pulumi.get(self, "resources") @resources.setter def resources(self, value: 'GetMdbKafkaClusterConfigZookeeperResourcesArgs'): pulumi.set(self, "resources", value) @pulumi.input_type class GetMdbKafkaClusterConfigZookeeperResourcesArgs: def __init__(__self__, *, disk_size: int, disk_type_id: str, resource_preset_id: str): """ :param int disk_size: (Optional) Volume of the storage available to a ZooKeeper host, in gigabytes. :param str disk_type_id: (Optional) Type of the storage of ZooKeeper hosts. For more information see [the official documentation](https://cloud.yandex.com/docs/managed-kafka/concepts/storage). """ pulumi.set(__self__, "disk_size", disk_size) pulumi.set(__self__, "disk_type_id", disk_type_id) pulumi.set(__self__, "resource_preset_id", resource_preset_id) @property @pulumi.getter(name="diskSize") def disk_size(self) -> int: """ (Optional) Volume of the storage available to a ZooKeeper host, in gigabytes. """ return pulumi.get(self, "disk_size") @disk_size.setter def disk_size(self, value: int): pulumi.set(self, "disk_size", value) @property @pulumi.getter(name="diskTypeId") def disk_type_id(self) -> str: """ (Optional) Type of the storage of ZooKeeper hosts. For more information see [the official documentation](https://cloud.yandex.com/docs/managed-kafka/concepts/storage). """ return pulumi.get(self, "disk_type_id") @disk_type_id.setter def disk_type_id(self, value: str): pulumi.set(self, "disk_type_id", value) @property @pulumi.getter(name="resourcePresetId") def resource_preset_id(self) -> str: return pulumi.get(self, "resource_preset_id") @resource_preset_id.setter def resource_preset_id(self, value: str): pulumi.set(self, "resource_preset_id", value) @pulumi.input_type class GetMdbKafkaClusterTopicArgs: def __init__(__self__, *, cluster_id: str, name: str, partitions: int, replication_factor: int, topic_config: Optional['GetMdbKafkaClusterTopicTopicConfigArgs'] = None): """ :param str cluster_id: The ID of the Kafka cluster. :param str name: The name of the Kafka cluster. :param int partitions: (Required) The number of the topic's partitions. :param int replication_factor: (Required) Amount of data copies (replicas) for the topic in the cluster. :param 'GetMdbKafkaClusterTopicTopicConfigArgs' topic_config: (Required) User-defined settings for the topic. The structure is documented below. """ pulumi.set(__self__, "cluster_id", cluster_id) pulumi.set(__self__, "name", name) pulumi.set(__self__, "partitions", partitions) pulumi.set(__self__, "replication_factor", replication_factor) if topic_config is not None: pulumi.set(__self__, "topic_config", topic_config) @property @pulumi.getter(name="clusterId") def cluster_id(self) -> str: """ The ID of the Kafka cluster. """ return pulumi.get(self, "cluster_id") @cluster_id.setter def cluster_id(self, value: str): pulumi.set(self, "cluster_id", value) @property @pulumi.getter def name(self) -> str: """ The name of the Kafka cluster. """ return pulumi.get(self, "name") @name.setter def name(self, value: str): pulumi.set(self, "name", value) @property @pulumi.getter def partitions(self) -> int: """ (Required) The number of the topic's partitions. """ return pulumi.get(self, "partitions") @partitions.setter def partitions(self, value: int): pulumi.set(self, "partitions", value) @property @pulumi.getter(name="replicationFactor") def replication_factor(self) -> int: """ (Required) Amount of data copies (replicas) for the topic in the cluster. """ return pulumi.get(self, "replication_factor") @replication_factor.setter def replication_factor(self, value: int): pulumi.set(self, "replication_factor", value) @property @pulumi.getter(name="topicConfig") def topic_config(self) -> Optional['GetMdbKafkaClusterTopicTopicConfigArgs']: """ (Required) User-defined settings for the topic. The structure is documented below. """ return pulumi.get(self, "topic_config") @topic_config.setter def topic_config(self, value: Optional['GetMdbKafkaClusterTopicTopicConfigArgs']): pulumi.set(self, "topic_config", value) @pulumi.input_type class GetMdbKafkaClusterTopicTopicConfigArgs: def __init__(__self__, *, cleanup_policy: Optional[str] = None, compression_type: Optional[str] = None, delete_retention_ms: Optional[str] = None, file_delete_delay_ms: Optional[str] = None, flush_messages: Optional[str] = None, flush_ms: Optional[str] = None, max_message_bytes: Optional[str] = None, min_compaction_lag_ms: Optional[str] = None, min_insync_replicas: Optional[str] = None, preallocate: Optional[bool] = None, retention_bytes: Optional[str] = None, retention_ms: Optional[str] = None, segment_bytes: Optional[str] = None): if cleanup_policy is not None: pulumi.set(__self__, "cleanup_policy", cleanup_policy) if compression_type is not None: pulumi.set(__self__, "compression_type", compression_type) if delete_retention_ms is not None: pulumi.set(__self__, "delete_retention_ms", delete_retention_ms) if file_delete_delay_ms is not None: pulumi.set(__self__, "file_delete_delay_ms", file_delete_delay_ms) if flush_messages is not None: pulumi.set(__self__, "flush_messages", flush_messages) if flush_ms is not None: pulumi.set(__self__, "flush_ms", flush_ms) if max_message_bytes is not None: pulumi.set(__self__, "max_message_bytes", max_message_bytes) if min_compaction_lag_ms is not None: pulumi.set(__self__, "min_compaction_lag_ms", min_compaction_lag_ms) if min_insync_replicas is not None: pulumi.set(__self__, "min_insync_replicas", min_insync_replicas) if preallocate is not None: pulumi.set(__self__, "preallocate", preallocate) if retention_bytes is not None: pulumi.set(__self__, "retention_bytes", retention_bytes) if retention_ms is not None: pulumi.set(__self__, "retention_ms", retention_ms) if segment_bytes is not None: pulumi.set(__self__, "segment_bytes", segment_bytes) @property @pulumi.getter(name="cleanupPolicy") def cleanup_policy(self) -> Optional[str]: return pulumi.get(self, "cleanup_policy") @cleanup_policy.setter def cleanup_policy(self, value: Optional[str]): pulumi.set(self, "cleanup_policy", value) @property @pulumi.getter(name="compressionType") def compression_type(self) -> Optional[str]: return pulumi.get(self, "compression_type") @compression_type.setter def compression_type(self, value: Optional[str]): pulumi.set(self, "compression_type", value) @property @pulumi.getter(name="deleteRetentionMs") def delete_retention_ms(self) -> Optional[str]: return pulumi.get(self, "delete_retention_ms") @delete_retention_ms.setter def delete_retention_ms(self, value: Optional[str]): pulumi.set(self, "delete_retention_ms", value) @property @pulumi.getter(name="fileDeleteDelayMs") def file_delete_delay_ms(self) -> Optional[str]: return pulumi.get(self, "file_delete_delay_ms") @file_delete_delay_ms.setter def file_delete_delay_ms(self, value: Optional[str]): pulumi.set(self, "file_delete_delay_ms", value) @property @pulumi.getter(name="flushMessages") def flush_messages(self) -> Optional[str]: return pulumi.get(self, "flush_messages") @flush_messages.setter def flush_messages(self, value: Optional[str]): pulumi.set(self, "flush_messages", value) @property @pulumi.getter(name="flushMs") def flush_ms(self) -> Optional[str]: return pulumi.get(self, "flush_ms") @flush_ms.setter def flush_ms(self, value: Optional[str]): pulumi.set(self, "flush_ms", value) @property @pulumi.getter(name="maxMessageBytes") def max_message_bytes(self) -> Optional[str]: return pulumi.get(self, "max_message_bytes") @max_message_bytes.setter def max_message_bytes(self, value: Optional[str]): pulumi.set(self, "max_message_bytes", value) @property @pulumi.getter(name="minCompactionLagMs") def min_compaction_lag_ms(self) -> Optional[str]: return pulumi.get(self, "min_compaction_lag_ms") @min_compaction_lag_ms.setter def min_compaction_lag_ms(self, value: Optional[str]): pulumi.set(self, "min_compaction_lag_ms", value) @property @pulumi.getter(name="minInsyncReplicas") def min_insync_replicas(self) -> Optional[str]: return pulumi.get(self, "min_insync_replicas") @min_insync_replicas.setter def min_insync_replicas(self, value: Optional[str]): pulumi.set(self, "min_insync_replicas", value) @property @pulumi.getter def preallocate(self) -> Optional[bool]: return pulumi.get(self, "preallocate") @preallocate.setter def preallocate(self, value: Optional[bool]): pulumi.set(self, "preallocate", value) @property @pulumi.getter(name="retentionBytes") def retention_bytes(self) -> Optional[str]: return pulumi.get(self, "retention_bytes") @retention_bytes.setter def retention_bytes(self, value: Optional[str]): pulumi.set(self, "retention_bytes", value) @property @pulumi.getter(name="retentionMs") def retention_ms(self) -> Optional[str]: return pulumi.get(self, "retention_ms") @retention_ms.setter def retention_ms(self, value: Optional[str]): pulumi.set(self, "retention_ms", value) @property @pulumi.getter(name="segmentBytes") def segment_bytes(self) -> Optional[str]: return pulumi.get(self, "segment_bytes") @segment_bytes.setter def segment_bytes(self, value: Optional[str]): pulumi.set(self, "segment_bytes", value) @pulumi.input_type class GetMdbKafkaClusterUserArgs: def __init__(__self__, *, name: str, password: str, permissions: Optional[Sequence['GetMdbKafkaClusterUserPermissionArgs']] = None): """ :param str name: The name of the Kafka cluster. :param str password: (Required) The password of the user. :param Sequence['GetMdbKafkaClusterUserPermissionArgs'] permissions: (Optional) Set of permissions granted to the user. The structure is documented below. """ pulumi.set(__self__, "name", name) pulumi.set(__self__, "password", password) if permissions is not None: pulumi.set(__self__, "permissions", permissions) @property @pulumi.getter def name(self) -> str: """ The name of the Kafka cluster. """ return pulumi.get(self, "name") @name.setter def name(self, value: str): pulumi.set(self, "name", value) @property @pulumi.getter def password(self) -> str: """ (Required) The password of the user. """ return pulumi.get(self, "password") @password.setter def password(self, value: str): pulumi.set(self, "password", value) @property @pulumi.getter def permissions(self) -> Optional[Sequence['GetMdbKafkaClusterUserPermissionArgs']]: """ (Optional) Set of permissions granted to the user. The structure is documented below. """ return pulumi.get(self, "permissions") @permissions.setter def permissions(self, value: Optional[Sequence['GetMdbKafkaClusterUserPermissionArgs']]): pulumi.set(self, "permissions", value) @pulumi.input_type class GetMdbKafkaClusterUserPermissionArgs: def __init__(__self__, *, role: str, topic_name: str): """ :param str role: Role of the host in the cluster. :param str topic_name: (Required) The name of the topic that the permission grants access to. """ pulumi.set(__self__, "role", role) pulumi.set(__self__, "topic_name", topic_name) @property @pulumi.getter def role(self) -> str: """ Role of the host in the cluster. """ return pulumi.get(self, "role") @role.setter def role(self, value: str): pulumi.set(self, "role", value) @property @pulumi.getter(name="topicName") def topic_name(self) -> str: """ (Required) The name of the topic that the permission grants access to. """ return pulumi.get(self, "topic_name") @topic_name.setter def topic_name(self, value: str): pulumi.set(self, "topic_name", value) @pulumi.input_type class GetMdbMysqlClusterAccessArgs: def __init__(__self__, *, data_lens: bool, web_sql: bool): """ :param bool data_lens: Allow access for [Yandex DataLens](https://cloud.yandex.com/services/datalens). :param bool web_sql: Allows access for [SQL queries in the management console](https://cloud.yandex.com/docs/managed-mysql/operations/web-sql-query). """ pulumi.set(__self__, "data_lens", data_lens) pulumi.set(__self__, "web_sql", web_sql) @property @pulumi.getter(name="dataLens") def data_lens(self) -> bool: """ Allow access for [Yandex DataLens](https://cloud.yandex.com/services/datalens). """ return pulumi.get(self, "data_lens") @data_lens.setter def data_lens(self, value: bool): pulumi.set(self, "data_lens", value) @property @pulumi.getter(name="webSql") def web_sql(self) -> bool: """ Allows access for [SQL queries in the management console](https://cloud.yandex.com/docs/managed-mysql/operations/web-sql-query). """ return pulumi.get(self, "web_sql") @web_sql.setter def web_sql(self, value: bool): pulumi.set(self, "web_sql", value)
[ "pulumi.get", "pulumi.getter", "pulumi.set", "warnings.warn", "pulumi.log.warn" ]
[((16397, 16433), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""targetGroupIds"""'}), "(name='targetGroupIds')\n", (16410, 16433), False, 'import pulumi\n'), ((17299, 17340), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""loadBalancingConfig"""'}), "(name='loadBalancingConfig')\n", (17312, 17340), False, 'import pulumi\n'), ((23050, 23087), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""grpcHealthcheck"""'}), "(name='grpcHealthcheck')\n", (23063, 23087), False, 'import pulumi\n'), ((23616, 23653), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""healthcheckPort"""'}), "(name='healthcheckPort')\n", (23629, 23653), False, 'import pulumi\n'), ((24018, 24056), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""healthyThreshold"""'}), "(name='healthyThreshold')\n", (24031, 24056), False, 'import pulumi\n'), ((24590, 24627), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""httpHealthcheck"""'}), "(name='httpHealthcheck')\n", (24603, 24627), False, 'import pulumi\n'), ((25156, 25199), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""intervalJitterPercent"""'}), "(name='intervalJitterPercent')\n", (25169, 25199), False, 'import pulumi\n'), ((25734, 25773), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""streamHealthcheck"""'}), "(name='streamHealthcheck')\n", (25747, 25773), False, 'import pulumi\n'), ((26318, 26358), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""unhealthyThreshold"""'}), "(name='unhealthyThreshold')\n", (26331, 26358), False, 'import pulumi\n'), ((27298, 27331), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""serviceName"""'}), "(name='serviceName')\n", (27311, 27331), False, 'import pulumi\n'), ((31878, 31927), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""localityAwareRoutingPercent"""'}), "(name='localityAwareRoutingPercent')\n", (31891, 31927), False, 'import pulumi\n'), ((32430, 32466), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""panicThreshold"""'}), "(name='panicThreshold')\n", (32443, 32466), False, 'import pulumi\n'), ((33037, 33073), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""strictLocality"""'}), "(name='strictLocality')\n", (33050, 33073), False, 'import pulumi\n'), ((34924, 34963), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""validationContext"""'}), "(name='validationContext')\n", (34937, 34963), False, 'import pulumi\n'), ((35821, 35857), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""trustedCaBytes"""'}), "(name='trustedCaBytes')\n", (35834, 35857), False, 'import pulumi\n'), ((36143, 36176), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""trustedCaId"""'}), "(name='trustedCaId')\n", (36156, 36176), False, 'import pulumi\n'), ((39114, 39150), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""targetGroupIds"""'}), "(name='targetGroupIds')\n", (39127, 39150), False, 'import pulumi\n'), ((40332, 40373), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""loadBalancingConfig"""'}), "(name='loadBalancingConfig')\n", (40345, 40373), False, 'import pulumi\n'), ((46083, 46120), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""grpcHealthcheck"""'}), "(name='grpcHealthcheck')\n", (46096, 46120), False, 'import pulumi\n'), ((46649, 46686), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""healthcheckPort"""'}), "(name='healthcheckPort')\n", (46662, 46686), False, 'import pulumi\n'), ((47051, 47089), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""healthyThreshold"""'}), "(name='healthyThreshold')\n", (47064, 47089), False, 'import pulumi\n'), ((47623, 47660), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""httpHealthcheck"""'}), "(name='httpHealthcheck')\n", (47636, 47660), False, 'import pulumi\n'), ((48189, 48232), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""intervalJitterPercent"""'}), "(name='intervalJitterPercent')\n", (48202, 48232), False, 'import pulumi\n'), ((48767, 48806), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""streamHealthcheck"""'}), "(name='streamHealthcheck')\n", (48780, 48806), False, 'import pulumi\n'), ((49351, 49391), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""unhealthyThreshold"""'}), "(name='unhealthyThreshold')\n", (49364, 49391), False, 'import pulumi\n'), ((50331, 50364), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""serviceName"""'}), "(name='serviceName')\n", (50344, 50364), False, 'import pulumi\n'), ((54911, 54960), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""localityAwareRoutingPercent"""'}), "(name='localityAwareRoutingPercent')\n", (54924, 54960), False, 'import pulumi\n'), ((55463, 55499), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""panicThreshold"""'}), "(name='panicThreshold')\n", (55476, 55499), False, 'import pulumi\n'), ((56070, 56106), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""strictLocality"""'}), "(name='strictLocality')\n", (56083, 56106), False, 'import pulumi\n'), ((57957, 57996), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""validationContext"""'}), "(name='validationContext')\n", (57970, 57996), False, 'import pulumi\n'), ((58854, 58890), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""trustedCaBytes"""'}), "(name='trustedCaBytes')\n", (58867, 58890), False, 'import pulumi\n'), ((59176, 59209), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""trustedCaId"""'}), "(name='trustedCaId')\n", (59189, 59209), False, 'import pulumi\n'), ((61207, 61237), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""subnetId"""'}), "(name='subnetId')\n", (61220, 61237), False, 'import pulumi\n'), ((61550, 61578), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""zoneId"""'}), "(name='zoneId')\n", (61563, 61578), False, 'import pulumi\n'), ((61875, 61911), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""disableTraffic"""'}), "(name='disableTraffic')\n", (61888, 61911), False, 'import pulumi\n'), ((67963, 68004), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""externalIpv4Address"""'}), "(name='externalIpv4Address')\n", (67976, 68004), False, 'import pulumi\n'), ((68525, 68566), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""externalIpv6Address"""'}), "(name='externalIpv6Address')\n", (68538, 68566), False, 'import pulumi\n'), ((69087, 69128), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""internalIpv4Address"""'}), "(name='internalIpv4Address')\n", (69100, 69128), False, 'import pulumi\n'), ((72058, 72088), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""subnetId"""'}), "(name='subnetId')\n", (72071, 72088), False, 'import pulumi\n'), ((75174, 75207), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""allowHttp10"""'}), "(name='allowHttp10')\n", (75187, 75207), False, 'import pulumi\n'), ((75569, 75603), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""http2Options"""'}), "(name='http2Options')\n", (75582, 75603), False, 'import pulumi\n'), ((76091, 76125), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""httpRouterId"""'}), "(name='httpRouterId')\n", (76104, 76125), False, 'import pulumi\n'), ((76885, 76927), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxConcurrentStreams"""'}), "(name='maxConcurrentStreams')\n", (76898, 76927), False, 'import pulumi\n'), ((77585, 77618), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""httpToHttps"""'}), "(name='httpToHttps')\n", (77598, 77618), False, 'import pulumi\n'), ((78711, 78747), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""defaultHandler"""'}), "(name='defaultHandler')\n", (78724, 78747), False, 'import pulumi\n'), ((79183, 79216), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""sniHandlers"""'}), "(name='sniHandlers')\n", (79196, 79216), False, 'import pulumi\n'), ((80640, 80676), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""certificateIds"""'}), "(name='certificateIds')\n", (80653, 80676), False, 'import pulumi\n'), ((81231, 81264), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""httpHandler"""'}), "(name='httpHandler')\n", (81244, 81264), False, 'import pulumi\n'), ((82823, 82856), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""allowHttp10"""'}), "(name='allowHttp10')\n", (82836, 82856), False, 'import pulumi\n'), ((83218, 83252), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""http2Options"""'}), "(name='http2Options')\n", (83231, 83252), False, 'import pulumi\n'), ((83774, 83808), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""httpRouterId"""'}), "(name='httpRouterId')\n", (83787, 83808), False, 'import pulumi\n'), ((84585, 84627), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxConcurrentStreams"""'}), "(name='maxConcurrentStreams')\n", (84598, 84627), False, 'import pulumi\n'), ((86525, 86558), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""serverNames"""'}), "(name='serverNames')\n", (86538, 86558), False, 'import pulumi\n'), ((87861, 87897), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""certificateIds"""'}), "(name='certificateIds')\n", (87874, 87897), False, 'import pulumi\n'), ((88452, 88485), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""httpHandler"""'}), "(name='httpHandler')\n", (88465, 88485), False, 'import pulumi\n'), ((90059, 90092), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""allowHttp10"""'}), "(name='allowHttp10')\n", (90072, 90092), False, 'import pulumi\n'), ((90454, 90488), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""http2Options"""'}), "(name='http2Options')\n", (90467, 90488), False, 'import pulumi\n'), ((91016, 91050), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""httpRouterId"""'}), "(name='httpRouterId')\n", (91029, 91050), False, 'import pulumi\n'), ((91830, 91872), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxConcurrentStreams"""'}), "(name='maxConcurrentStreams')\n", (91843, 91872), False, 'import pulumi\n'), ((92851, 92882), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""ipAddress"""'}), "(name='ipAddress')\n", (92864, 92882), False, 'import pulumi\n'), ((93176, 93206), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""subnetId"""'}), "(name='subnetId')\n", (93189, 93206), False, 'import pulumi\n'), ((99666, 99697), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""grpcRoute"""'}), "(name='grpcRoute')\n", (99679, 99697), False, 'import pulumi\n'), ((100103, 100134), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""httpRoute"""'}), "(name='httpRoute')\n", (100116, 100134), False, 'import pulumi\n'), ((102219, 102252), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""grpcMatches"""'}), "(name='grpcMatches')\n", (102232, 102252), False, 'import pulumi\n'), ((102743, 102780), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""grpcRouteAction"""'}), "(name='grpcRouteAction')\n", (102756, 102780), False, 'import pulumi\n'), ((103258, 103304), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""grpcStatusResponseAction"""'}), "(name='grpcStatusResponseAction')\n", (103271, 103304), False, 'import pulumi\n'), ((107411, 107447), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""backendGroupId"""'}), "(name='backendGroupId')\n", (107424, 107447), False, 'import pulumi\n'), ((107778, 107815), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""autoHostRewrite"""'}), "(name='autoHostRewrite')\n", (107791, 107815), False, 'import pulumi\n'), ((108181, 108214), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""hostRewrite"""'}), "(name='hostRewrite')\n", (108194, 108214), False, 'import pulumi\n'), ((108536, 108569), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""idleTimeout"""'}), "(name='idleTimeout')\n", (108549, 108569), False, 'import pulumi\n'), ((109205, 109237), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxTimeout"""'}), "(name='maxTimeout')\n", (109218, 109237), False, 'import pulumi\n'), ((112321, 112363), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""directResponseAction"""'}), "(name='directResponseAction')\n", (112334, 112363), False, 'import pulumi\n'), ((112881, 112914), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""httpMatches"""'}), "(name='httpMatches')\n", (112894, 112914), False, 'import pulumi\n'), ((113405, 113442), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""httpRouteAction"""'}), "(name='httpRouteAction')\n", (113418, 113442), False, 'import pulumi\n'), ((113920, 113956), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""redirectAction"""'}), "(name='redirectAction')\n", (113933, 113956), False, 'import pulumi\n'), ((116553, 116586), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""httpMethods"""'}), "(name='httpMethods')\n", (116566, 116586), False, 'import pulumi\n'), ((120667, 120703), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""backendGroupId"""'}), "(name='backendGroupId')\n", (120680, 120703), False, 'import pulumi\n'), ((121034, 121071), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""autoHostRewrite"""'}), "(name='autoHostRewrite')\n", (121047, 121071), False, 'import pulumi\n'), ((121437, 121470), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""hostRewrite"""'}), "(name='hostRewrite')\n", (121450, 121470), False, 'import pulumi\n'), ((121792, 121825), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""idleTimeout"""'}), "(name='idleTimeout')\n", (121805, 121825), False, 'import pulumi\n'), ((122461, 122496), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""prefixRewrite"""'}), "(name='prefixRewrite')\n", (122474, 122496), False, 'import pulumi\n'), ((123298, 123332), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""upgradeTypes"""'}), "(name='upgradeTypes')\n", (123311, 123332), False, 'import pulumi\n'), ((125925, 125958), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""removeQuery"""'}), "(name='removeQuery')\n", (125938, 125958), False, 'import pulumi\n'), ((126226, 126259), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""replaceHost"""'}), "(name='replaceHost')\n", (126239, 126259), False, 'import pulumi\n'), ((126576, 126609), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""replacePath"""'}), "(name='replacePath')\n", (126589, 126609), False, 'import pulumi\n'), ((126921, 126954), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""replacePort"""'}), "(name='replacePort')\n", (126934, 126954), False, 'import pulumi\n'), ((127267, 127302), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""replacePrefix"""'}), "(name='replacePrefix')\n", (127280, 127302), False, 'import pulumi\n'), ((127790, 127825), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""replaceScheme"""'}), "(name='replaceScheme')\n", (127803, 127825), False, 'import pulumi\n'), ((128253, 128287), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""responseCode"""'}), "(name='responseCode')\n", (128266, 128287), False, 'import pulumi\n'), ((129105, 129147), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""diskPlacementGroupId"""'}), "(name='diskPlacementGroupId')\n", (129118, 129147), False, 'import pulumi\n'), ((131245, 131277), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""autoDelete"""'}), "(name='autoDelete')\n", (131258, 131277), False, 'import pulumi\n'), ((131671, 131703), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""deviceName"""'}), "(name='deviceName')\n", (131684, 131703), False, 'import pulumi\n'), ((132079, 132107), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""diskId"""'}), "(name='diskId')\n", (132092, 132107), False, 'import pulumi\n'), ((132429, 132467), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""initializeParams"""'}), "(name='initializeParams')\n", (132442, 132467), False, 'import pulumi\n'), ((135175, 135204), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""imageId"""'}), "(name='imageId')\n", (135188, 135204), False, 'import pulumi\n'), ((136114, 136146), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""snapshotId"""'}), "(name='snapshotId')\n", (136127, 136146), False, 'import pulumi\n'), ((139380, 139427), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxOpeningTrafficDuration"""'}), "(name='maxOpeningTrafficDuration')\n", (139393, 139427), False, 'import pulumi\n'), ((139980, 140015), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""statusMessage"""'}), "(name='statusMessage')\n", (139993, 140015), False, 'import pulumi\n'), ((140359, 140403), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""targetGroupDescription"""'}), "(name='targetGroupDescription')\n", (140372, 140403), False, 'import pulumi\n'), ((140796, 140831), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""targetGroupId"""'}), "(name='targetGroupId')\n", (140809, 140831), False, 'import pulumi\n'), ((141112, 141151), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""targetGroupLabels"""'}), "(name='targetGroupLabels')\n", (141125, 141151), False, 'import pulumi\n'), ((141572, 141609), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""targetGroupName"""'}), "(name='targetGroupName')\n", (141585, 141609), False, 'import pulumi\n'), ((144258, 144292), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxExpansion"""'}), "(name='maxExpansion')\n", (144271, 144292), False, 'import pulumi\n'), ((144706, 144742), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxUnavailable"""'}), "(name='maxUnavailable')\n", (144719, 144742), False, 'import pulumi\n'), ((145174, 145207), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxCreating"""'}), "(name='maxCreating')\n", (145187, 145207), False, 'import pulumi\n'), ((145575, 145608), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxDeleting"""'}), "(name='maxDeleting')\n", (145588, 145608), False, 'import pulumi\n'), ((145976, 146013), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""startupDuration"""'}), "(name='startupDuration')\n", (145989, 146013), False, 'import pulumi\n'), ((149271, 149309), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""healthyThreshold"""'}), "(name='healthyThreshold')\n", (149284, 149309), False, 'import pulumi\n'), ((149720, 149753), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""httpOptions"""'}), "(name='httpOptions')\n", (149733, 149753), False, 'import pulumi\n'), ((150588, 150620), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""tcpOptions"""'}), "(name='tcpOptions')\n", (150601, 150620), False, 'import pulumi\n'), ((151475, 151515), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""unhealthyThreshold"""'}), "(name='unhealthyThreshold')\n", (151488, 151515), False, 'import pulumi\n'), ((156547, 156579), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""instanceId"""'}), "(name='instanceId')\n", (156560, 156579), False, 'import pulumi\n'), ((157621, 157660), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""networkInterfaces"""'}), "(name='networkInterfaces')\n", (157634, 157660), False, 'import pulumi\n'), ((158579, 158616), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""statusChangedAt"""'}), "(name='statusChangedAt')\n", (158592, 158616), False, 'import pulumi\n'), ((158907, 158942), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""statusMessage"""'}), "(name='statusMessage')\n", (158920, 158942), False, 'import pulumi\n'), ((159286, 159314), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""zoneId"""'}), "(name='zoneId')\n", (159299, 159314), False, 'import pulumi\n'), ((162561, 162592), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""ipAddress"""'}), "(name='ipAddress')\n", (162574, 162592), False, 'import pulumi\n'), ((163482, 163515), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""ipv6Address"""'}), "(name='ipv6Address')\n", (163495, 163515), False, 'import pulumi\n'), ((163845, 163877), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""macAddress"""'}), "(name='macAddress')\n", (163858, 163877), False, 'import pulumi\n'), ((164509, 164543), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""natIpAddress"""'}), "(name='natIpAddress')\n", (164522, 164543), False, 'import pulumi\n'), ((164942, 164976), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""natIpVersion"""'}), "(name='natIpVersion')\n", (164955, 164976), False, 'import pulumi\n'), ((165323, 165353), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""subnetId"""'}), "(name='subnetId')\n", (165336, 165353), False, 'import pulumi\n'), ((171891, 171921), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""bootDisk"""'}), "(name='bootDisk')\n", (171904, 171921), False, 'import pulumi\n'), ((172356, 172395), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""networkInterfaces"""'}), "(name='networkInterfaces')\n", (172369, 172395), False, 'import pulumi\n'), ((176343, 176380), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""networkSettings"""'}), "(name='networkSettings')\n", (176356, 176380), False, 'import pulumi\n'), ((176927, 176964), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""placementPolicy"""'}), "(name='placementPolicy')\n", (176940, 176964), False, 'import pulumi\n'), ((177461, 177493), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""platformId"""'}), "(name='platformId')\n", (177474, 177493), False, 'import pulumi\n'), ((177880, 177918), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""schedulingPolicy"""'}), "(name='schedulingPolicy')\n", (177893, 177918), False, 'import pulumi\n'), ((178423, 178459), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""secondaryDisks"""'}), "(name='secondaryDisks')\n", (178436, 178459), False, 'import pulumi\n'), ((179002, 179040), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""serviceAccountId"""'}), "(name='serviceAccountId')\n", (179015, 179040), False, 'import pulumi\n'), ((180797, 180829), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""deviceName"""'}), "(name='deviceName')\n", (180810, 180829), False, 'import pulumi\n'), ((181195, 181223), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""diskId"""'}), "(name='diskId')\n", (181208, 181223), False, 'import pulumi\n'), ((181543, 181581), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""initializeParams"""'}), "(name='initializeParams')\n", (181556, 181581), False, 'import pulumi\n'), ((184165, 184194), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""imageId"""'}), "(name='imageId')\n", (184178, 184194), False, 'import pulumi\n'), ((184835, 184867), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""snapshotId"""'}), "(name='snapshotId')\n", (184848, 184867), False, 'import pulumi\n'), ((189299, 189331), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""dnsRecords"""'}), "(name='dnsRecords')\n", (189312, 189331), False, 'import pulumi\n'), ((189857, 189888), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""ipAddress"""'}), "(name='ipAddress')\n", (189870, 189888), False, 'import pulumi\n'), ((190778, 190811), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""ipv6Address"""'}), "(name='ipv6Address')\n", (190791, 190811), False, 'import pulumi\n'), ((191141, 191177), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""ipv6DnsRecords"""'}), "(name='ipv6DnsRecords')\n", (191154, 191177), False, 'import pulumi\n'), ((192029, 192064), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""natDnsRecords"""'}), "(name='natDnsRecords')\n", (192042, 192064), False, 'import pulumi\n'), ((192620, 192654), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""natIpAddress"""'}), "(name='natIpAddress')\n", (192633, 192654), False, 'import pulumi\n'), ((193053, 193084), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""networkId"""'}), "(name='networkId')\n", (193066, 193084), False, 'import pulumi\n'), ((193395, 193433), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""securityGroupIds"""'}), "(name='securityGroupIds')\n", (193408, 193433), False, 'import pulumi\n'), ((193851, 193882), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""subnetIds"""'}), "(name='subnetIds')\n", (193864, 193882), False, 'import pulumi\n'), ((195537, 195568), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""dnsZoneId"""'}), "(name='dnsZoneId')\n", (195550, 195568), False, 'import pulumi\n'), ((197774, 197805), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""dnsZoneId"""'}), "(name='dnsZoneId')\n", (197787, 197805), False, 'import pulumi\n'), ((200010, 200041), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""dnsZoneId"""'}), "(name='dnsZoneId')\n", (200023, 200041), False, 'import pulumi\n'), ((202087, 202125), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""placementGroupId"""'}), "(name='placementGroupId')\n", (202100, 202125), False, 'import pulumi\n'), ((203966, 204000), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""coreFraction"""'}), "(name='coreFraction')\n", (203979, 204000), False, 'import pulumi\n'), ((206770, 206802), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""deviceName"""'}), "(name='deviceName')\n", (206783, 206802), False, 'import pulumi\n'), ((207168, 207196), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""diskId"""'}), "(name='diskId')\n", (207181, 207196), False, 'import pulumi\n'), ((207516, 207554), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""initializeParams"""'}), "(name='initializeParams')\n", (207529, 207554), False, 'import pulumi\n'), ((210153, 210182), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""imageId"""'}), "(name='imageId')\n", (210166, 210182), False, 'import pulumi\n'), ((210823, 210855), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""snapshotId"""'}), "(name='snapshotId')\n", (210836, 210855), False, 'import pulumi\n'), ((213476, 213523), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxOpeningTrafficDuration"""'}), "(name='maxOpeningTrafficDuration')\n", (213489, 213523), False, 'import pulumi\n'), ((214076, 214111), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""statusMessage"""'}), "(name='statusMessage')\n", (214089, 214111), False, 'import pulumi\n'), ((214455, 214499), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""targetGroupDescription"""'}), "(name='targetGroupDescription')\n", (214468, 214499), False, 'import pulumi\n'), ((214892, 214927), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""targetGroupId"""'}), "(name='targetGroupId')\n", (214905, 214927), False, 'import pulumi\n'), ((215208, 215247), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""targetGroupLabels"""'}), "(name='targetGroupLabels')\n", (215221, 215247), False, 'import pulumi\n'), ((215668, 215705), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""targetGroupName"""'}), "(name='targetGroupName')\n", (215681, 215705), False, 'import pulumi\n'), ((217380, 217411), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""autoScale"""'}), "(name='autoScale')\n", (217393, 217411), False, 'import pulumi\n'), ((217867, 217899), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""fixedScale"""'}), "(name='fixedScale')\n", (217880, 217899), False, 'import pulumi\n'), ((218363, 218398), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""testAutoScale"""'}), "(name='testAutoScale')\n", (218376, 218398), False, 'import pulumi\n'), ((221933, 221966), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""initialSize"""'}), "(name='initialSize')\n", (221946, 221966), False, 'import pulumi\n'), ((222299, 222340), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""measurementDuration"""'}), "(name='measurementDuration')\n", (222312, 222340), False, 'import pulumi\n'), ((222908, 222950), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""cpuUtilizationTarget"""'}), "(name='cpuUtilizationTarget')\n", (222921, 222950), False, 'import pulumi\n'), ((223325, 223358), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""customRules"""'}), "(name='customRules')\n", (223338, 223358), False, 'import pulumi\n'), ((223869, 223898), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxSize"""'}), "(name='maxSize')\n", (223882, 223898), False, 'import pulumi\n'), ((224229, 224262), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""minZoneSize"""'}), "(name='minZoneSize')\n", (224242, 224262), False, 'import pulumi\n'), ((224635, 224678), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""stabilizationDuration"""'}), "(name='stabilizationDuration')\n", (224648, 224678), False, 'import pulumi\n'), ((225304, 225340), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""warmupDuration"""'}), "(name='warmupDuration')\n", (225317, 225340), False, 'import pulumi\n'), ((227920, 227952), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""metricName"""'}), "(name='metricName')\n", (227933, 227952), False, 'import pulumi\n'), ((228245, 228277), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""metricType"""'}), "(name='metricType')\n", (228258, 228277), False, 'import pulumi\n'), ((228585, 228615), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""ruleType"""'}), "(name='ruleType')\n", (228598, 228615), False, 'import pulumi\n'), ((229627, 229657), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""folderId"""'}), "(name='folderId')\n", (229640, 229657), False, 'import pulumi\n'), ((234360, 234393), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""initialSize"""'}), "(name='initialSize')\n", (234373, 234393), False, 'import pulumi\n'), ((234726, 234767), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""measurementDuration"""'}), "(name='measurementDuration')\n", (234739, 234767), False, 'import pulumi\n'), ((235335, 235377), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""cpuUtilizationTarget"""'}), "(name='cpuUtilizationTarget')\n", (235348, 235377), False, 'import pulumi\n'), ((235752, 235785), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""customRules"""'}), "(name='customRules')\n", (235765, 235785), False, 'import pulumi\n'), ((236304, 236333), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxSize"""'}), "(name='maxSize')\n", (236317, 236333), False, 'import pulumi\n'), ((236664, 236697), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""minZoneSize"""'}), "(name='minZoneSize')\n", (236677, 236697), False, 'import pulumi\n'), ((237070, 237113), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""stabilizationDuration"""'}), "(name='stabilizationDuration')\n", (237083, 237113), False, 'import pulumi\n'), ((237739, 237775), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""warmupDuration"""'}), "(name='warmupDuration')\n", (237752, 237775), False, 'import pulumi\n'), ((240359, 240391), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""metricName"""'}), "(name='metricName')\n", (240372, 240391), False, 'import pulumi\n'), ((240684, 240716), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""metricType"""'}), "(name='metricType')\n", (240697, 240716), False, 'import pulumi\n'), ((241024, 241054), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""ruleType"""'}), "(name='ruleType')\n", (241037, 241054), False, 'import pulumi\n'), ((242066, 242096), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""folderId"""'}), "(name='folderId')\n", (242079, 242096), False, 'import pulumi\n'), ((247539, 247569), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""subnetId"""'}), "(name='subnetId')\n", (247552, 247569), False, 'import pulumi\n'), ((247970, 248002), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""dnsRecords"""'}), "(name='dnsRecords')\n", (247983, 248002), False, 'import pulumi\n'), ((248762, 248793), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""ipAddress"""'}), "(name='ipAddress')\n", (248775, 248793), False, 'import pulumi\n'), ((249969, 250002), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""ipv6Address"""'}), "(name='ipv6Address')\n", (249982, 250002), False, 'import pulumi\n'), ((250352, 250388), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""ipv6DnsRecords"""'}), "(name='ipv6DnsRecords')\n", (250365, 250388), False, 'import pulumi\n'), ((250937, 250969), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""macAddress"""'}), "(name='macAddress')\n", (250950, 250969), False, 'import pulumi\n'), ((251571, 251606), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""natDnsRecords"""'}), "(name='natDnsRecords')\n", (251584, 251606), False, 'import pulumi\n'), ((252152, 252186), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""natIpAddress"""'}), "(name='natIpAddress')\n", (252165, 252186), False, 'import pulumi\n'), ((252613, 252647), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""natIpVersion"""'}), "(name='natIpVersion')\n", (252626, 252647), False, 'import pulumi\n'), ((252923, 252961), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""securityGroupIds"""'}), "(name='securityGroupIds')\n", (252936, 252961), False, 'import pulumi\n'), ((254643, 254674), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""dnsZoneId"""'}), "(name='dnsZoneId')\n", (254656, 254674), False, 'import pulumi\n'), ((256889, 256920), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""dnsZoneId"""'}), "(name='dnsZoneId')\n", (256902, 256920), False, 'import pulumi\n'), ((259134, 259165), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""dnsZoneId"""'}), "(name='dnsZoneId')\n", (259147, 259165), False, 'import pulumi\n'), ((260477, 260515), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""placementGroupId"""'}), "(name='placementGroupId')\n", (260490, 260515), False, 'import pulumi\n'), ((262304, 262338), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""coreFraction"""'}), "(name='coreFraction')\n", (262317, 262338), False, 'import pulumi\n'), ((264890, 264918), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""diskId"""'}), "(name='diskId')\n", (264903, 264918), False, 'import pulumi\n'), ((265220, 265252), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""autoDelete"""'}), "(name='autoDelete')\n", (265233, 265252), False, 'import pulumi\n'), ((265646, 265678), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""deviceName"""'}), "(name='deviceName')\n", (265659, 265678), False, 'import pulumi\n'), ((267472, 267509), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""subclusterSpecs"""'}), "(name='subclusterSpecs')\n", (267485, 267509), False, 'import pulumi\n'), ((268441, 268472), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""versionId"""'}), "(name='versionId')\n", (268454, 268472), False, 'import pulumi\n'), ((270821, 270856), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""sshPublicKeys"""'}), "(name='sshPublicKeys')\n", (270834, 270856), False, 'import pulumi\n'), ((273328, 273360), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""hostsCount"""'}), "(name='hostsCount')\n", (273341, 273360), False, 'import pulumi\n'), ((274744, 274774), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""subnetId"""'}), "(name='subnetId')\n", (274757, 274774), False, 'import pulumi\n'), ((275168, 275207), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""autoscalingConfig"""'}), "(name='autoscalingConfig')\n", (275181, 275207), False, 'import pulumi\n'), ((278809, 278844), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxHostsCount"""'}), "(name='maxHostsCount')\n", (278822, 278844), False, 'import pulumi\n'), ((279189, 279231), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""cpuUtilizationTarget"""'}), "(name='cpuUtilizationTarget')\n", (279202, 279231), False, 'import pulumi\n'), ((279723, 279764), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""decommissionTimeout"""'}), "(name='decommissionTimeout')\n", (279736, 279764), False, 'import pulumi\n'), ((280175, 280216), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""measurementDuration"""'}), "(name='measurementDuration')\n", (280188, 280216), False, 'import pulumi\n'), ((281227, 281270), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""stabilizationDuration"""'}), "(name='stabilizationDuration')\n", (281240, 281270), False, 'import pulumi\n'), ((281854, 281890), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""warmupDuration"""'}), "(name='warmupDuration')\n", (281867, 281890), False, 'import pulumi\n'), ((283344, 283374), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""diskSize"""'}), "(name='diskSize')\n", (283357, 283374), False, 'import pulumi\n'), ((283694, 283732), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""resourcePresetId"""'}), "(name='resourcePresetId')\n", (283707, 283732), False, 'import pulumi\n'), ((284232, 284264), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""diskTypeId"""'}), "(name='diskTypeId')\n", (284245, 284264), False, 'import pulumi\n'), ((284830, 284863), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""zipFilename"""'}), "(name='zipFilename')\n", (284843, 284863), False, 'import pulumi\n'), ((285543, 285575), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""bucketName"""'}), "(name='bucketName')\n", (285556, 285575), False, 'import pulumi\n'), ((285816, 285848), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""objectName"""'}), "(name='objectName')\n", (285829, 285848), False, 'import pulumi\n'), ((287119, 287159), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""zoneInstancesLimit"""'}), "(name='zoneInstancesLimit')\n", (287132, 287159), False, 'import pulumi\n'), ((287465, 287504), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""zoneRequestsLimit"""'}), "(name='zoneRequestsLimit')\n", (287478, 287504), False, 'import pulumi\n'), ((288111, 288140), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""queueId"""'}), "(name='queueId')\n", (288124, 288140), False, 'import pulumi\n'), ((288366, 288404), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""serviceAccountId"""'}), "(name='serviceAccountId')\n", (288379, 288404), False, 'import pulumi\n'), ((289727, 289762), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""retryAttempts"""'}), "(name='retryAttempts')\n", (289740, 289762), False, 'import pulumi\n'), ((290038, 290073), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""retryInterval"""'}), "(name='retryInterval')\n", (290051, 290073), False, 'import pulumi\n'), ((290349, 290387), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""serviceAccountId"""'}), "(name='serviceAccountId')\n", (290362, 290387), False, 'import pulumi\n'), ((291399, 291431), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""registryId"""'}), "(name='registryId')\n", (291412, 291431), False, 'import pulumi\n'), ((291672, 291702), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""deviceId"""'}), "(name='deviceId')\n", (291685, 291702), False, 'import pulumi\n'), ((292688, 292721), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""batchCutoff"""'}), "(name='batchCutoff')\n", (292701, 292721), False, 'import pulumi\n'), ((292967, 293000), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""logGroupIds"""'}), "(name='logGroupIds')\n", (292980, 293000), False, 'import pulumi\n'), ((293299, 293330), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""batchSize"""'}), "(name='batchSize')\n", (293312, 293330), False, 'import pulumi\n'), ((294424, 294457), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""batchCutoff"""'}), "(name='batchCutoff')\n", (294437, 294457), False, 'import pulumi\n'), ((294703, 294732), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""groupId"""'}), "(name='groupId')\n", (294716, 294732), False, 'import pulumi\n'), ((295235, 295268), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""resourceIds"""'}), "(name='resourceIds')\n", (295248, 295268), False, 'import pulumi\n'), ((295562, 295597), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""resourceTypes"""'}), "(name='resourceTypes')\n", (295575, 295597), False, 'import pulumi\n'), ((295901, 295932), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""batchSize"""'}), "(name='batchSize')\n", (295914, 295932), False, 'import pulumi\n'), ((296962, 296995), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""batchCutoff"""'}), "(name='batchCutoff')\n", (296975, 296995), False, 'import pulumi\n'), ((297241, 297270), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""queueId"""'}), "(name='queueId')\n", (297254, 297270), False, 'import pulumi\n'), ((297496, 297534), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""serviceAccountId"""'}), "(name='serviceAccountId')\n", (297509, 297534), False, 'import pulumi\n'), ((297810, 297841), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""batchSize"""'}), "(name='batchSize')\n", (297823, 297841), False, 'import pulumi\n'), ((298097, 298136), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""visibilityTimeout"""'}), "(name='visibilityTimeout')\n", (298110, 298136), False, 'import pulumi\n'), ((299342, 299372), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""bucketId"""'}), "(name='bucketId')\n", (299355, 299372), False, 'import pulumi\n'), ((301056, 301092), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""cronExpression"""'}), "(name='cronExpression')\n", (301069, 301092), False, 'import pulumi\n'), ((301665, 301692), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""keyId"""'}), "(name='keyId')\n", (301678, 301692), False, 'import pulumi\n'), ((306583, 306625), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""clusterCaCertificate"""'}), "(name='clusterCaCertificate')\n", (306596, 306625), False, 'import pulumi\n'), ((307069, 307108), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""externalV4Address"""'}), "(name='externalV4Address')\n", (307082, 307108), False, 'import pulumi\n'), ((307517, 307557), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""externalV4Endpoint"""'}), "(name='externalV4Endpoint')\n", (307530, 307557), False, 'import pulumi\n'), ((308016, 308055), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""internalV4Address"""'}), "(name='internalV4Address')\n", (308029, 308055), False, 'import pulumi\n'), ((308464, 308504), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""internalV4Endpoint"""'}), "(name='internalV4Endpoint')\n", (308477, 308504), False, 'import pulumi\n'), ((308934, 308973), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maintenancePolicy"""'}), "(name='maintenancePolicy')\n", (308947, 308973), False, 'import pulumi\n'), ((309777, 309807), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""publicIp"""'}), "(name='publicIp')\n", (309790, 309807), False, 'import pulumi\n'), ((310665, 310703), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""securityGroupIds"""'}), "(name='securityGroupIds')\n", (310678, 310703), False, 'import pulumi\n'), ((311518, 311551), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""versionInfo"""'}), "(name='versionInfo')\n", (311531, 311551), False, 'import pulumi\n'), ((313626, 313659), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""autoUpgrade"""'}), "(name='autoUpgrade')\n", (313639, 313659), False, 'import pulumi\n'), ((314056, 314096), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maintenanceWindows"""'}), "(name='maintenanceWindows')\n", (314069, 314096), False, 'import pulumi\n'), ((315631, 315662), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""startTime"""'}), "(name='startTime')\n", (315644, 315662), False, 'import pulumi\n'), ((318372, 318402), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""subnetId"""'}), "(name='subnetId')\n", (318385, 318402), False, 'import pulumi\n'), ((320740, 320776), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""currentVersion"""'}), "(name='currentVersion')\n", (320753, 320776), False, 'import pulumi\n'), ((321142, 321184), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""newRevisionAvailable"""'}), "(name='newRevisionAvailable')\n", (321155, 321184), False, 'import pulumi\n'), ((321773, 321813), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""newRevisionSummary"""'}), "(name='newRevisionSummary')\n", (321786, 321813), False, 'import pulumi\n'), ((322295, 322334), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""versionDeprecated"""'}), "(name='versionDeprecated')\n", (322308, 322334), False, 'import pulumi\n'), ((323349, 323379), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""subnetId"""'}), "(name='subnetId')\n", (323362, 323379), False, 'import pulumi\n'), ((327209, 327239), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""subnetId"""'}), "(name='subnetId')\n", (327222, 327239), False, 'import pulumi\n'), ((328585, 328619), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxExpansion"""'}), "(name='maxExpansion')\n", (328598, 328619), False, 'import pulumi\n'), ((329017, 329053), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxUnavailable"""'}), "(name='maxUnavailable')\n", (329030, 329053), False, 'import pulumi\n'), ((333419, 333449), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""bootDisk"""'}), "(name='bootDisk')\n", (333432, 333449), False, 'import pulumi\n'), ((335052, 335097), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""networkAccelerationType"""'}), "(name='networkAccelerationType')\n", (335065, 335097), False, 'import pulumi\n'), ((335534, 335573), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""networkInterfaces"""'}), "(name='networkInterfaces')\n", (335547, 335573), False, 'import pulumi\n'), ((336168, 336205), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""placementPolicy"""'}), "(name='placementPolicy')\n", (336181, 336205), False, 'import pulumi\n'), ((336700, 336732), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""platformId"""'}), "(name='platformId')\n", (336713, 336732), False, 'import pulumi\n'), ((337467, 337505), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""schedulingPolicy"""'}), "(name='schedulingPolicy')\n", (337480, 337505), False, 'import pulumi\n'), ((340642, 340673), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""subnetIds"""'}), "(name='subnetIds')\n", (340655, 340673), False, 'import pulumi\n'), ((342090, 342128), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""securityGroupIds"""'}), "(name='securityGroupIds')\n", (342103, 342128), False, 'import pulumi\n'), ((342932, 342970), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""placementGroupId"""'}), "(name='placementGroupId')\n", (342945, 342970), False, 'import pulumi\n'), ((344044, 344078), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""coreFraction"""'}), "(name='coreFraction')\n", (344057, 344078), False, 'import pulumi\n'), ((347082, 347114), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""autoRepair"""'}), "(name='autoRepair')\n", (347095, 347114), False, 'import pulumi\n'), ((347499, 347532), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""autoUpgrade"""'}), "(name='autoUpgrade')\n", (347512, 347532), False, 'import pulumi\n'), ((347922, 347962), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maintenanceWindows"""'}), "(name='maintenanceWindows')\n", (347935, 347962), False, 'import pulumi\n'), ((349263, 349294), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""startTime"""'}), "(name='startTime')\n", (349276, 349294), False, 'import pulumi\n'), ((350621, 350652), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""autoScale"""'}), "(name='autoScale')\n", (350634, 350652), False, 'import pulumi\n'), ((351102, 351134), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""fixedScale"""'}), "(name='fixedScale')\n", (351115, 351134), False, 'import pulumi\n'), ((355510, 355546), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""currentVersion"""'}), "(name='currentVersion')\n", (355523, 355546), False, 'import pulumi\n'), ((355912, 355954), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""newRevisionAvailable"""'}), "(name='newRevisionAvailable')\n", (355925, 355954), False, 'import pulumi\n'), ((356546, 356586), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""newRevisionSummary"""'}), "(name='newRevisionSummary')\n", (356559, 356586), False, 'import pulumi\n'), ((357068, 357107), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""versionDeprecated"""'}), "(name='versionDeprecated')\n", (357081, 357107), False, 'import pulumi\n'), ((358764, 358799), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""targetGroupId"""'}), "(name='targetGroupId')\n", (358777, 358799), False, 'import pulumi\n'), ((361835, 361873), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""healthyThreshold"""'}), "(name='healthyThreshold')\n", (361848, 361873), False, 'import pulumi\n'), ((362293, 362326), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""httpOptions"""'}), "(name='httpOptions')\n", (362306, 362326), False, 'import pulumi\n'), ((363173, 363205), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""tcpOptions"""'}), "(name='tcpOptions')\n", (363186, 363205), False, 'import pulumi\n'), ((364066, 364106), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""unhealthyThreshold"""'}), "(name='unhealthyThreshold')\n", (364079, 364106), False, 'import pulumi\n'), ((368887, 368928), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""externalAddressSpec"""'}), "(name='externalAddressSpec')\n", (368900, 368928), False, 'import pulumi\n'), ((369443, 369484), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""internalAddressSpec"""'}), "(name='internalAddressSpec')\n", (369456, 369484), False, 'import pulumi\n'), ((370356, 370388), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""targetPort"""'}), "(name='targetPort')\n", (370369, 370388), False, 'import pulumi\n'), ((371974, 372005), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""ipVersion"""'}), "(name='ipVersion')\n", (371987, 372005), False, 'import pulumi\n'), ((373416, 373446), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""subnetId"""'}), "(name='subnetId')\n", (373429, 373446), False, 'import pulumi\n'), ((374201, 374232), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""ipVersion"""'}), "(name='ipVersion')\n", (374214, 374232), False, 'import pulumi\n'), ((375515, 375545), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""subnetId"""'}), "(name='subnetId')\n", (375528, 375545), False, 'import pulumi\n'), ((377119, 377149), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""dataLens"""'}), "(name='dataLens')\n", (377132, 377149), False, 'import pulumi\n'), ((378213, 378241), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""webSql"""'}), "(name='webSql')\n", (378226, 378241), False, 'import pulumi\n'), ((389775, 389815), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""backgroundPoolSize"""'}), "(name='backgroundPoolSize')\n", (389788, 389815), False, 'import pulumi\n'), ((390121, 390169), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""backgroundSchedulePoolSize"""'}), "(name='backgroundSchedulePoolSize')\n", (390134, 390169), False, 'import pulumi\n'), ((391046, 391078), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""geobaseUri"""'}), "(name='geobaseUri')\n", (391059, 391078), False, 'import pulumi\n'), ((391339, 391376), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""graphiteRollups"""'}), "(name='graphiteRollups')\n", (391352, 391376), False, 'import pulumi\n'), ((392345, 392378), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""kafkaTopics"""'}), "(name='kafkaTopics')\n", (392358, 392378), False, 'import pulumi\n'), ((392895, 392933), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""keepAliveTimeout"""'}), "(name='keepAliveTimeout')\n", (392908, 392933), False, 'import pulumi\n'), ((393229, 393259), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""logLevel"""'}), "(name='logLevel')\n", (393242, 393259), False, 'import pulumi\n'), ((393510, 393545), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""markCacheSize"""'}), "(name='markCacheSize')\n", (393523, 393545), False, 'import pulumi\n'), ((393826, 393868), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxConcurrentQueries"""'}), "(name='maxConcurrentQueries')\n", (393839, 393868), False, 'import pulumi\n'), ((394184, 394220), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxConnections"""'}), "(name='maxConnections')\n", (394197, 394220), False, 'import pulumi\n'), ((394501, 394545), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxPartitionSizeToDrop"""'}), "(name='maxPartitionSizeToDrop')\n", (394514, 394545), False, 'import pulumi\n'), ((394881, 394921), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxTableSizeToDrop"""'}), "(name='maxTableSizeToDrop')\n", (394894, 394921), False, 'import pulumi\n'), ((395237, 395268), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""mergeTree"""'}), "(name='mergeTree')\n", (395250, 395268), False, 'import pulumi\n'), ((395719, 395757), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""metricLogEnabled"""'}), "(name='metricLogEnabled')\n", (395732, 395757), False, 'import pulumi\n'), ((396055, 396099), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""metricLogRetentionSize"""'}), "(name='metricLogRetentionSize')\n", (396068, 396099), False, 'import pulumi\n'), ((396430, 396474), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""metricLogRetentionTime"""'}), "(name='metricLogRetentionTime')\n", (396443, 396474), False, 'import pulumi\n'), ((396805, 396847), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""partLogRetentionSize"""'}), "(name='partLogRetentionSize')\n", (396818, 396847), False, 'import pulumi\n'), ((397168, 397210), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""partLogRetentionTime"""'}), "(name='partLogRetentionTime')\n", (397181, 397210), False, 'import pulumi\n'), ((397531, 397574), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""queryLogRetentionSize"""'}), "(name='queryLogRetentionSize')\n", (397544, 397574), False, 'import pulumi\n'), ((397900, 397943), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""queryLogRetentionTime"""'}), "(name='queryLogRetentionTime')\n", (397913, 397943), False, 'import pulumi\n'), ((398269, 398312), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""queryThreadLogEnabled"""'}), "(name='queryThreadLogEnabled')\n", (398282, 398312), False, 'import pulumi\n'), ((398640, 398689), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""queryThreadLogRetentionSize"""'}), "(name='queryThreadLogRetentionSize')\n", (398653, 398689), False, 'import pulumi\n'), ((399050, 399099), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""queryThreadLogRetentionTime"""'}), "(name='queryThreadLogRetentionTime')\n", (399063, 399099), False, 'import pulumi\n'), ((399915, 399951), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""textLogEnabled"""'}), "(name='textLogEnabled')\n", (399928, 399951), False, 'import pulumi\n'), ((400239, 400273), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""textLogLevel"""'}), "(name='textLogLevel')\n", (400252, 400273), False, 'import pulumi\n'), ((400549, 400591), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""textLogRetentionSize"""'}), "(name='textLogRetentionSize')\n", (400562, 400591), False, 'import pulumi\n'), ((400912, 400954), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""textLogRetentionTime"""'}), "(name='textLogRetentionTime')\n", (400925, 400954), False, 'import pulumi\n'), ((401534, 401571), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""traceLogEnabled"""'}), "(name='traceLogEnabled')\n", (401547, 401571), False, 'import pulumi\n'), ((401864, 401907), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""traceLogRetentionSize"""'}), "(name='traceLogRetentionSize')\n", (401877, 401907), False, 'import pulumi\n'), ((402233, 402276), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""traceLogRetentionTime"""'}), "(name='traceLogRetentionTime')\n", (402246, 402276), False, 'import pulumi\n'), ((402602, 402645), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""uncompressedCacheSize"""'}), "(name='uncompressedCacheSize')\n", (402615, 402645), False, 'import pulumi\n'), ((404349, 404382), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""minPartSize"""'}), "(name='minPartSize')\n", (404362, 404382), False, 'import pulumi\n'), ((404840, 404878), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""minPartSizeRatio"""'}), "(name='minPartSizeRatio')\n", (404853, 404878), False, 'import pulumi\n'), ((411225, 411260), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""saslMechanism"""'}), "(name='saslMechanism')\n", (411238, 411260), False, 'import pulumi\n'), ((411613, 411647), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""saslPassword"""'}), "(name='saslPassword')\n", (411626, 411647), False, 'import pulumi\n'), ((411981, 412015), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""saslUsername"""'}), "(name='saslUsername')\n", (411994, 412015), False, 'import pulumi\n'), ((412344, 412382), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""securityProtocol"""'}), "(name='securityProtocol')\n", (412357, 412382), False, 'import pulumi\n'), ((415341, 415376), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""saslMechanism"""'}), "(name='saslMechanism')\n", (415354, 415376), False, 'import pulumi\n'), ((415729, 415763), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""saslPassword"""'}), "(name='saslPassword')\n", (415742, 415763), False, 'import pulumi\n'), ((416097, 416131), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""saslUsername"""'}), "(name='saslUsername')\n", (416110, 416131), False, 'import pulumi\n'), ((416460, 416498), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""securityProtocol"""'}), "(name='securityProtocol')\n", (416473, 416498), False, 'import pulumi\n'), ((420415, 420468), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxBytesToMergeAtMinSpaceInPool"""'}), "(name='maxBytesToMergeAtMinSpaceInPool')\n", (420428, 420468), False, 'import pulumi\n'), ((421052, 421100), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxReplicatedMergesInQueue"""'}), "(name='maxReplicatedMergesInQueue')\n", (421065, 421100), False, 'import pulumi\n'), ((421613, 421681), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""numberOfFreeEntriesInPoolToLowerMaxSizeOfMerge"""'}), "(name='numberOfFreeEntriesInPoolToLowerMaxSizeOfMerge')\n", (421626, 421681), False, 'import pulumi\n'), ((422519, 422559), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""partsToDelayInsert"""'}), "(name='partsToDelayInsert')\n", (422532, 422559), False, 'import pulumi\n'), ((423065, 423105), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""partsToThrowInsert"""'}), "(name='partsToThrowInsert')\n", (423078, 423105), False, 'import pulumi\n'), ((423593, 423644), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""replicatedDeduplicationWindow"""'}), "(name='replicatedDeduplicationWindow')\n", (423606, 423644), False, 'import pulumi\n'), ((424157, 424215), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""replicatedDeduplicationWindowSeconds"""'}), "(name='replicatedDeduplicationWindowSeconds')\n", (424170, 424215), False, 'import pulumi\n'), ((426745, 426775), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""diskSize"""'}), "(name='diskSize')\n", (426758, 426775), False, 'import pulumi\n'), ((427105, 427137), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""diskTypeId"""'}), "(name='diskTypeId')\n", (427118, 427137), False, 'import pulumi\n'), ((427585, 427623), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""resourcePresetId"""'}), "(name='resourcePresetId')\n", (427598, 427623), False, 'import pulumi\n'), ((433533, 433569), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""assignPublicIp"""'}), "(name='assignPublicIp')\n", (433546, 433569), False, 'import pulumi\n'), ((434304, 434335), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""shardName"""'}), "(name='shardName')\n", (434317, 434335), False, 'import pulumi\n'), ((434672, 434702), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""subnetId"""'}), "(name='subnetId')\n", (434685, 434702), False, 'import pulumi\n'), ((439941, 439973), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""shardNames"""'}), "(name='shardNames')\n", (439954, 439973), False, 'import pulumi\n'), ((444467, 444501), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""databaseName"""'}), "(name='databaseName')\n", (444480, 444501), False, 'import pulumi\n'), ((446468, 446506), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""intervalDuration"""'}), "(name='intervalDuration')\n", (446481, 446506), False, 'import pulumi\n'), ((447182, 447217), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""executionTime"""'}), "(name='executionTime')\n", (447195, 447217), False, 'import pulumi\n'), ((447901, 447931), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""readRows"""'}), "(name='readRows')\n", (447914, 447931), False, 'import pulumi\n'), ((448309, 448341), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""resultRows"""'}), "(name='resultRows')\n", (448322, 448341), False, 'import pulumi\n'), ((481080, 481119), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""addHttpCorsHeader"""'}), "(name='addHttpCorsHeader')\n", (481093, 481119), False, 'import pulumi\n'), ((481499, 481529), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""allowDdl"""'}), "(name='allowDdl')\n", (481512, 481529), False, 'import pulumi\n'), ((482163, 482203), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""compileExpressions"""'}), "(name='compileExpressions')\n", (482176, 482203), False, 'import pulumi\n'), ((482570, 482606), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""connectTimeout"""'}), "(name='connectTimeout')\n", (482583, 482606), False, 'import pulumi\n'), ((483005, 483054), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""countDistinctImplementation"""'}), "(name='countDistinctImplementation')\n", (483018, 483054), False, 'import pulumi\n'), ((483538, 483580), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""distinctOverflowMode"""'}), "(name='distinctOverflowMode')\n", (483551, 483580), False, 'import pulumi\n'), ((483993, 484052), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""distributedAggregationMemoryEfficient"""'}), "(name='distributedAggregationMemoryEfficient')\n", (484006, 484052), False, 'import pulumi\n'), ((484542, 484589), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""distributedDdlTaskTimeout"""'}), "(name='distributedDdlTaskTimeout')\n", (484555, 484589), False, 'import pulumi\n'), ((485009, 485053), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""distributedProductMode"""'}), "(name='distributedProductMode')\n", (485022, 485053), False, 'import pulumi\n'), ((485460, 485517), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""emptyResultForAggregationByEmptySet"""'}), "(name='emptyResultForAggregationByEmptySet')\n", (485473, 485517), False, 'import pulumi\n'), ((485993, 486036), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""enableHttpCompression"""'}), "(name='enableHttpCompression')\n", (486006, 486036), False, 'import pulumi\n'), ((486464, 486530), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""fallbackToStaleReplicasForDistributedQueries"""'}), "(name='fallbackToStaleReplicasForDistributedQueries')\n", (486477, 486530), False, 'import pulumi\n'), ((487095, 487133), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""forceIndexByDate"""'}), "(name='forceIndexByDate')\n", (487108, 487133), False, 'import pulumi\n'), ((487529, 487566), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""forcePrimaryKey"""'}), "(name='forcePrimaryKey')\n", (487542, 487566), False, 'import pulumi\n'), ((487964, 488005), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""groupByOverflowMode"""'}), "(name='groupByOverflowMode')\n", (487977, 488005), False, 'import pulumi\n'), ((488423, 488469), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""groupByTwoLevelThreshold"""'}), "(name='groupByTwoLevelThreshold')\n", (488436, 488469), False, 'import pulumi\n'), ((488942, 488993), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""groupByTwoLevelThresholdBytes"""'}), "(name='groupByTwoLevelThresholdBytes')\n", (488955, 488993), False, 'import pulumi\n'), ((489497, 489540), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""httpConnectionTimeout"""'}), "(name='httpConnectionTimeout')\n", (489510, 489540), False, 'import pulumi\n'), ((489938, 489987), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""httpHeadersProgressInterval"""'}), "(name='httpHeadersProgressInterval')\n", (489951, 489987), False, 'import pulumi\n'), ((490479, 490519), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""httpReceiveTimeout"""'}), "(name='httpReceiveTimeout')\n", (490492, 490519), False, 'import pulumi\n'), ((490902, 490939), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""httpSendTimeout"""'}), "(name='httpSendTimeout')\n", (490915, 490939), False, 'import pulumi\n'), ((491307, 491364), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""inputFormatDefaultsForOmittedFields"""'}), "(name='inputFormatDefaultsForOmittedFields')\n", (491320, 491364), False, 'import pulumi\n'), ((491919, 491978), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""inputFormatValuesInterpretExpressions"""'}), "(name='inputFormatValuesInterpretExpressions')\n", (491932, 491978), False, 'import pulumi\n'), ((492511, 492545), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""insertQuorum"""'}), "(name='insertQuorum')\n", (492524, 492545), False, 'import pulumi\n'), ((492875, 492916), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""insertQuorumTimeout"""'}), "(name='insertQuorumTimeout')\n", (492888, 492916), False, 'import pulumi\n'), ((493302, 493340), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""joinOverflowMode"""'}), "(name='joinOverflowMode')\n", (493315, 493340), False, 'import pulumi\n'), ((493721, 493755), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""joinUseNulls"""'}), "(name='joinUseNulls')\n", (493734, 493755), False, 'import pulumi\n'), ((494200, 494249), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""joinedSubqueryRequiresAlias"""'}), "(name='joinedSubqueryRequiresAlias')\n", (494213, 494249), False, 'import pulumi\n'), ((494735, 494790), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""lowCardinalityAllowInNativeFormat"""'}), "(name='lowCardinalityAllowInNativeFormat')\n", (494748, 494790), False, 'import pulumi\n'), ((495299, 495332), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxAstDepth"""'}), "(name='maxAstDepth')\n", (495312, 495332), False, 'import pulumi\n'), ((495671, 495707), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxAstElements"""'}), "(name='maxAstElements')\n", (495684, 495707), False, 'import pulumi\n'), ((496064, 496098), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxBlockSize"""'}), "(name='maxBlockSize')\n", (496077, 496098), False, 'import pulumi\n'), ((496492, 496543), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxBytesBeforeExternalGroupBy"""'}), "(name='maxBytesBeforeExternalGroupBy')\n", (496505, 496543), False, 'import pulumi\n'), ((497023, 497071), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxBytesBeforeExternalSort"""'}), "(name='maxBytesBeforeExternalSort')\n", (497036, 497071), False, 'import pulumi\n'), ((497602, 497642), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxBytesInDistinct"""'}), "(name='maxBytesInDistinct')\n", (497615, 497642), False, 'import pulumi\n'), ((498075, 498111), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxBytesInJoin"""'}), "(name='maxBytesInJoin')\n", (498088, 498111), False, 'import pulumi\n'), ((498494, 498529), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxBytesInSet"""'}), "(name='maxBytesInSet')\n", (498507, 498529), False, 'import pulumi\n'), ((498935, 498971), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxBytesToRead"""'}), "(name='maxBytesToRead')\n", (498948, 498971), False, 'import pulumi\n'), ((499401, 499437), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxBytesToSort"""'}), "(name='maxBytesToSort')\n", (499414, 499437), False, 'import pulumi\n'), ((499858, 499898), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxBytesToTransfer"""'}), "(name='maxBytesToTransfer')\n", (499871, 499898), False, 'import pulumi\n'), ((500386, 500424), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxColumnsToRead"""'}), "(name='maxColumnsToRead')\n", (500399, 500424), False, 'import pulumi\n'), ((500843, 500881), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxExecutionTime"""'}), "(name='maxExecutionTime')\n", (500856, 500881), False, 'import pulumi\n'), ((501266, 501310), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxExpandedAstElements"""'}), "(name='maxExpandedAstElements')\n", (501279, 501310), False, 'import pulumi\n'), ((501742, 501782), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxInsertBlockSize"""'}), "(name='maxInsertBlockSize')\n", (501755, 501782), False, 'import pulumi\n'), ((502201, 502237), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxMemoryUsage"""'}), "(name='maxMemoryUsage')\n", (502214, 502237), False, 'import pulumi\n'), ((502641, 502684), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxMemoryUsageForUser"""'}), "(name='maxMemoryUsageForUser')\n", (502654, 502684), False, 'import pulumi\n'), ((503143, 503184), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxNetworkBandwidth"""'}), "(name='maxNetworkBandwidth')\n", (503156, 503184), False, 'import pulumi\n'), ((503603, 503651), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxNetworkBandwidthForUser"""'}), "(name='maxNetworkBandwidthForUser')\n", (503616, 503651), False, 'import pulumi\n'), ((504115, 504149), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxQuerySize"""'}), "(name='maxQuerySize')\n", (504128, 504149), False, 'import pulumi\n'), ((504543, 504601), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxReplicaDelayForDistributedQueries"""'}), "(name='maxReplicaDelayForDistributedQueries')\n", (504556, 504601), False, 'import pulumi\n'), ((505095, 505131), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxResultBytes"""'}), "(name='maxResultBytes')\n", (505108, 505131), False, 'import pulumi\n'), ((505491, 505526), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxResultRows"""'}), "(name='maxResultRows')\n", (505504, 505526), False, 'import pulumi\n'), ((505880, 505919), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxRowsInDistinct"""'}), "(name='maxRowsInDistinct')\n", (505893, 505919), False, 'import pulumi\n'), ((506322, 506357), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxRowsInJoin"""'}), "(name='maxRowsInJoin')\n", (506335, 506357), False, 'import pulumi\n'), ((506734, 506768), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxRowsInSet"""'}), "(name='maxRowsInSet')\n", (506747, 506768), False, 'import pulumi\n'), ((507168, 507206), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxRowsToGroupBy"""'}), "(name='maxRowsToGroupBy')\n", (507181, 507206), False, 'import pulumi\n'), ((507621, 507656), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxRowsToRead"""'}), "(name='maxRowsToRead')\n", (507634, 507656), False, 'import pulumi\n'), ((508060, 508095), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxRowsToSort"""'}), "(name='maxRowsToSort')\n", (508073, 508095), False, 'import pulumi\n'), ((508490, 508529), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxRowsToTransfer"""'}), "(name='maxRowsToTransfer')\n", (508503, 508529), False, 'import pulumi\n'), ((508991, 509032), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxTemporaryColumns"""'}), "(name='maxTemporaryColumns')\n", (509004, 509032), False, 'import pulumi\n'), ((509514, 509563), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxTemporaryNonConstColumns"""'}), "(name='maxTemporaryNonConstColumns')\n", (509527, 509563), False, 'import pulumi\n'), ((510095, 510127), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxThreads"""'}), "(name='maxThreads')\n", (510108, 510127), False, 'import pulumi\n'), ((510527, 510576), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""mergeTreeMaxBytesToUseCache"""'}), "(name='mergeTreeMaxBytesToUseCache')\n", (510540, 510576), False, 'import pulumi\n'), ((511118, 511166), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""mergeTreeMaxRowsToUseCache"""'}), "(name='mergeTreeMaxRowsToUseCache')\n", (511131, 511166), False, 'import pulumi\n'), ((511701, 511757), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""mergeTreeMinBytesForConcurrentRead"""'}), "(name='mergeTreeMinBytesForConcurrentRead')\n", (511714, 511757), False, 'import pulumi\n'), ((512396, 512451), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""mergeTreeMinRowsForConcurrentRead"""'}), "(name='mergeTreeMinRowsForConcurrentRead')\n", (512409, 512451), False, 'import pulumi\n'), ((513087, 513130), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""minBytesToUseDirectIo"""'}), "(name='minBytesToUseDirectIo')\n", (513100, 513130), False, 'import pulumi\n'), ((513580, 513619), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""minCountToCompile"""'}), "(name='minCountToCompile')\n", (513593, 513619), False, 'import pulumi\n'), ((514044, 514093), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""minCountToCompileExpression"""'}), "(name='minCountToCompileExpression')\n", (514057, 514093), False, 'import pulumi\n'), ((514578, 514617), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""minExecutionSpeed"""'}), "(name='minExecutionSpeed')\n", (514591, 514617), False, 'import pulumi\n'), ((514994, 515038), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""minExecutionSpeedBytes"""'}), "(name='minExecutionSpeedBytes')\n", (515007, 515038), False, 'import pulumi\n'), ((515446, 515491), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""minInsertBlockSizeBytes"""'}), "(name='minInsertBlockSizeBytes')\n", (515459, 515491), False, 'import pulumi\n'), ((515965, 516009), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""minInsertBlockSizeRows"""'}), "(name='minInsertBlockSizeRows')\n", (515978, 516009), False, 'import pulumi\n'), ((516477, 516533), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""outputFormatJsonQuote64bitIntegers"""'}), "(name='outputFormatJsonQuote64bitIntegers')\n", (516490, 516533), False, 'import pulumi\n'), ((517162, 517214), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""outputFormatJsonQuoteDenormals"""'}), "(name='outputFormatJsonQuoteDenormals')\n", (517175, 517214), False, 'import pulumi\n'), ((517993, 518024), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""quotaMode"""'}), "(name='quotaMode')\n", (518006, 518024), False, 'import pulumi\n'), ((518335, 518373), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""readOverflowMode"""'}), "(name='readOverflowMode')\n", (518348, 518373), False, 'import pulumi\n'), ((519128, 519164), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""receiveTimeout"""'}), "(name='receiveTimeout')\n", (519141, 519164), False, 'import pulumi\n'), ((519563, 519615), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""replicationAlterPartitionsSync"""'}), "(name='replicationAlterPartitionsSync')\n", (519576, 519615), False, 'import pulumi\n'), ((520137, 520177), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""resultOverflowMode"""'}), "(name='resultOverflowMode')\n", (520150, 520177), False, 'import pulumi\n'), ((520570, 520619), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""selectSequentialConsistency"""'}), "(name='selectSequentialConsistency')\n", (520583, 520619), False, 'import pulumi\n'), ((521067, 521114), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""sendProgressInHttpHeaders"""'}), "(name='sendProgressInHttpHeaders')\n", (521080, 521114), False, 'import pulumi\n'), ((521595, 521628), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""sendTimeout"""'}), "(name='sendTimeout')\n", (521608, 521628), False, 'import pulumi\n'), ((522009, 522046), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""setOverflowMode"""'}), "(name='setOverflowMode')\n", (522022, 522046), False, 'import pulumi\n'), ((522435, 522478), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""skipUnavailableShards"""'}), "(name='skipUnavailableShards')\n", (522448, 522478), False, 'import pulumi\n'), ((522894, 522932), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""sortOverflowMode"""'}), "(name='sortOverflowMode')\n", (522907, 522932), False, 'import pulumi\n'), ((523316, 523357), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""timeoutOverflowMode"""'}), "(name='timeoutOverflowMode')\n", (523329, 523357), False, 'import pulumi\n'), ((523745, 523787), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""transferOverflowMode"""'}), "(name='transferOverflowMode')\n", (523758, 523787), False, 'import pulumi\n'), ((524180, 524217), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""transformNullIn"""'}), "(name='transformNullIn')\n", (524193, 524217), False, 'import pulumi\n'), ((524591, 524633), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""useUncompressedCache"""'}), "(name='useUncompressedCache')\n", (524604, 524633), False, 'import pulumi\n'), ((526966, 526996), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""diskSize"""'}), "(name='diskSize')\n", (526979, 526996), False, 'import pulumi\n'), ((527346, 527378), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""diskTypeId"""'}), "(name='diskTypeId')\n", (527359, 527378), False, 'import pulumi\n'), ((527846, 527884), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""resourcePresetId"""'}), "(name='resourcePresetId')\n", (527859, 527884), False, 'import pulumi\n'), ((530048, 530083), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""adminPassword"""'}), "(name='adminPassword')\n", (530061, 530083), False, 'import pulumi\n'), ((530413, 530443), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""dataNode"""'}), "(name='dataNode')\n", (530426, 530443), False, 'import pulumi\n'), ((531324, 531356), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""masterNode"""'}), "(name='masterNode')\n", (531337, 531356), False, 'import pulumi\n'), ((534137, 534167), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""diskSize"""'}), "(name='diskSize')\n", (534150, 534167), False, 'import pulumi\n'), ((534487, 534519), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""diskTypeId"""'}), "(name='diskTypeId')\n", (534500, 534519), False, 'import pulumi\n'), ((534841, 534879), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""resourcePresetId"""'}), "(name='resourcePresetId')\n", (534854, 534879), False, 'import pulumi\n'), ((536790, 536820), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""diskSize"""'}), "(name='diskSize')\n", (536803, 536820), False, 'import pulumi\n'), ((537140, 537172), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""diskTypeId"""'}), "(name='diskTypeId')\n", (537153, 537172), False, 'import pulumi\n'), ((537494, 537532), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""resourcePresetId"""'}), "(name='resourcePresetId')\n", (537507, 537532), False, 'import pulumi\n'), ((540533, 540569), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""assignPublicIp"""'}), "(name='assignPublicIp')\n", (540546, 540569), False, 'import pulumi\n'), ((541304, 541334), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""subnetId"""'}), "(name='subnetId')\n", (541317, 541334), False, 'import pulumi\n'), ((542386, 542416), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""dataLens"""'}), "(name='dataLens')\n", (542399, 542416), False, 'import pulumi\n'), ((542781, 542809), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""webSql"""'}), "(name='webSql')\n", (542794, 542809), False, 'import pulumi\n'), ((545124, 545160), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""assignPublicIp"""'}), "(name='assignPublicIp')\n", (545137, 545160), False, 'import pulumi\n'), ((547359, 547389), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""diskSize"""'}), "(name='diskSize')\n", (547372, 547389), False, 'import pulumi\n'), ((547620, 547652), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""diskTypeId"""'}), "(name='diskTypeId')\n", (547633, 547652), False, 'import pulumi\n'), ((547898, 547936), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""resourcePresetId"""'}), "(name='resourcePresetId')\n", (547911, 547936), False, 'import pulumi\n'), ((550294, 550324), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""diskSize"""'}), "(name='diskSize')\n", (550307, 550324), False, 'import pulumi\n'), ((550555, 550587), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""diskTypeId"""'}), "(name='diskTypeId')\n", (550568, 550587), False, 'import pulumi\n'), ((550833, 550871), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""resourcePresetId"""'}), "(name='resourcePresetId')\n", (550846, 550871), False, 'import pulumi\n'), ((554475, 554511), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""assignPublicIp"""'}), "(name='assignPublicIp')\n", (554488, 554511), False, 'import pulumi\n'), ((554924, 554958), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""brokersCount"""'}), "(name='brokersCount')\n", (554937, 554958), False, 'import pulumi\n'), ((555321, 555357), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""schemaRegistry"""'}), "(name='schemaRegistry')\n", (555334, 555357), False, 'import pulumi\n'), ((555740, 555777), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""unmanagedTopics"""'}), "(name='unmanagedTopics')\n", (555753, 555777), False, 'import pulumi\n'), ((557854, 557887), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""kafkaConfig"""'}), "(name='kafkaConfig')\n", (557867, 557887), False, 'import pulumi\n'), ((561505, 561549), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""autoCreateTopicsEnable"""'}), "(name='autoCreateTopicsEnable')\n", (561518, 561549), False, 'import pulumi\n'), ((561882, 561919), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""compressionType"""'}), "(name='compressionType')\n", (561895, 561919), False, 'import pulumi\n'), ((562205, 562251), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""defaultReplicationFactor"""'}), "(name='defaultReplicationFactor')\n", (562218, 562251), False, 'import pulumi\n'), ((562587, 562633), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""logFlushIntervalMessages"""'}), "(name='logFlushIntervalMessages')\n", (562600, 562633), False, 'import pulumi\n'), ((562974, 563014), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""logFlushIntervalMs"""'}), "(name='logFlushIntervalMs')\n", (562987, 563014), False, 'import pulumi\n'), ((563325, 563374), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""logFlushSchedulerIntervalMs"""'}), "(name='logFlushSchedulerIntervalMs')\n", (563338, 563374), False, 'import pulumi\n'), ((563735, 563771), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""logPreallocate"""'}), "(name='logPreallocate')\n", (563748, 563771), False, 'import pulumi\n'), ((564054, 564093), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""logRetentionBytes"""'}), "(name='logRetentionBytes')\n", (564067, 564093), False, 'import pulumi\n'), ((564394, 564433), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""logRetentionHours"""'}), "(name='logRetentionHours')\n", (564407, 564433), False, 'import pulumi\n'), ((564734, 564775), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""logRetentionMinutes"""'}), "(name='logRetentionMinutes')\n", (564747, 564775), False, 'import pulumi\n'), ((565086, 565122), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""logRetentionMs"""'}), "(name='logRetentionMs')\n", (565099, 565122), False, 'import pulumi\n'), ((565408, 565445), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""logSegmentBytes"""'}), "(name='logSegmentBytes')\n", (565421, 565445), False, 'import pulumi\n'), ((565736, 565771), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""numPartitions"""'}), "(name='numPartitions')\n", (565749, 565771), False, 'import pulumi\n'), ((566047, 566093), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""socketReceiveBufferBytes"""'}), "(name='socketReceiveBufferBytes')\n", (566060, 566093), False, 'import pulumi\n'), ((566434, 566477), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""socketSendBufferBytes"""'}), "(name='socketSendBufferBytes')\n", (566447, 566477), False, 'import pulumi\n'), ((567592, 567622), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""diskSize"""'}), "(name='diskSize')\n", (567605, 567622), False, 'import pulumi\n'), ((567952, 567984), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""diskTypeId"""'}), "(name='diskTypeId')\n", (567965, 567984), False, 'import pulumi\n'), ((568427, 568465), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""resourcePresetId"""'}), "(name='resourcePresetId')\n", (568440, 568465), False, 'import pulumi\n'), ((570678, 570708), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""diskSize"""'}), "(name='diskSize')\n", (570691, 570708), False, 'import pulumi\n'), ((571058, 571090), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""diskTypeId"""'}), "(name='diskTypeId')\n", (571071, 571090), False, 'import pulumi\n'), ((571553, 571591), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""resourcePresetId"""'}), "(name='resourcePresetId')\n", (571566, 571591), False, 'import pulumi\n'), ((573428, 573464), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""assignPublicIp"""'}), "(name='assignPublicIp')\n", (573441, 573464), False, 'import pulumi\n'), ((574780, 574810), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""subnetId"""'}), "(name='subnetId')\n", (574793, 574810), False, 'import pulumi\n'), ((575142, 575170), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""zoneId"""'}), "(name='zoneId')\n", (575155, 575170), False, 'import pulumi\n'), ((577130, 577169), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""replicationFactor"""'}), "(name='replicationFactor')\n", (577143, 577169), False, 'import pulumi\n'), ((577540, 577573), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""topicConfig"""'}), "(name='topicConfig')\n", (577553, 577573), False, 'import pulumi\n'), ((580450, 580485), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""cleanupPolicy"""'}), "(name='cleanupPolicy')\n", (580463, 580485), False, 'import pulumi\n'), ((580761, 580798), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""compressionType"""'}), "(name='compressionType')\n", (580774, 580798), False, 'import pulumi\n'), ((581084, 581123), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""deleteRetentionMs"""'}), "(name='deleteRetentionMs')\n", (581097, 581123), False, 'import pulumi\n'), ((581424, 581463), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""fileDeleteDelayMs"""'}), "(name='fileDeleteDelayMs')\n", (581437, 581463), False, 'import pulumi\n'), ((581769, 581804), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""flushMessages"""'}), "(name='flushMessages')\n", (581782, 581804), False, 'import pulumi\n'), ((582080, 582109), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""flushMs"""'}), "(name='flushMs')\n", (582093, 582109), False, 'import pulumi\n'), ((582355, 582392), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxMessageBytes"""'}), "(name='maxMessageBytes')\n", (582368, 582392), False, 'import pulumi\n'), ((582683, 582723), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""minCompactionLagMs"""'}), "(name='minCompactionLagMs')\n", (582696, 582723), False, 'import pulumi\n'), ((583034, 583073), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""minInsyncReplicas"""'}), "(name='minInsyncReplicas')\n", (583047, 583073), False, 'import pulumi\n'), ((583650, 583686), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""retentionBytes"""'}), "(name='retentionBytes')\n", (583663, 583686), False, 'import pulumi\n'), ((583967, 584000), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""retentionMs"""'}), "(name='retentionMs')\n", (583980, 584000), False, 'import pulumi\n'), ((584266, 584300), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""segmentBytes"""'}), "(name='segmentBytes')\n", (584279, 584300), False, 'import pulumi\n'), ((587230, 587261), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""topicName"""'}), "(name='topicName')\n", (587243, 587261), False, 'import pulumi\n'), ((590021, 590056), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""cleanupPolicy"""'}), "(name='cleanupPolicy')\n", (590034, 590056), False, 'import pulumi\n'), ((590332, 590369), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""compressionType"""'}), "(name='compressionType')\n", (590345, 590369), False, 'import pulumi\n'), ((590655, 590694), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""deleteRetentionMs"""'}), "(name='deleteRetentionMs')\n", (590668, 590694), False, 'import pulumi\n'), ((590995, 591034), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""fileDeleteDelayMs"""'}), "(name='fileDeleteDelayMs')\n", (591008, 591034), False, 'import pulumi\n'), ((591340, 591375), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""flushMessages"""'}), "(name='flushMessages')\n", (591353, 591375), False, 'import pulumi\n'), ((591651, 591680), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""flushMs"""'}), "(name='flushMs')\n", (591664, 591680), False, 'import pulumi\n'), ((591926, 591963), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxMessageBytes"""'}), "(name='maxMessageBytes')\n", (591939, 591963), False, 'import pulumi\n'), ((592254, 592294), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""minCompactionLagMs"""'}), "(name='minCompactionLagMs')\n", (592267, 592294), False, 'import pulumi\n'), ((592605, 592644), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""minInsyncReplicas"""'}), "(name='minInsyncReplicas')\n", (592618, 592644), False, 'import pulumi\n'), ((593221, 593257), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""retentionBytes"""'}), "(name='retentionBytes')\n", (593234, 593257), False, 'import pulumi\n'), ((593538, 593571), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""retentionMs"""'}), "(name='retentionMs')\n", (593551, 593571), False, 'import pulumi\n'), ((593837, 593871), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""segmentBytes"""'}), "(name='segmentBytes')\n", (593850, 593871), False, 'import pulumi\n'), ((596404, 596443), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""backupWindowStart"""'}), "(name='backupWindowStart')\n", (596417, 596443), False, 'import pulumi\n'), ((596964, 597013), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""featureCompatibilityVersion"""'}), "(name='featureCompatibilityVersion')\n", (596977, 597013), False, 'import pulumi\n'), ((597869, 597899), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""dataLens"""'}), "(name='dataLens')\n", (597882, 597899), False, 'import pulumi\n'), ((602363, 602393), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""subnetId"""'}), "(name='subnetId')\n", (602376, 602393), False, 'import pulumi\n'), ((602784, 602812), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""zoneId"""'}), "(name='zoneId')\n", (602797, 602812), False, 'import pulumi\n'), ((603249, 603285), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""assignPublicIp"""'}), "(name='assignPublicIp')\n", (603262, 603285), False, 'import pulumi\n'), ((604674, 604705), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""shardName"""'}), "(name='shardName')\n", (604687, 604705), False, 'import pulumi\n'), ((608288, 608318), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""diskSize"""'}), "(name='diskSize')\n", (608301, 608318), False, 'import pulumi\n'), ((608646, 608678), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""diskTypeId"""'}), "(name='diskTypeId')\n", (608659, 608678), False, 'import pulumi\n'), ((609124, 609162), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""resourcePresetId"""'}), "(name='resourcePresetId')\n", (609137, 609162), False, 'import pulumi\n'), ((611913, 611947), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""databaseName"""'}), "(name='databaseName')\n", (611926, 611947), False, 'import pulumi\n'), ((613295, 613325), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""dataLens"""'}), "(name='dataLens')\n", (613308, 613325), False, 'import pulumi\n'), ((613690, 613718), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""webSql"""'}), "(name='webSql')\n", (613703, 613718), False, 'import pulumi\n'), ((618908, 618944), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""assignPublicIp"""'}), "(name='assignPublicIp')\n", (618921, 618944), False, 'import pulumi\n'), ((620177, 620216), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""replicationSource"""'}), "(name='replicationSource')\n", (620190, 620216), False, 'import pulumi\n'), ((620635, 620678), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""replicationSourceName"""'}), "(name='replicationSourceName')\n", (620648, 620678), False, 'import pulumi\n'), ((621193, 621223), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""subnetId"""'}), "(name='subnetId')\n", (621206, 621223), False, 'import pulumi\n'), ((624251, 624281), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""diskSize"""'}), "(name='diskSize')\n", (624264, 624281), False, 'import pulumi\n'), ((624607, 624639), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""diskTypeId"""'}), "(name='diskTypeId')\n", (624620, 624639), False, 'import pulumi\n'), ((624953, 624991), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""resourcePresetId"""'}), "(name='resourcePresetId')\n", (624966, 624991), False, 'import pulumi\n'), ((626017, 626047), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""backupId"""'}), "(name='backupId')\n", (626030, 626047), False, 'import pulumi\n'), ((630058, 630100), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""authenticationPlugin"""'}), "(name='authenticationPlugin')\n", (630071, 630100), False, 'import pulumi\n'), ((630611, 630649), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""connectionLimits"""'}), "(name='connectionLimits')\n", (630624, 630649), False, 'import pulumi\n'), ((631177, 631216), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""globalPermissions"""'}), "(name='globalPermissions')\n", (631190, 631216), False, 'import pulumi\n'), ((633566, 633609), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxConnectionsPerHour"""'}), "(name='maxConnectionsPerHour')\n", (633579, 633609), False, 'import pulumi\n'), ((633993, 634034), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxQuestionsPerHour"""'}), "(name='maxQuestionsPerHour')\n", (634006, 634034), False, 'import pulumi\n'), ((634406, 634445), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxUpdatesPerHour"""'}), "(name='maxUpdatesPerHour')\n", (634419, 634445), False, 'import pulumi\n'), ((634805, 634845), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxUserConnections"""'}), "(name='maxUserConnections')\n", (634818, 634845), False, 'import pulumi\n'), ((636055, 636089), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""databaseName"""'}), "(name='databaseName')\n", (636068, 636089), False, 'import pulumi\n'), ((640330, 640367), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxmemoryPolicy"""'}), "(name='maxmemoryPolicy')\n", (640343, 640367), False, 'import pulumi\n'), ((640917, 640959), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""notifyKeyspaceEvents"""'}), "(name='notifyKeyspaceEvents')\n", (640930, 640959), False, 'import pulumi\n'), ((641372, 641414), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""slowlogLogSlowerThan"""'}), "(name='slowlogLogSlowerThan')\n", (641385, 641414), False, 'import pulumi\n'), ((641819, 641854), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""slowlogMaxLen"""'}), "(name='slowlogMaxLen')\n", (641832, 641854), False, 'import pulumi\n'), ((644512, 644543), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""shardName"""'}), "(name='shardName')\n", (644525, 644543), False, 'import pulumi\n'), ((644880, 644910), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""subnetId"""'}), "(name='subnetId')\n", (644893, 644910), False, 'import pulumi\n'), ((648113, 648143), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""diskSize"""'}), "(name='diskSize')\n", (648126, 648143), False, 'import pulumi\n'), ((648463, 648501), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""resourcePresetId"""'}), "(name='resourcePresetId')\n", (648476, 648501), False, 'import pulumi\n'), ((648777, 648809), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""diskTypeId"""'}), "(name='diskTypeId')\n", (648790, 648809), False, 'import pulumi\n'), ((652482, 652518), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""assignPublicIp"""'}), "(name='assignPublicIp')\n", (652495, 652518), False, 'import pulumi\n'), ((653296, 653326), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""subnetId"""'}), "(name='subnetId')\n", (653309, 653326), False, 'import pulumi\n'), ((654379, 654409), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""diskSize"""'}), "(name='diskSize')\n", (654392, 654409), False, 'import pulumi\n'), ((654739, 654771), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""diskTypeId"""'}), "(name='diskTypeId')\n", (654752, 654771), False, 'import pulumi\n'), ((655089, 655127), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""resourcePresetId"""'}), "(name='resourcePresetId')\n", (655102, 655127), False, 'import pulumi\n'), ((658066, 658100), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""databaseName"""'}), "(name='databaseName')\n", (658079, 658100), False, 'import pulumi\n'), ((660570, 660606), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""allowedMethods"""'}), "(name='allowedMethods')\n", (660583, 660606), False, 'import pulumi\n'), ((661033, 661069), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""allowedOrigins"""'}), "(name='allowedOrigins')\n", (661046, 661069), False, 'import pulumi\n'), ((661447, 661483), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""allowedHeaders"""'}), "(name='allowedHeaders')\n", (661460, 661483), False, 'import pulumi\n'), ((661881, 661916), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""exposeHeaders"""'}), "(name='exposeHeaders')\n", (661894, 661916), False, 'import pulumi\n'), ((662313, 662348), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxAgeSeconds"""'}), "(name='maxAgeSeconds')\n", (662326, 662348), False, 'import pulumi\n'), ((667873, 667929), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""abortIncompleteMultipartUploadDays"""'}), "(name='abortIncompleteMultipartUploadDays')\n", (667886, 667929), False, 'import pulumi\n'), ((669257, 669306), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""noncurrentVersionExpiration"""'}), "(name='noncurrentVersionExpiration')\n", (669270, 669306), False, 'import pulumi\n'), ((669870, 669920), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""noncurrentVersionTransitions"""'}), "(name='noncurrentVersionTransitions')\n", (669883, 669920), False, 'import pulumi\n'), ((673258, 673305), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""expiredObjectDeleteMarker"""'}), "(name='expiredObjectDeleteMarker')\n", (673271, 673305), False, 'import pulumi\n'), ((675268, 675302), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""storageClass"""'}), "(name='storageClass')\n", (675281, 675302), False, 'import pulumi\n'), ((676970, 677004), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""storageClass"""'}), "(name='storageClass')\n", (676983, 677004), False, 'import pulumi\n'), ((678700, 678734), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""targetBucket"""'}), "(name='targetBucket')\n", (678713, 678734), False, 'import pulumi\n'), ((679075, 679109), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""targetPrefix"""'}), "(name='targetPrefix')\n", (679088, 679109), False, 'import pulumi\n'), ((681008, 681064), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""applyServerSideEncryptionByDefault"""'}), "(name='applyServerSideEncryptionByDefault')\n", (681021, 681064), False, 'import pulumi\n'), ((682373, 682409), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""kmsMasterKeyId"""'}), "(name='kmsMasterKeyId')\n", (682386, 682409), False, 'import pulumi\n'), ((682767, 682801), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""sseAlgorithm"""'}), "(name='sseAlgorithm')\n", (682780, 682801), False, 'import pulumi\n'), ((685696, 685731), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""errorDocument"""'}), "(name='errorDocument')\n", (685709, 685731), False, 'import pulumi\n'), ((686106, 686141), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""indexDocument"""'}), "(name='indexDocument')\n", (686119, 686141), False, 'import pulumi\n'), ((686553, 686596), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""redirectAllRequestsTo"""'}), "(name='redirectAllRequestsTo')\n", (686566, 686596), False, 'import pulumi\n'), ((687196, 687230), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""routingRules"""'}), "(name='routingRules')\n", (687209, 687230), False, 'import pulumi\n'), ((689189, 689233), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""ddosProtectionProvider"""'}), "(name='ddosProtectionProvider')\n", (689202, 689233), False, 'import pulumi\n'), ((689645, 689689), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""outgoingSmtpCapability"""'}), "(name='outgoingSmtpCapability')\n", (689658, 689689), False, 'import pulumi\n'), ((690080, 690108), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""zoneId"""'}), "(name='zoneId')\n", (690093, 690108), False, 'import pulumi\n'), ((693130, 693160), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""fromPort"""'}), "(name='fromPort')\n", (693143, 693160), False, 'import pulumi\n'), ((694315, 694353), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""predefinedTarget"""'}), "(name='predefinedTarget')\n", (694328, 694353), False, 'import pulumi\n'), ((694644, 694681), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""securityGroupId"""'}), "(name='securityGroupId')\n", (694657, 694681), False, 'import pulumi\n'), ((694972, 695000), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""toPort"""'}), "(name='toPort')\n", (694985, 695000), False, 'import pulumi\n'), ((695241, 695275), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""v4CidrBlocks"""'}), "(name='v4CidrBlocks')\n", (695254, 695275), False, 'import pulumi\n'), ((695599, 695633), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""v6CidrBlocks"""'}), "(name='v6CidrBlocks')\n", (695612, 695633), False, 'import pulumi\n'), ((698676, 698706), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""fromPort"""'}), "(name='fromPort')\n", (698689, 698706), False, 'import pulumi\n'), ((699861, 699899), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""predefinedTarget"""'}), "(name='predefinedTarget')\n", (699874, 699899), False, 'import pulumi\n'), ((700190, 700227), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""securityGroupId"""'}), "(name='securityGroupId')\n", (700203, 700227), False, 'import pulumi\n'), ((700518, 700546), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""toPort"""'}), "(name='toPort')\n", (700531, 700546), False, 'import pulumi\n'), ((700787, 700821), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""v4CidrBlocks"""'}), "(name='v4CidrBlocks')\n", (700800, 700821), False, 'import pulumi\n'), ((701145, 701179), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""v6CidrBlocks"""'}), "(name='v6CidrBlocks')\n", (701158, 701179), False, 'import pulumi\n'), ((702149, 702188), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""destinationPrefix"""'}), "(name='destinationPrefix')\n", (702162, 702188), False, 'import pulumi\n'), ((702547, 702583), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""nextHopAddress"""'}), "(name='nextHopAddress')\n", (702560, 702583), False, 'import pulumi\n'), ((705627, 705657), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""fromPort"""'}), "(name='fromPort')\n", (705640, 705657), False, 'import pulumi\n'), ((706802, 706840), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""predefinedTarget"""'}), "(name='predefinedTarget')\n", (706815, 706840), False, 'import pulumi\n'), ((707131, 707168), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""securityGroupId"""'}), "(name='securityGroupId')\n", (707144, 707168), False, 'import pulumi\n'), ((707459, 707487), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""toPort"""'}), "(name='toPort')\n", (707472, 707487), False, 'import pulumi\n'), ((707728, 707762), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""v4CidrBlocks"""'}), "(name='v4CidrBlocks')\n", (707741, 707762), False, 'import pulumi\n'), ((708086, 708120), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""v6CidrBlocks"""'}), "(name='v6CidrBlocks')\n", (708099, 708120), False, 'import pulumi\n'), ((711146, 711176), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""fromPort"""'}), "(name='fromPort')\n", (711159, 711176), False, 'import pulumi\n'), ((712321, 712359), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""predefinedTarget"""'}), "(name='predefinedTarget')\n", (712334, 712359), False, 'import pulumi\n'), ((712650, 712687), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""securityGroupId"""'}), "(name='securityGroupId')\n", (712663, 712687), False, 'import pulumi\n'), ((712978, 713006), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""toPort"""'}), "(name='toPort')\n", (712991, 713006), False, 'import pulumi\n'), ((713247, 713281), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""v4CidrBlocks"""'}), "(name='v4CidrBlocks')\n", (713260, 713281), False, 'import pulumi\n'), ((713605, 713639), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""v6CidrBlocks"""'}), "(name='v6CidrBlocks')\n", (713618, 713639), False, 'import pulumi\n'), ((714905, 714937), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""domainName"""'}), "(name='domainName')\n", (714918, 714937), False, 'import pulumi\n'), ((715243, 715282), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""domainNameServers"""'}), "(name='domainNameServers')\n", (715256, 715282), False, 'import pulumi\n'), ((715696, 715728), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""ntpServers"""'}), "(name='ntpServers')\n", (715709, 715728), False, 'import pulumi\n'), ((718024, 718056), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""fixedScale"""'}), "(name='fixedScale')\n", (718037, 718056), False, 'import pulumi\n'), ((719778, 719810), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""groupCount"""'}), "(name='groupCount')\n", (719791, 719810), False, 'import pulumi\n'), ((720158, 720193), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""storageTypeId"""'}), "(name='storageTypeId')\n", (720171, 720193), False, 'import pulumi\n'), ((722759, 722800), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""loadBalancingConfig"""'}), "(name='loadBalancingConfig')\n", (722772, 722800), False, 'import pulumi\n'), ((723801, 723837), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""targetGroupIds"""'}), "(name='targetGroupIds')\n", (723814, 723837), False, 'import pulumi\n'), ((727661, 727698), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""grpcHealthcheck"""'}), "(name='grpcHealthcheck')\n", (727674, 727698), False, 'import pulumi\n'), ((728185, 728222), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""healthcheckPort"""'}), "(name='healthcheckPort')\n", (728198, 728222), False, 'import pulumi\n'), ((728539, 728577), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""healthyThreshold"""'}), "(name='healthyThreshold')\n", (728552, 728577), False, 'import pulumi\n'), ((729063, 729100), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""httpHealthcheck"""'}), "(name='httpHealthcheck')\n", (729076, 729100), False, 'import pulumi\n'), ((729862, 729905), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""intervalJitterPercent"""'}), "(name='intervalJitterPercent')\n", (729875, 729905), False, 'import pulumi\n'), ((730392, 730431), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""streamHealthcheck"""'}), "(name='streamHealthcheck')\n", (730405, 730431), False, 'import pulumi\n'), ((731214, 731254), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""unhealthyThreshold"""'}), "(name='unhealthyThreshold')\n", (731227, 731254), False, 'import pulumi\n'), ((732072, 732105), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""serviceName"""'}), "(name='serviceName')\n", (732085, 732105), False, 'import pulumi\n'), ((735830, 735879), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""localityAwareRoutingPercent"""'}), "(name='localityAwareRoutingPercent')\n", (735843, 735879), False, 'import pulumi\n'), ((736334, 736370), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""panicThreshold"""'}), "(name='panicThreshold')\n", (736347, 736370), False, 'import pulumi\n'), ((736893, 736929), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""strictLocality"""'}), "(name='strictLocality')\n", (736906, 736929), False, 'import pulumi\n'), ((738491, 738530), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""validationContext"""'}), "(name='validationContext')\n", (738504, 738530), False, 'import pulumi\n'), ((739200, 739236), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""trustedCaBytes"""'}), "(name='trustedCaBytes')\n", (739213, 739236), False, 'import pulumi\n'), ((739474, 739507), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""trustedCaId"""'}), "(name='trustedCaId')\n", (739487, 739507), False, 'import pulumi\n'), ((742280, 742321), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""loadBalancingConfig"""'}), "(name='loadBalancingConfig')\n", (742293, 742321), False, 'import pulumi\n'), ((743322, 743358), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""targetGroupIds"""'}), "(name='targetGroupIds')\n", (743335, 743358), False, 'import pulumi\n'), ((747182, 747219), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""grpcHealthcheck"""'}), "(name='grpcHealthcheck')\n", (747195, 747219), False, 'import pulumi\n'), ((747706, 747743), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""healthcheckPort"""'}), "(name='healthcheckPort')\n", (747719, 747743), False, 'import pulumi\n'), ((748060, 748098), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""healthyThreshold"""'}), "(name='healthyThreshold')\n", (748073, 748098), False, 'import pulumi\n'), ((748584, 748621), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""httpHealthcheck"""'}), "(name='httpHealthcheck')\n", (748597, 748621), False, 'import pulumi\n'), ((749383, 749426), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""intervalJitterPercent"""'}), "(name='intervalJitterPercent')\n", (749396, 749426), False, 'import pulumi\n'), ((749913, 749952), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""streamHealthcheck"""'}), "(name='streamHealthcheck')\n", (749926, 749952), False, 'import pulumi\n'), ((750735, 750775), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""unhealthyThreshold"""'}), "(name='unhealthyThreshold')\n", (750748, 750775), False, 'import pulumi\n'), ((751593, 751626), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""serviceName"""'}), "(name='serviceName')\n", (751606, 751626), False, 'import pulumi\n'), ((755351, 755400), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""localityAwareRoutingPercent"""'}), "(name='localityAwareRoutingPercent')\n", (755364, 755400), False, 'import pulumi\n'), ((755855, 755891), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""panicThreshold"""'}), "(name='panicThreshold')\n", (755868, 755891), False, 'import pulumi\n'), ((756414, 756450), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""strictLocality"""'}), "(name='strictLocality')\n", (756427, 756450), False, 'import pulumi\n'), ((758012, 758051), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""validationContext"""'}), "(name='validationContext')\n", (758025, 758051), False, 'import pulumi\n'), ((758721, 758757), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""trustedCaBytes"""'}), "(name='trustedCaBytes')\n", (758734, 758757), False, 'import pulumi\n'), ((758995, 759028), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""trustedCaId"""'}), "(name='trustedCaId')\n", (759008, 759028), False, 'import pulumi\n'), ((759476, 759518), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""diskPlacementGroupId"""'}), "(name='diskPlacementGroupId')\n", (759489, 759518), False, 'import pulumi\n'), ((760001, 760039), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""placementGroupId"""'}), "(name='placementGroupId')\n", (760014, 760039), False, 'import pulumi\n'), ((761730, 761770), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""zoneInstancesLimit"""'}), "(name='zoneInstancesLimit')\n", (761743, 761770), False, 'import pulumi\n'), ((762048, 762087), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""zoneRequestsLimit"""'}), "(name='zoneRequestsLimit')\n", (762061, 762087), False, 'import pulumi\n'), ((768194, 768230), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""assignPublicIp"""'}), "(name='assignPublicIp')\n", (768207, 768230), False, 'import pulumi\n'), ((768597, 768631), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""brokersCount"""'}), "(name='brokersCount')\n", (768610, 768631), False, 'import pulumi\n'), ((768957, 768993), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""schemaRegistry"""'}), "(name='schemaRegistry')\n", (768970, 768993), False, 'import pulumi\n'), ((769368, 769405), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""unmanagedTopics"""'}), "(name='unmanagedTopics')\n", (769381, 769405), False, 'import pulumi\n'), ((771004, 771037), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""kafkaConfig"""'}), "(name='kafkaConfig')\n", (771017, 771037), False, 'import pulumi\n'), ((774437, 774481), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""autoCreateTopicsEnable"""'}), "(name='autoCreateTopicsEnable')\n", (774450, 774481), False, 'import pulumi\n'), ((774786, 774823), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""compressionType"""'}), "(name='compressionType')\n", (774799, 774823), False, 'import pulumi\n'), ((775081, 775127), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""defaultReplicationFactor"""'}), "(name='defaultReplicationFactor')\n", (775094, 775127), False, 'import pulumi\n'), ((775435, 775481), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""logFlushIntervalMessages"""'}), "(name='logFlushIntervalMessages')\n", (775448, 775481), False, 'import pulumi\n'), ((775794, 775834), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""logFlushIntervalMs"""'}), "(name='logFlushIntervalMs')\n", (775807, 775834), False, 'import pulumi\n'), ((776117, 776166), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""logFlushSchedulerIntervalMs"""'}), "(name='logFlushSchedulerIntervalMs')\n", (776130, 776166), False, 'import pulumi\n'), ((776499, 776535), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""logPreallocate"""'}), "(name='logPreallocate')\n", (776512, 776535), False, 'import pulumi\n'), ((776790, 776829), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""logRetentionBytes"""'}), "(name='logRetentionBytes')\n", (776803, 776829), False, 'import pulumi\n'), ((777102, 777141), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""logRetentionHours"""'}), "(name='logRetentionHours')\n", (777115, 777141), False, 'import pulumi\n'), ((777414, 777455), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""logRetentionMinutes"""'}), "(name='logRetentionMinutes')\n", (777427, 777455), False, 'import pulumi\n'), ((777738, 777774), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""logRetentionMs"""'}), "(name='logRetentionMs')\n", (777751, 777774), False, 'import pulumi\n'), ((778032, 778069), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""logSegmentBytes"""'}), "(name='logSegmentBytes')\n", (778045, 778069), False, 'import pulumi\n'), ((778332, 778367), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""numPartitions"""'}), "(name='numPartitions')\n", (778345, 778367), False, 'import pulumi\n'), ((778615, 778661), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""socketReceiveBufferBytes"""'}), "(name='socketReceiveBufferBytes')\n", (778628, 778661), False, 'import pulumi\n'), ((778974, 779017), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""socketSendBufferBytes"""'}), "(name='socketSendBufferBytes')\n", (778987, 779017), False, 'import pulumi\n'), ((780059, 780089), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""diskSize"""'}), "(name='diskSize')\n", (780072, 780089), False, 'import pulumi\n'), ((780402, 780434), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""diskTypeId"""'}), "(name='diskTypeId')\n", (780415, 780434), False, 'import pulumi\n'), ((780860, 780898), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""resourcePresetId"""'}), "(name='resourcePresetId')\n", (780873, 780898), False, 'import pulumi\n'), ((782767, 782797), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""diskSize"""'}), "(name='diskSize')\n", (782780, 782797), False, 'import pulumi\n'), ((783110, 783142), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""diskTypeId"""'}), "(name='diskTypeId')\n", (783123, 783142), False, 'import pulumi\n'), ((783568, 783606), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""resourcePresetId"""'}), "(name='resourcePresetId')\n", (783581, 783606), False, 'import pulumi\n'), ((784979, 785010), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""clusterId"""'}), "(name='clusterId')\n", (784992, 785010), False, 'import pulumi\n'), ((785835, 785874), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""replicationFactor"""'}), "(name='replicationFactor')\n", (785848, 785874), False, 'import pulumi\n'), ((786228, 786261), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""topicConfig"""'}), "(name='topicConfig')\n", (786241, 786261), False, 'import pulumi\n'), ((788948, 788983), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""cleanupPolicy"""'}), "(name='cleanupPolicy')\n", (788961, 788983), False, 'import pulumi\n'), ((789231, 789268), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""compressionType"""'}), "(name='compressionType')\n", (789244, 789268), False, 'import pulumi\n'), ((789526, 789565), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""deleteRetentionMs"""'}), "(name='deleteRetentionMs')\n", (789539, 789565), False, 'import pulumi\n'), ((789838, 789877), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""fileDeleteDelayMs"""'}), "(name='fileDeleteDelayMs')\n", (789851, 789877), False, 'import pulumi\n'), ((790155, 790190), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""flushMessages"""'}), "(name='flushMessages')\n", (790168, 790190), False, 'import pulumi\n'), ((790438, 790467), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""flushMs"""'}), "(name='flushMs')\n", (790451, 790467), False, 'import pulumi\n'), ((790685, 790722), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""maxMessageBytes"""'}), "(name='maxMessageBytes')\n", (790698, 790722), False, 'import pulumi\n'), ((790985, 791025), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""minCompactionLagMs"""'}), "(name='minCompactionLagMs')\n", (790998, 791025), False, 'import pulumi\n'), ((791308, 791347), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""minInsyncReplicas"""'}), "(name='minInsyncReplicas')\n", (791321, 791347), False, 'import pulumi\n'), ((791868, 791904), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""retentionBytes"""'}), "(name='retentionBytes')\n", (791881, 791904), False, 'import pulumi\n'), ((792157, 792190), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""retentionMs"""'}), "(name='retentionMs')\n", (792170, 792190), False, 'import pulumi\n'), ((792428, 792462), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""segmentBytes"""'}), "(name='segmentBytes')\n", (792441, 792462), False, 'import pulumi\n'), ((795132, 795163), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""topicName"""'}), "(name='topicName')\n", (795145, 795163), False, 'import pulumi\n'), ((796023, 796053), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""dataLens"""'}), "(name='dataLens')\n", (796036, 796053), False, 'import pulumi\n'), ((796370, 796398), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""webSql"""'}), "(name='webSql')\n", (796383, 796398), False, 'import pulumi\n'), ((15548, 15582), 'pulumi.set', 'pulumi.set', (['__self__', '"""name"""', 'name'], {}), "(__self__, 'name', name)\n", (15558, 15582), False, 'import pulumi\n'), ((15591, 15649), 'pulumi.set', 'pulumi.set', (['__self__', '"""target_group_ids"""', 'target_group_ids'], {}), "(__self__, 'target_group_ids', target_group_ids)\n", (15601, 15649), False, 'import pulumi\n'), ((16248, 16272), 'pulumi.get', 'pulumi.get', (['self', '"""name"""'], {}), "(self, 'name')\n", (16258, 16272), False, 'import pulumi\n'), ((16345, 16376), 'pulumi.set', 'pulumi.set', (['self', '"""name"""', 'value'], {}), "(self, 'name', value)\n", (16355, 16376), False, 'import pulumi\n'), ((16600, 16636), 'pulumi.get', 'pulumi.get', (['self', '"""target_group_ids"""'], {}), "(self, 'target_group_ids')\n", (16610, 16636), False, 'import pulumi\n'), ((16757, 16800), 'pulumi.set', 'pulumi.set', (['self', '"""target_group_ids"""', 'value'], {}), "(self, 'target_group_ids', value)\n", (16767, 16800), False, 'import pulumi\n'), ((17072, 17103), 'pulumi.get', 'pulumi.get', (['self', '"""healthcheck"""'], {}), "(self, 'healthcheck')\n", (17082, 17103), False, 'import pulumi\n'), ((17240, 17278), 'pulumi.set', 'pulumi.set', (['self', '"""healthcheck"""', 'value'], {}), "(self, 'healthcheck', value)\n", (17250, 17278), False, 'import pulumi\n'), ((17606, 17647), 'pulumi.get', 'pulumi.get', (['self', '"""load_balancing_config"""'], {}), "(self, 'load_balancing_config')\n", (17616, 17647), False, 'import pulumi\n'), ((17812, 17860), 'pulumi.set', 'pulumi.set', (['self', '"""load_balancing_config"""', 'value'], {}), "(self, 'load_balancing_config', value)\n", (17822, 17860), False, 'import pulumi\n'), ((18020, 18044), 'pulumi.get', 'pulumi.get', (['self', '"""port"""'], {}), "(self, 'port')\n", (18030, 18044), False, 'import pulumi\n'), ((18127, 18158), 'pulumi.set', 'pulumi.set', (['self', '"""port"""', 'value'], {}), "(self, 'port', value)\n", (18137, 18158), False, 'import pulumi\n'), ((18406, 18429), 'pulumi.get', 'pulumi.get', (['self', '"""tls"""'], {}), "(self, 'tls')\n", (18416, 18429), False, 'import pulumi\n'), ((18542, 18572), 'pulumi.set', 'pulumi.set', (['self', '"""tls"""', 'value'], {}), "(self, 'tls', value)\n", (18552, 18572), False, 'import pulumi\n'), ((18822, 18848), 'pulumi.get', 'pulumi.get', (['self', '"""weight"""'], {}), "(self, 'weight')\n", (18832, 18848), False, 'import pulumi\n'), ((18935, 18968), 'pulumi.set', 'pulumi.set', (['self', '"""weight"""', 'value'], {}), "(self, 'weight', value)\n", (18945, 18968), False, 'import pulumi\n'), ((21504, 21546), 'pulumi.set', 'pulumi.set', (['__self__', '"""interval"""', 'interval'], {}), "(__self__, 'interval', interval)\n", (21514, 21546), False, 'import pulumi\n'), ((21555, 21595), 'pulumi.set', 'pulumi.set', (['__self__', '"""timeout"""', 'timeout'], {}), "(__self__, 'timeout', timeout)\n", (21565, 21595), False, 'import pulumi\n'), ((22577, 22605), 'pulumi.get', 'pulumi.get', (['self', '"""interval"""'], {}), "(self, 'interval')\n", (22587, 22605), False, 'import pulumi\n'), ((22686, 22721), 'pulumi.set', 'pulumi.set', (['self', '"""interval"""', 'value'], {}), "(self, 'interval', value)\n", (22696, 22721), False, 'import pulumi\n'), ((22889, 22916), 'pulumi.get', 'pulumi.get', (['self', '"""timeout"""'], {}), "(self, 'timeout')\n", (22899, 22916), False, 'import pulumi\n'), ((22995, 23029), 'pulumi.set', 'pulumi.set', (['self', '"""timeout"""', 'value'], {}), "(self, 'timeout', value)\n", (23005, 23029), False, 'import pulumi\n'), ((23354, 23390), 'pulumi.get', 'pulumi.get', (['self', '"""grpc_healthcheck"""'], {}), "(self, 'grpc_healthcheck')\n", (23364, 23390), False, 'import pulumi\n'), ((23552, 23595), 'pulumi.set', 'pulumi.set', (['self', '"""grpc_healthcheck"""', 'value'], {}), "(self, 'grpc_healthcheck', value)\n", (23562, 23595), False, 'import pulumi\n'), ((23811, 23847), 'pulumi.get', 'pulumi.get', (['self', '"""healthcheck_port"""'], {}), "(self, 'healthcheck_port')\n", (23821, 23847), False, 'import pulumi\n'), ((23954, 23997), 'pulumi.set', 'pulumi.set', (['self', '"""healthcheck_port"""', 'value'], {}), "(self, 'healthcheck_port', value)\n", (23964, 23997), False, 'import pulumi\n'), ((24379, 24416), 'pulumi.get', 'pulumi.get', (['self', '"""healthy_threshold"""'], {}), "(self, 'healthy_threshold')\n", (24389, 24416), False, 'import pulumi\n'), ((24525, 24569), 'pulumi.set', 'pulumi.set', (['self', '"""healthy_threshold"""', 'value'], {}), "(self, 'healthy_threshold', value)\n", (24535, 24569), False, 'import pulumi\n'), ((24894, 24930), 'pulumi.get', 'pulumi.get', (['self', '"""http_healthcheck"""'], {}), "(self, 'http_healthcheck')\n", (24904, 24930), False, 'import pulumi\n'), ((25092, 25135), 'pulumi.set', 'pulumi.set', (['self', '"""http_healthcheck"""', 'value'], {}), "(self, 'http_healthcheck', value)\n", (25102, 25135), False, 'import pulumi\n'), ((25497, 25540), 'pulumi.get', 'pulumi.get', (['self', '"""interval_jitter_percent"""'], {}), "(self, 'interval_jitter_percent')\n", (25507, 25540), False, 'import pulumi\n'), ((25663, 25713), 'pulumi.set', 'pulumi.set', (['self', '"""interval_jitter_percent"""', 'value'], {}), "(self, 'interval_jitter_percent', value)\n", (25673, 25713), False, 'import pulumi\n'), ((26046, 26084), 'pulumi.get', 'pulumi.get', (['self', '"""stream_healthcheck"""'], {}), "(self, 'stream_healthcheck')\n", (26056, 26084), False, 'import pulumi\n'), ((26252, 26297), 'pulumi.set', 'pulumi.set', (['self', '"""stream_healthcheck"""', 'value'], {}), "(self, 'stream_healthcheck', value)\n", (26262, 26297), False, 'import pulumi\n'), ((26665, 26704), 'pulumi.get', 'pulumi.get', (['self', '"""unhealthy_threshold"""'], {}), "(self, 'unhealthy_threshold')\n", (26675, 26704), False, 'import pulumi\n'), ((26817, 26863), 'pulumi.set', 'pulumi.set', (['self', '"""unhealthy_threshold"""', 'value'], {}), "(self, 'unhealthy_threshold', value)\n", (26827, 26863), False, 'import pulumi\n'), ((27498, 27530), 'pulumi.get', 'pulumi.get', (['self', '"""service_name"""'], {}), "(self, 'service_name')\n", (27508, 27530), False, 'import pulumi\n'), ((27629, 27668), 'pulumi.set', 'pulumi.set', (['self', '"""service_name"""', 'value'], {}), "(self, 'service_name', value)\n", (27639, 27668), False, 'import pulumi\n'), ((28173, 28207), 'pulumi.set', 'pulumi.set', (['__self__', '"""path"""', 'path'], {}), "(__self__, 'path', path)\n", (28183, 28207), False, 'import pulumi\n'), ((28496, 28520), 'pulumi.get', 'pulumi.get', (['self', '"""path"""'], {}), "(self, 'path')\n", (28506, 28520), False, 'import pulumi\n'), ((28593, 28624), 'pulumi.set', 'pulumi.set', (['self', '"""path"""', 'value'], {}), "(self, 'path', value)\n", (28603, 28624), False, 'import pulumi\n'), ((28783, 28807), 'pulumi.get', 'pulumi.get', (['self', '"""host"""'], {}), "(self, 'host')\n", (28793, 28807), False, 'import pulumi\n'), ((28890, 28921), 'pulumi.set', 'pulumi.set', (['self', '"""host"""', 'value'], {}), "(self, 'host', value)\n", (28900, 28921), False, 'import pulumi\n'), ((29094, 29119), 'pulumi.get', 'pulumi.get', (['self', '"""http2"""'], {}), "(self, 'http2')\n", (29104, 29119), False, 'import pulumi\n'), ((29205, 29237), 'pulumi.set', 'pulumi.set', (['self', '"""http2"""', 'value'], {}), "(self, 'http2', value)\n", (29215, 29237), False, 'import pulumi\n'), ((29989, 30016), 'pulumi.get', 'pulumi.get', (['self', '"""receive"""'], {}), "(self, 'receive')\n", (29999, 30016), False, 'import pulumi\n'), ((30105, 30139), 'pulumi.set', 'pulumi.set', (['self', '"""receive"""', 'value'], {}), "(self, 'receive', value)\n", (30115, 30139), False, 'import pulumi\n'), ((30333, 30357), 'pulumi.get', 'pulumi.get', (['self', '"""send"""'], {}), "(self, 'send')\n", (30343, 30357), False, 'import pulumi\n'), ((30440, 30471), 'pulumi.set', 'pulumi.set', (['self', '"""send"""', 'value'], {}), "(self, 'send', value)\n", (30450, 30471), False, 'import pulumi\n'), ((32167, 32217), 'pulumi.get', 'pulumi.get', (['self', '"""locality_aware_routing_percent"""'], {}), "(self, 'locality_aware_routing_percent')\n", (32177, 32217), False, 'import pulumi\n'), ((32352, 32409), 'pulumi.set', 'pulumi.set', (['self', '"""locality_aware_routing_percent"""', 'value'], {}), "(self, 'locality_aware_routing_percent', value)\n", (32362, 32409), False, 'import pulumi\n'), ((32834, 32869), 'pulumi.get', 'pulumi.get', (['self', '"""panic_threshold"""'], {}), "(self, 'panic_threshold')\n", (32844, 32869), False, 'import pulumi\n'), ((32974, 33016), 'pulumi.set', 'pulumi.set', (['self', '"""panic_threshold"""', 'value'], {}), "(self, 'panic_threshold', value)\n", (32984, 33016), False, 'import pulumi\n'), ((33300, 33335), 'pulumi.get', 'pulumi.get', (['self', '"""strict_locality"""'], {}), "(self, 'strict_locality')\n", (33310, 33335), False, 'import pulumi\n'), ((33441, 33483), 'pulumi.set', 'pulumi.set', (['self', '"""strict_locality"""', 'value'], {}), "(self, 'strict_locality', value)\n", (33451, 33483), False, 'import pulumi\n'), ((34769, 34792), 'pulumi.get', 'pulumi.get', (['self', '"""sni"""'], {}), "(self, 'sni')\n", (34779, 34792), False, 'import pulumi\n'), ((34873, 34903), 'pulumi.set', 'pulumi.set', (['self', '"""sni"""', 'value'], {}), "(self, 'sni', value)\n", (34883, 34903), False, 'import pulumi\n'), ((35093, 35131), 'pulumi.get', 'pulumi.get', (['self', '"""validation_context"""'], {}), "(self, 'validation_context')\n", (35103, 35131), False, 'import pulumi\n'), ((35291, 35336), 'pulumi.set', 'pulumi.set', (['self', '"""validation_context"""', 'value'], {}), "(self, 'validation_context', value)\n", (35301, 35336), False, 'import pulumi\n'), ((35936, 35972), 'pulumi.get', 'pulumi.get', (['self', '"""trusted_ca_bytes"""'], {}), "(self, 'trusted_ca_bytes')\n", (35946, 35972), False, 'import pulumi\n'), ((36079, 36122), 'pulumi.set', 'pulumi.set', (['self', '"""trusted_ca_bytes"""', 'value'], {}), "(self, 'trusted_ca_bytes', value)\n", (36089, 36122), False, 'import pulumi\n'), ((36252, 36285), 'pulumi.get', 'pulumi.get', (['self', '"""trusted_ca_id"""'], {}), "(self, 'trusted_ca_id')\n", (36262, 36285), False, 'import pulumi\n'), ((36386, 36426), 'pulumi.set', 'pulumi.set', (['self', '"""trusted_ca_id"""', 'value'], {}), "(self, 'trusted_ca_id', value)\n", (36396, 36426), False, 'import pulumi\n'), ((38186, 38220), 'pulumi.set', 'pulumi.set', (['__self__', '"""name"""', 'name'], {}), "(__self__, 'name', name)\n", (38196, 38220), False, 'import pulumi\n'), ((38229, 38287), 'pulumi.set', 'pulumi.set', (['__self__', '"""target_group_ids"""', 'target_group_ids'], {}), "(__self__, 'target_group_ids', target_group_ids)\n", (38239, 38287), False, 'import pulumi\n'), ((38965, 38989), 'pulumi.get', 'pulumi.get', (['self', '"""name"""'], {}), "(self, 'name')\n", (38975, 38989), False, 'import pulumi\n'), ((39062, 39093), 'pulumi.set', 'pulumi.set', (['self', '"""name"""', 'value'], {}), "(self, 'name', value)\n", (39072, 39093), False, 'import pulumi\n'), ((39317, 39353), 'pulumi.get', 'pulumi.get', (['self', '"""target_group_ids"""'], {}), "(self, 'target_group_ids')\n", (39327, 39353), False, 'import pulumi\n'), ((39474, 39517), 'pulumi.set', 'pulumi.set', (['self', '"""target_group_ids"""', 'value'], {}), "(self, 'target_group_ids', value)\n", (39484, 39517), False, 'import pulumi\n'), ((39789, 39820), 'pulumi.get', 'pulumi.get', (['self', '"""healthcheck"""'], {}), "(self, 'healthcheck')\n", (39799, 39820), False, 'import pulumi\n'), ((39957, 39995), 'pulumi.set', 'pulumi.set', (['self', '"""healthcheck"""', 'value'], {}), "(self, 'healthcheck', value)\n", (39967, 39995), False, 'import pulumi\n'), ((40168, 40193), 'pulumi.get', 'pulumi.get', (['self', '"""http2"""'], {}), "(self, 'http2')\n", (40178, 40193), False, 'import pulumi\n'), ((40279, 40311), 'pulumi.set', 'pulumi.set', (['self', '"""http2"""', 'value'], {}), "(self, 'http2', value)\n", (40289, 40311), False, 'import pulumi\n'), ((40639, 40680), 'pulumi.get', 'pulumi.get', (['self', '"""load_balancing_config"""'], {}), "(self, 'load_balancing_config')\n", (40649, 40680), False, 'import pulumi\n'), ((40845, 40893), 'pulumi.set', 'pulumi.set', (['self', '"""load_balancing_config"""', 'value'], {}), "(self, 'load_balancing_config', value)\n", (40855, 40893), False, 'import pulumi\n'), ((41053, 41077), 'pulumi.get', 'pulumi.get', (['self', '"""port"""'], {}), "(self, 'port')\n", (41063, 41077), False, 'import pulumi\n'), ((41160, 41191), 'pulumi.set', 'pulumi.set', (['self', '"""port"""', 'value'], {}), "(self, 'port', value)\n", (41170, 41191), False, 'import pulumi\n'), ((41439, 41462), 'pulumi.get', 'pulumi.get', (['self', '"""tls"""'], {}), "(self, 'tls')\n", (41449, 41462), False, 'import pulumi\n'), ((41575, 41605), 'pulumi.set', 'pulumi.set', (['self', '"""tls"""', 'value'], {}), "(self, 'tls', value)\n", (41585, 41605), False, 'import pulumi\n'), ((41855, 41881), 'pulumi.get', 'pulumi.get', (['self', '"""weight"""'], {}), "(self, 'weight')\n", (41865, 41881), False, 'import pulumi\n'), ((41968, 42001), 'pulumi.set', 'pulumi.set', (['self', '"""weight"""', 'value'], {}), "(self, 'weight', value)\n", (41978, 42001), False, 'import pulumi\n'), ((44537, 44579), 'pulumi.set', 'pulumi.set', (['__self__', '"""interval"""', 'interval'], {}), "(__self__, 'interval', interval)\n", (44547, 44579), False, 'import pulumi\n'), ((44588, 44628), 'pulumi.set', 'pulumi.set', (['__self__', '"""timeout"""', 'timeout'], {}), "(__self__, 'timeout', timeout)\n", (44598, 44628), False, 'import pulumi\n'), ((45610, 45638), 'pulumi.get', 'pulumi.get', (['self', '"""interval"""'], {}), "(self, 'interval')\n", (45620, 45638), False, 'import pulumi\n'), ((45719, 45754), 'pulumi.set', 'pulumi.set', (['self', '"""interval"""', 'value'], {}), "(self, 'interval', value)\n", (45729, 45754), False, 'import pulumi\n'), ((45922, 45949), 'pulumi.get', 'pulumi.get', (['self', '"""timeout"""'], {}), "(self, 'timeout')\n", (45932, 45949), False, 'import pulumi\n'), ((46028, 46062), 'pulumi.set', 'pulumi.set', (['self', '"""timeout"""', 'value'], {}), "(self, 'timeout', value)\n", (46038, 46062), False, 'import pulumi\n'), ((46387, 46423), 'pulumi.get', 'pulumi.get', (['self', '"""grpc_healthcheck"""'], {}), "(self, 'grpc_healthcheck')\n", (46397, 46423), False, 'import pulumi\n'), ((46585, 46628), 'pulumi.set', 'pulumi.set', (['self', '"""grpc_healthcheck"""', 'value'], {}), "(self, 'grpc_healthcheck', value)\n", (46595, 46628), False, 'import pulumi\n'), ((46844, 46880), 'pulumi.get', 'pulumi.get', (['self', '"""healthcheck_port"""'], {}), "(self, 'healthcheck_port')\n", (46854, 46880), False, 'import pulumi\n'), ((46987, 47030), 'pulumi.set', 'pulumi.set', (['self', '"""healthcheck_port"""', 'value'], {}), "(self, 'healthcheck_port', value)\n", (46997, 47030), False, 'import pulumi\n'), ((47412, 47449), 'pulumi.get', 'pulumi.get', (['self', '"""healthy_threshold"""'], {}), "(self, 'healthy_threshold')\n", (47422, 47449), False, 'import pulumi\n'), ((47558, 47602), 'pulumi.set', 'pulumi.set', (['self', '"""healthy_threshold"""', 'value'], {}), "(self, 'healthy_threshold', value)\n", (47568, 47602), False, 'import pulumi\n'), ((47927, 47963), 'pulumi.get', 'pulumi.get', (['self', '"""http_healthcheck"""'], {}), "(self, 'http_healthcheck')\n", (47937, 47963), False, 'import pulumi\n'), ((48125, 48168), 'pulumi.set', 'pulumi.set', (['self', '"""http_healthcheck"""', 'value'], {}), "(self, 'http_healthcheck', value)\n", (48135, 48168), False, 'import pulumi\n'), ((48530, 48573), 'pulumi.get', 'pulumi.get', (['self', '"""interval_jitter_percent"""'], {}), "(self, 'interval_jitter_percent')\n", (48540, 48573), False, 'import pulumi\n'), ((48696, 48746), 'pulumi.set', 'pulumi.set', (['self', '"""interval_jitter_percent"""', 'value'], {}), "(self, 'interval_jitter_percent', value)\n", (48706, 48746), False, 'import pulumi\n'), ((49079, 49117), 'pulumi.get', 'pulumi.get', (['self', '"""stream_healthcheck"""'], {}), "(self, 'stream_healthcheck')\n", (49089, 49117), False, 'import pulumi\n'), ((49285, 49330), 'pulumi.set', 'pulumi.set', (['self', '"""stream_healthcheck"""', 'value'], {}), "(self, 'stream_healthcheck', value)\n", (49295, 49330), False, 'import pulumi\n'), ((49698, 49737), 'pulumi.get', 'pulumi.get', (['self', '"""unhealthy_threshold"""'], {}), "(self, 'unhealthy_threshold')\n", (49708, 49737), False, 'import pulumi\n'), ((49850, 49896), 'pulumi.set', 'pulumi.set', (['self', '"""unhealthy_threshold"""', 'value'], {}), "(self, 'unhealthy_threshold', value)\n", (49860, 49896), False, 'import pulumi\n'), ((50531, 50563), 'pulumi.get', 'pulumi.get', (['self', '"""service_name"""'], {}), "(self, 'service_name')\n", (50541, 50563), False, 'import pulumi\n'), ((50662, 50701), 'pulumi.set', 'pulumi.set', (['self', '"""service_name"""', 'value'], {}), "(self, 'service_name', value)\n", (50672, 50701), False, 'import pulumi\n'), ((51206, 51240), 'pulumi.set', 'pulumi.set', (['__self__', '"""path"""', 'path'], {}), "(__self__, 'path', path)\n", (51216, 51240), False, 'import pulumi\n'), ((51529, 51553), 'pulumi.get', 'pulumi.get', (['self', '"""path"""'], {}), "(self, 'path')\n", (51539, 51553), False, 'import pulumi\n'), ((51626, 51657), 'pulumi.set', 'pulumi.set', (['self', '"""path"""', 'value'], {}), "(self, 'path', value)\n", (51636, 51657), False, 'import pulumi\n'), ((51816, 51840), 'pulumi.get', 'pulumi.get', (['self', '"""host"""'], {}), "(self, 'host')\n", (51826, 51840), False, 'import pulumi\n'), ((51923, 51954), 'pulumi.set', 'pulumi.set', (['self', '"""host"""', 'value'], {}), "(self, 'host', value)\n", (51933, 51954), False, 'import pulumi\n'), ((52127, 52152), 'pulumi.get', 'pulumi.get', (['self', '"""http2"""'], {}), "(self, 'http2')\n", (52137, 52152), False, 'import pulumi\n'), ((52238, 52270), 'pulumi.set', 'pulumi.set', (['self', '"""http2"""', 'value'], {}), "(self, 'http2', value)\n", (52248, 52270), False, 'import pulumi\n'), ((53022, 53049), 'pulumi.get', 'pulumi.get', (['self', '"""receive"""'], {}), "(self, 'receive')\n", (53032, 53049), False, 'import pulumi\n'), ((53138, 53172), 'pulumi.set', 'pulumi.set', (['self', '"""receive"""', 'value'], {}), "(self, 'receive', value)\n", (53148, 53172), False, 'import pulumi\n'), ((53366, 53390), 'pulumi.get', 'pulumi.get', (['self', '"""send"""'], {}), "(self, 'send')\n", (53376, 53390), False, 'import pulumi\n'), ((53473, 53504), 'pulumi.set', 'pulumi.set', (['self', '"""send"""', 'value'], {}), "(self, 'send', value)\n", (53483, 53504), False, 'import pulumi\n'), ((55200, 55250), 'pulumi.get', 'pulumi.get', (['self', '"""locality_aware_routing_percent"""'], {}), "(self, 'locality_aware_routing_percent')\n", (55210, 55250), False, 'import pulumi\n'), ((55385, 55442), 'pulumi.set', 'pulumi.set', (['self', '"""locality_aware_routing_percent"""', 'value'], {}), "(self, 'locality_aware_routing_percent', value)\n", (55395, 55442), False, 'import pulumi\n'), ((55867, 55902), 'pulumi.get', 'pulumi.get', (['self', '"""panic_threshold"""'], {}), "(self, 'panic_threshold')\n", (55877, 55902), False, 'import pulumi\n'), ((56007, 56049), 'pulumi.set', 'pulumi.set', (['self', '"""panic_threshold"""', 'value'], {}), "(self, 'panic_threshold', value)\n", (56017, 56049), False, 'import pulumi\n'), ((56333, 56368), 'pulumi.get', 'pulumi.get', (['self', '"""strict_locality"""'], {}), "(self, 'strict_locality')\n", (56343, 56368), False, 'import pulumi\n'), ((56474, 56516), 'pulumi.set', 'pulumi.set', (['self', '"""strict_locality"""', 'value'], {}), "(self, 'strict_locality', value)\n", (56484, 56516), False, 'import pulumi\n'), ((57802, 57825), 'pulumi.get', 'pulumi.get', (['self', '"""sni"""'], {}), "(self, 'sni')\n", (57812, 57825), False, 'import pulumi\n'), ((57906, 57936), 'pulumi.set', 'pulumi.set', (['self', '"""sni"""', 'value'], {}), "(self, 'sni', value)\n", (57916, 57936), False, 'import pulumi\n'), ((58126, 58164), 'pulumi.get', 'pulumi.get', (['self', '"""validation_context"""'], {}), "(self, 'validation_context')\n", (58136, 58164), False, 'import pulumi\n'), ((58324, 58369), 'pulumi.set', 'pulumi.set', (['self', '"""validation_context"""', 'value'], {}), "(self, 'validation_context', value)\n", (58334, 58369), False, 'import pulumi\n'), ((58969, 59005), 'pulumi.get', 'pulumi.get', (['self', '"""trusted_ca_bytes"""'], {}), "(self, 'trusted_ca_bytes')\n", (58979, 59005), False, 'import pulumi\n'), ((59112, 59155), 'pulumi.set', 'pulumi.set', (['self', '"""trusted_ca_bytes"""', 'value'], {}), "(self, 'trusted_ca_bytes', value)\n", (59122, 59155), False, 'import pulumi\n'), ((59285, 59318), 'pulumi.get', 'pulumi.get', (['self', '"""trusted_ca_id"""'], {}), "(self, 'trusted_ca_id')\n", (59295, 59318), False, 'import pulumi\n'), ((59419, 59459), 'pulumi.set', 'pulumi.set', (['self', '"""trusted_ca_id"""', 'value'], {}), "(self, 'trusted_ca_id', value)\n", (59429, 59459), False, 'import pulumi\n'), ((59870, 59914), 'pulumi.set', 'pulumi.set', (['__self__', '"""locations"""', 'locations'], {}), "(__self__, 'locations', locations)\n", (59880, 59914), False, 'import pulumi\n'), ((60168, 60197), 'pulumi.get', 'pulumi.get', (['self', '"""locations"""'], {}), "(self, 'locations')\n", (60178, 60197), False, 'import pulumi\n'), ((60346, 60382), 'pulumi.set', 'pulumi.set', (['self', '"""locations"""', 'value'], {}), "(self, 'locations', value)\n", (60356, 60382), False, 'import pulumi\n'), ((60984, 61028), 'pulumi.set', 'pulumi.set', (['__self__', '"""subnet_id"""', 'subnet_id'], {}), "(__self__, 'subnet_id', subnet_id)\n", (60994, 61028), False, 'import pulumi\n'), ((61037, 61077), 'pulumi.set', 'pulumi.set', (['__self__', '"""zone_id"""', 'zone_id'], {}), "(__self__, 'zone_id', zone_id)\n", (61047, 61077), False, 'import pulumi\n'), ((61381, 61410), 'pulumi.get', 'pulumi.get', (['self', '"""subnet_id"""'], {}), "(self, 'subnet_id')\n", (61391, 61410), False, 'import pulumi\n'), ((61493, 61529), 'pulumi.set', 'pulumi.set', (['self', '"""subnet_id"""', 'value'], {}), "(self, 'subnet_id', value)\n", (61503, 61529), False, 'import pulumi\n'), ((61714, 61741), 'pulumi.get', 'pulumi.get', (['self', '"""zone_id"""'], {}), "(self, 'zone_id')\n", (61724, 61741), False, 'import pulumi\n'), ((61820, 61854), 'pulumi.set', 'pulumi.set', (['self', '"""zone_id"""', 'value'], {}), "(self, 'zone_id', value)\n", (61830, 61854), False, 'import pulumi\n'), ((62094, 62129), 'pulumi.get', 'pulumi.get', (['self', '"""disable_traffic"""'], {}), "(self, 'disable_traffic')\n", (62104, 62129), False, 'import pulumi\n'), ((62235, 62277), 'pulumi.set', 'pulumi.set', (['self', '"""disable_traffic"""', 'value'], {}), "(self, 'disable_traffic', value)\n", (62245, 62277), False, 'import pulumi\n'), ((63242, 63276), 'pulumi.set', 'pulumi.set', (['__self__', '"""name"""', 'name'], {}), "(__self__, 'name', name)\n", (63252, 63276), False, 'import pulumi\n'), ((63658, 63682), 'pulumi.get', 'pulumi.get', (['self', '"""name"""'], {}), "(self, 'name')\n", (63668, 63682), False, 'import pulumi\n'), ((63755, 63786), 'pulumi.set', 'pulumi.set', (['self', '"""name"""', 'value'], {}), "(self, 'name', value)\n", (63765, 63786), False, 'import pulumi\n'), ((64074, 64103), 'pulumi.get', 'pulumi.get', (['self', '"""endpoints"""'], {}), "(self, 'endpoints')\n", (64084, 64103), False, 'import pulumi\n'), ((64254, 64290), 'pulumi.set', 'pulumi.set', (['self', '"""endpoints"""', 'value'], {}), "(self, 'endpoints', value)\n", (64264, 64290), False, 'import pulumi\n'), ((64512, 64536), 'pulumi.get', 'pulumi.get', (['self', '"""http"""'], {}), "(self, 'http')\n", (64522, 64536), False, 'import pulumi\n'), ((64649, 64680), 'pulumi.set', 'pulumi.set', (['self', '"""http"""', 'value'], {}), "(self, 'http', value)\n", (64659, 64680), False, 'import pulumi\n'), ((64899, 64922), 'pulumi.get', 'pulumi.get', (['self', '"""tls"""'], {}), "(self, 'tls')\n", (64909, 64922), False, 'import pulumi\n'), ((65032, 65062), 'pulumi.set', 'pulumi.set', (['self', '"""tls"""', 'value'], {}), "(self, 'tls', value)\n", (65042, 65062), False, 'import pulumi\n'), ((65624, 65668), 'pulumi.set', 'pulumi.set', (['__self__', '"""addresses"""', 'addresses'], {}), "(__self__, 'addresses', addresses)\n", (65634, 65668), False, 'import pulumi\n'), ((65677, 65713), 'pulumi.set', 'pulumi.set', (['__self__', '"""ports"""', 'ports'], {}), "(__self__, 'ports', ports)\n", (65687, 65713), False, 'import pulumi\n'), ((65956, 65985), 'pulumi.get', 'pulumi.get', (['self', '"""addresses"""'], {}), "(self, 'addresses')\n", (65966, 65985), False, 'import pulumi\n'), ((66133, 66169), 'pulumi.set', 'pulumi.set', (['self', '"""addresses"""', 'value'], {}), "(self, 'addresses', value)\n", (66143, 66169), False, 'import pulumi\n'), ((66349, 66374), 'pulumi.get', 'pulumi.get', (['self', '"""ports"""'], {}), "(self, 'ports')\n", (66359, 66374), False, 'import pulumi\n'), ((66473, 66505), 'pulumi.set', 'pulumi.set', (['self', '"""ports"""', 'value'], {}), "(self, 'ports', value)\n", (66483, 66505), False, 'import pulumi\n'), ((68238, 68279), 'pulumi.get', 'pulumi.get', (['self', '"""external_ipv4_address"""'], {}), "(self, 'external_ipv4_address')\n", (68248, 68279), False, 'import pulumi\n'), ((68456, 68504), 'pulumi.set', 'pulumi.set', (['self', '"""external_ipv4_address"""', 'value'], {}), "(self, 'external_ipv4_address', value)\n", (68466, 68504), False, 'import pulumi\n'), ((68800, 68841), 'pulumi.get', 'pulumi.get', (['self', '"""external_ipv6_address"""'], {}), "(self, 'external_ipv6_address')\n", (68810, 68841), False, 'import pulumi\n'), ((69018, 69066), 'pulumi.set', 'pulumi.set', (['self', '"""external_ipv6_address"""', 'value'], {}), "(self, 'external_ipv6_address', value)\n", (69028, 69066), False, 'import pulumi\n'), ((69362, 69403), 'pulumi.get', 'pulumi.get', (['self', '"""internal_ipv4_address"""'], {}), "(self, 'internal_ipv4_address')\n", (69372, 69403), False, 'import pulumi\n'), ((69580, 69628), 'pulumi.set', 'pulumi.set', (['self', '"""internal_ipv4_address"""', 'value'], {}), "(self, 'internal_ipv4_address', value)\n", (69590, 69628), False, 'import pulumi\n'), ((70198, 70225), 'pulumi.get', 'pulumi.get', (['self', '"""address"""'], {}), "(self, 'address')\n", (70208, 70225), False, 'import pulumi\n'), ((70314, 70348), 'pulumi.set', 'pulumi.set', (['self', '"""address"""', 'value'], {}), "(self, 'address', value)\n", (70324, 70348), False, 'import pulumi\n'), ((70918, 70945), 'pulumi.get', 'pulumi.get', (['self', '"""address"""'], {}), "(self, 'address')\n", (70928, 70945), False, 'import pulumi\n'), ((71034, 71068), 'pulumi.set', 'pulumi.set', (['self', '"""address"""', 'value'], {}), "(self, 'address', value)\n", (71044, 71068), False, 'import pulumi\n'), ((71887, 71914), 'pulumi.get', 'pulumi.get', (['self', '"""address"""'], {}), "(self, 'address')\n", (71897, 71914), False, 'import pulumi\n'), ((72003, 72037), 'pulumi.set', 'pulumi.set', (['self', '"""address"""', 'value'], {}), "(self, 'address', value)\n", (72013, 72037), False, 'import pulumi\n'), ((72242, 72271), 'pulumi.get', 'pulumi.get', (['self', '"""subnet_id"""'], {}), "(self, 'subnet_id')\n", (72252, 72271), False, 'import pulumi\n'), ((72364, 72400), 'pulumi.set', 'pulumi.set', (['self', '"""subnet_id"""', 'value'], {}), "(self, 'subnet_id', value)\n", (72374, 72400), False, 'import pulumi\n'), ((73469, 73496), 'pulumi.get', 'pulumi.get', (['self', '"""handler"""'], {}), "(self, 'handler')\n", (73479, 73496), False, 'import pulumi\n'), ((73622, 73656), 'pulumi.set', 'pulumi.set', (['self', '"""handler"""', 'value'], {}), "(self, 'handler', value)\n", (73632, 73656), False, 'import pulumi\n'), ((73912, 73941), 'pulumi.get', 'pulumi.get', (['self', '"""redirects"""'], {}), "(self, 'redirects')\n", (73922, 73941), False, 'import pulumi\n'), ((74073, 74109), 'pulumi.set', 'pulumi.set', (['self', '"""redirects"""', 'value'], {}), "(self, 'redirects', value)\n", (74083, 74109), False, 'import pulumi\n'), ((75377, 75409), 'pulumi.get', 'pulumi.get', (['self', '"""allow_http10"""'], {}), "(self, 'allow_http10')\n", (75387, 75409), False, 'import pulumi\n'), ((75509, 75548), 'pulumi.set', 'pulumi.set', (['self', '"""allow_http10"""', 'value'], {}), "(self, 'allow_http10', value)\n", (75519, 75548), False, 'import pulumi\n'), ((75847, 75880), 'pulumi.get', 'pulumi.get', (['self', '"""http2_options"""'], {}), "(self, 'http2_options')\n", (75857, 75880), False, 'import pulumi\n'), ((76030, 76070), 'pulumi.set', 'pulumi.set', (['self', '"""http2_options"""', 'value'], {}), "(self, 'http2_options', value)\n", (76040, 76070), False, 'import pulumi\n'), ((76250, 76284), 'pulumi.get', 'pulumi.get', (['self', '"""http_router_id"""'], {}), "(self, 'http_router_id')\n", (76260, 76284), False, 'import pulumi\n'), ((76387, 76428), 'pulumi.set', 'pulumi.set', (['self', '"""http_router_id"""', 'value'], {}), "(self, 'http_router_id', value)\n", (76397, 76428), False, 'import pulumi\n'), ((77082, 77124), 'pulumi.get', 'pulumi.get', (['self', '"""max_concurrent_streams"""'], {}), "(self, 'max_concurrent_streams')\n", (77092, 77124), False, 'import pulumi\n'), ((77243, 77292), 'pulumi.set', 'pulumi.set', (['self', '"""max_concurrent_streams"""', 'value'], {}), "(self, 'max_concurrent_streams', value)\n", (77253, 77292), False, 'import pulumi\n'), ((77695, 77728), 'pulumi.get', 'pulumi.get', (['self', '"""http_to_https"""'], {}), "(self, 'http_to_https')\n", (77705, 77728), False, 'import pulumi\n'), ((77830, 77870), 'pulumi.set', 'pulumi.set', (['self', '"""http_to_https"""', 'value'], {}), "(self, 'http_to_https', value)\n", (77840, 77870), False, 'import pulumi\n'), ((78534, 78590), 'pulumi.set', 'pulumi.set', (['__self__', '"""default_handler"""', 'default_handler'], {}), "(__self__, 'default_handler', default_handler)\n", (78544, 78590), False, 'import pulumi\n'), ((78947, 78982), 'pulumi.get', 'pulumi.get', (['self', '"""default_handler"""'], {}), "(self, 'default_handler')\n", (78957, 78982), False, 'import pulumi\n'), ((79120, 79162), 'pulumi.set', 'pulumi.set', (['self', '"""default_handler"""', 'value'], {}), "(self, 'default_handler', value)\n", (79130, 79162), False, 'import pulumi\n'), ((79441, 79473), 'pulumi.get', 'pulumi.get', (['self', '"""sni_handlers"""'], {}), "(self, 'sni_handlers')\n", (79451, 79473), False, 'import pulumi\n'), ((79635, 79674), 'pulumi.set', 'pulumi.set', (['self', '"""sni_handlers"""', 'value'], {}), "(self, 'sni_handlers', value)\n", (79645, 79674), False, 'import pulumi\n'), ((80463, 80519), 'pulumi.set', 'pulumi.set', (['__self__', '"""certificate_ids"""', 'certificate_ids'], {}), "(__self__, 'certificate_ids', certificate_ids)\n", (80473, 80519), False, 'import pulumi\n'), ((81014, 81049), 'pulumi.get', 'pulumi.get', (['self', '"""certificate_ids"""'], {}), "(self, 'certificate_ids')\n", (81024, 81049), False, 'import pulumi\n'), ((81168, 81210), 'pulumi.set', 'pulumi.set', (['self', '"""certificate_ids"""', 'value'], {}), "(self, 'certificate_ids', value)\n", (81178, 81210), False, 'import pulumi\n'), ((81483, 81515), 'pulumi.get', 'pulumi.get', (['self', '"""http_handler"""'], {}), "(self, 'http_handler')\n", (81493, 81515), False, 'import pulumi\n'), ((81668, 81707), 'pulumi.set', 'pulumi.set', (['self', '"""http_handler"""', 'value'], {}), "(self, 'http_handler', value)\n", (81678, 81707), False, 'import pulumi\n'), ((83026, 83058), 'pulumi.get', 'pulumi.get', (['self', '"""allow_http10"""'], {}), "(self, 'allow_http10')\n", (83036, 83058), False, 'import pulumi\n'), ((83158, 83197), 'pulumi.set', 'pulumi.set', (['self', '"""allow_http10"""', 'value'], {}), "(self, 'allow_http10', value)\n", (83168, 83197), False, 'import pulumi\n'), ((83513, 83546), 'pulumi.get', 'pulumi.get', (['self', '"""http2_options"""'], {}), "(self, 'http2_options')\n", (83523, 83546), False, 'import pulumi\n'), ((83713, 83753), 'pulumi.set', 'pulumi.set', (['self', '"""http2_options"""', 'value'], {}), "(self, 'http2_options', value)\n", (83723, 83753), False, 'import pulumi\n'), ((83933, 83967), 'pulumi.get', 'pulumi.get', (['self', '"""http_router_id"""'], {}), "(self, 'http_router_id')\n", (83943, 83967), False, 'import pulumi\n'), ((84070, 84111), 'pulumi.set', 'pulumi.set', (['self', '"""http_router_id"""', 'value'], {}), "(self, 'http_router_id', value)\n", (84080, 84111), False, 'import pulumi\n'), ((84782, 84824), 'pulumi.get', 'pulumi.get', (['self', '"""max_concurrent_streams"""'], {}), "(self, 'max_concurrent_streams')\n", (84792, 84824), False, 'import pulumi\n'), ((84943, 84992), 'pulumi.set', 'pulumi.set', (['self', '"""max_concurrent_streams"""', 'value'], {}), "(self, 'max_concurrent_streams', value)\n", (84953, 84992), False, 'import pulumi\n'), ((85653, 85693), 'pulumi.set', 'pulumi.set', (['__self__', '"""handler"""', 'handler'], {}), "(__self__, 'handler', handler)\n", (85663, 85693), False, 'import pulumi\n'), ((85702, 85736), 'pulumi.set', 'pulumi.set', (['__self__', '"""name"""', 'name'], {}), "(__self__, 'name', name)\n", (85712, 85736), False, 'import pulumi\n'), ((85745, 85795), 'pulumi.set', 'pulumi.set', (['__self__', '"""server_names"""', 'server_names'], {}), "(__self__, 'server_names', server_names)\n", (85755, 85795), False, 'import pulumi\n'), ((86048, 86075), 'pulumi.get', 'pulumi.get', (['self', '"""handler"""'], {}), "(self, 'handler')\n", (86058, 86075), False, 'import pulumi\n'), ((86200, 86234), 'pulumi.set', 'pulumi.set', (['self', '"""handler"""', 'value'], {}), "(self, 'handler', value)\n", (86210, 86234), False, 'import pulumi\n'), ((86376, 86400), 'pulumi.get', 'pulumi.get', (['self', '"""name"""'], {}), "(self, 'name')\n", (86386, 86400), False, 'import pulumi\n'), ((86473, 86504), 'pulumi.set', 'pulumi.set', (['self', '"""name"""', 'value'], {}), "(self, 'name', value)\n", (86483, 86504), False, 'import pulumi\n'), ((86702, 86734), 'pulumi.get', 'pulumi.get', (['self', '"""server_names"""'], {}), "(self, 'server_names')\n", (86712, 86734), False, 'import pulumi\n'), ((86847, 86886), 'pulumi.set', 'pulumi.set', (['self', '"""server_names"""', 'value'], {}), "(self, 'server_names', value)\n", (86857, 86886), False, 'import pulumi\n'), ((87684, 87740), 'pulumi.set', 'pulumi.set', (['__self__', '"""certificate_ids"""', 'certificate_ids'], {}), "(__self__, 'certificate_ids', certificate_ids)\n", (87694, 87740), False, 'import pulumi\n'), ((88235, 88270), 'pulumi.get', 'pulumi.get', (['self', '"""certificate_ids"""'], {}), "(self, 'certificate_ids')\n", (88245, 88270), False, 'import pulumi\n'), ((88389, 88431), 'pulumi.set', 'pulumi.set', (['self', '"""certificate_ids"""', 'value'], {}), "(self, 'certificate_ids', value)\n", (88399, 88431), False, 'import pulumi\n'), ((88707, 88739), 'pulumi.get', 'pulumi.get', (['self', '"""http_handler"""'], {}), "(self, 'http_handler')\n", (88717, 88739), False, 'import pulumi\n'), ((88895, 88934), 'pulumi.set', 'pulumi.set', (['self', '"""http_handler"""', 'value'], {}), "(self, 'http_handler', value)\n", (88905, 88934), False, 'import pulumi\n'), ((90262, 90294), 'pulumi.get', 'pulumi.get', (['self', '"""allow_http10"""'], {}), "(self, 'allow_http10')\n", (90272, 90294), False, 'import pulumi\n'), ((90394, 90433), 'pulumi.set', 'pulumi.set', (['self', '"""allow_http10"""', 'value'], {}), "(self, 'allow_http10', value)\n", (90404, 90433), False, 'import pulumi\n'), ((90752, 90785), 'pulumi.get', 'pulumi.get', (['self', '"""http2_options"""'], {}), "(self, 'http2_options')\n", (90762, 90785), False, 'import pulumi\n'), ((90955, 90995), 'pulumi.set', 'pulumi.set', (['self', '"""http2_options"""', 'value'], {}), "(self, 'http2_options', value)\n", (90965, 90995), False, 'import pulumi\n'), ((91175, 91209), 'pulumi.get', 'pulumi.get', (['self', '"""http_router_id"""'], {}), "(self, 'http_router_id')\n", (91185, 91209), False, 'import pulumi\n'), ((91312, 91353), 'pulumi.set', 'pulumi.set', (['self', '"""http_router_id"""', 'value'], {}), "(self, 'http_router_id', value)\n", (91322, 91353), False, 'import pulumi\n'), ((92027, 92069), 'pulumi.get', 'pulumi.get', (['self', '"""max_concurrent_streams"""'], {}), "(self, 'max_concurrent_streams')\n", (92037, 92069), False, 'import pulumi\n'), ((92188, 92237), 'pulumi.set', 'pulumi.set', (['self', '"""max_concurrent_streams"""', 'value'], {}), "(self, 'max_concurrent_streams', value)\n", (92198, 92237), False, 'import pulumi\n'), ((92731, 92777), 'pulumi.set', 'pulumi.set', (['__self__', '"""ip_address"""', 'ip_address'], {}), "(__self__, 'ip_address', ip_address)\n", (92741, 92777), False, 'import pulumi\n'), ((92786, 92830), 'pulumi.set', 'pulumi.set', (['__self__', '"""subnet_id"""', 'subnet_id'], {}), "(__self__, 'subnet_id', subnet_id)\n", (92796, 92830), False, 'import pulumi\n'), ((93003, 93033), 'pulumi.get', 'pulumi.get', (['self', '"""ip_address"""'], {}), "(self, 'ip_address')\n", (93013, 93033), False, 'import pulumi\n'), ((93118, 93155), 'pulumi.set', 'pulumi.set', (['self', '"""ip_address"""', 'value'], {}), "(self, 'ip_address', value)\n", (93128, 93155), False, 'import pulumi\n'), ((93460, 93489), 'pulumi.get', 'pulumi.get', (['self', '"""subnet_id"""'], {}), "(self, 'subnet_id')\n", (93470, 93489), False, 'import pulumi\n'), ((93572, 93608), 'pulumi.set', 'pulumi.set', (['self', '"""subnet_id"""', 'value'], {}), "(self, 'subnet_id', value)\n", (93582, 93608), False, 'import pulumi\n'), ((94921, 94947), 'pulumi.get', 'pulumi.get', (['self', '"""append"""'], {}), "(self, 'append')\n", (94931, 94947), False, 'import pulumi\n'), ((95034, 95067), 'pulumi.set', 'pulumi.set', (['self', '"""append"""', 'value'], {}), "(self, 'append', value)\n", (95044, 95067), False, 'import pulumi\n'), ((95219, 95243), 'pulumi.get', 'pulumi.get', (['self', '"""name"""'], {}), "(self, 'name')\n", (95229, 95243), False, 'import pulumi\n'), ((95326, 95357), 'pulumi.set', 'pulumi.set', (['self', '"""name"""', 'value'], {}), "(self, 'name', value)\n", (95336, 95357), False, 'import pulumi\n'), ((95520, 95546), 'pulumi.get', 'pulumi.get', (['self', '"""remove"""'], {}), "(self, 'remove')\n", (95530, 95546), False, 'import pulumi\n'), ((95634, 95667), 'pulumi.set', 'pulumi.set', (['self', '"""remove"""', 'value'], {}), "(self, 'remove', value)\n", (95644, 95667), False, 'import pulumi\n'), ((96004, 96031), 'pulumi.get', 'pulumi.get', (['self', '"""replace"""'], {}), "(self, 'replace')\n", (96014, 96031), False, 'import pulumi\n'), ((96120, 96154), 'pulumi.set', 'pulumi.set', (['self', '"""replace"""', 'value'], {}), "(self, 'replace', value)\n", (96130, 96154), False, 'import pulumi\n'), ((97468, 97494), 'pulumi.get', 'pulumi.get', (['self', '"""append"""'], {}), "(self, 'append')\n", (97478, 97494), False, 'import pulumi\n'), ((97581, 97614), 'pulumi.set', 'pulumi.set', (['self', '"""append"""', 'value'], {}), "(self, 'append', value)\n", (97591, 97614), False, 'import pulumi\n'), ((97766, 97790), 'pulumi.get', 'pulumi.get', (['self', '"""name"""'], {}), "(self, 'name')\n", (97776, 97790), False, 'import pulumi\n'), ((97873, 97904), 'pulumi.set', 'pulumi.set', (['self', '"""name"""', 'value'], {}), "(self, 'name', value)\n", (97883, 97904), False, 'import pulumi\n'), ((98067, 98093), 'pulumi.get', 'pulumi.get', (['self', '"""remove"""'], {}), "(self, 'remove')\n", (98077, 98093), False, 'import pulumi\n'), ((98181, 98214), 'pulumi.set', 'pulumi.set', (['self', '"""remove"""', 'value'], {}), "(self, 'remove', value)\n", (98191, 98214), False, 'import pulumi\n'), ((98551, 98578), 'pulumi.get', 'pulumi.get', (['self', '"""replace"""'], {}), "(self, 'replace')\n", (98561, 98578), False, 'import pulumi\n'), ((98667, 98701), 'pulumi.set', 'pulumi.set', (['self', '"""replace"""', 'value'], {}), "(self, 'replace', value)\n", (98677, 98701), False, 'import pulumi\n'), ((99889, 99919), 'pulumi.get', 'pulumi.get', (['self', '"""grpc_route"""'], {}), "(self, 'grpc_route')\n", (99899, 99919), False, 'import pulumi\n'), ((100045, 100082), 'pulumi.set', 'pulumi.set', (['self', '"""grpc_route"""', 'value'], {}), "(self, 'grpc_route', value)\n", (100055, 100082), False, 'import pulumi\n'), ((100326, 100356), 'pulumi.get', 'pulumi.get', (['self', '"""http_route"""'], {}), "(self, 'http_route')\n", (100336, 100356), False, 'import pulumi\n'), ((100482, 100519), 'pulumi.set', 'pulumi.set', (['self', '"""http_route"""', 'value'], {}), "(self, 'http_route', value)\n", (100492, 100519), False, 'import pulumi\n'), ((100671, 100695), 'pulumi.get', 'pulumi.get', (['self', '"""name"""'], {}), "(self, 'name')\n", (100681, 100695), False, 'import pulumi\n'), ((100778, 100809), 'pulumi.set', 'pulumi.set', (['self', '"""name"""', 'value'], {}), "(self, 'name', value)\n", (100788, 100809), False, 'import pulumi\n'), ((102488, 102520), 'pulumi.get', 'pulumi.get', (['self', '"""grpc_matches"""'], {}), "(self, 'grpc_matches')\n", (102498, 102520), False, 'import pulumi\n'), ((102683, 102722), 'pulumi.set', 'pulumi.set', (['self', '"""grpc_matches"""', 'value'], {}), "(self, 'grpc_matches', value)\n", (102693, 102722), False, 'import pulumi\n'), ((103001, 103038), 'pulumi.get', 'pulumi.get', (['self', '"""grpc_route_action"""'], {}), "(self, 'grpc_route_action')\n", (103011, 103038), False, 'import pulumi\n'), ((103193, 103237), 'pulumi.set', 'pulumi.set', (['self', '"""grpc_route_action"""', 'value'], {}), "(self, 'grpc_route_action', value)\n", (103203, 103237), False, 'import pulumi\n'), ((103554, 103601), 'pulumi.get', 'pulumi.get', (['self', '"""grpc_status_response_action"""'], {}), "(self, 'grpc_status_response_action')\n", (103564, 103601), False, 'import pulumi\n'), ((103785, 103839), 'pulumi.set', 'pulumi.set', (['self', '"""grpc_status_response_action"""', 'value'], {}), "(self, 'grpc_status_response_action', value)\n", (103795, 103839), False, 'import pulumi\n'), ((104565, 104589), 'pulumi.get', 'pulumi.get', (['self', '"""fqmn"""'], {}), "(self, 'fqmn')\n", (104575, 104589), False, 'import pulumi\n'), ((104716, 104747), 'pulumi.set', 'pulumi.set', (['self', '"""fqmn"""', 'value'], {}), "(self, 'fqmn', value)\n", (104726, 104747), False, 'import pulumi\n'), ((105236, 105261), 'pulumi.get', 'pulumi.get', (['self', '"""exact"""'], {}), "(self, 'exact')\n", (105246, 105261), False, 'import pulumi\n'), ((105346, 105378), 'pulumi.set', 'pulumi.set', (['self', '"""exact"""', 'value'], {}), "(self, 'exact', value)\n", (105356, 105378), False, 'import pulumi\n'), ((105481, 105507), 'pulumi.get', 'pulumi.get', (['self', '"""prefix"""'], {}), "(self, 'prefix')\n", (105491, 105507), False, 'import pulumi\n'), ((105594, 105627), 'pulumi.set', 'pulumi.set', (['self', '"""prefix"""', 'value'], {}), "(self, 'prefix', value)\n", (105604, 105627), False, 'import pulumi\n'), ((106920, 106978), 'pulumi.set', 'pulumi.set', (['__self__', '"""backend_group_id"""', 'backend_group_id'], {}), "(__self__, 'backend_group_id', backend_group_id)\n", (106930, 106978), False, 'import pulumi\n'), ((107581, 107617), 'pulumi.get', 'pulumi.get', (['self', '"""backend_group_id"""'], {}), "(self, 'backend_group_id')\n", (107591, 107617), False, 'import pulumi\n'), ((107714, 107757), 'pulumi.set', 'pulumi.set', (['self', '"""backend_group_id"""', 'value'], {}), "(self, 'backend_group_id', value)\n", (107724, 107757), False, 'import pulumi\n'), ((107969, 108006), 'pulumi.get', 'pulumi.get', (['self', '"""auto_host_rewrite"""'], {}), "(self, 'auto_host_rewrite')\n", (107979, 108006), False, 'import pulumi\n'), ((108116, 108160), 'pulumi.set', 'pulumi.set', (['self', '"""auto_host_rewrite"""', 'value'], {}), "(self, 'auto_host_rewrite', value)\n", (108126, 108160), False, 'import pulumi\n'), ((108345, 108377), 'pulumi.get', 'pulumi.get', (['self', '"""host_rewrite"""'], {}), "(self, 'host_rewrite')\n", (108355, 108377), False, 'import pulumi\n'), ((108476, 108515), 'pulumi.set', 'pulumi.set', (['self', '"""host_rewrite"""', 'value'], {}), "(self, 'host_rewrite', value)\n", (108486, 108515), False, 'import pulumi\n'), ((109014, 109046), 'pulumi.get', 'pulumi.get', (['self', '"""idle_timeout"""'], {}), "(self, 'idle_timeout')\n", (109024, 109046), False, 'import pulumi\n'), ((109145, 109184), 'pulumi.set', 'pulumi.set', (['self', '"""idle_timeout"""', 'value'], {}), "(self, 'idle_timeout', value)\n", (109155, 109184), False, 'import pulumi\n'), ((109461, 109492), 'pulumi.get', 'pulumi.get', (['self', '"""max_timeout"""'], {}), "(self, 'max_timeout')\n", (109471, 109492), False, 'import pulumi\n'), ((109589, 109627), 'pulumi.set', 'pulumi.set', (['self', '"""max_timeout"""', 'value'], {}), "(self, 'max_timeout', value)\n", (109599, 109627), False, 'import pulumi\n'), ((110423, 110449), 'pulumi.get', 'pulumi.get', (['self', '"""status"""'], {}), "(self, 'status')\n", (110433, 110449), False, 'import pulumi\n'), ((110536, 110569), 'pulumi.set', 'pulumi.set', (['self', '"""status"""', 'value'], {}), "(self, 'status', value)\n", (110546, 110569), False, 'import pulumi\n'), ((112599, 112641), 'pulumi.get', 'pulumi.get', (['self', '"""direct_response_action"""'], {}), "(self, 'direct_response_action')\n", (112609, 112641), False, 'import pulumi\n'), ((112811, 112860), 'pulumi.set', 'pulumi.set', (['self', '"""direct_response_action"""', 'value'], {}), "(self, 'direct_response_action', value)\n", (112821, 112860), False, 'import pulumi\n'), ((113150, 113182), 'pulumi.get', 'pulumi.get', (['self', '"""http_matches"""'], {}), "(self, 'http_matches')\n", (113160, 113182), False, 'import pulumi\n'), ((113345, 113384), 'pulumi.set', 'pulumi.set', (['self', '"""http_matches"""', 'value'], {}), "(self, 'http_matches', value)\n", (113355, 113384), False, 'import pulumi\n'), ((113663, 113700), 'pulumi.get', 'pulumi.get', (['self', '"""http_route_action"""'], {}), "(self, 'http_route_action')\n", (113673, 113700), False, 'import pulumi\n'), ((113855, 113899), 'pulumi.set', 'pulumi.set', (['self', '"""http_route_action"""', 'value'], {}), "(self, 'http_route_action', value)\n", (113865, 113899), False, 'import pulumi\n'), ((114172, 114207), 'pulumi.get', 'pulumi.get', (['self', '"""redirect_action"""'], {}), "(self, 'redirect_action')\n", (114182, 114207), False, 'import pulumi\n'), ((114357, 114399), 'pulumi.set', 'pulumi.set', (['self', '"""redirect_action"""', 'value'], {}), "(self, 'redirect_action', value)\n", (114367, 114399), False, 'import pulumi\n'), ((115238, 115262), 'pulumi.get', 'pulumi.get', (['self', '"""body"""'], {}), "(self, 'body')\n", (115248, 115262), False, 'import pulumi\n'), ((115345, 115376), 'pulumi.set', 'pulumi.set', (['self', '"""body"""', 'value'], {}), "(self, 'body', value)\n", (115355, 115376), False, 'import pulumi\n'), ((115676, 115702), 'pulumi.get', 'pulumi.get', (['self', '"""status"""'], {}), "(self, 'status')\n", (115686, 115702), False, 'import pulumi\n'), ((115789, 115822), 'pulumi.set', 'pulumi.set', (['self', '"""status"""', 'value'], {}), "(self, 'status', value)\n", (115799, 115822), False, 'import pulumi\n'), ((116729, 116761), 'pulumi.get', 'pulumi.get', (['self', '"""http_methods"""'], {}), "(self, 'http_methods')\n", (116739, 116761), False, 'import pulumi\n'), ((116870, 116909), 'pulumi.set', 'pulumi.set', (['self', '"""http_methods"""', 'value'], {}), "(self, 'http_methods', value)\n", (116880, 116909), False, 'import pulumi\n'), ((117149, 117173), 'pulumi.get', 'pulumi.get', (['self', '"""path"""'], {}), "(self, 'path')\n", (117159, 117173), False, 'import pulumi\n'), ((117300, 117331), 'pulumi.set', 'pulumi.set', (['self', '"""path"""', 'value'], {}), "(self, 'path', value)\n", (117310, 117331), False, 'import pulumi\n'), ((117820, 117845), 'pulumi.get', 'pulumi.get', (['self', '"""exact"""'], {}), "(self, 'exact')\n", (117830, 117845), False, 'import pulumi\n'), ((117930, 117962), 'pulumi.set', 'pulumi.set', (['self', '"""exact"""', 'value'], {}), "(self, 'exact', value)\n", (117940, 117962), False, 'import pulumi\n'), ((118065, 118091), 'pulumi.get', 'pulumi.get', (['self', '"""prefix"""'], {}), "(self, 'prefix')\n", (118075, 118091), False, 'import pulumi\n'), ((118178, 118211), 'pulumi.set', 'pulumi.set', (['self', '"""prefix"""', 'value'], {}), "(self, 'prefix', value)\n", (118188, 118211), False, 'import pulumi\n'), ((119979, 120037), 'pulumi.set', 'pulumi.set', (['__self__', '"""backend_group_id"""', 'backend_group_id'], {}), "(__self__, 'backend_group_id', backend_group_id)\n", (119989, 120037), False, 'import pulumi\n'), ((120837, 120873), 'pulumi.get', 'pulumi.get', (['self', '"""backend_group_id"""'], {}), "(self, 'backend_group_id')\n", (120847, 120873), False, 'import pulumi\n'), ((120970, 121013), 'pulumi.set', 'pulumi.set', (['self', '"""backend_group_id"""', 'value'], {}), "(self, 'backend_group_id', value)\n", (120980, 121013), False, 'import pulumi\n'), ((121225, 121262), 'pulumi.get', 'pulumi.get', (['self', '"""auto_host_rewrite"""'], {}), "(self, 'auto_host_rewrite')\n", (121235, 121262), False, 'import pulumi\n'), ((121372, 121416), 'pulumi.set', 'pulumi.set', (['self', '"""auto_host_rewrite"""', 'value'], {}), "(self, 'auto_host_rewrite', value)\n", (121382, 121416), False, 'import pulumi\n'), ((121601, 121633), 'pulumi.get', 'pulumi.get', (['self', '"""host_rewrite"""'], {}), "(self, 'host_rewrite')\n", (121611, 121633), False, 'import pulumi\n'), ((121732, 121771), 'pulumi.set', 'pulumi.set', (['self', '"""host_rewrite"""', 'value'], {}), "(self, 'host_rewrite', value)\n", (121742, 121771), False, 'import pulumi\n'), ((122270, 122302), 'pulumi.get', 'pulumi.get', (['self', '"""idle_timeout"""'], {}), "(self, 'idle_timeout')\n", (122280, 122302), False, 'import pulumi\n'), ((122401, 122440), 'pulumi.set', 'pulumi.set', (['self', '"""idle_timeout"""', 'value'], {}), "(self, 'idle_timeout', value)\n", (122411, 122440), False, 'import pulumi\n'), ((122671, 122705), 'pulumi.get', 'pulumi.get', (['self', '"""prefix_rewrite"""'], {}), "(self, 'prefix_rewrite')\n", (122681, 122705), False, 'import pulumi\n'), ((122808, 122849), 'pulumi.set', 'pulumi.set', (['self', '"""prefix_rewrite"""', 'value'], {}), "(self, 'prefix_rewrite', value)\n", (122818, 122849), False, 'import pulumi\n'), ((123127, 123154), 'pulumi.get', 'pulumi.get', (['self', '"""timeout"""'], {}), "(self, 'timeout')\n", (123137, 123154), False, 'import pulumi\n'), ((123243, 123277), 'pulumi.set', 'pulumi.set', (['self', '"""timeout"""', 'value'], {}), "(self, 'timeout', value)\n", (123253, 123277), False, 'import pulumi\n'), ((123568, 123601), 'pulumi.get', 'pulumi.get', (['self', '"""upgrade_types"""'], {}), "(self, 'upgrade_types')\n", (123578, 123601), False, 'import pulumi\n'), ((123726, 123766), 'pulumi.set', 'pulumi.set', (['self', '"""upgrade_types"""', 'value'], {}), "(self, 'upgrade_types', value)\n", (123736, 123766), False, 'import pulumi\n'), ((126034, 126066), 'pulumi.get', 'pulumi.get', (['self', '"""remove_query"""'], {}), "(self, 'remove_query')\n", (126044, 126066), False, 'import pulumi\n'), ((126166, 126205), 'pulumi.set', 'pulumi.set', (['self', '"""remove_query"""', 'value'], {}), "(self, 'remove_query', value)\n", (126176, 126205), False, 'import pulumi\n'), ((126385, 126417), 'pulumi.get', 'pulumi.get', (['self', '"""replace_host"""'], {}), "(self, 'replace_host')\n", (126395, 126417), False, 'import pulumi\n'), ((126516, 126555), 'pulumi.set', 'pulumi.set', (['self', '"""replace_host"""', 'value'], {}), "(self, 'replace_host', value)\n", (126526, 126555), False, 'import pulumi\n'), ((126730, 126762), 'pulumi.get', 'pulumi.get', (['self', '"""replace_path"""'], {}), "(self, 'replace_path')\n", (126740, 126762), False, 'import pulumi\n'), ((126861, 126900), 'pulumi.set', 'pulumi.set', (['self', '"""replace_path"""', 'value'], {}), "(self, 'replace_path', value)\n", (126871, 126900), False, 'import pulumi\n'), ((127076, 127108), 'pulumi.get', 'pulumi.get', (['self', '"""replace_port"""'], {}), "(self, 'replace_port')\n", (127086, 127108), False, 'import pulumi\n'), ((127207, 127246), 'pulumi.set', 'pulumi.set', (['self', '"""replace_port"""', 'value'], {}), "(self, 'replace_port', value)\n", (127217, 127246), False, 'import pulumi\n'), ((127591, 127625), 'pulumi.get', 'pulumi.get', (['self', '"""replace_prefix"""'], {}), "(self, 'replace_prefix')\n", (127601, 127625), False, 'import pulumi\n'), ((127728, 127769), 'pulumi.set', 'pulumi.set', (['self', '"""replace_prefix"""', 'value'], {}), "(self, 'replace_prefix', value)\n", (127738, 127769), False, 'import pulumi\n'), ((128054, 128088), 'pulumi.get', 'pulumi.get', (['self', '"""replace_scheme"""'], {}), "(self, 'replace_scheme')\n", (128064, 128088), False, 'import pulumi\n'), ((128191, 128232), 'pulumi.set', 'pulumi.set', (['self', '"""replace_scheme"""', 'value'], {}), "(self, 'replace_scheme', value)\n", (128201, 128232), False, 'import pulumi\n'), ((128557, 128590), 'pulumi.get', 'pulumi.get', (['self', '"""response_code"""'], {}), "(self, 'response_code')\n", (128567, 128590), False, 'import pulumi\n'), ((128691, 128731), 'pulumi.set', 'pulumi.set', (['self', '"""response_code"""', 'value'], {}), "(self, 'response_code', value)\n", (128701, 128731), False, 'import pulumi\n'), ((129012, 129084), 'pulumi.set', 'pulumi.set', (['__self__', '"""disk_placement_group_id"""', 'disk_placement_group_id'], {}), "(__self__, 'disk_placement_group_id', disk_placement_group_id)\n", (129022, 129084), False, 'import pulumi\n'), ((129290, 129333), 'pulumi.get', 'pulumi.get', (['self', '"""disk_placement_group_id"""'], {}), "(self, 'disk_placement_group_id')\n", (129300, 129333), False, 'import pulumi\n'), ((129444, 129494), 'pulumi.set', 'pulumi.set', (['self', '"""disk_placement_group_id"""', 'value'], {}), "(self, 'disk_placement_group_id', value)\n", (129454, 129494), False, 'import pulumi\n'), ((131483, 131514), 'pulumi.get', 'pulumi.get', (['self', '"""auto_delete"""'], {}), "(self, 'auto_delete')\n", (131493, 131514), False, 'import pulumi\n'), ((131612, 131650), 'pulumi.set', 'pulumi.set', (['self', '"""auto_delete"""', 'value'], {}), "(self, 'auto_delete', value)\n", (131622, 131650), False, 'import pulumi\n'), ((131892, 131923), 'pulumi.get', 'pulumi.get', (['self', '"""device_name"""'], {}), "(self, 'device_name')\n", (131902, 131923), False, 'import pulumi\n'), ((132020, 132058), 'pulumi.set', 'pulumi.set', (['self', '"""device_name"""', 'value'], {}), "(self, 'device_name', value)\n", (132030, 132058), False, 'import pulumi\n'), ((132258, 132285), 'pulumi.get', 'pulumi.get', (['self', '"""disk_id"""'], {}), "(self, 'disk_id')\n", (132268, 132285), False, 'import pulumi\n'), ((132374, 132408), 'pulumi.set', 'pulumi.set', (['self', '"""disk_id"""', 'value'], {}), "(self, 'disk_id', value)\n", (132384, 132408), False, 'import pulumi\n'), ((132792, 132829), 'pulumi.get', 'pulumi.get', (['self', '"""initialize_params"""'], {}), "(self, 'initialize_params')\n", (132802, 132829), False, 'import pulumi\n'), ((132980, 133024), 'pulumi.set', 'pulumi.set', (['self', '"""initialize_params"""', 'value'], {}), "(self, 'initialize_params', value)\n", (132990, 133024), False, 'import pulumi\n'), ((133247, 133271), 'pulumi.get', 'pulumi.get', (['self', '"""mode"""'], {}), "(self, 'mode')\n", (133257, 133271), False, 'import pulumi\n'), ((133354, 133385), 'pulumi.set', 'pulumi.set', (['self', '"""mode"""', 'value'], {}), "(self, 'mode', value)\n", (133364, 133385), False, 'import pulumi\n'), ((134988, 135019), 'pulumi.get', 'pulumi.get', (['self', '"""description"""'], {}), "(self, 'description')\n", (134998, 135019), False, 'import pulumi\n'), ((135116, 135154), 'pulumi.set', 'pulumi.set', (['self', '"""description"""', 'value'], {}), "(self, 'description', value)\n", (135126, 135154), False, 'import pulumi\n'), ((135350, 135378), 'pulumi.get', 'pulumi.get', (['self', '"""image_id"""'], {}), "(self, 'image_id')\n", (135360, 135378), False, 'import pulumi\n'), ((135469, 135504), 'pulumi.set', 'pulumi.set', (['self', '"""image_id"""', 'value'], {}), "(self, 'image_id', value)\n", (135479, 135504), False, 'import pulumi\n'), ((135660, 135684), 'pulumi.get', 'pulumi.get', (['self', '"""name"""'], {}), "(self, 'name')\n", (135670, 135684), False, 'import pulumi\n'), ((135767, 135798), 'pulumi.set', 'pulumi.set', (['self', '"""name"""', 'value'], {}), "(self, 'name', value)\n", (135777, 135798), False, 'import pulumi\n'), ((135955, 135979), 'pulumi.get', 'pulumi.get', (['self', '"""size"""'], {}), "(self, 'size')\n", (135965, 135979), False, 'import pulumi\n'), ((136062, 136093), 'pulumi.set', 'pulumi.set', (['self', '"""size"""', 'value'], {}), "(self, 'size', value)\n", (136072, 136093), False, 'import pulumi\n'), ((136293, 136324), 'pulumi.get', 'pulumi.get', (['self', '"""snapshot_id"""'], {}), "(self, 'snapshot_id')\n", (136303, 136324), False, 'import pulumi\n'), ((136421, 136459), 'pulumi.set', 'pulumi.set', (['self', '"""snapshot_id"""', 'value'], {}), "(self, 'snapshot_id', value)\n", (136431, 136459), False, 'import pulumi\n'), ((136603, 136627), 'pulumi.get', 'pulumi.get', (['self', '"""type"""'], {}), "(self, 'type')\n", (136613, 136627), False, 'import pulumi\n'), ((136710, 136741), 'pulumi.set', 'pulumi.set', (['self', '"""type"""', 'value'], {}), "(self, 'type', value)\n", (136720, 136741), False, 'import pulumi\n'), ((137035, 137071), 'pulumi.set', 'pulumi.set', (['__self__', '"""zones"""', 'zones'], {}), "(__self__, 'zones', zones)\n", (137045, 137071), False, 'import pulumi\n'), ((137249, 137274), 'pulumi.get', 'pulumi.get', (['self', '"""zones"""'], {}), "(self, 'zones')\n", (137259, 137274), False, 'import pulumi\n'), ((137373, 137405), 'pulumi.set', 'pulumi.set', (['self', '"""zones"""', 'value'], {}), "(self, 'zones', value)\n", (137383, 137405), False, 'import pulumi\n'), ((139725, 139773), 'pulumi.get', 'pulumi.get', (['self', '"""max_opening_traffic_duration"""'], {}), "(self, 'max_opening_traffic_duration')\n", (139735, 139773), False, 'import pulumi\n'), ((139904, 139959), 'pulumi.set', 'pulumi.set', (['self', '"""max_opening_traffic_duration"""', 'value'], {}), "(self, 'max_opening_traffic_duration', value)\n", (139914, 139959), False, 'import pulumi\n'), ((140160, 140194), 'pulumi.get', 'pulumi.get', (['self', '"""status_message"""'], {}), "(self, 'status_message')\n", (140170, 140194), False, 'import pulumi\n'), ((140297, 140338), 'pulumi.set', 'pulumi.set', (['self', '"""status_message"""', 'value'], {}), "(self, 'status_message', value)\n", (140307, 140338), False, 'import pulumi\n'), ((140557, 140601), 'pulumi.get', 'pulumi.get', (['self', '"""target_group_description"""'], {}), "(self, 'target_group_description')\n", (140567, 140601), False, 'import pulumi\n'), ((140724, 140775), 'pulumi.set', 'pulumi.set', (['self', '"""target_group_description"""', 'value'], {}), "(self, 'target_group_description', value)\n", (140734, 140775), False, 'import pulumi\n'), ((140909, 140944), 'pulumi.get', 'pulumi.get', (['self', '"""target_group_id"""'], {}), "(self, 'target_group_id')\n", (140919, 140944), False, 'import pulumi\n'), ((141049, 141091), 'pulumi.set', 'pulumi.set', (['self', '"""target_group_id"""', 'value'], {}), "(self, 'target_group_id', value)\n", (141059, 141091), False, 'import pulumi\n'), ((141325, 141364), 'pulumi.get', 'pulumi.get', (['self', '"""target_group_labels"""'], {}), "(self, 'target_group_labels')\n", (141335, 141364), False, 'import pulumi\n'), ((141505, 141551), 'pulumi.set', 'pulumi.set', (['self', '"""target_group_labels"""', 'value'], {}), "(self, 'target_group_labels', value)\n", (141515, 141551), False, 'import pulumi\n'), ((141751, 141788), 'pulumi.get', 'pulumi.get', (['self', '"""target_group_name"""'], {}), "(self, 'target_group_name')\n", (141761, 141788), False, 'import pulumi\n'), ((141897, 141941), 'pulumi.set', 'pulumi.set', (['self', '"""target_group_name"""', 'value'], {}), "(self, 'target_group_name', value)\n", (141907, 141941), False, 'import pulumi\n'), ((143720, 143772), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_expansion"""', 'max_expansion'], {}), "(__self__, 'max_expansion', max_expansion)\n", (143730, 143772), False, 'import pulumi\n'), ((143781, 143837), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_unavailable"""', 'max_unavailable'], {}), "(__self__, 'max_unavailable', max_unavailable)\n", (143791, 143837), False, 'import pulumi\n'), ((144521, 144554), 'pulumi.get', 'pulumi.get', (['self', '"""max_expansion"""'], {}), "(self, 'max_expansion')\n", (144531, 144554), False, 'import pulumi\n'), ((144645, 144685), 'pulumi.set', 'pulumi.set', (['self', '"""max_expansion"""', 'value'], {}), "(self, 'max_expansion', value)\n", (144655, 144685), False, 'import pulumi\n'), ((144981, 145016), 'pulumi.get', 'pulumi.get', (['self', '"""max_unavailable"""'], {}), "(self, 'max_unavailable')\n", (144991, 145016), False, 'import pulumi\n'), ((145111, 145153), 'pulumi.set', 'pulumi.set', (['self', '"""max_unavailable"""', 'value'], {}), "(self, 'max_unavailable', value)\n", (145121, 145153), False, 'import pulumi\n'), ((145384, 145416), 'pulumi.get', 'pulumi.get', (['self', '"""max_creating"""'], {}), "(self, 'max_creating')\n", (145394, 145416), False, 'import pulumi\n'), ((145515, 145554), 'pulumi.set', 'pulumi.set', (['self', '"""max_creating"""', 'value'], {}), "(self, 'max_creating', value)\n", (145525, 145554), False, 'import pulumi\n'), ((145785, 145817), 'pulumi.get', 'pulumi.get', (['self', '"""max_deleting"""'], {}), "(self, 'max_deleting')\n", (145795, 145817), False, 'import pulumi\n'), ((145916, 145955), 'pulumi.set', 'pulumi.set', (['self', '"""max_deleting"""', 'value'], {}), "(self, 'max_deleting', value)\n", (145926, 145955), False, 'import pulumi\n'), ((146356, 146392), 'pulumi.get', 'pulumi.get', (['self', '"""startup_duration"""'], {}), "(self, 'startup_duration')\n", (146366, 146392), False, 'import pulumi\n'), ((146499, 146542), 'pulumi.set', 'pulumi.set', (['self', '"""startup_duration"""', 'value'], {}), "(self, 'startup_duration', value)\n", (146509, 146542), False, 'import pulumi\n'), ((146987, 147015), 'pulumi.get', 'pulumi.get', (['self', '"""strategy"""'], {}), "(self, 'strategy')\n", (146997, 147015), False, 'import pulumi\n'), ((147106, 147141), 'pulumi.set', 'pulumi.set', (['self', '"""strategy"""', 'value'], {}), "(self, 'strategy', value)\n", (147116, 147141), False, 'import pulumi\n'), ((149509, 149546), 'pulumi.get', 'pulumi.get', (['self', '"""healthy_threshold"""'], {}), "(self, 'healthy_threshold')\n", (149519, 149546), False, 'import pulumi\n'), ((149655, 149699), 'pulumi.set', 'pulumi.set', (['self', '"""healthy_threshold"""', 'value'], {}), "(self, 'healthy_threshold', value)\n", (149665, 149699), False, 'import pulumi\n'), ((149983, 150015), 'pulumi.get', 'pulumi.get', (['self', '"""http_options"""'], {}), "(self, 'http_options')\n", (149993, 150015), False, 'import pulumi\n'), ((150182, 150221), 'pulumi.set', 'pulumi.set', (['self', '"""http_options"""', 'value'], {}), "(self, 'http_options', value)\n", (150192, 150221), False, 'import pulumi\n'), ((150413, 150441), 'pulumi.get', 'pulumi.get', (['self', '"""interval"""'], {}), "(self, 'interval')\n", (150423, 150441), False, 'import pulumi\n'), ((150532, 150567), 'pulumi.set', 'pulumi.set', (['self', '"""interval"""', 'value'], {}), "(self, 'interval', value)\n", (150542, 150567), False, 'import pulumi\n'), ((150847, 150878), 'pulumi.get', 'pulumi.get', (['self', '"""tcp_options"""'], {}), "(self, 'tcp_options')\n", (150857, 150878), False, 'import pulumi\n'), ((151042, 151080), 'pulumi.set', 'pulumi.set', (['self', '"""tcp_options"""', 'value'], {}), "(self, 'tcp_options', value)\n", (151052, 151080), False, 'import pulumi\n'), ((151304, 151331), 'pulumi.get', 'pulumi.get', (['self', '"""timeout"""'], {}), "(self, 'timeout')\n", (151314, 151331), False, 'import pulumi\n'), ((151420, 151454), 'pulumi.set', 'pulumi.set', (['self', '"""timeout"""', 'value'], {}), "(self, 'timeout', value)\n", (151430, 151454), False, 'import pulumi\n'), ((151715, 151754), 'pulumi.get', 'pulumi.get', (['self', '"""unhealthy_threshold"""'], {}), "(self, 'unhealthy_threshold')\n", (151725, 151754), False, 'import pulumi\n'), ((151867, 151913), 'pulumi.set', 'pulumi.set', (['self', '"""unhealthy_threshold"""', 'value'], {}), "(self, 'unhealthy_threshold', value)\n", (151877, 151913), False, 'import pulumi\n'), ((152295, 152329), 'pulumi.set', 'pulumi.set', (['__self__', '"""path"""', 'path'], {}), "(__self__, 'path', path)\n", (152305, 152329), False, 'import pulumi\n'), ((152338, 152372), 'pulumi.set', 'pulumi.set', (['__self__', '"""port"""', 'port'], {}), "(__self__, 'port', port)\n", (152348, 152372), False, 'import pulumi\n'), ((152540, 152564), 'pulumi.get', 'pulumi.get', (['self', '"""path"""'], {}), "(self, 'path')\n", (152550, 152564), False, 'import pulumi\n'), ((152637, 152668), 'pulumi.set', 'pulumi.set', (['self', '"""path"""', 'value'], {}), "(self, 'path', value)\n", (152647, 152668), False, 'import pulumi\n'), ((152828, 152852), 'pulumi.get', 'pulumi.get', (['self', '"""port"""'], {}), "(self, 'port')\n", (152838, 152852), False, 'import pulumi\n'), ((152925, 152956), 'pulumi.set', 'pulumi.set', (['self', '"""port"""', 'value'], {}), "(self, 'port', value)\n", (152935, 152956), False, 'import pulumi\n'), ((153211, 153245), 'pulumi.set', 'pulumi.set', (['__self__', '"""port"""', 'port'], {}), "(__self__, 'port', port)\n", (153221, 153245), False, 'import pulumi\n'), ((153405, 153429), 'pulumi.get', 'pulumi.get', (['self', '"""port"""'], {}), "(self, 'port')\n", (153415, 153429), False, 'import pulumi\n'), ((153502, 153533), 'pulumi.set', 'pulumi.set', (['self', '"""port"""', 'value'], {}), "(self, 'port', value)\n", (153512, 153533), False, 'import pulumi\n'), ((156388, 156412), 'pulumi.get', 'pulumi.get', (['self', '"""fqdn"""'], {}), "(self, 'fqdn')\n", (156398, 156412), False, 'import pulumi\n'), ((156495, 156526), 'pulumi.set', 'pulumi.set', (['self', '"""fqdn"""', 'value'], {}), "(self, 'fqdn', value)\n", (156505, 156526), False, 'import pulumi\n'), ((156709, 156740), 'pulumi.get', 'pulumi.get', (['self', '"""instance_id"""'], {}), "(self, 'instance_id')\n", (156719, 156740), False, 'import pulumi\n'), ((156837, 156875), 'pulumi.set', 'pulumi.set', (['self', '"""instance_id"""', 'value'], {}), "(self, 'instance_id', value)\n", (156847, 156875), False, 'import pulumi\n'), ((157462, 157486), 'pulumi.get', 'pulumi.get', (['self', '"""name"""'], {}), "(self, 'name')\n", (157472, 157486), False, 'import pulumi\n'), ((157569, 157600), 'pulumi.set', 'pulumi.set', (['self', '"""name"""', 'value'], {}), "(self, 'name', value)\n", (157579, 157600), False, 'import pulumi\n'), ((157984, 158022), 'pulumi.get', 'pulumi.get', (['self', '"""network_interfaces"""'], {}), "(self, 'network_interfaces')\n", (157994, 158022), False, 'import pulumi\n'), ((158204, 158249), 'pulumi.set', 'pulumi.set', (['self', '"""network_interfaces"""', 'value'], {}), "(self, 'network_interfaces', value)\n", (158214, 158249), False, 'import pulumi\n'), ((158412, 158438), 'pulumi.get', 'pulumi.get', (['self', '"""status"""'], {}), "(self, 'status')\n", (158422, 158438), False, 'import pulumi\n'), ((158525, 158558), 'pulumi.set', 'pulumi.set', (['self', '"""status"""', 'value'], {}), "(self, 'status', value)\n", (158535, 158558), False, 'import pulumi\n'), ((158696, 158733), 'pulumi.get', 'pulumi.get', (['self', '"""status_changed_at"""'], {}), "(self, 'status_changed_at')\n", (158706, 158733), False, 'import pulumi\n'), ((158842, 158886), 'pulumi.set', 'pulumi.set', (['self', '"""status_changed_at"""', 'value'], {}), "(self, 'status_changed_at', value)\n", (158852, 158886), False, 'import pulumi\n'), ((159087, 159121), 'pulumi.get', 'pulumi.get', (['self', '"""status_message"""'], {}), "(self, 'status_message')\n", (159097, 159121), False, 'import pulumi\n'), ((159224, 159265), 'pulumi.set', 'pulumi.set', (['self', '"""status_message"""', 'value'], {}), "(self, 'status_message', value)\n", (159234, 159265), False, 'import pulumi\n'), ((159476, 159503), 'pulumi.get', 'pulumi.get', (['self', '"""zone_id"""'], {}), "(self, 'zone_id')\n", (159486, 159503), False, 'import pulumi\n'), ((159592, 159626), 'pulumi.set', 'pulumi.set', (['self', '"""zone_id"""', 'value'], {}), "(self, 'zone_id', value)\n", (159602, 159626), False, 'import pulumi\n'), ((162398, 162423), 'pulumi.get', 'pulumi.get', (['self', '"""index"""'], {}), "(self, 'index')\n", (162408, 162423), False, 'import pulumi\n'), ((162508, 162540), 'pulumi.set', 'pulumi.set', (['self', '"""index"""', 'value'], {}), "(self, 'index', value)\n", (162518, 162540), False, 'import pulumi\n'), ((162727, 162757), 'pulumi.get', 'pulumi.get', (['self', '"""ip_address"""'], {}), "(self, 'ip_address')\n", (162737, 162757), False, 'import pulumi\n'), ((162852, 162889), 'pulumi.set', 'pulumi.set', (['self', '"""ip_address"""', 'value'], {}), "(self, 'ip_address', value)\n", (162862, 162889), False, 'import pulumi\n'), ((163081, 163105), 'pulumi.get', 'pulumi.get', (['self', '"""ipv4"""'], {}), "(self, 'ipv4')\n", (163091, 163105), False, 'import pulumi\n'), ((163189, 163220), 'pulumi.set', 'pulumi.set', (['self', '"""ipv4"""', 'value'], {}), "(self, 'ipv4', value)\n", (163199, 163220), False, 'import pulumi\n'), ((163322, 163346), 'pulumi.get', 'pulumi.get', (['self', '"""ipv6"""'], {}), "(self, 'ipv6')\n", (163332, 163346), False, 'import pulumi\n'), ((163430, 163461), 'pulumi.set', 'pulumi.set', (['self', '"""ipv6"""', 'value'], {}), "(self, 'ipv6', value)\n", (163440, 163461), False, 'import pulumi\n'), ((163654, 163686), 'pulumi.get', 'pulumi.get', (['self', '"""ipv6_address"""'], {}), "(self, 'ipv6_address')\n", (163664, 163686), False, 'import pulumi\n'), ((163785, 163824), 'pulumi.set', 'pulumi.set', (['self', '"""ipv6_address"""', 'value'], {}), "(self, 'ipv6_address', value)\n", (163795, 163824), False, 'import pulumi\n'), ((164034, 164065), 'pulumi.get', 'pulumi.get', (['self', '"""mac_address"""'], {}), "(self, 'mac_address')\n", (164044, 164065), False, 'import pulumi\n'), ((164162, 164200), 'pulumi.set', 'pulumi.set', (['self', '"""mac_address"""', 'value'], {}), "(self, 'mac_address', value)\n", (164172, 164200), False, 'import pulumi\n'), ((164353, 164376), 'pulumi.get', 'pulumi.get', (['self', '"""nat"""'], {}), "(self, 'nat')\n", (164363, 164376), False, 'import pulumi\n'), ((164458, 164488), 'pulumi.set', 'pulumi.set', (['self', '"""nat"""', 'value'], {}), "(self, 'nat', value)\n", (164468, 164488), False, 'import pulumi\n'), ((164743, 164777), 'pulumi.get', 'pulumi.get', (['self', '"""nat_ip_address"""'], {}), "(self, 'nat_ip_address')\n", (164753, 164777), False, 'import pulumi\n'), ((164880, 164921), 'pulumi.set', 'pulumi.set', (['self', '"""nat_ip_address"""', 'value'], {}), "(self, 'nat_ip_address', value)\n", (164890, 164921), False, 'import pulumi\n'), ((165124, 165158), 'pulumi.get', 'pulumi.get', (['self', '"""nat_ip_version"""'], {}), "(self, 'nat_ip_version')\n", (165134, 165158), False, 'import pulumi\n'), ((165261, 165302), 'pulumi.set', 'pulumi.set', (['self', '"""nat_ip_version"""', 'value'], {}), "(self, 'nat_ip_version', value)\n", (165271, 165302), False, 'import pulumi\n'), ((165580, 165609), 'pulumi.get', 'pulumi.get', (['self', '"""subnet_id"""'], {}), "(self, 'subnet_id')\n", (165590, 165609), False, 'import pulumi\n'), ((165702, 165738), 'pulumi.set', 'pulumi.set', (['self', '"""subnet_id"""', 'value'], {}), "(self, 'subnet_id', value)\n", (165712, 165738), False, 'import pulumi\n'), ((170608, 170652), 'pulumi.set', 'pulumi.set', (['__self__', '"""boot_disk"""', 'boot_disk'], {}), "(__self__, 'boot_disk', boot_disk)\n", (170618, 170652), False, 'import pulumi\n'), ((170661, 170723), 'pulumi.set', 'pulumi.set', (['__self__', '"""network_interfaces"""', 'network_interfaces'], {}), "(__self__, 'network_interfaces', network_interfaces)\n", (170671, 170723), False, 'import pulumi\n'), ((170732, 170776), 'pulumi.set', 'pulumi.set', (['__self__', '"""resources"""', 'resources'], {}), "(__self__, 'resources', resources)\n", (170742, 170776), False, 'import pulumi\n'), ((172140, 172169), 'pulumi.get', 'pulumi.get', (['self', '"""boot_disk"""'], {}), "(self, 'boot_disk')\n", (172150, 172169), False, 'import pulumi\n'), ((172299, 172335), 'pulumi.set', 'pulumi.set', (['self', '"""boot_disk"""', 'value'], {}), "(self, 'boot_disk', value)\n", (172309, 172335), False, 'import pulumi\n'), ((172717, 172755), 'pulumi.get', 'pulumi.get', (['self', '"""network_interfaces"""'], {}), "(self, 'network_interfaces')\n", (172727, 172755), False, 'import pulumi\n'), ((172935, 172980), 'pulumi.set', 'pulumi.set', (['self', '"""network_interfaces"""', 'value'], {}), "(self, 'network_interfaces', value)\n", (172945, 172980), False, 'import pulumi\n'), ((173241, 173270), 'pulumi.get', 'pulumi.get', (['self', '"""resources"""'], {}), "(self, 'resources')\n", (173251, 173270), False, 'import pulumi\n'), ((173401, 173437), 'pulumi.set', 'pulumi.set', (['self', '"""resources"""', 'value'], {}), "(self, 'resources', value)\n", (173411, 173437), False, 'import pulumi\n'), ((173609, 173640), 'pulumi.get', 'pulumi.get', (['self', '"""description"""'], {}), "(self, 'description')\n", (173619, 173640), False, 'import pulumi\n'), ((173737, 173775), 'pulumi.set', 'pulumi.set', (['self', '"""description"""', 'value'], {}), "(self, 'description', value)\n", (173747, 173775), False, 'import pulumi\n'), ((174654, 174682), 'pulumi.get', 'pulumi.get', (['self', '"""hostname"""'], {}), "(self, 'hostname')\n", (174664, 174682), False, 'import pulumi\n'), ((174773, 174808), 'pulumi.set', 'pulumi.set', (['self', '"""hostname"""', 'value'], {}), "(self, 'hostname', value)\n", (174783, 174808), False, 'import pulumi\n'), ((174998, 175024), 'pulumi.get', 'pulumi.get', (['self', '"""labels"""'], {}), "(self, 'labels')\n", (175008, 175024), False, 'import pulumi\n'), ((175139, 175172), 'pulumi.set', 'pulumi.set', (['self', '"""labels"""', 'value'], {}), "(self, 'labels', value)\n", (175149, 175172), False, 'import pulumi\n'), ((175415, 175443), 'pulumi.get', 'pulumi.get', (['self', '"""metadata"""'], {}), "(self, 'metadata')\n", (175425, 175443), False, 'import pulumi\n'), ((175562, 175597), 'pulumi.set', 'pulumi.set', (['self', '"""metadata"""', 'value'], {}), "(self, 'metadata', value)\n", (175572, 175597), False, 'import pulumi\n'), ((176184, 176208), 'pulumi.get', 'pulumi.get', (['self', '"""name"""'], {}), "(self, 'name')\n", (176194, 176208), False, 'import pulumi\n'), ((176291, 176322), 'pulumi.set', 'pulumi.set', (['self', '"""name"""', 'value'], {}), "(self, 'name', value)\n", (176301, 176322), False, 'import pulumi\n'), ((176643, 176679), 'pulumi.get', 'pulumi.get', (['self', '"""network_settings"""'], {}), "(self, 'network_settings')\n", (176653, 176679), False, 'import pulumi\n'), ((176863, 176906), 'pulumi.set', 'pulumi.set', (['self', '"""network_settings"""', 'value'], {}), "(self, 'network_settings', value)\n", (176873, 176906), False, 'import pulumi\n'), ((177200, 177236), 'pulumi.get', 'pulumi.get', (['self', '"""placement_policy"""'], {}), "(self, 'placement_policy')\n", (177210, 177236), False, 'import pulumi\n'), ((177397, 177440), 'pulumi.set', 'pulumi.set', (['self', '"""placement_policy"""', 'value'], {}), "(self, 'placement_policy', value)\n", (177407, 177440), False, 'import pulumi\n'), ((177693, 177724), 'pulumi.get', 'pulumi.get', (['self', '"""platform_id"""'], {}), "(self, 'platform_id')\n", (177703, 177724), False, 'import pulumi\n'), ((177821, 177859), 'pulumi.set', 'pulumi.set', (['self', '"""platform_id"""', 'value'], {}), "(self, 'platform_id', value)\n", (177831, 177859), False, 'import pulumi\n'), ((178157, 178194), 'pulumi.get', 'pulumi.get', (['self', '"""scheduling_policy"""'], {}), "(self, 'scheduling_policy')\n", (178167, 178194), False, 'import pulumi\n'), ((178358, 178402), 'pulumi.set', 'pulumi.set', (['self', '"""scheduling_policy"""', 'value'], {}), "(self, 'scheduling_policy', value)\n", (178368, 178402), False, 'import pulumi\n'), ((178723, 178758), 'pulumi.get', 'pulumi.get', (['self', '"""secondary_disks"""'], {}), "(self, 'secondary_disks')\n", (178733, 178758), False, 'import pulumi\n'), ((178939, 178981), 'pulumi.set', 'pulumi.set', (['self', '"""secondary_disks"""', 'value'], {}), "(self, 'secondary_disks', value)\n", (178949, 178981), False, 'import pulumi\n'), ((179213, 179251), 'pulumi.get', 'pulumi.get', (['self', '"""service_account_id"""'], {}), "(self, 'service_account_id')\n", (179223, 179251), False, 'import pulumi\n'), ((179362, 179407), 'pulumi.set', 'pulumi.set', (['self', '"""service_account_id"""', 'value'], {}), "(self, 'service_account_id', value)\n", (179372, 179407), False, 'import pulumi\n'), ((181008, 181039), 'pulumi.get', 'pulumi.get', (['self', '"""device_name"""'], {}), "(self, 'device_name')\n", (181018, 181039), False, 'import pulumi\n'), ((181136, 181174), 'pulumi.set', 'pulumi.set', (['self', '"""device_name"""', 'value'], {}), "(self, 'device_name', value)\n", (181146, 181174), False, 'import pulumi\n'), ((181372, 181399), 'pulumi.get', 'pulumi.get', (['self', '"""disk_id"""'], {}), "(self, 'disk_id')\n", (181382, 181399), False, 'import pulumi\n'), ((181488, 181522), 'pulumi.set', 'pulumi.set', (['self', '"""disk_id"""', 'value'], {}), "(self, 'disk_id', value)\n", (181498, 181522), False, 'import pulumi\n'), ((181846, 181883), 'pulumi.get', 'pulumi.get', (['self', '"""initialize_params"""'], {}), "(self, 'initialize_params')\n", (181856, 181883), False, 'import pulumi\n'), ((182055, 182099), 'pulumi.set', 'pulumi.set', (['self', '"""initialize_params"""', 'value'], {}), "(self, 'initialize_params', value)\n", (182065, 182099), False, 'import pulumi\n'), ((182322, 182346), 'pulumi.get', 'pulumi.get', (['self', '"""mode"""'], {}), "(self, 'mode')\n", (182332, 182346), False, 'import pulumi\n'), ((182429, 182460), 'pulumi.set', 'pulumi.set', (['self', '"""mode"""', 'value'], {}), "(self, 'mode', value)\n", (182439, 182460), False, 'import pulumi\n'), ((183978, 184009), 'pulumi.get', 'pulumi.get', (['self', '"""description"""'], {}), "(self, 'description')\n", (183988, 184009), False, 'import pulumi\n'), ((184106, 184144), 'pulumi.set', 'pulumi.set', (['self', '"""description"""', 'value'], {}), "(self, 'description', value)\n", (184116, 184144), False, 'import pulumi\n'), ((184342, 184370), 'pulumi.get', 'pulumi.get', (['self', '"""image_id"""'], {}), "(self, 'image_id')\n", (184352, 184370), False, 'import pulumi\n'), ((184461, 184496), 'pulumi.set', 'pulumi.set', (['self', '"""image_id"""', 'value'], {}), "(self, 'image_id', value)\n", (184471, 184496), False, 'import pulumi\n'), ((184676, 184700), 'pulumi.get', 'pulumi.get', (['self', '"""size"""'], {}), "(self, 'size')\n", (184686, 184700), False, 'import pulumi\n'), ((184783, 184814), 'pulumi.set', 'pulumi.set', (['self', '"""size"""', 'value'], {}), "(self, 'size', value)\n", (184793, 184814), False, 'import pulumi\n'), ((185016, 185047), 'pulumi.get', 'pulumi.get', (['self', '"""snapshot_id"""'], {}), "(self, 'snapshot_id')\n", (185026, 185047), False, 'import pulumi\n'), ((185144, 185182), 'pulumi.set', 'pulumi.set', (['self', '"""snapshot_id"""', 'value'], {}), "(self, 'snapshot_id', value)\n", (185154, 185182), False, 'import pulumi\n'), ((185386, 185410), 'pulumi.get', 'pulumi.get', (['self', '"""type"""'], {}), "(self, 'type')\n", (185396, 185410), False, 'import pulumi\n'), ((185493, 185524), 'pulumi.set', 'pulumi.set', (['self', '"""type"""', 'value'], {}), "(self, 'type', value)\n", (185503, 185524), False, 'import pulumi\n'), ((189582, 189613), 'pulumi.get', 'pulumi.get', (['self', '"""dns_records"""'], {}), "(self, 'dns_records')\n", (189592, 189613), False, 'import pulumi\n'), ((189798, 189836), 'pulumi.set', 'pulumi.set', (['self', '"""dns_records"""', 'value'], {}), "(self, 'dns_records', value)\n", (189808, 189836), False, 'import pulumi\n'), ((190023, 190053), 'pulumi.get', 'pulumi.get', (['self', '"""ip_address"""'], {}), "(self, 'ip_address')\n", (190033, 190053), False, 'import pulumi\n'), ((190148, 190185), 'pulumi.set', 'pulumi.set', (['self', '"""ip_address"""', 'value'], {}), "(self, 'ip_address', value)\n", (190158, 190185), False, 'import pulumi\n'), ((190377, 190401), 'pulumi.get', 'pulumi.get', (['self', '"""ipv4"""'], {}), "(self, 'ipv4')\n", (190387, 190401), False, 'import pulumi\n'), ((190485, 190516), 'pulumi.set', 'pulumi.set', (['self', '"""ipv4"""', 'value'], {}), "(self, 'ipv4', value)\n", (190495, 190516), False, 'import pulumi\n'), ((190618, 190642), 'pulumi.get', 'pulumi.get', (['self', '"""ipv6"""'], {}), "(self, 'ipv6')\n", (190628, 190642), False, 'import pulumi\n'), ((190726, 190757), 'pulumi.set', 'pulumi.set', (['self', '"""ipv6"""', 'value'], {}), "(self, 'ipv6', value)\n", (190736, 190757), False, 'import pulumi\n'), ((190950, 190982), 'pulumi.get', 'pulumi.get', (['self', '"""ipv6_address"""'], {}), "(self, 'ipv6_address')\n", (190960, 190982), False, 'import pulumi\n'), ((191081, 191120), 'pulumi.set', 'pulumi.set', (['self', '"""ipv6_address"""', 'value'], {}), "(self, 'ipv6_address', value)\n", (191091, 191120), False, 'import pulumi\n'), ((191442, 191478), 'pulumi.get', 'pulumi.get', (['self', '"""ipv6_dns_records"""'], {}), "(self, 'ipv6_dns_records')\n", (191452, 191478), False, 'import pulumi\n'), ((191677, 191720), 'pulumi.set', 'pulumi.set', (['self', '"""ipv6_dns_records"""', 'value'], {}), "(self, 'ipv6_dns_records', value)\n", (191687, 191720), False, 'import pulumi\n'), ((191873, 191896), 'pulumi.get', 'pulumi.get', (['self', '"""nat"""'], {}), "(self, 'nat')\n", (191883, 191896), False, 'import pulumi\n'), ((191978, 192008), 'pulumi.set', 'pulumi.set', (['self', '"""nat"""', 'value'], {}), "(self, 'nat', value)\n", (191988, 192008), False, 'import pulumi\n'), ((192326, 192361), 'pulumi.get', 'pulumi.get', (['self', '"""nat_dns_records"""'], {}), "(self, 'nat_dns_records')\n", (192336, 192361), False, 'import pulumi\n'), ((192557, 192599), 'pulumi.set', 'pulumi.set', (['self', '"""nat_dns_records"""', 'value'], {}), "(self, 'nat_dns_records', value)\n", (192567, 192599), False, 'import pulumi\n'), ((192854, 192888), 'pulumi.get', 'pulumi.get', (['self', '"""nat_ip_address"""'], {}), "(self, 'nat_ip_address')\n", (192864, 192888), False, 'import pulumi\n'), ((192991, 193032), 'pulumi.set', 'pulumi.set', (['self', '"""nat_ip_address"""', 'value'], {}), "(self, 'nat_ip_address', value)\n", (193001, 193032), False, 'import pulumi\n'), ((193212, 193242), 'pulumi.get', 'pulumi.get', (['self', '"""network_id"""'], {}), "(self, 'network_id')\n", (193222, 193242), False, 'import pulumi\n'), ((193337, 193374), 'pulumi.set', 'pulumi.set', (['self', '"""network_id"""', 'value'], {}), "(self, 'network_id', value)\n", (193347, 193374), False, 'import pulumi\n'), ((193612, 193650), 'pulumi.get', 'pulumi.get', (['self', '"""security_group_ids"""'], {}), "(self, 'security_group_ids')\n", (193622, 193650), False, 'import pulumi\n'), ((193785, 193830), 'pulumi.set', 'pulumi.set', (['self', '"""security_group_ids"""', 'value'], {}), "(self, 'security_group_ids', value)\n", (193795, 193830), False, 'import pulumi\n'), ((194062, 194092), 'pulumi.get', 'pulumi.get', (['self', '"""subnet_ids"""'], {}), "(self, 'subnet_ids')\n", (194072, 194092), False, 'import pulumi\n'), ((194211, 194248), 'pulumi.set', 'pulumi.set', (['self', '"""subnet_ids"""', 'value'], {}), "(self, 'subnet_ids', value)\n", (194221, 194248), False, 'import pulumi\n'), ((194944, 194978), 'pulumi.set', 'pulumi.set', (['__self__', '"""fqdn"""', 'fqdn'], {}), "(__self__, 'fqdn', fqdn)\n", (194954, 194978), False, 'import pulumi\n'), ((195388, 195412), 'pulumi.get', 'pulumi.get', (['self', '"""fqdn"""'], {}), "(self, 'fqdn')\n", (195398, 195412), False, 'import pulumi\n'), ((195485, 195516), 'pulumi.set', 'pulumi.set', (['self', '"""fqdn"""', 'value'], {}), "(self, 'fqdn', value)\n", (195495, 195516), False, 'import pulumi\n'), ((195719, 195750), 'pulumi.get', 'pulumi.get', (['self', '"""dns_zone_id"""'], {}), "(self, 'dns_zone_id')\n", (195729, 195750), False, 'import pulumi\n'), ((195847, 195885), 'pulumi.set', 'pulumi.set', (['self', '"""dns_zone_id"""', 'value'], {}), "(self, 'dns_zone_id', value)\n", (195857, 195885), False, 'import pulumi\n'), ((196064, 196087), 'pulumi.get', 'pulumi.get', (['self', '"""ptr"""'], {}), "(self, 'ptr')\n", (196074, 196087), False, 'import pulumi\n'), ((196169, 196199), 'pulumi.set', 'pulumi.set', (['self', '"""ptr"""', 'value'], {}), "(self, 'ptr', value)\n", (196179, 196199), False, 'import pulumi\n'), ((196347, 196370), 'pulumi.get', 'pulumi.get', (['self', '"""ttl"""'], {}), "(self, 'ttl')\n", (196357, 196370), False, 'import pulumi\n'), ((196451, 196481), 'pulumi.set', 'pulumi.set', (['self', '"""ttl"""', 'value'], {}), "(self, 'ttl', value)\n", (196461, 196481), False, 'import pulumi\n'), ((197181, 197215), 'pulumi.set', 'pulumi.set', (['__self__', '"""fqdn"""', 'fqdn'], {}), "(__self__, 'fqdn', fqdn)\n", (197191, 197215), False, 'import pulumi\n'), ((197625, 197649), 'pulumi.get', 'pulumi.get', (['self', '"""fqdn"""'], {}), "(self, 'fqdn')\n", (197635, 197649), False, 'import pulumi\n'), ((197722, 197753), 'pulumi.set', 'pulumi.set', (['self', '"""fqdn"""', 'value'], {}), "(self, 'fqdn', value)\n", (197732, 197753), False, 'import pulumi\n'), ((197956, 197987), 'pulumi.get', 'pulumi.get', (['self', '"""dns_zone_id"""'], {}), "(self, 'dns_zone_id')\n", (197966, 197987), False, 'import pulumi\n'), ((198084, 198122), 'pulumi.set', 'pulumi.set', (['self', '"""dns_zone_id"""', 'value'], {}), "(self, 'dns_zone_id', value)\n", (198094, 198122), False, 'import pulumi\n'), ((198301, 198324), 'pulumi.get', 'pulumi.get', (['self', '"""ptr"""'], {}), "(self, 'ptr')\n", (198311, 198324), False, 'import pulumi\n'), ((198406, 198436), 'pulumi.set', 'pulumi.set', (['self', '"""ptr"""', 'value'], {}), "(self, 'ptr', value)\n", (198416, 198436), False, 'import pulumi\n'), ((198584, 198607), 'pulumi.get', 'pulumi.get', (['self', '"""ttl"""'], {}), "(self, 'ttl')\n", (198594, 198607), False, 'import pulumi\n'), ((198688, 198718), 'pulumi.set', 'pulumi.set', (['self', '"""ttl"""', 'value'], {}), "(self, 'ttl', value)\n", (198698, 198718), False, 'import pulumi\n'), ((199417, 199451), 'pulumi.set', 'pulumi.set', (['__self__', '"""fqdn"""', 'fqdn'], {}), "(__self__, 'fqdn', fqdn)\n", (199427, 199451), False, 'import pulumi\n'), ((199861, 199885), 'pulumi.get', 'pulumi.get', (['self', '"""fqdn"""'], {}), "(self, 'fqdn')\n", (199871, 199885), False, 'import pulumi\n'), ((199958, 199989), 'pulumi.set', 'pulumi.set', (['self', '"""fqdn"""', 'value'], {}), "(self, 'fqdn', value)\n", (199968, 199989), False, 'import pulumi\n'), ((200192, 200223), 'pulumi.get', 'pulumi.get', (['self', '"""dns_zone_id"""'], {}), "(self, 'dns_zone_id')\n", (200202, 200223), False, 'import pulumi\n'), ((200320, 200358), 'pulumi.set', 'pulumi.set', (['self', '"""dns_zone_id"""', 'value'], {}), "(self, 'dns_zone_id', value)\n", (200330, 200358), False, 'import pulumi\n'), ((200537, 200560), 'pulumi.get', 'pulumi.get', (['self', '"""ptr"""'], {}), "(self, 'ptr')\n", (200547, 200560), False, 'import pulumi\n'), ((200642, 200672), 'pulumi.set', 'pulumi.set', (['self', '"""ptr"""', 'value'], {}), "(self, 'ptr', value)\n", (200652, 200672), False, 'import pulumi\n'), ((200820, 200843), 'pulumi.get', 'pulumi.get', (['self', '"""ttl"""'], {}), "(self, 'ttl')\n", (200830, 200843), False, 'import pulumi\n'), ((200924, 200954), 'pulumi.set', 'pulumi.set', (['self', '"""ttl"""', 'value'], {}), "(self, 'ttl', value)\n", (200934, 200954), False, 'import pulumi\n'), ((201541, 201565), 'pulumi.get', 'pulumi.get', (['self', '"""type"""'], {}), "(self, 'type')\n", (201551, 201565), False, 'import pulumi\n'), ((201648, 201679), 'pulumi.set', 'pulumi.set', (['self', '"""type"""', 'value'], {}), "(self, 'type', value)\n", (201658, 201679), False, 'import pulumi\n'), ((202004, 202066), 'pulumi.set', 'pulumi.set', (['__self__', '"""placement_group_id"""', 'placement_group_id'], {}), "(__self__, 'placement_group_id', placement_group_id)\n", (202014, 202066), False, 'import pulumi\n'), ((202296, 202334), 'pulumi.get', 'pulumi.get', (['self', '"""placement_group_id"""'], {}), "(self, 'placement_group_id')\n", (202306, 202334), False, 'import pulumi\n'), ((202435, 202480), 'pulumi.set', 'pulumi.set', (['self', '"""placement_group_id"""', 'value'], {}), "(self, 'placement_group_id', value)\n", (202445, 202480), False, 'import pulumi\n'), ((203097, 203133), 'pulumi.set', 'pulumi.set', (['__self__', '"""cores"""', 'cores'], {}), "(__self__, 'cores', cores)\n", (203107, 203133), False, 'import pulumi\n'), ((203142, 203180), 'pulumi.set', 'pulumi.set', (['__self__', '"""memory"""', 'memory'], {}), "(__self__, 'memory', memory)\n", (203152, 203180), False, 'import pulumi\n'), ((203525, 203550), 'pulumi.get', 'pulumi.get', (['self', '"""cores"""'], {}), "(self, 'cores')\n", (203535, 203550), False, 'import pulumi\n'), ((203625, 203657), 'pulumi.set', 'pulumi.set', (['self', '"""cores"""', 'value'], {}), "(self, 'cores', value)\n", (203635, 203657), False, 'import pulumi\n'), ((203807, 203833), 'pulumi.get', 'pulumi.get', (['self', '"""memory"""'], {}), "(self, 'memory')\n", (203817, 203833), False, 'import pulumi\n'), ((203912, 203945), 'pulumi.set', 'pulumi.set', (['self', '"""memory"""', 'value'], {}), "(self, 'memory', value)\n", (203922, 203945), False, 'import pulumi\n'), ((204171, 204204), 'pulumi.get', 'pulumi.get', (['self', '"""core_fraction"""'], {}), "(self, 'core_fraction')\n", (204181, 204204), False, 'import pulumi\n'), ((204305, 204345), 'pulumi.set', 'pulumi.set', (['self', '"""core_fraction"""', 'value'], {}), "(self, 'core_fraction', value)\n", (204315, 204345), False, 'import pulumi\n'), ((204446, 204470), 'pulumi.get', 'pulumi.get', (['self', '"""gpus"""'], {}), "(self, 'gpus')\n", (204456, 204470), False, 'import pulumi\n'), ((204553, 204584), 'pulumi.set', 'pulumi.set', (['self', '"""gpus"""', 'value'], {}), "(self, 'gpus', value)\n", (204563, 204584), False, 'import pulumi\n'), ((205198, 205229), 'pulumi.get', 'pulumi.get', (['self', '"""preemptible"""'], {}), "(self, 'preemptible')\n", (205208, 205229), False, 'import pulumi\n'), ((205327, 205365), 'pulumi.set', 'pulumi.set', (['self', '"""preemptible"""', 'value'], {}), "(self, 'preemptible', value)\n", (205337, 205365), False, 'import pulumi\n'), ((206981, 207012), 'pulumi.get', 'pulumi.get', (['self', '"""device_name"""'], {}), "(self, 'device_name')\n", (206991, 207012), False, 'import pulumi\n'), ((207109, 207147), 'pulumi.set', 'pulumi.set', (['self', '"""device_name"""', 'value'], {}), "(self, 'device_name', value)\n", (207119, 207147), False, 'import pulumi\n'), ((207345, 207372), 'pulumi.get', 'pulumi.get', (['self', '"""disk_id"""'], {}), "(self, 'disk_id')\n", (207355, 207372), False, 'import pulumi\n'), ((207461, 207495), 'pulumi.set', 'pulumi.set', (['self', '"""disk_id"""', 'value'], {}), "(self, 'disk_id', value)\n", (207471, 207495), False, 'import pulumi\n'), ((207824, 207861), 'pulumi.get', 'pulumi.get', (['self', '"""initialize_params"""'], {}), "(self, 'initialize_params')\n", (207834, 207861), False, 'import pulumi\n'), ((208038, 208082), 'pulumi.set', 'pulumi.set', (['self', '"""initialize_params"""', 'value'], {}), "(self, 'initialize_params', value)\n", (208048, 208082), False, 'import pulumi\n'), ((208305, 208329), 'pulumi.get', 'pulumi.get', (['self', '"""mode"""'], {}), "(self, 'mode')\n", (208315, 208329), False, 'import pulumi\n'), ((208412, 208443), 'pulumi.set', 'pulumi.set', (['self', '"""mode"""', 'value'], {}), "(self, 'mode', value)\n", (208422, 208443), False, 'import pulumi\n'), ((209966, 209997), 'pulumi.get', 'pulumi.get', (['self', '"""description"""'], {}), "(self, 'description')\n", (209976, 209997), False, 'import pulumi\n'), ((210094, 210132), 'pulumi.set', 'pulumi.set', (['self', '"""description"""', 'value'], {}), "(self, 'description', value)\n", (210104, 210132), False, 'import pulumi\n'), ((210330, 210358), 'pulumi.get', 'pulumi.get', (['self', '"""image_id"""'], {}), "(self, 'image_id')\n", (210340, 210358), False, 'import pulumi\n'), ((210449, 210484), 'pulumi.set', 'pulumi.set', (['self', '"""image_id"""', 'value'], {}), "(self, 'image_id', value)\n", (210459, 210484), False, 'import pulumi\n'), ((210664, 210688), 'pulumi.get', 'pulumi.get', (['self', '"""size"""'], {}), "(self, 'size')\n", (210674, 210688), False, 'import pulumi\n'), ((210771, 210802), 'pulumi.set', 'pulumi.set', (['self', '"""size"""', 'value'], {}), "(self, 'size', value)\n", (210781, 210802), False, 'import pulumi\n'), ((211004, 211035), 'pulumi.get', 'pulumi.get', (['self', '"""snapshot_id"""'], {}), "(self, 'snapshot_id')\n", (211014, 211035), False, 'import pulumi\n'), ((211132, 211170), 'pulumi.set', 'pulumi.set', (['self', '"""snapshot_id"""', 'value'], {}), "(self, 'snapshot_id', value)\n", (211142, 211170), False, 'import pulumi\n'), ((211374, 211398), 'pulumi.get', 'pulumi.get', (['self', '"""type"""'], {}), "(self, 'type')\n", (211384, 211398), False, 'import pulumi\n'), ((211481, 211512), 'pulumi.set', 'pulumi.set', (['self', '"""type"""', 'value'], {}), "(self, 'type', value)\n", (211491, 211512), False, 'import pulumi\n'), ((213821, 213869), 'pulumi.get', 'pulumi.get', (['self', '"""max_opening_traffic_duration"""'], {}), "(self, 'max_opening_traffic_duration')\n", (213831, 213869), False, 'import pulumi\n'), ((214000, 214055), 'pulumi.set', 'pulumi.set', (['self', '"""max_opening_traffic_duration"""', 'value'], {}), "(self, 'max_opening_traffic_duration', value)\n", (214010, 214055), False, 'import pulumi\n'), ((214256, 214290), 'pulumi.get', 'pulumi.get', (['self', '"""status_message"""'], {}), "(self, 'status_message')\n", (214266, 214290), False, 'import pulumi\n'), ((214393, 214434), 'pulumi.set', 'pulumi.set', (['self', '"""status_message"""', 'value'], {}), "(self, 'status_message', value)\n", (214403, 214434), False, 'import pulumi\n'), ((214653, 214697), 'pulumi.get', 'pulumi.get', (['self', '"""target_group_description"""'], {}), "(self, 'target_group_description')\n", (214663, 214697), False, 'import pulumi\n'), ((214820, 214871), 'pulumi.set', 'pulumi.set', (['self', '"""target_group_description"""', 'value'], {}), "(self, 'target_group_description', value)\n", (214830, 214871), False, 'import pulumi\n'), ((215005, 215040), 'pulumi.get', 'pulumi.get', (['self', '"""target_group_id"""'], {}), "(self, 'target_group_id')\n", (215015, 215040), False, 'import pulumi\n'), ((215145, 215187), 'pulumi.set', 'pulumi.set', (['self', '"""target_group_id"""', 'value'], {}), "(self, 'target_group_id', value)\n", (215155, 215187), False, 'import pulumi\n'), ((215421, 215460), 'pulumi.get', 'pulumi.get', (['self', '"""target_group_labels"""'], {}), "(self, 'target_group_labels')\n", (215431, 215460), False, 'import pulumi\n'), ((215601, 215647), 'pulumi.set', 'pulumi.set', (['self', '"""target_group_labels"""', 'value'], {}), "(self, 'target_group_labels', value)\n", (215611, 215647), False, 'import pulumi\n'), ((215847, 215884), 'pulumi.get', 'pulumi.get', (['self', '"""target_group_name"""'], {}), "(self, 'target_group_name')\n", (215857, 215884), False, 'import pulumi\n'), ((215993, 216037), 'pulumi.set', 'pulumi.set', (['self', '"""target_group_name"""', 'value'], {}), "(self, 'target_group_name', value)\n", (216003, 216037), False, 'import pulumi\n'), ((217641, 217671), 'pulumi.get', 'pulumi.get', (['self', '"""auto_scale"""'], {}), "(self, 'auto_scale')\n", (217651, 217671), False, 'import pulumi\n'), ((217809, 217846), 'pulumi.set', 'pulumi.set', (['self', '"""auto_scale"""', 'value'], {}), "(self, 'auto_scale', value)\n", (217819, 217846), False, 'import pulumi\n'), ((218132, 218163), 'pulumi.get', 'pulumi.get', (['self', '"""fixed_scale"""'], {}), "(self, 'fixed_scale')\n", (218142, 218163), False, 'import pulumi\n'), ((218304, 218342), 'pulumi.set', 'pulumi.set', (['self', '"""fixed_scale"""', 'value'], {}), "(self, 'fixed_scale', value)\n", (218314, 218342), False, 'import pulumi\n'), ((218683, 218718), 'pulumi.get', 'pulumi.get', (['self', '"""test_auto_scale"""'], {}), "(self, 'test_auto_scale')\n", (218693, 218718), False, 'import pulumi\n'), ((218870, 218912), 'pulumi.set', 'pulumi.set', (['self', '"""test_auto_scale"""', 'value'], {}), "(self, 'test_auto_scale', value)\n", (218880, 218912), False, 'import pulumi\n'), ((221127, 221177), 'pulumi.set', 'pulumi.set', (['__self__', '"""initial_size"""', 'initial_size'], {}), "(__self__, 'initial_size', initial_size)\n", (221137, 221177), False, 'import pulumi\n'), ((221186, 221252), 'pulumi.set', 'pulumi.set', (['__self__', '"""measurement_duration"""', 'measurement_duration'], {}), "(__self__, 'measurement_duration', measurement_duration)\n", (221196, 221252), False, 'import pulumi\n'), ((222118, 222150), 'pulumi.get', 'pulumi.get', (['self', '"""initial_size"""'], {}), "(self, 'initial_size')\n", (222128, 222150), False, 'import pulumi\n'), ((222239, 222278), 'pulumi.set', 'pulumi.set', (['self', '"""initial_size"""', 'value'], {}), "(self, 'initial_size', value)\n", (222249, 222278), False, 'import pulumi\n'), ((222695, 222735), 'pulumi.get', 'pulumi.get', (['self', '"""measurement_duration"""'], {}), "(self, 'measurement_duration')\n", (222705, 222735), False, 'import pulumi\n'), ((222840, 222887), 'pulumi.set', 'pulumi.set', (['self', '"""measurement_duration"""', 'value'], {}), "(self, 'measurement_duration', value)\n", (222850, 222887), False, 'import pulumi\n'), ((223092, 223134), 'pulumi.get', 'pulumi.get', (['self', '"""cpu_utilization_target"""'], {}), "(self, 'cpu_utilization_target')\n", (223102, 223134), False, 'import pulumi\n'), ((223255, 223304), 'pulumi.set', 'pulumi.set', (['self', '"""cpu_utilization_target"""', 'value'], {}), "(self, 'cpu_utilization_target', value)\n", (223265, 223304), False, 'import pulumi\n'), ((223601, 223633), 'pulumi.get', 'pulumi.get', (['self', '"""custom_rules"""'], {}), "(self, 'custom_rules')\n", (223611, 223633), False, 'import pulumi\n'), ((223809, 223848), 'pulumi.set', 'pulumi.set', (['self', '"""custom_rules"""', 'value'], {}), "(self, 'custom_rules', value)\n", (223819, 223848), False, 'import pulumi\n'), ((224054, 224082), 'pulumi.get', 'pulumi.get', (['self', '"""max_size"""'], {}), "(self, 'max_size')\n", (224064, 224082), False, 'import pulumi\n'), ((224173, 224208), 'pulumi.set', 'pulumi.set', (['self', '"""max_size"""', 'value'], {}), "(self, 'max_size', value)\n", (224183, 224208), False, 'import pulumi\n'), ((224440, 224473), 'pulumi.get', 'pulumi.get', (['self', '"""min_zone_size"""'], {}), "(self, 'min_zone_size')\n", (224450, 224473), False, 'import pulumi\n'), ((224574, 224614), 'pulumi.set', 'pulumi.set', (['self', '"""min_zone_size"""', 'value'], {}), "(self, 'min_zone_size', value)\n", (224584, 224614), False, 'import pulumi\n'), ((225073, 225115), 'pulumi.get', 'pulumi.get', (['self', '"""stabilization_duration"""'], {}), "(self, 'stabilization_duration')\n", (225083, 225115), False, 'import pulumi\n'), ((225234, 225283), 'pulumi.set', 'pulumi.set', (['self', '"""stabilization_duration"""', 'value'], {}), "(self, 'stabilization_duration', value)\n", (225244, 225283), False, 'import pulumi\n'), ((225613, 225648), 'pulumi.get', 'pulumi.get', (['self', '"""warmup_duration"""'], {}), "(self, 'warmup_duration')\n", (225623, 225648), False, 'import pulumi\n'), ((225753, 225795), 'pulumi.set', 'pulumi.set', (['self', '"""warmup_duration"""', 'value'], {}), "(self, 'warmup_duration', value)\n", (225763, 225795), False, 'import pulumi\n'), ((227436, 227484), 'pulumi.set', 'pulumi.set', (['__self__', '"""metric_name"""', 'metric_name'], {}), "(__self__, 'metric_name', metric_name)\n", (227446, 227484), False, 'import pulumi\n'), ((227493, 227541), 'pulumi.set', 'pulumi.set', (['__self__', '"""metric_type"""', 'metric_type'], {}), "(__self__, 'metric_type', metric_type)\n", (227503, 227541), False, 'import pulumi\n'), ((227550, 227594), 'pulumi.set', 'pulumi.set', (['__self__', '"""rule_type"""', 'rule_type'], {}), "(__self__, 'rule_type', rule_type)\n", (227560, 227594), False, 'import pulumi\n'), ((227603, 227641), 'pulumi.set', 'pulumi.set', (['__self__', '"""target"""', 'target'], {}), "(__self__, 'target', target)\n", (227613, 227641), False, 'import pulumi\n'), ((228068, 228099), 'pulumi.get', 'pulumi.get', (['self', '"""metric_name"""'], {}), "(self, 'metric_name')\n", (228078, 228099), False, 'import pulumi\n'), ((228186, 228224), 'pulumi.set', 'pulumi.set', (['self', '"""metric_name"""', 'value'], {}), "(self, 'metric_name', value)\n", (228196, 228224), False, 'import pulumi\n'), ((228408, 228439), 'pulumi.get', 'pulumi.get', (['self', '"""metric_type"""'], {}), "(self, 'metric_type')\n", (228418, 228439), False, 'import pulumi\n'), ((228526, 228564), 'pulumi.set', 'pulumi.set', (['self', '"""metric_type"""', 'value'], {}), "(self, 'metric_type', value)\n", (228536, 228564), False, 'import pulumi\n'), ((229166, 229195), 'pulumi.get', 'pulumi.get', (['self', '"""rule_type"""'], {}), "(self, 'rule_type')\n", (229176, 229195), False, 'import pulumi\n'), ((229278, 229314), 'pulumi.set', 'pulumi.set', (['self', '"""rule_type"""', 'value'], {}), "(self, 'rule_type', value)\n", (229288, 229314), False, 'import pulumi\n'), ((229468, 229494), 'pulumi.get', 'pulumi.get', (['self', '"""target"""'], {}), "(self, 'target')\n", (229478, 229494), False, 'import pulumi\n'), ((229573, 229606), 'pulumi.set', 'pulumi.set', (['self', '"""target"""', 'value'], {}), "(self, 'target', value)\n", (229583, 229606), False, 'import pulumi\n'), ((229842, 229871), 'pulumi.get', 'pulumi.get', (['self', '"""folder_id"""'], {}), "(self, 'folder_id')\n", (229852, 229871), False, 'import pulumi\n'), ((229964, 230000), 'pulumi.set', 'pulumi.set', (['self', '"""folder_id"""', 'value'], {}), "(self, 'folder_id', value)\n", (229974, 230000), False, 'import pulumi\n'), ((230190, 230216), 'pulumi.get', 'pulumi.get', (['self', '"""labels"""'], {}), "(self, 'labels')\n", (230200, 230216), False, 'import pulumi\n'), ((230331, 230364), 'pulumi.set', 'pulumi.set', (['self', '"""labels"""', 'value'], {}), "(self, 'labels', value)\n", (230341, 230364), False, 'import pulumi\n'), ((230579, 230606), 'pulumi.get', 'pulumi.get', (['self', '"""service"""'], {}), "(self, 'service')\n", (230589, 230606), False, 'import pulumi\n'), ((230695, 230729), 'pulumi.set', 'pulumi.set', (['self', '"""service"""', 'value'], {}), "(self, 'service', value)\n", (230705, 230729), False, 'import pulumi\n'), ((230995, 231029), 'pulumi.set', 'pulumi.set', (['__self__', '"""size"""', 'size'], {}), "(__self__, 'size', size)\n", (231005, 231029), False, 'import pulumi\n'), ((231199, 231223), 'pulumi.get', 'pulumi.get', (['self', '"""size"""'], {}), "(self, 'size')\n", (231209, 231223), False, 'import pulumi\n'), ((231296, 231327), 'pulumi.set', 'pulumi.set', (['self', '"""size"""', 'value'], {}), "(self, 'size', value)\n", (231306, 231327), False, 'import pulumi\n'), ((233554, 233604), 'pulumi.set', 'pulumi.set', (['__self__', '"""initial_size"""', 'initial_size'], {}), "(__self__, 'initial_size', initial_size)\n", (233564, 233604), False, 'import pulumi\n'), ((233613, 233679), 'pulumi.set', 'pulumi.set', (['__self__', '"""measurement_duration"""', 'measurement_duration'], {}), "(__self__, 'measurement_duration', measurement_duration)\n", (233623, 233679), False, 'import pulumi\n'), ((234545, 234577), 'pulumi.get', 'pulumi.get', (['self', '"""initial_size"""'], {}), "(self, 'initial_size')\n", (234555, 234577), False, 'import pulumi\n'), ((234666, 234705), 'pulumi.set', 'pulumi.set', (['self', '"""initial_size"""', 'value'], {}), "(self, 'initial_size', value)\n", (234676, 234705), False, 'import pulumi\n'), ((235122, 235162), 'pulumi.get', 'pulumi.get', (['self', '"""measurement_duration"""'], {}), "(self, 'measurement_duration')\n", (235132, 235162), False, 'import pulumi\n'), ((235267, 235314), 'pulumi.set', 'pulumi.set', (['self', '"""measurement_duration"""', 'value'], {}), "(self, 'measurement_duration', value)\n", (235277, 235314), False, 'import pulumi\n'), ((235519, 235561), 'pulumi.get', 'pulumi.get', (['self', '"""cpu_utilization_target"""'], {}), "(self, 'cpu_utilization_target')\n", (235529, 235561), False, 'import pulumi\n'), ((235682, 235731), 'pulumi.set', 'pulumi.set', (['self', '"""cpu_utilization_target"""', 'value'], {}), "(self, 'cpu_utilization_target', value)\n", (235692, 235731), False, 'import pulumi\n'), ((236032, 236064), 'pulumi.get', 'pulumi.get', (['self', '"""custom_rules"""'], {}), "(self, 'custom_rules')\n", (236042, 236064), False, 'import pulumi\n'), ((236244, 236283), 'pulumi.set', 'pulumi.set', (['self', '"""custom_rules"""', 'value'], {}), "(self, 'custom_rules', value)\n", (236254, 236283), False, 'import pulumi\n'), ((236489, 236517), 'pulumi.get', 'pulumi.get', (['self', '"""max_size"""'], {}), "(self, 'max_size')\n", (236499, 236517), False, 'import pulumi\n'), ((236608, 236643), 'pulumi.set', 'pulumi.set', (['self', '"""max_size"""', 'value'], {}), "(self, 'max_size', value)\n", (236618, 236643), False, 'import pulumi\n'), ((236875, 236908), 'pulumi.get', 'pulumi.get', (['self', '"""min_zone_size"""'], {}), "(self, 'min_zone_size')\n", (236885, 236908), False, 'import pulumi\n'), ((237009, 237049), 'pulumi.set', 'pulumi.set', (['self', '"""min_zone_size"""', 'value'], {}), "(self, 'min_zone_size', value)\n", (237019, 237049), False, 'import pulumi\n'), ((237508, 237550), 'pulumi.get', 'pulumi.get', (['self', '"""stabilization_duration"""'], {}), "(self, 'stabilization_duration')\n", (237518, 237550), False, 'import pulumi\n'), ((237669, 237718), 'pulumi.set', 'pulumi.set', (['self', '"""stabilization_duration"""', 'value'], {}), "(self, 'stabilization_duration', value)\n", (237679, 237718), False, 'import pulumi\n'), ((238048, 238083), 'pulumi.get', 'pulumi.get', (['self', '"""warmup_duration"""'], {}), "(self, 'warmup_duration')\n", (238058, 238083), False, 'import pulumi\n'), ((238188, 238230), 'pulumi.set', 'pulumi.set', (['self', '"""warmup_duration"""', 'value'], {}), "(self, 'warmup_duration', value)\n", (238198, 238230), False, 'import pulumi\n'), ((239875, 239923), 'pulumi.set', 'pulumi.set', (['__self__', '"""metric_name"""', 'metric_name'], {}), "(__self__, 'metric_name', metric_name)\n", (239885, 239923), False, 'import pulumi\n'), ((239932, 239980), 'pulumi.set', 'pulumi.set', (['__self__', '"""metric_type"""', 'metric_type'], {}), "(__self__, 'metric_type', metric_type)\n", (239942, 239980), False, 'import pulumi\n'), ((239989, 240033), 'pulumi.set', 'pulumi.set', (['__self__', '"""rule_type"""', 'rule_type'], {}), "(__self__, 'rule_type', rule_type)\n", (239999, 240033), False, 'import pulumi\n'), ((240042, 240080), 'pulumi.set', 'pulumi.set', (['__self__', '"""target"""', 'target'], {}), "(__self__, 'target', target)\n", (240052, 240080), False, 'import pulumi\n'), ((240507, 240538), 'pulumi.get', 'pulumi.get', (['self', '"""metric_name"""'], {}), "(self, 'metric_name')\n", (240517, 240538), False, 'import pulumi\n'), ((240625, 240663), 'pulumi.set', 'pulumi.set', (['self', '"""metric_name"""', 'value'], {}), "(self, 'metric_name', value)\n", (240635, 240663), False, 'import pulumi\n'), ((240847, 240878), 'pulumi.get', 'pulumi.get', (['self', '"""metric_type"""'], {}), "(self, 'metric_type')\n", (240857, 240878), False, 'import pulumi\n'), ((240965, 241003), 'pulumi.set', 'pulumi.set', (['self', '"""metric_type"""', 'value'], {}), "(self, 'metric_type', value)\n", (240975, 241003), False, 'import pulumi\n'), ((241605, 241634), 'pulumi.get', 'pulumi.get', (['self', '"""rule_type"""'], {}), "(self, 'rule_type')\n", (241615, 241634), False, 'import pulumi\n'), ((241717, 241753), 'pulumi.set', 'pulumi.set', (['self', '"""rule_type"""', 'value'], {}), "(self, 'rule_type', value)\n", (241727, 241753), False, 'import pulumi\n'), ((241907, 241933), 'pulumi.get', 'pulumi.get', (['self', '"""target"""'], {}), "(self, 'target')\n", (241917, 241933), False, 'import pulumi\n'), ((242012, 242045), 'pulumi.set', 'pulumi.set', (['self', '"""target"""', 'value'], {}), "(self, 'target', value)\n", (242022, 242045), False, 'import pulumi\n'), ((242281, 242310), 'pulumi.get', 'pulumi.get', (['self', '"""folder_id"""'], {}), "(self, 'folder_id')\n", (242291, 242310), False, 'import pulumi\n'), ((242403, 242439), 'pulumi.set', 'pulumi.set', (['self', '"""folder_id"""', 'value'], {}), "(self, 'folder_id', value)\n", (242413, 242439), False, 'import pulumi\n'), ((242629, 242655), 'pulumi.get', 'pulumi.get', (['self', '"""labels"""'], {}), "(self, 'labels')\n", (242639, 242655), False, 'import pulumi\n'), ((242770, 242803), 'pulumi.set', 'pulumi.set', (['self', '"""labels"""', 'value'], {}), "(self, 'labels', value)\n", (242780, 242803), False, 'import pulumi\n'), ((243018, 243045), 'pulumi.get', 'pulumi.get', (['self', '"""service"""'], {}), "(self, 'service')\n", (243028, 243045), False, 'import pulumi\n'), ((243134, 243168), 'pulumi.set', 'pulumi.set', (['self', '"""service"""', 'value'], {}), "(self, 'service', value)\n", (243144, 243168), False, 'import pulumi\n'), ((246231, 246275), 'pulumi.set', 'pulumi.set', (['__self__', '"""subnet_id"""', 'subnet_id'], {}), "(__self__, 'subnet_id', subnet_id)\n", (246241, 246275), False, 'import pulumi\n'), ((247801, 247830), 'pulumi.get', 'pulumi.get', (['self', '"""subnet_id"""'], {}), "(self, 'subnet_id')\n", (247811, 247830), False, 'import pulumi\n'), ((247913, 247949), 'pulumi.set', 'pulumi.set', (['self', '"""subnet_id"""', 'value'], {}), "(self, 'subnet_id', value)\n", (247923, 247949), False, 'import pulumi\n'), ((248264, 248295), 'pulumi.get', 'pulumi.get', (['self', '"""dns_records"""'], {}), "(self, 'dns_records')\n", (248274, 248295), False, 'import pulumi\n'), ((248459, 248497), 'pulumi.set', 'pulumi.set', (['self', '"""dns_records"""', 'value'], {}), "(self, 'dns_records', value)\n", (248469, 248497), False, 'import pulumi\n'), ((248599, 248624), 'pulumi.get', 'pulumi.get', (['self', '"""index"""'], {}), "(self, 'index')\n", (248609, 248624), False, 'import pulumi\n'), ((248709, 248741), 'pulumi.set', 'pulumi.set', (['self', '"""index"""', 'value'], {}), "(self, 'index', value)\n", (248719, 248741), False, 'import pulumi\n'), ((249036, 249066), 'pulumi.get', 'pulumi.get', (['self', '"""ip_address"""'], {}), "(self, 'ip_address')\n", (249046, 249066), False, 'import pulumi\n'), ((249161, 249198), 'pulumi.set', 'pulumi.set', (['self', '"""ip_address"""', 'value'], {}), "(self, 'ip_address', value)\n", (249171, 249198), False, 'import pulumi\n'), ((249405, 249429), 'pulumi.get', 'pulumi.get', (['self', '"""ipv4"""'], {}), "(self, 'ipv4')\n", (249415, 249429), False, 'import pulumi\n'), ((249513, 249544), 'pulumi.set', 'pulumi.set', (['self', '"""ipv4"""', 'value'], {}), "(self, 'ipv4', value)\n", (249523, 249544), False, 'import pulumi\n'), ((249809, 249833), 'pulumi.get', 'pulumi.get', (['self', '"""ipv6"""'], {}), "(self, 'ipv6')\n", (249819, 249833), False, 'import pulumi\n'), ((249917, 249948), 'pulumi.set', 'pulumi.set', (['self', '"""ipv6"""', 'value'], {}), "(self, 'ipv6', value)\n", (249927, 249948), False, 'import pulumi\n'), ((250161, 250193), 'pulumi.get', 'pulumi.get', (['self', '"""ipv6_address"""'], {}), "(self, 'ipv6_address')\n", (250171, 250193), False, 'import pulumi\n'), ((250292, 250331), 'pulumi.set', 'pulumi.set', (['self', '"""ipv6_address"""', 'value'], {}), "(self, 'ipv6_address', value)\n", (250302, 250331), False, 'import pulumi\n'), ((250659, 250695), 'pulumi.get', 'pulumi.get', (['self', '"""ipv6_dns_records"""'], {}), "(self, 'ipv6_dns_records')\n", (250669, 250695), False, 'import pulumi\n'), ((250873, 250916), 'pulumi.set', 'pulumi.set', (['self', '"""ipv6_dns_records"""', 'value'], {}), "(self, 'ipv6_dns_records', value)\n", (250883, 250916), False, 'import pulumi\n'), ((251043, 251074), 'pulumi.get', 'pulumi.get', (['self', '"""mac_address"""'], {}), "(self, 'mac_address')\n", (251053, 251074), False, 'import pulumi\n'), ((251171, 251209), 'pulumi.set', 'pulumi.set', (['self', '"""mac_address"""', 'value'], {}), "(self, 'mac_address', value)\n", (251181, 251209), False, 'import pulumi\n'), ((251415, 251438), 'pulumi.get', 'pulumi.get', (['self', '"""nat"""'], {}), "(self, 'nat')\n", (251425, 251438), False, 'import pulumi\n'), ((251520, 251550), 'pulumi.set', 'pulumi.set', (['self', '"""nat"""', 'value'], {}), "(self, 'nat', value)\n", (251530, 251550), False, 'import pulumi\n'), ((251879, 251914), 'pulumi.get', 'pulumi.get', (['self', '"""nat_dns_records"""'], {}), "(self, 'nat_dns_records')\n", (251889, 251914), False, 'import pulumi\n'), ((252089, 252131), 'pulumi.set', 'pulumi.set', (['self', '"""nat_dns_records"""', 'value'], {}), "(self, 'nat_dns_records', value)\n", (252099, 252131), False, 'import pulumi\n'), ((252414, 252448), 'pulumi.get', 'pulumi.get', (['self', '"""nat_ip_address"""'], {}), "(self, 'nat_ip_address')\n", (252424, 252448), False, 'import pulumi\n'), ((252551, 252592), 'pulumi.set', 'pulumi.set', (['self', '"""nat_ip_address"""', 'value'], {}), "(self, 'nat_ip_address', value)\n", (252561, 252592), False, 'import pulumi\n'), ((252724, 252758), 'pulumi.get', 'pulumi.get', (['self', '"""nat_ip_version"""'], {}), "(self, 'nat_ip_version')\n", (252734, 252758), False, 'import pulumi\n'), ((252861, 252902), 'pulumi.set', 'pulumi.set', (['self', '"""nat_ip_version"""', 'value'], {}), "(self, 'nat_ip_version', value)\n", (252871, 252902), False, 'import pulumi\n'), ((253140, 253178), 'pulumi.get', 'pulumi.get', (['self', '"""security_group_ids"""'], {}), "(self, 'security_group_ids')\n", (253150, 253178), False, 'import pulumi\n'), ((253313, 253358), 'pulumi.set', 'pulumi.set', (['self', '"""security_group_ids"""', 'value'], {}), "(self, 'security_group_ids', value)\n", (253323, 253358), False, 'import pulumi\n'), ((254048, 254082), 'pulumi.set', 'pulumi.set', (['__self__', '"""fqdn"""', 'fqdn'], {}), "(__self__, 'fqdn', fqdn)\n", (254058, 254082), False, 'import pulumi\n'), ((254494, 254518), 'pulumi.get', 'pulumi.get', (['self', '"""fqdn"""'], {}), "(self, 'fqdn')\n", (254504, 254518), False, 'import pulumi\n'), ((254591, 254622), 'pulumi.set', 'pulumi.set', (['self', '"""fqdn"""', 'value'], {}), "(self, 'fqdn', value)\n", (254601, 254622), False, 'import pulumi\n'), ((254825, 254856), 'pulumi.get', 'pulumi.get', (['self', '"""dns_zone_id"""'], {}), "(self, 'dns_zone_id')\n", (254835, 254856), False, 'import pulumi\n'), ((254953, 254991), 'pulumi.set', 'pulumi.set', (['self', '"""dns_zone_id"""', 'value'], {}), "(self, 'dns_zone_id', value)\n", (254963, 254991), False, 'import pulumi\n'), ((255172, 255195), 'pulumi.get', 'pulumi.get', (['self', '"""ptr"""'], {}), "(self, 'ptr')\n", (255182, 255195), False, 'import pulumi\n'), ((255277, 255307), 'pulumi.set', 'pulumi.set', (['self', '"""ptr"""', 'value'], {}), "(self, 'ptr', value)\n", (255287, 255307), False, 'import pulumi\n'), ((255466, 255489), 'pulumi.get', 'pulumi.get', (['self', '"""ttl"""'], {}), "(self, 'ttl')\n", (255476, 255489), False, 'import pulumi\n'), ((255570, 255600), 'pulumi.set', 'pulumi.set', (['self', '"""ttl"""', 'value'], {}), "(self, 'ttl', value)\n", (255580, 255600), False, 'import pulumi\n'), ((256294, 256328), 'pulumi.set', 'pulumi.set', (['__self__', '"""fqdn"""', 'fqdn'], {}), "(__self__, 'fqdn', fqdn)\n", (256304, 256328), False, 'import pulumi\n'), ((256740, 256764), 'pulumi.get', 'pulumi.get', (['self', '"""fqdn"""'], {}), "(self, 'fqdn')\n", (256750, 256764), False, 'import pulumi\n'), ((256837, 256868), 'pulumi.set', 'pulumi.set', (['self', '"""fqdn"""', 'value'], {}), "(self, 'fqdn', value)\n", (256847, 256868), False, 'import pulumi\n'), ((257071, 257102), 'pulumi.get', 'pulumi.get', (['self', '"""dns_zone_id"""'], {}), "(self, 'dns_zone_id')\n", (257081, 257102), False, 'import pulumi\n'), ((257199, 257237), 'pulumi.set', 'pulumi.set', (['self', '"""dns_zone_id"""', 'value'], {}), "(self, 'dns_zone_id', value)\n", (257209, 257237), False, 'import pulumi\n'), ((257418, 257441), 'pulumi.get', 'pulumi.get', (['self', '"""ptr"""'], {}), "(self, 'ptr')\n", (257428, 257441), False, 'import pulumi\n'), ((257523, 257553), 'pulumi.set', 'pulumi.set', (['self', '"""ptr"""', 'value'], {}), "(self, 'ptr', value)\n", (257533, 257553), False, 'import pulumi\n'), ((257712, 257735), 'pulumi.get', 'pulumi.get', (['self', '"""ttl"""'], {}), "(self, 'ttl')\n", (257722, 257735), False, 'import pulumi\n'), ((257816, 257846), 'pulumi.set', 'pulumi.set', (['self', '"""ttl"""', 'value'], {}), "(self, 'ttl', value)\n", (257826, 257846), False, 'import pulumi\n'), ((258539, 258573), 'pulumi.set', 'pulumi.set', (['__self__', '"""fqdn"""', 'fqdn'], {}), "(__self__, 'fqdn', fqdn)\n", (258549, 258573), False, 'import pulumi\n'), ((258985, 259009), 'pulumi.get', 'pulumi.get', (['self', '"""fqdn"""'], {}), "(self, 'fqdn')\n", (258995, 259009), False, 'import pulumi\n'), ((259082, 259113), 'pulumi.set', 'pulumi.set', (['self', '"""fqdn"""', 'value'], {}), "(self, 'fqdn', value)\n", (259092, 259113), False, 'import pulumi\n'), ((259316, 259347), 'pulumi.get', 'pulumi.get', (['self', '"""dns_zone_id"""'], {}), "(self, 'dns_zone_id')\n", (259326, 259347), False, 'import pulumi\n'), ((259444, 259482), 'pulumi.set', 'pulumi.set', (['self', '"""dns_zone_id"""', 'value'], {}), "(self, 'dns_zone_id', value)\n", (259454, 259482), False, 'import pulumi\n'), ((259663, 259686), 'pulumi.get', 'pulumi.get', (['self', '"""ptr"""'], {}), "(self, 'ptr')\n", (259673, 259686), False, 'import pulumi\n'), ((259768, 259798), 'pulumi.set', 'pulumi.set', (['self', '"""ptr"""', 'value'], {}), "(self, 'ptr', value)\n", (259778, 259798), False, 'import pulumi\n'), ((259957, 259980), 'pulumi.get', 'pulumi.get', (['self', '"""ttl"""'], {}), "(self, 'ttl')\n", (259967, 259980), False, 'import pulumi\n'), ((260061, 260091), 'pulumi.set', 'pulumi.set', (['self', '"""ttl"""', 'value'], {}), "(self, 'ttl', value)\n", (260071, 260091), False, 'import pulumi\n'), ((260394, 260456), 'pulumi.set', 'pulumi.set', (['__self__', '"""placement_group_id"""', 'placement_group_id'], {}), "(__self__, 'placement_group_id', placement_group_id)\n", (260404, 260456), False, 'import pulumi\n'), ((260685, 260723), 'pulumi.get', 'pulumi.get', (['self', '"""placement_group_id"""'], {}), "(self, 'placement_group_id')\n", (260695, 260723), False, 'import pulumi\n'), ((260824, 260869), 'pulumi.set', 'pulumi.set', (['self', '"""placement_group_id"""', 'value'], {}), "(self, 'placement_group_id', value)\n", (260834, 260869), False, 'import pulumi\n'), ((261453, 261489), 'pulumi.set', 'pulumi.set', (['__self__', '"""cores"""', 'cores'], {}), "(__self__, 'cores', cores)\n", (261463, 261489), False, 'import pulumi\n'), ((261498, 261536), 'pulumi.set', 'pulumi.set', (['__self__', '"""memory"""', 'memory'], {}), "(__self__, 'memory', memory)\n", (261508, 261536), False, 'import pulumi\n'), ((261867, 261892), 'pulumi.get', 'pulumi.get', (['self', '"""cores"""'], {}), "(self, 'cores')\n", (261877, 261892), False, 'import pulumi\n'), ((261967, 261999), 'pulumi.set', 'pulumi.set', (['self', '"""cores"""', 'value'], {}), "(self, 'cores', value)\n", (261977, 261999), False, 'import pulumi\n'), ((262145, 262171), 'pulumi.get', 'pulumi.get', (['self', '"""memory"""'], {}), "(self, 'memory')\n", (262155, 262171), False, 'import pulumi\n'), ((262250, 262283), 'pulumi.set', 'pulumi.set', (['self', '"""memory"""', 'value'], {}), "(self, 'memory', value)\n", (262260, 262283), False, 'import pulumi\n'), ((262515, 262548), 'pulumi.get', 'pulumi.get', (['self', '"""core_fraction"""'], {}), "(self, 'core_fraction')\n", (262525, 262548), False, 'import pulumi\n'), ((262649, 262689), 'pulumi.set', 'pulumi.set', (['self', '"""core_fraction"""', 'value'], {}), "(self, 'core_fraction', value)\n", (262659, 262689), False, 'import pulumi\n'), ((262790, 262814), 'pulumi.get', 'pulumi.get', (['self', '"""gpus"""'], {}), "(self, 'gpus')\n", (262800, 262814), False, 'import pulumi\n'), ((262897, 262928), 'pulumi.set', 'pulumi.set', (['self', '"""gpus"""', 'value'], {}), "(self, 'gpus', value)\n", (262907, 262928), False, 'import pulumi\n'), ((263521, 263552), 'pulumi.get', 'pulumi.get', (['self', '"""preemptible"""'], {}), "(self, 'preemptible')\n", (263531, 263552), False, 'import pulumi\n'), ((263650, 263688), 'pulumi.set', 'pulumi.set', (['self', '"""preemptible"""', 'value'], {}), "(self, 'preemptible', value)\n", (263660, 263688), False, 'import pulumi\n'), ((264559, 264599), 'pulumi.set', 'pulumi.set', (['__self__', '"""disk_id"""', 'disk_id'], {}), "(__self__, 'disk_id', disk_id)\n", (264569, 264599), False, 'import pulumi\n'), ((265059, 265086), 'pulumi.get', 'pulumi.get', (['self', '"""disk_id"""'], {}), "(self, 'disk_id')\n", (265069, 265086), False, 'import pulumi\n'), ((265165, 265199), 'pulumi.set', 'pulumi.set', (['self', '"""disk_id"""', 'value'], {}), "(self, 'disk_id', value)\n", (265175, 265199), False, 'import pulumi\n'), ((265458, 265489), 'pulumi.get', 'pulumi.get', (['self', '"""auto_delete"""'], {}), "(self, 'auto_delete')\n", (265468, 265489), False, 'import pulumi\n'), ((265587, 265625), 'pulumi.set', 'pulumi.set', (['self', '"""auto_delete"""', 'value'], {}), "(self, 'auto_delete', value)\n", (265597, 265625), False, 'import pulumi\n'), ((265867, 265898), 'pulumi.get', 'pulumi.get', (['self', '"""device_name"""'], {}), "(self, 'device_name')\n", (265877, 265898), False, 'import pulumi\n'), ((265995, 266033), 'pulumi.set', 'pulumi.set', (['self', '"""device_name"""', 'value'], {}), "(self, 'device_name', value)\n", (266005, 266033), False, 'import pulumi\n'), ((266256, 266280), 'pulumi.get', 'pulumi.get', (['self', '"""mode"""'], {}), "(self, 'mode')\n", (266266, 266280), False, 'import pulumi\n'), ((266363, 266394), 'pulumi.set', 'pulumi.set', (['self', '"""mode"""', 'value'], {}), "(self, 'mode', value)\n", (266373, 266394), False, 'import pulumi\n'), ((267217, 267275), 'pulumi.set', 'pulumi.set', (['__self__', '"""subcluster_specs"""', 'subcluster_specs'], {}), "(__self__, 'subcluster_specs', subcluster_specs)\n", (267227, 267275), False, 'import pulumi\n'), ((267757, 267793), 'pulumi.get', 'pulumi.get', (['self', '"""subcluster_specs"""'], {}), "(self, 'subcluster_specs')\n", (267767, 267793), False, 'import pulumi\n'), ((267959, 268002), 'pulumi.set', 'pulumi.set', (['self', '"""subcluster_specs"""', 'value'], {}), "(self, 'subcluster_specs', value)\n", (267969, 268002), False, 'import pulumi\n'), ((268237, 268263), 'pulumi.get', 'pulumi.get', (['self', '"""hadoop"""'], {}), "(self, 'hadoop')\n", (268247, 268263), False, 'import pulumi\n'), ((268387, 268420), 'pulumi.set', 'pulumi.set', (['self', '"""hadoop"""', 'value'], {}), "(self, 'hadoop', value)\n", (268397, 268420), False, 'import pulumi\n'), ((268605, 268635), 'pulumi.get', 'pulumi.get', (['self', '"""version_id"""'], {}), "(self, 'version_id')\n", (268615, 268635), False, 'import pulumi\n'), ((268730, 268767), 'pulumi.set', 'pulumi.set', (['self', '"""version_id"""', 'value'], {}), "(self, 'version_id', value)\n", (268740, 268767), False, 'import pulumi\n'), ((270225, 270255), 'pulumi.get', 'pulumi.get', (['self', '"""properties"""'], {}), "(self, 'properties')\n", (270235, 270255), False, 'import pulumi\n'), ((270378, 270415), 'pulumi.set', 'pulumi.set', (['self', '"""properties"""', 'value'], {}), "(self, 'properties', value)\n", (270388, 270415), False, 'import pulumi\n'), ((270622, 270650), 'pulumi.get', 'pulumi.get', (['self', '"""services"""'], {}), "(self, 'services')\n", (270632, 270650), False, 'import pulumi\n'), ((270765, 270800), 'pulumi.set', 'pulumi.set', (['self', '"""services"""', 'value'], {}), "(self, 'services', value)\n", (270775, 270800), False, 'import pulumi\n'), ((271194, 271229), 'pulumi.get', 'pulumi.get', (['self', '"""ssh_public_keys"""'], {}), "(self, 'ssh_public_keys')\n", (271204, 271229), False, 'import pulumi\n'), ((271358, 271400), 'pulumi.set', 'pulumi.set', (['self', '"""ssh_public_keys"""', 'value'], {}), "(self, 'ssh_public_keys', value)\n", (271368, 271400), False, 'import pulumi\n'), ((272879, 272927), 'pulumi.set', 'pulumi.set', (['__self__', '"""hosts_count"""', 'hosts_count'], {}), "(__self__, 'hosts_count', hosts_count)\n", (272889, 272927), False, 'import pulumi\n'), ((272936, 272970), 'pulumi.set', 'pulumi.set', (['__self__', '"""name"""', 'name'], {}), "(__self__, 'name', name)\n", (272946, 272970), False, 'import pulumi\n'), ((272979, 273023), 'pulumi.set', 'pulumi.set', (['__self__', '"""resources"""', 'resources'], {}), "(__self__, 'resources', resources)\n", (272989, 273023), False, 'import pulumi\n'), ((273032, 273066), 'pulumi.set', 'pulumi.set', (['__self__', '"""role"""', 'role'], {}), "(__self__, 'role', role)\n", (273042, 273066), False, 'import pulumi\n'), ((273075, 273119), 'pulumi.set', 'pulumi.set', (['__self__', '"""subnet_id"""', 'subnet_id'], {}), "(__self__, 'subnet_id', subnet_id)\n", (273085, 273119), False, 'import pulumi\n'), ((273501, 273532), 'pulumi.get', 'pulumi.get', (['self', '"""hosts_count"""'], {}), "(self, 'hosts_count')\n", (273511, 273532), False, 'import pulumi\n'), ((273619, 273657), 'pulumi.set', 'pulumi.set', (['self', '"""hosts_count"""', 'value'], {}), "(self, 'hosts_count', value)\n", (273629, 273657), False, 'import pulumi\n'), ((273814, 273838), 'pulumi.get', 'pulumi.get', (['self', '"""name"""'], {}), "(self, 'name')\n", (273824, 273838), False, 'import pulumi\n'), ((273911, 273942), 'pulumi.set', 'pulumi.set', (['self', '"""name"""', 'value'], {}), "(self, 'name', value)\n", (273921, 273942), False, 'import pulumi\n'), ((274221, 274250), 'pulumi.get', 'pulumi.get', (['self', '"""resources"""'], {}), "(self, 'resources')\n", (274231, 274250), False, 'import pulumi\n'), ((274387, 274423), 'pulumi.set', 'pulumi.set', (['self', '"""resources"""', 'value'], {}), "(self, 'resources', value)\n", (274397, 274423), False, 'import pulumi\n'), ((274595, 274619), 'pulumi.get', 'pulumi.get', (['self', '"""role"""'], {}), "(self, 'role')\n", (274605, 274619), False, 'import pulumi\n'), ((274692, 274723), 'pulumi.set', 'pulumi.set', (['self', '"""role"""', 'value'], {}), "(self, 'role', value)\n", (274702, 274723), False, 'import pulumi\n'), ((274999, 275028), 'pulumi.get', 'pulumi.get', (['self', '"""subnet_id"""'], {}), "(self, 'subnet_id')\n", (275009, 275028), False, 'import pulumi\n'), ((275111, 275147), 'pulumi.set', 'pulumi.set', (['self', '"""subnet_id"""', 'value'], {}), "(self, 'subnet_id', value)\n", (275121, 275147), False, 'import pulumi\n'), ((275433, 275471), 'pulumi.get', 'pulumi.get', (['self', '"""autoscaling_config"""'], {}), "(self, 'autoscaling_config')\n", (275443, 275471), False, 'import pulumi\n'), ((275644, 275689), 'pulumi.set', 'pulumi.set', (['self', '"""autoscaling_config"""', 'value'], {}), "(self, 'autoscaling_config', value)\n", (275654, 275689), False, 'import pulumi\n'), ((275862, 275884), 'pulumi.get', 'pulumi.get', (['self', '"""id"""'], {}), "(self, 'id')\n", (275872, 275884), False, 'import pulumi\n'), ((275963, 275992), 'pulumi.set', 'pulumi.set', (['self', '"""id"""', 'value'], {}), "(self, 'id', value)\n", (275973, 275992), False, 'import pulumi\n'), ((278018, 278074), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_hosts_count"""', 'max_hosts_count'], {}), "(__self__, 'max_hosts_count', max_hosts_count)\n", (278028, 278074), False, 'import pulumi\n'), ((278996, 279031), 'pulumi.get', 'pulumi.get', (['self', '"""max_hosts_count"""'], {}), "(self, 'max_hosts_count')\n", (279006, 279031), False, 'import pulumi\n'), ((279126, 279168), 'pulumi.set', 'pulumi.set', (['self', '"""max_hosts_count"""', 'value'], {}), "(self, 'max_hosts_count', value)\n", (279136, 279168), False, 'import pulumi\n'), ((279490, 279532), 'pulumi.get', 'pulumi.get', (['self', '"""cpu_utilization_target"""'], {}), "(self, 'cpu_utilization_target')\n", (279500, 279532), False, 'import pulumi\n'), ((279653, 279702), 'pulumi.set', 'pulumi.set', (['self', '"""cpu_utilization_target"""', 'value'], {}), "(self, 'cpu_utilization_target', value)\n", (279663, 279702), False, 'import pulumi\n'), ((279952, 279992), 'pulumi.get', 'pulumi.get', (['self', '"""decommission_timeout"""'], {}), "(self, 'decommission_timeout')\n", (279962, 279992), False, 'import pulumi\n'), ((280107, 280154), 'pulumi.set', 'pulumi.set', (['self', '"""decommission_timeout"""', 'value'], {}), "(self, 'decommission_timeout', value)\n", (280117, 280154), False, 'import pulumi\n'), ((280379, 280419), 'pulumi.get', 'pulumi.get', (['self', '"""measurement_duration"""'], {}), "(self, 'measurement_duration')\n", (280389, 280419), False, 'import pulumi\n'), ((280534, 280581), 'pulumi.set', 'pulumi.set', (['self', '"""measurement_duration"""', 'value'], {}), "(self, 'measurement_duration', value)\n", (280544, 280581), False, 'import pulumi\n'), ((281039, 281070), 'pulumi.get', 'pulumi.get', (['self', '"""preemptible"""'], {}), "(self, 'preemptible')\n", (281049, 281070), False, 'import pulumi\n'), ((281168, 281206), 'pulumi.set', 'pulumi.set', (['self', '"""preemptible"""', 'value'], {}), "(self, 'preemptible', value)\n", (281178, 281206), False, 'import pulumi\n'), ((281623, 281665), 'pulumi.get', 'pulumi.get', (['self', '"""stabilization_duration"""'], {}), "(self, 'stabilization_duration')\n", (281633, 281665), False, 'import pulumi\n'), ((281784, 281833), 'pulumi.set', 'pulumi.set', (['self', '"""stabilization_duration"""', 'value'], {}), "(self, 'stabilization_duration', value)\n", (281794, 281833), False, 'import pulumi\n'), ((282135, 282170), 'pulumi.get', 'pulumi.get', (['self', '"""warmup_duration"""'], {}), "(self, 'warmup_duration')\n", (282145, 282170), False, 'import pulumi\n'), ((282275, 282317), 'pulumi.set', 'pulumi.set', (['self', '"""warmup_duration"""', 'value'], {}), "(self, 'warmup_duration', value)\n", (282285, 282317), False, 'import pulumi\n'), ((283108, 283152), 'pulumi.set', 'pulumi.set', (['__self__', '"""disk_size"""', 'disk_size'], {}), "(__self__, 'disk_size', disk_size)\n", (283118, 283152), False, 'import pulumi\n'), ((283161, 283223), 'pulumi.set', 'pulumi.set', (['__self__', '"""resource_preset_id"""', 'resource_preset_id'], {}), "(__self__, 'resource_preset_id', resource_preset_id)\n", (283171, 283223), False, 'import pulumi\n'), ((283525, 283554), 'pulumi.get', 'pulumi.get', (['self', '"""disk_size"""'], {}), "(self, 'disk_size')\n", (283535, 283554), False, 'import pulumi\n'), ((283637, 283673), 'pulumi.set', 'pulumi.set', (['self', '"""disk_size"""', 'value'], {}), "(self, 'disk_size', value)\n", (283647, 283673), False, 'import pulumi\n'), ((284027, 284065), 'pulumi.get', 'pulumi.get', (['self', '"""resource_preset_id"""'], {}), "(self, 'resource_preset_id')\n", (284037, 284065), False, 'import pulumi\n'), ((284166, 284211), 'pulumi.set', 'pulumi.set', (['self', '"""resource_preset_id"""', 'value'], {}), "(self, 'resource_preset_id', value)\n", (284176, 284211), False, 'import pulumi\n'), ((284451, 284483), 'pulumi.get', 'pulumi.get', (['self', '"""disk_type_id"""'], {}), "(self, 'disk_type_id')\n", (284461, 284483), False, 'import pulumi\n'), ((284582, 284621), 'pulumi.set', 'pulumi.set', (['self', '"""disk_type_id"""', 'value'], {}), "(self, 'disk_type_id', value)\n", (284592, 284621), False, 'import pulumi\n'), ((284759, 284809), 'pulumi.set', 'pulumi.set', (['__self__', '"""zip_filename"""', 'zip_filename'], {}), "(__self__, 'zip_filename', zip_filename)\n", (284769, 284809), False, 'import pulumi\n'), ((284928, 284960), 'pulumi.get', 'pulumi.get', (['self', '"""zip_filename"""'], {}), "(self, 'zip_filename')\n", (284938, 284960), False, 'import pulumi\n'), ((285049, 285088), 'pulumi.set', 'pulumi.set', (['self', '"""zip_filename"""', 'value'], {}), "(self, 'zip_filename', value)\n", (285059, 285088), False, 'import pulumi\n'), ((285335, 285383), 'pulumi.set', 'pulumi.set', (['__self__', '"""bucket_name"""', 'bucket_name'], {}), "(__self__, 'bucket_name', bucket_name)\n", (285345, 285383), False, 'import pulumi\n'), ((285392, 285440), 'pulumi.set', 'pulumi.set', (['__self__', '"""object_name"""', 'object_name'], {}), "(__self__, 'object_name', object_name)\n", (285402, 285440), False, 'import pulumi\n'), ((285639, 285670), 'pulumi.get', 'pulumi.get', (['self', '"""bucket_name"""'], {}), "(self, 'bucket_name')\n", (285649, 285670), False, 'import pulumi\n'), ((285757, 285795), 'pulumi.set', 'pulumi.set', (['self', '"""bucket_name"""', 'value'], {}), "(self, 'bucket_name', value)\n", (285767, 285795), False, 'import pulumi\n'), ((285912, 285943), 'pulumi.get', 'pulumi.get', (['self', '"""object_name"""'], {}), "(self, 'object_name')\n", (285922, 285943), False, 'import pulumi\n'), ((286030, 286068), 'pulumi.set', 'pulumi.set', (['self', '"""object_name"""', 'value'], {}), "(self, 'object_name', value)\n", (286040, 286068), False, 'import pulumi\n'), ((286171, 286197), 'pulumi.get', 'pulumi.get', (['self', '"""sha256"""'], {}), "(self, 'sha256')\n", (286181, 286197), False, 'import pulumi\n'), ((286284, 286317), 'pulumi.set', 'pulumi.set', (['self', '"""sha256"""', 'value'], {}), "(self, 'sha256', value)\n", (286294, 286317), False, 'import pulumi\n'), ((286607, 286639), 'pulumi.set', 'pulumi.set', (['__self__', '"""tag"""', 'tag'], {}), "(__self__, 'tag', tag)\n", (286617, 286639), False, 'import pulumi\n'), ((286974, 286997), 'pulumi.get', 'pulumi.get', (['self', '"""tag"""'], {}), "(self, 'tag')\n", (286984, 286997), False, 'import pulumi\n'), ((287068, 287098), 'pulumi.set', 'pulumi.set', (['self', '"""tag"""', 'value'], {}), "(self, 'tag', value)\n", (287078, 287098), False, 'import pulumi\n'), ((287242, 287282), 'pulumi.get', 'pulumi.get', (['self', '"""zone_instances_limit"""'], {}), "(self, 'zone_instances_limit')\n", (287252, 287282), False, 'import pulumi\n'), ((287397, 287444), 'pulumi.set', 'pulumi.set', (['self', '"""zone_instances_limit"""', 'value'], {}), "(self, 'zone_instances_limit', value)\n", (287407, 287444), False, 'import pulumi\n'), ((287586, 287625), 'pulumi.get', 'pulumi.get', (['self', '"""zone_requests_limit"""'], {}), "(self, 'zone_requests_limit')\n", (287596, 287625), False, 'import pulumi\n'), ((287738, 287784), 'pulumi.set', 'pulumi.set', (['self', '"""zone_requests_limit"""', 'value'], {}), "(self, 'zone_requests_limit', value)\n", (287748, 287784), False, 'import pulumi\n'), ((287977, 288019), 'pulumi.set', 'pulumi.set', (['__self__', '"""queue_id"""', 'queue_id'], {}), "(__self__, 'queue_id', queue_id)\n", (287987, 288019), False, 'import pulumi\n'), ((288028, 288090), 'pulumi.set', 'pulumi.set', (['__self__', '"""service_account_id"""', 'service_account_id'], {}), "(__self__, 'service_account_id', service_account_id)\n", (288038, 288090), False, 'import pulumi\n'), ((288201, 288229), 'pulumi.get', 'pulumi.get', (['self', '"""queue_id"""'], {}), "(self, 'queue_id')\n", (288211, 288229), False, 'import pulumi\n'), ((288310, 288345), 'pulumi.set', 'pulumi.set', (['self', '"""queue_id"""', 'value'], {}), "(self, 'queue_id', value)\n", (288320, 288345), False, 'import pulumi\n'), ((288475, 288513), 'pulumi.get', 'pulumi.get', (['self', '"""service_account_id"""'], {}), "(self, 'service_account_id')\n", (288485, 288513), False, 'import pulumi\n'), ((288614, 288659), 'pulumi.set', 'pulumi.set', (['self', '"""service_account_id"""', 'value'], {}), "(self, 'service_account_id', value)\n", (288624, 288659), False, 'import pulumi\n'), ((289064, 289094), 'pulumi.set', 'pulumi.set', (['__self__', '"""id"""', 'id'], {}), "(__self__, 'id', id)\n", (289074, 289094), False, 'import pulumi\n'), ((289586, 289608), 'pulumi.get', 'pulumi.get', (['self', '"""id"""'], {}), "(self, 'id')\n", (289596, 289608), False, 'import pulumi\n'), ((289677, 289706), 'pulumi.set', 'pulumi.set', (['self', '"""id"""', 'value'], {}), "(self, 'id', value)\n", (289687, 289706), False, 'import pulumi\n'), ((289839, 289873), 'pulumi.get', 'pulumi.get', (['self', '"""retry_attempts"""'], {}), "(self, 'retry_attempts')\n", (289849, 289873), False, 'import pulumi\n'), ((289976, 290017), 'pulumi.set', 'pulumi.set', (['self', '"""retry_attempts"""', 'value'], {}), "(self, 'retry_attempts', value)\n", (289986, 290017), False, 'import pulumi\n'), ((290150, 290184), 'pulumi.get', 'pulumi.get', (['self', '"""retry_interval"""'], {}), "(self, 'retry_interval')\n", (290160, 290184), False, 'import pulumi\n'), ((290287, 290328), 'pulumi.set', 'pulumi.set', (['self', '"""retry_interval"""', 'value'], {}), "(self, 'retry_interval', value)\n", (290297, 290328), False, 'import pulumi\n'), ((290468, 290506), 'pulumi.get', 'pulumi.get', (['self', '"""service_account_id"""'], {}), "(self, 'service_account_id')\n", (290478, 290506), False, 'import pulumi\n'), ((290617, 290662), 'pulumi.set', 'pulumi.set', (['self', '"""service_account_id"""', 'value'], {}), "(self, 'service_account_id', value)\n", (290627, 290662), False, 'import pulumi\n'), ((290762, 290785), 'pulumi.get', 'pulumi.get', (['self', '"""tag"""'], {}), "(self, 'tag')\n", (290772, 290785), False, 'import pulumi\n'), ((290866, 290896), 'pulumi.set', 'pulumi.set', (['self', '"""tag"""', 'value'], {}), "(self, 'tag', value)\n", (290876, 290896), False, 'import pulumi\n'), ((291160, 291208), 'pulumi.set', 'pulumi.set', (['__self__', '"""registry_id"""', 'registry_id'], {}), "(__self__, 'registry_id', registry_id)\n", (291170, 291208), False, 'import pulumi\n'), ((291495, 291526), 'pulumi.get', 'pulumi.get', (['self', '"""registry_id"""'], {}), "(self, 'registry_id')\n", (291505, 291526), False, 'import pulumi\n'), ((291613, 291651), 'pulumi.set', 'pulumi.set', (['self', '"""registry_id"""', 'value'], {}), "(self, 'registry_id', value)\n", (291623, 291651), False, 'import pulumi\n'), ((291774, 291803), 'pulumi.get', 'pulumi.get', (['self', '"""device_id"""'], {}), "(self, 'device_id')\n", (291784, 291803), False, 'import pulumi\n'), ((291896, 291932), 'pulumi.set', 'pulumi.set', (['self', '"""device_id"""', 'value'], {}), "(self, 'device_id', value)\n", (291906, 291932), False, 'import pulumi\n'), ((292034, 292059), 'pulumi.get', 'pulumi.get', (['self', '"""topic"""'], {}), "(self, 'topic')\n", (292044, 292059), False, 'import pulumi\n'), ((292144, 292176), 'pulumi.set', 'pulumi.set', (['self', '"""topic"""', 'value'], {}), "(self, 'topic', value)\n", (292154, 292176), False, 'import pulumi\n'), ((292462, 292512), 'pulumi.set', 'pulumi.set', (['__self__', '"""batch_cutoff"""', 'batch_cutoff'], {}), "(__self__, 'batch_cutoff', batch_cutoff)\n", (292472, 292512), False, 'import pulumi\n'), ((292521, 292573), 'pulumi.set', 'pulumi.set', (['__self__', '"""log_group_ids"""', 'log_group_ids'], {}), "(__self__, 'log_group_ids', log_group_ids)\n", (292531, 292573), False, 'import pulumi\n'), ((292786, 292818), 'pulumi.get', 'pulumi.get', (['self', '"""batch_cutoff"""'], {}), "(self, 'batch_cutoff')\n", (292796, 292818), False, 'import pulumi\n'), ((292907, 292946), 'pulumi.set', 'pulumi.set', (['self', '"""batch_cutoff"""', 'value'], {}), "(self, 'batch_cutoff', value)\n", (292917, 292946), False, 'import pulumi\n'), ((293090, 293123), 'pulumi.get', 'pulumi.get', (['self', '"""log_group_ids"""'], {}), "(self, 'log_group_ids')\n", (293100, 293123), False, 'import pulumi\n'), ((293238, 293278), 'pulumi.set', 'pulumi.set', (['self', '"""log_group_ids"""', 'value'], {}), "(self, 'log_group_ids', value)\n", (293248, 293278), False, 'import pulumi\n'), ((293403, 293433), 'pulumi.get', 'pulumi.get', (['self', '"""batch_size"""'], {}), "(self, 'batch_size')\n", (293413, 293433), False, 'import pulumi\n'), ((293528, 293565), 'pulumi.set', 'pulumi.set', (['self', '"""batch_size"""', 'value'], {}), "(self, 'batch_size', value)\n", (293538, 293565), False, 'import pulumi\n'), ((294039, 294089), 'pulumi.set', 'pulumi.set', (['__self__', '"""batch_cutoff"""', 'batch_cutoff'], {}), "(__self__, 'batch_cutoff', batch_cutoff)\n", (294049, 294089), False, 'import pulumi\n'), ((294098, 294140), 'pulumi.set', 'pulumi.set', (['__self__', '"""group_id"""', 'group_id'], {}), "(__self__, 'group_id', group_id)\n", (294108, 294140), False, 'import pulumi\n'), ((294149, 294187), 'pulumi.set', 'pulumi.set', (['__self__', '"""levels"""', 'levels'], {}), "(__self__, 'levels', levels)\n", (294159, 294187), False, 'import pulumi\n'), ((294196, 294246), 'pulumi.set', 'pulumi.set', (['__self__', '"""resource_ids"""', 'resource_ids'], {}), "(__self__, 'resource_ids', resource_ids)\n", (294206, 294246), False, 'import pulumi\n'), ((294255, 294309), 'pulumi.set', 'pulumi.set', (['__self__', '"""resource_types"""', 'resource_types'], {}), "(__self__, 'resource_types', resource_types)\n", (294265, 294309), False, 'import pulumi\n'), ((294522, 294554), 'pulumi.get', 'pulumi.get', (['self', '"""batch_cutoff"""'], {}), "(self, 'batch_cutoff')\n", (294532, 294554), False, 'import pulumi\n'), ((294643, 294682), 'pulumi.set', 'pulumi.set', (['self', '"""batch_cutoff"""', 'value'], {}), "(self, 'batch_cutoff', value)\n", (294653, 294682), False, 'import pulumi\n'), ((294793, 294821), 'pulumi.get', 'pulumi.get', (['self', '"""group_id"""'], {}), "(self, 'group_id')\n", (294803, 294821), False, 'import pulumi\n'), ((294902, 294937), 'pulumi.set', 'pulumi.set', (['self', '"""group_id"""', 'value'], {}), "(self, 'group_id', value)\n", (294912, 294937), False, 'import pulumi\n'), ((295054, 295080), 'pulumi.get', 'pulumi.get', (['self', '"""levels"""'], {}), "(self, 'levels')\n", (295064, 295080), False, 'import pulumi\n'), ((295181, 295214), 'pulumi.set', 'pulumi.set', (['self', '"""levels"""', 'value'], {}), "(self, 'levels', value)\n", (295191, 295214), False, 'import pulumi\n'), ((295357, 295389), 'pulumi.get', 'pulumi.get', (['self', '"""resource_ids"""'], {}), "(self, 'resource_ids')\n", (295367, 295389), False, 'import pulumi\n'), ((295502, 295541), 'pulumi.set', 'pulumi.set', (['self', '"""resource_ids"""', 'value'], {}), "(self, 'resource_ids', value)\n", (295512, 295541), False, 'import pulumi\n'), ((295688, 295722), 'pulumi.get', 'pulumi.get', (['self', '"""resource_types"""'], {}), "(self, 'resource_types')\n", (295698, 295722), False, 'import pulumi\n'), ((295839, 295880), 'pulumi.set', 'pulumi.set', (['self', '"""resource_types"""', 'value'], {}), "(self, 'resource_types', value)\n", (295849, 295880), False, 'import pulumi\n'), ((296005, 296035), 'pulumi.get', 'pulumi.get', (['self', '"""batch_size"""'], {}), "(self, 'batch_size')\n", (296015, 296035), False, 'import pulumi\n'), ((296130, 296167), 'pulumi.set', 'pulumi.set', (['self', '"""batch_size"""', 'value'], {}), "(self, 'batch_size', value)\n", (296140, 296167), False, 'import pulumi\n'), ((296557, 296607), 'pulumi.set', 'pulumi.set', (['__self__', '"""batch_cutoff"""', 'batch_cutoff'], {}), "(__self__, 'batch_cutoff', batch_cutoff)\n", (296567, 296607), False, 'import pulumi\n'), ((296616, 296658), 'pulumi.set', 'pulumi.set', (['__self__', '"""queue_id"""', 'queue_id'], {}), "(__self__, 'queue_id', queue_id)\n", (296626, 296658), False, 'import pulumi\n'), ((296667, 296729), 'pulumi.set', 'pulumi.set', (['__self__', '"""service_account_id"""', 'service_account_id'], {}), "(__self__, 'service_account_id', service_account_id)\n", (296677, 296729), False, 'import pulumi\n'), ((297060, 297092), 'pulumi.get', 'pulumi.get', (['self', '"""batch_cutoff"""'], {}), "(self, 'batch_cutoff')\n", (297070, 297092), False, 'import pulumi\n'), ((297181, 297220), 'pulumi.set', 'pulumi.set', (['self', '"""batch_cutoff"""', 'value'], {}), "(self, 'batch_cutoff', value)\n", (297191, 297220), False, 'import pulumi\n'), ((297331, 297359), 'pulumi.get', 'pulumi.get', (['self', '"""queue_id"""'], {}), "(self, 'queue_id')\n", (297341, 297359), False, 'import pulumi\n'), ((297440, 297475), 'pulumi.set', 'pulumi.set', (['self', '"""queue_id"""', 'value'], {}), "(self, 'queue_id', value)\n", (297450, 297475), False, 'import pulumi\n'), ((297605, 297643), 'pulumi.get', 'pulumi.get', (['self', '"""service_account_id"""'], {}), "(self, 'service_account_id')\n", (297615, 297643), False, 'import pulumi\n'), ((297744, 297789), 'pulumi.set', 'pulumi.set', (['self', '"""service_account_id"""', 'value'], {}), "(self, 'service_account_id', value)\n", (297754, 297789), False, 'import pulumi\n'), ((297914, 297944), 'pulumi.get', 'pulumi.get', (['self', '"""batch_size"""'], {}), "(self, 'batch_size')\n", (297924, 297944), False, 'import pulumi\n'), ((298039, 298076), 'pulumi.set', 'pulumi.set', (['self', '"""batch_size"""', 'value'], {}), "(self, 'batch_size', value)\n", (298049, 298076), False, 'import pulumi\n'), ((298217, 298255), 'pulumi.get', 'pulumi.get', (['self', '"""visibility_timeout"""'], {}), "(self, 'visibility_timeout')\n", (298227, 298255), False, 'import pulumi\n'), ((298366, 298411), 'pulumi.set', 'pulumi.set', (['self', '"""visibility_timeout"""', 'value'], {}), "(self, 'visibility_timeout', value)\n", (298376, 298411), False, 'import pulumi\n'), ((298867, 298911), 'pulumi.set', 'pulumi.set', (['__self__', '"""bucket_id"""', 'bucket_id'], {}), "(__self__, 'bucket_id', bucket_id)\n", (298877, 298911), False, 'import pulumi\n'), ((299434, 299463), 'pulumi.get', 'pulumi.get', (['self', '"""bucket_id"""'], {}), "(self, 'bucket_id')\n", (299444, 299463), False, 'import pulumi\n'), ((299546, 299582), 'pulumi.set', 'pulumi.set', (['self', '"""bucket_id"""', 'value'], {}), "(self, 'bucket_id', value)\n", (299556, 299582), False, 'import pulumi\n'), ((299686, 299712), 'pulumi.get', 'pulumi.get', (['self', '"""create"""'], {}), "(self, 'create')\n", (299696, 299712), False, 'import pulumi\n'), ((299800, 299833), 'pulumi.set', 'pulumi.set', (['self', '"""create"""', 'value'], {}), "(self, 'create', value)\n", (299810, 299833), False, 'import pulumi\n'), ((299937, 299963), 'pulumi.get', 'pulumi.get', (['self', '"""delete"""'], {}), "(self, 'delete')\n", (299947, 299963), False, 'import pulumi\n'), ((300051, 300084), 'pulumi.set', 'pulumi.set', (['self', '"""delete"""', 'value'], {}), "(self, 'delete', value)\n", (300061, 300084), False, 'import pulumi\n'), ((300187, 300213), 'pulumi.get', 'pulumi.get', (['self', '"""prefix"""'], {}), "(self, 'prefix')\n", (300197, 300213), False, 'import pulumi\n'), ((300300, 300333), 'pulumi.set', 'pulumi.set', (['self', '"""prefix"""', 'value'], {}), "(self, 'prefix', value)\n", (300310, 300333), False, 'import pulumi\n'), ((300436, 300462), 'pulumi.get', 'pulumi.get', (['self', '"""suffix"""'], {}), "(self, 'suffix')\n", (300446, 300462), False, 'import pulumi\n'), ((300549, 300582), 'pulumi.set', 'pulumi.set', (['self', '"""suffix"""', 'value'], {}), "(self, 'suffix', value)\n", (300559, 300582), False, 'import pulumi\n'), ((300686, 300712), 'pulumi.get', 'pulumi.get', (['self', '"""update"""'], {}), "(self, 'update')\n", (300696, 300712), False, 'import pulumi\n'), ((300800, 300833), 'pulumi.set', 'pulumi.set', (['self', '"""update"""', 'value'], {}), "(self, 'update', value)\n", (300810, 300833), False, 'import pulumi\n'), ((300979, 301035), 'pulumi.set', 'pulumi.set', (['__self__', '"""cron_expression"""', 'cron_expression'], {}), "(__self__, 'cron_expression', cron_expression)\n", (300989, 301035), False, 'import pulumi\n'), ((301160, 301195), 'pulumi.get', 'pulumi.get', (['self', '"""cron_expression"""'], {}), "(self, 'cron_expression')\n", (301170, 301195), False, 'import pulumi\n'), ((301290, 301332), 'pulumi.set', 'pulumi.set', (['self', '"""cron_expression"""', 'value'], {}), "(self, 'cron_expression', value)\n", (301300, 301332), False, 'import pulumi\n'), ((301805, 301831), 'pulumi.get', 'pulumi.get', (['self', '"""key_id"""'], {}), "(self, 'key_id')\n", (301815, 301831), False, 'import pulumi\n'), ((301918, 301951), 'pulumi.set', 'pulumi.set', (['self', '"""key_id"""', 'value'], {}), "(self, 'key_id', value)\n", (301928, 301951), False, 'import pulumi\n'), ((306838, 306880), 'pulumi.get', 'pulumi.get', (['self', '"""cluster_ca_certificate"""'], {}), "(self, 'cluster_ca_certificate')\n", (306848, 306880), False, 'import pulumi\n'), ((306999, 307048), 'pulumi.set', 'pulumi.set', (['self', '"""cluster_ca_certificate"""', 'value'], {}), "(self, 'cluster_ca_certificate', value)\n", (307009, 307048), False, 'import pulumi\n'), ((307298, 307337), 'pulumi.get', 'pulumi.get', (['self', '"""external_v4_address"""'], {}), "(self, 'external_v4_address')\n", (307308, 307337), False, 'import pulumi\n'), ((307450, 307496), 'pulumi.set', 'pulumi.set', (['self', '"""external_v4_address"""', 'value'], {}), "(self, 'external_v4_address', value)\n", (307460, 307496), False, 'import pulumi\n'), ((307793, 307833), 'pulumi.get', 'pulumi.get', (['self', '"""external_v4_endpoint"""'], {}), "(self, 'external_v4_endpoint')\n", (307803, 307833), False, 'import pulumi\n'), ((307948, 307995), 'pulumi.set', 'pulumi.set', (['self', '"""external_v4_endpoint"""', 'value'], {}), "(self, 'external_v4_endpoint', value)\n", (307958, 307995), False, 'import pulumi\n'), ((308245, 308284), 'pulumi.get', 'pulumi.get', (['self', '"""internal_v4_address"""'], {}), "(self, 'internal_v4_address')\n", (308255, 308284), False, 'import pulumi\n'), ((308397, 308443), 'pulumi.set', 'pulumi.set', (['self', '"""internal_v4_address"""', 'value'], {}), "(self, 'internal_v4_address', value)\n", (308407, 308443), False, 'import pulumi\n'), ((308711, 308751), 'pulumi.get', 'pulumi.get', (['self', '"""internal_v4_endpoint"""'], {}), "(self, 'internal_v4_endpoint')\n", (308721, 308751), False, 'import pulumi\n'), ((308866, 308913), 'pulumi.set', 'pulumi.set', (['self', '"""internal_v4_endpoint"""', 'value'], {}), "(self, 'internal_v4_endpoint', value)\n", (308876, 308913), False, 'import pulumi\n'), ((309519, 309557), 'pulumi.get', 'pulumi.get', (['self', '"""maintenance_policy"""'], {}), "(self, 'maintenance_policy')\n", (309529, 309557), False, 'import pulumi\n'), ((309711, 309756), 'pulumi.set', 'pulumi.set', (['self', '"""maintenance_policy"""', 'value'], {}), "(self, 'maintenance_policy', value)\n", (309721, 309756), False, 'import pulumi\n'), ((310011, 310040), 'pulumi.get', 'pulumi.get', (['self', '"""public_ip"""'], {}), "(self, 'public_ip')\n", (310021, 310040), False, 'import pulumi\n'), ((310134, 310170), 'pulumi.set', 'pulumi.set', (['self', '"""public_ip"""', 'value'], {}), "(self, 'public_ip', value)\n", (310144, 310170), False, 'import pulumi\n'), ((310456, 310484), 'pulumi.get', 'pulumi.get', (['self', '"""regional"""'], {}), "(self, 'regional')\n", (310466, 310484), False, 'import pulumi\n'), ((310609, 310644), 'pulumi.set', 'pulumi.set', (['self', '"""regional"""', 'value'], {}), "(self, 'regional', value)\n", (310619, 310644), False, 'import pulumi\n'), ((310919, 310957), 'pulumi.get', 'pulumi.get', (['self', '"""security_group_ids"""'], {}), "(self, 'security_group_ids')\n", (310929, 310957), False, 'import pulumi\n'), ((311092, 311137), 'pulumi.set', 'pulumi.set', (['self', '"""security_group_ids"""', 'value'], {}), "(self, 'security_group_ids', value)\n", (311102, 311137), False, 'import pulumi\n'), ((311347, 311374), 'pulumi.get', 'pulumi.get', (['self', '"""version"""'], {}), "(self, 'version')\n", (311357, 311374), False, 'import pulumi\n'), ((311463, 311497), 'pulumi.set', 'pulumi.set', (['self', '"""version"""', 'value'], {}), "(self, 'version', value)\n", (311473, 311497), False, 'import pulumi\n'), ((311776, 311808), 'pulumi.get', 'pulumi.get', (['self', '"""version_info"""'], {}), "(self, 'version_info')\n", (311786, 311808), False, 'import pulumi\n'), ((311944, 311983), 'pulumi.set', 'pulumi.set', (['self', '"""version_info"""', 'value'], {}), "(self, 'version_info', value)\n", (311954, 311983), False, 'import pulumi\n'), ((312255, 312280), 'pulumi.get', 'pulumi.get', (['self', '"""zonal"""'], {}), "(self, 'zonal')\n", (312265, 312280), False, 'import pulumi\n'), ((312396, 312428), 'pulumi.set', 'pulumi.set', (['self', '"""zonal"""', 'value'], {}), "(self, 'zonal', value)\n", (312406, 312428), False, 'import pulumi\n'), ((313434, 313484), 'pulumi.set', 'pulumi.set', (['__self__', '"""auto_upgrade"""', 'auto_upgrade'], {}), "(__self__, 'auto_upgrade', auto_upgrade)\n", (313444, 313484), False, 'import pulumi\n'), ((313874, 313906), 'pulumi.get', 'pulumi.get', (['self', '"""auto_upgrade"""'], {}), "(self, 'auto_upgrade')\n", (313884, 313906), False, 'import pulumi\n'), ((313996, 314035), 'pulumi.set', 'pulumi.set', (['self', '"""auto_upgrade"""', 'value'], {}), "(self, 'auto_upgrade', value)\n", (314006, 314035), False, 'import pulumi\n'), ((314637, 314676), 'pulumi.get', 'pulumi.get', (['self', '"""maintenance_windows"""'], {}), "(self, 'maintenance_windows')\n", (314647, 314676), False, 'import pulumi\n'), ((314873, 314919), 'pulumi.set', 'pulumi.set', (['self', '"""maintenance_windows"""', 'value'], {}), "(self, 'maintenance_windows', value)\n", (314883, 314919), False, 'import pulumi\n'), ((315201, 315243), 'pulumi.set', 'pulumi.set', (['__self__', '"""duration"""', 'duration'], {}), "(__self__, 'duration', duration)\n", (315211, 315243), False, 'import pulumi\n'), ((315252, 315298), 'pulumi.set', 'pulumi.set', (['__self__', '"""start_time"""', 'start_time'], {}), "(__self__, 'start_time', start_time)\n", (315262, 315298), False, 'import pulumi\n'), ((315466, 315494), 'pulumi.get', 'pulumi.get', (['self', '"""duration"""'], {}), "(self, 'duration')\n", (315476, 315494), False, 'import pulumi\n'), ((315575, 315610), 'pulumi.set', 'pulumi.set', (['self', '"""duration"""', 'value'], {}), "(self, 'duration', value)\n", (315585, 315610), False, 'import pulumi\n'), ((315725, 315755), 'pulumi.get', 'pulumi.get', (['self', '"""start_time"""'], {}), "(self, 'start_time')\n", (315735, 315755), False, 'import pulumi\n'), ((315840, 315877), 'pulumi.set', 'pulumi.set', (['self', '"""start_time"""', 'value'], {}), "(self, 'start_time', value)\n", (315850, 315877), False, 'import pulumi\n'), ((315977, 316000), 'pulumi.get', 'pulumi.get', (['self', '"""day"""'], {}), "(self, 'day')\n", (315987, 316000), False, 'import pulumi\n'), ((316081, 316111), 'pulumi.set', 'pulumi.set', (['self', '"""day"""', 'value'], {}), "(self, 'day', value)\n", (316091, 316111), False, 'import pulumi\n'), ((316764, 316802), 'pulumi.set', 'pulumi.set', (['__self__', '"""region"""', 'region'], {}), "(__self__, 'region', region)\n", (316774, 316802), False, 'import pulumi\n'), ((317121, 317147), 'pulumi.get', 'pulumi.get', (['self', '"""region"""'], {}), "(self, 'region')\n", (317131, 317147), False, 'import pulumi\n'), ((317224, 317257), 'pulumi.set', 'pulumi.set', (['self', '"""region"""', 'value'], {}), "(self, 'region', value)\n", (317234, 317257), False, 'import pulumi\n'), ((317558, 317587), 'pulumi.get', 'pulumi.get', (['self', '"""locations"""'], {}), "(self, 'locations')\n", (317568, 317587), False, 'import pulumi\n'), ((317746, 317782), 'pulumi.set', 'pulumi.set', (['self', '"""locations"""', 'value'], {}), "(self, 'locations', value)\n", (317756, 317782), False, 'import pulumi\n'), ((318535, 318564), 'pulumi.get', 'pulumi.get', (['self', '"""subnet_id"""'], {}), "(self, 'subnet_id')\n", (318545, 318564), False, 'import pulumi\n'), ((318657, 318693), 'pulumi.set', 'pulumi.set', (['self', '"""subnet_id"""', 'value'], {}), "(self, 'subnet_id', value)\n", (318667, 318693), False, 'import pulumi\n'), ((318866, 318890), 'pulumi.get', 'pulumi.get', (['self', '"""zone"""'], {}), "(self, 'zone')\n", (318876, 318890), False, 'import pulumi\n'), ((318973, 319004), 'pulumi.set', 'pulumi.set', (['self', '"""zone"""', 'value'], {}), "(self, 'zone', value)\n", (318983, 319004), False, 'import pulumi\n'), ((320939, 320974), 'pulumi.get', 'pulumi.get', (['self', '"""current_version"""'], {}), "(self, 'current_version')\n", (320949, 320974), False, 'import pulumi\n'), ((321079, 321121), 'pulumi.set', 'pulumi.set', (['self', '"""current_version"""', 'value'], {}), "(self, 'current_version', value)\n", (321089, 321121), False, 'import pulumi\n'), ((321541, 321583), 'pulumi.get', 'pulumi.get', (['self', '"""new_revision_available"""'], {}), "(self, 'new_revision_available')\n", (321551, 321583), False, 'import pulumi\n'), ((321703, 321752), 'pulumi.set', 'pulumi.set', (['self', '"""new_revision_available"""', 'value'], {}), "(self, 'new_revision_available', value)\n", (321713, 321752), False, 'import pulumi\n'), ((322072, 322112), 'pulumi.get', 'pulumi.get', (['self', '"""new_revision_summary"""'], {}), "(self, 'new_revision_summary')\n", (322082, 322112), False, 'import pulumi\n'), ((322227, 322274), 'pulumi.set', 'pulumi.set', (['self', '"""new_revision_summary"""', 'value'], {}), "(self, 'new_revision_summary', value)\n", (322237, 322274), False, 'import pulumi\n'), ((322575, 322613), 'pulumi.get', 'pulumi.get', (['self', '"""version_deprecated"""'], {}), "(self, 'version_deprecated')\n", (322585, 322613), False, 'import pulumi\n'), ((322725, 322770), 'pulumi.set', 'pulumi.set', (['self', '"""version_deprecated"""', 'value'], {}), "(self, 'version_deprecated', value)\n", (322735, 322770), False, 'import pulumi\n'), ((323512, 323541), 'pulumi.get', 'pulumi.get', (['self', '"""subnet_id"""'], {}), "(self, 'subnet_id')\n", (323522, 323541), False, 'import pulumi\n'), ((323634, 323670), 'pulumi.set', 'pulumi.set', (['self', '"""subnet_id"""', 'value'], {}), "(self, 'subnet_id', value)\n", (323644, 323670), False, 'import pulumi\n'), ((323843, 323867), 'pulumi.get', 'pulumi.get', (['self', '"""zone"""'], {}), "(self, 'zone')\n", (323853, 323867), False, 'import pulumi\n'), ((323950, 323981), 'pulumi.set', 'pulumi.set', (['self', '"""zone"""', 'value'], {}), "(self, 'zone', value)\n", (323960, 323981), False, 'import pulumi\n'), ((324715, 324741), 'pulumi.get', 'pulumi.get', (['self', '"""cilium"""'], {}), "(self, 'cilium')\n", (324725, 324741), False, 'import pulumi\n'), ((324875, 324908), 'pulumi.set', 'pulumi.set', (['self', '"""cilium"""', 'value'], {}), "(self, 'cilium', value)\n", (324885, 324908), False, 'import pulumi\n'), ((325960, 325989), 'pulumi.get', 'pulumi.get', (['self', '"""locations"""'], {}), "(self, 'locations')\n", (325970, 325989), False, 'import pulumi\n'), ((326152, 326188), 'pulumi.set', 'pulumi.set', (['self', '"""locations"""', 'value'], {}), "(self, 'locations', value)\n", (326162, 326188), False, 'import pulumi\n'), ((327418, 327447), 'pulumi.get', 'pulumi.get', (['self', '"""subnet_id"""'], {}), "(self, 'subnet_id')\n", (327428, 327447), False, 'import pulumi\n'), ((327540, 327576), 'pulumi.set', 'pulumi.set', (['self', '"""subnet_id"""', 'value'], {}), "(self, 'subnet_id', value)\n", (327550, 327576), False, 'import pulumi\n'), ((327783, 327807), 'pulumi.get', 'pulumi.get', (['self', '"""zone"""'], {}), "(self, 'zone')\n", (327793, 327807), False, 'import pulumi\n'), ((327890, 327921), 'pulumi.set', 'pulumi.set', (['self', '"""zone"""', 'value'], {}), "(self, 'zone', value)\n", (327900, 327921), False, 'import pulumi\n'), ((328447, 328499), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_expansion"""', 'max_expansion'], {}), "(__self__, 'max_expansion', max_expansion)\n", (328457, 328499), False, 'import pulumi\n'), ((328508, 328564), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_unavailable"""', 'max_unavailable'], {}), "(__self__, 'max_unavailable', max_unavailable)\n", (328518, 328564), False, 'import pulumi\n'), ((328832, 328865), 'pulumi.get', 'pulumi.get', (['self', '"""max_expansion"""'], {}), "(self, 'max_expansion')\n", (328842, 328865), False, 'import pulumi\n'), ((328956, 328996), 'pulumi.set', 'pulumi.set', (['self', '"""max_expansion"""', 'value'], {}), "(self, 'max_expansion', value)\n", (328966, 328996), False, 'import pulumi\n'), ((329234, 329269), 'pulumi.get', 'pulumi.get', (['self', '"""max_unavailable"""'], {}), "(self, 'max_unavailable')\n", (329244, 329269), False, 'import pulumi\n'), ((329364, 329406), 'pulumi.set', 'pulumi.set', (['self', '"""max_unavailable"""', 'value'], {}), "(self, 'max_unavailable', value)\n", (329374, 329406), False, 'import pulumi\n'), ((333707, 333736), 'pulumi.get', 'pulumi.get', (['self', '"""boot_disk"""'], {}), "(self, 'boot_disk')\n", (333717, 333736), False, 'import pulumi\n'), ((333875, 333911), 'pulumi.set', 'pulumi.set', (['self', '"""boot_disk"""', 'value'], {}), "(self, 'boot_disk', value)\n", (333885, 333911), False, 'import pulumi\n'), ((334514, 334542), 'pulumi.get', 'pulumi.get', (['self', '"""metadata"""'], {}), "(self, 'metadata')\n", (334524, 334542), False, 'import pulumi\n'), ((334661, 334696), 'pulumi.set', 'pulumi.set', (['self', '"""metadata"""', 'value'], {}), "(self, 'metadata', value)\n", (334671, 334696), False, 'import pulumi\n'), ((334896, 334919), 'pulumi.get', 'pulumi.get', (['self', '"""nat"""'], {}), "(self, 'nat')\n", (334906, 334919), False, 'import pulumi\n'), ((335001, 335031), 'pulumi.set', 'pulumi.set', (['self', '"""nat"""', 'value'], {}), "(self, 'nat', value)\n", (335011, 335031), False, 'import pulumi\n'), ((335291, 335336), 'pulumi.get', 'pulumi.get', (['self', '"""network_acceleration_type"""'], {}), "(self, 'network_acceleration_type')\n", (335301, 335336), False, 'import pulumi\n'), ((335461, 335513), 'pulumi.set', 'pulumi.set', (['self', '"""network_acceleration_type"""', 'value'], {}), "(self, 'network_acceleration_type', value)\n", (335471, 335513), False, 'import pulumi\n'), ((335875, 335913), 'pulumi.get', 'pulumi.get', (['self', '"""network_interfaces"""'], {}), "(self, 'network_interfaces')\n", (335885, 335913), False, 'import pulumi\n'), ((336102, 336147), 'pulumi.set', 'pulumi.set', (['self', '"""network_interfaces"""', 'value'], {}), "(self, 'network_interfaces', value)\n", (336112, 336147), False, 'import pulumi\n'), ((336440, 336476), 'pulumi.get', 'pulumi.get', (['self', '"""placement_policy"""'], {}), "(self, 'placement_policy')\n", (336450, 336476), False, 'import pulumi\n'), ((336636, 336679), 'pulumi.set', 'pulumi.set', (['self', '"""placement_policy"""', 'value'], {}), "(self, 'placement_policy', value)\n", (336646, 336679), False, 'import pulumi\n'), ((336922, 336953), 'pulumi.get', 'pulumi.get', (['self', '"""platform_id"""'], {}), "(self, 'platform_id')\n", (336932, 336953), False, 'import pulumi\n'), ((337050, 337088), 'pulumi.set', 'pulumi.set', (['self', '"""platform_id"""', 'value'], {}), "(self, 'platform_id', value)\n", (337060, 337088), False, 'import pulumi\n'), ((337241, 337270), 'pulumi.get', 'pulumi.get', (['self', '"""resources"""'], {}), "(self, 'resources')\n", (337251, 337270), False, 'import pulumi\n'), ((337410, 337446), 'pulumi.set', 'pulumi.set', (['self', '"""resources"""', 'value'], {}), "(self, 'resources', value)\n", (337420, 337446), False, 'import pulumi\n'), ((337761, 337798), 'pulumi.get', 'pulumi.get', (['self', '"""scheduling_policy"""'], {}), "(self, 'scheduling_policy')\n", (337771, 337798), False, 'import pulumi\n'), ((337961, 338005), 'pulumi.set', 'pulumi.set', (['self', '"""scheduling_policy"""', 'value'], {}), "(self, 'scheduling_policy', value)\n", (337971, 338005), False, 'import pulumi\n'), ((338718, 338742), 'pulumi.get', 'pulumi.get', (['self', '"""size"""'], {}), "(self, 'size')\n", (338728, 338742), False, 'import pulumi\n'), ((338825, 338856), 'pulumi.set', 'pulumi.set', (['self', '"""size"""', 'value'], {}), "(self, 'size', value)\n", (338835, 338856), False, 'import pulumi\n'), ((339004, 339028), 'pulumi.get', 'pulumi.get', (['self', '"""type"""'], {}), "(self, 'type')\n", (339014, 339028), False, 'import pulumi\n'), ((339111, 339142), 'pulumi.set', 'pulumi.set', (['self', '"""type"""', 'value'], {}), "(self, 'type', value)\n", (339121, 339142), False, 'import pulumi\n'), ((340232, 340278), 'pulumi.set', 'pulumi.set', (['__self__', '"""subnet_ids"""', 'subnet_ids'], {}), "(__self__, 'subnet_ids', subnet_ids)\n", (340242, 340278), False, 'import pulumi\n'), ((340816, 340846), 'pulumi.get', 'pulumi.get', (['self', '"""subnet_ids"""'], {}), "(self, 'subnet_ids')\n", (340826, 340846), False, 'import pulumi\n'), ((340955, 340992), 'pulumi.set', 'pulumi.set', (['self', '"""subnet_ids"""', 'value'], {}), "(self, 'subnet_ids', value)\n", (340965, 340992), False, 'import pulumi\n'), ((341199, 341223), 'pulumi.get', 'pulumi.get', (['self', '"""ipv4"""'], {}), "(self, 'ipv4')\n", (341209, 341223), False, 'import pulumi\n'), ((341307, 341338), 'pulumi.set', 'pulumi.set', (['self', '"""ipv4"""', 'value'], {}), "(self, 'ipv4', value)\n", (341317, 341338), False, 'import pulumi\n'), ((341595, 341619), 'pulumi.get', 'pulumi.get', (['self', '"""ipv6"""'], {}), "(self, 'ipv6')\n", (341605, 341619), False, 'import pulumi\n'), ((341703, 341734), 'pulumi.set', 'pulumi.set', (['self', '"""ipv6"""', 'value'], {}), "(self, 'ipv6', value)\n", (341713, 341734), False, 'import pulumi\n'), ((341934, 341957), 'pulumi.get', 'pulumi.get', (['self', '"""nat"""'], {}), "(self, 'nat')\n", (341944, 341957), False, 'import pulumi\n'), ((342039, 342069), 'pulumi.set', 'pulumi.set', (['self', '"""nat"""', 'value'], {}), "(self, 'nat', value)\n", (342049, 342069), False, 'import pulumi\n'), ((342307, 342345), 'pulumi.get', 'pulumi.get', (['self', '"""security_group_ids"""'], {}), "(self, 'security_group_ids')\n", (342317, 342345), False, 'import pulumi\n'), ((342480, 342525), 'pulumi.set', 'pulumi.set', (['self', '"""security_group_ids"""', 'value'], {}), "(self, 'security_group_ids', value)\n", (342490, 342525), False, 'import pulumi\n'), ((342849, 342911), 'pulumi.set', 'pulumi.set', (['__self__', '"""placement_group_id"""', 'placement_group_id'], {}), "(__self__, 'placement_group_id', placement_group_id)\n", (342859, 342911), False, 'import pulumi\n'), ((343141, 343179), 'pulumi.get', 'pulumi.get', (['self', '"""placement_group_id"""'], {}), "(self, 'placement_group_id')\n", (343151, 343179), False, 'import pulumi\n'), ((343280, 343325), 'pulumi.set', 'pulumi.set', (['self', '"""placement_group_id"""', 'value'], {}), "(self, 'placement_group_id', value)\n", (343290, 343325), False, 'import pulumi\n'), ((344154, 344187), 'pulumi.get', 'pulumi.get', (['self', '"""core_fraction"""'], {}), "(self, 'core_fraction')\n", (344164, 344187), False, 'import pulumi\n'), ((344288, 344328), 'pulumi.set', 'pulumi.set', (['self', '"""core_fraction"""', 'value'], {}), "(self, 'core_fraction', value)\n", (344298, 344328), False, 'import pulumi\n'), ((344430, 344455), 'pulumi.get', 'pulumi.get', (['self', '"""cores"""'], {}), "(self, 'cores')\n", (344440, 344455), False, 'import pulumi\n'), ((344540, 344572), 'pulumi.set', 'pulumi.set', (['self', '"""cores"""', 'value'], {}), "(self, 'cores', value)\n", (344550, 344572), False, 'import pulumi\n'), ((344673, 344697), 'pulumi.get', 'pulumi.get', (['self', '"""gpus"""'], {}), "(self, 'gpus')\n", (344683, 344697), False, 'import pulumi\n'), ((344780, 344811), 'pulumi.set', 'pulumi.set', (['self', '"""gpus"""', 'value'], {}), "(self, 'gpus', value)\n", (344790, 344811), False, 'import pulumi\n'), ((344916, 344942), 'pulumi.get', 'pulumi.get', (['self', '"""memory"""'], {}), "(self, 'memory')\n", (344926, 344942), False, 'import pulumi\n'), ((345031, 345064), 'pulumi.set', 'pulumi.set', (['self', '"""memory"""', 'value'], {}), "(self, 'memory', value)\n", (345041, 345064), False, 'import pulumi\n'), ((345708, 345739), 'pulumi.get', 'pulumi.get', (['self', '"""preemptible"""'], {}), "(self, 'preemptible')\n", (345718, 345739), False, 'import pulumi\n'), ((345837, 345875), 'pulumi.set', 'pulumi.set', (['self', '"""preemptible"""', 'value'], {}), "(self, 'preemptible', value)\n", (345847, 345875), False, 'import pulumi\n'), ((346833, 346881), 'pulumi.set', 'pulumi.set', (['__self__', '"""auto_repair"""', 'auto_repair'], {}), "(__self__, 'auto_repair', auto_repair)\n", (346843, 346881), False, 'import pulumi\n'), ((346890, 346940), 'pulumi.set', 'pulumi.set', (['__self__', '"""auto_upgrade"""', 'auto_upgrade'], {}), "(__self__, 'auto_upgrade', auto_upgrade)\n", (346900, 346940), False, 'import pulumi\n'), ((347321, 347352), 'pulumi.get', 'pulumi.get', (['self', '"""auto_repair"""'], {}), "(self, 'auto_repair')\n", (347331, 347352), False, 'import pulumi\n'), ((347440, 347478), 'pulumi.set', 'pulumi.set', (['self', '"""auto_repair"""', 'value'], {}), "(self, 'auto_repair', value)\n", (347450, 347478), False, 'import pulumi\n'), ((347740, 347772), 'pulumi.get', 'pulumi.get', (['self', '"""auto_upgrade"""'], {}), "(self, 'auto_upgrade')\n", (347750, 347772), False, 'import pulumi\n'), ((347862, 347901), 'pulumi.set', 'pulumi.set', (['self', '"""auto_upgrade"""', 'value'], {}), "(self, 'auto_upgrade', value)\n", (347872, 347901), False, 'import pulumi\n'), ((348277, 348316), 'pulumi.get', 'pulumi.get', (['self', '"""maintenance_windows"""'], {}), "(self, 'maintenance_windows')\n", (348287, 348316), False, 'import pulumi\n'), ((348509, 348555), 'pulumi.set', 'pulumi.set', (['self', '"""maintenance_windows"""', 'value'], {}), "(self, 'maintenance_windows', value)\n", (348519, 348555), False, 'import pulumi\n'), ((348833, 348875), 'pulumi.set', 'pulumi.set', (['__self__', '"""duration"""', 'duration'], {}), "(__self__, 'duration', duration)\n", (348843, 348875), False, 'import pulumi\n'), ((348884, 348930), 'pulumi.set', 'pulumi.set', (['__self__', '"""start_time"""', 'start_time'], {}), "(__self__, 'start_time', start_time)\n", (348894, 348930), False, 'import pulumi\n'), ((349098, 349126), 'pulumi.get', 'pulumi.get', (['self', '"""duration"""'], {}), "(self, 'duration')\n", (349108, 349126), False, 'import pulumi\n'), ((349207, 349242), 'pulumi.set', 'pulumi.set', (['self', '"""duration"""', 'value'], {}), "(self, 'duration', value)\n", (349217, 349242), False, 'import pulumi\n'), ((349357, 349387), 'pulumi.get', 'pulumi.get', (['self', '"""start_time"""'], {}), "(self, 'start_time')\n", (349367, 349387), False, 'import pulumi\n'), ((349472, 349509), 'pulumi.set', 'pulumi.set', (['self', '"""start_time"""', 'value'], {}), "(self, 'start_time', value)\n", (349482, 349509), False, 'import pulumi\n'), ((349609, 349632), 'pulumi.get', 'pulumi.get', (['self', '"""day"""'], {}), "(self, 'day')\n", (349619, 349632), False, 'import pulumi\n'), ((349713, 349743), 'pulumi.set', 'pulumi.set', (['self', '"""day"""', 'value'], {}), "(self, 'day', value)\n", (349723, 349743), False, 'import pulumi\n'), ((350877, 350907), 'pulumi.get', 'pulumi.get', (['self', '"""auto_scale"""'], {}), "(self, 'auto_scale')\n", (350887, 350907), False, 'import pulumi\n'), ((351044, 351081), 'pulumi.set', 'pulumi.set', (['self', '"""auto_scale"""', 'value'], {}), "(self, 'auto_scale', value)\n", (351054, 351081), False, 'import pulumi\n'), ((351361, 351392), 'pulumi.get', 'pulumi.get', (['self', '"""fixed_scale"""'], {}), "(self, 'fixed_scale')\n", (351371, 351392), False, 'import pulumi\n'), ((351532, 351570), 'pulumi.set', 'pulumi.set', (['self', '"""fixed_scale"""', 'value'], {}), "(self, 'fixed_scale', value)\n", (351542, 351570), False, 'import pulumi\n'), ((352092, 352132), 'pulumi.set', 'pulumi.set', (['__self__', '"""initial"""', 'initial'], {}), "(__self__, 'initial', initial)\n", (352102, 352132), False, 'import pulumi\n'), ((352141, 352173), 'pulumi.set', 'pulumi.set', (['__self__', '"""max"""', 'max'], {}), "(__self__, 'max', max)\n", (352151, 352173), False, 'import pulumi\n'), ((352182, 352214), 'pulumi.set', 'pulumi.set', (['__self__', '"""min"""', 'min'], {}), "(__self__, 'min', min)\n", (352192, 352214), False, 'import pulumi\n'), ((352387, 352414), 'pulumi.get', 'pulumi.get', (['self', '"""initial"""'], {}), "(self, 'initial')\n", (352397, 352414), False, 'import pulumi\n'), ((352493, 352527), 'pulumi.set', 'pulumi.set', (['self', '"""initial"""', 'value'], {}), "(self, 'initial', value)\n", (352503, 352527), False, 'import pulumi\n'), ((352696, 352719), 'pulumi.get', 'pulumi.get', (['self', '"""max"""'], {}), "(self, 'max')\n", (352706, 352719), False, 'import pulumi\n'), ((352790, 352820), 'pulumi.set', 'pulumi.set', (['self', '"""max"""', 'value'], {}), "(self, 'max', value)\n", (352800, 352820), False, 'import pulumi\n'), ((352989, 353012), 'pulumi.get', 'pulumi.get', (['self', '"""min"""'], {}), "(self, 'min')\n", (352999, 353012), False, 'import pulumi\n'), ((353083, 353113), 'pulumi.set', 'pulumi.set', (['self', '"""min"""', 'value'], {}), "(self, 'min', value)\n", (353093, 353113), False, 'import pulumi\n'), ((353634, 353658), 'pulumi.get', 'pulumi.get', (['self', '"""size"""'], {}), "(self, 'size')\n", (353644, 353658), False, 'import pulumi\n'), ((353741, 353772), 'pulumi.set', 'pulumi.set', (['self', '"""size"""', 'value'], {}), "(self, 'size', value)\n", (353751, 353772), False, 'import pulumi\n'), ((355709, 355744), 'pulumi.get', 'pulumi.get', (['self', '"""current_version"""'], {}), "(self, 'current_version')\n", (355719, 355744), False, 'import pulumi\n'), ((355849, 355891), 'pulumi.set', 'pulumi.set', (['self', '"""current_version"""', 'value'], {}), "(self, 'current_version', value)\n", (355859, 355891), False, 'import pulumi\n'), ((356314, 356356), 'pulumi.get', 'pulumi.get', (['self', '"""new_revision_available"""'], {}), "(self, 'new_revision_available')\n", (356324, 356356), False, 'import pulumi\n'), ((356476, 356525), 'pulumi.set', 'pulumi.set', (['self', '"""new_revision_available"""', 'value'], {}), "(self, 'new_revision_available', value)\n", (356486, 356525), False, 'import pulumi\n'), ((356845, 356885), 'pulumi.get', 'pulumi.get', (['self', '"""new_revision_summary"""'], {}), "(self, 'new_revision_summary')\n", (356855, 356885), False, 'import pulumi\n'), ((357000, 357047), 'pulumi.set', 'pulumi.set', (['self', '"""new_revision_summary"""', 'value'], {}), "(self, 'new_revision_summary', value)\n", (357010, 357047), False, 'import pulumi\n'), ((357351, 357389), 'pulumi.get', 'pulumi.get', (['self', '"""version_deprecated"""'], {}), "(self, 'version_deprecated')\n", (357361, 357389), False, 'import pulumi\n'), ((357501, 357546), 'pulumi.set', 'pulumi.set', (['self', '"""version_deprecated"""', 'value'], {}), "(self, 'version_deprecated', value)\n", (357511, 357546), False, 'import pulumi\n'), ((358122, 358172), 'pulumi.set', 'pulumi.set', (['__self__', '"""healthchecks"""', 'healthchecks'], {}), "(__self__, 'healthchecks', healthchecks)\n", (358132, 358172), False, 'import pulumi\n'), ((358181, 358237), 'pulumi.set', 'pulumi.set', (['__self__', '"""target_group_id"""', 'target_group_id'], {}), "(__self__, 'target_group_id', target_group_id)\n", (358191, 358237), False, 'import pulumi\n'), ((358505, 358537), 'pulumi.get', 'pulumi.get', (['self', '"""healthchecks"""'], {}), "(self, 'healthchecks')\n", (358515, 358537), False, 'import pulumi\n'), ((358704, 358743), 'pulumi.set', 'pulumi.set', (['self', '"""healthchecks"""', 'value'], {}), "(self, 'healthchecks', value)\n", (358714, 358743), False, 'import pulumi\n'), ((358923, 358958), 'pulumi.get', 'pulumi.get', (['self', '"""target_group_id"""'], {}), "(self, 'target_group_id')\n", (358933, 358958), False, 'import pulumi\n'), ((359053, 359095), 'pulumi.set', 'pulumi.set', (['self', '"""target_group_id"""', 'value'], {}), "(self, 'target_group_id', value)\n", (359063, 359095), False, 'import pulumi\n'), ((360832, 360866), 'pulumi.set', 'pulumi.set', (['__self__', '"""name"""', 'name'], {}), "(__self__, 'name', name)\n", (360842, 360866), False, 'import pulumi\n'), ((361686, 361710), 'pulumi.get', 'pulumi.get', (['self', '"""name"""'], {}), "(self, 'name')\n", (361696, 361710), False, 'import pulumi\n'), ((361783, 361814), 'pulumi.set', 'pulumi.set', (['self', '"""name"""', 'value'], {}), "(self, 'name', value)\n", (361793, 361814), False, 'import pulumi\n'), ((362082, 362119), 'pulumi.get', 'pulumi.get', (['self', '"""healthy_threshold"""'], {}), "(self, 'healthy_threshold')\n", (362092, 362119), False, 'import pulumi\n'), ((362228, 362272), 'pulumi.set', 'pulumi.set', (['self', '"""healthy_threshold"""', 'value'], {}), "(self, 'healthy_threshold', value)\n", (362238, 362272), False, 'import pulumi\n'), ((362564, 362596), 'pulumi.get', 'pulumi.get', (['self', '"""http_options"""'], {}), "(self, 'http_options')\n", (362574, 362596), False, 'import pulumi\n'), ((362760, 362799), 'pulumi.set', 'pulumi.set', (['self', '"""http_options"""', 'value'], {}), "(self, 'http_options', value)\n", (362770, 362799), False, 'import pulumi\n'), ((362998, 363026), 'pulumi.get', 'pulumi.get', (['self', '"""interval"""'], {}), "(self, 'interval')\n", (363008, 363026), False, 'import pulumi\n'), ((363117, 363152), 'pulumi.set', 'pulumi.set', (['self', '"""interval"""', 'value'], {}), "(self, 'interval', value)\n", (363127, 363152), False, 'import pulumi\n'), ((363440, 363471), 'pulumi.get', 'pulumi.get', (['self', '"""tcp_options"""'], {}), "(self, 'tcp_options')\n", (363450, 363471), False, 'import pulumi\n'), ((363632, 363670), 'pulumi.set', 'pulumi.set', (['self', '"""tcp_options"""', 'value'], {}), "(self, 'tcp_options', value)\n", (363642, 363670), False, 'import pulumi\n'), ((363895, 363922), 'pulumi.get', 'pulumi.get', (['self', '"""timeout"""'], {}), "(self, 'timeout')\n", (363905, 363922), False, 'import pulumi\n'), ((364011, 364045), 'pulumi.set', 'pulumi.set', (['self', '"""timeout"""', 'value'], {}), "(self, 'timeout', value)\n", (364021, 364045), False, 'import pulumi\n'), ((364312, 364351), 'pulumi.get', 'pulumi.get', (['self', '"""unhealthy_threshold"""'], {}), "(self, 'unhealthy_threshold')\n", (364322, 364351), False, 'import pulumi\n'), ((364464, 364510), 'pulumi.set', 'pulumi.set', (['self', '"""unhealthy_threshold"""', 'value'], {}), "(self, 'unhealthy_threshold', value)\n", (364474, 364510), False, 'import pulumi\n'), ((365004, 365038), 'pulumi.set', 'pulumi.set', (['__self__', '"""port"""', 'port'], {}), "(__self__, 'port', port)\n", (365014, 365038), False, 'import pulumi\n'), ((365264, 365288), 'pulumi.get', 'pulumi.get', (['self', '"""port"""'], {}), "(self, 'port')\n", (365274, 365288), False, 'import pulumi\n'), ((365361, 365392), 'pulumi.set', 'pulumi.set', (['self', '"""port"""', 'value'], {}), "(self, 'port', value)\n", (365371, 365392), False, 'import pulumi\n'), ((365654, 365678), 'pulumi.get', 'pulumi.get', (['self', '"""path"""'], {}), "(self, 'path')\n", (365664, 365678), False, 'import pulumi\n'), ((365761, 365792), 'pulumi.set', 'pulumi.set', (['self', '"""path"""', 'value'], {}), "(self, 'path', value)\n", (365771, 365792), False, 'import pulumi\n'), ((366058, 366092), 'pulumi.set', 'pulumi.set', (['__self__', '"""port"""', 'port'], {}), "(__self__, 'port', port)\n", (366068, 366092), False, 'import pulumi\n'), ((366242, 366266), 'pulumi.get', 'pulumi.get', (['self', '"""port"""'], {}), "(self, 'port')\n", (366252, 366266), False, 'import pulumi\n'), ((366339, 366370), 'pulumi.set', 'pulumi.set', (['self', '"""port"""', 'value'], {}), "(self, 'port', value)\n", (366349, 366370), False, 'import pulumi\n'), ((367730, 367764), 'pulumi.set', 'pulumi.set', (['__self__', '"""name"""', 'name'], {}), "(__self__, 'name', name)\n", (367740, 367764), False, 'import pulumi\n'), ((367773, 367807), 'pulumi.set', 'pulumi.set', (['__self__', '"""port"""', 'port'], {}), "(__self__, 'port', port)\n", (367783, 367807), False, 'import pulumi\n'), ((368460, 368484), 'pulumi.get', 'pulumi.get', (['self', '"""name"""'], {}), "(self, 'name')\n", (368470, 368484), False, 'import pulumi\n'), ((368557, 368588), 'pulumi.set', 'pulumi.set', (['self', '"""name"""', 'value'], {}), "(self, 'name', value)\n", (368567, 368588), False, 'import pulumi\n'), ((368738, 368762), 'pulumi.get', 'pulumi.get', (['self', '"""port"""'], {}), "(self, 'port')\n", (368748, 368762), False, 'import pulumi\n'), ((368835, 368866), 'pulumi.set', 'pulumi.set', (['self', '"""port"""', 'value'], {}), "(self, 'port', value)\n", (368845, 368866), False, 'import pulumi\n'), ((369165, 369206), 'pulumi.get', 'pulumi.get', (['self', '"""external_address_spec"""'], {}), "(self, 'external_address_spec')\n", (369175, 369206), False, 'import pulumi\n'), ((369374, 369422), 'pulumi.set', 'pulumi.set', (['self', '"""external_address_spec"""', 'value'], {}), "(self, 'external_address_spec', value)\n", (369384, 369422), False, 'import pulumi\n'), ((369721, 369762), 'pulumi.get', 'pulumi.get', (['self', '"""internal_address_spec"""'], {}), "(self, 'internal_address_spec')\n", (369731, 369762), False, 'import pulumi\n'), ((369930, 369978), 'pulumi.set', 'pulumi.set', (['self', '"""internal_address_spec"""', 'value'], {}), "(self, 'internal_address_spec', value)\n", (369940, 369978), False, 'import pulumi\n'), ((370181, 370209), 'pulumi.get', 'pulumi.get', (['self', '"""protocol"""'], {}), "(self, 'protocol')\n", (370191, 370209), False, 'import pulumi\n'), ((370300, 370335), 'pulumi.set', 'pulumi.set', (['self', '"""protocol"""', 'value'], {}), "(self, 'protocol', value)\n", (370310, 370335), False, 'import pulumi\n'), ((370556, 370587), 'pulumi.get', 'pulumi.get', (['self', '"""target_port"""'], {}), "(self, 'target_port')\n", (370566, 370587), False, 'import pulumi\n'), ((370684, 370722), 'pulumi.set', 'pulumi.set', (['self', '"""target_port"""', 'value'], {}), "(self, 'target_port', value)\n", (370694, 370722), False, 'import pulumi\n'), ((371803, 371830), 'pulumi.get', 'pulumi.get', (['self', '"""address"""'], {}), "(self, 'address')\n", (371813, 371830), False, 'import pulumi\n'), ((371919, 371953), 'pulumi.set', 'pulumi.set', (['self', '"""address"""', 'value'], {}), "(self, 'address', value)\n", (371929, 371953), False, 'import pulumi\n'), ((372232, 372262), 'pulumi.get', 'pulumi.get', (['self', '"""ip_version"""'], {}), "(self, 'ip_version')\n", (372242, 372262), False, 'import pulumi\n'), ((372357, 372394), 'pulumi.set', 'pulumi.set', (['self', '"""ip_version"""', 'value'], {}), "(self, 'ip_version', value)\n", (372367, 372394), False, 'import pulumi\n'), ((373172, 373216), 'pulumi.set', 'pulumi.set', (['__self__', '"""subnet_id"""', 'subnet_id'], {}), "(__self__, 'subnet_id', subnet_id)\n", (373182, 373216), False, 'import pulumi\n'), ((373599, 373628), 'pulumi.get', 'pulumi.get', (['self', '"""subnet_id"""'], {}), "(self, 'subnet_id')\n", (373609, 373628), False, 'import pulumi\n'), ((373711, 373747), 'pulumi.set', 'pulumi.set', (['self', '"""subnet_id"""', 'value'], {}), "(self, 'subnet_id', value)\n", (373721, 373747), False, 'import pulumi\n'), ((374030, 374057), 'pulumi.get', 'pulumi.get', (['self', '"""address"""'], {}), "(self, 'address')\n", (374040, 374057), False, 'import pulumi\n'), ((374146, 374180), 'pulumi.set', 'pulumi.set', (['self', '"""address"""', 'value'], {}), "(self, 'address', value)\n", (374156, 374180), False, 'import pulumi\n'), ((374459, 374489), 'pulumi.get', 'pulumi.get', (['self', '"""ip_version"""'], {}), "(self, 'ip_version')\n", (374469, 374489), False, 'import pulumi\n'), ((374584, 374621), 'pulumi.set', 'pulumi.set', (['self', '"""ip_version"""', 'value'], {}), "(self, 'ip_version', value)\n", (374594, 374621), False, 'import pulumi\n'), ((375109, 375149), 'pulumi.set', 'pulumi.set', (['__self__', '"""address"""', 'address'], {}), "(__self__, 'address', address)\n", (375119, 375149), False, 'import pulumi\n'), ((375158, 375202), 'pulumi.set', 'pulumi.set', (['__self__', '"""subnet_id"""', 'subnet_id'], {}), "(__self__, 'subnet_id', subnet_id)\n", (375168, 375202), False, 'import pulumi\n'), ((375354, 375381), 'pulumi.get', 'pulumi.get', (['self', '"""address"""'], {}), "(self, 'address')\n", (375364, 375381), False, 'import pulumi\n'), ((375460, 375494), 'pulumi.set', 'pulumi.set', (['self', '"""address"""', 'value'], {}), "(self, 'address', value)\n", (375470, 375494), False, 'import pulumi\n'), ((375800, 375829), 'pulumi.get', 'pulumi.get', (['self', '"""subnet_id"""'], {}), "(self, 'subnet_id')\n", (375810, 375829), False, 'import pulumi\n'), ((375912, 375948), 'pulumi.set', 'pulumi.set', (['self', '"""subnet_id"""', 'value'], {}), "(self, 'subnet_id', value)\n", (375922, 375948), False, 'import pulumi\n'), ((377314, 377343), 'pulumi.get', 'pulumi.get', (['self', '"""data_lens"""'], {}), "(self, 'data_lens')\n", (377324, 377343), False, 'import pulumi\n'), ((377437, 377473), 'pulumi.set', 'pulumi.set', (['self', '"""data_lens"""', 'value'], {}), "(self, 'data_lens', value)\n", (377447, 377473), False, 'import pulumi\n'), ((377676, 377703), 'pulumi.get', 'pulumi.get', (['self', '"""metrika"""'], {}), "(self, 'metrika')\n", (377686, 377703), False, 'import pulumi\n'), ((377793, 377827), 'pulumi.set', 'pulumi.set', (['self', '"""metrika"""', 'value'], {}), "(self, 'metrika', value)\n", (377803, 377827), False, 'import pulumi\n'), ((378029, 378059), 'pulumi.get', 'pulumi.get', (['self', '"""serverless"""'], {}), "(self, 'serverless')\n", (378039, 378059), False, 'import pulumi\n'), ((378155, 378192), 'pulumi.set', 'pulumi.set', (['self', '"""serverless"""', 'value'], {}), "(self, 'serverless', value)\n", (378165, 378192), False, 'import pulumi\n'), ((378403, 378430), 'pulumi.get', 'pulumi.get', (['self', '"""web_sql"""'], {}), "(self, 'web_sql')\n", (378413, 378430), False, 'import pulumi\n'), ((378520, 378554), 'pulumi.set', 'pulumi.set', (['self', '"""web_sql"""', 'value'], {}), "(self, 'web_sql', value)\n", (378530, 378554), False, 'import pulumi\n'), ((379309, 379334), 'pulumi.get', 'pulumi.get', (['self', '"""hours"""'], {}), "(self, 'hours')\n", (379319, 379334), False, 'import pulumi\n'), ((379419, 379451), 'pulumi.set', 'pulumi.set', (['self', '"""hours"""', 'value'], {}), "(self, 'hours', value)\n", (379429, 379451), False, 'import pulumi\n'), ((379631, 379658), 'pulumi.get', 'pulumi.get', (['self', '"""minutes"""'], {}), "(self, 'minutes')\n", (379641, 379658), False, 'import pulumi\n'), ((379747, 379781), 'pulumi.set', 'pulumi.set', (['self', '"""minutes"""', 'value'], {}), "(self, 'minutes', value)\n", (379757, 379781), False, 'import pulumi\n'), ((380395, 380439), 'pulumi.set', 'pulumi.set', (['__self__', '"""resources"""', 'resources'], {}), "(__self__, 'resources', resources)\n", (380405, 380439), False, 'import pulumi\n'), ((380784, 380813), 'pulumi.get', 'pulumi.get', (['self', '"""resources"""'], {}), "(self, 'resources')\n", (380794, 380813), False, 'import pulumi\n'), ((380938, 380974), 'pulumi.set', 'pulumi.set', (['self', '"""resources"""', 'value'], {}), "(self, 'resources', value)\n", (380948, 380974), False, 'import pulumi\n'), ((381187, 381213), 'pulumi.get', 'pulumi.get', (['self', '"""config"""'], {}), "(self, 'config')\n", (381197, 381213), False, 'import pulumi\n'), ((381339, 381372), 'pulumi.set', 'pulumi.set', (['self', '"""config"""', 'value'], {}), "(self, 'config', value)\n", (381349, 381372), False, 'import pulumi\n'), ((389898, 389938), 'pulumi.get', 'pulumi.get', (['self', '"""background_pool_size"""'], {}), "(self, 'background_pool_size')\n", (389908, 389938), False, 'import pulumi\n'), ((390053, 390100), 'pulumi.set', 'pulumi.set', (['self', '"""background_pool_size"""', 'value'], {}), "(self, 'background_pool_size', value)\n", (390063, 390100), False, 'import pulumi\n'), ((390261, 390310), 'pulumi.get', 'pulumi.get', (['self', '"""background_schedule_pool_size"""'], {}), "(self, 'background_schedule_pool_size')\n", (390271, 390310), False, 'import pulumi\n'), ((390443, 390499), 'pulumi.set', 'pulumi.set', (['self', '"""background_schedule_pool_size"""', 'value'], {}), "(self, 'background_schedule_pool_size', value)\n", (390453, 390499), False, 'import pulumi\n'), ((390781, 390813), 'pulumi.get', 'pulumi.get', (['self', '"""compressions"""'], {}), "(self, 'compressions')\n", (390791, 390813), False, 'import pulumi\n'), ((390986, 391025), 'pulumi.set', 'pulumi.set', (['self', '"""compressions"""', 'value'], {}), "(self, 'compressions', value)\n", (390996, 391025), False, 'import pulumi\n'), ((391152, 391183), 'pulumi.get', 'pulumi.get', (['self', '"""geobase_uri"""'], {}), "(self, 'geobase_uri')\n", (391162, 391183), False, 'import pulumi\n'), ((391280, 391318), 'pulumi.set', 'pulumi.set', (['self', '"""geobase_uri"""', 'value'], {}), "(self, 'geobase_uri', value)\n", (391290, 391318), False, 'import pulumi\n'), ((391630, 391666), 'pulumi.get', 'pulumi.get', (['self', '"""graphite_rollups"""'], {}), "(self, 'graphite_rollups')\n", (391640, 391666), False, 'import pulumi\n'), ((391850, 391893), 'pulumi.set', 'pulumi.set', (['self', '"""graphite_rollups"""', 'value'], {}), "(self, 'graphite_rollups', value)\n", (391860, 391893), False, 'import pulumi\n'), ((392138, 392163), 'pulumi.get', 'pulumi.get', (['self', '"""kafka"""'], {}), "(self, 'kafka')\n", (392148, 392163), False, 'import pulumi\n'), ((392292, 392324), 'pulumi.set', 'pulumi.set', (['self', '"""kafka"""', 'value'], {}), "(self, 'kafka', value)\n", (392302, 392324), False, 'import pulumi\n'), ((392631, 392663), 'pulumi.get', 'pulumi.get', (['self', '"""kafka_topics"""'], {}), "(self, 'kafka_topics')\n", (392641, 392663), False, 'import pulumi\n'), ((392835, 392874), 'pulumi.set', 'pulumi.set', (['self', '"""kafka_topics"""', 'value'], {}), "(self, 'kafka_topics', value)\n", (392845, 392874), False, 'import pulumi\n'), ((393014, 393052), 'pulumi.get', 'pulumi.get', (['self', '"""keep_alive_timeout"""'], {}), "(self, 'keep_alive_timeout')\n", (393024, 393052), False, 'import pulumi\n'), ((393163, 393208), 'pulumi.set', 'pulumi.set', (['self', '"""keep_alive_timeout"""', 'value'], {}), "(self, 'keep_alive_timeout', value)\n", (393173, 393208), False, 'import pulumi\n'), ((393331, 393360), 'pulumi.get', 'pulumi.get', (['self', '"""log_level"""'], {}), "(self, 'log_level')\n", (393341, 393360), False, 'import pulumi\n'), ((393453, 393489), 'pulumi.set', 'pulumi.set', (['self', '"""log_level"""', 'value'], {}), "(self, 'log_level', value)\n", (393463, 393489), False, 'import pulumi\n'), ((393623, 393658), 'pulumi.get', 'pulumi.get', (['self', '"""mark_cache_size"""'], {}), "(self, 'mark_cache_size')\n", (393633, 393658), False, 'import pulumi\n'), ((393763, 393805), 'pulumi.set', 'pulumi.set', (['self', '"""mark_cache_size"""', 'value'], {}), "(self, 'mark_cache_size', value)\n", (393773, 393805), False, 'import pulumi\n'), ((393953, 393995), 'pulumi.get', 'pulumi.get', (['self', '"""max_concurrent_queries"""'], {}), "(self, 'max_concurrent_queries')\n", (393963, 393995), False, 'import pulumi\n'), ((394114, 394163), 'pulumi.set', 'pulumi.set', (['self', '"""max_concurrent_queries"""', 'value'], {}), "(self, 'max_concurrent_queries', value)\n", (394124, 394163), False, 'import pulumi\n'), ((394298, 394333), 'pulumi.get', 'pulumi.get', (['self', '"""max_connections"""'], {}), "(self, 'max_connections')\n", (394308, 394333), False, 'import pulumi\n'), ((394438, 394480), 'pulumi.set', 'pulumi.set', (['self', '"""max_connections"""', 'value'], {}), "(self, 'max_connections', value)\n", (394448, 394480), False, 'import pulumi\n'), ((394634, 394680), 'pulumi.get', 'pulumi.get', (['self', '"""max_partition_size_to_drop"""'], {}), "(self, 'max_partition_size_to_drop')\n", (394644, 394680), False, 'import pulumi\n'), ((394807, 394860), 'pulumi.set', 'pulumi.set', (['self', '"""max_partition_size_to_drop"""', 'value'], {}), "(self, 'max_partition_size_to_drop', value)\n", (394817, 394860), False, 'import pulumi\n'), ((395006, 395048), 'pulumi.get', 'pulumi.get', (['self', '"""max_table_size_to_drop"""'], {}), "(self, 'max_table_size_to_drop')\n", (395016, 395048), False, 'import pulumi\n'), ((395167, 395216), 'pulumi.set', 'pulumi.set', (['self', '"""max_table_size_to_drop"""', 'value'], {}), "(self, 'max_table_size_to_drop', value)\n", (395177, 395216), False, 'import pulumi\n'), ((395488, 395518), 'pulumi.get', 'pulumi.get', (['self', '"""merge_tree"""'], {}), "(self, 'merge_tree')\n", (395498, 395518), False, 'import pulumi\n'), ((395661, 395698), 'pulumi.set', 'pulumi.set', (['self', '"""merge_tree"""', 'value'], {}), "(self, 'merge_tree', value)\n", (395671, 395698), False, 'import pulumi\n'), ((395839, 395877), 'pulumi.get', 'pulumi.get', (['self', '"""metric_log_enabled"""'], {}), "(self, 'metric_log_enabled')\n", (395849, 395877), False, 'import pulumi\n'), ((395989, 396034), 'pulumi.set', 'pulumi.set', (['self', '"""metric_log_enabled"""', 'value'], {}), "(self, 'metric_log_enabled', value)\n", (395999, 396034), False, 'import pulumi\n'), ((396187, 396232), 'pulumi.get', 'pulumi.get', (['self', '"""metric_log_retention_size"""'], {}), "(self, 'metric_log_retention_size')\n", (396197, 396232), False, 'import pulumi\n'), ((396357, 396409), 'pulumi.set', 'pulumi.set', (['self', '"""metric_log_retention_size"""', 'value'], {}), "(self, 'metric_log_retention_size', value)\n", (396367, 396409), False, 'import pulumi\n'), ((396562, 396607), 'pulumi.get', 'pulumi.get', (['self', '"""metric_log_retention_time"""'], {}), "(self, 'metric_log_retention_time')\n", (396572, 396607), False, 'import pulumi\n'), ((396732, 396784), 'pulumi.set', 'pulumi.set', (['self', '"""metric_log_retention_time"""', 'value'], {}), "(self, 'metric_log_retention_time', value)\n", (396742, 396784), False, 'import pulumi\n'), ((396933, 396976), 'pulumi.get', 'pulumi.get', (['self', '"""part_log_retention_size"""'], {}), "(self, 'part_log_retention_size')\n", (396943, 396976), False, 'import pulumi\n'), ((397097, 397147), 'pulumi.set', 'pulumi.set', (['self', '"""part_log_retention_size"""', 'value'], {}), "(self, 'part_log_retention_size', value)\n", (397107, 397147), False, 'import pulumi\n'), ((397296, 397339), 'pulumi.get', 'pulumi.get', (['self', '"""part_log_retention_time"""'], {}), "(self, 'part_log_retention_time')\n", (397306, 397339), False, 'import pulumi\n'), ((397460, 397510), 'pulumi.set', 'pulumi.set', (['self', '"""part_log_retention_time"""', 'value'], {}), "(self, 'part_log_retention_time', value)\n", (397470, 397510), False, 'import pulumi\n'), ((397661, 397705), 'pulumi.get', 'pulumi.get', (['self', '"""query_log_retention_size"""'], {}), "(self, 'query_log_retention_size')\n", (397671, 397705), False, 'import pulumi\n'), ((397828, 397879), 'pulumi.set', 'pulumi.set', (['self', '"""query_log_retention_size"""', 'value'], {}), "(self, 'query_log_retention_size', value)\n", (397838, 397879), False, 'import pulumi\n'), ((398030, 398074), 'pulumi.get', 'pulumi.get', (['self', '"""query_log_retention_time"""'], {}), "(self, 'query_log_retention_time')\n", (398040, 398074), False, 'import pulumi\n'), ((398197, 398248), 'pulumi.set', 'pulumi.set', (['self', '"""query_log_retention_time"""', 'value'], {}), "(self, 'query_log_retention_time', value)\n", (398207, 398248), False, 'import pulumi\n'), ((398400, 398444), 'pulumi.get', 'pulumi.get', (['self', '"""query_thread_log_enabled"""'], {}), "(self, 'query_thread_log_enabled')\n", (398410, 398444), False, 'import pulumi\n'), ((398568, 398619), 'pulumi.set', 'pulumi.set', (['self', '"""query_thread_log_enabled"""', 'value'], {}), "(self, 'query_thread_log_enabled', value)\n", (398578, 398619), False, 'import pulumi\n'), ((398783, 398834), 'pulumi.get', 'pulumi.get', (['self', '"""query_thread_log_retention_size"""'], {}), "(self, 'query_thread_log_retention_size')\n", (398793, 398834), False, 'import pulumi\n'), ((398971, 399029), 'pulumi.set', 'pulumi.set', (['self', '"""query_thread_log_retention_size"""', 'value'], {}), "(self, 'query_thread_log_retention_size', value)\n", (398981, 399029), False, 'import pulumi\n'), ((399193, 399244), 'pulumi.get', 'pulumi.get', (['self', '"""query_thread_log_retention_time"""'], {}), "(self, 'query_thread_log_retention_time')\n", (399203, 399244), False, 'import pulumi\n'), ((399381, 399439), 'pulumi.set', 'pulumi.set', (['self', '"""query_thread_log_retention_time"""', 'value'], {}), "(self, 'query_thread_log_retention_time', value)\n", (399391, 399439), False, 'import pulumi\n'), ((399693, 399721), 'pulumi.get', 'pulumi.get', (['self', '"""rabbitmq"""'], {}), "(self, 'rabbitmq')\n", (399703, 399721), False, 'import pulumi\n'), ((399859, 399894), 'pulumi.set', 'pulumi.set', (['self', '"""rabbitmq"""', 'value'], {}), "(self, 'rabbitmq', value)\n", (399869, 399894), False, 'import pulumi\n'), ((400031, 400067), 'pulumi.get', 'pulumi.get', (['self', '"""text_log_enabled"""'], {}), "(self, 'text_log_enabled')\n", (400041, 400067), False, 'import pulumi\n'), ((400175, 400218), 'pulumi.set', 'pulumi.set', (['self', '"""text_log_enabled"""', 'value'], {}), "(self, 'text_log_enabled', value)\n", (400185, 400218), False, 'import pulumi\n'), ((400350, 400384), 'pulumi.get', 'pulumi.get', (['self', '"""text_log_level"""'], {}), "(self, 'text_log_level')\n", (400360, 400384), False, 'import pulumi\n'), ((400487, 400528), 'pulumi.set', 'pulumi.set', (['self', '"""text_log_level"""', 'value'], {}), "(self, 'text_log_level', value)\n", (400497, 400528), False, 'import pulumi\n'), ((400677, 400720), 'pulumi.get', 'pulumi.get', (['self', '"""text_log_retention_size"""'], {}), "(self, 'text_log_retention_size')\n", (400687, 400720), False, 'import pulumi\n'), ((400841, 400891), 'pulumi.set', 'pulumi.set', (['self', '"""text_log_retention_size"""', 'value'], {}), "(self, 'text_log_retention_size', value)\n", (400851, 400891), False, 'import pulumi\n'), ((401040, 401083), 'pulumi.get', 'pulumi.get', (['self', '"""text_log_retention_time"""'], {}), "(self, 'text_log_retention_time')\n", (401050, 401083), False, 'import pulumi\n'), ((401204, 401254), 'pulumi.set', 'pulumi.set', (['self', '"""text_log_retention_time"""', 'value'], {}), "(self, 'text_log_retention_time', value)\n", (401214, 401254), False, 'import pulumi\n'), ((401359, 401387), 'pulumi.get', 'pulumi.get', (['self', '"""timezone"""'], {}), "(self, 'timezone')\n", (401369, 401387), False, 'import pulumi\n'), ((401478, 401513), 'pulumi.set', 'pulumi.set', (['self', '"""timezone"""', 'value'], {}), "(self, 'timezone', value)\n", (401488, 401513), False, 'import pulumi\n'), ((401652, 401689), 'pulumi.get', 'pulumi.get', (['self', '"""trace_log_enabled"""'], {}), "(self, 'trace_log_enabled')\n", (401662, 401689), False, 'import pulumi\n'), ((401799, 401843), 'pulumi.set', 'pulumi.set', (['self', '"""trace_log_enabled"""', 'value'], {}), "(self, 'trace_log_enabled', value)\n", (401809, 401843), False, 'import pulumi\n'), ((401994, 402038), 'pulumi.get', 'pulumi.get', (['self', '"""trace_log_retention_size"""'], {}), "(self, 'trace_log_retention_size')\n", (402004, 402038), False, 'import pulumi\n'), ((402161, 402212), 'pulumi.set', 'pulumi.set', (['self', '"""trace_log_retention_size"""', 'value'], {}), "(self, 'trace_log_retention_size', value)\n", (402171, 402212), False, 'import pulumi\n'), ((402363, 402407), 'pulumi.get', 'pulumi.get', (['self', '"""trace_log_retention_time"""'], {}), "(self, 'trace_log_retention_time')\n", (402373, 402407), False, 'import pulumi\n'), ((402530, 402581), 'pulumi.set', 'pulumi.set', (['self', '"""trace_log_retention_time"""', 'value'], {}), "(self, 'trace_log_retention_time', value)\n", (402540, 402581), False, 'import pulumi\n'), ((402731, 402774), 'pulumi.get', 'pulumi.get', (['self', '"""uncompressed_cache_size"""'], {}), "(self, 'uncompressed_cache_size')\n", (402741, 402774), False, 'import pulumi\n'), ((402895, 402945), 'pulumi.set', 'pulumi.set', (['self', '"""uncompressed_cache_size"""', 'value'], {}), "(self, 'uncompressed_cache_size', value)\n", (402905, 402945), False, 'import pulumi\n'), ((403826, 403864), 'pulumi.set', 'pulumi.set', (['__self__', '"""method"""', 'method'], {}), "(__self__, 'method', method)\n", (403836, 403864), False, 'import pulumi\n'), ((403873, 403925), 'pulumi.set', 'pulumi.set', (['__self__', '"""min_part_size"""', 'min_part_size'], {}), "(__self__, 'min_part_size', min_part_size)\n", (403883, 403925), False, 'import pulumi\n'), ((403934, 403998), 'pulumi.set', 'pulumi.set', (['__self__', '"""min_part_size_ratio"""', 'min_part_size_ratio'], {}), "(__self__, 'min_part_size_ratio', min_part_size_ratio)\n", (403944, 403998), False, 'import pulumi\n'), ((404192, 404218), 'pulumi.get', 'pulumi.get', (['self', '"""method"""'], {}), "(self, 'method')\n", (404202, 404218), False, 'import pulumi\n'), ((404295, 404328), 'pulumi.set', 'pulumi.set', (['self', '"""method"""', 'value'], {}), "(self, 'method', value)\n", (404305, 404328), False, 'import pulumi\n'), ((404655, 404688), 'pulumi.get', 'pulumi.get', (['self', '"""min_part_size"""'], {}), "(self, 'min_part_size')\n", (404665, 404688), False, 'import pulumi\n'), ((404779, 404819), 'pulumi.set', 'pulumi.set', (['self', '"""min_part_size"""', 'value'], {}), "(self, 'min_part_size', value)\n", (404789, 404819), False, 'import pulumi\n'), ((405178, 405217), 'pulumi.get', 'pulumi.get', (['self', '"""min_part_size_ratio"""'], {}), "(self, 'min_part_size_ratio')\n", (405188, 405217), False, 'import pulumi\n'), ((405322, 405368), 'pulumi.set', 'pulumi.set', (['self', '"""min_part_size_ratio"""', 'value'], {}), "(self, 'min_part_size_ratio', value)\n", (405332, 405368), False, 'import pulumi\n'), ((405929, 405963), 'pulumi.set', 'pulumi.set', (['__self__', '"""name"""', 'name'], {}), "(__self__, 'name', name)\n", (405939, 405963), False, 'import pulumi\n'), ((406210, 406234), 'pulumi.get', 'pulumi.get', (['self', '"""name"""'], {}), "(self, 'name')\n", (406220, 406234), False, 'import pulumi\n'), ((406307, 406338), 'pulumi.set', 'pulumi.set', (['self', '"""name"""', 'value'], {}), "(self, 'name', value)\n", (406317, 406338), False, 'import pulumi\n'), ((406582, 406610), 'pulumi.get', 'pulumi.get', (['self', '"""patterns"""'], {}), "(self, 'patterns')\n", (406592, 406610), False, 'import pulumi\n'), ((406785, 406820), 'pulumi.set', 'pulumi.set', (['self', '"""patterns"""', 'value'], {}), "(self, 'patterns', value)\n", (406795, 406820), False, 'import pulumi\n'), ((407559, 407601), 'pulumi.set', 'pulumi.set', (['__self__', '"""function"""', 'function'], {}), "(__self__, 'function', function)\n", (407569, 407601), False, 'import pulumi\n'), ((407931, 407959), 'pulumi.get', 'pulumi.get', (['self', '"""function"""'], {}), "(self, 'function')\n", (407941, 407959), False, 'import pulumi\n'), ((408040, 408075), 'pulumi.set', 'pulumi.set', (['self', '"""function"""', 'value'], {}), "(self, 'function', value)\n", (408050, 408075), False, 'import pulumi\n'), ((408262, 408288), 'pulumi.get', 'pulumi.get', (['self', '"""regexp"""'], {}), "(self, 'regexp')\n", (408272, 408288), False, 'import pulumi\n'), ((408375, 408408), 'pulumi.set', 'pulumi.set', (['self', '"""regexp"""', 'value'], {}), "(self, 'regexp', value)\n", (408385, 408408), False, 'import pulumi\n'), ((408659, 408689), 'pulumi.get', 'pulumi.get', (['self', '"""retentions"""'], {}), "(self, 'retentions')\n", (408669, 408689), False, 'import pulumi\n'), ((408877, 408914), 'pulumi.set', 'pulumi.set', (['self', '"""retentions"""', 'value'], {}), "(self, 'retentions', value)\n", (408887, 408914), False, 'import pulumi\n'), ((409332, 409364), 'pulumi.set', 'pulumi.set', (['__self__', '"""age"""', 'age'], {}), "(__self__, 'age', age)\n", (409342, 409364), False, 'import pulumi\n'), ((409373, 409417), 'pulumi.set', 'pulumi.set', (['__self__', '"""precision"""', 'precision'], {}), "(__self__, 'precision', precision)\n", (409383, 409417), False, 'import pulumi\n'), ((409568, 409591), 'pulumi.get', 'pulumi.get', (['self', '"""age"""'], {}), "(self, 'age')\n", (409578, 409591), False, 'import pulumi\n'), ((409662, 409692), 'pulumi.set', 'pulumi.set', (['self', '"""age"""', 'value'], {}), "(self, 'age', value)\n", (409672, 409692), False, 'import pulumi\n'), ((409876, 409905), 'pulumi.get', 'pulumi.get', (['self', '"""precision"""'], {}), "(self, 'precision')\n", (409886, 409905), False, 'import pulumi\n'), ((409988, 410024), 'pulumi.set', 'pulumi.set', (['self', '"""precision"""', 'value'], {}), "(self, 'precision', value)\n", (409998, 410024), False, 'import pulumi\n'), ((411414, 411448), 'pulumi.get', 'pulumi.get', (['self', '"""sasl_mechanism"""'], {}), "(self, 'sasl_mechanism')\n", (411424, 411448), False, 'import pulumi\n'), ((411551, 411592), 'pulumi.set', 'pulumi.set', (['self', '"""sasl_mechanism"""', 'value'], {}), "(self, 'sasl_mechanism', value)\n", (411561, 411592), False, 'import pulumi\n'), ((411786, 411819), 'pulumi.get', 'pulumi.get', (['self', '"""sasl_password"""'], {}), "(self, 'sasl_password')\n", (411796, 411819), False, 'import pulumi\n'), ((411920, 411960), 'pulumi.set', 'pulumi.set', (['self', '"""sasl_password"""', 'value'], {}), "(self, 'sasl_password', value)\n", (411930, 411960), False, 'import pulumi\n'), ((412149, 412182), 'pulumi.get', 'pulumi.get', (['self', '"""sasl_username"""'], {}), "(self, 'sasl_username')\n", (412159, 412182), False, 'import pulumi\n'), ((412283, 412323), 'pulumi.set', 'pulumi.set', (['self', '"""sasl_username"""', 'value'], {}), "(self, 'sasl_username', value)\n", (412293, 412323), False, 'import pulumi\n'), ((412545, 412582), 'pulumi.get', 'pulumi.get', (['self', '"""security_protocol"""'], {}), "(self, 'security_protocol')\n", (412555, 412582), False, 'import pulumi\n'), ((412691, 412735), 'pulumi.set', 'pulumi.set', (['self', '"""security_protocol"""', 'value'], {}), "(self, 'security_protocol', value)\n", (412701, 412735), False, 'import pulumi\n'), ((413264, 413298), 'pulumi.set', 'pulumi.set', (['__self__', '"""name"""', 'name'], {}), "(__self__, 'name', name)\n", (413274, 413298), False, 'import pulumi\n'), ((413545, 413569), 'pulumi.get', 'pulumi.get', (['self', '"""name"""'], {}), "(self, 'name')\n", (413555, 413569), False, 'import pulumi\n'), ((413642, 413673), 'pulumi.set', 'pulumi.set', (['self', '"""name"""', 'value'], {}), "(self, 'name', value)\n", (413652, 413673), False, 'import pulumi\n'), ((413916, 413944), 'pulumi.get', 'pulumi.get', (['self', '"""settings"""'], {}), "(self, 'settings')\n", (413926, 413944), False, 'import pulumi\n'), ((414092, 414127), 'pulumi.set', 'pulumi.set', (['self', '"""settings"""', 'value'], {}), "(self, 'settings', value)\n", (414102, 414127), False, 'import pulumi\n'), ((415530, 415564), 'pulumi.get', 'pulumi.get', (['self', '"""sasl_mechanism"""'], {}), "(self, 'sasl_mechanism')\n", (415540, 415564), False, 'import pulumi\n'), ((415667, 415708), 'pulumi.set', 'pulumi.set', (['self', '"""sasl_mechanism"""', 'value'], {}), "(self, 'sasl_mechanism', value)\n", (415677, 415708), False, 'import pulumi\n'), ((415902, 415935), 'pulumi.get', 'pulumi.get', (['self', '"""sasl_password"""'], {}), "(self, 'sasl_password')\n", (415912, 415935), False, 'import pulumi\n'), ((416036, 416076), 'pulumi.set', 'pulumi.set', (['self', '"""sasl_password"""', 'value'], {}), "(self, 'sasl_password', value)\n", (416046, 416076), False, 'import pulumi\n'), ((416265, 416298), 'pulumi.get', 'pulumi.get', (['self', '"""sasl_username"""'], {}), "(self, 'sasl_username')\n", (416275, 416298), False, 'import pulumi\n'), ((416399, 416439), 'pulumi.set', 'pulumi.set', (['self', '"""sasl_username"""', 'value'], {}), "(self, 'sasl_username', value)\n", (416409, 416439), False, 'import pulumi\n'), ((416661, 416698), 'pulumi.get', 'pulumi.get', (['self', '"""security_protocol"""'], {}), "(self, 'security_protocol')\n", (416671, 416698), False, 'import pulumi\n'), ((416807, 416851), 'pulumi.set', 'pulumi.set', (['self', '"""security_protocol"""', 'value'], {}), "(self, 'security_protocol', value)\n", (416817, 416851), False, 'import pulumi\n'), ((420753, 420812), 'pulumi.get', 'pulumi.get', (['self', '"""max_bytes_to_merge_at_min_space_in_pool"""'], {}), "(self, 'max_bytes_to_merge_at_min_space_in_pool')\n", (420763, 420812), False, 'import pulumi\n'), ((420965, 421031), 'pulumi.set', 'pulumi.set', (['self', '"""max_bytes_to_merge_at_min_space_in_pool"""', 'value'], {}), "(self, 'max_bytes_to_merge_at_min_space_in_pool', value)\n", (420975, 421031), False, 'import pulumi\n'), ((421350, 421400), 'pulumi.get', 'pulumi.get', (['self', '"""max_replicated_merges_in_queue"""'], {}), "(self, 'max_replicated_merges_in_queue')\n", (421360, 421400), False, 'import pulumi\n'), ((421535, 421592), 'pulumi.set', 'pulumi.set', (['self', '"""max_replicated_merges_in_queue"""', 'value'], {}), "(self, 'max_replicated_merges_in_queue', value)\n", (421545, 421592), False, 'import pulumi\n'), ((422148, 422225), 'pulumi.get', 'pulumi.get', (['self', '"""number_of_free_entries_in_pool_to_lower_max_size_of_merge"""'], {}), "(self, 'number_of_free_entries_in_pool_to_lower_max_size_of_merge')\n", (422158, 422225), False, 'import pulumi\n'), ((422414, 422502), 'pulumi.set', 'pulumi.set', (['self', '"""number_of_free_entries_in_pool_to_lower_max_size_of_merge"""', 'value'], {}), "(self,\n 'number_of_free_entries_in_pool_to_lower_max_size_of_merge', value)\n", (422424, 422502), False, 'import pulumi\n'), ((422838, 422879), 'pulumi.get', 'pulumi.get', (['self', '"""parts_to_delay_insert"""'], {}), "(self, 'parts_to_delay_insert')\n", (422848, 422879), False, 'import pulumi\n'), ((422996, 423044), 'pulumi.set', 'pulumi.set', (['self', '"""parts_to_delay_insert"""', 'value'], {}), "(self, 'parts_to_delay_insert', value)\n", (423006, 423044), False, 'import pulumi\n'), ((423366, 423407), 'pulumi.get', 'pulumi.get', (['self', '"""parts_to_throw_insert"""'], {}), "(self, 'parts_to_throw_insert')\n", (423376, 423407), False, 'import pulumi\n'), ((423524, 423572), 'pulumi.set', 'pulumi.set', (['self', '"""parts_to_throw_insert"""', 'value'], {}), "(self, 'parts_to_throw_insert', value)\n", (423534, 423572), False, 'import pulumi\n'), ((423890, 423941), 'pulumi.get', 'pulumi.get', (['self', '"""replicated_deduplication_window"""'], {}), "(self, 'replicated_deduplication_window')\n", (423900, 423941), False, 'import pulumi\n'), ((424078, 424136), 'pulumi.set', 'pulumi.set', (['self', '"""replicated_deduplication_window"""', 'value'], {}), "(self, 'replicated_deduplication_window', value)\n", (424088, 424136), False, 'import pulumi\n'), ((424472, 424531), 'pulumi.get', 'pulumi.get', (['self', '"""replicated_deduplication_window_seconds"""'], {}), "(self, 'replicated_deduplication_window_seconds')\n", (424482, 424531), False, 'import pulumi\n'), ((424684, 424750), 'pulumi.set', 'pulumi.set', (['self', '"""replicated_deduplication_window_seconds"""', 'value'], {}), "(self, 'replicated_deduplication_window_seconds', value)\n", (424694, 424750), False, 'import pulumi\n'), ((425462, 425490), 'pulumi.get', 'pulumi.get', (['self', '"""password"""'], {}), "(self, 'password')\n", (425472, 425490), False, 'import pulumi\n'), ((425581, 425616), 'pulumi.set', 'pulumi.set', (['self', '"""password"""', 'value'], {}), "(self, 'password', value)\n", (425591, 425616), False, 'import pulumi\n'), ((425772, 425800), 'pulumi.get', 'pulumi.get', (['self', '"""username"""'], {}), "(self, 'username')\n", (425782, 425800), False, 'import pulumi\n'), ((425891, 425926), 'pulumi.set', 'pulumi.set', (['self', '"""username"""', 'value'], {}), "(self, 'username', value)\n", (425901, 425926), False, 'import pulumi\n'), ((426550, 426594), 'pulumi.set', 'pulumi.set', (['__self__', '"""disk_size"""', 'disk_size'], {}), "(__self__, 'disk_size', disk_size)\n", (426560, 426594), False, 'import pulumi\n'), ((426603, 426653), 'pulumi.set', 'pulumi.set', (['__self__', '"""disk_type_id"""', 'disk_type_id'], {}), "(__self__, 'disk_type_id', disk_type_id)\n", (426613, 426653), False, 'import pulumi\n'), ((426662, 426724), 'pulumi.set', 'pulumi.set', (['__self__', '"""resource_preset_id"""', 'resource_preset_id'], {}), "(__self__, 'resource_preset_id', resource_preset_id)\n", (426672, 426724), False, 'import pulumi\n'), ((426936, 426965), 'pulumi.get', 'pulumi.get', (['self', '"""disk_size"""'], {}), "(self, 'disk_size')\n", (426946, 426965), False, 'import pulumi\n'), ((427048, 427084), 'pulumi.set', 'pulumi.set', (['self', '"""disk_size"""', 'value'], {}), "(self, 'disk_size', value)\n", (427058, 427084), False, 'import pulumi\n'), ((427404, 427436), 'pulumi.get', 'pulumi.get', (['self', '"""disk_type_id"""'], {}), "(self, 'disk_type_id')\n", (427414, 427436), False, 'import pulumi\n'), ((427525, 427564), 'pulumi.set', 'pulumi.set', (['self', '"""disk_type_id"""', 'value'], {}), "(self, 'disk_type_id', value)\n", (427535, 427564), False, 'import pulumi\n'), ((427694, 427732), 'pulumi.get', 'pulumi.get', (['self', '"""resource_preset_id"""'], {}), "(self, 'resource_preset_id')\n", (427704, 427732), False, 'import pulumi\n'), ((427833, 427878), 'pulumi.set', 'pulumi.set', (['self', '"""resource_preset_id"""', 'value'], {}), "(self, 'resource_preset_id', value)\n", (427843, 427878), False, 'import pulumi\n'), ((428195, 428235), 'pulumi.set', 'pulumi.set', (['__self__', '"""enabled"""', 'enabled'], {}), "(__self__, 'enabled', enabled)\n", (428205, 428235), False, 'import pulumi\n'), ((428461, 428488), 'pulumi.get', 'pulumi.get', (['self', '"""enabled"""'], {}), "(self, 'enabled')\n", (428471, 428488), False, 'import pulumi\n'), ((428568, 428602), 'pulumi.set', 'pulumi.set', (['self', '"""enabled"""', 'value'], {}), "(self, 'enabled', value)\n", (428578, 428602), False, 'import pulumi\n'), ((428844, 428878), 'pulumi.set', 'pulumi.set', (['__self__', '"""name"""', 'name'], {}), "(__self__, 'name', name)\n", (428854, 428878), False, 'import pulumi\n'), ((429037, 429061), 'pulumi.get', 'pulumi.get', (['self', '"""name"""'], {}), "(self, 'name')\n", (429047, 429061), False, 'import pulumi\n'), ((429134, 429165), 'pulumi.set', 'pulumi.set', (['self', '"""name"""', 'value'], {}), "(self, 'name', value)\n", (429144, 429165), False, 'import pulumi\n'), ((429775, 429809), 'pulumi.set', 'pulumi.set', (['__self__', '"""name"""', 'name'], {}), "(__self__, 'name', name)\n", (429785, 429809), False, 'import pulumi\n'), ((429818, 429852), 'pulumi.set', 'pulumi.set', (['__self__', '"""type"""', 'type'], {}), "(__self__, 'type', type)\n", (429828, 429852), False, 'import pulumi\n'), ((429861, 429893), 'pulumi.set', 'pulumi.set', (['__self__', '"""uri"""', 'uri'], {}), "(__self__, 'uri', uri)\n", (429871, 429893), False, 'import pulumi\n'), ((430052, 430076), 'pulumi.get', 'pulumi.get', (['self', '"""name"""'], {}), "(self, 'name')\n", (430062, 430076), False, 'import pulumi\n'), ((430149, 430180), 'pulumi.set', 'pulumi.set', (['self', '"""name"""', 'value'], {}), "(self, 'name', value)\n", (430159, 430180), False, 'import pulumi\n'), ((430434, 430458), 'pulumi.get', 'pulumi.get', (['self', '"""type"""'], {}), "(self, 'type')\n", (430444, 430458), False, 'import pulumi\n'), ((430531, 430562), 'pulumi.set', 'pulumi.set', (['self', '"""type"""', 'value'], {}), "(self, 'type', value)\n", (430541, 430562), False, 'import pulumi\n'), ((430757, 430780), 'pulumi.get', 'pulumi.get', (['self', '"""uri"""'], {}), "(self, 'uri')\n", (430767, 430780), False, 'import pulumi\n'), ((430851, 430881), 'pulumi.set', 'pulumi.set', (['self', '"""uri"""', 'value'], {}), "(self, 'uri', value)\n", (430861, 430881), False, 'import pulumi\n'), ((432242, 432276), 'pulumi.set', 'pulumi.set', (['__self__', '"""type"""', 'type'], {}), "(__self__, 'type', type)\n", (432252, 432276), False, 'import pulumi\n'), ((432285, 432319), 'pulumi.set', 'pulumi.set', (['__self__', '"""zone"""', 'zone'], {}), "(__self__, 'zone', zone)\n", (432295, 432319), False, 'import pulumi\n'), ((432946, 432970), 'pulumi.get', 'pulumi.get', (['self', '"""type"""'], {}), "(self, 'type')\n", (432956, 432970), False, 'import pulumi\n'), ((433043, 433074), 'pulumi.set', 'pulumi.set', (['self', '"""type"""', 'value'], {}), "(self, 'type', value)\n", (433053, 433074), False, 'import pulumi\n'), ((433384, 433408), 'pulumi.get', 'pulumi.get', (['self', '"""zone"""'], {}), "(self, 'zone')\n", (433394, 433408), False, 'import pulumi\n'), ((433481, 433512), 'pulumi.set', 'pulumi.set', (['self', '"""zone"""', 'value'], {}), "(self, 'zone', value)\n", (433491, 433512), False, 'import pulumi\n'), ((433780, 433816), 'pulumi.get', 'pulumi.get', (['self', '"""assign_public_ip"""'], {}), "(self, 'assign_public_ip')\n", (433790, 433816), False, 'import pulumi\n'), ((433924, 433967), 'pulumi.set', 'pulumi.set', (['self', '"""assign_public_ip"""', 'value'], {}), "(self, 'assign_public_ip', value)\n", (433934, 433967), False, 'import pulumi\n'), ((434145, 434169), 'pulumi.get', 'pulumi.get', (['self', '"""fqdn"""'], {}), "(self, 'fqdn')\n", (434155, 434169), False, 'import pulumi\n'), ((434252, 434283), 'pulumi.set', 'pulumi.set', (['self', '"""fqdn"""', 'value'], {}), "(self, 'fqdn', value)\n", (434262, 434283), False, 'import pulumi\n'), ((434489, 434519), 'pulumi.get', 'pulumi.get', (['self', '"""shard_name"""'], {}), "(self, 'shard_name')\n", (434499, 434519), False, 'import pulumi\n'), ((434614, 434651), 'pulumi.set', 'pulumi.set', (['self', '"""shard_name"""', 'value'], {}), "(self, 'shard_name', value)\n", (434624, 434651), False, 'import pulumi\n'), ((434926, 434955), 'pulumi.get', 'pulumi.get', (['self', '"""subnet_id"""'], {}), "(self, 'subnet_id')\n", (434936, 434955), False, 'import pulumi\n'), ((435048, 435084), 'pulumi.set', 'pulumi.set', (['self', '"""subnet_id"""', 'value'], {}), "(self, 'subnet_id', value)\n", (435058, 435084), False, 'import pulumi\n'), ((435836, 435870), 'pulumi.set', 'pulumi.set', (['__self__', '"""type"""', 'type'], {}), "(__self__, 'type', type)\n", (435846, 435870), False, 'import pulumi\n'), ((436273, 436297), 'pulumi.get', 'pulumi.get', (['self', '"""type"""'], {}), "(self, 'type')\n", (436283, 436297), False, 'import pulumi\n'), ((436370, 436401), 'pulumi.set', 'pulumi.set', (['self', '"""type"""', 'value'], {}), "(self, 'type', value)\n", (436380, 436401), False, 'import pulumi\n'), ((436660, 436683), 'pulumi.get', 'pulumi.get', (['self', '"""day"""'], {}), "(self, 'day')\n", (436670, 436683), False, 'import pulumi\n'), ((436764, 436794), 'pulumi.set', 'pulumi.set', (['self', '"""day"""', 'value'], {}), "(self, 'day', value)\n", (436774, 436794), False, 'import pulumi\n'), ((437012, 437036), 'pulumi.get', 'pulumi.get', (['self', '"""hour"""'], {}), "(self, 'hour')\n", (437022, 437036), False, 'import pulumi\n'), ((437119, 437150), 'pulumi.set', 'pulumi.set', (['self', '"""hour"""', 'value'], {}), "(self, 'hour', value)\n", (437129, 437150), False, 'import pulumi\n'), ((437755, 437789), 'pulumi.set', 'pulumi.set', (['__self__', '"""name"""', 'name'], {}), "(__self__, 'name', name)\n", (437765, 437789), False, 'import pulumi\n'), ((437798, 437832), 'pulumi.set', 'pulumi.set', (['__self__', '"""type"""', 'type'], {}), "(__self__, 'type', type)\n", (437808, 437832), False, 'import pulumi\n'), ((437841, 437873), 'pulumi.set', 'pulumi.set', (['__self__', '"""uri"""', 'uri'], {}), "(__self__, 'uri', uri)\n", (437851, 437873), False, 'import pulumi\n'), ((438032, 438056), 'pulumi.get', 'pulumi.get', (['self', '"""name"""'], {}), "(self, 'name')\n", (438042, 438056), False, 'import pulumi\n'), ((438129, 438160), 'pulumi.set', 'pulumi.set', (['self', '"""name"""', 'value'], {}), "(self, 'name', value)\n", (438139, 438160), False, 'import pulumi\n'), ((438414, 438438), 'pulumi.get', 'pulumi.get', (['self', '"""type"""'], {}), "(self, 'type')\n", (438424, 438438), False, 'import pulumi\n'), ((438511, 438542), 'pulumi.set', 'pulumi.set', (['self', '"""type"""', 'value'], {}), "(self, 'type', value)\n", (438521, 438542), False, 'import pulumi\n'), ((438737, 438760), 'pulumi.get', 'pulumi.get', (['self', '"""uri"""'], {}), "(self, 'uri')\n", (438747, 438760), False, 'import pulumi\n'), ((438831, 438861), 'pulumi.set', 'pulumi.set', (['self', '"""uri"""', 'value'], {}), "(self, 'uri', value)\n", (438841, 438861), False, 'import pulumi\n'), ((439445, 439479), 'pulumi.set', 'pulumi.set', (['__self__', '"""name"""', 'name'], {}), "(__self__, 'name', name)\n", (439455, 439479), False, 'import pulumi\n'), ((439488, 439536), 'pulumi.set', 'pulumi.set', (['__self__', '"""shard_names"""', 'shard_names'], {}), "(__self__, 'shard_names', shard_names)\n", (439498, 439536), False, 'import pulumi\n'), ((439792, 439816), 'pulumi.get', 'pulumi.get', (['self', '"""name"""'], {}), "(self, 'name')\n", (439802, 439816), False, 'import pulumi\n'), ((439889, 439920), 'pulumi.set', 'pulumi.set', (['self', '"""name"""', 'value'], {}), "(self, 'name', value)\n", (439899, 439920), False, 'import pulumi\n'), ((440146, 440177), 'pulumi.get', 'pulumi.get', (['self', '"""shard_names"""'], {}), "(self, 'shard_names')\n", (440156, 440177), False, 'import pulumi\n'), ((440288, 440326), 'pulumi.set', 'pulumi.set', (['self', '"""shard_names"""', 'value'], {}), "(self, 'shard_names', value)\n", (440298, 440326), False, 'import pulumi\n'), ((440498, 440529), 'pulumi.get', 'pulumi.get', (['self', '"""description"""'], {}), "(self, 'description')\n", (440508, 440529), False, 'import pulumi\n'), ((440626, 440664), 'pulumi.set', 'pulumi.set', (['self', '"""description"""', 'value'], {}), "(self, 'description', value)\n", (440636, 440664), False, 'import pulumi\n'), ((441806, 441840), 'pulumi.set', 'pulumi.set', (['__self__', '"""name"""', 'name'], {}), "(__self__, 'name', name)\n", (441816, 441840), False, 'import pulumi\n'), ((441849, 441891), 'pulumi.set', 'pulumi.set', (['__self__', '"""password"""', 'password'], {}), "(__self__, 'password', password)\n", (441859, 441891), False, 'import pulumi\n'), ((442317, 442341), 'pulumi.get', 'pulumi.get', (['self', '"""name"""'], {}), "(self, 'name')\n", (442327, 442341), False, 'import pulumi\n'), ((442414, 442445), 'pulumi.set', 'pulumi.set', (['self', '"""name"""', 'value'], {}), "(self, 'name', value)\n", (442424, 442445), False, 'import pulumi\n'), ((442596, 442624), 'pulumi.get', 'pulumi.get', (['self', '"""password"""'], {}), "(self, 'password')\n", (442606, 442624), False, 'import pulumi\n'), ((442705, 442740), 'pulumi.set', 'pulumi.set', (['self', '"""password"""', 'value'], {}), "(self, 'password', value)\n", (442715, 442740), False, 'import pulumi\n'), ((443016, 443047), 'pulumi.get', 'pulumi.get', (['self', '"""permissions"""'], {}), "(self, 'permissions')\n", (443026, 443047), False, 'import pulumi\n'), ((443205, 443243), 'pulumi.set', 'pulumi.set', (['self', '"""permissions"""', 'value'], {}), "(self, 'permissions', value)\n", (443215, 443243), False, 'import pulumi\n'), ((443489, 443515), 'pulumi.get', 'pulumi.get', (['self', '"""quotas"""'], {}), "(self, 'quotas')\n", (443499, 443515), False, 'import pulumi\n'), ((443658, 443691), 'pulumi.set', 'pulumi.set', (['self', '"""quotas"""', 'value'], {}), "(self, 'quotas', value)\n", (443668, 443691), False, 'import pulumi\n'), ((443912, 443940), 'pulumi.get', 'pulumi.get', (['self', '"""settings"""'], {}), "(self, 'settings')\n", (443922, 443940), False, 'import pulumi\n'), ((444066, 444101), 'pulumi.set', 'pulumi.set', (['self', '"""settings"""', 'value'], {}), "(self, 'settings', value)\n", (444076, 444101), False, 'import pulumi\n'), ((444394, 444446), 'pulumi.set', 'pulumi.set', (['__self__', '"""database_name"""', 'database_name'], {}), "(__self__, 'database_name', database_name)\n", (444404, 444446), False, 'import pulumi\n'), ((444662, 444695), 'pulumi.get', 'pulumi.get', (['self', '"""database_name"""'], {}), "(self, 'database_name')\n", (444672, 444695), False, 'import pulumi\n'), ((444786, 444826), 'pulumi.set', 'pulumi.set', (['self', '"""database_name"""', 'value'], {}), "(self, 'database_name', value)\n", (444796, 444826), False, 'import pulumi\n'), ((445926, 445986), 'pulumi.set', 'pulumi.set', (['__self__', '"""interval_duration"""', 'interval_duration'], {}), "(__self__, 'interval_duration', interval_duration)\n", (445936, 445986), False, 'import pulumi\n'), ((446656, 446693), 'pulumi.get', 'pulumi.get', (['self', '"""interval_duration"""'], {}), "(self, 'interval_duration')\n", (446666, 446693), False, 'import pulumi\n'), ((446792, 446836), 'pulumi.set', 'pulumi.set', (['self', '"""interval_duration"""', 'value'], {}), "(self, 'interval_duration', value)\n", (446802, 446836), False, 'import pulumi\n'), ((447015, 447041), 'pulumi.get', 'pulumi.get', (['self', '"""errors"""'], {}), "(self, 'errors')\n", (447025, 447041), False, 'import pulumi\n'), ((447128, 447161), 'pulumi.set', 'pulumi.set', (['self', '"""errors"""', 'value'], {}), "(self, 'errors', value)\n", (447138, 447161), False, 'import pulumi\n'), ((447387, 447421), 'pulumi.get', 'pulumi.get', (['self', '"""execution_time"""'], {}), "(self, 'execution_time')\n", (447397, 447421), False, 'import pulumi\n'), ((447524, 447565), 'pulumi.set', 'pulumi.set', (['self', '"""execution_time"""', 'value'], {}), "(self, 'execution_time', value)\n", (447534, 447565), False, 'import pulumi\n'), ((447730, 447757), 'pulumi.get', 'pulumi.get', (['self', '"""queries"""'], {}), "(self, 'queries')\n", (447740, 447757), False, 'import pulumi\n'), ((447846, 447880), 'pulumi.set', 'pulumi.set', (['self', '"""queries"""', 'value'], {}), "(self, 'queries', value)\n", (447856, 447880), False, 'import pulumi\n'), ((448130, 448159), 'pulumi.get', 'pulumi.get', (['self', '"""read_rows"""'], {}), "(self, 'read_rows')\n", (448140, 448159), False, 'import pulumi\n'), ((448252, 448288), 'pulumi.set', 'pulumi.set', (['self', '"""read_rows"""', 'value'], {}), "(self, 'read_rows', value)\n", (448262, 448288), False, 'import pulumi\n'), ((448493, 448524), 'pulumi.get', 'pulumi.get', (['self', '"""result_rows"""'], {}), "(self, 'result_rows')\n", (448503, 448524), False, 'import pulumi\n'), ((448621, 448659), 'pulumi.set', 'pulumi.set', (['self', '"""result_rows"""', 'value'], {}), "(self, 'result_rows', value)\n", (448631, 448659), False, 'import pulumi\n'), ((481275, 481315), 'pulumi.get', 'pulumi.get', (['self', '"""add_http_cors_header"""'], {}), "(self, 'add_http_cors_header')\n", (481285, 481315), False, 'import pulumi\n'), ((481431, 481478), 'pulumi.set', 'pulumi.set', (['self', '"""add_http_cors_header"""', 'value'], {}), "(self, 'add_http_cors_header', value)\n", (481441, 481478), False, 'import pulumi\n'), ((481664, 481693), 'pulumi.get', 'pulumi.get', (['self', '"""allow_ddl"""'], {}), "(self, 'allow_ddl')\n", (481674, 481693), False, 'import pulumi\n'), ((481787, 481823), 'pulumi.set', 'pulumi.set', (['self', '"""allow_ddl"""', 'value'], {}), "(self, 'allow_ddl', value)\n", (481797, 481823), False, 'import pulumi\n'), ((481991, 482018), 'pulumi.get', 'pulumi.get', (['self', '"""compile"""'], {}), "(self, 'compile')\n", (482001, 482018), False, 'import pulumi\n'), ((482108, 482142), 'pulumi.set', 'pulumi.set', (['self', '"""compile"""', 'value'], {}), "(self, 'compile', value)\n", (482118, 482142), False, 'import pulumi\n'), ((482350, 482389), 'pulumi.get', 'pulumi.get', (['self', '"""compile_expressions"""'], {}), "(self, 'compile_expressions')\n", (482360, 482389), False, 'import pulumi\n'), ((482503, 482549), 'pulumi.set', 'pulumi.set', (['self', '"""compile_expressions"""', 'value'], {}), "(self, 'compile_expressions', value)\n", (482513, 482549), False, 'import pulumi\n'), ((482802, 482837), 'pulumi.get', 'pulumi.get', (['self', '"""connect_timeout"""'], {}), "(self, 'connect_timeout')\n", (482812, 482837), False, 'import pulumi\n'), ((482942, 482984), 'pulumi.set', 'pulumi.set', (['self', '"""connect_timeout"""', 'value'], {}), "(self, 'connect_timeout', value)\n", (482952, 482984), False, 'import pulumi\n'), ((483279, 483328), 'pulumi.get', 'pulumi.get', (['self', '"""count_distinct_implementation"""'], {}), "(self, 'count_distinct_implementation')\n", (483289, 483328), False, 'import pulumi\n'), ((483461, 483517), 'pulumi.set', 'pulumi.set', (['self', '"""count_distinct_implementation"""', 'value'], {}), "(self, 'count_distinct_implementation', value)\n", (483471, 483517), False, 'import pulumi\n'), ((483762, 483804), 'pulumi.get', 'pulumi.get', (['self', '"""distinct_overflow_mode"""'], {}), "(self, 'distinct_overflow_mode')\n", (483772, 483804), False, 'import pulumi\n'), ((483923, 483972), 'pulumi.set', 'pulumi.set', (['self', '"""distinct_overflow_mode"""', 'value'], {}), "(self, 'distinct_overflow_mode', value)\n", (483933, 483972), False, 'import pulumi\n'), ((484238, 484298), 'pulumi.get', 'pulumi.get', (['self', '"""distributed_aggregation_memory_efficient"""'], {}), "(self, 'distributed_aggregation_memory_efficient')\n", (484248, 484298), False, 'import pulumi\n'), ((484454, 484521), 'pulumi.set', 'pulumi.set', (['self', '"""distributed_aggregation_memory_efficient"""', 'value'], {}), "(self, 'distributed_aggregation_memory_efficient', value)\n", (484464, 484521), False, 'import pulumi\n'), ((484754, 484802), 'pulumi.get', 'pulumi.get', (['self', '"""distributed_ddl_task_timeout"""'], {}), "(self, 'distributed_ddl_task_timeout')\n", (484764, 484802), False, 'import pulumi\n'), ((484933, 484988), 'pulumi.set', 'pulumi.set', (['self', '"""distributed_ddl_task_timeout"""', 'value'], {}), "(self, 'distributed_ddl_task_timeout', value)\n", (484943, 484988), False, 'import pulumi\n'), ((485221, 485265), 'pulumi.get', 'pulumi.get', (['self', '"""distributed_product_mode"""'], {}), "(self, 'distributed_product_mode')\n", (485231, 485265), False, 'import pulumi\n'), ((485388, 485439), 'pulumi.set', 'pulumi.set', (['self', '"""distributed_product_mode"""', 'value'], {}), "(self, 'distributed_product_mode', value)\n", (485398, 485439), False, 'import pulumi\n'), ((485685, 485746), 'pulumi.get', 'pulumi.get', (['self', '"""empty_result_for_aggregation_by_empty_set"""'], {}), "(self, 'empty_result_for_aggregation_by_empty_set')\n", (485695, 485746), False, 'import pulumi\n'), ((485904, 485972), 'pulumi.set', 'pulumi.set', (['self', '"""empty_result_for_aggregation_by_empty_set"""', 'value'], {}), "(self, 'empty_result_for_aggregation_by_empty_set', value)\n", (485914, 485972), False, 'import pulumi\n'), ((486228, 486271), 'pulumi.get', 'pulumi.get', (['self', '"""enable_http_compression"""'], {}), "(self, 'enable_http_compression')\n", (486238, 486271), False, 'import pulumi\n'), ((486393, 486443), 'pulumi.set', 'pulumi.set', (['self', '"""enable_http_compression"""', 'value'], {}), "(self, 'enable_http_compression', value)\n", (486403, 486443), False, 'import pulumi\n'), ((486751, 486821), 'pulumi.get', 'pulumi.get', (['self', '"""fallback_to_stale_replicas_for_distributed_queries"""'], {}), "(self, 'fallback_to_stale_replicas_for_distributed_queries')\n", (486761, 486821), False, 'import pulumi\n'), ((486997, 487074), 'pulumi.set', 'pulumi.set', (['self', '"""fallback_to_stale_replicas_for_distributed_queries"""', 'value'], {}), "(self, 'fallback_to_stale_replicas_for_distributed_queries', value)\n", (487007, 487074), False, 'import pulumi\n'), ((487309, 487348), 'pulumi.get', 'pulumi.get', (['self', '"""force_index_by_date"""'], {}), "(self, 'force_index_by_date')\n", (487319, 487348), False, 'import pulumi\n'), ((487462, 487508), 'pulumi.set', 'pulumi.set', (['self', '"""force_index_by_date"""', 'value'], {}), "(self, 'force_index_by_date', value)\n", (487472, 487508), False, 'import pulumi\n'), ((487752, 487789), 'pulumi.get', 'pulumi.get', (['self', '"""force_primary_key"""'], {}), "(self, 'force_primary_key')\n", (487762, 487789), False, 'import pulumi\n'), ((487899, 487943), 'pulumi.set', 'pulumi.set', (['self', '"""force_primary_key"""', 'value'], {}), "(self, 'force_primary_key', value)\n", (487909, 487943), False, 'import pulumi\n'), ((488192, 488234), 'pulumi.get', 'pulumi.get', (['self', '"""group_by_overflow_mode"""'], {}), "(self, 'group_by_overflow_mode')\n", (488202, 488234), False, 'import pulumi\n'), ((488353, 488402), 'pulumi.set', 'pulumi.set', (['self', '"""group_by_overflow_mode"""', 'value'], {}), "(self, 'group_by_overflow_mode', value)\n", (488363, 488402), False, 'import pulumi\n'), ((488687, 488735), 'pulumi.get', 'pulumi.get', (['self', '"""group_by_two_level_threshold"""'], {}), "(self, 'group_by_two_level_threshold')\n", (488697, 488735), False, 'import pulumi\n'), ((488866, 488921), 'pulumi.set', 'pulumi.set', (['self', '"""group_by_two_level_threshold"""', 'value'], {}), "(self, 'group_by_two_level_threshold', value)\n", (488876, 488921), False, 'import pulumi\n'), ((489218, 489272), 'pulumi.get', 'pulumi.get', (['self', '"""group_by_two_level_threshold_bytes"""'], {}), "(self, 'group_by_two_level_threshold_bytes')\n", (489228, 489272), False, 'import pulumi\n'), ((489415, 489476), 'pulumi.set', 'pulumi.set', (['self', '"""group_by_two_level_threshold_bytes"""', 'value'], {}), "(self, 'group_by_two_level_threshold_bytes', value)\n", (489425, 489476), False, 'import pulumi\n'), ((489703, 489746), 'pulumi.get', 'pulumi.get', (['self', '"""http_connection_timeout"""'], {}), "(self, 'http_connection_timeout')\n", (489713, 489746), False, 'import pulumi\n'), ((489867, 489917), 'pulumi.set', 'pulumi.set', (['self', '"""http_connection_timeout"""', 'value'], {}), "(self, 'http_connection_timeout', value)\n", (489877, 489917), False, 'import pulumi\n'), ((490216, 490266), 'pulumi.get', 'pulumi.get', (['self', '"""http_headers_progress_interval"""'], {}), "(self, 'http_headers_progress_interval')\n", (490226, 490266), False, 'import pulumi\n'), ((490401, 490458), 'pulumi.set', 'pulumi.set', (['self', '"""http_headers_progress_interval"""', 'value'], {}), "(self, 'http_headers_progress_interval', value)\n", (490411, 490458), False, 'import pulumi\n'), ((490679, 490719), 'pulumi.get', 'pulumi.get', (['self', '"""http_receive_timeout"""'], {}), "(self, 'http_receive_timeout')\n", (490689, 490719), False, 'import pulumi\n'), ((490834, 490881), 'pulumi.set', 'pulumi.set', (['self', '"""http_receive_timeout"""', 'value'], {}), "(self, 'http_receive_timeout', value)\n", (490844, 490881), False, 'import pulumi\n'), ((491096, 491133), 'pulumi.get', 'pulumi.get', (['self', '"""http_send_timeout"""'], {}), "(self, 'http_send_timeout')\n", (491106, 491133), False, 'import pulumi\n'), ((491242, 491286), 'pulumi.set', 'pulumi.set', (['self', '"""http_send_timeout"""', 'value'], {}), "(self, 'http_send_timeout', value)\n", (491252, 491286), False, 'import pulumi\n'), ((491615, 491675), 'pulumi.get', 'pulumi.get', (['self', '"""input_format_defaults_for_omitted_fields"""'], {}), "(self, 'input_format_defaults_for_omitted_fields')\n", (491625, 491675), False, 'import pulumi\n'), ((491831, 491898), 'pulumi.set', 'pulumi.set', (['self', '"""input_format_defaults_for_omitted_fields"""', 'value'], {}), "(self, 'input_format_defaults_for_omitted_fields', value)\n", (491841, 491898), False, 'import pulumi\n'), ((492203, 492264), 'pulumi.get', 'pulumi.get', (['self', '"""input_format_values_interpret_expressions"""'], {}), "(self, 'input_format_values_interpret_expressions')\n", (492213, 492264), False, 'import pulumi\n'), ((492422, 492490), 'pulumi.set', 'pulumi.set', (['self', '"""input_format_values_interpret_expressions"""', 'value'], {}), "(self, 'input_format_values_interpret_expressions', value)\n", (492432, 492490), False, 'import pulumi\n'), ((492680, 492713), 'pulumi.get', 'pulumi.get', (['self', '"""insert_quorum"""'], {}), "(self, 'insert_quorum')\n", (492690, 492713), False, 'import pulumi\n'), ((492814, 492854), 'pulumi.set', 'pulumi.set', (['self', '"""insert_quorum"""', 'value'], {}), "(self, 'insert_quorum', value)\n", (492824, 492854), False, 'import pulumi\n'), ((493075, 493116), 'pulumi.get', 'pulumi.get', (['self', '"""insert_quorum_timeout"""'], {}), "(self, 'insert_quorum_timeout')\n", (493085, 493116), False, 'import pulumi\n'), ((493233, 493281), 'pulumi.set', 'pulumi.set', (['self', '"""insert_quorum_timeout"""', 'value'], {}), "(self, 'insert_quorum_timeout', value)\n", (493243, 493281), False, 'import pulumi\n'), ((493506, 493544), 'pulumi.get', 'pulumi.get', (['self', '"""join_overflow_mode"""'], {}), "(self, 'join_overflow_mode')\n", (493516, 493544), False, 'import pulumi\n'), ((493655, 493700), 'pulumi.set', 'pulumi.set', (['self', '"""join_overflow_mode"""', 'value'], {}), "(self, 'join_overflow_mode', value)\n", (493665, 493700), False, 'import pulumi\n'), ((494000, 494034), 'pulumi.get', 'pulumi.get', (['self', '"""join_use_nulls"""'], {}), "(self, 'join_use_nulls')\n", (494010, 494034), False, 'import pulumi\n'), ((494138, 494179), 'pulumi.set', 'pulumi.set', (['self', '"""join_use_nulls"""', 'value'], {}), "(self, 'join_use_nulls', value)\n", (494148, 494179), False, 'import pulumi\n'), ((494471, 494521), 'pulumi.get', 'pulumi.get', (['self', '"""joined_subquery_requires_alias"""'], {}), "(self, 'joined_subquery_requires_alias')\n", (494481, 494521), False, 'import pulumi\n'), ((494657, 494714), 'pulumi.set', 'pulumi.set', (['self', '"""joined_subquery_requires_alias"""', 'value'], {}), "(self, 'joined_subquery_requires_alias', value)\n", (494667, 494714), False, 'import pulumi\n'), ((495003, 495061), 'pulumi.get', 'pulumi.get', (['self', '"""low_cardinality_allow_in_native_format"""'], {}), "(self, 'low_cardinality_allow_in_native_format')\n", (495013, 495061), False, 'import pulumi\n'), ((495213, 495278), 'pulumi.set', 'pulumi.set', (['self', '"""low_cardinality_allow_in_native_format"""', 'value'], {}), "(self, 'low_cardinality_allow_in_native_format', value)\n", (495223, 495278), False, 'import pulumi\n'), ((495476, 495509), 'pulumi.get', 'pulumi.get', (['self', '"""max_ast_depth"""'], {}), "(self, 'max_ast_depth')\n", (495486, 495509), False, 'import pulumi\n'), ((495610, 495650), 'pulumi.set', 'pulumi.set', (['self', '"""max_ast_depth"""', 'value'], {}), "(self, 'max_ast_depth', value)\n", (495620, 495650), False, 'import pulumi\n'), ((495857, 495893), 'pulumi.get', 'pulumi.get', (['self', '"""max_ast_elements"""'], {}), "(self, 'max_ast_elements')\n", (495867, 495893), False, 'import pulumi\n'), ((496000, 496043), 'pulumi.set', 'pulumi.set', (['self', '"""max_ast_elements"""', 'value'], {}), "(self, 'max_ast_elements', value)\n", (496010, 496043), False, 'import pulumi\n'), ((496293, 496327), 'pulumi.get', 'pulumi.get', (['self', '"""max_block_size"""'], {}), "(self, 'max_block_size')\n", (496303, 496327), False, 'import pulumi\n'), ((496430, 496471), 'pulumi.set', 'pulumi.set', (['self', '"""max_block_size"""', 'value'], {}), "(self, 'max_block_size', value)\n", (496440, 496471), False, 'import pulumi\n'), ((496744, 496798), 'pulumi.get', 'pulumi.get', (['self', '"""max_bytes_before_external_group_by"""'], {}), "(self, 'max_bytes_before_external_group_by')\n", (496754, 496798), False, 'import pulumi\n'), ((496941, 497002), 'pulumi.set', 'pulumi.set', (['self', '"""max_bytes_before_external_group_by"""', 'value'], {}), "(self, 'max_bytes_before_external_group_by', value)\n", (496951, 497002), False, 'import pulumi\n'), ((497339, 497389), 'pulumi.get', 'pulumi.get', (['self', '"""max_bytes_before_external_sort"""'], {}), "(self, 'max_bytes_before_external_sort')\n", (497349, 497389), False, 'import pulumi\n'), ((497524, 497581), 'pulumi.set', 'pulumi.set', (['self', '"""max_bytes_before_external_sort"""', 'value'], {}), "(self, 'max_bytes_before_external_sort', value)\n", (497534, 497581), False, 'import pulumi\n'), ((497848, 497889), 'pulumi.get', 'pulumi.get', (['self', '"""max_bytes_in_distinct"""'], {}), "(self, 'max_bytes_in_distinct')\n", (497858, 497889), False, 'import pulumi\n'), ((498006, 498054), 'pulumi.set', 'pulumi.set', (['self', '"""max_bytes_in_distinct"""', 'value'], {}), "(self, 'max_bytes_in_distinct', value)\n", (498016, 498054), False, 'import pulumi\n'), ((498283, 498320), 'pulumi.get', 'pulumi.get', (['self', '"""max_bytes_in_join"""'], {}), "(self, 'max_bytes_in_join')\n", (498293, 498320), False, 'import pulumi\n'), ((498429, 498473), 'pulumi.set', 'pulumi.set', (['self', '"""max_bytes_in_join"""', 'value'], {}), "(self, 'max_bytes_in_join', value)\n", (498439, 498473), False, 'import pulumi\n'), ((498728, 498764), 'pulumi.get', 'pulumi.get', (['self', '"""max_bytes_in_set"""'], {}), "(self, 'max_bytes_in_set')\n", (498738, 498764), False, 'import pulumi\n'), ((498871, 498914), 'pulumi.set', 'pulumi.set', (['self', '"""max_bytes_in_set"""', 'value'], {}), "(self, 'max_bytes_in_set', value)\n", (498881, 498914), False, 'import pulumi\n'), ((499190, 499227), 'pulumi.get', 'pulumi.get', (['self', '"""max_bytes_to_read"""'], {}), "(self, 'max_bytes_to_read')\n", (499200, 499227), False, 'import pulumi\n'), ((499336, 499380), 'pulumi.set', 'pulumi.set', (['self', '"""max_bytes_to_read"""', 'value'], {}), "(self, 'max_bytes_to_read', value)\n", (499346, 499380), False, 'import pulumi\n'), ((499647, 499684), 'pulumi.get', 'pulumi.get', (['self', '"""max_bytes_to_sort"""'], {}), "(self, 'max_bytes_to_sort')\n", (499657, 499684), False, 'import pulumi\n'), ((499793, 499837), 'pulumi.set', 'pulumi.set', (['self', '"""max_bytes_to_sort"""', 'value'], {}), "(self, 'max_bytes_to_sort', value)\n", (499803, 499837), False, 'import pulumi\n'), ((500159, 500200), 'pulumi.get', 'pulumi.get', (['self', '"""max_bytes_to_transfer"""'], {}), "(self, 'max_bytes_to_transfer')\n", (500169, 500200), False, 'import pulumi\n'), ((500317, 500365), 'pulumi.set', 'pulumi.set', (['self', '"""max_bytes_to_transfer"""', 'value'], {}), "(self, 'max_bytes_to_transfer', value)\n", (500327, 500365), False, 'import pulumi\n'), ((500624, 500663), 'pulumi.get', 'pulumi.get', (['self', '"""max_columns_to_read"""'], {}), "(self, 'max_columns_to_read')\n", (500634, 500663), False, 'import pulumi\n'), ((500776, 500822), 'pulumi.set', 'pulumi.set', (['self', '"""max_columns_to_read"""', 'value'], {}), "(self, 'max_columns_to_read', value)\n", (500786, 500822), False, 'import pulumi\n'), ((501051, 501089), 'pulumi.get', 'pulumi.get', (['self', '"""max_execution_time"""'], {}), "(self, 'max_execution_time')\n", (501061, 501089), False, 'import pulumi\n'), ((501200, 501245), 'pulumi.set', 'pulumi.set', (['self', '"""max_execution_time"""', 'value'], {}), "(self, 'max_execution_time', value)\n", (501210, 501245), False, 'import pulumi\n'), ((501499, 501544), 'pulumi.get', 'pulumi.get', (['self', '"""max_expanded_ast_elements"""'], {}), "(self, 'max_expanded_ast_elements')\n", (501509, 501544), False, 'import pulumi\n'), ((501669, 501721), 'pulumi.set', 'pulumi.set', (['self', '"""max_expanded_ast_elements"""', 'value'], {}), "(self, 'max_expanded_ast_elements', value)\n", (501679, 501721), False, 'import pulumi\n'), ((501974, 502015), 'pulumi.get', 'pulumi.get', (['self', '"""max_insert_block_size"""'], {}), "(self, 'max_insert_block_size')\n", (501984, 502015), False, 'import pulumi\n'), ((502132, 502180), 'pulumi.set', 'pulumi.set', (['self', '"""max_insert_block_size"""', 'value'], {}), "(self, 'max_insert_block_size', value)\n", (502142, 502180), False, 'import pulumi\n'), ((502434, 502470), 'pulumi.get', 'pulumi.get', (['self', '"""max_memory_usage"""'], {}), "(self, 'max_memory_usage')\n", (502444, 502470), False, 'import pulumi\n'), ((502577, 502620), 'pulumi.set', 'pulumi.set', (['self', '"""max_memory_usage"""', 'value'], {}), "(self, 'max_memory_usage', value)\n", (502587, 502620), False, 'import pulumi\n'), ((502900, 502945), 'pulumi.get', 'pulumi.get', (['self', '"""max_memory_usage_for_user"""'], {}), "(self, 'max_memory_usage_for_user')\n", (502910, 502945), False, 'import pulumi\n'), ((503070, 503122), 'pulumi.set', 'pulumi.set', (['self', '"""max_memory_usage_for_user"""', 'value'], {}), "(self, 'max_memory_usage_for_user', value)\n", (503080, 503122), False, 'import pulumi\n'), ((503376, 503417), 'pulumi.get', 'pulumi.get', (['self', '"""max_network_bandwidth"""'], {}), "(self, 'max_network_bandwidth')\n", (503386, 503417), False, 'import pulumi\n'), ((503534, 503582), 'pulumi.set', 'pulumi.set', (['self', '"""max_network_bandwidth"""', 'value'], {}), "(self, 'max_network_bandwidth', value)\n", (503544, 503582), False, 'import pulumi\n'), ((503852, 503902), 'pulumi.get', 'pulumi.get', (['self', '"""max_network_bandwidth_for_user"""'], {}), "(self, 'max_network_bandwidth_for_user')\n", (503862, 503902), False, 'import pulumi\n'), ((504037, 504094), 'pulumi.set', 'pulumi.set', (['self', '"""max_network_bandwidth_for_user"""', 'value'], {}), "(self, 'max_network_bandwidth_for_user', value)\n", (504047, 504094), False, 'import pulumi\n'), ((504344, 504378), 'pulumi.get', 'pulumi.get', (['self', '"""max_query_size"""'], {}), "(self, 'max_query_size')\n", (504354, 504378), False, 'import pulumi\n'), ((504481, 504522), 'pulumi.set', 'pulumi.set', (['self', '"""max_query_size"""', 'value'], {}), "(self, 'max_query_size', value)\n", (504491, 504522), False, 'import pulumi\n'), ((504788, 504849), 'pulumi.get', 'pulumi.get', (['self', '"""max_replica_delay_for_distributed_queries"""'], {}), "(self, 'max_replica_delay_for_distributed_queries')\n", (504798, 504849), False, 'import pulumi\n'), ((505006, 505074), 'pulumi.set', 'pulumi.set', (['self', '"""max_replica_delay_for_distributed_queries"""', 'value'], {}), "(self, 'max_replica_delay_for_distributed_queries', value)\n", (505016, 505074), False, 'import pulumi\n'), ((505284, 505320), 'pulumi.get', 'pulumi.get', (['self', '"""max_result_bytes"""'], {}), "(self, 'max_result_bytes')\n", (505294, 505320), False, 'import pulumi\n'), ((505427, 505470), 'pulumi.set', 'pulumi.set', (['self', '"""max_result_bytes"""', 'value'], {}), "(self, 'max_result_bytes', value)\n", (505437, 505470), False, 'import pulumi\n'), ((505677, 505712), 'pulumi.get', 'pulumi.get', (['self', '"""max_result_rows"""'], {}), "(self, 'max_result_rows')\n", (505687, 505712), False, 'import pulumi\n'), ((505817, 505859), 'pulumi.set', 'pulumi.set', (['self', '"""max_result_rows"""', 'value'], {}), "(self, 'max_result_rows', value)\n", (505827, 505859), False, 'import pulumi\n'), ((506099, 506139), 'pulumi.get', 'pulumi.get', (['self', '"""max_rows_in_distinct"""'], {}), "(self, 'max_rows_in_distinct')\n", (506109, 506139), False, 'import pulumi\n'), ((506254, 506301), 'pulumi.set', 'pulumi.set', (['self', '"""max_rows_in_distinct"""', 'value'], {}), "(self, 'max_rows_in_distinct', value)\n", (506264, 506301), False, 'import pulumi\n'), ((506527, 506563), 'pulumi.get', 'pulumi.get', (['self', '"""max_rows_in_join"""'], {}), "(self, 'max_rows_in_join')\n", (506537, 506563), False, 'import pulumi\n'), ((506670, 506713), 'pulumi.set', 'pulumi.set', (['self', '"""max_rows_in_join"""', 'value'], {}), "(self, 'max_rows_in_join', value)\n", (506680, 506713), False, 'import pulumi\n'), ((506965, 507000), 'pulumi.get', 'pulumi.get', (['self', '"""max_rows_in_set"""'], {}), "(self, 'max_rows_in_set')\n", (506975, 507000), False, 'import pulumi\n'), ((507105, 507147), 'pulumi.set', 'pulumi.set', (['self', '"""max_rows_in_set"""', 'value'], {}), "(self, 'max_rows_in_set', value)\n", (507115, 507147), False, 'import pulumi\n'), ((507398, 507438), 'pulumi.get', 'pulumi.get', (['self', '"""max_rows_to_group_by"""'], {}), "(self, 'max_rows_to_group_by')\n", (507408, 507438), False, 'import pulumi\n'), ((507553, 507600), 'pulumi.set', 'pulumi.set', (['self', '"""max_rows_to_group_by"""', 'value'], {}), "(self, 'max_rows_to_group_by', value)\n", (507563, 507600), False, 'import pulumi\n'), ((507853, 507889), 'pulumi.get', 'pulumi.get', (['self', '"""max_rows_to_read"""'], {}), "(self, 'max_rows_to_read')\n", (507863, 507889), False, 'import pulumi\n'), ((507996, 508039), 'pulumi.set', 'pulumi.set', (['self', '"""max_rows_to_read"""', 'value'], {}), "(self, 'max_rows_to_read', value)\n", (508006, 508039), False, 'import pulumi\n'), ((508283, 508319), 'pulumi.get', 'pulumi.get', (['self', '"""max_rows_to_sort"""'], {}), "(self, 'max_rows_to_sort')\n", (508293, 508319), False, 'import pulumi\n'), ((508426, 508469), 'pulumi.set', 'pulumi.set', (['self', '"""max_rows_to_sort"""', 'value'], {}), "(self, 'max_rows_to_sort', value)\n", (508436, 508469), False, 'import pulumi\n'), ((508768, 508808), 'pulumi.get', 'pulumi.get', (['self', '"""max_rows_to_transfer"""'], {}), "(self, 'max_rows_to_transfer')\n", (508778, 508808), False, 'import pulumi\n'), ((508923, 508970), 'pulumi.set', 'pulumi.set', (['self', '"""max_rows_to_transfer"""', 'value'], {}), "(self, 'max_rows_to_transfer', value)\n", (508933, 508970), False, 'import pulumi\n'), ((509287, 509328), 'pulumi.get', 'pulumi.get', (['self', '"""max_temporary_columns"""'], {}), "(self, 'max_temporary_columns')\n", (509297, 509328), False, 'import pulumi\n'), ((509445, 509493), 'pulumi.set', 'pulumi.set', (['self', '"""max_temporary_columns"""', 'value'], {}), "(self, 'max_temporary_columns', value)\n", (509455, 509493), False, 'import pulumi\n'), ((509828, 509879), 'pulumi.get', 'pulumi.get', (['self', '"""max_temporary_non_const_columns"""'], {}), "(self, 'max_temporary_non_const_columns')\n", (509838, 509879), False, 'import pulumi\n'), ((510016, 510074), 'pulumi.set', 'pulumi.set', (['self', '"""max_temporary_non_const_columns"""', 'value'], {}), "(self, 'max_temporary_non_const_columns', value)\n", (510026, 510074), False, 'import pulumi\n'), ((510340, 510371), 'pulumi.get', 'pulumi.get', (['self', '"""max_threads"""'], {}), "(self, 'max_threads')\n", (510350, 510371), False, 'import pulumi\n'), ((510468, 510506), 'pulumi.set', 'pulumi.set', (['self', '"""max_threads"""', 'value'], {}), "(self, 'max_threads', value)\n", (510478, 510506), False, 'import pulumi\n'), ((510843, 510896), 'pulumi.get', 'pulumi.get', (['self', '"""merge_tree_max_bytes_to_use_cache"""'], {}), "(self, 'merge_tree_max_bytes_to_use_cache')\n", (510853, 510896), False, 'import pulumi\n'), ((511037, 511097), 'pulumi.set', 'pulumi.set', (['self', '"""merge_tree_max_bytes_to_use_cache"""', 'value'], {}), "(self, 'merge_tree_max_bytes_to_use_cache', value)\n", (511047, 511097), False, 'import pulumi\n'), ((511430, 511482), 'pulumi.get', 'pulumi.get', (['self', '"""merge_tree_max_rows_to_use_cache"""'], {}), "(self, 'merge_tree_max_rows_to_use_cache')\n", (511440, 511482), False, 'import pulumi\n'), ((511621, 511680), 'pulumi.set', 'pulumi.set', (['self', '"""merge_tree_max_rows_to_use_cache"""', 'value'], {}), "(self, 'merge_tree_max_rows_to_use_cache', value)\n", (511631, 511680), False, 'import pulumi\n'), ((512093, 512153), 'pulumi.get', 'pulumi.get', (['self', '"""merge_tree_min_bytes_for_concurrent_read"""'], {}), "(self, 'merge_tree_min_bytes_for_concurrent_read')\n", (512103, 512153), False, 'import pulumi\n'), ((512308, 512375), 'pulumi.set', 'pulumi.set', (['self', '"""merge_tree_min_bytes_for_concurrent_read"""', 'value'], {}), "(self, 'merge_tree_min_bytes_for_concurrent_read', value)\n", (512318, 512375), False, 'import pulumi\n'), ((512788, 512847), 'pulumi.get', 'pulumi.get', (['self', '"""merge_tree_min_rows_for_concurrent_read"""'], {}), "(self, 'merge_tree_min_rows_for_concurrent_read')\n", (512798, 512847), False, 'import pulumi\n'), ((513000, 513066), 'pulumi.set', 'pulumi.set', (['self', '"""merge_tree_min_rows_for_concurrent_read"""', 'value'], {}), "(self, 'merge_tree_min_rows_for_concurrent_read', value)\n", (513010, 513066), False, 'import pulumi\n'), ((513333, 513379), 'pulumi.get', 'pulumi.get', (['self', '"""min_bytes_to_use_direct_io"""'], {}), "(self, 'min_bytes_to_use_direct_io')\n", (513343, 513379), False, 'import pulumi\n'), ((513506, 513559), 'pulumi.set', 'pulumi.set', (['self', '"""min_bytes_to_use_direct_io"""', 'value'], {}), "(self, 'min_bytes_to_use_direct_io', value)\n", (513516, 513559), False, 'import pulumi\n'), ((513821, 513861), 'pulumi.get', 'pulumi.get', (['self', '"""min_count_to_compile"""'], {}), "(self, 'min_count_to_compile')\n", (513831, 513861), False, 'import pulumi\n'), ((513976, 514023), 'pulumi.set', 'pulumi.set', (['self', '"""min_count_to_compile"""', 'value'], {}), "(self, 'min_count_to_compile', value)\n", (513986, 514023), False, 'import pulumi\n'), ((514311, 514362), 'pulumi.get', 'pulumi.get', (['self', '"""min_count_to_compile_expression"""'], {}), "(self, 'min_count_to_compile_expression')\n", (514321, 514362), False, 'import pulumi\n'), ((514499, 514557), 'pulumi.set', 'pulumi.set', (['self', '"""min_count_to_compile_expression"""', 'value'], {}), "(self, 'min_count_to_compile_expression', value)\n", (514509, 514557), False, 'import pulumi\n'), ((514775, 514814), 'pulumi.get', 'pulumi.get', (['self', '"""min_execution_speed"""'], {}), "(self, 'min_execution_speed')\n", (514785, 514814), False, 'import pulumi\n'), ((514927, 514973), 'pulumi.set', 'pulumi.set', (['self', '"""min_execution_speed"""', 'value'], {}), "(self, 'min_execution_speed', value)\n", (514937, 514973), False, 'import pulumi\n'), ((515203, 515248), 'pulumi.get', 'pulumi.get', (['self', '"""min_execution_speed_bytes"""'], {}), "(self, 'min_execution_speed_bytes')\n", (515213, 515248), False, 'import pulumi\n'), ((515373, 515425), 'pulumi.set', 'pulumi.set', (['self', '"""min_execution_speed_bytes"""', 'value'], {}), "(self, 'min_execution_speed_bytes', value)\n", (515383, 515425), False, 'import pulumi\n'), ((515714, 515761), 'pulumi.get', 'pulumi.get', (['self', '"""min_insert_block_size_bytes"""'], {}), "(self, 'min_insert_block_size_bytes')\n", (515724, 515761), False, 'import pulumi\n'), ((515890, 515944), 'pulumi.set', 'pulumi.set', (['self', '"""min_insert_block_size_bytes"""', 'value'], {}), "(self, 'min_insert_block_size_bytes', value)\n", (515900, 515944), False, 'import pulumi\n'), ((516230, 516276), 'pulumi.get', 'pulumi.get', (['self', '"""min_insert_block_size_rows"""'], {}), "(self, 'min_insert_block_size_rows')\n", (516240, 516276), False, 'import pulumi\n'), ((516403, 516456), 'pulumi.set', 'pulumi.set', (['self', '"""min_insert_block_size_rows"""', 'value'], {}), "(self, 'min_insert_block_size_rows', value)\n", (516413, 516456), False, 'import pulumi\n'), ((516866, 516924), 'pulumi.get', 'pulumi.get', (['self', '"""output_format_json_quote64bit_integers"""'], {}), "(self, 'output_format_json_quote64bit_integers')\n", (516876, 516924), False, 'import pulumi\n'), ((517076, 517141), 'pulumi.set', 'pulumi.set', (['self', '"""output_format_json_quote64bit_integers"""', 'value'], {}), "(self, 'output_format_json_quote64bit_integers', value)\n", (517086, 517141), False, 'import pulumi\n'), ((517406, 517460), 'pulumi.get', 'pulumi.get', (['self', '"""output_format_json_quote_denormals"""'], {}), "(self, 'output_format_json_quote_denormals')\n", (517416, 517460), False, 'import pulumi\n'), ((517604, 517665), 'pulumi.set', 'pulumi.set', (['self', '"""output_format_json_quote_denormals"""', 'value'], {}), "(self, 'output_format_json_quote_denormals', value)\n", (517614, 517665), False, 'import pulumi\n'), ((517818, 517846), 'pulumi.get', 'pulumi.get', (['self', '"""priority"""'], {}), "(self, 'priority')\n", (517828, 517846), False, 'import pulumi\n'), ((517937, 517972), 'pulumi.set', 'pulumi.set', (['self', '"""priority"""', 'value'], {}), "(self, 'priority', value)\n", (517947, 517972), False, 'import pulumi\n'), ((518152, 518182), 'pulumi.get', 'pulumi.get', (['self', '"""quota_mode"""'], {}), "(self, 'quota_mode')\n", (518162, 518182), False, 'import pulumi\n'), ((518277, 518314), 'pulumi.set', 'pulumi.set', (['self', '"""quota_mode"""', 'value'], {}), "(self, 'quota_mode', value)\n", (518287, 518314), False, 'import pulumi\n'), ((518542, 518580), 'pulumi.get', 'pulumi.get', (['self', '"""read_overflow_mode"""'], {}), "(self, 'read_overflow_mode')\n", (518552, 518580), False, 'import pulumi\n'), ((518691, 518736), 'pulumi.set', 'pulumi.set', (['self', '"""read_overflow_mode"""', 'value'], {}), "(self, 'read_overflow_mode', value)\n", (518701, 518736), False, 'import pulumi\n'), ((518953, 518981), 'pulumi.get', 'pulumi.get', (['self', '"""readonly"""'], {}), "(self, 'readonly')\n", (518963, 518981), False, 'import pulumi\n'), ((519072, 519107), 'pulumi.set', 'pulumi.set', (['self', '"""readonly"""', 'value'], {}), "(self, 'readonly', value)\n", (519082, 519107), False, 'import pulumi\n'), ((519360, 519395), 'pulumi.get', 'pulumi.get', (['self', '"""receive_timeout"""'], {}), "(self, 'receive_timeout')\n", (519370, 519395), False, 'import pulumi\n'), ((519500, 519542), 'pulumi.set', 'pulumi.set', (['self', '"""receive_timeout"""', 'value'], {}), "(self, 'receive_timeout', value)\n", (519510, 519542), False, 'import pulumi\n'), ((519862, 519915), 'pulumi.get', 'pulumi.get', (['self', '"""replication_alter_partitions_sync"""'], {}), "(self, 'replication_alter_partitions_sync')\n", (519872, 519915), False, 'import pulumi\n'), ((520056, 520116), 'pulumi.set', 'pulumi.set', (['self', '"""replication_alter_partitions_sync"""', 'value'], {}), "(self, 'replication_alter_partitions_sync', value)\n", (520066, 520116), False, 'import pulumi\n'), ((520347, 520387), 'pulumi.get', 'pulumi.get', (['self', '"""result_overflow_mode"""'], {}), "(self, 'result_overflow_mode')\n", (520357, 520387), False, 'import pulumi\n'), ((520502, 520549), 'pulumi.set', 'pulumi.set', (['self', '"""result_overflow_mode"""', 'value'], {}), "(self, 'result_overflow_mode', value)\n", (520512, 520549), False, 'import pulumi\n'), ((520807, 520856), 'pulumi.get', 'pulumi.get', (['self', '"""select_sequential_consistency"""'], {}), "(self, 'select_sequential_consistency')\n", (520817, 520856), False, 'import pulumi\n'), ((520990, 521046), 'pulumi.set', 'pulumi.set', (['self', '"""select_sequential_consistency"""', 'value'], {}), "(self, 'select_sequential_consistency', value)\n", (521000, 521046), False, 'import pulumi\n'), ((521335, 521384), 'pulumi.get', 'pulumi.get', (['self', '"""send_progress_in_http_headers"""'], {}), "(self, 'send_progress_in_http_headers')\n", (521345, 521384), False, 'import pulumi\n'), ((521518, 521574), 'pulumi.set', 'pulumi.set', (['self', '"""send_progress_in_http_headers"""', 'value'], {}), "(self, 'send_progress_in_http_headers', value)\n", (521528, 521574), False, 'import pulumi\n'), ((521818, 521850), 'pulumi.get', 'pulumi.get', (['self', '"""send_timeout"""'], {}), "(self, 'send_timeout')\n", (521828, 521850), False, 'import pulumi\n'), ((521949, 521988), 'pulumi.set', 'pulumi.set', (['self', '"""send_timeout"""', 'value'], {}), "(self, 'send_timeout', value)\n", (521959, 521988), False, 'import pulumi\n'), ((522224, 522261), 'pulumi.get', 'pulumi.get', (['self', '"""set_overflow_mode"""'], {}), "(self, 'set_overflow_mode')\n", (522234, 522261), False, 'import pulumi\n'), ((522370, 522414), 'pulumi.set', 'pulumi.set', (['self', '"""set_overflow_mode"""', 'value'], {}), "(self, 'set_overflow_mode', value)\n", (522380, 522414), False, 'import pulumi\n'), ((522658, 522701), 'pulumi.get', 'pulumi.get', (['self', '"""skip_unavailable_shards"""'], {}), "(self, 'skip_unavailable_shards')\n", (522668, 522701), False, 'import pulumi\n'), ((522823, 522873), 'pulumi.set', 'pulumi.set', (['self', '"""skip_unavailable_shards"""', 'value'], {}), "(self, 'skip_unavailable_shards', value)\n", (522833, 522873), False, 'import pulumi\n'), ((523101, 523139), 'pulumi.get', 'pulumi.get', (['self', '"""sort_overflow_mode"""'], {}), "(self, 'sort_overflow_mode')\n", (523111, 523139), False, 'import pulumi\n'), ((523250, 523295), 'pulumi.set', 'pulumi.set', (['self', '"""sort_overflow_mode"""', 'value'], {}), "(self, 'sort_overflow_mode', value)\n", (523260, 523295), False, 'import pulumi\n'), ((523518, 523559), 'pulumi.get', 'pulumi.get', (['self', '"""timeout_overflow_mode"""'], {}), "(self, 'timeout_overflow_mode')\n", (523528, 523559), False, 'import pulumi\n'), ((523676, 523724), 'pulumi.set', 'pulumi.set', (['self', '"""timeout_overflow_mode"""', 'value'], {}), "(self, 'timeout_overflow_mode', value)\n", (523686, 523724), False, 'import pulumi\n'), ((523949, 523991), 'pulumi.get', 'pulumi.get', (['self', '"""transfer_overflow_mode"""'], {}), "(self, 'transfer_overflow_mode')\n", (523959, 523991), False, 'import pulumi\n'), ((524110, 524159), 'pulumi.set', 'pulumi.set', (['self', '"""transfer_overflow_mode"""', 'value'], {}), "(self, 'transfer_overflow_mode', value)\n", (524120, 524159), False, 'import pulumi\n'), ((524379, 524416), 'pulumi.get', 'pulumi.get', (['self', '"""transform_null_in"""'], {}), "(self, 'transform_null_in')\n", (524389, 524416), False, 'import pulumi\n'), ((524526, 524570), 'pulumi.set', 'pulumi.set', (['self', '"""transform_null_in"""', 'value'], {}), "(self, 'transform_null_in', value)\n", (524536, 524570), False, 'import pulumi\n'), ((524798, 524840), 'pulumi.get', 'pulumi.get', (['self', '"""use_uncompressed_cache"""'], {}), "(self, 'use_uncompressed_cache')\n", (524808, 524840), False, 'import pulumi\n'), ((524960, 525009), 'pulumi.set', 'pulumi.set', (['self', '"""use_uncompressed_cache"""', 'value'], {}), "(self, 'use_uncompressed_cache', value)\n", (524970, 525009), False, 'import pulumi\n'), ((525772, 525801), 'pulumi.get', 'pulumi.get', (['self', '"""resources"""'], {}), "(self, 'resources')\n", (525782, 525801), False, 'import pulumi\n'), ((525935, 525971), 'pulumi.set', 'pulumi.set', (['self', '"""resources"""', 'value'], {}), "(self, 'resources', value)\n", (525945, 525971), False, 'import pulumi\n'), ((527167, 527196), 'pulumi.get', 'pulumi.get', (['self', '"""disk_size"""'], {}), "(self, 'disk_size')\n", (527177, 527196), False, 'import pulumi\n'), ((527289, 527325), 'pulumi.set', 'pulumi.set', (['self', '"""disk_size"""', 'value'], {}), "(self, 'disk_size', value)\n", (527299, 527325), False, 'import pulumi\n'), ((527655, 527687), 'pulumi.get', 'pulumi.get', (['self', '"""disk_type_id"""'], {}), "(self, 'disk_type_id')\n", (527665, 527687), False, 'import pulumi\n'), ((527786, 527825), 'pulumi.set', 'pulumi.set', (['self', '"""disk_type_id"""', 'value'], {}), "(self, 'disk_type_id', value)\n", (527796, 527825), False, 'import pulumi\n'), ((527965, 528003), 'pulumi.get', 'pulumi.get', (['self', '"""resource_preset_id"""'], {}), "(self, 'resource_preset_id')\n", (527975, 528003), False, 'import pulumi\n'), ((528114, 528159), 'pulumi.set', 'pulumi.set', (['self', '"""resource_preset_id"""', 'value'], {}), "(self, 'resource_preset_id', value)\n", (528124, 528159), False, 'import pulumi\n'), ((529568, 529622), 'pulumi.set', 'pulumi.set', (['__self__', '"""admin_password"""', 'admin_password'], {}), "(__self__, 'admin_password', admin_password)\n", (529578, 529622), False, 'import pulumi\n'), ((529631, 529675), 'pulumi.set', 'pulumi.set', (['__self__', '"""data_node"""', 'data_node'], {}), "(__self__, 'data_node', data_node)\n", (529641, 529675), False, 'import pulumi\n'), ((530224, 530258), 'pulumi.get', 'pulumi.get', (['self', '"""admin_password"""'], {}), "(self, 'admin_password')\n", (530234, 530258), False, 'import pulumi\n'), ((530351, 530392), 'pulumi.set', 'pulumi.set', (['self', '"""admin_password"""', 'value'], {}), "(self, 'admin_password', value)\n", (530361, 530392), False, 'import pulumi\n'), ((530667, 530696), 'pulumi.get', 'pulumi.get', (['self', '"""data_node"""'], {}), "(self, 'data_node')\n", (530677, 530696), False, 'import pulumi\n'), ((530819, 530855), 'pulumi.set', 'pulumi.set', (['self', '"""data_node"""', 'value'], {}), "(self, 'data_node', value)\n", (530829, 530855), False, 'import pulumi\n'), ((531153, 531180), 'pulumi.get', 'pulumi.get', (['self', '"""edition"""'], {}), "(self, 'edition')\n", (531163, 531180), False, 'import pulumi\n'), ((531269, 531303), 'pulumi.set', 'pulumi.set', (['self', '"""edition"""', 'value'], {}), "(self, 'edition', value)\n", (531279, 531303), False, 'import pulumi\n'), ((531596, 531627), 'pulumi.get', 'pulumi.get', (['self', '"""master_node"""'], {}), "(self, 'master_node')\n", (531606, 531627), False, 'import pulumi\n'), ((531766, 531804), 'pulumi.set', 'pulumi.set', (['self', '"""master_node"""', 'value'], {}), "(self, 'master_node', value)\n", (531776, 531804), False, 'import pulumi\n'), ((532007, 532034), 'pulumi.get', 'pulumi.get', (['self', '"""plugins"""'], {}), "(self, 'plugins')\n", (532017, 532034), False, 'import pulumi\n'), ((532147, 532181), 'pulumi.set', 'pulumi.set', (['self', '"""plugins"""', 'value'], {}), "(self, 'plugins', value)\n", (532157, 532181), False, 'import pulumi\n'), ((532343, 532370), 'pulumi.get', 'pulumi.get', (['self', '"""version"""'], {}), "(self, 'version')\n", (532353, 532370), False, 'import pulumi\n'), ((532459, 532493), 'pulumi.set', 'pulumi.set', (['self', '"""version"""', 'value'], {}), "(self, 'version', value)\n", (532469, 532493), False, 'import pulumi\n'), ((532926, 532970), 'pulumi.set', 'pulumi.set', (['__self__', '"""resources"""', 'resources'], {}), "(__self__, 'resources', resources)\n", (532936, 532970), False, 'import pulumi\n'), ((533257, 533286), 'pulumi.get', 'pulumi.get', (['self', '"""resources"""'], {}), "(self, 'resources')\n", (533267, 533286), False, 'import pulumi\n'), ((533418, 533454), 'pulumi.set', 'pulumi.set', (['self', '"""resources"""', 'value'], {}), "(self, 'resources', value)\n", (533428, 533454), False, 'import pulumi\n'), ((533942, 533986), 'pulumi.set', 'pulumi.set', (['__self__', '"""disk_size"""', 'disk_size'], {}), "(__self__, 'disk_size', disk_size)\n", (533952, 533986), False, 'import pulumi\n'), ((533995, 534045), 'pulumi.set', 'pulumi.set', (['__self__', '"""disk_type_id"""', 'disk_type_id'], {}), "(__self__, 'disk_type_id', disk_type_id)\n", (534005, 534045), False, 'import pulumi\n'), ((534054, 534116), 'pulumi.set', 'pulumi.set', (['__self__', '"""resource_preset_id"""', 'resource_preset_id'], {}), "(__self__, 'resource_preset_id', resource_preset_id)\n", (534064, 534116), False, 'import pulumi\n'), ((534318, 534347), 'pulumi.get', 'pulumi.get', (['self', '"""disk_size"""'], {}), "(self, 'disk_size')\n", (534328, 534347), False, 'import pulumi\n'), ((534430, 534466), 'pulumi.set', 'pulumi.set', (['self', '"""disk_size"""', 'value'], {}), "(self, 'disk_size', value)\n", (534440, 534466), False, 'import pulumi\n'), ((534660, 534692), 'pulumi.get', 'pulumi.get', (['self', '"""disk_type_id"""'], {}), "(self, 'disk_type_id')\n", (534670, 534692), False, 'import pulumi\n'), ((534781, 534820), 'pulumi.set', 'pulumi.set', (['self', '"""disk_type_id"""', 'value'], {}), "(self, 'disk_type_id', value)\n", (534791, 534820), False, 'import pulumi\n'), ((534950, 534988), 'pulumi.get', 'pulumi.get', (['self', '"""resource_preset_id"""'], {}), "(self, 'resource_preset_id')\n", (534960, 534988), False, 'import pulumi\n'), ((535089, 535134), 'pulumi.set', 'pulumi.set', (['self', '"""resource_preset_id"""', 'value'], {}), "(self, 'resource_preset_id', value)\n", (535099, 535134), False, 'import pulumi\n'), ((535573, 535617), 'pulumi.set', 'pulumi.set', (['__self__', '"""resources"""', 'resources'], {}), "(__self__, 'resources', resources)\n", (535583, 535617), False, 'import pulumi\n'), ((535906, 535935), 'pulumi.get', 'pulumi.get', (['self', '"""resources"""'], {}), "(self, 'resources')\n", (535916, 535935), False, 'import pulumi\n'), ((536069, 536105), 'pulumi.set', 'pulumi.set', (['self', '"""resources"""', 'value'], {}), "(self, 'resources', value)\n", (536079, 536105), False, 'import pulumi\n'), ((536595, 536639), 'pulumi.set', 'pulumi.set', (['__self__', '"""disk_size"""', 'disk_size'], {}), "(__self__, 'disk_size', disk_size)\n", (536605, 536639), False, 'import pulumi\n'), ((536648, 536698), 'pulumi.set', 'pulumi.set', (['__self__', '"""disk_type_id"""', 'disk_type_id'], {}), "(__self__, 'disk_type_id', disk_type_id)\n", (536658, 536698), False, 'import pulumi\n'), ((536707, 536769), 'pulumi.set', 'pulumi.set', (['__self__', '"""resource_preset_id"""', 'resource_preset_id'], {}), "(__self__, 'resource_preset_id', resource_preset_id)\n", (536717, 536769), False, 'import pulumi\n'), ((536971, 537000), 'pulumi.get', 'pulumi.get', (['self', '"""disk_size"""'], {}), "(self, 'disk_size')\n", (536981, 537000), False, 'import pulumi\n'), ((537083, 537119), 'pulumi.set', 'pulumi.set', (['self', '"""disk_size"""', 'value'], {}), "(self, 'disk_size', value)\n", (537093, 537119), False, 'import pulumi\n'), ((537313, 537345), 'pulumi.get', 'pulumi.get', (['self', '"""disk_type_id"""'], {}), "(self, 'disk_type_id')\n", (537323, 537345), False, 'import pulumi\n'), ((537434, 537473), 'pulumi.set', 'pulumi.set', (['self', '"""disk_type_id"""', 'value'], {}), "(self, 'disk_type_id', value)\n", (537444, 537473), False, 'import pulumi\n'), ((537603, 537641), 'pulumi.get', 'pulumi.get', (['self', '"""resource_preset_id"""'], {}), "(self, 'resource_preset_id')\n", (537613, 537641), False, 'import pulumi\n'), ((537742, 537787), 'pulumi.set', 'pulumi.set', (['self', '"""resource_preset_id"""', 'value'], {}), "(self, 'resource_preset_id', value)\n", (537752, 537787), False, 'import pulumi\n'), ((539065, 539099), 'pulumi.set', 'pulumi.set', (['__self__', '"""name"""', 'name'], {}), "(__self__, 'name', name)\n", (539075, 539099), False, 'import pulumi\n'), ((539108, 539142), 'pulumi.set', 'pulumi.set', (['__self__', '"""type"""', 'type'], {}), "(__self__, 'type', type)\n", (539118, 539142), False, 'import pulumi\n'), ((539151, 539185), 'pulumi.set', 'pulumi.set', (['__self__', '"""zone"""', 'zone'], {}), "(__self__, 'zone', zone)\n", (539161, 539185), False, 'import pulumi\n'), ((539611, 539635), 'pulumi.get', 'pulumi.get', (['self', '"""name"""'], {}), "(self, 'name')\n", (539621, 539635), False, 'import pulumi\n'), ((539708, 539739), 'pulumi.set', 'pulumi.set', (['self', '"""name"""', 'value'], {}), "(self, 'name', value)\n", (539718, 539739), False, 'import pulumi\n'), ((539943, 539967), 'pulumi.get', 'pulumi.get', (['self', '"""type"""'], {}), "(self, 'type')\n", (539953, 539967), False, 'import pulumi\n'), ((540040, 540071), 'pulumi.set', 'pulumi.set', (['self', '"""type"""', 'value'], {}), "(self, 'type', value)\n", (540050, 540071), False, 'import pulumi\n'), ((540384, 540408), 'pulumi.get', 'pulumi.get', (['self', '"""zone"""'], {}), "(self, 'zone')\n", (540394, 540408), False, 'import pulumi\n'), ((540481, 540512), 'pulumi.set', 'pulumi.set', (['self', '"""zone"""', 'value'], {}), "(self, 'zone', value)\n", (540491, 540512), False, 'import pulumi\n'), ((540780, 540816), 'pulumi.get', 'pulumi.get', (['self', '"""assign_public_ip"""'], {}), "(self, 'assign_public_ip')\n", (540790, 540816), False, 'import pulumi\n'), ((540924, 540967), 'pulumi.set', 'pulumi.set', (['self', '"""assign_public_ip"""', 'value'], {}), "(self, 'assign_public_ip', value)\n", (540934, 540967), False, 'import pulumi\n'), ((541145, 541169), 'pulumi.get', 'pulumi.get', (['self', '"""fqdn"""'], {}), "(self, 'fqdn')\n", (541155, 541169), False, 'import pulumi\n'), ((541252, 541283), 'pulumi.set', 'pulumi.set', (['self', '"""fqdn"""', 'value'], {}), "(self, 'fqdn', value)\n", (541262, 541283), False, 'import pulumi\n'), ((541566, 541595), 'pulumi.get', 'pulumi.get', (['self', '"""subnet_id"""'], {}), "(self, 'subnet_id')\n", (541576, 541595), False, 'import pulumi\n'), ((541688, 541724), 'pulumi.set', 'pulumi.set', (['self', '"""subnet_id"""', 'value'], {}), "(self, 'subnet_id', value)\n", (541698, 541724), False, 'import pulumi\n'), ((542601, 542630), 'pulumi.get', 'pulumi.get', (['self', '"""data_lens"""'], {}), "(self, 'data_lens')\n", (542611, 542630), False, 'import pulumi\n'), ((542724, 542760), 'pulumi.set', 'pulumi.set', (['self', '"""data_lens"""', 'value'], {}), "(self, 'data_lens', value)\n", (542734, 542760), False, 'import pulumi\n'), ((542968, 542995), 'pulumi.get', 'pulumi.get', (['self', '"""web_sql"""'], {}), "(self, 'web_sql')\n", (542978, 542995), False, 'import pulumi\n'), ((543085, 543119), 'pulumi.set', 'pulumi.set', (['self', '"""web_sql"""', 'value'], {}), "(self, 'web_sql', value)\n", (543095, 543119), False, 'import pulumi\n'), ((543891, 543916), 'pulumi.get', 'pulumi.get', (['self', '"""hours"""'], {}), "(self, 'hours')\n", (543901, 543916), False, 'import pulumi\n'), ((544001, 544033), 'pulumi.set', 'pulumi.set', (['self', '"""hours"""', 'value'], {}), "(self, 'hours', value)\n", (544011, 544033), False, 'import pulumi\n'), ((544219, 544246), 'pulumi.get', 'pulumi.get', (['self', '"""minutes"""'], {}), "(self, 'minutes')\n", (544229, 544246), False, 'import pulumi\n'), ((544335, 544369), 'pulumi.set', 'pulumi.set', (['self', '"""minutes"""', 'value'], {}), "(self, 'minutes', value)\n", (544345, 544369), False, 'import pulumi\n'), ((545423, 545459), 'pulumi.get', 'pulumi.get', (['self', '"""assign_public_ip"""'], {}), "(self, 'assign_public_ip')\n", (545433, 545459), False, 'import pulumi\n'), ((545567, 545610), 'pulumi.set', 'pulumi.set', (['self', '"""assign_public_ip"""', 'value'], {}), "(self, 'assign_public_ip', value)\n", (545577, 545610), False, 'import pulumi\n'), ((545799, 545823), 'pulumi.get', 'pulumi.get', (['self', '"""fqdn"""'], {}), "(self, 'fqdn')\n", (545809, 545823), False, 'import pulumi\n'), ((545906, 545937), 'pulumi.set', 'pulumi.set', (['self', '"""fqdn"""', 'value'], {}), "(self, 'fqdn', value)\n", (545916, 545937), False, 'import pulumi\n'), ((546367, 546411), 'pulumi.set', 'pulumi.set', (['__self__', '"""resources"""', 'resources'], {}), "(__self__, 'resources', resources)\n", (546377, 546411), False, 'import pulumi\n'), ((546699, 546728), 'pulumi.get', 'pulumi.get', (['self', '"""resources"""'], {}), "(self, 'resources')\n", (546709, 546728), False, 'import pulumi\n'), ((546858, 546894), 'pulumi.set', 'pulumi.set', (['self', '"""resources"""', 'value'], {}), "(self, 'resources', value)\n", (546868, 546894), False, 'import pulumi\n'), ((547164, 547208), 'pulumi.set', 'pulumi.set', (['__self__', '"""disk_size"""', 'disk_size'], {}), "(__self__, 'disk_size', disk_size)\n", (547174, 547208), False, 'import pulumi\n'), ((547217, 547267), 'pulumi.set', 'pulumi.set', (['__self__', '"""disk_type_id"""', 'disk_type_id'], {}), "(__self__, 'disk_type_id', disk_type_id)\n", (547227, 547267), False, 'import pulumi\n'), ((547276, 547338), 'pulumi.set', 'pulumi.set', (['__self__', '"""resource_preset_id"""', 'resource_preset_id'], {}), "(__self__, 'resource_preset_id', resource_preset_id)\n", (547286, 547338), False, 'import pulumi\n'), ((547451, 547480), 'pulumi.get', 'pulumi.get', (['self', '"""disk_size"""'], {}), "(self, 'disk_size')\n", (547461, 547480), False, 'import pulumi\n'), ((547563, 547599), 'pulumi.set', 'pulumi.set', (['self', '"""disk_size"""', 'value'], {}), "(self, 'disk_size', value)\n", (547573, 547599), False, 'import pulumi\n'), ((547717, 547749), 'pulumi.get', 'pulumi.get', (['self', '"""disk_type_id"""'], {}), "(self, 'disk_type_id')\n", (547727, 547749), False, 'import pulumi\n'), ((547838, 547877), 'pulumi.set', 'pulumi.set', (['self', '"""disk_type_id"""', 'value'], {}), "(self, 'disk_type_id', value)\n", (547848, 547877), False, 'import pulumi\n'), ((548007, 548045), 'pulumi.get', 'pulumi.get', (['self', '"""resource_preset_id"""'], {}), "(self, 'resource_preset_id')\n", (548017, 548045), False, 'import pulumi\n'), ((548146, 548191), 'pulumi.set', 'pulumi.set', (['self', '"""resource_preset_id"""', 'value'], {}), "(self, 'resource_preset_id', value)\n", (548156, 548191), False, 'import pulumi\n'), ((548728, 548752), 'pulumi.get', 'pulumi.get', (['self', '"""fqdn"""'], {}), "(self, 'fqdn')\n", (548738, 548752), False, 'import pulumi\n'), ((548835, 548866), 'pulumi.set', 'pulumi.set', (['self', '"""fqdn"""', 'value'], {}), "(self, 'fqdn', value)\n", (548845, 548866), False, 'import pulumi\n'), ((549299, 549343), 'pulumi.set', 'pulumi.set', (['__self__', '"""resources"""', 'resources'], {}), "(__self__, 'resources', resources)\n", (549309, 549343), False, 'import pulumi\n'), ((549632, 549661), 'pulumi.get', 'pulumi.get', (['self', '"""resources"""'], {}), "(self, 'resources')\n", (549642, 549661), False, 'import pulumi\n'), ((549792, 549828), 'pulumi.set', 'pulumi.set', (['self', '"""resources"""', 'value'], {}), "(self, 'resources', value)\n", (549802, 549828), False, 'import pulumi\n'), ((550099, 550143), 'pulumi.set', 'pulumi.set', (['__self__', '"""disk_size"""', 'disk_size'], {}), "(__self__, 'disk_size', disk_size)\n", (550109, 550143), False, 'import pulumi\n'), ((550152, 550202), 'pulumi.set', 'pulumi.set', (['__self__', '"""disk_type_id"""', 'disk_type_id'], {}), "(__self__, 'disk_type_id', disk_type_id)\n", (550162, 550202), False, 'import pulumi\n'), ((550211, 550273), 'pulumi.set', 'pulumi.set', (['__self__', '"""resource_preset_id"""', 'resource_preset_id'], {}), "(__self__, 'resource_preset_id', resource_preset_id)\n", (550221, 550273), False, 'import pulumi\n'), ((550386, 550415), 'pulumi.get', 'pulumi.get', (['self', '"""disk_size"""'], {}), "(self, 'disk_size')\n", (550396, 550415), False, 'import pulumi\n'), ((550498, 550534), 'pulumi.set', 'pulumi.set', (['self', '"""disk_size"""', 'value'], {}), "(self, 'disk_size', value)\n", (550508, 550534), False, 'import pulumi\n'), ((550652, 550684), 'pulumi.get', 'pulumi.get', (['self', '"""disk_type_id"""'], {}), "(self, 'disk_type_id')\n", (550662, 550684), False, 'import pulumi\n'), ((550773, 550812), 'pulumi.set', 'pulumi.set', (['self', '"""disk_type_id"""', 'value'], {}), "(self, 'disk_type_id', value)\n", (550783, 550812), False, 'import pulumi\n'), ((550942, 550980), 'pulumi.get', 'pulumi.get', (['self', '"""resource_preset_id"""'], {}), "(self, 'resource_preset_id')\n", (550952, 550980), False, 'import pulumi\n'), ((551081, 551126), 'pulumi.set', 'pulumi.set', (['self', '"""resource_preset_id"""', 'value'], {}), "(self, 'resource_preset_id', value)\n", (551091, 551126), False, 'import pulumi\n'), ((552773, 552809), 'pulumi.set', 'pulumi.set', (['__self__', '"""kafka"""', 'kafka'], {}), "(__self__, 'kafka', kafka)\n", (552783, 552809), False, 'import pulumi\n'), ((552818, 552858), 'pulumi.set', 'pulumi.set', (['__self__', '"""version"""', 'version'], {}), "(__self__, 'version', version)\n", (552828, 552858), False, 'import pulumi\n'), ((552867, 552903), 'pulumi.set', 'pulumi.set', (['__self__', '"""zones"""', 'zones'], {}), "(__self__, 'zones', zones)\n", (552877, 552903), False, 'import pulumi\n'), ((553657, 553682), 'pulumi.get', 'pulumi.get', (['self', '"""kafka"""'], {}), "(self, 'kafka')\n", (553667, 553682), False, 'import pulumi\n'), ((553786, 553818), 'pulumi.set', 'pulumi.set', (['self', '"""kafka"""', 'value'], {}), "(self, 'kafka', value)\n", (553796, 553818), False, 'import pulumi\n'), ((553982, 554009), 'pulumi.get', 'pulumi.get', (['self', '"""version"""'], {}), "(self, 'version')\n", (553992, 554009), False, 'import pulumi\n'), ((554088, 554122), 'pulumi.set', 'pulumi.set', (['self', '"""version"""', 'value'], {}), "(self, 'version', value)\n", (554098, 554122), False, 'import pulumi\n'), ((554298, 554323), 'pulumi.get', 'pulumi.get', (['self', '"""zones"""'], {}), "(self, 'zones')\n", (554308, 554323), False, 'import pulumi\n'), ((554422, 554454), 'pulumi.set', 'pulumi.set', (['self', '"""zones"""', 'value'], {}), "(self, 'zones', value)\n", (554432, 554454), False, 'import pulumi\n'), ((554716, 554752), 'pulumi.get', 'pulumi.get', (['self', '"""assign_public_ip"""'], {}), "(self, 'assign_public_ip')\n", (554726, 554752), False, 'import pulumi\n'), ((554860, 554903), 'pulumi.set', 'pulumi.set', (['self', '"""assign_public_ip"""', 'value'], {}), "(self, 'assign_public_ip', value)\n", (554870, 554903), False, 'import pulumi\n'), ((555126, 555159), 'pulumi.get', 'pulumi.get', (['self', '"""brokers_count"""'], {}), "(self, 'brokers_count')\n", (555136, 555159), False, 'import pulumi\n'), ((555260, 555300), 'pulumi.set', 'pulumi.set', (['self', '"""brokers_count"""', 'value'], {}), "(self, 'brokers_count', value)\n", (555270, 555300), False, 'import pulumi\n'), ((555536, 555571), 'pulumi.get', 'pulumi.get', (['self', '"""schema_registry"""'], {}), "(self, 'schema_registry')\n", (555546, 555571), False, 'import pulumi\n'), ((555677, 555719), 'pulumi.set', 'pulumi.set', (['self', '"""schema_registry"""', 'value'], {}), "(self, 'schema_registry', value)\n", (555687, 555719), False, 'import pulumi\n'), ((555960, 555996), 'pulumi.get', 'pulumi.get', (['self', '"""unmanaged_topics"""'], {}), "(self, 'unmanaged_topics')\n", (555970, 555996), False, 'import pulumi\n'), ((556104, 556147), 'pulumi.set', 'pulumi.set', (['self', '"""unmanaged_topics"""', 'value'], {}), "(self, 'unmanaged_topics', value)\n", (556114, 556147), False, 'import pulumi\n'), ((556396, 556425), 'pulumi.get', 'pulumi.get', (['self', '"""zookeeper"""'], {}), "(self, 'zookeeper')\n", (556406, 556425), False, 'import pulumi\n'), ((556551, 556587), 'pulumi.set', 'pulumi.set', (['self', '"""zookeeper"""', 'value'], {}), "(self, 'zookeeper', value)\n", (556561, 556587), False, 'import pulumi\n'), ((557244, 557288), 'pulumi.set', 'pulumi.set', (['__self__', '"""resources"""', 'resources'], {}), "(__self__, 'resources', resources)\n", (557254, 557288), False, 'import pulumi\n'), ((557647, 557676), 'pulumi.get', 'pulumi.get', (['self', '"""resources"""'], {}), "(self, 'resources')\n", (557657, 557676), False, 'import pulumi\n'), ((557797, 557833), 'pulumi.set', 'pulumi.set', (['self', '"""resources"""', 'value'], {}), "(self, 'resources', value)\n", (557807, 557833), False, 'import pulumi\n'), ((558114, 558146), 'pulumi.get', 'pulumi.get', (['self', '"""kafka_config"""'], {}), "(self, 'kafka_config')\n", (558124, 558146), False, 'import pulumi\n'), ((558285, 558324), 'pulumi.set', 'pulumi.set', (['self', '"""kafka_config"""', 'value'], {}), "(self, 'kafka_config', value)\n", (558295, 558324), False, 'import pulumi\n'), ((561638, 561683), 'pulumi.get', 'pulumi.get', (['self', '"""auto_create_topics_enable"""'], {}), "(self, 'auto_create_topics_enable')\n", (561648, 561683), False, 'import pulumi\n'), ((561809, 561861), 'pulumi.set', 'pulumi.set', (['self', '"""auto_create_topics_enable"""', 'value'], {}), "(self, 'auto_create_topics_enable', value)\n", (561819, 561861), False, 'import pulumi\n'), ((561998, 562034), 'pulumi.get', 'pulumi.get', (['self', '"""compression_type"""'], {}), "(self, 'compression_type')\n", (562008, 562034), False, 'import pulumi\n'), ((562141, 562184), 'pulumi.set', 'pulumi.set', (['self', '"""compression_type"""', 'value'], {}), "(self, 'compression_type', value)\n", (562151, 562184), False, 'import pulumi\n'), ((562340, 562386), 'pulumi.get', 'pulumi.get', (['self', '"""default_replication_factor"""'], {}), "(self, 'default_replication_factor')\n", (562350, 562386), False, 'import pulumi\n'), ((562513, 562566), 'pulumi.set', 'pulumi.set', (['self', '"""default_replication_factor"""', 'value'], {}), "(self, 'default_replication_factor', value)\n", (562523, 562566), False, 'import pulumi\n'), ((562723, 562770), 'pulumi.get', 'pulumi.get', (['self', '"""log_flush_interval_messages"""'], {}), "(self, 'log_flush_interval_messages')\n", (562733, 562770), False, 'import pulumi\n'), ((562899, 562953), 'pulumi.set', 'pulumi.set', (['self', '"""log_flush_interval_messages"""', 'value'], {}), "(self, 'log_flush_interval_messages', value)\n", (562909, 562953), False, 'import pulumi\n'), ((563098, 563139), 'pulumi.get', 'pulumi.get', (['self', '"""log_flush_interval_ms"""'], {}), "(self, 'log_flush_interval_ms')\n", (563108, 563139), False, 'import pulumi\n'), ((563256, 563304), 'pulumi.set', 'pulumi.set', (['self', '"""log_flush_interval_ms"""', 'value'], {}), "(self, 'log_flush_interval_ms', value)\n", (563266, 563304), False, 'import pulumi\n'), ((563468, 563519), 'pulumi.get', 'pulumi.get', (['self', '"""log_flush_scheduler_interval_ms"""'], {}), "(self, 'log_flush_scheduler_interval_ms')\n", (563478, 563519), False, 'import pulumi\n'), ((563656, 563714), 'pulumi.set', 'pulumi.set', (['self', '"""log_flush_scheduler_interval_ms"""', 'value'], {}), "(self, 'log_flush_scheduler_interval_ms', value)\n", (563666, 563714), False, 'import pulumi\n'), ((563850, 563885), 'pulumi.get', 'pulumi.get', (['self', '"""log_preallocate"""'], {}), "(self, 'log_preallocate')\n", (563860, 563885), False, 'import pulumi\n'), ((563991, 564033), 'pulumi.set', 'pulumi.set', (['self', '"""log_preallocate"""', 'value'], {}), "(self, 'log_preallocate', value)\n", (564001, 564033), False, 'import pulumi\n'), ((564175, 564214), 'pulumi.get', 'pulumi.get', (['self', '"""log_retention_bytes"""'], {}), "(self, 'log_retention_bytes')\n", (564185, 564214), False, 'import pulumi\n'), ((564327, 564373), 'pulumi.set', 'pulumi.set', (['self', '"""log_retention_bytes"""', 'value'], {}), "(self, 'log_retention_bytes', value)\n", (564337, 564373), False, 'import pulumi\n'), ((564515, 564554), 'pulumi.get', 'pulumi.get', (['self', '"""log_retention_hours"""'], {}), "(self, 'log_retention_hours')\n", (564525, 564554), False, 'import pulumi\n'), ((564667, 564713), 'pulumi.set', 'pulumi.set', (['self', '"""log_retention_hours"""', 'value'], {}), "(self, 'log_retention_hours', value)\n", (564677, 564713), False, 'import pulumi\n'), ((564859, 564900), 'pulumi.get', 'pulumi.get', (['self', '"""log_retention_minutes"""'], {}), "(self, 'log_retention_minutes')\n", (564869, 564900), False, 'import pulumi\n'), ((565017, 565065), 'pulumi.set', 'pulumi.set', (['self', '"""log_retention_minutes"""', 'value'], {}), "(self, 'log_retention_minutes', value)\n", (565027, 565065), False, 'import pulumi\n'), ((565201, 565237), 'pulumi.get', 'pulumi.get', (['self', '"""log_retention_ms"""'], {}), "(self, 'log_retention_ms')\n", (565211, 565237), False, 'import pulumi\n'), ((565344, 565387), 'pulumi.set', 'pulumi.set', (['self', '"""log_retention_ms"""', 'value'], {}), "(self, 'log_retention_ms', value)\n", (565354, 565387), False, 'import pulumi\n'), ((565525, 565562), 'pulumi.get', 'pulumi.get', (['self', '"""log_segment_bytes"""'], {}), "(self, 'log_segment_bytes')\n", (565535, 565562), False, 'import pulumi\n'), ((565671, 565715), 'pulumi.set', 'pulumi.set', (['self', '"""log_segment_bytes"""', 'value'], {}), "(self, 'log_segment_bytes', value)\n", (565681, 565715), False, 'import pulumi\n'), ((565848, 565882), 'pulumi.get', 'pulumi.get', (['self', '"""num_partitions"""'], {}), "(self, 'num_partitions')\n", (565858, 565882), False, 'import pulumi\n'), ((565985, 566026), 'pulumi.set', 'pulumi.set', (['self', '"""num_partitions"""', 'value'], {}), "(self, 'num_partitions', value)\n", (565995, 566026), False, 'import pulumi\n'), ((566183, 566230), 'pulumi.get', 'pulumi.get', (['self', '"""socket_receive_buffer_bytes"""'], {}), "(self, 'socket_receive_buffer_bytes')\n", (566193, 566230), False, 'import pulumi\n'), ((566359, 566413), 'pulumi.set', 'pulumi.set', (['self', '"""socket_receive_buffer_bytes"""', 'value'], {}), "(self, 'socket_receive_buffer_bytes', value)\n", (566369, 566413), False, 'import pulumi\n'), ((566564, 566608), 'pulumi.get', 'pulumi.get', (['self', '"""socket_send_buffer_bytes"""'], {}), "(self, 'socket_send_buffer_bytes')\n", (566574, 566608), False, 'import pulumi\n'), ((566731, 566782), 'pulumi.set', 'pulumi.set', (['self', '"""socket_send_buffer_bytes"""', 'value'], {}), "(self, 'socket_send_buffer_bytes', value)\n", (566741, 566782), False, 'import pulumi\n'), ((567397, 567441), 'pulumi.set', 'pulumi.set', (['__self__', '"""disk_size"""', 'disk_size'], {}), "(__self__, 'disk_size', disk_size)\n", (567407, 567441), False, 'import pulumi\n'), ((567450, 567500), 'pulumi.set', 'pulumi.set', (['__self__', '"""disk_type_id"""', 'disk_type_id'], {}), "(__self__, 'disk_type_id', disk_type_id)\n", (567460, 567500), False, 'import pulumi\n'), ((567509, 567571), 'pulumi.set', 'pulumi.set', (['__self__', '"""resource_preset_id"""', 'resource_preset_id'], {}), "(__self__, 'resource_preset_id', resource_preset_id)\n", (567519, 567571), False, 'import pulumi\n'), ((567783, 567812), 'pulumi.get', 'pulumi.get', (['self', '"""disk_size"""'], {}), "(self, 'disk_size')\n", (567793, 567812), False, 'import pulumi\n'), ((567895, 567931), 'pulumi.set', 'pulumi.set', (['self', '"""disk_size"""', 'value'], {}), "(self, 'disk_size', value)\n", (567905, 567931), False, 'import pulumi\n'), ((568246, 568278), 'pulumi.get', 'pulumi.get', (['self', '"""disk_type_id"""'], {}), "(self, 'disk_type_id')\n", (568256, 568278), False, 'import pulumi\n'), ((568367, 568406), 'pulumi.set', 'pulumi.set', (['self', '"""disk_type_id"""', 'value'], {}), "(self, 'disk_type_id', value)\n", (568377, 568406), False, 'import pulumi\n'), ((568536, 568574), 'pulumi.get', 'pulumi.get', (['self', '"""resource_preset_id"""'], {}), "(self, 'resource_preset_id')\n", (568546, 568574), False, 'import pulumi\n'), ((568675, 568720), 'pulumi.set', 'pulumi.set', (['self', '"""resource_preset_id"""', 'value'], {}), "(self, 'resource_preset_id', value)\n", (568685, 568720), False, 'import pulumi\n'), ((569487, 569516), 'pulumi.get', 'pulumi.get', (['self', '"""resources"""'], {}), "(self, 'resources')\n", (569497, 569516), False, 'import pulumi\n'), ((569651, 569687), 'pulumi.set', 'pulumi.set', (['self', '"""resources"""', 'value'], {}), "(self, 'resources', value)\n", (569661, 569687), False, 'import pulumi\n'), ((570879, 570908), 'pulumi.get', 'pulumi.get', (['self', '"""disk_size"""'], {}), "(self, 'disk_size')\n", (570889, 570908), False, 'import pulumi\n'), ((571001, 571037), 'pulumi.set', 'pulumi.set', (['self', '"""disk_size"""', 'value'], {}), "(self, 'disk_size', value)\n", (571011, 571037), False, 'import pulumi\n'), ((571362, 571394), 'pulumi.get', 'pulumi.get', (['self', '"""disk_type_id"""'], {}), "(self, 'disk_type_id')\n", (571372, 571394), False, 'import pulumi\n'), ((571493, 571532), 'pulumi.set', 'pulumi.set', (['self', '"""disk_type_id"""', 'value'], {}), "(self, 'disk_type_id', value)\n", (571503, 571532), False, 'import pulumi\n'), ((571672, 571710), 'pulumi.get', 'pulumi.get', (['self', '"""resource_preset_id"""'], {}), "(self, 'resource_preset_id')\n", (571682, 571710), False, 'import pulumi\n'), ((571821, 571866), 'pulumi.set', 'pulumi.set', (['self', '"""resource_preset_id"""', 'value'], {}), "(self, 'resource_preset_id', value)\n", (571831, 571866), False, 'import pulumi\n'), ((573669, 573705), 'pulumi.get', 'pulumi.get', (['self', '"""assign_public_ip"""'], {}), "(self, 'assign_public_ip')\n", (573679, 573705), False, 'import pulumi\n'), ((573813, 573856), 'pulumi.set', 'pulumi.set', (['self', '"""assign_public_ip"""', 'value'], {}), "(self, 'assign_public_ip', value)\n", (573823, 573856), False, 'import pulumi\n'), ((574011, 574037), 'pulumi.get', 'pulumi.get', (['self', '"""health"""'], {}), "(self, 'health')\n", (574021, 574037), False, 'import pulumi\n'), ((574124, 574157), 'pulumi.set', 'pulumi.set', (['self', '"""health"""', 'value'], {}), "(self, 'health', value)\n", (574134, 574157), False, 'import pulumi\n'), ((574313, 574337), 'pulumi.get', 'pulumi.get', (['self', '"""name"""'], {}), "(self, 'name')\n", (574323, 574337), False, 'import pulumi\n'), ((574420, 574451), 'pulumi.set', 'pulumi.set', (['self', '"""name"""', 'value'], {}), "(self, 'name', value)\n", (574430, 574451), False, 'import pulumi\n'), ((574621, 574645), 'pulumi.get', 'pulumi.get', (['self', '"""role"""'], {}), "(self, 'role')\n", (574631, 574645), False, 'import pulumi\n'), ((574728, 574759), 'pulumi.set', 'pulumi.set', (['self', '"""role"""', 'value'], {}), "(self, 'role', value)\n", (574738, 574759), False, 'import pulumi\n'), ((574963, 574992), 'pulumi.get', 'pulumi.get', (['self', '"""subnet_id"""'], {}), "(self, 'subnet_id')\n", (574973, 574992), False, 'import pulumi\n'), ((575085, 575121), 'pulumi.set', 'pulumi.set', (['self', '"""subnet_id"""', 'value'], {}), "(self, 'subnet_id', value)\n", (575095, 575121), False, 'import pulumi\n'), ((575328, 575355), 'pulumi.get', 'pulumi.get', (['self', '"""zone_id"""'], {}), "(self, 'zone_id')\n", (575338, 575355), False, 'import pulumi\n'), ((575444, 575478), 'pulumi.set', 'pulumi.set', (['self', '"""zone_id"""', 'value'], {}), "(self, 'zone_id', value)\n", (575454, 575478), False, 'import pulumi\n'), ((576256, 576290), 'pulumi.set', 'pulumi.set', (['__self__', '"""name"""', 'name'], {}), "(__self__, 'name', name)\n", (576266, 576290), False, 'import pulumi\n'), ((576299, 576345), 'pulumi.set', 'pulumi.set', (['__self__', '"""partitions"""', 'partitions'], {}), "(__self__, 'partitions', partitions)\n", (576309, 576345), False, 'import pulumi\n'), ((576354, 576416), 'pulumi.set', 'pulumi.set', (['__self__', '"""replication_factor"""', 'replication_factor'], {}), "(__self__, 'replication_factor', replication_factor)\n", (576364, 576416), False, 'import pulumi\n'), ((576662, 576686), 'pulumi.get', 'pulumi.get', (['self', '"""name"""'], {}), "(self, 'name')\n", (576672, 576686), False, 'import pulumi\n'), ((576759, 576790), 'pulumi.set', 'pulumi.set', (['self', '"""name"""', 'value'], {}), "(self, 'name', value)\n", (576769, 576790), False, 'import pulumi\n'), ((576957, 576987), 'pulumi.get', 'pulumi.get', (['self', '"""partitions"""'], {}), "(self, 'partitions')\n", (576967, 576987), False, 'import pulumi\n'), ((577072, 577109), 'pulumi.set', 'pulumi.set', (['self', '"""partitions"""', 'value'], {}), "(self, 'partitions', value)\n", (577082, 577109), False, 'import pulumi\n'), ((577335, 577373), 'pulumi.get', 'pulumi.get', (['self', '"""replication_factor"""'], {}), "(self, 'replication_factor')\n", (577345, 577373), False, 'import pulumi\n'), ((577474, 577519), 'pulumi.set', 'pulumi.set', (['self', '"""replication_factor"""', 'value'], {}), "(self, 'replication_factor', value)\n", (577484, 577519), False, 'import pulumi\n'), ((577786, 577818), 'pulumi.get', 'pulumi.get', (['self', '"""topic_config"""'], {}), "(self, 'topic_config')\n", (577796, 577818), False, 'import pulumi\n'), ((577951, 577990), 'pulumi.set', 'pulumi.set', (['self', '"""topic_config"""', 'value'], {}), "(self, 'topic_config', value)\n", (577961, 577990), False, 'import pulumi\n'), ((580562, 580596), 'pulumi.get', 'pulumi.get', (['self', '"""cleanup_policy"""'], {}), "(self, 'cleanup_policy')\n", (580572, 580596), False, 'import pulumi\n'), ((580699, 580740), 'pulumi.set', 'pulumi.set', (['self', '"""cleanup_policy"""', 'value'], {}), "(self, 'cleanup_policy', value)\n", (580709, 580740), False, 'import pulumi\n'), ((580877, 580913), 'pulumi.get', 'pulumi.get', (['self', '"""compression_type"""'], {}), "(self, 'compression_type')\n", (580887, 580913), False, 'import pulumi\n'), ((581020, 581063), 'pulumi.set', 'pulumi.set', (['self', '"""compression_type"""', 'value'], {}), "(self, 'compression_type', value)\n", (581030, 581063), False, 'import pulumi\n'), ((581205, 581244), 'pulumi.get', 'pulumi.get', (['self', '"""delete_retention_ms"""'], {}), "(self, 'delete_retention_ms')\n", (581215, 581244), False, 'import pulumi\n'), ((581357, 581403), 'pulumi.set', 'pulumi.set', (['self', '"""delete_retention_ms"""', 'value'], {}), "(self, 'delete_retention_ms', value)\n", (581367, 581403), False, 'import pulumi\n'), ((581546, 581586), 'pulumi.get', 'pulumi.get', (['self', '"""file_delete_delay_ms"""'], {}), "(self, 'file_delete_delay_ms')\n", (581556, 581586), False, 'import pulumi\n'), ((581701, 581748), 'pulumi.set', 'pulumi.set', (['self', '"""file_delete_delay_ms"""', 'value'], {}), "(self, 'file_delete_delay_ms', value)\n", (581711, 581748), False, 'import pulumi\n'), ((581881, 581915), 'pulumi.get', 'pulumi.get', (['self', '"""flush_messages"""'], {}), "(self, 'flush_messages')\n", (581891, 581915), False, 'import pulumi\n'), ((582018, 582059), 'pulumi.set', 'pulumi.set', (['self', '"""flush_messages"""', 'value'], {}), "(self, 'flush_messages', value)\n", (582028, 582059), False, 'import pulumi\n'), ((582180, 582208), 'pulumi.get', 'pulumi.get', (['self', '"""flush_ms"""'], {}), "(self, 'flush_ms')\n", (582190, 582208), False, 'import pulumi\n'), ((582299, 582334), 'pulumi.set', 'pulumi.set', (['self', '"""flush_ms"""', 'value'], {}), "(self, 'flush_ms', value)\n", (582309, 582334), False, 'import pulumi\n'), ((582472, 582509), 'pulumi.get', 'pulumi.get', (['self', '"""max_message_bytes"""'], {}), "(self, 'max_message_bytes')\n", (582482, 582509), False, 'import pulumi\n'), ((582618, 582662), 'pulumi.set', 'pulumi.set', (['self', '"""max_message_bytes"""', 'value'], {}), "(self, 'max_message_bytes', value)\n", (582628, 582662), False, 'import pulumi\n'), ((582807, 582848), 'pulumi.get', 'pulumi.get', (['self', '"""min_compaction_lag_ms"""'], {}), "(self, 'min_compaction_lag_ms')\n", (582817, 582848), False, 'import pulumi\n'), ((582965, 583013), 'pulumi.set', 'pulumi.set', (['self', '"""min_compaction_lag_ms"""', 'value'], {}), "(self, 'min_compaction_lag_ms', value)\n", (582975, 583013), False, 'import pulumi\n'), ((583155, 583194), 'pulumi.get', 'pulumi.get', (['self', '"""min_insync_replicas"""'], {}), "(self, 'min_insync_replicas')\n", (583165, 583194), False, 'import pulumi\n'), ((583307, 583353), 'pulumi.set', 'pulumi.set', (['self', '"""min_insync_replicas"""', 'value'], {}), "(self, 'min_insync_replicas', value)\n", (583317, 583353), False, 'import pulumi\n'), ((583462, 583493), 'pulumi.get', 'pulumi.get', (['self', '"""preallocate"""'], {}), "(self, 'preallocate')\n", (583472, 583493), False, 'import pulumi\n'), ((583591, 583629), 'pulumi.set', 'pulumi.set', (['self', '"""preallocate"""', 'value'], {}), "(self, 'preallocate', value)\n", (583601, 583629), False, 'import pulumi\n'), ((583764, 583799), 'pulumi.get', 'pulumi.get', (['self', '"""retention_bytes"""'], {}), "(self, 'retention_bytes')\n", (583774, 583799), False, 'import pulumi\n'), ((583904, 583946), 'pulumi.set', 'pulumi.set', (['self', '"""retention_bytes"""', 'value'], {}), "(self, 'retention_bytes', value)\n", (583914, 583946), False, 'import pulumi\n'), ((584075, 584107), 'pulumi.get', 'pulumi.get', (['self', '"""retention_ms"""'], {}), "(self, 'retention_ms')\n", (584085, 584107), False, 'import pulumi\n'), ((584206, 584245), 'pulumi.set', 'pulumi.set', (['self', '"""retention_ms"""', 'value'], {}), "(self, 'retention_ms', value)\n", (584216, 584245), False, 'import pulumi\n'), ((584376, 584409), 'pulumi.get', 'pulumi.get', (['self', '"""segment_bytes"""'], {}), "(self, 'segment_bytes')\n", (584386, 584409), False, 'import pulumi\n'), ((584510, 584550), 'pulumi.set', 'pulumi.set', (['self', '"""segment_bytes"""', 'value'], {}), "(self, 'segment_bytes', value)\n", (584520, 584550), False, 'import pulumi\n'), ((585189, 585223), 'pulumi.set', 'pulumi.set', (['__self__', '"""name"""', 'name'], {}), "(__self__, 'name', name)\n", (585199, 585223), False, 'import pulumi\n'), ((585232, 585274), 'pulumi.set', 'pulumi.set', (['__self__', '"""password"""', 'password'], {}), "(__self__, 'password', password)\n", (585242, 585274), False, 'import pulumi\n'), ((585517, 585541), 'pulumi.get', 'pulumi.get', (['self', '"""name"""'], {}), "(self, 'name')\n", (585527, 585541), False, 'import pulumi\n'), ((585614, 585645), 'pulumi.set', 'pulumi.set', (['self', '"""name"""', 'value'], {}), "(self, 'name', value)\n", (585624, 585645), False, 'import pulumi\n'), ((585798, 585826), 'pulumi.get', 'pulumi.get', (['self', '"""password"""'], {}), "(self, 'password')\n", (585808, 585826), False, 'import pulumi\n'), ((585907, 585942), 'pulumi.set', 'pulumi.set', (['self', '"""password"""', 'value'], {}), "(self, 'password', value)\n", (585917, 585942), False, 'import pulumi\n'), ((586213, 586244), 'pulumi.get', 'pulumi.get', (['self', '"""permissions"""'], {}), "(self, 'permissions')\n", (586223, 586244), False, 'import pulumi\n'), ((586397, 586435), 'pulumi.set', 'pulumi.set', (['self', '"""permissions"""', 'value'], {}), "(self, 'permissions', value)\n", (586407, 586435), False, 'import pulumi\n'), ((586832, 586866), 'pulumi.set', 'pulumi.set', (['__self__', '"""role"""', 'role'], {}), "(__self__, 'role', role)\n", (586842, 586866), False, 'import pulumi\n'), ((586875, 586921), 'pulumi.set', 'pulumi.set', (['__self__', '"""topic_name"""', 'topic_name'], {}), "(__self__, 'topic_name', topic_name)\n", (586885, 586921), False, 'import pulumi\n'), ((587081, 587105), 'pulumi.get', 'pulumi.get', (['self', '"""role"""'], {}), "(self, 'role')\n", (587091, 587105), False, 'import pulumi\n'), ((587178, 587209), 'pulumi.set', 'pulumi.set', (['self', '"""role"""', 'value'], {}), "(self, 'role', value)\n", (587188, 587209), False, 'import pulumi\n'), ((587416, 587446), 'pulumi.get', 'pulumi.get', (['self', '"""topic_name"""'], {}), "(self, 'topic_name')\n", (587426, 587446), False, 'import pulumi\n'), ((587531, 587568), 'pulumi.set', 'pulumi.set', (['self', '"""topic_name"""', 'value'], {}), "(self, 'topic_name', value)\n", (587541, 587568), False, 'import pulumi\n'), ((590133, 590167), 'pulumi.get', 'pulumi.get', (['self', '"""cleanup_policy"""'], {}), "(self, 'cleanup_policy')\n", (590143, 590167), False, 'import pulumi\n'), ((590270, 590311), 'pulumi.set', 'pulumi.set', (['self', '"""cleanup_policy"""', 'value'], {}), "(self, 'cleanup_policy', value)\n", (590280, 590311), False, 'import pulumi\n'), ((590448, 590484), 'pulumi.get', 'pulumi.get', (['self', '"""compression_type"""'], {}), "(self, 'compression_type')\n", (590458, 590484), False, 'import pulumi\n'), ((590591, 590634), 'pulumi.set', 'pulumi.set', (['self', '"""compression_type"""', 'value'], {}), "(self, 'compression_type', value)\n", (590601, 590634), False, 'import pulumi\n'), ((590776, 590815), 'pulumi.get', 'pulumi.get', (['self', '"""delete_retention_ms"""'], {}), "(self, 'delete_retention_ms')\n", (590786, 590815), False, 'import pulumi\n'), ((590928, 590974), 'pulumi.set', 'pulumi.set', (['self', '"""delete_retention_ms"""', 'value'], {}), "(self, 'delete_retention_ms', value)\n", (590938, 590974), False, 'import pulumi\n'), ((591117, 591157), 'pulumi.get', 'pulumi.get', (['self', '"""file_delete_delay_ms"""'], {}), "(self, 'file_delete_delay_ms')\n", (591127, 591157), False, 'import pulumi\n'), ((591272, 591319), 'pulumi.set', 'pulumi.set', (['self', '"""file_delete_delay_ms"""', 'value'], {}), "(self, 'file_delete_delay_ms', value)\n", (591282, 591319), False, 'import pulumi\n'), ((591452, 591486), 'pulumi.get', 'pulumi.get', (['self', '"""flush_messages"""'], {}), "(self, 'flush_messages')\n", (591462, 591486), False, 'import pulumi\n'), ((591589, 591630), 'pulumi.set', 'pulumi.set', (['self', '"""flush_messages"""', 'value'], {}), "(self, 'flush_messages', value)\n", (591599, 591630), False, 'import pulumi\n'), ((591751, 591779), 'pulumi.get', 'pulumi.get', (['self', '"""flush_ms"""'], {}), "(self, 'flush_ms')\n", (591761, 591779), False, 'import pulumi\n'), ((591870, 591905), 'pulumi.set', 'pulumi.set', (['self', '"""flush_ms"""', 'value'], {}), "(self, 'flush_ms', value)\n", (591880, 591905), False, 'import pulumi\n'), ((592043, 592080), 'pulumi.get', 'pulumi.get', (['self', '"""max_message_bytes"""'], {}), "(self, 'max_message_bytes')\n", (592053, 592080), False, 'import pulumi\n'), ((592189, 592233), 'pulumi.set', 'pulumi.set', (['self', '"""max_message_bytes"""', 'value'], {}), "(self, 'max_message_bytes', value)\n", (592199, 592233), False, 'import pulumi\n'), ((592378, 592419), 'pulumi.get', 'pulumi.get', (['self', '"""min_compaction_lag_ms"""'], {}), "(self, 'min_compaction_lag_ms')\n", (592388, 592419), False, 'import pulumi\n'), ((592536, 592584), 'pulumi.set', 'pulumi.set', (['self', '"""min_compaction_lag_ms"""', 'value'], {}), "(self, 'min_compaction_lag_ms', value)\n", (592546, 592584), False, 'import pulumi\n'), ((592726, 592765), 'pulumi.get', 'pulumi.get', (['self', '"""min_insync_replicas"""'], {}), "(self, 'min_insync_replicas')\n", (592736, 592765), False, 'import pulumi\n'), ((592878, 592924), 'pulumi.set', 'pulumi.set', (['self', '"""min_insync_replicas"""', 'value'], {}), "(self, 'min_insync_replicas', value)\n", (592888, 592924), False, 'import pulumi\n'), ((593033, 593064), 'pulumi.get', 'pulumi.get', (['self', '"""preallocate"""'], {}), "(self, 'preallocate')\n", (593043, 593064), False, 'import pulumi\n'), ((593162, 593200), 'pulumi.set', 'pulumi.set', (['self', '"""preallocate"""', 'value'], {}), "(self, 'preallocate', value)\n", (593172, 593200), False, 'import pulumi\n'), ((593335, 593370), 'pulumi.get', 'pulumi.get', (['self', '"""retention_bytes"""'], {}), "(self, 'retention_bytes')\n", (593345, 593370), False, 'import pulumi\n'), ((593475, 593517), 'pulumi.set', 'pulumi.set', (['self', '"""retention_bytes"""', 'value'], {}), "(self, 'retention_bytes', value)\n", (593485, 593517), False, 'import pulumi\n'), ((593646, 593678), 'pulumi.get', 'pulumi.get', (['self', '"""retention_ms"""'], {}), "(self, 'retention_ms')\n", (593656, 593678), False, 'import pulumi\n'), ((593777, 593816), 'pulumi.set', 'pulumi.set', (['self', '"""retention_ms"""', 'value'], {}), "(self, 'retention_ms', value)\n", (593787, 593816), False, 'import pulumi\n'), ((593947, 593980), 'pulumi.get', 'pulumi.get', (['self', '"""segment_bytes"""'], {}), "(self, 'segment_bytes')\n", (593957, 593980), False, 'import pulumi\n'), ((594081, 594121), 'pulumi.set', 'pulumi.set', (['self', '"""segment_bytes"""', 'value'], {}), "(self, 'segment_bytes', value)\n", (594091, 594121), False, 'import pulumi\n'), ((595232, 595272), 'pulumi.set', 'pulumi.set', (['__self__', '"""version"""', 'version'], {}), "(__self__, 'version', version)\n", (595242, 595272), False, 'import pulumi\n'), ((595802, 595829), 'pulumi.get', 'pulumi.get', (['self', '"""version"""'], {}), "(self, 'version')\n", (595812, 595829), False, 'import pulumi\n'), ((595908, 595942), 'pulumi.set', 'pulumi.set', (['self', '"""version"""', 'value'], {}), "(self, 'version', value)\n", (595918, 595942), False, 'import pulumi\n'), ((596198, 596224), 'pulumi.get', 'pulumi.get', (['self', '"""access"""'], {}), "(self, 'access')\n", (596208, 596224), False, 'import pulumi\n'), ((596350, 596383), 'pulumi.set', 'pulumi.set', (['self', '"""access"""', 'value'], {}), "(self, 'access', value)\n", (596360, 596383), False, 'import pulumi\n'), ((596695, 596734), 'pulumi.get', 'pulumi.get', (['self', '"""backup_window_start"""'], {}), "(self, 'backup_window_start')\n", (596705, 596734), False, 'import pulumi\n'), ((596897, 596943), 'pulumi.set', 'pulumi.set', (['self', '"""backup_window_start"""', 'value'], {}), "(self, 'backup_window_start', value)\n", (596907, 596943), False, 'import pulumi\n'), ((597258, 597307), 'pulumi.get', 'pulumi.get', (['self', '"""feature_compatibility_version"""'], {}), "(self, 'feature_compatibility_version')\n", (597268, 597307), False, 'import pulumi\n'), ((597440, 597496), 'pulumi.set', 'pulumi.set', (['self', '"""feature_compatibility_version"""', 'value'], {}), "(self, 'feature_compatibility_version', value)\n", (597450, 597496), False, 'import pulumi\n'), ((598031, 598060), 'pulumi.get', 'pulumi.get', (['self', '"""data_lens"""'], {}), "(self, 'data_lens')\n", (598041, 598060), False, 'import pulumi\n'), ((598154, 598190), 'pulumi.set', 'pulumi.set', (['self', '"""data_lens"""', 'value'], {}), "(self, 'data_lens', value)\n", (598164, 598190), False, 'import pulumi\n'), ((598955, 598980), 'pulumi.get', 'pulumi.get', (['self', '"""hours"""'], {}), "(self, 'hours')\n", (598965, 598980), False, 'import pulumi\n'), ((599065, 599097), 'pulumi.set', 'pulumi.set', (['self', '"""hours"""', 'value'], {}), "(self, 'hours', value)\n", (599075, 599097), False, 'import pulumi\n'), ((599277, 599304), 'pulumi.get', 'pulumi.get', (['self', '"""minutes"""'], {}), "(self, 'minutes')\n", (599287, 599304), False, 'import pulumi\n'), ((599393, 599427), 'pulumi.set', 'pulumi.set', (['self', '"""minutes"""', 'value'], {}), "(self, 'minutes', value)\n", (599403, 599427), False, 'import pulumi\n'), ((599700, 599734), 'pulumi.set', 'pulumi.set', (['__self__', '"""name"""', 'name'], {}), "(__self__, 'name', name)\n", (599710, 599734), False, 'import pulumi\n'), ((599927, 599951), 'pulumi.get', 'pulumi.get', (['self', '"""name"""'], {}), "(self, 'name')\n", (599937, 599951), False, 'import pulumi\n'), ((600024, 600055), 'pulumi.set', 'pulumi.set', (['self', '"""name"""', 'value'], {}), "(self, 'name', value)\n", (600034, 600055), False, 'import pulumi\n'), ((601733, 601777), 'pulumi.set', 'pulumi.set', (['__self__', '"""subnet_id"""', 'subnet_id'], {}), "(__self__, 'subnet_id', subnet_id)\n", (601743, 601777), False, 'import pulumi\n'), ((601786, 601826), 'pulumi.set', 'pulumi.set', (['__self__', '"""zone_id"""', 'zone_id'], {}), "(__self__, 'zone_id', zone_id)\n", (601796, 601826), False, 'import pulumi\n'), ((602615, 602644), 'pulumi.get', 'pulumi.get', (['self', '"""subnet_id"""'], {}), "(self, 'subnet_id')\n", (602625, 602644), False, 'import pulumi\n'), ((602727, 602763), 'pulumi.set', 'pulumi.set', (['self', '"""subnet_id"""', 'value'], {}), "(self, 'subnet_id', value)\n", (602737, 602763), False, 'import pulumi\n'), ((603088, 603115), 'pulumi.get', 'pulumi.get', (['self', '"""zone_id"""'], {}), "(self, 'zone_id')\n", (603098, 603115), False, 'import pulumi\n'), ((603194, 603228), 'pulumi.set', 'pulumi.set', (['self', '"""zone_id"""', 'value'], {}), "(self, 'zone_id', value)\n", (603204, 603228), False, 'import pulumi\n'), ((603494, 603530), 'pulumi.get', 'pulumi.get', (['self', '"""assign_public_ip"""'], {}), "(self, 'assign_public_ip')\n", (603504, 603530), False, 'import pulumi\n'), ((603638, 603681), 'pulumi.set', 'pulumi.set', (['self', '"""assign_public_ip"""', 'value'], {}), "(self, 'assign_public_ip', value)\n", (603648, 603681), False, 'import pulumi\n'), ((603840, 603866), 'pulumi.get', 'pulumi.get', (['self', '"""health"""'], {}), "(self, 'health')\n", (603850, 603866), False, 'import pulumi\n'), ((603953, 603986), 'pulumi.set', 'pulumi.set', (['self', '"""health"""', 'value'], {}), "(self, 'health', value)\n", (603963, 603986), False, 'import pulumi\n'), ((604189, 604213), 'pulumi.get', 'pulumi.get', (['self', '"""name"""'], {}), "(self, 'name')\n", (604199, 604213), False, 'import pulumi\n'), ((604296, 604327), 'pulumi.set', 'pulumi.set', (['self', '"""name"""', 'value'], {}), "(self, 'name', value)\n", (604306, 604327), False, 'import pulumi\n'), ((604515, 604539), 'pulumi.get', 'pulumi.get', (['self', '"""role"""'], {}), "(self, 'role')\n", (604525, 604539), False, 'import pulumi\n'), ((604622, 604653), 'pulumi.set', 'pulumi.set', (['self', '"""role"""', 'value'], {}), "(self, 'role', value)\n", (604632, 604653), False, 'import pulumi\n'), ((604859, 604889), 'pulumi.get', 'pulumi.get', (['self', '"""shard_name"""'], {}), "(self, 'shard_name')\n", (604869, 604889), False, 'import pulumi\n'), ((604984, 605021), 'pulumi.set', 'pulumi.set', (['self', '"""shard_name"""', 'value'], {}), "(self, 'shard_name', value)\n", (604994, 605021), False, 'import pulumi\n'), ((605285, 605309), 'pulumi.get', 'pulumi.get', (['self', '"""type"""'], {}), "(self, 'type')\n", (605295, 605309), False, 'import pulumi\n'), ((605392, 605423), 'pulumi.set', 'pulumi.set', (['self', '"""type"""', 'value'], {}), "(self, 'type', value)\n", (605402, 605423), False, 'import pulumi\n'), ((606172, 606206), 'pulumi.set', 'pulumi.set', (['__self__', '"""type"""', 'type'], {}), "(__self__, 'type', type)\n", (606182, 606206), False, 'import pulumi\n'), ((606609, 606633), 'pulumi.get', 'pulumi.get', (['self', '"""type"""'], {}), "(self, 'type')\n", (606619, 606633), False, 'import pulumi\n'), ((606706, 606737), 'pulumi.set', 'pulumi.set', (['self', '"""type"""', 'value'], {}), "(self, 'type', value)\n", (606716, 606737), False, 'import pulumi\n'), ((606996, 607019), 'pulumi.get', 'pulumi.get', (['self', '"""day"""'], {}), "(self, 'day')\n", (607006, 607019), False, 'import pulumi\n'), ((607100, 607130), 'pulumi.set', 'pulumi.set', (['self', '"""day"""', 'value'], {}), "(self, 'day', value)\n", (607110, 607130), False, 'import pulumi\n'), ((607348, 607372), 'pulumi.get', 'pulumi.get', (['self', '"""hour"""'], {}), "(self, 'hour')\n", (607358, 607372), False, 'import pulumi\n'), ((607455, 607486), 'pulumi.set', 'pulumi.set', (['self', '"""hour"""', 'value'], {}), "(self, 'hour', value)\n", (607465, 607486), False, 'import pulumi\n'), ((608093, 608137), 'pulumi.set', 'pulumi.set', (['__self__', '"""disk_size"""', 'disk_size'], {}), "(__self__, 'disk_size', disk_size)\n", (608103, 608137), False, 'import pulumi\n'), ((608146, 608196), 'pulumi.set', 'pulumi.set', (['__self__', '"""disk_type_id"""', 'disk_type_id'], {}), "(__self__, 'disk_type_id', disk_type_id)\n", (608156, 608196), False, 'import pulumi\n'), ((608205, 608267), 'pulumi.set', 'pulumi.set', (['__self__', '"""resource_preset_id"""', 'resource_preset_id'], {}), "(__self__, 'resource_preset_id', resource_preset_id)\n", (608215, 608267), False, 'import pulumi\n'), ((608477, 608506), 'pulumi.get', 'pulumi.get', (['self', '"""disk_size"""'], {}), "(self, 'disk_size')\n", (608487, 608506), False, 'import pulumi\n'), ((608589, 608625), 'pulumi.set', 'pulumi.set', (['self', '"""disk_size"""', 'value'], {}), "(self, 'disk_size', value)\n", (608599, 608625), False, 'import pulumi\n'), ((608943, 608975), 'pulumi.get', 'pulumi.get', (['self', '"""disk_type_id"""'], {}), "(self, 'disk_type_id')\n", (608953, 608975), False, 'import pulumi\n'), ((609064, 609103), 'pulumi.set', 'pulumi.set', (['self', '"""disk_type_id"""', 'value'], {}), "(self, 'disk_type_id', value)\n", (609074, 609103), False, 'import pulumi\n'), ((609233, 609271), 'pulumi.get', 'pulumi.get', (['self', '"""resource_preset_id"""'], {}), "(self, 'resource_preset_id')\n", (609243, 609271), False, 'import pulumi\n'), ((609372, 609417), 'pulumi.set', 'pulumi.set', (['self', '"""resource_preset_id"""', 'value'], {}), "(self, 'resource_preset_id', value)\n", (609382, 609417), False, 'import pulumi\n'), ((610090, 610124), 'pulumi.set', 'pulumi.set', (['__self__', '"""name"""', 'name'], {}), "(__self__, 'name', name)\n", (610100, 610124), False, 'import pulumi\n'), ((610133, 610175), 'pulumi.set', 'pulumi.set', (['__self__', '"""password"""', 'password'], {}), "(__self__, 'password', password)\n", (610143, 610175), False, 'import pulumi\n'), ((610465, 610489), 'pulumi.get', 'pulumi.get', (['self', '"""name"""'], {}), "(self, 'name')\n", (610475, 610489), False, 'import pulumi\n'), ((610562, 610593), 'pulumi.set', 'pulumi.set', (['self', '"""name"""', 'value'], {}), "(self, 'name', value)\n", (610572, 610593), False, 'import pulumi\n'), ((610746, 610774), 'pulumi.get', 'pulumi.get', (['self', '"""password"""'], {}), "(self, 'password')\n", (610756, 610774), False, 'import pulumi\n'), ((610855, 610890), 'pulumi.set', 'pulumi.set', (['self', '"""password"""', 'value'], {}), "(self, 'password', value)\n", (610865, 610890), False, 'import pulumi\n'), ((611163, 611194), 'pulumi.get', 'pulumi.get', (['self', '"""permissions"""'], {}), "(self, 'permissions')\n", (611173, 611194), False, 'import pulumi\n'), ((611349, 611387), 'pulumi.set', 'pulumi.set', (['self', '"""permissions"""', 'value'], {}), "(self, 'permissions', value)\n", (611359, 611387), False, 'import pulumi\n'), ((611761, 611813), 'pulumi.set', 'pulumi.set', (['__self__', '"""database_name"""', 'database_name'], {}), "(__self__, 'database_name', database_name)\n", (611771, 611813), False, 'import pulumi\n'), ((612108, 612141), 'pulumi.get', 'pulumi.get', (['self', '"""database_name"""'], {}), "(self, 'database_name')\n", (612118, 612141), False, 'import pulumi\n'), ((612232, 612272), 'pulumi.set', 'pulumi.set', (['self', '"""database_name"""', 'value'], {}), "(self, 'database_name', value)\n", (612242, 612272), False, 'import pulumi\n'), ((612398, 612423), 'pulumi.get', 'pulumi.get', (['self', '"""roles"""'], {}), "(self, 'roles')\n", (612408, 612423), False, 'import pulumi\n'), ((612532, 612564), 'pulumi.set', 'pulumi.set', (['self', '"""roles"""', 'value'], {}), "(self, 'roles', value)\n", (612542, 612564), False, 'import pulumi\n'), ((613510, 613539), 'pulumi.get', 'pulumi.get', (['self', '"""data_lens"""'], {}), "(self, 'data_lens')\n", (613520, 613539), False, 'import pulumi\n'), ((613633, 613669), 'pulumi.set', 'pulumi.set', (['self', '"""data_lens"""', 'value'], {}), "(self, 'data_lens', value)\n", (613643, 613669), False, 'import pulumi\n'), ((613950, 613977), 'pulumi.get', 'pulumi.get', (['self', '"""web_sql"""'], {}), "(self, 'web_sql')\n", (613960, 613977), False, 'import pulumi\n'), ((614067, 614101), 'pulumi.set', 'pulumi.set', (['self', '"""web_sql"""', 'value'], {}), "(self, 'web_sql', value)\n", (614077, 614101), False, 'import pulumi\n'), ((614851, 614876), 'pulumi.get', 'pulumi.get', (['self', '"""hours"""'], {}), "(self, 'hours')\n", (614861, 614876), False, 'import pulumi\n'), ((614961, 614993), 'pulumi.set', 'pulumi.set', (['self', '"""hours"""', 'value'], {}), "(self, 'hours', value)\n", (614971, 614993), False, 'import pulumi\n'), ((615173, 615200), 'pulumi.get', 'pulumi.get', (['self', '"""minutes"""'], {}), "(self, 'minutes')\n", (615183, 615200), False, 'import pulumi\n'), ((615289, 615323), 'pulumi.set', 'pulumi.set', (['self', '"""minutes"""', 'value'], {}), "(self, 'minutes', value)\n", (615299, 615323), False, 'import pulumi\n'), ((615742, 615776), 'pulumi.set', 'pulumi.set', (['__self__', '"""name"""', 'name'], {}), "(__self__, 'name', name)\n", (615752, 615776), False, 'import pulumi\n'), ((616117, 616141), 'pulumi.get', 'pulumi.get', (['self', '"""name"""'], {}), "(self, 'name')\n", (616127, 616141), False, 'import pulumi\n'), ((616214, 616245), 'pulumi.set', 'pulumi.set', (['self', '"""name"""', 'value'], {}), "(self, 'name', value)\n", (616224, 616245), False, 'import pulumi\n'), ((617936, 617970), 'pulumi.set', 'pulumi.set', (['__self__', '"""zone"""', 'zone'], {}), "(__self__, 'zone', zone)\n", (617946, 617970), False, 'import pulumi\n'), ((618759, 618783), 'pulumi.get', 'pulumi.get', (['self', '"""zone"""'], {}), "(self, 'zone')\n", (618769, 618783), False, 'import pulumi\n'), ((618856, 618887), 'pulumi.set', 'pulumi.set', (['self', '"""zone"""', 'value'], {}), "(self, 'zone', value)\n", (618866, 618887), False, 'import pulumi\n'), ((619164, 619200), 'pulumi.get', 'pulumi.get', (['self', '"""assign_public_ip"""'], {}), "(self, 'assign_public_ip')\n", (619174, 619200), False, 'import pulumi\n'), ((619308, 619351), 'pulumi.set', 'pulumi.set', (['self', '"""assign_public_ip"""', 'value'], {}), "(self, 'assign_public_ip', value)\n", (619318, 619351), False, 'import pulumi\n'), ((619529, 619553), 'pulumi.get', 'pulumi.get', (['self', '"""fqdn"""'], {}), "(self, 'fqdn')\n", (619539, 619553), False, 'import pulumi\n'), ((619636, 619667), 'pulumi.set', 'pulumi.set', (['self', '"""fqdn"""', 'value'], {}), "(self, 'fqdn', value)\n", (619646, 619667), False, 'import pulumi\n'), ((620018, 620042), 'pulumi.get', 'pulumi.get', (['self', '"""name"""'], {}), "(self, 'name')\n", (620028, 620042), False, 'import pulumi\n'), ((620125, 620156), 'pulumi.set', 'pulumi.set', (['self', '"""name"""', 'value'], {}), "(self, 'name', value)\n", (620135, 620156), False, 'import pulumi\n'), ((620420, 620458), 'pulumi.get', 'pulumi.get', (['self', '"""replication_source"""'], {}), "(self, 'replication_source')\n", (620430, 620458), False, 'import pulumi\n'), ((620569, 620614), 'pulumi.set', 'pulumi.set', (['self', '"""replication_source"""', 'value'], {}), "(self, 'replication_source', value)\n", (620579, 620614), False, 'import pulumi\n'), ((620958, 621001), 'pulumi.get', 'pulumi.get', (['self', '"""replication_source_name"""'], {}), "(self, 'replication_source_name')\n", (620968, 621001), False, 'import pulumi\n'), ((621122, 621172), 'pulumi.set', 'pulumi.set', (['self', '"""replication_source_name"""', 'value'], {}), "(self, 'replication_source_name', value)\n", (621132, 621172), False, 'import pulumi\n'), ((621447, 621476), 'pulumi.get', 'pulumi.get', (['self', '"""subnet_id"""'], {}), "(self, 'subnet_id')\n", (621457, 621476), False, 'import pulumi\n'), ((621569, 621605), 'pulumi.set', 'pulumi.set', (['self', '"""subnet_id"""', 'value'], {}), "(self, 'subnet_id', value)\n", (621579, 621605), False, 'import pulumi\n'), ((622315, 622349), 'pulumi.set', 'pulumi.set', (['__self__', '"""type"""', 'type'], {}), "(__self__, 'type', type)\n", (622325, 622349), False, 'import pulumi\n'), ((622752, 622776), 'pulumi.get', 'pulumi.get', (['self', '"""type"""'], {}), "(self, 'type')\n", (622762, 622776), False, 'import pulumi\n'), ((622849, 622880), 'pulumi.set', 'pulumi.set', (['self', '"""type"""', 'value'], {}), "(self, 'type', value)\n", (622859, 622880), False, 'import pulumi\n'), ((623111, 623134), 'pulumi.get', 'pulumi.get', (['self', '"""day"""'], {}), "(self, 'day')\n", (623121, 623134), False, 'import pulumi\n'), ((623215, 623245), 'pulumi.set', 'pulumi.set', (['self', '"""day"""', 'value'], {}), "(self, 'day', value)\n", (623225, 623245), False, 'import pulumi\n'), ((623454, 623478), 'pulumi.get', 'pulumi.get', (['self', '"""hour"""'], {}), "(self, 'hour')\n", (623464, 623478), False, 'import pulumi\n'), ((623561, 623592), 'pulumi.set', 'pulumi.set', (['self', '"""hour"""', 'value'], {}), "(self, 'hour', value)\n", (623571, 623592), False, 'import pulumi\n'), ((624056, 624100), 'pulumi.set', 'pulumi.set', (['__self__', '"""disk_size"""', 'disk_size'], {}), "(__self__, 'disk_size', disk_size)\n", (624066, 624100), False, 'import pulumi\n'), ((624109, 624159), 'pulumi.set', 'pulumi.set', (['__self__', '"""disk_type_id"""', 'disk_type_id'], {}), "(__self__, 'disk_type_id', disk_type_id)\n", (624119, 624159), False, 'import pulumi\n'), ((624168, 624230), 'pulumi.set', 'pulumi.set', (['__self__', '"""resource_preset_id"""', 'resource_preset_id'], {}), "(__self__, 'resource_preset_id', resource_preset_id)\n", (624178, 624230), False, 'import pulumi\n'), ((624438, 624467), 'pulumi.get', 'pulumi.get', (['self', '"""disk_size"""'], {}), "(self, 'disk_size')\n", (624448, 624467), False, 'import pulumi\n'), ((624550, 624586), 'pulumi.set', 'pulumi.set', (['self', '"""disk_size"""', 'value'], {}), "(self, 'disk_size', value)\n", (624560, 624586), False, 'import pulumi\n'), ((624772, 624804), 'pulumi.get', 'pulumi.get', (['self', '"""disk_type_id"""'], {}), "(self, 'disk_type_id')\n", (624782, 624804), False, 'import pulumi\n'), ((624893, 624932), 'pulumi.set', 'pulumi.set', (['self', '"""disk_type_id"""', 'value'], {}), "(self, 'disk_type_id', value)\n", (624903, 624932), False, 'import pulumi\n'), ((625062, 625100), 'pulumi.get', 'pulumi.get', (['self', '"""resource_preset_id"""'], {}), "(self, 'resource_preset_id')\n", (625072, 625100), False, 'import pulumi\n'), ((625201, 625246), 'pulumi.set', 'pulumi.set', (['self', '"""resource_preset_id"""', 'value'], {}), "(self, 'resource_preset_id', value)\n", (625211, 625246), False, 'import pulumi\n'), ((625876, 625920), 'pulumi.set', 'pulumi.set', (['__self__', '"""backup_id"""', 'backup_id'], {}), "(__self__, 'backup_id', backup_id)\n", (625886, 625920), False, 'import pulumi\n'), ((626317, 626346), 'pulumi.get', 'pulumi.get', (['self', '"""backup_id"""'], {}), "(self, 'backup_id')\n", (626327, 626346), False, 'import pulumi\n'), ((626429, 626465), 'pulumi.set', 'pulumi.set', (['self', '"""backup_id"""', 'value'], {}), "(self, 'backup_id', value)\n", (626439, 626465), False, 'import pulumi\n'), ((626744, 626768), 'pulumi.get', 'pulumi.get', (['self', '"""time"""'], {}), "(self, 'time')\n", (626754, 626768), False, 'import pulumi\n'), ((626851, 626882), 'pulumi.set', 'pulumi.set', (['self', '"""time"""', 'value'], {}), "(self, 'time', value)\n", (626861, 626882), False, 'import pulumi\n'), ((628729, 628763), 'pulumi.set', 'pulumi.set', (['__self__', '"""name"""', 'name'], {}), "(__self__, 'name', name)\n", (628739, 628763), False, 'import pulumi\n'), ((628772, 628814), 'pulumi.set', 'pulumi.set', (['__self__', '"""password"""', 'password'], {}), "(__self__, 'password', password)\n", (628782, 628814), False, 'import pulumi\n'), ((629612, 629636), 'pulumi.get', 'pulumi.get', (['self', '"""name"""'], {}), "(self, 'name')\n", (629622, 629636), False, 'import pulumi\n'), ((629709, 629740), 'pulumi.set', 'pulumi.set', (['self', '"""name"""', 'value'], {}), "(self, 'name', value)\n", (629719, 629740), False, 'import pulumi\n'), ((629893, 629921), 'pulumi.get', 'pulumi.get', (['self', '"""password"""'], {}), "(self, 'password')\n", (629903, 629921), False, 'import pulumi\n'), ((630002, 630037), 'pulumi.set', 'pulumi.set', (['self', '"""password"""', 'value'], {}), "(self, 'password', value)\n", (630012, 630037), False, 'import pulumi\n'), ((630384, 630425), 'pulumi.get', 'pulumi.get', (['self', '"""authentication_plugin"""'], {}), "(self, 'authentication_plugin')\n", (630394, 630425), False, 'import pulumi\n'), ((630542, 630590), 'pulumi.set', 'pulumi.set', (['self', '"""authentication_plugin"""', 'value'], {}), "(self, 'authentication_plugin', value)\n", (630552, 630590), False, 'import pulumi\n'), ((630928, 630965), 'pulumi.get', 'pulumi.get', (['self', '"""connection_limits"""'], {}), "(self, 'connection_limits')\n", (630938, 630965), False, 'import pulumi\n'), ((631112, 631156), 'pulumi.set', 'pulumi.set', (['self', '"""connection_limits"""', 'value'], {}), "(self, 'connection_limits', value)\n", (631122, 631156), False, 'import pulumi\n'), ((631571, 631609), 'pulumi.get', 'pulumi.get', (['self', '"""global_permissions"""'], {}), "(self, 'global_permissions')\n", (631581, 631609), False, 'import pulumi\n'), ((631744, 631789), 'pulumi.set', 'pulumi.set', (['self', '"""global_permissions"""', 'value'], {}), "(self, 'global_permissions', value)\n", (631754, 631789), False, 'import pulumi\n'), ((632060, 632091), 'pulumi.get', 'pulumi.get', (['self', '"""permissions"""'], {}), "(self, 'permissions')\n", (632070, 632091), False, 'import pulumi\n'), ((632244, 632282), 'pulumi.set', 'pulumi.set', (['self', '"""permissions"""', 'value'], {}), "(self, 'permissions', value)\n", (632254, 632282), False, 'import pulumi\n'), ((633754, 633798), 'pulumi.get', 'pulumi.get', (['self', '"""max_connections_per_hour"""'], {}), "(self, 'max_connections_per_hour')\n", (633764, 633798), False, 'import pulumi\n'), ((633921, 633972), 'pulumi.set', 'pulumi.set', (['self', '"""max_connections_per_hour"""', 'value'], {}), "(self, 'max_connections_per_hour', value)\n", (633931, 633972), False, 'import pulumi\n'), ((634175, 634217), 'pulumi.get', 'pulumi.get', (['self', '"""max_questions_per_hour"""'], {}), "(self, 'max_questions_per_hour')\n", (634185, 634217), False, 'import pulumi\n'), ((634336, 634385), 'pulumi.set', 'pulumi.set', (['self', '"""max_questions_per_hour"""', 'value'], {}), "(self, 'max_questions_per_hour', value)\n", (634346, 634385), False, 'import pulumi\n'), ((634582, 634622), 'pulumi.get', 'pulumi.get', (['self', '"""max_updates_per_hour"""'], {}), "(self, 'max_updates_per_hour')\n", (634592, 634622), False, 'import pulumi\n'), ((634737, 634784), 'pulumi.set', 'pulumi.set', (['self', '"""max_updates_per_hour"""', 'value'], {}), "(self, 'max_updates_per_hour', value)\n", (634747, 634784), False, 'import pulumi\n'), ((634982, 635022), 'pulumi.get', 'pulumi.get', (['self', '"""max_user_connections"""'], {}), "(self, 'max_user_connections')\n", (634992, 635022), False, 'import pulumi\n'), ((635137, 635184), 'pulumi.set', 'pulumi.set', (['self', '"""max_user_connections"""', 'value'], {}), "(self, 'max_user_connections', value)\n", (635147, 635184), False, 'import pulumi\n'), ((635903, 635955), 'pulumi.set', 'pulumi.set', (['__self__', '"""database_name"""', 'database_name'], {}), "(__self__, 'database_name', database_name)\n", (635913, 635955), False, 'import pulumi\n'), ((636250, 636283), 'pulumi.get', 'pulumi.get', (['self', '"""database_name"""'], {}), "(self, 'database_name')\n", (636260, 636283), False, 'import pulumi\n'), ((636374, 636414), 'pulumi.set', 'pulumi.set', (['self', '"""database_name"""', 'value'], {}), "(self, 'database_name', value)\n", (636384, 636414), False, 'import pulumi\n'), ((636841, 636866), 'pulumi.get', 'pulumi.get', (['self', '"""roles"""'], {}), "(self, 'roles')\n", (636851, 636866), False, 'import pulumi\n'), ((636975, 637007), 'pulumi.set', 'pulumi.set', (['self', '"""roles"""', 'value'], {}), "(self, 'roles', value)\n", (636985, 637007), False, 'import pulumi\n'), ((638595, 638637), 'pulumi.set', 'pulumi.set', (['__self__', '"""password"""', 'password'], {}), "(__self__, 'password', password)\n", (638605, 638637), False, 'import pulumi\n'), ((638646, 638686), 'pulumi.set', 'pulumi.set', (['__self__', '"""version"""', 'version'], {}), "(__self__, 'version', version)\n", (638656, 638686), False, 'import pulumi\n'), ((639505, 639533), 'pulumi.get', 'pulumi.get', (['self', '"""password"""'], {}), "(self, 'password')\n", (639515, 639533), False, 'import pulumi\n'), ((639614, 639649), 'pulumi.set', 'pulumi.set', (['self', '"""password"""', 'value'], {}), "(self, 'password', value)\n", (639624, 639649), False, 'import pulumi\n'), ((639811, 639838), 'pulumi.get', 'pulumi.get', (['self', '"""version"""'], {}), "(self, 'version')\n", (639821, 639838), False, 'import pulumi\n'), ((639917, 639951), 'pulumi.set', 'pulumi.set', (['self', '"""version"""', 'value'], {}), "(self, 'version', value)\n", (639927, 639951), False, 'import pulumi\n'), ((640151, 640180), 'pulumi.get', 'pulumi.get', (['self', '"""databases"""'], {}), "(self, 'databases')\n", (640161, 640180), False, 'import pulumi\n'), ((640273, 640309), 'pulumi.set', 'pulumi.set', (['self', '"""databases"""', 'value'], {}), "(self, 'databases', value)\n", (640283, 640309), False, 'import pulumi\n'), ((640710, 640746), 'pulumi.get', 'pulumi.get', (['self', '"""maxmemory_policy"""'], {}), "(self, 'maxmemory_policy')\n", (640720, 640746), False, 'import pulumi\n'), ((640853, 640896), 'pulumi.set', 'pulumi.set', (['self', '"""maxmemory_policy"""', 'value'], {}), "(self, 'maxmemory_policy', value)\n", (640863, 640896), False, 'import pulumi\n'), ((641141, 641183), 'pulumi.get', 'pulumi.get', (['self', '"""notify_keyspace_events"""'], {}), "(self, 'notify_keyspace_events')\n", (641151, 641183), False, 'import pulumi\n'), ((641302, 641351), 'pulumi.set', 'pulumi.set', (['self', '"""notify_keyspace_events"""', 'value'], {}), "(self, 'notify_keyspace_events', value)\n", (641312, 641351), False, 'import pulumi\n'), ((641584, 641627), 'pulumi.get', 'pulumi.get', (['self', '"""slowlog_log_slower_than"""'], {}), "(self, 'slowlog_log_slower_than')\n", (641594, 641627), False, 'import pulumi\n'), ((641748, 641798), 'pulumi.set', 'pulumi.set', (['self', '"""slowlog_log_slower_than"""', 'value'], {}), "(self, 'slowlog_log_slower_than', value)\n", (641758, 641798), False, 'import pulumi\n'), ((641989, 642024), 'pulumi.get', 'pulumi.get', (['self', '"""slowlog_max_len"""'], {}), "(self, 'slowlog_max_len')\n", (641999, 642024), False, 'import pulumi\n'), ((642129, 642171), 'pulumi.set', 'pulumi.set', (['self', '"""slowlog_max_len"""', 'value'], {}), "(self, 'slowlog_max_len', value)\n", (642139, 642171), False, 'import pulumi\n'), ((642366, 642393), 'pulumi.get', 'pulumi.get', (['self', '"""timeout"""'], {}), "(self, 'timeout')\n", (642376, 642393), False, 'import pulumi\n'), ((642482, 642516), 'pulumi.set', 'pulumi.set', (['self', '"""timeout"""', 'value'], {}), "(self, 'timeout', value)\n", (642492, 642516), False, 'import pulumi\n'), ((643447, 643481), 'pulumi.set', 'pulumi.set', (['__self__', '"""zone"""', 'zone'], {}), "(__self__, 'zone', zone)\n", (643457, 643481), False, 'import pulumi\n'), ((644047, 644071), 'pulumi.get', 'pulumi.get', (['self', '"""zone"""'], {}), "(self, 'zone')\n", (644057, 644071), False, 'import pulumi\n'), ((644144, 644175), 'pulumi.set', 'pulumi.set', (['self', '"""zone"""', 'value'], {}), "(self, 'zone', value)\n", (644154, 644175), False, 'import pulumi\n'), ((644353, 644377), 'pulumi.get', 'pulumi.get', (['self', '"""fqdn"""'], {}), "(self, 'fqdn')\n", (644363, 644377), False, 'import pulumi\n'), ((644460, 644491), 'pulumi.set', 'pulumi.set', (['self', '"""fqdn"""', 'value'], {}), "(self, 'fqdn', value)\n", (644470, 644491), False, 'import pulumi\n'), ((644697, 644727), 'pulumi.get', 'pulumi.get', (['self', '"""shard_name"""'], {}), "(self, 'shard_name')\n", (644707, 644727), False, 'import pulumi\n'), ((644822, 644859), 'pulumi.set', 'pulumi.set', (['self', '"""shard_name"""', 'value'], {}), "(self, 'shard_name', value)\n", (644832, 644859), False, 'import pulumi\n'), ((645142, 645171), 'pulumi.get', 'pulumi.get', (['self', '"""subnet_id"""'], {}), "(self, 'subnet_id')\n", (645152, 645171), False, 'import pulumi\n'), ((645264, 645300), 'pulumi.set', 'pulumi.set', (['self', '"""subnet_id"""', 'value'], {}), "(self, 'subnet_id', value)\n", (645274, 645300), False, 'import pulumi\n'), ((646047, 646081), 'pulumi.set', 'pulumi.set', (['__self__', '"""type"""', 'type'], {}), "(__self__, 'type', type)\n", (646057, 646081), False, 'import pulumi\n'), ((646484, 646508), 'pulumi.get', 'pulumi.get', (['self', '"""type"""'], {}), "(self, 'type')\n", (646494, 646508), False, 'import pulumi\n'), ((646581, 646612), 'pulumi.set', 'pulumi.set', (['self', '"""type"""', 'value'], {}), "(self, 'type', value)\n", (646591, 646612), False, 'import pulumi\n'), ((646871, 646894), 'pulumi.get', 'pulumi.get', (['self', '"""day"""'], {}), "(self, 'day')\n", (646881, 646894), False, 'import pulumi\n'), ((646975, 647005), 'pulumi.set', 'pulumi.set', (['self', '"""day"""', 'value'], {}), "(self, 'day', value)\n", (646985, 647005), False, 'import pulumi\n'), ((647223, 647247), 'pulumi.get', 'pulumi.get', (['self', '"""hour"""'], {}), "(self, 'hour')\n", (647233, 647247), False, 'import pulumi\n'), ((647330, 647361), 'pulumi.set', 'pulumi.set', (['self', '"""hour"""', 'value'], {}), "(self, 'hour', value)\n", (647340, 647361), False, 'import pulumi\n'), ((647877, 647921), 'pulumi.set', 'pulumi.set', (['__self__', '"""disk_size"""', 'disk_size'], {}), "(__self__, 'disk_size', disk_size)\n", (647887, 647921), False, 'import pulumi\n'), ((647930, 647992), 'pulumi.set', 'pulumi.set', (['__self__', '"""resource_preset_id"""', 'resource_preset_id'], {}), "(__self__, 'resource_preset_id', resource_preset_id)\n", (647940, 647992), False, 'import pulumi\n'), ((648294, 648323), 'pulumi.get', 'pulumi.get', (['self', '"""disk_size"""'], {}), "(self, 'disk_size')\n", (648304, 648323), False, 'import pulumi\n'), ((648406, 648442), 'pulumi.set', 'pulumi.set', (['self', '"""disk_size"""', 'value'], {}), "(self, 'disk_size', value)\n", (648416, 648442), False, 'import pulumi\n'), ((648572, 648610), 'pulumi.get', 'pulumi.get', (['self', '"""resource_preset_id"""'], {}), "(self, 'resource_preset_id')\n", (648582, 648610), False, 'import pulumi\n'), ((648711, 648756), 'pulumi.set', 'pulumi.set', (['self', '"""resource_preset_id"""', 'value'], {}), "(self, 'resource_preset_id', value)\n", (648721, 648756), False, 'import pulumi\n'), ((648993, 649025), 'pulumi.get', 'pulumi.get', (['self', '"""disk_type_id"""'], {}), "(self, 'disk_type_id')\n", (649003, 649025), False, 'import pulumi\n'), ((649124, 649163), 'pulumi.set', 'pulumi.set', (['self', '"""disk_type_id"""', 'value'], {}), "(self, 'disk_type_id', value)\n", (649134, 649163), False, 'import pulumi\n'), ((649917, 649942), 'pulumi.get', 'pulumi.get', (['self', '"""hours"""'], {}), "(self, 'hours')\n", (649927, 649942), False, 'import pulumi\n'), ((650027, 650059), 'pulumi.set', 'pulumi.set', (['self', '"""hours"""', 'value'], {}), "(self, 'hours', value)\n", (650037, 650059), False, 'import pulumi\n'), ((650239, 650266), 'pulumi.get', 'pulumi.get', (['self', '"""minutes"""'], {}), "(self, 'minutes')\n", (650249, 650266), False, 'import pulumi\n'), ((650355, 650389), 'pulumi.set', 'pulumi.set', (['self', '"""minutes"""', 'value'], {}), "(self, 'minutes', value)\n", (650365, 650389), False, 'import pulumi\n'), ((650620, 650654), 'pulumi.set', 'pulumi.set', (['__self__', '"""name"""', 'name'], {}), "(__self__, 'name', name)\n", (650630, 650654), False, 'import pulumi\n'), ((650803, 650827), 'pulumi.get', 'pulumi.get', (['self', '"""name"""'], {}), "(self, 'name')\n", (650813, 650827), False, 'import pulumi\n'), ((650900, 650931), 'pulumi.set', 'pulumi.set', (['self', '"""name"""', 'value'], {}), "(self, 'name', value)\n", (650910, 650931), False, 'import pulumi\n'), ((651833, 651867), 'pulumi.set', 'pulumi.set', (['__self__', '"""zone"""', 'zone'], {}), "(__self__, 'zone', zone)\n", (651843, 651867), False, 'import pulumi\n'), ((652333, 652357), 'pulumi.get', 'pulumi.get', (['self', '"""zone"""'], {}), "(self, 'zone')\n", (652343, 652357), False, 'import pulumi\n'), ((652430, 652461), 'pulumi.set', 'pulumi.set', (['self', '"""zone"""', 'value'], {}), "(self, 'zone', value)\n", (652440, 652461), False, 'import pulumi\n'), ((652772, 652808), 'pulumi.get', 'pulumi.get', (['self', '"""assign_public_ip"""'], {}), "(self, 'assign_public_ip')\n", (652782, 652808), False, 'import pulumi\n'), ((652916, 652959), 'pulumi.set', 'pulumi.set', (['self', '"""assign_public_ip"""', 'value'], {}), "(self, 'assign_public_ip', value)\n", (652926, 652959), False, 'import pulumi\n'), ((653137, 653161), 'pulumi.get', 'pulumi.get', (['self', '"""fqdn"""'], {}), "(self, 'fqdn')\n", (653147, 653161), False, 'import pulumi\n'), ((653244, 653275), 'pulumi.set', 'pulumi.set', (['self', '"""fqdn"""', 'value'], {}), "(self, 'fqdn', value)\n", (653254, 653275), False, 'import pulumi\n'), ((653550, 653579), 'pulumi.get', 'pulumi.get', (['self', '"""subnet_id"""'], {}), "(self, 'subnet_id')\n", (653560, 653579), False, 'import pulumi\n'), ((653672, 653708), 'pulumi.set', 'pulumi.set', (['self', '"""subnet_id"""', 'value'], {}), "(self, 'subnet_id', value)\n", (653682, 653708), False, 'import pulumi\n'), ((654184, 654228), 'pulumi.set', 'pulumi.set', (['__self__', '"""disk_size"""', 'disk_size'], {}), "(__self__, 'disk_size', disk_size)\n", (654194, 654228), False, 'import pulumi\n'), ((654237, 654287), 'pulumi.set', 'pulumi.set', (['__self__', '"""disk_type_id"""', 'disk_type_id'], {}), "(__self__, 'disk_type_id', disk_type_id)\n", (654247, 654287), False, 'import pulumi\n'), ((654296, 654358), 'pulumi.set', 'pulumi.set', (['__self__', '"""resource_preset_id"""', 'resource_preset_id'], {}), "(__self__, 'resource_preset_id', resource_preset_id)\n", (654306, 654358), False, 'import pulumi\n'), ((654570, 654599), 'pulumi.get', 'pulumi.get', (['self', '"""disk_size"""'], {}), "(self, 'disk_size')\n", (654580, 654599), False, 'import pulumi\n'), ((654682, 654718), 'pulumi.set', 'pulumi.set', (['self', '"""disk_size"""', 'value'], {}), "(self, 'disk_size', value)\n", (654692, 654718), False, 'import pulumi\n'), ((654908, 654940), 'pulumi.get', 'pulumi.get', (['self', '"""disk_type_id"""'], {}), "(self, 'disk_type_id')\n", (654918, 654940), False, 'import pulumi\n'), ((655029, 655068), 'pulumi.set', 'pulumi.set', (['self', '"""disk_type_id"""', 'value'], {}), "(self, 'disk_type_id', value)\n", (655039, 655068), False, 'import pulumi\n'), ((655198, 655236), 'pulumi.get', 'pulumi.get', (['self', '"""resource_preset_id"""'], {}), "(self, 'resource_preset_id')\n", (655208, 655236), False, 'import pulumi\n'), ((655337, 655382), 'pulumi.set', 'pulumi.set', (['self', '"""resource_preset_id"""', 'value'], {}), "(self, 'resource_preset_id', value)\n", (655347, 655382), False, 'import pulumi\n'), ((656017, 656051), 'pulumi.set', 'pulumi.set', (['__self__', '"""name"""', 'name'], {}), "(__self__, 'name', name)\n", (656027, 656051), False, 'import pulumi\n'), ((656060, 656102), 'pulumi.set', 'pulumi.set', (['__self__', '"""password"""', 'password'], {}), "(__self__, 'password', password)\n", (656070, 656102), False, 'import pulumi\n'), ((656348, 656372), 'pulumi.get', 'pulumi.get', (['self', '"""name"""'], {}), "(self, 'name')\n", (656358, 656372), False, 'import pulumi\n'), ((656445, 656476), 'pulumi.set', 'pulumi.set', (['self', '"""name"""', 'value'], {}), "(self, 'name', value)\n", (656455, 656476), False, 'import pulumi\n'), ((656629, 656657), 'pulumi.get', 'pulumi.get', (['self', '"""password"""'], {}), "(self, 'password')\n", (656639, 656657), False, 'import pulumi\n'), ((656738, 656773), 'pulumi.set', 'pulumi.set', (['self', '"""password"""', 'value'], {}), "(self, 'password', value)\n", (656748, 656773), False, 'import pulumi\n'), ((657048, 657079), 'pulumi.get', 'pulumi.get', (['self', '"""permissions"""'], {}), "(self, 'permissions')\n", (657058, 657079), False, 'import pulumi\n'), ((657236, 657274), 'pulumi.set', 'pulumi.set', (['self', '"""permissions"""', 'value'], {}), "(self, 'permissions', value)\n", (657246, 657274), False, 'import pulumi\n'), ((657914, 657966), 'pulumi.set', 'pulumi.set', (['__self__', '"""database_name"""', 'database_name'], {}), "(__self__, 'database_name', database_name)\n", (657924, 657966), False, 'import pulumi\n'), ((658261, 658294), 'pulumi.get', 'pulumi.get', (['self', '"""database_name"""'], {}), "(self, 'database_name')\n", (658271, 658294), False, 'import pulumi\n'), ((658385, 658425), 'pulumi.set', 'pulumi.set', (['self', '"""database_name"""', 'value'], {}), "(self, 'database_name', value)\n", (658395, 658425), False, 'import pulumi\n'), ((658776, 658801), 'pulumi.get', 'pulumi.get', (['self', '"""roles"""'], {}), "(self, 'roles')\n", (658786, 658801), False, 'import pulumi\n'), ((658910, 658942), 'pulumi.set', 'pulumi.set', (['self', '"""roles"""', 'value'], {}), "(self, 'roles', value)\n", (658920, 658942), False, 'import pulumi\n'), ((660104, 660160), 'pulumi.set', 'pulumi.set', (['__self__', '"""allowed_methods"""', 'allowed_methods'], {}), "(__self__, 'allowed_methods', allowed_methods)\n", (660114, 660160), False, 'import pulumi\n'), ((660169, 660225), 'pulumi.set', 'pulumi.set', (['__self__', '"""allowed_origins"""', 'allowed_origins'], {}), "(__self__, 'allowed_origins', allowed_origins)\n", (660179, 660225), False, 'import pulumi\n'), ((660816, 660851), 'pulumi.get', 'pulumi.get', (['self', '"""allowed_methods"""'], {}), "(self, 'allowed_methods')\n", (660826, 660851), False, 'import pulumi\n'), ((660970, 661012), 'pulumi.set', 'pulumi.set', (['self', '"""allowed_methods"""', 'value'], {}), "(self, 'allowed_methods', value)\n", (660980, 661012), False, 'import pulumi\n'), ((661230, 661265), 'pulumi.get', 'pulumi.get', (['self', '"""allowed_origins"""'], {}), "(self, 'allowed_origins')\n", (661240, 661265), False, 'import pulumi\n'), ((661384, 661426), 'pulumi.set', 'pulumi.set', (['self', '"""allowed_origins"""', 'value'], {}), "(self, 'allowed_origins', value)\n", (661394, 661426), False, 'import pulumi\n'), ((661654, 661689), 'pulumi.get', 'pulumi.get', (['self', '"""allowed_headers"""'], {}), "(self, 'allowed_headers')\n", (661664, 661689), False, 'import pulumi\n'), ((661818, 661860), 'pulumi.set', 'pulumi.set', (['self', '"""allowed_headers"""', 'value'], {}), "(self, 'allowed_headers', value)\n", (661828, 661860), False, 'import pulumi\n'), ((662090, 662124), 'pulumi.get', 'pulumi.get', (['self', '"""expose_headers"""'], {}), "(self, 'expose_headers')\n", (662100, 662124), False, 'import pulumi\n'), ((662251, 662292), 'pulumi.set', 'pulumi.set', (['self', '"""expose_headers"""', 'value'], {}), "(self, 'expose_headers', value)\n", (662261, 662292), False, 'import pulumi\n'), ((662545, 662580), 'pulumi.get', 'pulumi.get', (['self', '"""max_age_seconds"""'], {}), "(self, 'max_age_seconds')\n", (662555, 662580), False, 'import pulumi\n'), ((662685, 662727), 'pulumi.set', 'pulumi.set', (['self', '"""max_age_seconds"""', 'value'], {}), "(self, 'max_age_seconds', value)\n", (662695, 662727), False, 'import pulumi\n'), ((663197, 663245), 'pulumi.set', 'pulumi.set', (['__self__', '"""permissions"""', 'permissions'], {}), "(__self__, 'permissions', permissions)\n", (663207, 663245), False, 'import pulumi\n'), ((663254, 663288), 'pulumi.set', 'pulumi.set', (['__self__', '"""type"""', 'type'], {}), "(__self__, 'type', type)\n", (663264, 663288), False, 'import pulumi\n'), ((663553, 663584), 'pulumi.get', 'pulumi.get', (['self', '"""permissions"""'], {}), "(self, 'permissions')\n", (663563, 663584), False, 'import pulumi\n'), ((663695, 663733), 'pulumi.set', 'pulumi.set', (['self', '"""permissions"""', 'value'], {}), "(self, 'permissions', value)\n", (663705, 663733), False, 'import pulumi\n'), ((663824, 663848), 'pulumi.get', 'pulumi.get', (['self', '"""type"""'], {}), "(self, 'type')\n", (663834, 663848), False, 'import pulumi\n'), ((663921, 663952), 'pulumi.set', 'pulumi.set', (['self', '"""type"""', 'value'], {}), "(self, 'type', value)\n", (663931, 663952), False, 'import pulumi\n'), ((664171, 664193), 'pulumi.get', 'pulumi.get', (['self', '"""id"""'], {}), "(self, 'id')\n", (664181, 664193), False, 'import pulumi\n'), ((664272, 664301), 'pulumi.set', 'pulumi.set', (['self', '"""id"""', 'value'], {}), "(self, 'id', value)\n", (664282, 664301), False, 'import pulumi\n'), ((664401, 664424), 'pulumi.get', 'pulumi.get', (['self', '"""uri"""'], {}), "(self, 'uri')\n", (664411, 664424), False, 'import pulumi\n'), ((664505, 664535), 'pulumi.set', 'pulumi.set', (['self', '"""uri"""', 'value'], {}), "(self, 'uri', value)\n", (664515, 664535), False, 'import pulumi\n'), ((666685, 666725), 'pulumi.set', 'pulumi.set', (['__self__', '"""enabled"""', 'enabled'], {}), "(__self__, 'enabled', enabled)\n", (666695, 666725), False, 'import pulumi\n'), ((667711, 667738), 'pulumi.get', 'pulumi.get', (['self', '"""enabled"""'], {}), "(self, 'enabled')\n", (667721, 667738), False, 'import pulumi\n'), ((667818, 667852), 'pulumi.set', 'pulumi.set', (['self', '"""enabled"""', 'value'], {}), "(self, 'enabled', value)\n", (667828, 667852), False, 'import pulumi\n'), ((668172, 668230), 'pulumi.get', 'pulumi.get', (['self', '"""abort_incomplete_multipart_upload_days"""'], {}), "(self, 'abort_incomplete_multipart_upload_days')\n", (668182, 668230), False, 'import pulumi\n'), ((668381, 668446), 'pulumi.set', 'pulumi.set', (['self', '"""abort_incomplete_multipart_upload_days"""', 'value'], {}), "(self, 'abort_incomplete_multipart_upload_days', value)\n", (668391, 668446), False, 'import pulumi\n'), ((668686, 668716), 'pulumi.get', 'pulumi.get', (['self', '"""expiration"""'], {}), "(self, 'expiration')\n", (668696, 668716), False, 'import pulumi\n'), ((668850, 668887), 'pulumi.set', 'pulumi.set', (['self', '"""expiration"""', 'value'], {}), "(self, 'expiration', value)\n", (668860, 668887), False, 'import pulumi\n'), ((669106, 669128), 'pulumi.get', 'pulumi.get', (['self', '"""id"""'], {}), "(self, 'id')\n", (669116, 669128), False, 'import pulumi\n'), ((669207, 669236), 'pulumi.set', 'pulumi.set', (['self', '"""id"""', 'value'], {}), "(self, 'id', value)\n", (669217, 669236), False, 'import pulumi\n'), ((669555, 669604), 'pulumi.get', 'pulumi.get', (['self', '"""noncurrent_version_expiration"""'], {}), "(self, 'noncurrent_version_expiration')\n", (669565, 669604), False, 'import pulumi\n'), ((669793, 669849), 'pulumi.set', 'pulumi.set', (['self', '"""noncurrent_version_expiration"""', 'value'], {}), "(self, 'noncurrent_version_expiration', value)\n", (669803, 669849), False, 'import pulumi\n'), ((670199, 670249), 'pulumi.get', 'pulumi.get', (['self', '"""noncurrent_version_transitions"""'], {}), "(self, 'noncurrent_version_transitions')\n", (670209, 670249), False, 'import pulumi\n'), ((670464, 670521), 'pulumi.set', 'pulumi.set', (['self', '"""noncurrent_version_transitions"""', 'value'], {}), "(self, 'noncurrent_version_transitions', value)\n", (670474, 670521), False, 'import pulumi\n'), ((670733, 670759), 'pulumi.get', 'pulumi.get', (['self', '"""prefix"""'], {}), "(self, 'prefix')\n", (670743, 670759), False, 'import pulumi\n'), ((670846, 670879), 'pulumi.set', 'pulumi.set', (['self', '"""prefix"""', 'value'], {}), "(self, 'prefix', value)\n", (670856, 670879), False, 'import pulumi\n'), ((671149, 671180), 'pulumi.get', 'pulumi.get', (['self', '"""transitions"""'], {}), "(self, 'transitions')\n", (671159, 671180), False, 'import pulumi\n'), ((671340, 671378), 'pulumi.set', 'pulumi.set', (['self', '"""transitions"""', 'value'], {}), "(self, 'transitions', value)\n", (671350, 671378), False, 'import pulumi\n'), ((672733, 672757), 'pulumi.get', 'pulumi.get', (['self', '"""date"""'], {}), "(self, 'date')\n", (672743, 672757), False, 'import pulumi\n'), ((672840, 672871), 'pulumi.set', 'pulumi.set', (['self', '"""date"""', 'value'], {}), "(self, 'date', value)\n", (672850, 672871), False, 'import pulumi\n'), ((673099, 673123), 'pulumi.get', 'pulumi.get', (['self', '"""days"""'], {}), "(self, 'days')\n", (673109, 673123), False, 'import pulumi\n'), ((673206, 673237), 'pulumi.set', 'pulumi.set', (['self', '"""days"""', 'value'], {}), "(self, 'days', value)\n", (673216, 673237), False, 'import pulumi\n'), ((673626, 673674), 'pulumi.get', 'pulumi.get', (['self', '"""expired_object_delete_marker"""'], {}), "(self, 'expired_object_delete_marker')\n", (673636, 673674), False, 'import pulumi\n'), ((673806, 673861), 'pulumi.set', 'pulumi.set', (['self', '"""expired_object_delete_marker"""', 'value'], {}), "(self, 'expired_object_delete_marker', value)\n", (673816, 673861), False, 'import pulumi\n'), ((674437, 674461), 'pulumi.get', 'pulumi.get', (['self', '"""days"""'], {}), "(self, 'days')\n", (674447, 674461), False, 'import pulumi\n'), ((674544, 674575), 'pulumi.set', 'pulumi.set', (['self', '"""days"""', 'value'], {}), "(self, 'days', value)\n", (674554, 674575), False, 'import pulumi\n'), ((675119, 675171), 'pulumi.set', 'pulumi.set', (['__self__', '"""storage_class"""', 'storage_class'], {}), "(__self__, 'storage_class', storage_class)\n", (675129, 675171), False, 'import pulumi\n'), ((675529, 675562), 'pulumi.get', 'pulumi.get', (['self', '"""storage_class"""'], {}), "(self, 'storage_class')\n", (675539, 675562), False, 'import pulumi\n'), ((675653, 675693), 'pulumi.set', 'pulumi.set', (['self', '"""storage_class"""', 'value'], {}), "(self, 'storage_class', value)\n", (675663, 675693), False, 'import pulumi\n'), ((675894, 675918), 'pulumi.get', 'pulumi.get', (['self', '"""days"""'], {}), "(self, 'days')\n", (675904, 675918), False, 'import pulumi\n'), ((676001, 676032), 'pulumi.set', 'pulumi.set', (['self', '"""days"""', 'value'], {}), "(self, 'days', value)\n", (676011, 676032), False, 'import pulumi\n'), ((676745, 676797), 'pulumi.set', 'pulumi.set', (['__self__', '"""storage_class"""', 'storage_class'], {}), "(__self__, 'storage_class', storage_class)\n", (676755, 676797), False, 'import pulumi\n'), ((677211, 677244), 'pulumi.get', 'pulumi.get', (['self', '"""storage_class"""'], {}), "(self, 'storage_class')\n", (677221, 677244), False, 'import pulumi\n'), ((677335, 677375), 'pulumi.set', 'pulumi.set', (['self', '"""storage_class"""', 'value'], {}), "(self, 'storage_class', value)\n", (677345, 677375), False, 'import pulumi\n'), ((677589, 677613), 'pulumi.get', 'pulumi.get', (['self', '"""date"""'], {}), "(self, 'date')\n", (677599, 677613), False, 'import pulumi\n'), ((677696, 677727), 'pulumi.set', 'pulumi.set', (['self', '"""date"""', 'value'], {}), "(self, 'date', value)\n", (677706, 677727), False, 'import pulumi\n'), ((677955, 677979), 'pulumi.get', 'pulumi.get', (['self', '"""days"""'], {}), "(self, 'days')\n", (677965, 677979), False, 'import pulumi\n'), ((678062, 678093), 'pulumi.set', 'pulumi.set', (['self', '"""days"""', 'value'], {}), "(self, 'days', value)\n", (678072, 678093), False, 'import pulumi\n'), ((678524, 678576), 'pulumi.set', 'pulumi.set', (['__self__', '"""target_bucket"""', 'target_bucket'], {}), "(__self__, 'target_bucket', target_bucket)\n", (678534, 678576), False, 'import pulumi\n'), ((678890, 678923), 'pulumi.get', 'pulumi.get', (['self', '"""target_bucket"""'], {}), "(self, 'target_bucket')\n", (678900, 678923), False, 'import pulumi\n'), ((679014, 679054), 'pulumi.set', 'pulumi.set', (['self', '"""target_bucket"""', 'value'], {}), "(self, 'target_bucket', value)\n", (679024, 679054), False, 'import pulumi\n'), ((679258, 679291), 'pulumi.get', 'pulumi.get', (['self', '"""target_prefix"""'], {}), "(self, 'target_prefix')\n", (679268, 679291), False, 'import pulumi\n'), ((679392, 679432), 'pulumi.set', 'pulumi.set', (['self', '"""target_prefix"""', 'value'], {}), "(self, 'target_prefix', value)\n", (679402, 679432), False, 'import pulumi\n'), ((679850, 679884), 'pulumi.set', 'pulumi.set', (['__self__', '"""rule"""', 'rule'], {}), "(__self__, 'rule', rule)\n", (679860, 679884), False, 'import pulumi\n'), ((680148, 680172), 'pulumi.get', 'pulumi.get', (['self', '"""rule"""'], {}), "(self, 'rule')\n", (680158, 680172), False, 'import pulumi\n'), ((680298, 680329), 'pulumi.set', 'pulumi.set', (['self', '"""rule"""', 'value'], {}), "(self, 'rule', value)\n", (680308, 680329), False, 'import pulumi\n'), ((680883, 680991), 'pulumi.set', 'pulumi.set', (['__self__', '"""apply_server_side_encryption_by_default"""', 'apply_server_side_encryption_by_default'], {}), "(__self__, 'apply_server_side_encryption_by_default',\n apply_server_side_encryption_by_default)\n", (680893, 680991), False, 'import pulumi\n'), ((681357, 681416), 'pulumi.get', 'pulumi.get', (['self', '"""apply_server_side_encryption_by_default"""'], {}), "(self, 'apply_server_side_encryption_by_default')\n", (681367, 681416), False, 'import pulumi\n'), ((681646, 681712), 'pulumi.set', 'pulumi.set', (['self', '"""apply_server_side_encryption_by_default"""', 'value'], {}), "(self, 'apply_server_side_encryption_by_default', value)\n", (681656, 681712), False, 'import pulumi\n'), ((682231, 682291), 'pulumi.set', 'pulumi.set', (['__self__', '"""kms_master_key_id"""', 'kms_master_key_id'], {}), "(__self__, 'kms_master_key_id', kms_master_key_id)\n", (682241, 682291), False, 'import pulumi\n'), ((682300, 682352), 'pulumi.set', 'pulumi.set', (['__self__', '"""sse_algorithm"""', 'sse_algorithm'], {}), "(__self__, 'sse_algorithm', sse_algorithm)\n", (682310, 682352), False, 'import pulumi\n'), ((682566, 682603), 'pulumi.get', 'pulumi.get', (['self', '"""kms_master_key_id"""'], {}), "(self, 'kms_master_key_id')\n", (682576, 682603), False, 'import pulumi\n'), ((682702, 682746), 'pulumi.set', 'pulumi.set', (['self', '"""kms_master_key_id"""', 'value'], {}), "(self, 'kms_master_key_id', value)\n", (682712, 682746), False, 'import pulumi\n'), ((682976, 683009), 'pulumi.get', 'pulumi.get', (['self', '"""sse_algorithm"""'], {}), "(self, 'sse_algorithm')\n", (682986, 683009), False, 'import pulumi\n'), ((683100, 683140), 'pulumi.set', 'pulumi.set', (['self', '"""sse_algorithm"""', 'value'], {}), "(self, 'sse_algorithm', value)\n", (683110, 683140), False, 'import pulumi\n'), ((683881, 683908), 'pulumi.get', 'pulumi.get', (['self', '"""enabled"""'], {}), "(self, 'enabled')\n", (683891, 683908), False, 'import pulumi\n'), ((683998, 684032), 'pulumi.set', 'pulumi.set', (['self', '"""enabled"""', 'value'], {}), "(self, 'enabled', value)\n", (684008, 684032), False, 'import pulumi\n'), ((685907, 685941), 'pulumi.get', 'pulumi.get', (['self', '"""error_document"""'], {}), "(self, 'error_document')\n", (685917, 685941), False, 'import pulumi\n'), ((686044, 686085), 'pulumi.set', 'pulumi.set', (['self', '"""error_document"""', 'value'], {}), "(self, 'error_document', value)\n", (686054, 686085), False, 'import pulumi\n'), ((686354, 686388), 'pulumi.get', 'pulumi.get', (['self', '"""index_document"""'], {}), "(self, 'index_document')\n", (686364, 686388), False, 'import pulumi\n'), ((686491, 686532), 'pulumi.set', 'pulumi.set', (['self', '"""index_document"""', 'value'], {}), "(self, 'index_document', value)\n", (686501, 686532), False, 'import pulumi\n'), ((686957, 687001), 'pulumi.get', 'pulumi.get', (['self', '"""redirect_all_requests_to"""'], {}), "(self, 'redirect_all_requests_to')\n", (686967, 687001), False, 'import pulumi\n'), ((687124, 687175), 'pulumi.set', 'pulumi.set', (['self', '"""redirect_all_requests_to"""', 'value'], {}), "(self, 'redirect_all_requests_to', value)\n", (687134, 687175), False, 'import pulumi\n'), ((687519, 687552), 'pulumi.get', 'pulumi.get', (['self', '"""routing_rules"""'], {}), "(self, 'routing_rules')\n", (687529, 687552), False, 'import pulumi\n'), ((687653, 687693), 'pulumi.set', 'pulumi.set', (['self', '"""routing_rules"""', 'value'], {}), "(self, 'routing_rules', value)\n", (687663, 687693), False, 'import pulumi\n'), ((689018, 689045), 'pulumi.get', 'pulumi.get', (['self', '"""address"""'], {}), "(self, 'address')\n", (689028, 689045), False, 'import pulumi\n'), ((689134, 689168), 'pulumi.set', 'pulumi.set', (['self', '"""address"""', 'value'], {}), "(self, 'address', value)\n", (689144, 689168), False, 'import pulumi\n'), ((689406, 689450), 'pulumi.get', 'pulumi.get', (['self', '"""ddos_protection_provider"""'], {}), "(self, 'ddos_protection_provider')\n", (689416, 689450), False, 'import pulumi\n'), ((689573, 689624), 'pulumi.set', 'pulumi.set', (['self', '"""ddos_protection_provider"""', 'value'], {}), "(self, 'ddos_protection_provider', value)\n", (689583, 689624), False, 'import pulumi\n'), ((689841, 689885), 'pulumi.get', 'pulumi.get', (['self', '"""outgoing_smtp_capability"""'], {}), "(self, 'outgoing_smtp_capability')\n", (689851, 689885), False, 'import pulumi\n'), ((690008, 690059), 'pulumi.set', 'pulumi.set', (['self', '"""outgoing_smtp_capability"""', 'value'], {}), "(self, 'outgoing_smtp_capability', value)\n", (690018, 690059), False, 'import pulumi\n'), ((690241, 690268), 'pulumi.get', 'pulumi.get', (['self', '"""zone_id"""'], {}), "(self, 'zone_id')\n", (690251, 690268), False, 'import pulumi\n'), ((690357, 690391), 'pulumi.set', 'pulumi.set', (['self', '"""zone_id"""', 'value'], {}), "(self, 'zone_id', value)\n", (690367, 690391), False, 'import pulumi\n'), ((691544, 691586), 'pulumi.set', 'pulumi.set', (['__self__', '"""protocol"""', 'protocol'], {}), "(__self__, 'protocol', protocol)\n", (691554, 691586), False, 'import pulumi\n'), ((692624, 692652), 'pulumi.get', 'pulumi.get', (['self', '"""protocol"""'], {}), "(self, 'protocol')\n", (692634, 692652), False, 'import pulumi\n'), ((692733, 692768), 'pulumi.set', 'pulumi.set', (['self', '"""protocol"""', 'value'], {}), "(self, 'protocol', value)\n", (692743, 692768), False, 'import pulumi\n'), ((692943, 692974), 'pulumi.get', 'pulumi.get', (['self', '"""description"""'], {}), "(self, 'description')\n", (692953, 692974), False, 'import pulumi\n'), ((693071, 693109), 'pulumi.set', 'pulumi.set', (['self', '"""description"""', 'value'], {}), "(self, 'description', value)\n", (693081, 693109), False, 'import pulumi\n'), ((693232, 693261), 'pulumi.get', 'pulumi.get', (['self', '"""from_port"""'], {}), "(self, 'from_port')\n", (693242, 693261), False, 'import pulumi\n'), ((693354, 693390), 'pulumi.set', 'pulumi.set', (['self', '"""from_port"""', 'value'], {}), "(self, 'from_port', value)\n", (693364, 693390), False, 'import pulumi\n'), ((693547, 693569), 'pulumi.get', 'pulumi.get', (['self', '"""id"""'], {}), "(self, 'id')\n", (693557, 693569), False, 'import pulumi\n'), ((693648, 693677), 'pulumi.set', 'pulumi.set', (['self', '"""id"""', 'value'], {}), "(self, 'id', value)\n", (693658, 693677), False, 'import pulumi\n'), ((693881, 693907), 'pulumi.get', 'pulumi.get', (['self', '"""labels"""'], {}), "(self, 'labels')\n", (693891, 693907), False, 'import pulumi\n'), ((694022, 694055), 'pulumi.set', 'pulumi.set', (['self', '"""labels"""', 'value'], {}), "(self, 'labels', value)\n", (694032, 694055), False, 'import pulumi\n'), ((694156, 694180), 'pulumi.get', 'pulumi.get', (['self', '"""port"""'], {}), "(self, 'port')\n", (694166, 694180), False, 'import pulumi\n'), ((694263, 694294), 'pulumi.set', 'pulumi.set', (['self', '"""port"""', 'value'], {}), "(self, 'port', value)\n", (694273, 694294), False, 'import pulumi\n'), ((694433, 694470), 'pulumi.get', 'pulumi.get', (['self', '"""predefined_target"""'], {}), "(self, 'predefined_target')\n", (694443, 694470), False, 'import pulumi\n'), ((694579, 694623), 'pulumi.set', 'pulumi.set', (['self', '"""predefined_target"""', 'value'], {}), "(self, 'predefined_target', value)\n", (694589, 694623), False, 'import pulumi\n'), ((694761, 694798), 'pulumi.get', 'pulumi.get', (['self', '"""security_group_id"""'], {}), "(self, 'security_group_id')\n", (694771, 694798), False, 'import pulumi\n'), ((694907, 694951), 'pulumi.set', 'pulumi.set', (['self', '"""security_group_id"""', 'value'], {}), "(self, 'security_group_id', value)\n", (694917, 694951), False, 'import pulumi\n'), ((695070, 695097), 'pulumi.get', 'pulumi.get', (['self', '"""to_port"""'], {}), "(self, 'to_port')\n", (695080, 695097), False, 'import pulumi\n'), ((695186, 695220), 'pulumi.set', 'pulumi.set', (['self', '"""to_port"""', 'value'], {}), "(self, 'to_port', value)\n", (695196, 695220), False, 'import pulumi\n'), ((695376, 695410), 'pulumi.get', 'pulumi.get', (['self', '"""v4_cidr_blocks"""'], {}), "(self, 'v4_cidr_blocks')\n", (695386, 695410), False, 'import pulumi\n'), ((695537, 695578), 'pulumi.set', 'pulumi.set', (['self', '"""v4_cidr_blocks"""', 'value'], {}), "(self, 'v4_cidr_blocks', value)\n", (695547, 695578), False, 'import pulumi\n'), ((695734, 695768), 'pulumi.get', 'pulumi.get', (['self', '"""v6_cidr_blocks"""'], {}), "(self, 'v6_cidr_blocks')\n", (695744, 695768), False, 'import pulumi\n'), ((695895, 695936), 'pulumi.set', 'pulumi.set', (['self', '"""v6_cidr_blocks"""', 'value'], {}), "(self, 'v6_cidr_blocks', value)\n", (695905, 695936), False, 'import pulumi\n'), ((697090, 697132), 'pulumi.set', 'pulumi.set', (['__self__', '"""protocol"""', 'protocol'], {}), "(__self__, 'protocol', protocol)\n", (697100, 697132), False, 'import pulumi\n'), ((698170, 698198), 'pulumi.get', 'pulumi.get', (['self', '"""protocol"""'], {}), "(self, 'protocol')\n", (698180, 698198), False, 'import pulumi\n'), ((698279, 698314), 'pulumi.set', 'pulumi.set', (['self', '"""protocol"""', 'value'], {}), "(self, 'protocol', value)\n", (698289, 698314), False, 'import pulumi\n'), ((698489, 698520), 'pulumi.get', 'pulumi.get', (['self', '"""description"""'], {}), "(self, 'description')\n", (698499, 698520), False, 'import pulumi\n'), ((698617, 698655), 'pulumi.set', 'pulumi.set', (['self', '"""description"""', 'value'], {}), "(self, 'description', value)\n", (698627, 698655), False, 'import pulumi\n'), ((698778, 698807), 'pulumi.get', 'pulumi.get', (['self', '"""from_port"""'], {}), "(self, 'from_port')\n", (698788, 698807), False, 'import pulumi\n'), ((698900, 698936), 'pulumi.set', 'pulumi.set', (['self', '"""from_port"""', 'value'], {}), "(self, 'from_port', value)\n", (698910, 698936), False, 'import pulumi\n'), ((699093, 699115), 'pulumi.get', 'pulumi.get', (['self', '"""id"""'], {}), "(self, 'id')\n", (699103, 699115), False, 'import pulumi\n'), ((699194, 699223), 'pulumi.set', 'pulumi.set', (['self', '"""id"""', 'value'], {}), "(self, 'id', value)\n", (699204, 699223), False, 'import pulumi\n'), ((699427, 699453), 'pulumi.get', 'pulumi.get', (['self', '"""labels"""'], {}), "(self, 'labels')\n", (699437, 699453), False, 'import pulumi\n'), ((699568, 699601), 'pulumi.set', 'pulumi.set', (['self', '"""labels"""', 'value'], {}), "(self, 'labels', value)\n", (699578, 699601), False, 'import pulumi\n'), ((699702, 699726), 'pulumi.get', 'pulumi.get', (['self', '"""port"""'], {}), "(self, 'port')\n", (699712, 699726), False, 'import pulumi\n'), ((699809, 699840), 'pulumi.set', 'pulumi.set', (['self', '"""port"""', 'value'], {}), "(self, 'port', value)\n", (699819, 699840), False, 'import pulumi\n'), ((699979, 700016), 'pulumi.get', 'pulumi.get', (['self', '"""predefined_target"""'], {}), "(self, 'predefined_target')\n", (699989, 700016), False, 'import pulumi\n'), ((700125, 700169), 'pulumi.set', 'pulumi.set', (['self', '"""predefined_target"""', 'value'], {}), "(self, 'predefined_target', value)\n", (700135, 700169), False, 'import pulumi\n'), ((700307, 700344), 'pulumi.get', 'pulumi.get', (['self', '"""security_group_id"""'], {}), "(self, 'security_group_id')\n", (700317, 700344), False, 'import pulumi\n'), ((700453, 700497), 'pulumi.set', 'pulumi.set', (['self', '"""security_group_id"""', 'value'], {}), "(self, 'security_group_id', value)\n", (700463, 700497), False, 'import pulumi\n'), ((700616, 700643), 'pulumi.get', 'pulumi.get', (['self', '"""to_port"""'], {}), "(self, 'to_port')\n", (700626, 700643), False, 'import pulumi\n'), ((700732, 700766), 'pulumi.set', 'pulumi.set', (['self', '"""to_port"""', 'value'], {}), "(self, 'to_port', value)\n", (700742, 700766), False, 'import pulumi\n'), ((700922, 700956), 'pulumi.get', 'pulumi.get', (['self', '"""v4_cidr_blocks"""'], {}), "(self, 'v4_cidr_blocks')\n", (700932, 700956), False, 'import pulumi\n'), ((701083, 701124), 'pulumi.set', 'pulumi.set', (['self', '"""v4_cidr_blocks"""', 'value'], {}), "(self, 'v4_cidr_blocks', value)\n", (701093, 701124), False, 'import pulumi\n'), ((701280, 701314), 'pulumi.get', 'pulumi.get', (['self', '"""v6_cidr_blocks"""'], {}), "(self, 'v6_cidr_blocks')\n", (701290, 701314), False, 'import pulumi\n'), ((701441, 701482), 'pulumi.set', 'pulumi.set', (['self', '"""v6_cidr_blocks"""', 'value'], {}), "(self, 'v6_cidr_blocks', value)\n", (701451, 701482), False, 'import pulumi\n'), ((702332, 702370), 'pulumi.get', 'pulumi.get', (['self', '"""destination_prefix"""'], {}), "(self, 'destination_prefix')\n", (702342, 702370), False, 'import pulumi\n'), ((702481, 702526), 'pulumi.set', 'pulumi.set', (['self', '"""destination_prefix"""', 'value'], {}), "(self, 'destination_prefix', value)\n", (702491, 702526), False, 'import pulumi\n'), ((702719, 702755), 'pulumi.get', 'pulumi.get', (['self', '"""next_hop_address"""'], {}), "(self, 'next_hop_address')\n", (702729, 702755), False, 'import pulumi\n'), ((702862, 702905), 'pulumi.set', 'pulumi.set', (['self', '"""next_hop_address"""', 'value'], {}), "(self, 'next_hop_address', value)\n", (702872, 702905), False, 'import pulumi\n'), ((704041, 704083), 'pulumi.set', 'pulumi.set', (['__self__', '"""protocol"""', 'protocol'], {}), "(__self__, 'protocol', protocol)\n", (704051, 704083), False, 'import pulumi\n'), ((705121, 705149), 'pulumi.get', 'pulumi.get', (['self', '"""protocol"""'], {}), "(self, 'protocol')\n", (705131, 705149), False, 'import pulumi\n'), ((705230, 705265), 'pulumi.set', 'pulumi.set', (['self', '"""protocol"""', 'value'], {}), "(self, 'protocol', value)\n", (705240, 705265), False, 'import pulumi\n'), ((705440, 705471), 'pulumi.get', 'pulumi.get', (['self', '"""description"""'], {}), "(self, 'description')\n", (705450, 705471), False, 'import pulumi\n'), ((705568, 705606), 'pulumi.set', 'pulumi.set', (['self', '"""description"""', 'value'], {}), "(self, 'description', value)\n", (705578, 705606), False, 'import pulumi\n'), ((705729, 705758), 'pulumi.get', 'pulumi.get', (['self', '"""from_port"""'], {}), "(self, 'from_port')\n", (705739, 705758), False, 'import pulumi\n'), ((705851, 705887), 'pulumi.set', 'pulumi.set', (['self', '"""from_port"""', 'value'], {}), "(self, 'from_port', value)\n", (705861, 705887), False, 'import pulumi\n'), ((706034, 706056), 'pulumi.get', 'pulumi.get', (['self', '"""id"""'], {}), "(self, 'id')\n", (706044, 706056), False, 'import pulumi\n'), ((706135, 706164), 'pulumi.set', 'pulumi.set', (['self', '"""id"""', 'value'], {}), "(self, 'id', value)\n", (706145, 706164), False, 'import pulumi\n'), ((706368, 706394), 'pulumi.get', 'pulumi.get', (['self', '"""labels"""'], {}), "(self, 'labels')\n", (706378, 706394), False, 'import pulumi\n'), ((706509, 706542), 'pulumi.set', 'pulumi.set', (['self', '"""labels"""', 'value'], {}), "(self, 'labels', value)\n", (706519, 706542), False, 'import pulumi\n'), ((706643, 706667), 'pulumi.get', 'pulumi.get', (['self', '"""port"""'], {}), "(self, 'port')\n", (706653, 706667), False, 'import pulumi\n'), ((706750, 706781), 'pulumi.set', 'pulumi.set', (['self', '"""port"""', 'value'], {}), "(self, 'port', value)\n", (706760, 706781), False, 'import pulumi\n'), ((706920, 706957), 'pulumi.get', 'pulumi.get', (['self', '"""predefined_target"""'], {}), "(self, 'predefined_target')\n", (706930, 706957), False, 'import pulumi\n'), ((707066, 707110), 'pulumi.set', 'pulumi.set', (['self', '"""predefined_target"""', 'value'], {}), "(self, 'predefined_target', value)\n", (707076, 707110), False, 'import pulumi\n'), ((707248, 707285), 'pulumi.get', 'pulumi.get', (['self', '"""security_group_id"""'], {}), "(self, 'security_group_id')\n", (707258, 707285), False, 'import pulumi\n'), ((707394, 707438), 'pulumi.set', 'pulumi.set', (['self', '"""security_group_id"""', 'value'], {}), "(self, 'security_group_id', value)\n", (707404, 707438), False, 'import pulumi\n'), ((707557, 707584), 'pulumi.get', 'pulumi.get', (['self', '"""to_port"""'], {}), "(self, 'to_port')\n", (707567, 707584), False, 'import pulumi\n'), ((707673, 707707), 'pulumi.set', 'pulumi.set', (['self', '"""to_port"""', 'value'], {}), "(self, 'to_port', value)\n", (707683, 707707), False, 'import pulumi\n'), ((707863, 707897), 'pulumi.get', 'pulumi.get', (['self', '"""v4_cidr_blocks"""'], {}), "(self, 'v4_cidr_blocks')\n", (707873, 707897), False, 'import pulumi\n'), ((708024, 708065), 'pulumi.set', 'pulumi.set', (['self', '"""v4_cidr_blocks"""', 'value'], {}), "(self, 'v4_cidr_blocks', value)\n", (708034, 708065), False, 'import pulumi\n'), ((708221, 708255), 'pulumi.get', 'pulumi.get', (['self', '"""v6_cidr_blocks"""'], {}), "(self, 'v6_cidr_blocks')\n", (708231, 708255), False, 'import pulumi\n'), ((708382, 708423), 'pulumi.set', 'pulumi.set', (['self', '"""v6_cidr_blocks"""', 'value'], {}), "(self, 'v6_cidr_blocks', value)\n", (708392, 708423), False, 'import pulumi\n'), ((709560, 709602), 'pulumi.set', 'pulumi.set', (['__self__', '"""protocol"""', 'protocol'], {}), "(__self__, 'protocol', protocol)\n", (709570, 709602), False, 'import pulumi\n'), ((710640, 710668), 'pulumi.get', 'pulumi.get', (['self', '"""protocol"""'], {}), "(self, 'protocol')\n", (710650, 710668), False, 'import pulumi\n'), ((710749, 710784), 'pulumi.set', 'pulumi.set', (['self', '"""protocol"""', 'value'], {}), "(self, 'protocol', value)\n", (710759, 710784), False, 'import pulumi\n'), ((710959, 710990), 'pulumi.get', 'pulumi.get', (['self', '"""description"""'], {}), "(self, 'description')\n", (710969, 710990), False, 'import pulumi\n'), ((711087, 711125), 'pulumi.set', 'pulumi.set', (['self', '"""description"""', 'value'], {}), "(self, 'description', value)\n", (711097, 711125), False, 'import pulumi\n'), ((711248, 711277), 'pulumi.get', 'pulumi.get', (['self', '"""from_port"""'], {}), "(self, 'from_port')\n", (711258, 711277), False, 'import pulumi\n'), ((711370, 711406), 'pulumi.set', 'pulumi.set', (['self', '"""from_port"""', 'value'], {}), "(self, 'from_port', value)\n", (711380, 711406), False, 'import pulumi\n'), ((711553, 711575), 'pulumi.get', 'pulumi.get', (['self', '"""id"""'], {}), "(self, 'id')\n", (711563, 711575), False, 'import pulumi\n'), ((711654, 711683), 'pulumi.set', 'pulumi.set', (['self', '"""id"""', 'value'], {}), "(self, 'id', value)\n", (711664, 711683), False, 'import pulumi\n'), ((711887, 711913), 'pulumi.get', 'pulumi.get', (['self', '"""labels"""'], {}), "(self, 'labels')\n", (711897, 711913), False, 'import pulumi\n'), ((712028, 712061), 'pulumi.set', 'pulumi.set', (['self', '"""labels"""', 'value'], {}), "(self, 'labels', value)\n", (712038, 712061), False, 'import pulumi\n'), ((712162, 712186), 'pulumi.get', 'pulumi.get', (['self', '"""port"""'], {}), "(self, 'port')\n", (712172, 712186), False, 'import pulumi\n'), ((712269, 712300), 'pulumi.set', 'pulumi.set', (['self', '"""port"""', 'value'], {}), "(self, 'port', value)\n", (712279, 712300), False, 'import pulumi\n'), ((712439, 712476), 'pulumi.get', 'pulumi.get', (['self', '"""predefined_target"""'], {}), "(self, 'predefined_target')\n", (712449, 712476), False, 'import pulumi\n'), ((712585, 712629), 'pulumi.set', 'pulumi.set', (['self', '"""predefined_target"""', 'value'], {}), "(self, 'predefined_target', value)\n", (712595, 712629), False, 'import pulumi\n'), ((712767, 712804), 'pulumi.get', 'pulumi.get', (['self', '"""security_group_id"""'], {}), "(self, 'security_group_id')\n", (712777, 712804), False, 'import pulumi\n'), ((712913, 712957), 'pulumi.set', 'pulumi.set', (['self', '"""security_group_id"""', 'value'], {}), "(self, 'security_group_id', value)\n", (712923, 712957), False, 'import pulumi\n'), ((713076, 713103), 'pulumi.get', 'pulumi.get', (['self', '"""to_port"""'], {}), "(self, 'to_port')\n", (713086, 713103), False, 'import pulumi\n'), ((713192, 713226), 'pulumi.set', 'pulumi.set', (['self', '"""to_port"""', 'value'], {}), "(self, 'to_port', value)\n", (713202, 713226), False, 'import pulumi\n'), ((713382, 713416), 'pulumi.get', 'pulumi.get', (['self', '"""v4_cidr_blocks"""'], {}), "(self, 'v4_cidr_blocks')\n", (713392, 713416), False, 'import pulumi\n'), ((713543, 713584), 'pulumi.set', 'pulumi.set', (['self', '"""v4_cidr_blocks"""', 'value'], {}), "(self, 'v4_cidr_blocks', value)\n", (713553, 713584), False, 'import pulumi\n'), ((713740, 713774), 'pulumi.get', 'pulumi.get', (['self', '"""v6_cidr_blocks"""'], {}), "(self, 'v6_cidr_blocks')\n", (713750, 713774), False, 'import pulumi\n'), ((713901, 713942), 'pulumi.set', 'pulumi.set', (['self', '"""v6_cidr_blocks"""', 'value'], {}), "(self, 'v6_cidr_blocks', value)\n", (713911, 713942), False, 'import pulumi\n'), ((715056, 715087), 'pulumi.get', 'pulumi.get', (['self', '"""domain_name"""'], {}), "(self, 'domain_name')\n", (715066, 715087), False, 'import pulumi\n'), ((715184, 715222), 'pulumi.set', 'pulumi.set', (['self', '"""domain_name"""', 'value'], {}), "(self, 'domain_name', value)\n", (715194, 715222), False, 'import pulumi\n'), ((715453, 715492), 'pulumi.get', 'pulumi.get', (['self', '"""domain_name_servers"""'], {}), "(self, 'domain_name_servers')\n", (715463, 715492), False, 'import pulumi\n'), ((715629, 715675), 'pulumi.set', 'pulumi.set', (['self', '"""domain_name_servers"""', 'value'], {}), "(self, 'domain_name_servers', value)\n", (715639, 715675), False, 'import pulumi\n'), ((715883, 715914), 'pulumi.get', 'pulumi.get', (['self', '"""ntp_servers"""'], {}), "(self, 'ntp_servers')\n", (715893, 715914), False, 'import pulumi\n'), ((716035, 716073), 'pulumi.set', 'pulumi.set', (['self', '"""ntp_servers"""', 'value'], {}), "(self, 'ntp_servers', value)\n", (716045, 716073), False, 'import pulumi\n'), ((716792, 716818), 'pulumi.get', 'pulumi.get', (['self', '"""region"""'], {}), "(self, 'region')\n", (716802, 716818), False, 'import pulumi\n'), ((716942, 716975), 'pulumi.set', 'pulumi.set', (['self', '"""region"""', 'value'], {}), "(self, 'region', value)\n", (716952, 716975), False, 'import pulumi\n'), ((717226, 717256), 'pulumi.set', 'pulumi.set', (['__self__', '"""id"""', 'id'], {}), "(__self__, 'id', id)\n", (717236, 717256), False, 'import pulumi\n'), ((717420, 717442), 'pulumi.get', 'pulumi.get', (['self', '"""id"""'], {}), "(self, 'id')\n", (717430, 717442), False, 'import pulumi\n'), ((717511, 717540), 'pulumi.set', 'pulumi.set', (['self', '"""id"""', 'value'], {}), "(self, 'id', value)\n", (717521, 717540), False, 'import pulumi\n'), ((717955, 718003), 'pulumi.set', 'pulumi.set', (['__self__', '"""fixed_scale"""', 'fixed_scale'], {}), "(__self__, 'fixed_scale', fixed_scale)\n", (717965, 718003), False, 'import pulumi\n'), ((718293, 718324), 'pulumi.get', 'pulumi.get', (['self', '"""fixed_scale"""'], {}), "(self, 'fixed_scale')\n", (718303, 718324), False, 'import pulumi\n'), ((718455, 718493), 'pulumi.set', 'pulumi.set', (['self', '"""fixed_scale"""', 'value'], {}), "(self, 'fixed_scale', value)\n", (718465, 718493), False, 'import pulumi\n'), ((718765, 718799), 'pulumi.set', 'pulumi.set', (['__self__', '"""size"""', 'size'], {}), "(__self__, 'size', size)\n", (718775, 718799), False, 'import pulumi\n'), ((718975, 718999), 'pulumi.get', 'pulumi.get', (['self', '"""size"""'], {}), "(self, 'size')\n", (718985, 718999), False, 'import pulumi\n'), ((719072, 719103), 'pulumi.set', 'pulumi.set', (['self', '"""size"""', 'value'], {}), "(self, 'size', value)\n", (719082, 719103), False, 'import pulumi\n'), ((719644, 719692), 'pulumi.set', 'pulumi.set', (['__self__', '"""group_count"""', 'group_count'], {}), "(__self__, 'group_count', group_count)\n", (719654, 719692), False, 'import pulumi\n'), ((719701, 719757), 'pulumi.set', 'pulumi.set', (['__self__', '"""storage_type_id"""', 'storage_type_id'], {}), "(__self__, 'storage_type_id', storage_type_id)\n", (719711, 719757), False, 'import pulumi\n'), ((719981, 720012), 'pulumi.get', 'pulumi.get', (['self', '"""group_count"""'], {}), "(self, 'group_count')\n", (719991, 720012), False, 'import pulumi\n'), ((720099, 720137), 'pulumi.set', 'pulumi.set', (['self', '"""group_count"""', 'value'], {}), "(self, 'group_count', value)\n", (720109, 720137), False, 'import pulumi\n'), ((720424, 720459), 'pulumi.get', 'pulumi.get', (['self', '"""storage_type_id"""'], {}), "(self, 'storage_type_id')\n", (720434, 720459), False, 'import pulumi\n'), ((720554, 720596), 'pulumi.set', 'pulumi.set', (['self', '"""storage_type_id"""', 'value'], {}), "(self, 'storage_type_id', value)\n", (720564, 720596), False, 'import pulumi\n'), ((721936, 721984), 'pulumi.set', 'pulumi.set', (['__self__', '"""healthcheck"""', 'healthcheck'], {}), "(__self__, 'healthcheck', healthcheck)\n", (721946, 721984), False, 'import pulumi\n'), ((721993, 722061), 'pulumi.set', 'pulumi.set', (['__self__', '"""load_balancing_config"""', 'load_balancing_config'], {}), "(__self__, 'load_balancing_config', load_balancing_config)\n", (722003, 722061), False, 'import pulumi\n'), ((722070, 722104), 'pulumi.set', 'pulumi.set', (['__self__', '"""name"""', 'name'], {}), "(__self__, 'name', name)\n", (722080, 722104), False, 'import pulumi\n'), ((722113, 722147), 'pulumi.set', 'pulumi.set', (['__self__', '"""port"""', 'port'], {}), "(__self__, 'port', port)\n", (722123, 722147), False, 'import pulumi\n'), ((722156, 722214), 'pulumi.set', 'pulumi.set', (['__self__', '"""target_group_ids"""', 'target_group_ids'], {}), "(__self__, 'target_group_ids', target_group_ids)\n", (722166, 722214), False, 'import pulumi\n'), ((722223, 722255), 'pulumi.set', 'pulumi.set', (['__self__', '"""tls"""', 'tls'], {}), "(__self__, 'tls', tls)\n", (722233, 722255), False, 'import pulumi\n'), ((722264, 722302), 'pulumi.set', 'pulumi.set', (['__self__', '"""weight"""', 'weight'], {}), "(__self__, 'weight', weight)\n", (722274, 722302), False, 'import pulumi\n'), ((722553, 722584), 'pulumi.get', 'pulumi.get', (['self', '"""healthcheck"""'], {}), "(self, 'healthcheck')\n", (722563, 722584), False, 'import pulumi\n'), ((722700, 722738), 'pulumi.set', 'pulumi.set', (['self', '"""healthcheck"""', 'value'], {}), "(self, 'healthcheck', value)\n", (722710, 722738), False, 'import pulumi\n'), ((723045, 723086), 'pulumi.get', 'pulumi.get', (['self', '"""load_balancing_config"""'], {}), "(self, 'load_balancing_config')\n", (723055, 723086), False, 'import pulumi\n'), ((723230, 723278), 'pulumi.set', 'pulumi.set', (['self', '"""load_balancing_config"""', 'value'], {}), "(self, 'load_balancing_config', value)\n", (723240, 723278), False, 'import pulumi\n'), ((723416, 723440), 'pulumi.get', 'pulumi.get', (['self', '"""name"""'], {}), "(self, 'name')\n", (723426, 723440), False, 'import pulumi\n'), ((723499, 723530), 'pulumi.set', 'pulumi.set', (['self', '"""name"""', 'value'], {}), "(self, 'name', value)\n", (723509, 723530), False, 'import pulumi\n'), ((723666, 723690), 'pulumi.get', 'pulumi.get', (['self', '"""port"""'], {}), "(self, 'port')\n", (723676, 723690), False, 'import pulumi\n'), ((723749, 723780), 'pulumi.set', 'pulumi.set', (['self', '"""port"""', 'value'], {}), "(self, 'port', value)\n", (723759, 723780), False, 'import pulumi\n'), ((723976, 724012), 'pulumi.get', 'pulumi.get', (['self', '"""target_group_ids"""'], {}), "(self, 'target_group_ids')\n", (723986, 724012), False, 'import pulumi\n'), ((724105, 724148), 'pulumi.set', 'pulumi.set', (['self', '"""target_group_ids"""', 'value'], {}), "(self, 'target_group_ids', value)\n", (724115, 724148), False, 'import pulumi\n'), ((724375, 724398), 'pulumi.get', 'pulumi.get', (['self', '"""tls"""'], {}), "(self, 'tls')\n", (724385, 724398), False, 'import pulumi\n'), ((724490, 724520), 'pulumi.set', 'pulumi.set', (['self', '"""tls"""', 'value'], {}), "(self, 'tls', value)\n", (724500, 724520), False, 'import pulumi\n'), ((724746, 724772), 'pulumi.get', 'pulumi.get', (['self', '"""weight"""'], {}), "(self, 'weight')\n", (724756, 724772), False, 'import pulumi\n'), ((724835, 724868), 'pulumi.set', 'pulumi.set', (['self', '"""weight"""', 'value'], {}), "(self, 'weight', value)\n", (724845, 724868), False, 'import pulumi\n'), ((727054, 727112), 'pulumi.set', 'pulumi.set', (['__self__', '"""grpc_healthcheck"""', 'grpc_healthcheck'], {}), "(__self__, 'grpc_healthcheck', grpc_healthcheck)\n", (727064, 727112), False, 'import pulumi\n'), ((727121, 727179), 'pulumi.set', 'pulumi.set', (['__self__', '"""healthcheck_port"""', 'healthcheck_port'], {}), "(__self__, 'healthcheck_port', healthcheck_port)\n", (727131, 727179), False, 'import pulumi\n'), ((727188, 727248), 'pulumi.set', 'pulumi.set', (['__self__', '"""healthy_threshold"""', 'healthy_threshold'], {}), "(__self__, 'healthy_threshold', healthy_threshold)\n", (727198, 727248), False, 'import pulumi\n'), ((727257, 727315), 'pulumi.set', 'pulumi.set', (['__self__', '"""http_healthcheck"""', 'http_healthcheck'], {}), "(__self__, 'http_healthcheck', http_healthcheck)\n", (727267, 727315), False, 'import pulumi\n'), ((727324, 727366), 'pulumi.set', 'pulumi.set', (['__self__', '"""interval"""', 'interval'], {}), "(__self__, 'interval', interval)\n", (727334, 727366), False, 'import pulumi\n'), ((727375, 727447), 'pulumi.set', 'pulumi.set', (['__self__', '"""interval_jitter_percent"""', 'interval_jitter_percent'], {}), "(__self__, 'interval_jitter_percent', interval_jitter_percent)\n", (727385, 727447), False, 'import pulumi\n'), ((727456, 727518), 'pulumi.set', 'pulumi.set', (['__self__', '"""stream_healthcheck"""', 'stream_healthcheck'], {}), "(__self__, 'stream_healthcheck', stream_healthcheck)\n", (727466, 727518), False, 'import pulumi\n'), ((727527, 727567), 'pulumi.set', 'pulumi.set', (['__self__', '"""timeout"""', 'timeout'], {}), "(__self__, 'timeout', timeout)\n", (727537, 727567), False, 'import pulumi\n'), ((727576, 727640), 'pulumi.set', 'pulumi.set', (['__self__', '"""unhealthy_threshold"""', 'unhealthy_threshold'], {}), "(__self__, 'unhealthy_threshold', unhealthy_threshold)\n", (727586, 727640), False, 'import pulumi\n'), ((727944, 727980), 'pulumi.get', 'pulumi.get', (['self', '"""grpc_healthcheck"""'], {}), "(self, 'grpc_healthcheck')\n", (727954, 727980), False, 'import pulumi\n'), ((728121, 728164), 'pulumi.set', 'pulumi.set', (['self', '"""grpc_healthcheck"""', 'value'], {}), "(self, 'grpc_healthcheck', value)\n", (728131, 728164), False, 'import pulumi\n'), ((728356, 728392), 'pulumi.get', 'pulumi.get', (['self', '"""healthcheck_port"""'], {}), "(self, 'healthcheck_port')\n", (728366, 728392), False, 'import pulumi\n'), ((728475, 728518), 'pulumi.set', 'pulumi.set', (['self', '"""healthcheck_port"""', 'value'], {}), "(self, 'healthcheck_port', value)\n", (728485, 728518), False, 'import pulumi\n'), ((728876, 728913), 'pulumi.get', 'pulumi.get', (['self', '"""healthy_threshold"""'], {}), "(self, 'healthy_threshold')\n", (728886, 728913), False, 'import pulumi\n'), ((728998, 729042), 'pulumi.set', 'pulumi.set', (['self', '"""healthy_threshold"""', 'value'], {}), "(self, 'healthy_threshold', value)\n", (729008, 729042), False, 'import pulumi\n'), ((729346, 729382), 'pulumi.get', 'pulumi.get', (['self', '"""http_healthcheck"""'], {}), "(self, 'http_healthcheck')\n", (729356, 729382), False, 'import pulumi\n'), ((729523, 729566), 'pulumi.set', 'pulumi.set', (['self', '"""http_healthcheck"""', 'value'], {}), "(self, 'http_healthcheck', value)\n", (729533, 729566), False, 'import pulumi\n'), ((729711, 729739), 'pulumi.get', 'pulumi.get', (['self', '"""interval"""'], {}), "(self, 'interval')\n", (729721, 729739), False, 'import pulumi\n'), ((729806, 729841), 'pulumi.set', 'pulumi.set', (['self', '"""interval"""', 'value'], {}), "(self, 'interval', value)\n", (729816, 729841), False, 'import pulumi\n'), ((730179, 730222), 'pulumi.get', 'pulumi.get', (['self', '"""interval_jitter_percent"""'], {}), "(self, 'interval_jitter_percent')\n", (730189, 730222), False, 'import pulumi\n'), ((730321, 730371), 'pulumi.set', 'pulumi.set', (['self', '"""interval_jitter_percent"""', 'value'], {}), "(self, 'interval_jitter_percent', value)\n", (730331, 730371), False, 'import pulumi\n'), ((730683, 730721), 'pulumi.get', 'pulumi.get', (['self', '"""stream_healthcheck"""'], {}), "(self, 'stream_healthcheck')\n", (730693, 730721), False, 'import pulumi\n'), ((730868, 730913), 'pulumi.set', 'pulumi.set', (['self', '"""stream_healthcheck"""', 'value'], {}), "(self, 'stream_healthcheck', value)\n", (730878, 730913), False, 'import pulumi\n'), ((731067, 731094), 'pulumi.get', 'pulumi.get', (['self', '"""timeout"""'], {}), "(self, 'timeout')\n", (731077, 731094), False, 'import pulumi\n'), ((731159, 731193), 'pulumi.set', 'pulumi.set', (['self', '"""timeout"""', 'value'], {}), "(self, 'timeout', value)\n", (731169, 731193), False, 'import pulumi\n'), ((731537, 731576), 'pulumi.get', 'pulumi.get', (['self', '"""unhealthy_threshold"""'], {}), "(self, 'unhealthy_threshold')\n", (731547, 731576), False, 'import pulumi\n'), ((731665, 731711), 'pulumi.set', 'pulumi.set', (['self', '"""unhealthy_threshold"""', 'value'], {}), "(self, 'unhealthy_threshold', value)\n", (731675, 731711), False, 'import pulumi\n'), ((732001, 732051), 'pulumi.set', 'pulumi.set', (['__self__', '"""service_name"""', 'service_name'], {}), "(__self__, 'service_name', service_name)\n", (732011, 732051), False, 'import pulumi\n'), ((732257, 732289), 'pulumi.get', 'pulumi.get', (['self', '"""service_name"""'], {}), "(self, 'service_name')\n", (732267, 732289), False, 'import pulumi\n'), ((732364, 732403), 'pulumi.set', 'pulumi.set', (['self', '"""service_name"""', 'value'], {}), "(self, 'service_name', value)\n", (732374, 732403), False, 'import pulumi\n'), ((732802, 732836), 'pulumi.set', 'pulumi.set', (['__self__', '"""host"""', 'host'], {}), "(__self__, 'host', host)\n", (732812, 732836), False, 'import pulumi\n'), ((732845, 732881), 'pulumi.set', 'pulumi.set', (['__self__', '"""http2"""', 'http2'], {}), "(__self__, 'http2', http2)\n", (732855, 732881), False, 'import pulumi\n'), ((732890, 732924), 'pulumi.set', 'pulumi.set', (['__self__', '"""path"""', 'path'], {}), "(__self__, 'path', path)\n", (732900, 732924), False, 'import pulumi\n'), ((733068, 733092), 'pulumi.get', 'pulumi.get', (['self', '"""host"""'], {}), "(self, 'host')\n", (733078, 733092), False, 'import pulumi\n'), ((733151, 733182), 'pulumi.set', 'pulumi.set', (['self', '"""host"""', 'value'], {}), "(self, 'host', value)\n", (733161, 733182), False, 'import pulumi\n'), ((733331, 733356), 'pulumi.get', 'pulumi.get', (['self', '"""http2"""'], {}), "(self, 'http2')\n", (733341, 733356), False, 'import pulumi\n'), ((733418, 733450), 'pulumi.set', 'pulumi.set', (['self', '"""http2"""', 'value'], {}), "(self, 'http2', value)\n", (733428, 733450), False, 'import pulumi\n'), ((733570, 733594), 'pulumi.get', 'pulumi.get', (['self', '"""path"""'], {}), "(self, 'path')\n", (733580, 733594), False, 'import pulumi\n'), ((733653, 733684), 'pulumi.set', 'pulumi.set', (['self', '"""path"""', 'value'], {}), "(self, 'path', value)\n", (733663, 733684), False, 'import pulumi\n'), ((734054, 734094), 'pulumi.set', 'pulumi.set', (['__self__', '"""receive"""', 'receive'], {}), "(__self__, 'receive', receive)\n", (734064, 734094), False, 'import pulumi\n'), ((734103, 734137), 'pulumi.set', 'pulumi.set', (['__self__', '"""send"""', 'send'], {}), "(__self__, 'send', send)\n", (734113, 734137), False, 'import pulumi\n'), ((734283, 734310), 'pulumi.get', 'pulumi.get', (['self', '"""receive"""'], {}), "(self, 'receive')\n", (734293, 734310), False, 'import pulumi\n'), ((734375, 734409), 'pulumi.set', 'pulumi.set', (['self', '"""receive"""', 'value'], {}), "(self, 'receive', value)\n", (734385, 734409), False, 'import pulumi\n'), ((734588, 734612), 'pulumi.get', 'pulumi.get', (['self', '"""send"""'], {}), "(self, 'send')\n", (734598, 734612), False, 'import pulumi\n'), ((734671, 734702), 'pulumi.set', 'pulumi.set', (['self', '"""send"""', 'value'], {}), "(self, 'send', value)\n", (734681, 734702), False, 'import pulumi\n'), ((735593, 735683), 'pulumi.set', 'pulumi.set', (['__self__', '"""locality_aware_routing_percent"""', 'locality_aware_routing_percent'], {}), "(__self__, 'locality_aware_routing_percent',\n locality_aware_routing_percent)\n", (735603, 735683), False, 'import pulumi\n'), ((735688, 735744), 'pulumi.set', 'pulumi.set', (['__self__', '"""panic_threshold"""', 'panic_threshold'], {}), "(__self__, 'panic_threshold', panic_threshold)\n", (735698, 735744), False, 'import pulumi\n'), ((735753, 735809), 'pulumi.set', 'pulumi.set', (['__self__', '"""strict_locality"""', 'strict_locality'], {}), "(__self__, 'strict_locality', strict_locality)\n", (735763, 735809), False, 'import pulumi\n'), ((736095, 736145), 'pulumi.get', 'pulumi.get', (['self', '"""locality_aware_routing_percent"""'], {}), "(self, 'locality_aware_routing_percent')\n", (736105, 736145), False, 'import pulumi\n'), ((736256, 736313), 'pulumi.set', 'pulumi.set', (['self', '"""locality_aware_routing_percent"""', 'value'], {}), "(self, 'locality_aware_routing_percent', value)\n", (736266, 736313), False, 'import pulumi\n'), ((736714, 736749), 'pulumi.get', 'pulumi.get', (['self', '"""panic_threshold"""'], {}), "(self, 'panic_threshold')\n", (736724, 736749), False, 'import pulumi\n'), ((736830, 736872), 'pulumi.set', 'pulumi.set', (['self', '"""panic_threshold"""', 'value'], {}), "(self, 'panic_threshold', value)\n", (736840, 736872), False, 'import pulumi\n'), ((737132, 737167), 'pulumi.get', 'pulumi.get', (['self', '"""strict_locality"""'], {}), "(self, 'strict_locality')\n", (737142, 737167), False, 'import pulumi\n'), ((737249, 737291), 'pulumi.set', 'pulumi.set', (['self', '"""strict_locality"""', 'value'], {}), "(self, 'strict_locality', value)\n", (737259, 737291), False, 'import pulumi\n'), ((737864, 737896), 'pulumi.set', 'pulumi.set', (['__self__', '"""sni"""', 'sni'], {}), "(__self__, 'sni', sni)\n", (737874, 737896), False, 'import pulumi\n'), ((737905, 737967), 'pulumi.set', 'pulumi.set', (['__self__', '"""validation_context"""', 'validation_context'], {}), "(__self__, 'validation_context', validation_context)\n", (737915, 737967), False, 'import pulumi\n'), ((738360, 738383), 'pulumi.get', 'pulumi.get', (['self', '"""sni"""'], {}), "(self, 'sni')\n", (738370, 738383), False, 'import pulumi\n'), ((738440, 738470), 'pulumi.set', 'pulumi.set', (['self', '"""sni"""', 'value'], {}), "(self, 'sni', value)\n", (738450, 738470), False, 'import pulumi\n'), ((738639, 738677), 'pulumi.get', 'pulumi.get', (['self', '"""validation_context"""'], {}), "(self, 'validation_context')\n", (738649, 738677), False, 'import pulumi\n'), ((738816, 738861), 'pulumi.set', 'pulumi.set', (['self', '"""validation_context"""', 'value'], {}), "(self, 'validation_context', value)\n", (738826, 738861), False, 'import pulumi\n'), ((739060, 739118), 'pulumi.set', 'pulumi.set', (['__self__', '"""trusted_ca_bytes"""', 'trusted_ca_bytes'], {}), "(__self__, 'trusted_ca_bytes', trusted_ca_bytes)\n", (739070, 739118), False, 'import pulumi\n'), ((739127, 739179), 'pulumi.set', 'pulumi.set', (['__self__', '"""trusted_ca_id"""', 'trusted_ca_id'], {}), "(__self__, 'trusted_ca_id', trusted_ca_id)\n", (739137, 739179), False, 'import pulumi\n'), ((739291, 739327), 'pulumi.get', 'pulumi.get', (['self', '"""trusted_ca_bytes"""'], {}), "(self, 'trusted_ca_bytes')\n", (739301, 739327), False, 'import pulumi\n'), ((739410, 739453), 'pulumi.set', 'pulumi.set', (['self', '"""trusted_ca_bytes"""', 'value'], {}), "(self, 'trusted_ca_bytes', value)\n", (739420, 739453), False, 'import pulumi\n'), ((739559, 739592), 'pulumi.get', 'pulumi.get', (['self', '"""trusted_ca_id"""'], {}), "(self, 'trusted_ca_id')\n", (739569, 739592), False, 'import pulumi\n'), ((739669, 739709), 'pulumi.set', 'pulumi.set', (['self', '"""trusted_ca_id"""', 'value'], {}), "(self, 'trusted_ca_id', value)\n", (739679, 739709), False, 'import pulumi\n'), ((741144, 741192), 'pulumi.set', 'pulumi.set', (['__self__', '"""healthcheck"""', 'healthcheck'], {}), "(__self__, 'healthcheck', healthcheck)\n", (741154, 741192), False, 'import pulumi\n'), ((741201, 741237), 'pulumi.set', 'pulumi.set', (['__self__', '"""http2"""', 'http2'], {}), "(__self__, 'http2', http2)\n", (741211, 741237), False, 'import pulumi\n'), ((741246, 741314), 'pulumi.set', 'pulumi.set', (['__self__', '"""load_balancing_config"""', 'load_balancing_config'], {}), "(__self__, 'load_balancing_config', load_balancing_config)\n", (741256, 741314), False, 'import pulumi\n'), ((741323, 741357), 'pulumi.set', 'pulumi.set', (['__self__', '"""name"""', 'name'], {}), "(__self__, 'name', name)\n", (741333, 741357), False, 'import pulumi\n'), ((741366, 741400), 'pulumi.set', 'pulumi.set', (['__self__', '"""port"""', 'port'], {}), "(__self__, 'port', port)\n", (741376, 741400), False, 'import pulumi\n'), ((741409, 741467), 'pulumi.set', 'pulumi.set', (['__self__', '"""target_group_ids"""', 'target_group_ids'], {}), "(__self__, 'target_group_ids', target_group_ids)\n", (741419, 741467), False, 'import pulumi\n'), ((741476, 741508), 'pulumi.set', 'pulumi.set', (['__self__', '"""tls"""', 'tls'], {}), "(__self__, 'tls', tls)\n", (741486, 741508), False, 'import pulumi\n'), ((741517, 741555), 'pulumi.set', 'pulumi.set', (['__self__', '"""weight"""', 'weight'], {}), "(__self__, 'weight', weight)\n", (741527, 741555), False, 'import pulumi\n'), ((741806, 741837), 'pulumi.get', 'pulumi.get', (['self', '"""healthcheck"""'], {}), "(self, 'healthcheck')\n", (741816, 741837), False, 'import pulumi\n'), ((741953, 741991), 'pulumi.set', 'pulumi.set', (['self', '"""healthcheck"""', 'value'], {}), "(self, 'healthcheck', value)\n", (741963, 741991), False, 'import pulumi\n'), ((742140, 742165), 'pulumi.get', 'pulumi.get', (['self', '"""http2"""'], {}), "(self, 'http2')\n", (742150, 742165), False, 'import pulumi\n'), ((742227, 742259), 'pulumi.set', 'pulumi.set', (['self', '"""http2"""', 'value'], {}), "(self, 'http2', value)\n", (742237, 742259), False, 'import pulumi\n'), ((742566, 742607), 'pulumi.get', 'pulumi.get', (['self', '"""load_balancing_config"""'], {}), "(self, 'load_balancing_config')\n", (742576, 742607), False, 'import pulumi\n'), ((742751, 742799), 'pulumi.set', 'pulumi.set', (['self', '"""load_balancing_config"""', 'value'], {}), "(self, 'load_balancing_config', value)\n", (742761, 742799), False, 'import pulumi\n'), ((742937, 742961), 'pulumi.get', 'pulumi.get', (['self', '"""name"""'], {}), "(self, 'name')\n", (742947, 742961), False, 'import pulumi\n'), ((743020, 743051), 'pulumi.set', 'pulumi.set', (['self', '"""name"""', 'value'], {}), "(self, 'name', value)\n", (743030, 743051), False, 'import pulumi\n'), ((743187, 743211), 'pulumi.get', 'pulumi.get', (['self', '"""port"""'], {}), "(self, 'port')\n", (743197, 743211), False, 'import pulumi\n'), ((743270, 743301), 'pulumi.set', 'pulumi.set', (['self', '"""port"""', 'value'], {}), "(self, 'port', value)\n", (743280, 743301), False, 'import pulumi\n'), ((743497, 743533), 'pulumi.get', 'pulumi.get', (['self', '"""target_group_ids"""'], {}), "(self, 'target_group_ids')\n", (743507, 743533), False, 'import pulumi\n'), ((743626, 743669), 'pulumi.set', 'pulumi.set', (['self', '"""target_group_ids"""', 'value'], {}), "(self, 'target_group_ids', value)\n", (743636, 743669), False, 'import pulumi\n'), ((743896, 743919), 'pulumi.get', 'pulumi.get', (['self', '"""tls"""'], {}), "(self, 'tls')\n", (743906, 743919), False, 'import pulumi\n'), ((744011, 744041), 'pulumi.set', 'pulumi.set', (['self', '"""tls"""', 'value'], {}), "(self, 'tls', value)\n", (744021, 744041), False, 'import pulumi\n'), ((744267, 744293), 'pulumi.get', 'pulumi.get', (['self', '"""weight"""'], {}), "(self, 'weight')\n", (744277, 744293), False, 'import pulumi\n'), ((744356, 744389), 'pulumi.set', 'pulumi.set', (['self', '"""weight"""', 'value'], {}), "(self, 'weight', value)\n", (744366, 744389), False, 'import pulumi\n'), ((746575, 746633), 'pulumi.set', 'pulumi.set', (['__self__', '"""grpc_healthcheck"""', 'grpc_healthcheck'], {}), "(__self__, 'grpc_healthcheck', grpc_healthcheck)\n", (746585, 746633), False, 'import pulumi\n'), ((746642, 746700), 'pulumi.set', 'pulumi.set', (['__self__', '"""healthcheck_port"""', 'healthcheck_port'], {}), "(__self__, 'healthcheck_port', healthcheck_port)\n", (746652, 746700), False, 'import pulumi\n'), ((746709, 746769), 'pulumi.set', 'pulumi.set', (['__self__', '"""healthy_threshold"""', 'healthy_threshold'], {}), "(__self__, 'healthy_threshold', healthy_threshold)\n", (746719, 746769), False, 'import pulumi\n'), ((746778, 746836), 'pulumi.set', 'pulumi.set', (['__self__', '"""http_healthcheck"""', 'http_healthcheck'], {}), "(__self__, 'http_healthcheck', http_healthcheck)\n", (746788, 746836), False, 'import pulumi\n'), ((746845, 746887), 'pulumi.set', 'pulumi.set', (['__self__', '"""interval"""', 'interval'], {}), "(__self__, 'interval', interval)\n", (746855, 746887), False, 'import pulumi\n'), ((746896, 746968), 'pulumi.set', 'pulumi.set', (['__self__', '"""interval_jitter_percent"""', 'interval_jitter_percent'], {}), "(__self__, 'interval_jitter_percent', interval_jitter_percent)\n", (746906, 746968), False, 'import pulumi\n'), ((746977, 747039), 'pulumi.set', 'pulumi.set', (['__self__', '"""stream_healthcheck"""', 'stream_healthcheck'], {}), "(__self__, 'stream_healthcheck', stream_healthcheck)\n", (746987, 747039), False, 'import pulumi\n'), ((747048, 747088), 'pulumi.set', 'pulumi.set', (['__self__', '"""timeout"""', 'timeout'], {}), "(__self__, 'timeout', timeout)\n", (747058, 747088), False, 'import pulumi\n'), ((747097, 747161), 'pulumi.set', 'pulumi.set', (['__self__', '"""unhealthy_threshold"""', 'unhealthy_threshold'], {}), "(__self__, 'unhealthy_threshold', unhealthy_threshold)\n", (747107, 747161), False, 'import pulumi\n'), ((747465, 747501), 'pulumi.get', 'pulumi.get', (['self', '"""grpc_healthcheck"""'], {}), "(self, 'grpc_healthcheck')\n", (747475, 747501), False, 'import pulumi\n'), ((747642, 747685), 'pulumi.set', 'pulumi.set', (['self', '"""grpc_healthcheck"""', 'value'], {}), "(self, 'grpc_healthcheck', value)\n", (747652, 747685), False, 'import pulumi\n'), ((747877, 747913), 'pulumi.get', 'pulumi.get', (['self', '"""healthcheck_port"""'], {}), "(self, 'healthcheck_port')\n", (747887, 747913), False, 'import pulumi\n'), ((747996, 748039), 'pulumi.set', 'pulumi.set', (['self', '"""healthcheck_port"""', 'value'], {}), "(self, 'healthcheck_port', value)\n", (748006, 748039), False, 'import pulumi\n'), ((748397, 748434), 'pulumi.get', 'pulumi.get', (['self', '"""healthy_threshold"""'], {}), "(self, 'healthy_threshold')\n", (748407, 748434), False, 'import pulumi\n'), ((748519, 748563), 'pulumi.set', 'pulumi.set', (['self', '"""healthy_threshold"""', 'value'], {}), "(self, 'healthy_threshold', value)\n", (748529, 748563), False, 'import pulumi\n'), ((748867, 748903), 'pulumi.get', 'pulumi.get', (['self', '"""http_healthcheck"""'], {}), "(self, 'http_healthcheck')\n", (748877, 748903), False, 'import pulumi\n'), ((749044, 749087), 'pulumi.set', 'pulumi.set', (['self', '"""http_healthcheck"""', 'value'], {}), "(self, 'http_healthcheck', value)\n", (749054, 749087), False, 'import pulumi\n'), ((749232, 749260), 'pulumi.get', 'pulumi.get', (['self', '"""interval"""'], {}), "(self, 'interval')\n", (749242, 749260), False, 'import pulumi\n'), ((749327, 749362), 'pulumi.set', 'pulumi.set', (['self', '"""interval"""', 'value'], {}), "(self, 'interval', value)\n", (749337, 749362), False, 'import pulumi\n'), ((749700, 749743), 'pulumi.get', 'pulumi.get', (['self', '"""interval_jitter_percent"""'], {}), "(self, 'interval_jitter_percent')\n", (749710, 749743), False, 'import pulumi\n'), ((749842, 749892), 'pulumi.set', 'pulumi.set', (['self', '"""interval_jitter_percent"""', 'value'], {}), "(self, 'interval_jitter_percent', value)\n", (749852, 749892), False, 'import pulumi\n'), ((750204, 750242), 'pulumi.get', 'pulumi.get', (['self', '"""stream_healthcheck"""'], {}), "(self, 'stream_healthcheck')\n", (750214, 750242), False, 'import pulumi\n'), ((750389, 750434), 'pulumi.set', 'pulumi.set', (['self', '"""stream_healthcheck"""', 'value'], {}), "(self, 'stream_healthcheck', value)\n", (750399, 750434), False, 'import pulumi\n'), ((750588, 750615), 'pulumi.get', 'pulumi.get', (['self', '"""timeout"""'], {}), "(self, 'timeout')\n", (750598, 750615), False, 'import pulumi\n'), ((750680, 750714), 'pulumi.set', 'pulumi.set', (['self', '"""timeout"""', 'value'], {}), "(self, 'timeout', value)\n", (750690, 750714), False, 'import pulumi\n'), ((751058, 751097), 'pulumi.get', 'pulumi.get', (['self', '"""unhealthy_threshold"""'], {}), "(self, 'unhealthy_threshold')\n", (751068, 751097), False, 'import pulumi\n'), ((751186, 751232), 'pulumi.set', 'pulumi.set', (['self', '"""unhealthy_threshold"""', 'value'], {}), "(self, 'unhealthy_threshold', value)\n", (751196, 751232), False, 'import pulumi\n'), ((751522, 751572), 'pulumi.set', 'pulumi.set', (['__self__', '"""service_name"""', 'service_name'], {}), "(__self__, 'service_name', service_name)\n", (751532, 751572), False, 'import pulumi\n'), ((751778, 751810), 'pulumi.get', 'pulumi.get', (['self', '"""service_name"""'], {}), "(self, 'service_name')\n", (751788, 751810), False, 'import pulumi\n'), ((751885, 751924), 'pulumi.set', 'pulumi.set', (['self', '"""service_name"""', 'value'], {}), "(self, 'service_name', value)\n", (751895, 751924), False, 'import pulumi\n'), ((752323, 752357), 'pulumi.set', 'pulumi.set', (['__self__', '"""host"""', 'host'], {}), "(__self__, 'host', host)\n", (752333, 752357), False, 'import pulumi\n'), ((752366, 752402), 'pulumi.set', 'pulumi.set', (['__self__', '"""http2"""', 'http2'], {}), "(__self__, 'http2', http2)\n", (752376, 752402), False, 'import pulumi\n'), ((752411, 752445), 'pulumi.set', 'pulumi.set', (['__self__', '"""path"""', 'path'], {}), "(__self__, 'path', path)\n", (752421, 752445), False, 'import pulumi\n'), ((752589, 752613), 'pulumi.get', 'pulumi.get', (['self', '"""host"""'], {}), "(self, 'host')\n", (752599, 752613), False, 'import pulumi\n'), ((752672, 752703), 'pulumi.set', 'pulumi.set', (['self', '"""host"""', 'value'], {}), "(self, 'host', value)\n", (752682, 752703), False, 'import pulumi\n'), ((752852, 752877), 'pulumi.get', 'pulumi.get', (['self', '"""http2"""'], {}), "(self, 'http2')\n", (752862, 752877), False, 'import pulumi\n'), ((752939, 752971), 'pulumi.set', 'pulumi.set', (['self', '"""http2"""', 'value'], {}), "(self, 'http2', value)\n", (752949, 752971), False, 'import pulumi\n'), ((753091, 753115), 'pulumi.get', 'pulumi.get', (['self', '"""path"""'], {}), "(self, 'path')\n", (753101, 753115), False, 'import pulumi\n'), ((753174, 753205), 'pulumi.set', 'pulumi.set', (['self', '"""path"""', 'value'], {}), "(self, 'path', value)\n", (753184, 753205), False, 'import pulumi\n'), ((753575, 753615), 'pulumi.set', 'pulumi.set', (['__self__', '"""receive"""', 'receive'], {}), "(__self__, 'receive', receive)\n", (753585, 753615), False, 'import pulumi\n'), ((753624, 753658), 'pulumi.set', 'pulumi.set', (['__self__', '"""send"""', 'send'], {}), "(__self__, 'send', send)\n", (753634, 753658), False, 'import pulumi\n'), ((753804, 753831), 'pulumi.get', 'pulumi.get', (['self', '"""receive"""'], {}), "(self, 'receive')\n", (753814, 753831), False, 'import pulumi\n'), ((753896, 753930), 'pulumi.set', 'pulumi.set', (['self', '"""receive"""', 'value'], {}), "(self, 'receive', value)\n", (753906, 753930), False, 'import pulumi\n'), ((754109, 754133), 'pulumi.get', 'pulumi.get', (['self', '"""send"""'], {}), "(self, 'send')\n", (754119, 754133), False, 'import pulumi\n'), ((754192, 754223), 'pulumi.set', 'pulumi.set', (['self', '"""send"""', 'value'], {}), "(self, 'send', value)\n", (754202, 754223), False, 'import pulumi\n'), ((755114, 755204), 'pulumi.set', 'pulumi.set', (['__self__', '"""locality_aware_routing_percent"""', 'locality_aware_routing_percent'], {}), "(__self__, 'locality_aware_routing_percent',\n locality_aware_routing_percent)\n", (755124, 755204), False, 'import pulumi\n'), ((755209, 755265), 'pulumi.set', 'pulumi.set', (['__self__', '"""panic_threshold"""', 'panic_threshold'], {}), "(__self__, 'panic_threshold', panic_threshold)\n", (755219, 755265), False, 'import pulumi\n'), ((755274, 755330), 'pulumi.set', 'pulumi.set', (['__self__', '"""strict_locality"""', 'strict_locality'], {}), "(__self__, 'strict_locality', strict_locality)\n", (755284, 755330), False, 'import pulumi\n'), ((755616, 755666), 'pulumi.get', 'pulumi.get', (['self', '"""locality_aware_routing_percent"""'], {}), "(self, 'locality_aware_routing_percent')\n", (755626, 755666), False, 'import pulumi\n'), ((755777, 755834), 'pulumi.set', 'pulumi.set', (['self', '"""locality_aware_routing_percent"""', 'value'], {}), "(self, 'locality_aware_routing_percent', value)\n", (755787, 755834), False, 'import pulumi\n'), ((756235, 756270), 'pulumi.get', 'pulumi.get', (['self', '"""panic_threshold"""'], {}), "(self, 'panic_threshold')\n", (756245, 756270), False, 'import pulumi\n'), ((756351, 756393), 'pulumi.set', 'pulumi.set', (['self', '"""panic_threshold"""', 'value'], {}), "(self, 'panic_threshold', value)\n", (756361, 756393), False, 'import pulumi\n'), ((756653, 756688), 'pulumi.get', 'pulumi.get', (['self', '"""strict_locality"""'], {}), "(self, 'strict_locality')\n", (756663, 756688), False, 'import pulumi\n'), ((756770, 756812), 'pulumi.set', 'pulumi.set', (['self', '"""strict_locality"""', 'value'], {}), "(self, 'strict_locality', value)\n", (756780, 756812), False, 'import pulumi\n'), ((757385, 757417), 'pulumi.set', 'pulumi.set', (['__self__', '"""sni"""', 'sni'], {}), "(__self__, 'sni', sni)\n", (757395, 757417), False, 'import pulumi\n'), ((757426, 757488), 'pulumi.set', 'pulumi.set', (['__self__', '"""validation_context"""', 'validation_context'], {}), "(__self__, 'validation_context', validation_context)\n", (757436, 757488), False, 'import pulumi\n'), ((757881, 757904), 'pulumi.get', 'pulumi.get', (['self', '"""sni"""'], {}), "(self, 'sni')\n", (757891, 757904), False, 'import pulumi\n'), ((757961, 757991), 'pulumi.set', 'pulumi.set', (['self', '"""sni"""', 'value'], {}), "(self, 'sni', value)\n", (757971, 757991), False, 'import pulumi\n'), ((758160, 758198), 'pulumi.get', 'pulumi.get', (['self', '"""validation_context"""'], {}), "(self, 'validation_context')\n", (758170, 758198), False, 'import pulumi\n'), ((758337, 758382), 'pulumi.set', 'pulumi.set', (['self', '"""validation_context"""', 'value'], {}), "(self, 'validation_context', value)\n", (758347, 758382), False, 'import pulumi\n'), ((758581, 758639), 'pulumi.set', 'pulumi.set', (['__self__', '"""trusted_ca_bytes"""', 'trusted_ca_bytes'], {}), "(__self__, 'trusted_ca_bytes', trusted_ca_bytes)\n", (758591, 758639), False, 'import pulumi\n'), ((758648, 758700), 'pulumi.set', 'pulumi.set', (['__self__', '"""trusted_ca_id"""', 'trusted_ca_id'], {}), "(__self__, 'trusted_ca_id', trusted_ca_id)\n", (758658, 758700), False, 'import pulumi\n'), ((758812, 758848), 'pulumi.get', 'pulumi.get', (['self', '"""trusted_ca_bytes"""'], {}), "(self, 'trusted_ca_bytes')\n", (758822, 758848), False, 'import pulumi\n'), ((758931, 758974), 'pulumi.set', 'pulumi.set', (['self', '"""trusted_ca_bytes"""', 'value'], {}), "(self, 'trusted_ca_bytes', value)\n", (758941, 758974), False, 'import pulumi\n'), ((759080, 759113), 'pulumi.get', 'pulumi.get', (['self', '"""trusted_ca_id"""'], {}), "(self, 'trusted_ca_id')\n", (759090, 759113), False, 'import pulumi\n'), ((759190, 759230), 'pulumi.set', 'pulumi.set', (['self', '"""trusted_ca_id"""', 'value'], {}), "(self, 'trusted_ca_id', value)\n", (759200, 759230), False, 'import pulumi\n'), ((759383, 759455), 'pulumi.set', 'pulumi.set', (['__self__', '"""disk_placement_group_id"""', 'disk_placement_group_id'], {}), "(__self__, 'disk_placement_group_id', disk_placement_group_id)\n", (759393, 759455), False, 'import pulumi\n'), ((759580, 759623), 'pulumi.get', 'pulumi.get', (['self', '"""disk_placement_group_id"""'], {}), "(self, 'disk_placement_group_id')\n", (759590, 759623), False, 'import pulumi\n'), ((759720, 759770), 'pulumi.set', 'pulumi.set', (['self', '"""disk_placement_group_id"""', 'value'], {}), "(self, 'disk_placement_group_id', value)\n", (759730, 759770), False, 'import pulumi\n'), ((759918, 759980), 'pulumi.set', 'pulumi.set', (['__self__', '"""placement_group_id"""', 'placement_group_id'], {}), "(__self__, 'placement_group_id', placement_group_id)\n", (759928, 759980), False, 'import pulumi\n'), ((760096, 760134), 'pulumi.get', 'pulumi.get', (['self', '"""placement_group_id"""'], {}), "(self, 'placement_group_id')\n", (760106, 760134), False, 'import pulumi\n'), ((760221, 760266), 'pulumi.set', 'pulumi.set', (['self', '"""placement_group_id"""', 'value'], {}), "(self, 'placement_group_id', value)\n", (760231, 760266), False, 'import pulumi\n'), ((760842, 760873), 'pulumi.get', 'pulumi.get', (['self', '"""preemptible"""'], {}), "(self, 'preemptible')\n", (760852, 760873), False, 'import pulumi\n'), ((760957, 760995), 'pulumi.set', 'pulumi.set', (['self', '"""preemptible"""', 'value'], {}), "(self, 'preemptible', value)\n", (760967, 760995), False, 'import pulumi\n'), ((761246, 761278), 'pulumi.set', 'pulumi.set', (['__self__', '"""tag"""', 'tag'], {}), "(__self__, 'tag', tag)\n", (761256, 761278), False, 'import pulumi\n'), ((761599, 761622), 'pulumi.get', 'pulumi.get', (['self', '"""tag"""'], {}), "(self, 'tag')\n", (761609, 761622), False, 'import pulumi\n'), ((761679, 761709), 'pulumi.set', 'pulumi.set', (['self', '"""tag"""', 'value'], {}), "(self, 'tag', value)\n", (761689, 761709), False, 'import pulumi\n'), ((761839, 761879), 'pulumi.get', 'pulumi.get', (['self', '"""zone_instances_limit"""'], {}), "(self, 'zone_instances_limit')\n", (761849, 761879), False, 'import pulumi\n'), ((761980, 762027), 'pulumi.set', 'pulumi.set', (['self', '"""zone_instances_limit"""', 'value'], {}), "(self, 'zone_instances_limit', value)\n", (761990, 762027), False, 'import pulumi\n'), ((762155, 762194), 'pulumi.get', 'pulumi.get', (['self', '"""zone_requests_limit"""'], {}), "(self, 'zone_requests_limit')\n", (762165, 762194), False, 'import pulumi\n'), ((762293, 762339), 'pulumi.set', 'pulumi.set', (['self', '"""zone_requests_limit"""', 'value'], {}), "(self, 'zone_requests_limit', value)\n", (762303, 762339), False, 'import pulumi\n'), ((763051, 763091), 'pulumi.set', 'pulumi.set', (['__self__', '"""members"""', 'members'], {}), "(__self__, 'members', members)\n", (763061, 763091), False, 'import pulumi\n'), ((763100, 763134), 'pulumi.set', 'pulumi.set', (['__self__', '"""role"""', 'role'], {}), "(__self__, 'role', role)\n", (763110, 763134), False, 'import pulumi\n'), ((763563, 763590), 'pulumi.get', 'pulumi.get', (['self', '"""members"""'], {}), "(self, 'members')\n", (763573, 763590), False, 'import pulumi\n'), ((763665, 763699), 'pulumi.set', 'pulumi.set', (['self', '"""members"""', 'value'], {}), "(self, 'members', value)\n", (763675, 763699), False, 'import pulumi\n'), ((763937, 763961), 'pulumi.get', 'pulumi.get', (['self', '"""role"""'], {}), "(self, 'role')\n", (763947, 763961), False, 'import pulumi\n'), ((764020, 764051), 'pulumi.set', 'pulumi.set', (['self', '"""role"""', 'value'], {}), "(self, 'role', value)\n", (764030, 764051), False, 'import pulumi\n'), ((764354, 764394), 'pulumi.set', 'pulumi.set', (['__self__', '"""enabled"""', 'enabled'], {}), "(__self__, 'enabled', enabled)\n", (764364, 764394), False, 'import pulumi\n'), ((764617, 764644), 'pulumi.get', 'pulumi.get', (['self', '"""enabled"""'], {}), "(self, 'enabled')\n", (764627, 764644), False, 'import pulumi\n'), ((764710, 764744), 'pulumi.set', 'pulumi.set', (['self', '"""enabled"""', 'value'], {}), "(self, 'enabled', value)\n", (764720, 764744), False, 'import pulumi\n'), ((766194, 766230), 'pulumi.set', 'pulumi.set', (['__self__', '"""kafka"""', 'kafka'], {}), "(__self__, 'kafka', kafka)\n", (766204, 766230), False, 'import pulumi\n'), ((766239, 766279), 'pulumi.set', 'pulumi.set', (['__self__', '"""version"""', 'version'], {}), "(__self__, 'version', version)\n", (766249, 766279), False, 'import pulumi\n'), ((766288, 766324), 'pulumi.set', 'pulumi.set', (['__self__', '"""zones"""', 'zones'], {}), "(__self__, 'zones', zones)\n", (766298, 766324), False, 'import pulumi\n'), ((766333, 766377), 'pulumi.set', 'pulumi.set', (['__self__', '"""zookeeper"""', 'zookeeper'], {}), "(__self__, 'zookeeper', zookeeper)\n", (766343, 766377), False, 'import pulumi\n'), ((767040, 767065), 'pulumi.get', 'pulumi.get', (['self', '"""kafka"""'], {}), "(self, 'kafka')\n", (767050, 767065), False, 'import pulumi\n'), ((767158, 767190), 'pulumi.set', 'pulumi.set', (['self', '"""kafka"""', 'value'], {}), "(self, 'kafka', value)\n", (767168, 767190), False, 'import pulumi\n'), ((767351, 767378), 'pulumi.get', 'pulumi.get', (['self', '"""version"""'], {}), "(self, 'version')\n", (767361, 767378), False, 'import pulumi\n'), ((767443, 767477), 'pulumi.set', 'pulumi.set', (['self', '"""version"""', 'value'], {}), "(self, 'version', value)\n", (767453, 767477), False, 'import pulumi\n'), ((767636, 767661), 'pulumi.get', 'pulumi.get', (['self', '"""zones"""'], {}), "(self, 'zones')\n", (767646, 767661), False, 'import pulumi\n'), ((767732, 767764), 'pulumi.set', 'pulumi.set', (['self', '"""zones"""', 'value'], {}), "(self, 'zones', value)\n", (767742, 767764), False, 'import pulumi\n'), ((768003, 768032), 'pulumi.get', 'pulumi.get', (['self', '"""zookeeper"""'], {}), "(self, 'zookeeper')\n", (768013, 768032), False, 'import pulumi\n'), ((768137, 768173), 'pulumi.set', 'pulumi.set', (['self', '"""zookeeper"""', 'value'], {}), "(self, 'zookeeper', value)\n", (768147, 768173), False, 'import pulumi\n'), ((768403, 768439), 'pulumi.get', 'pulumi.get', (['self', '"""assign_public_ip"""'], {}), "(self, 'assign_public_ip')\n", (768413, 768439), False, 'import pulumi\n'), ((768533, 768576), 'pulumi.set', 'pulumi.set', (['self', '"""assign_public_ip"""', 'value'], {}), "(self, 'assign_public_ip', value)\n", (768543, 768576), False, 'import pulumi\n'), ((768776, 768809), 'pulumi.get', 'pulumi.get', (['self', '"""brokers_count"""'], {}), "(self, 'brokers_count')\n", (768786, 768809), False, 'import pulumi\n'), ((768896, 768936), 'pulumi.set', 'pulumi.set', (['self', '"""brokers_count"""', 'value'], {}), "(self, 'brokers_count', value)\n", (768906, 768936), False, 'import pulumi\n'), ((769178, 769213), 'pulumi.get', 'pulumi.get', (['self', '"""schema_registry"""'], {}), "(self, 'schema_registry')\n", (769188, 769213), False, 'import pulumi\n'), ((769305, 769347), 'pulumi.set', 'pulumi.set', (['self', '"""schema_registry"""', 'value'], {}), "(self, 'schema_registry', value)\n", (769315, 769347), False, 'import pulumi\n'), ((769594, 769630), 'pulumi.get', 'pulumi.get', (['self', '"""unmanaged_topics"""'], {}), "(self, 'unmanaged_topics')\n", (769604, 769630), False, 'import pulumi\n'), ((769724, 769767), 'pulumi.set', 'pulumi.set', (['self', '"""unmanaged_topics"""', 'value'], {}), "(self, 'unmanaged_topics', value)\n", (769734, 769767), False, 'import pulumi\n'), ((770405, 770449), 'pulumi.set', 'pulumi.set', (['__self__', '"""resources"""', 'resources'], {}), "(__self__, 'resources', resources)\n", (770415, 770449), False, 'import pulumi\n'), ((770808, 770837), 'pulumi.get', 'pulumi.get', (['self', '"""resources"""'], {}), "(self, 'resources')\n", (770818, 770837), False, 'import pulumi\n'), ((770947, 770983), 'pulumi.set', 'pulumi.set', (['self', '"""resources"""', 'value'], {}), "(self, 'resources', value)\n", (770957, 770983), False, 'import pulumi\n'), ((771264, 771296), 'pulumi.get', 'pulumi.get', (['self', '"""kafka_config"""'], {}), "(self, 'kafka_config')\n", (771274, 771296), False, 'import pulumi\n'), ((771424, 771463), 'pulumi.set', 'pulumi.set', (['self', '"""kafka_config"""', 'value'], {}), "(self, 'kafka_config', value)\n", (771434, 771463), False, 'import pulumi\n'), ((774556, 774601), 'pulumi.get', 'pulumi.get', (['self', '"""auto_create_topics_enable"""'], {}), "(self, 'auto_create_topics_enable')\n", (774566, 774601), False, 'import pulumi\n'), ((774713, 774765), 'pulumi.set', 'pulumi.set', (['self', '"""auto_create_topics_enable"""', 'value'], {}), "(self, 'auto_create_topics_enable', value)\n", (774723, 774765), False, 'import pulumi\n'), ((774888, 774924), 'pulumi.get', 'pulumi.get', (['self', '"""compression_type"""'], {}), "(self, 'compression_type')\n", (774898, 774924), False, 'import pulumi\n'), ((775017, 775060), 'pulumi.set', 'pulumi.set', (['self', '"""compression_type"""', 'value'], {}), "(self, 'compression_type', value)\n", (775027, 775060), False, 'import pulumi\n'), ((775202, 775248), 'pulumi.get', 'pulumi.get', (['self', '"""default_replication_factor"""'], {}), "(self, 'default_replication_factor')\n", (775212, 775248), False, 'import pulumi\n'), ((775361, 775414), 'pulumi.set', 'pulumi.set', (['self', '"""default_replication_factor"""', 'value'], {}), "(self, 'default_replication_factor', value)\n", (775371, 775414), False, 'import pulumi\n'), ((775557, 775604), 'pulumi.get', 'pulumi.get', (['self', '"""log_flush_interval_messages"""'], {}), "(self, 'log_flush_interval_messages')\n", (775567, 775604), False, 'import pulumi\n'), ((775719, 775773), 'pulumi.set', 'pulumi.set', (['self', '"""log_flush_interval_messages"""', 'value'], {}), "(self, 'log_flush_interval_messages', value)\n", (775729, 775773), False, 'import pulumi\n'), ((775904, 775945), 'pulumi.get', 'pulumi.get', (['self', '"""log_flush_interval_ms"""'], {}), "(self, 'log_flush_interval_ms')\n", (775914, 775945), False, 'import pulumi\n'), ((776048, 776096), 'pulumi.set', 'pulumi.set', (['self', '"""log_flush_interval_ms"""', 'value'], {}), "(self, 'log_flush_interval_ms', value)\n", (776058, 776096), False, 'import pulumi\n'), ((776246, 776297), 'pulumi.get', 'pulumi.get', (['self', '"""log_flush_scheduler_interval_ms"""'], {}), "(self, 'log_flush_scheduler_interval_ms')\n", (776256, 776297), False, 'import pulumi\n'), ((776420, 776478), 'pulumi.set', 'pulumi.set', (['self', '"""log_flush_scheduler_interval_ms"""', 'value'], {}), "(self, 'log_flush_scheduler_interval_ms', value)\n", (776430, 776478), False, 'import pulumi\n'), ((776600, 776635), 'pulumi.get', 'pulumi.get', (['self', '"""log_preallocate"""'], {}), "(self, 'log_preallocate')\n", (776610, 776635), False, 'import pulumi\n'), ((776727, 776769), 'pulumi.set', 'pulumi.set', (['self', '"""log_preallocate"""', 'value'], {}), "(self, 'log_preallocate', value)\n", (776737, 776769), False, 'import pulumi\n'), ((776897, 776936), 'pulumi.get', 'pulumi.get', (['self', '"""log_retention_bytes"""'], {}), "(self, 'log_retention_bytes')\n", (776907, 776936), False, 'import pulumi\n'), ((777035, 777081), 'pulumi.set', 'pulumi.set', (['self', '"""log_retention_bytes"""', 'value'], {}), "(self, 'log_retention_bytes', value)\n", (777045, 777081), False, 'import pulumi\n'), ((777209, 777248), 'pulumi.get', 'pulumi.get', (['self', '"""log_retention_hours"""'], {}), "(self, 'log_retention_hours')\n", (777219, 777248), False, 'import pulumi\n'), ((777347, 777393), 'pulumi.set', 'pulumi.set', (['self', '"""log_retention_hours"""', 'value'], {}), "(self, 'log_retention_hours', value)\n", (777357, 777393), False, 'import pulumi\n'), ((777525, 777566), 'pulumi.get', 'pulumi.get', (['self', '"""log_retention_minutes"""'], {}), "(self, 'log_retention_minutes')\n", (777535, 777566), False, 'import pulumi\n'), ((777669, 777717), 'pulumi.set', 'pulumi.set', (['self', '"""log_retention_minutes"""', 'value'], {}), "(self, 'log_retention_minutes', value)\n", (777679, 777717), False, 'import pulumi\n'), ((777839, 777875), 'pulumi.get', 'pulumi.get', (['self', '"""log_retention_ms"""'], {}), "(self, 'log_retention_ms')\n", (777849, 777875), False, 'import pulumi\n'), ((777968, 778011), 'pulumi.set', 'pulumi.set', (['self', '"""log_retention_ms"""', 'value'], {}), "(self, 'log_retention_ms', value)\n", (777978, 778011), False, 'import pulumi\n'), ((778135, 778172), 'pulumi.get', 'pulumi.get', (['self', '"""log_segment_bytes"""'], {}), "(self, 'log_segment_bytes')\n", (778145, 778172), False, 'import pulumi\n'), ((778267, 778311), 'pulumi.set', 'pulumi.set', (['self', '"""log_segment_bytes"""', 'value'], {}), "(self, 'log_segment_bytes', value)\n", (778277, 778311), False, 'import pulumi\n'), ((778430, 778464), 'pulumi.get', 'pulumi.get', (['self', '"""num_partitions"""'], {}), "(self, 'num_partitions')\n", (778440, 778464), False, 'import pulumi\n'), ((778553, 778594), 'pulumi.set', 'pulumi.set', (['self', '"""num_partitions"""', 'value'], {}), "(self, 'num_partitions', value)\n", (778563, 778594), False, 'import pulumi\n'), ((778737, 778784), 'pulumi.get', 'pulumi.get', (['self', '"""socket_receive_buffer_bytes"""'], {}), "(self, 'socket_receive_buffer_bytes')\n", (778747, 778784), False, 'import pulumi\n'), ((778899, 778953), 'pulumi.set', 'pulumi.set', (['self', '"""socket_receive_buffer_bytes"""', 'value'], {}), "(self, 'socket_receive_buffer_bytes', value)\n", (778909, 778953), False, 'import pulumi\n'), ((779090, 779134), 'pulumi.get', 'pulumi.get', (['self', '"""socket_send_buffer_bytes"""'], {}), "(self, 'socket_send_buffer_bytes')\n", (779100, 779134), False, 'import pulumi\n'), ((779243, 779294), 'pulumi.set', 'pulumi.set', (['self', '"""socket_send_buffer_bytes"""', 'value'], {}), "(self, 'socket_send_buffer_bytes', value)\n", (779253, 779294), False, 'import pulumi\n'), ((779864, 779908), 'pulumi.set', 'pulumi.set', (['__self__', '"""disk_size"""', 'disk_size'], {}), "(__self__, 'disk_size', disk_size)\n", (779874, 779908), False, 'import pulumi\n'), ((779917, 779967), 'pulumi.set', 'pulumi.set', (['__self__', '"""disk_type_id"""', 'disk_type_id'], {}), "(__self__, 'disk_type_id', disk_type_id)\n", (779927, 779967), False, 'import pulumi\n'), ((779976, 780038), 'pulumi.set', 'pulumi.set', (['__self__', '"""resource_preset_id"""', 'resource_preset_id'], {}), "(__self__, 'resource_preset_id', resource_preset_id)\n", (779986, 780038), False, 'import pulumi\n'), ((780247, 780276), 'pulumi.get', 'pulumi.get', (['self', '"""disk_size"""'], {}), "(self, 'disk_size')\n", (780257, 780276), False, 'import pulumi\n'), ((780345, 780381), 'pulumi.set', 'pulumi.set', (['self', '"""disk_size"""', 'value'], {}), "(self, 'disk_size', value)\n", (780355, 780381), False, 'import pulumi\n'), ((780693, 780725), 'pulumi.get', 'pulumi.get', (['self', '"""disk_type_id"""'], {}), "(self, 'disk_type_id')\n", (780703, 780725), False, 'import pulumi\n'), ((780800, 780839), 'pulumi.set', 'pulumi.set', (['self', '"""disk_type_id"""', 'value'], {}), "(self, 'disk_type_id', value)\n", (780810, 780839), False, 'import pulumi\n'), ((780955, 780993), 'pulumi.get', 'pulumi.get', (['self', '"""resource_preset_id"""'], {}), "(self, 'resource_preset_id')\n", (780965, 780993), False, 'import pulumi\n'), ((781080, 781125), 'pulumi.set', 'pulumi.set', (['self', '"""resource_preset_id"""', 'value'], {}), "(self, 'resource_preset_id', value)\n", (781090, 781125), False, 'import pulumi\n'), ((781512, 781556), 'pulumi.set', 'pulumi.set', (['__self__', '"""resources"""', 'resources'], {}), "(__self__, 'resources', resources)\n", (781522, 781556), False, 'import pulumi\n'), ((781819, 781848), 'pulumi.get', 'pulumi.get', (['self', '"""resources"""'], {}), "(self, 'resources')\n", (781829, 781848), False, 'import pulumi\n'), ((781962, 781998), 'pulumi.set', 'pulumi.set', (['self', '"""resources"""', 'value'], {}), "(self, 'resources', value)\n", (781972, 781998), False, 'import pulumi\n'), ((782572, 782616), 'pulumi.set', 'pulumi.set', (['__self__', '"""disk_size"""', 'disk_size'], {}), "(__self__, 'disk_size', disk_size)\n", (782582, 782616), False, 'import pulumi\n'), ((782625, 782675), 'pulumi.set', 'pulumi.set', (['__self__', '"""disk_type_id"""', 'disk_type_id'], {}), "(__self__, 'disk_type_id', disk_type_id)\n", (782635, 782675), False, 'import pulumi\n'), ((782684, 782746), 'pulumi.set', 'pulumi.set', (['__self__', '"""resource_preset_id"""', 'resource_preset_id'], {}), "(__self__, 'resource_preset_id', resource_preset_id)\n", (782694, 782746), False, 'import pulumi\n'), ((782955, 782984), 'pulumi.get', 'pulumi.get', (['self', '"""disk_size"""'], {}), "(self, 'disk_size')\n", (782965, 782984), False, 'import pulumi\n'), ((783053, 783089), 'pulumi.set', 'pulumi.set', (['self', '"""disk_size"""', 'value'], {}), "(self, 'disk_size', value)\n", (783063, 783089), False, 'import pulumi\n'), ((783401, 783433), 'pulumi.get', 'pulumi.get', (['self', '"""disk_type_id"""'], {}), "(self, 'disk_type_id')\n", (783411, 783433), False, 'import pulumi\n'), ((783508, 783547), 'pulumi.set', 'pulumi.set', (['self', '"""disk_type_id"""', 'value'], {}), "(self, 'disk_type_id', value)\n", (783518, 783547), False, 'import pulumi\n'), ((783663, 783701), 'pulumi.get', 'pulumi.get', (['self', '"""resource_preset_id"""'], {}), "(self, 'resource_preset_id')\n", (783673, 783701), False, 'import pulumi\n'), ((783788, 783833), 'pulumi.set', 'pulumi.set', (['self', '"""resource_preset_id"""', 'value'], {}), "(self, 'resource_preset_id', value)\n", (783798, 783833), False, 'import pulumi\n'), ((784643, 784689), 'pulumi.set', 'pulumi.set', (['__self__', '"""cluster_id"""', 'cluster_id'], {}), "(__self__, 'cluster_id', cluster_id)\n", (784653, 784689), False, 'import pulumi\n'), ((784698, 784732), 'pulumi.set', 'pulumi.set', (['__self__', '"""name"""', 'name'], {}), "(__self__, 'name', name)\n", (784708, 784732), False, 'import pulumi\n'), ((784741, 784787), 'pulumi.set', 'pulumi.set', (['__self__', '"""partitions"""', 'partitions'], {}), "(__self__, 'partitions', partitions)\n", (784751, 784787), False, 'import pulumi\n'), ((784796, 784858), 'pulumi.set', 'pulumi.set', (['__self__', '"""replication_factor"""', 'replication_factor'], {}), "(__self__, 'replication_factor', replication_factor)\n", (784806, 784858), False, 'import pulumi\n'), ((785120, 785150), 'pulumi.get', 'pulumi.get', (['self', '"""cluster_id"""'], {}), "(self, 'cluster_id')\n", (785130, 785150), False, 'import pulumi\n'), ((785221, 785258), 'pulumi.set', 'pulumi.set', (['self', '"""cluster_id"""', 'value'], {}), "(self, 'cluster_id', value)\n", (785231, 785258), False, 'import pulumi\n'), ((785398, 785422), 'pulumi.get', 'pulumi.get', (['self', '"""name"""'], {}), "(self, 'name')\n", (785408, 785422), False, 'import pulumi\n'), ((785481, 785512), 'pulumi.set', 'pulumi.set', (['self', '"""name"""', 'value'], {}), "(self, 'name', value)\n", (785491, 785512), False, 'import pulumi\n'), ((785676, 785706), 'pulumi.get', 'pulumi.get', (['self', '"""partitions"""'], {}), "(self, 'partitions')\n", (785686, 785706), False, 'import pulumi\n'), ((785777, 785814), 'pulumi.set', 'pulumi.set', (['self', '"""partitions"""', 'value'], {}), "(self, 'partitions', value)\n", (785787, 785814), False, 'import pulumi\n'), ((786037, 786075), 'pulumi.get', 'pulumi.get', (['self', '"""replication_factor"""'], {}), "(self, 'replication_factor')\n", (786047, 786075), False, 'import pulumi\n'), ((786162, 786207), 'pulumi.set', 'pulumi.set', (['self', '"""replication_factor"""', 'value'], {}), "(self, 'replication_factor', value)\n", (786172, 786207), False, 'import pulumi\n'), ((786474, 786506), 'pulumi.get', 'pulumi.get', (['self', '"""topic_config"""'], {}), "(self, 'topic_config')\n", (786484, 786506), False, 'import pulumi\n'), ((786628, 786667), 'pulumi.set', 'pulumi.set', (['self', '"""topic_config"""', 'value'], {}), "(self, 'topic_config', value)\n", (786638, 786667), False, 'import pulumi\n'), ((789046, 789080), 'pulumi.get', 'pulumi.get', (['self', '"""cleanup_policy"""'], {}), "(self, 'cleanup_policy')\n", (789056, 789080), False, 'import pulumi\n'), ((789169, 789210), 'pulumi.set', 'pulumi.set', (['self', '"""cleanup_policy"""', 'value'], {}), "(self, 'cleanup_policy', value)\n", (789179, 789210), False, 'import pulumi\n'), ((789333, 789369), 'pulumi.get', 'pulumi.get', (['self', '"""compression_type"""'], {}), "(self, 'compression_type')\n", (789343, 789369), False, 'import pulumi\n'), ((789462, 789505), 'pulumi.set', 'pulumi.set', (['self', '"""compression_type"""', 'value'], {}), "(self, 'compression_type', value)\n", (789472, 789505), False, 'import pulumi\n'), ((789633, 789672), 'pulumi.get', 'pulumi.get', (['self', '"""delete_retention_ms"""'], {}), "(self, 'delete_retention_ms')\n", (789643, 789672), False, 'import pulumi\n'), ((789771, 789817), 'pulumi.set', 'pulumi.set', (['self', '"""delete_retention_ms"""', 'value'], {}), "(self, 'delete_retention_ms', value)\n", (789781, 789817), False, 'import pulumi\n'), ((789946, 789986), 'pulumi.get', 'pulumi.get', (['self', '"""file_delete_delay_ms"""'], {}), "(self, 'file_delete_delay_ms')\n", (789956, 789986), False, 'import pulumi\n'), ((790087, 790134), 'pulumi.set', 'pulumi.set', (['self', '"""file_delete_delay_ms"""', 'value'], {}), "(self, 'file_delete_delay_ms', value)\n", (790097, 790134), False, 'import pulumi\n'), ((790253, 790287), 'pulumi.get', 'pulumi.get', (['self', '"""flush_messages"""'], {}), "(self, 'flush_messages')\n", (790263, 790287), False, 'import pulumi\n'), ((790376, 790417), 'pulumi.set', 'pulumi.set', (['self', '"""flush_messages"""', 'value'], {}), "(self, 'flush_messages', value)\n", (790386, 790417), False, 'import pulumi\n'), ((790524, 790552), 'pulumi.get', 'pulumi.get', (['self', '"""flush_ms"""'], {}), "(self, 'flush_ms')\n", (790534, 790552), False, 'import pulumi\n'), ((790629, 790664), 'pulumi.set', 'pulumi.set', (['self', '"""flush_ms"""', 'value'], {}), "(self, 'flush_ms', value)\n", (790639, 790664), False, 'import pulumi\n'), ((790788, 790825), 'pulumi.get', 'pulumi.get', (['self', '"""max_message_bytes"""'], {}), "(self, 'max_message_bytes')\n", (790798, 790825), False, 'import pulumi\n'), ((790920, 790964), 'pulumi.set', 'pulumi.set', (['self', '"""max_message_bytes"""', 'value'], {}), "(self, 'max_message_bytes', value)\n", (790930, 790964), False, 'import pulumi\n'), ((791095, 791136), 'pulumi.get', 'pulumi.get', (['self', '"""min_compaction_lag_ms"""'], {}), "(self, 'min_compaction_lag_ms')\n", (791105, 791136), False, 'import pulumi\n'), ((791239, 791287), 'pulumi.set', 'pulumi.set', (['self', '"""min_compaction_lag_ms"""', 'value'], {}), "(self, 'min_compaction_lag_ms', value)\n", (791249, 791287), False, 'import pulumi\n'), ((791415, 791454), 'pulumi.get', 'pulumi.get', (['self', '"""min_insync_replicas"""'], {}), "(self, 'min_insync_replicas')\n", (791425, 791454), False, 'import pulumi\n'), ((791553, 791599), 'pulumi.set', 'pulumi.set', (['self', '"""min_insync_replicas"""', 'value'], {}), "(self, 'min_insync_replicas', value)\n", (791563, 791599), False, 'import pulumi\n'), ((791694, 791725), 'pulumi.get', 'pulumi.get', (['self', '"""preallocate"""'], {}), "(self, 'preallocate')\n", (791704, 791725), False, 'import pulumi\n'), ((791809, 791847), 'pulumi.set', 'pulumi.set', (['self', '"""preallocate"""', 'value'], {}), "(self, 'preallocate', value)\n", (791819, 791847), False, 'import pulumi\n'), ((791968, 792003), 'pulumi.get', 'pulumi.get', (['self', '"""retention_bytes"""'], {}), "(self, 'retention_bytes')\n", (791978, 792003), False, 'import pulumi\n'), ((792094, 792136), 'pulumi.set', 'pulumi.set', (['self', '"""retention_bytes"""', 'value'], {}), "(self, 'retention_bytes', value)\n", (792104, 792136), False, 'import pulumi\n'), ((792251, 792283), 'pulumi.get', 'pulumi.get', (['self', '"""retention_ms"""'], {}), "(self, 'retention_ms')\n", (792261, 792283), False, 'import pulumi\n'), ((792368, 792407), 'pulumi.set', 'pulumi.set', (['self', '"""retention_ms"""', 'value'], {}), "(self, 'retention_ms', value)\n", (792378, 792407), False, 'import pulumi\n'), ((792524, 792557), 'pulumi.get', 'pulumi.get', (['self', '"""segment_bytes"""'], {}), "(self, 'segment_bytes')\n", (792534, 792557), False, 'import pulumi\n'), ((792644, 792684), 'pulumi.set', 'pulumi.set', (['self', '"""segment_bytes"""', 'value'], {}), "(self, 'segment_bytes', value)\n", (792654, 792684), False, 'import pulumi\n'), ((793245, 793279), 'pulumi.set', 'pulumi.set', (['__self__', '"""name"""', 'name'], {}), "(__self__, 'name', name)\n", (793255, 793279), False, 'import pulumi\n'), ((793288, 793330), 'pulumi.set', 'pulumi.set', (['__self__', '"""password"""', 'password'], {}), "(__self__, 'password', password)\n", (793298, 793330), False, 'import pulumi\n'), ((793567, 793591), 'pulumi.get', 'pulumi.get', (['self', '"""name"""'], {}), "(self, 'name')\n", (793577, 793591), False, 'import pulumi\n'), ((793650, 793681), 'pulumi.set', 'pulumi.set', (['self', '"""name"""', 'value'], {}), "(self, 'name', value)\n", (793660, 793681), False, 'import pulumi\n'), ((793831, 793859), 'pulumi.get', 'pulumi.get', (['self', '"""password"""'], {}), "(self, 'password')\n", (793841, 793859), False, 'import pulumi\n'), ((793926, 793961), 'pulumi.set', 'pulumi.set', (['self', '"""password"""', 'value'], {}), "(self, 'password', value)\n", (793936, 793961), False, 'import pulumi\n'), ((794218, 794249), 'pulumi.get', 'pulumi.get', (['self', '"""permissions"""'], {}), "(self, 'permissions')\n", (794228, 794249), False, 'import pulumi\n'), ((794377, 794415), 'pulumi.set', 'pulumi.set', (['self', '"""permissions"""', 'value'], {}), "(self, 'permissions', value)\n", (794387, 794415), False, 'import pulumi\n'), ((794766, 794800), 'pulumi.set', 'pulumi.set', (['__self__', '"""role"""', 'role'], {}), "(__self__, 'role', role)\n", (794776, 794800), False, 'import pulumi\n'), ((794809, 794855), 'pulumi.set', 'pulumi.set', (['__self__', '"""topic_name"""', 'topic_name'], {}), "(__self__, 'topic_name', topic_name)\n", (794819, 794855), False, 'import pulumi\n'), ((794997, 795021), 'pulumi.get', 'pulumi.get', (['self', '"""role"""'], {}), "(self, 'role')\n", (795007, 795021), False, 'import pulumi\n'), ((795080, 795111), 'pulumi.set', 'pulumi.set', (['self', '"""role"""', 'value'], {}), "(self, 'role', value)\n", (795090, 795111), False, 'import pulumi\n'), ((795315, 795345), 'pulumi.get', 'pulumi.get', (['self', '"""topic_name"""'], {}), "(self, 'topic_name')\n", (795325, 795345), False, 'import pulumi\n'), ((795416, 795453), 'pulumi.set', 'pulumi.set', (['self', '"""topic_name"""', 'value'], {}), "(self, 'topic_name', value)\n", (795426, 795453), False, 'import pulumi\n'), ((795909, 795953), 'pulumi.set', 'pulumi.set', (['__self__', '"""data_lens"""', 'data_lens'], {}), "(__self__, 'data_lens', data_lens)\n", (795919, 795953), False, 'import pulumi\n'), ((795962, 796002), 'pulumi.set', 'pulumi.set', (['__self__', '"""web_sql"""', 'web_sql'], {}), "(__self__, 'web_sql', web_sql)\n", (795972, 796002), False, 'import pulumi\n'), ((796214, 796243), 'pulumi.get', 'pulumi.get', (['self', '"""data_lens"""'], {}), "(self, 'data_lens')\n", (796224, 796243), False, 'import pulumi\n'), ((796313, 796349), 'pulumi.set', 'pulumi.set', (['self', '"""data_lens"""', 'value'], {}), "(self, 'data_lens', value)\n", (796323, 796349), False, 'import pulumi\n'), ((796606, 796633), 'pulumi.get', 'pulumi.get', (['self', '"""web_sql"""'], {}), "(self, 'web_sql')\n", (796616, 796633), False, 'import pulumi\n'), ((796699, 796733), 'pulumi.set', 'pulumi.set', (['self', '"""web_sql"""', 'value'], {}), "(self, 'web_sql', value)\n", (796709, 796733), False, 'import pulumi\n'), ((15698, 15746), 'pulumi.set', 'pulumi.set', (['__self__', '"""healthcheck"""', 'healthcheck'], {}), "(__self__, 'healthcheck', healthcheck)\n", (15708, 15746), False, 'import pulumi\n'), ((15805, 15873), 'pulumi.set', 'pulumi.set', (['__self__', '"""load_balancing_config"""', 'load_balancing_config'], {}), "(__self__, 'load_balancing_config', load_balancing_config)\n", (15815, 15873), False, 'import pulumi\n'), ((15915, 15949), 'pulumi.set', 'pulumi.set', (['__self__', '"""port"""', 'port'], {}), "(__self__, 'port', port)\n", (15925, 15949), False, 'import pulumi\n'), ((15990, 16022), 'pulumi.set', 'pulumi.set', (['__self__', '"""tls"""', 'tls'], {}), "(__self__, 'tls', tls)\n", (16000, 16022), False, 'import pulumi\n'), ((16066, 16104), 'pulumi.set', 'pulumi.set', (['__self__', '"""weight"""', 'weight'], {}), "(__self__, 'weight', weight)\n", (16076, 16104), False, 'import pulumi\n'), ((21649, 21707), 'pulumi.set', 'pulumi.set', (['__self__', '"""grpc_healthcheck"""', 'grpc_healthcheck'], {}), "(__self__, 'grpc_healthcheck', grpc_healthcheck)\n", (21659, 21707), False, 'import pulumi\n'), ((21761, 21819), 'pulumi.set', 'pulumi.set', (['__self__', '"""healthcheck_port"""', 'healthcheck_port'], {}), "(__self__, 'healthcheck_port', healthcheck_port)\n", (21771, 21819), False, 'import pulumi\n'), ((21874, 21934), 'pulumi.set', 'pulumi.set', (['__self__', '"""healthy_threshold"""', 'healthy_threshold'], {}), "(__self__, 'healthy_threshold', healthy_threshold)\n", (21884, 21934), False, 'import pulumi\n'), ((21988, 22046), 'pulumi.set', 'pulumi.set', (['__self__', '"""http_healthcheck"""', 'http_healthcheck'], {}), "(__self__, 'http_healthcheck', http_healthcheck)\n", (21998, 22046), False, 'import pulumi\n'), ((22107, 22179), 'pulumi.set', 'pulumi.set', (['__self__', '"""interval_jitter_percent"""', 'interval_jitter_percent'], {}), "(__self__, 'interval_jitter_percent', interval_jitter_percent)\n", (22117, 22179), False, 'import pulumi\n'), ((22235, 22297), 'pulumi.set', 'pulumi.set', (['__self__', '"""stream_healthcheck"""', 'stream_healthcheck'], {}), "(__self__, 'stream_healthcheck', stream_healthcheck)\n", (22245, 22297), False, 'import pulumi\n'), ((22354, 22418), 'pulumi.set', 'pulumi.set', (['__self__', '"""unhealthy_threshold"""', 'unhealthy_threshold'], {}), "(__self__, 'unhealthy_threshold', unhealthy_threshold)\n", (22364, 22418), False, 'import pulumi\n'), ((27227, 27277), 'pulumi.set', 'pulumi.set', (['__self__', '"""service_name"""', 'service_name'], {}), "(__self__, 'service_name', service_name)\n", (27237, 27277), False, 'import pulumi\n'), ((28249, 28283), 'pulumi.set', 'pulumi.set', (['__self__', '"""host"""', 'host'], {}), "(__self__, 'host', host)\n", (28259, 28283), False, 'import pulumi\n'), ((28326, 28362), 'pulumi.set', 'pulumi.set', (['__self__', '"""http2"""', 'http2'], {}), "(__self__, 'http2', http2)\n", (28336, 28362), False, 'import pulumi\n'), ((29712, 29752), 'pulumi.set', 'pulumi.set', (['__self__', '"""receive"""', 'receive'], {}), "(__self__, 'receive', receive)\n", (29722, 29752), False, 'import pulumi\n'), ((29794, 29828), 'pulumi.set', 'pulumi.set', (['__self__', '"""send"""', 'send'], {}), "(__self__, 'send', send)\n", (29804, 29828), False, 'import pulumi\n'), ((31553, 31643), 'pulumi.set', 'pulumi.set', (['__self__', '"""locality_aware_routing_percent"""', 'locality_aware_routing_percent'], {}), "(__self__, 'locality_aware_routing_percent',\n locality_aware_routing_percent)\n", (31563, 31643), False, 'import pulumi\n'), ((31692, 31748), 'pulumi.set', 'pulumi.set', (['__self__', '"""panic_threshold"""', 'panic_threshold'], {}), "(__self__, 'panic_threshold', panic_threshold)\n", (31702, 31748), False, 'import pulumi\n'), ((31801, 31857), 'pulumi.set', 'pulumi.set', (['__self__', '"""strict_locality"""', 'strict_locality'], {}), "(__self__, 'strict_locality', strict_locality)\n", (31811, 31857), False, 'import pulumi\n'), ((34180, 34212), 'pulumi.set', 'pulumi.set', (['__self__', '"""sni"""', 'sni'], {}), "(__self__, 'sni', sni)\n", (34190, 34212), False, 'import pulumi\n'), ((34268, 34330), 'pulumi.set', 'pulumi.set', (['__self__', '"""validation_context"""', 'validation_context'], {}), "(__self__, 'validation_context', validation_context)\n", (34278, 34330), False, 'import pulumi\n'), ((35639, 35697), 'pulumi.set', 'pulumi.set', (['__self__', '"""trusted_ca_bytes"""', 'trusted_ca_bytes'], {}), "(__self__, 'trusted_ca_bytes', trusted_ca_bytes)\n", (35649, 35697), False, 'import pulumi\n'), ((35748, 35800), 'pulumi.set', 'pulumi.set', (['__self__', '"""trusted_ca_id"""', 'trusted_ca_id'], {}), "(__self__, 'trusted_ca_id', trusted_ca_id)\n", (35758, 35800), False, 'import pulumi\n'), ((38336, 38384), 'pulumi.set', 'pulumi.set', (['__self__', '"""healthcheck"""', 'healthcheck'], {}), "(__self__, 'healthcheck', healthcheck)\n", (38346, 38384), False, 'import pulumi\n'), ((38427, 38463), 'pulumi.set', 'pulumi.set', (['__self__', '"""http2"""', 'http2'], {}), "(__self__, 'http2', http2)\n", (38437, 38463), False, 'import pulumi\n'), ((38522, 38590), 'pulumi.set', 'pulumi.set', (['__self__', '"""load_balancing_config"""', 'load_balancing_config'], {}), "(__self__, 'load_balancing_config', load_balancing_config)\n", (38532, 38590), False, 'import pulumi\n'), ((38632, 38666), 'pulumi.set', 'pulumi.set', (['__self__', '"""port"""', 'port'], {}), "(__self__, 'port', port)\n", (38642, 38666), False, 'import pulumi\n'), ((38707, 38739), 'pulumi.set', 'pulumi.set', (['__self__', '"""tls"""', 'tls'], {}), "(__self__, 'tls', tls)\n", (38717, 38739), False, 'import pulumi\n'), ((38783, 38821), 'pulumi.set', 'pulumi.set', (['__self__', '"""weight"""', 'weight'], {}), "(__self__, 'weight', weight)\n", (38793, 38821), False, 'import pulumi\n'), ((44682, 44740), 'pulumi.set', 'pulumi.set', (['__self__', '"""grpc_healthcheck"""', 'grpc_healthcheck'], {}), "(__self__, 'grpc_healthcheck', grpc_healthcheck)\n", (44692, 44740), False, 'import pulumi\n'), ((44794, 44852), 'pulumi.set', 'pulumi.set', (['__self__', '"""healthcheck_port"""', 'healthcheck_port'], {}), "(__self__, 'healthcheck_port', healthcheck_port)\n", (44804, 44852), False, 'import pulumi\n'), ((44907, 44967), 'pulumi.set', 'pulumi.set', (['__self__', '"""healthy_threshold"""', 'healthy_threshold'], {}), "(__self__, 'healthy_threshold', healthy_threshold)\n", (44917, 44967), False, 'import pulumi\n'), ((45021, 45079), 'pulumi.set', 'pulumi.set', (['__self__', '"""http_healthcheck"""', 'http_healthcheck'], {}), "(__self__, 'http_healthcheck', http_healthcheck)\n", (45031, 45079), False, 'import pulumi\n'), ((45140, 45212), 'pulumi.set', 'pulumi.set', (['__self__', '"""interval_jitter_percent"""', 'interval_jitter_percent'], {}), "(__self__, 'interval_jitter_percent', interval_jitter_percent)\n", (45150, 45212), False, 'import pulumi\n'), ((45268, 45330), 'pulumi.set', 'pulumi.set', (['__self__', '"""stream_healthcheck"""', 'stream_healthcheck'], {}), "(__self__, 'stream_healthcheck', stream_healthcheck)\n", (45278, 45330), False, 'import pulumi\n'), ((45387, 45451), 'pulumi.set', 'pulumi.set', (['__self__', '"""unhealthy_threshold"""', 'unhealthy_threshold'], {}), "(__self__, 'unhealthy_threshold', unhealthy_threshold)\n", (45397, 45451), False, 'import pulumi\n'), ((50260, 50310), 'pulumi.set', 'pulumi.set', (['__self__', '"""service_name"""', 'service_name'], {}), "(__self__, 'service_name', service_name)\n", (50270, 50310), False, 'import pulumi\n'), ((51282, 51316), 'pulumi.set', 'pulumi.set', (['__self__', '"""host"""', 'host'], {}), "(__self__, 'host', host)\n", (51292, 51316), False, 'import pulumi\n'), ((51359, 51395), 'pulumi.set', 'pulumi.set', (['__self__', '"""http2"""', 'http2'], {}), "(__self__, 'http2', http2)\n", (51369, 51395), False, 'import pulumi\n'), ((52745, 52785), 'pulumi.set', 'pulumi.set', (['__self__', '"""receive"""', 'receive'], {}), "(__self__, 'receive', receive)\n", (52755, 52785), False, 'import pulumi\n'), ((52827, 52861), 'pulumi.set', 'pulumi.set', (['__self__', '"""send"""', 'send'], {}), "(__self__, 'send', send)\n", (52837, 52861), False, 'import pulumi\n'), ((54586, 54676), 'pulumi.set', 'pulumi.set', (['__self__', '"""locality_aware_routing_percent"""', 'locality_aware_routing_percent'], {}), "(__self__, 'locality_aware_routing_percent',\n locality_aware_routing_percent)\n", (54596, 54676), False, 'import pulumi\n'), ((54725, 54781), 'pulumi.set', 'pulumi.set', (['__self__', '"""panic_threshold"""', 'panic_threshold'], {}), "(__self__, 'panic_threshold', panic_threshold)\n", (54735, 54781), False, 'import pulumi\n'), ((54834, 54890), 'pulumi.set', 'pulumi.set', (['__self__', '"""strict_locality"""', 'strict_locality'], {}), "(__self__, 'strict_locality', strict_locality)\n", (54844, 54890), False, 'import pulumi\n'), ((57213, 57245), 'pulumi.set', 'pulumi.set', (['__self__', '"""sni"""', 'sni'], {}), "(__self__, 'sni', sni)\n", (57223, 57245), False, 'import pulumi\n'), ((57301, 57363), 'pulumi.set', 'pulumi.set', (['__self__', '"""validation_context"""', 'validation_context'], {}), "(__self__, 'validation_context', validation_context)\n", (57311, 57363), False, 'import pulumi\n'), ((58672, 58730), 'pulumi.set', 'pulumi.set', (['__self__', '"""trusted_ca_bytes"""', 'trusted_ca_bytes'], {}), "(__self__, 'trusted_ca_bytes', trusted_ca_bytes)\n", (58682, 58730), False, 'import pulumi\n'), ((58781, 58833), 'pulumi.set', 'pulumi.set', (['__self__', '"""trusted_ca_id"""', 'trusted_ca_id'], {}), "(__self__, 'trusted_ca_id', trusted_ca_id)\n", (58791, 58833), False, 'import pulumi\n'), ((61130, 61186), 'pulumi.set', 'pulumi.set', (['__self__', '"""disable_traffic"""', 'disable_traffic'], {}), "(__self__, 'disable_traffic', disable_traffic)\n", (61140, 61186), False, 'import pulumi\n'), ((63323, 63367), 'pulumi.set', 'pulumi.set', (['__self__', '"""endpoints"""', 'endpoints'], {}), "(__self__, 'endpoints', endpoints)\n", (63333, 63367), False, 'import pulumi\n'), ((63409, 63443), 'pulumi.set', 'pulumi.set', (['__self__', '"""http"""', 'http'], {}), "(__self__, 'http', http)\n", (63419, 63443), False, 'import pulumi\n'), ((63484, 63516), 'pulumi.set', 'pulumi.set', (['__self__', '"""tls"""', 'tls'], {}), "(__self__, 'tls', tls)\n", (63494, 63516), False, 'import pulumi\n'), ((67620, 67688), 'pulumi.set', 'pulumi.set', (['__self__', '"""external_ipv4_address"""', 'external_ipv4_address'], {}), "(__self__, 'external_ipv4_address', external_ipv4_address)\n", (67630, 67688), False, 'import pulumi\n'), ((67747, 67815), 'pulumi.set', 'pulumi.set', (['__self__', '"""external_ipv6_address"""', 'external_ipv6_address'], {}), "(__self__, 'external_ipv6_address', external_ipv6_address)\n", (67757, 67815), False, 'import pulumi\n'), ((67874, 67942), 'pulumi.set', 'pulumi.set', (['__self__', '"""internal_ipv4_address"""', 'internal_ipv4_address'], {}), "(__self__, 'internal_ipv4_address', internal_ipv4_address)\n", (67884, 67942), False, 'import pulumi\n'), ((69972, 70012), 'pulumi.set', 'pulumi.set', (['__self__', '"""address"""', 'address'], {}), "(__self__, 'address', address)\n", (69982, 70012), False, 'import pulumi\n'), ((70692, 70732), 'pulumi.set', 'pulumi.set', (['__self__', '"""address"""', 'address'], {}), "(__self__, 'address', address)\n", (70702, 70732), False, 'import pulumi\n'), ((71570, 71610), 'pulumi.set', 'pulumi.set', (['__self__', '"""address"""', 'address'], {}), "(__self__, 'address', address)\n", (71580, 71610), False, 'import pulumi\n'), ((71657, 71701), 'pulumi.set', 'pulumi.set', (['__self__', '"""subnet_id"""', 'subnet_id'], {}), "(__self__, 'subnet_id', subnet_id)\n", (71667, 71701), False, 'import pulumi\n'), ((73084, 73124), 'pulumi.set', 'pulumi.set', (['__self__', '"""handler"""', 'handler'], {}), "(__self__, 'handler', handler)\n", (73094, 73124), False, 'import pulumi\n'), ((73171, 73215), 'pulumi.set', 'pulumi.set', (['__self__', '"""redirects"""', 'redirects'], {}), "(__self__, 'redirects', redirects)\n", (73181, 73215), False, 'import pulumi\n'), ((74894, 74944), 'pulumi.set', 'pulumi.set', (['__self__', '"""allow_http10"""', 'allow_http10'], {}), "(__self__, 'allow_http10', allow_http10)\n", (74904, 74944), False, 'import pulumi\n'), ((74995, 75047), 'pulumi.set', 'pulumi.set', (['__self__', '"""http2_options"""', 'http2_options'], {}), "(__self__, 'http2_options', http2_options)\n", (75005, 75047), False, 'import pulumi\n'), ((75099, 75153), 'pulumi.set', 'pulumi.set', (['__self__', '"""http_router_id"""', 'http_router_id'], {}), "(__self__, 'http_router_id', http_router_id)\n", (75109, 75153), False, 'import pulumi\n'), ((76794, 76864), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_concurrent_streams"""', 'max_concurrent_streams'], {}), "(__self__, 'max_concurrent_streams', max_concurrent_streams)\n", (76804, 76864), False, 'import pulumi\n'), ((77512, 77564), 'pulumi.set', 'pulumi.set', (['__self__', '"""http_to_https"""', 'http_to_https'], {}), "(__self__, 'http_to_https', http_to_https)\n", (77522, 77564), False, 'import pulumi\n'), ((78640, 78690), 'pulumi.set', 'pulumi.set', (['__self__', '"""sni_handlers"""', 'sni_handlers'], {}), "(__self__, 'sni_handlers', sni_handlers)\n", (78650, 78690), False, 'import pulumi\n'), ((80569, 80619), 'pulumi.set', 'pulumi.set', (['__self__', '"""http_handler"""', 'http_handler'], {}), "(__self__, 'http_handler', http_handler)\n", (80579, 80619), False, 'import pulumi\n'), ((82543, 82593), 'pulumi.set', 'pulumi.set', (['__self__', '"""allow_http10"""', 'allow_http10'], {}), "(__self__, 'allow_http10', allow_http10)\n", (82553, 82593), False, 'import pulumi\n'), ((82644, 82696), 'pulumi.set', 'pulumi.set', (['__self__', '"""http2_options"""', 'http2_options'], {}), "(__self__, 'http2_options', http2_options)\n", (82654, 82696), False, 'import pulumi\n'), ((82748, 82802), 'pulumi.set', 'pulumi.set', (['__self__', '"""http_router_id"""', 'http_router_id'], {}), "(__self__, 'http_router_id', http_router_id)\n", (82758, 82802), False, 'import pulumi\n'), ((84494, 84564), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_concurrent_streams"""', 'max_concurrent_streams'], {}), "(__self__, 'max_concurrent_streams', max_concurrent_streams)\n", (84504, 84564), False, 'import pulumi\n'), ((87790, 87840), 'pulumi.set', 'pulumi.set', (['__self__', '"""http_handler"""', 'http_handler'], {}), "(__self__, 'http_handler', http_handler)\n", (87800, 87840), False, 'import pulumi\n'), ((89779, 89829), 'pulumi.set', 'pulumi.set', (['__self__', '"""allow_http10"""', 'allow_http10'], {}), "(__self__, 'allow_http10', allow_http10)\n", (89789, 89829), False, 'import pulumi\n'), ((89880, 89932), 'pulumi.set', 'pulumi.set', (['__self__', '"""http2_options"""', 'http2_options'], {}), "(__self__, 'http2_options', http2_options)\n", (89890, 89932), False, 'import pulumi\n'), ((89984, 90038), 'pulumi.set', 'pulumi.set', (['__self__', '"""http_router_id"""', 'http_router_id'], {}), "(__self__, 'http_router_id', http_router_id)\n", (89994, 90038), False, 'import pulumi\n'), ((91739, 91809), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_concurrent_streams"""', 'max_concurrent_streams'], {}), "(__self__, 'max_concurrent_streams', max_concurrent_streams)\n", (91749, 91809), False, 'import pulumi\n'), ((94470, 94508), 'pulumi.set', 'pulumi.set', (['__self__', '"""append"""', 'append'], {}), "(__self__, 'append', append)\n", (94480, 94508), False, 'import pulumi\n'), ((94550, 94584), 'pulumi.set', 'pulumi.set', (['__self__', '"""name"""', 'name'], {}), "(__self__, 'name', name)\n", (94560, 94584), False, 'import pulumi\n'), ((94628, 94666), 'pulumi.set', 'pulumi.set', (['__self__', '"""remove"""', 'remove'], {}), "(__self__, 'remove', remove)\n", (94638, 94666), False, 'import pulumi\n'), ((94711, 94751), 'pulumi.set', 'pulumi.set', (['__self__', '"""replace"""', 'replace'], {}), "(__self__, 'replace', replace)\n", (94721, 94751), False, 'import pulumi\n'), ((97017, 97055), 'pulumi.set', 'pulumi.set', (['__self__', '"""append"""', 'append'], {}), "(__self__, 'append', append)\n", (97027, 97055), False, 'import pulumi\n'), ((97097, 97131), 'pulumi.set', 'pulumi.set', (['__self__', '"""name"""', 'name'], {}), "(__self__, 'name', name)\n", (97107, 97131), False, 'import pulumi\n'), ((97175, 97213), 'pulumi.set', 'pulumi.set', (['__self__', '"""remove"""', 'remove'], {}), "(__self__, 'remove', remove)\n", (97185, 97213), False, 'import pulumi\n'), ((97258, 97298), 'pulumi.set', 'pulumi.set', (['__self__', '"""replace"""', 'replace'], {}), "(__self__, 'replace', replace)\n", (97268, 97298), False, 'import pulumi\n'), ((99429, 99475), 'pulumi.set', 'pulumi.set', (['__self__', '"""grpc_route"""', 'grpc_route'], {}), "(__self__, 'grpc_route', grpc_route)\n", (99439, 99475), False, 'import pulumi\n'), ((99523, 99569), 'pulumi.set', 'pulumi.set', (['__self__', '"""http_route"""', 'http_route'], {}), "(__self__, 'http_route', http_route)\n", (99533, 99569), False, 'import pulumi\n'), ((99611, 99645), 'pulumi.set', 'pulumi.set', (['__self__', '"""name"""', 'name'], {}), "(__self__, 'name', name)\n", (99621, 99645), False, 'import pulumi\n'), ((101888, 101938), 'pulumi.set', 'pulumi.set', (['__self__', '"""grpc_matches"""', 'grpc_matches'], {}), "(__self__, 'grpc_matches', grpc_matches)\n", (101898, 101938), False, 'import pulumi\n'), ((101993, 102053), 'pulumi.set', 'pulumi.set', (['__self__', '"""grpc_route_action"""', 'grpc_route_action'], {}), "(__self__, 'grpc_route_action', grpc_route_action)\n", (102003, 102053), False, 'import pulumi\n'), ((102118, 102203), 'pulumi.set', 'pulumi.set', (['__self__', '"""grpc_status_response_action"""', 'grpc_status_response_action'], {}), "(__self__, 'grpc_status_response_action', grpc_status_response_action\n )\n", (102128, 102203), False, 'import pulumi\n'), ((104273, 104307), 'pulumi.set', 'pulumi.set', (['__self__', '"""fqmn"""', 'fqmn'], {}), "(__self__, 'fqmn', fqmn)\n", (104283, 104307), False, 'import pulumi\n'), ((105016, 105052), 'pulumi.set', 'pulumi.set', (['__self__', '"""exact"""', 'exact'], {}), "(__self__, 'exact', exact)\n", (105026, 105052), False, 'import pulumi\n'), ((105096, 105134), 'pulumi.set', 'pulumi.set', (['__self__', '"""prefix"""', 'prefix'], {}), "(__self__, 'prefix', prefix)\n", (105106, 105134), False, 'import pulumi\n'), ((107033, 107093), 'pulumi.set', 'pulumi.set', (['__self__', '"""auto_host_rewrite"""', 'auto_host_rewrite'], {}), "(__self__, 'auto_host_rewrite', auto_host_rewrite)\n", (107043, 107093), False, 'import pulumi\n'), ((107143, 107193), 'pulumi.set', 'pulumi.set', (['__self__', '"""host_rewrite"""', 'host_rewrite'], {}), "(__self__, 'host_rewrite', host_rewrite)\n", (107153, 107193), False, 'import pulumi\n'), ((107243, 107293), 'pulumi.set', 'pulumi.set', (['__self__', '"""idle_timeout"""', 'idle_timeout'], {}), "(__self__, 'idle_timeout', idle_timeout)\n", (107253, 107293), False, 'import pulumi\n'), ((107342, 107390), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_timeout"""', 'max_timeout'], {}), "(__self__, 'max_timeout', max_timeout)\n", (107352, 107390), False, 'import pulumi\n'), ((110085, 110123), 'pulumi.set', 'pulumi.set', (['__self__', '"""status"""', 'status'], {}), "(__self__, 'status', status)\n", (110095, 110123), False, 'import pulumi\n'), ((111906, 111976), 'pulumi.set', 'pulumi.set', (['__self__', '"""direct_response_action"""', 'direct_response_action'], {}), "(__self__, 'direct_response_action', direct_response_action)\n", (111916, 111976), False, 'import pulumi\n'), ((112026, 112076), 'pulumi.set', 'pulumi.set', (['__self__', '"""http_matches"""', 'http_matches'], {}), "(__self__, 'http_matches', http_matches)\n", (112036, 112076), False, 'import pulumi\n'), ((112131, 112191), 'pulumi.set', 'pulumi.set', (['__self__', '"""http_route_action"""', 'http_route_action'], {}), "(__self__, 'http_route_action', http_route_action)\n", (112141, 112191), False, 'import pulumi\n'), ((112244, 112300), 'pulumi.set', 'pulumi.set', (['__self__', '"""redirect_action"""', 'redirect_action'], {}), "(__self__, 'redirect_action', redirect_action)\n", (112254, 112300), False, 'import pulumi\n'), ((114969, 115003), 'pulumi.set', 'pulumi.set', (['__self__', '"""body"""', 'body'], {}), "(__self__, 'body', body)\n", (114979, 115003), False, 'import pulumi\n'), ((115047, 115085), 'pulumi.set', 'pulumi.set', (['__self__', '"""status"""', 'status'], {}), "(__self__, 'status', status)\n", (115057, 115085), False, 'import pulumi\n'), ((116406, 116456), 'pulumi.set', 'pulumi.set', (['__self__', '"""http_methods"""', 'http_methods'], {}), "(__self__, 'http_methods', http_methods)\n", (116416, 116456), False, 'import pulumi\n'), ((116498, 116532), 'pulumi.set', 'pulumi.set', (['__self__', '"""path"""', 'path'], {}), "(__self__, 'path', path)\n", (116508, 116532), False, 'import pulumi\n'), ((117600, 117636), 'pulumi.set', 'pulumi.set', (['__self__', '"""exact"""', 'exact'], {}), "(__self__, 'exact', exact)\n", (117610, 117636), False, 'import pulumi\n'), ((117680, 117718), 'pulumi.set', 'pulumi.set', (['__self__', '"""prefix"""', 'prefix'], {}), "(__self__, 'prefix', prefix)\n", (117690, 117718), False, 'import pulumi\n'), ((120092, 120152), 'pulumi.set', 'pulumi.set', (['__self__', '"""auto_host_rewrite"""', 'auto_host_rewrite'], {}), "(__self__, 'auto_host_rewrite', auto_host_rewrite)\n", (120102, 120152), False, 'import pulumi\n'), ((120202, 120252), 'pulumi.set', 'pulumi.set', (['__self__', '"""host_rewrite"""', 'host_rewrite'], {}), "(__self__, 'host_rewrite', host_rewrite)\n", (120212, 120252), False, 'import pulumi\n'), ((120302, 120352), 'pulumi.set', 'pulumi.set', (['__self__', '"""idle_timeout"""', 'idle_timeout'], {}), "(__self__, 'idle_timeout', idle_timeout)\n", (120312, 120352), False, 'import pulumi\n'), ((120404, 120458), 'pulumi.set', 'pulumi.set', (['__self__', '"""prefix_rewrite"""', 'prefix_rewrite'], {}), "(__self__, 'prefix_rewrite', prefix_rewrite)\n", (120414, 120458), False, 'import pulumi\n'), ((120503, 120543), 'pulumi.set', 'pulumi.set', (['__self__', '"""timeout"""', 'timeout'], {}), "(__self__, 'timeout', timeout)\n", (120513, 120543), False, 'import pulumi\n'), ((120594, 120646), 'pulumi.set', 'pulumi.set', (['__self__', '"""upgrade_types"""', 'upgrade_types'], {}), "(__self__, 'upgrade_types', upgrade_types)\n", (120604, 120646), False, 'import pulumi\n'), ((125239, 125289), 'pulumi.set', 'pulumi.set', (['__self__', '"""remove_query"""', 'remove_query'], {}), "(__self__, 'remove_query', remove_query)\n", (125249, 125289), False, 'import pulumi\n'), ((125339, 125389), 'pulumi.set', 'pulumi.set', (['__self__', '"""replace_host"""', 'replace_host'], {}), "(__self__, 'replace_host', replace_host)\n", (125349, 125389), False, 'import pulumi\n'), ((125439, 125489), 'pulumi.set', 'pulumi.set', (['__self__', '"""replace_path"""', 'replace_path'], {}), "(__self__, 'replace_path', replace_path)\n", (125449, 125489), False, 'import pulumi\n'), ((125539, 125589), 'pulumi.set', 'pulumi.set', (['__self__', '"""replace_port"""', 'replace_port'], {}), "(__self__, 'replace_port', replace_port)\n", (125549, 125589), False, 'import pulumi\n'), ((125641, 125695), 'pulumi.set', 'pulumi.set', (['__self__', '"""replace_prefix"""', 'replace_prefix'], {}), "(__self__, 'replace_prefix', replace_prefix)\n", (125651, 125695), False, 'import pulumi\n'), ((125747, 125801), 'pulumi.set', 'pulumi.set', (['__self__', '"""replace_scheme"""', 'replace_scheme'], {}), "(__self__, 'replace_scheme', replace_scheme)\n", (125757, 125801), False, 'import pulumi\n'), ((125852, 125904), 'pulumi.set', 'pulumi.set', (['__self__', '"""response_code"""', 'response_code'], {}), "(__self__, 'response_code', response_code)\n", (125862, 125904), False, 'import pulumi\n'), ((130803, 130851), 'pulumi.set', 'pulumi.set', (['__self__', '"""auto_delete"""', 'auto_delete'], {}), "(__self__, 'auto_delete', auto_delete)\n", (130813, 130851), False, 'import pulumi\n'), ((130900, 130948), 'pulumi.set', 'pulumi.set', (['__self__', '"""device_name"""', 'device_name'], {}), "(__self__, 'device_name', device_name)\n", (130910, 130948), False, 'import pulumi\n'), ((130993, 131033), 'pulumi.set', 'pulumi.set', (['__self__', '"""disk_id"""', 'disk_id'], {}), "(__self__, 'disk_id', disk_id)\n", (131003, 131033), False, 'import pulumi\n'), ((131088, 131148), 'pulumi.set', 'pulumi.set', (['__self__', '"""initialize_params"""', 'initialize_params'], {}), "(__self__, 'initialize_params', initialize_params)\n", (131098, 131148), False, 'import pulumi\n'), ((131190, 131224), 'pulumi.set', 'pulumi.set', (['__self__', '"""mode"""', 'mode'], {}), "(__self__, 'mode', mode)\n", (131200, 131224), False, 'import pulumi\n'), ((134357, 134405), 'pulumi.set', 'pulumi.set', (['__self__', '"""description"""', 'description'], {}), "(__self__, 'description', description)\n", (134367, 134405), False, 'import pulumi\n'), ((134451, 134493), 'pulumi.set', 'pulumi.set', (['__self__', '"""image_id"""', 'image_id'], {}), "(__self__, 'image_id', image_id)\n", (134461, 134493), False, 'import pulumi\n'), ((134535, 134569), 'pulumi.set', 'pulumi.set', (['__self__', '"""name"""', 'name'], {}), "(__self__, 'name', name)\n", (134545, 134569), False, 'import pulumi\n'), ((134611, 134645), 'pulumi.set', 'pulumi.set', (['__self__', '"""size"""', 'size'], {}), "(__self__, 'size', size)\n", (134621, 134645), False, 'import pulumi\n'), ((134694, 134742), 'pulumi.set', 'pulumi.set', (['__self__', '"""snapshot_id"""', 'snapshot_id'], {}), "(__self__, 'snapshot_id', snapshot_id)\n", (134704, 134742), False, 'import pulumi\n'), ((134784, 134818), 'pulumi.set', 'pulumi.set', (['__self__', '"""type"""', 'type'], {}), "(__self__, 'type', type)\n", (134794, 134818), False, 'import pulumi\n'), ((138690, 138776), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_opening_traffic_duration"""', 'max_opening_traffic_duration'], {}), "(__self__, 'max_opening_traffic_duration',\n max_opening_traffic_duration)\n", (138700, 138776), False, 'import pulumi\n'), ((138824, 138878), 'pulumi.set', 'pulumi.set', (['__self__', '"""status_message"""', 'status_message'], {}), "(__self__, 'status_message', status_message)\n", (138834, 138878), False, 'import pulumi\n'), ((138940, 139014), 'pulumi.set', 'pulumi.set', (['__self__', '"""target_group_description"""', 'target_group_description'], {}), "(__self__, 'target_group_description', target_group_description)\n", (138950, 139014), False, 'import pulumi\n'), ((139067, 139123), 'pulumi.set', 'pulumi.set', (['__self__', '"""target_group_id"""', 'target_group_id'], {}), "(__self__, 'target_group_id', target_group_id)\n", (139077, 139123), False, 'import pulumi\n'), ((139180, 139244), 'pulumi.set', 'pulumi.set', (['__self__', '"""target_group_labels"""', 'target_group_labels'], {}), "(__self__, 'target_group_labels', target_group_labels)\n", (139190, 139244), False, 'import pulumi\n'), ((139299, 139359), 'pulumi.set', 'pulumi.set', (['__self__', '"""target_group_name"""', 'target_group_name'], {}), "(__self__, 'target_group_name', target_group_name)\n", (139309, 139359), False, 'import pulumi\n'), ((143887, 143937), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_creating"""', 'max_creating'], {}), "(__self__, 'max_creating', max_creating)\n", (143897, 143937), False, 'import pulumi\n'), ((143987, 144037), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_deleting"""', 'max_deleting'], {}), "(__self__, 'max_deleting', max_deleting)\n", (143997, 144037), False, 'import pulumi\n'), ((144091, 144149), 'pulumi.set', 'pulumi.set', (['__self__', '"""startup_duration"""', 'startup_duration'], {}), "(__self__, 'startup_duration', startup_duration)\n", (144101, 144149), False, 'import pulumi\n'), ((144195, 144237), 'pulumi.set', 'pulumi.set', (['__self__', '"""strategy"""', 'strategy'], {}), "(__self__, 'strategy', strategy)\n", (144205, 144237), False, 'import pulumi\n'), ((148699, 148759), 'pulumi.set', 'pulumi.set', (['__self__', '"""healthy_threshold"""', 'healthy_threshold'], {}), "(__self__, 'healthy_threshold', healthy_threshold)\n", (148709, 148759), False, 'import pulumi\n'), ((148809, 148859), 'pulumi.set', 'pulumi.set', (['__self__', '"""http_options"""', 'http_options'], {}), "(__self__, 'http_options', http_options)\n", (148819, 148859), False, 'import pulumi\n'), ((148905, 148947), 'pulumi.set', 'pulumi.set', (['__self__', '"""interval"""', 'interval'], {}), "(__self__, 'interval', interval)\n", (148915, 148947), False, 'import pulumi\n'), ((148996, 149044), 'pulumi.set', 'pulumi.set', (['__self__', '"""tcp_options"""', 'tcp_options'], {}), "(__self__, 'tcp_options', tcp_options)\n", (149006, 149044), False, 'import pulumi\n'), ((149089, 149129), 'pulumi.set', 'pulumi.set', (['__self__', '"""timeout"""', 'timeout'], {}), "(__self__, 'timeout', timeout)\n", (149099, 149129), False, 'import pulumi\n'), ((149186, 149250), 'pulumi.set', 'pulumi.set', (['__self__', '"""unhealthy_threshold"""', 'unhealthy_threshold'], {}), "(__self__, 'unhealthy_threshold', unhealthy_threshold)\n", (149196, 149250), False, 'import pulumi\n'), ((155498, 155532), 'pulumi.set', 'pulumi.set', (['__self__', '"""fqdn"""', 'fqdn'], {}), "(__self__, 'fqdn', fqdn)\n", (155508, 155532), False, 'import pulumi\n'), ((155581, 155629), 'pulumi.set', 'pulumi.set', (['__self__', '"""instance_id"""', 'instance_id'], {}), "(__self__, 'instance_id', instance_id)\n", (155591, 155629), False, 'import pulumi\n'), ((155671, 155705), 'pulumi.set', 'pulumi.set', (['__self__', '"""name"""', 'name'], {}), "(__self__, 'name', name)\n", (155681, 155705), False, 'import pulumi\n'), ((155761, 155823), 'pulumi.set', 'pulumi.set', (['__self__', '"""network_interfaces"""', 'network_interfaces'], {}), "(__self__, 'network_interfaces', network_interfaces)\n", (155771, 155823), False, 'import pulumi\n'), ((155867, 155905), 'pulumi.set', 'pulumi.set', (['__self__', '"""status"""', 'status'], {}), "(__self__, 'status', status)\n", (155877, 155905), False, 'import pulumi\n'), ((155960, 156020), 'pulumi.set', 'pulumi.set', (['__self__', '"""status_changed_at"""', 'status_changed_at'], {}), "(__self__, 'status_changed_at', status_changed_at)\n", (155970, 156020), False, 'import pulumi\n'), ((156072, 156126), 'pulumi.set', 'pulumi.set', (['__self__', '"""status_message"""', 'status_message'], {}), "(__self__, 'status_message', status_message)\n", (156082, 156126), False, 'import pulumi\n'), ((156171, 156211), 'pulumi.set', 'pulumi.set', (['__self__', '"""zone_id"""', 'zone_id'], {}), "(__self__, 'zone_id', zone_id)\n", (156181, 156211), False, 'import pulumi\n'), ((161346, 161382), 'pulumi.set', 'pulumi.set', (['__self__', '"""index"""', 'index'], {}), "(__self__, 'index', index)\n", (161356, 161382), False, 'import pulumi\n'), ((161430, 161476), 'pulumi.set', 'pulumi.set', (['__self__', '"""ip_address"""', 'ip_address'], {}), "(__self__, 'ip_address', ip_address)\n", (161440, 161476), False, 'import pulumi\n'), ((161518, 161552), 'pulumi.set', 'pulumi.set', (['__self__', '"""ipv4"""', 'ipv4'], {}), "(__self__, 'ipv4', ipv4)\n", (161528, 161552), False, 'import pulumi\n'), ((161594, 161628), 'pulumi.set', 'pulumi.set', (['__self__', '"""ipv6"""', 'ipv6'], {}), "(__self__, 'ipv6', ipv6)\n", (161604, 161628), False, 'import pulumi\n'), ((161678, 161728), 'pulumi.set', 'pulumi.set', (['__self__', '"""ipv6_address"""', 'ipv6_address'], {}), "(__self__, 'ipv6_address', ipv6_address)\n", (161688, 161728), False, 'import pulumi\n'), ((161777, 161825), 'pulumi.set', 'pulumi.set', (['__self__', '"""mac_address"""', 'mac_address'], {}), "(__self__, 'mac_address', mac_address)\n", (161787, 161825), False, 'import pulumi\n'), ((161866, 161898), 'pulumi.set', 'pulumi.set', (['__self__', '"""nat"""', 'nat'], {}), "(__self__, 'nat', nat)\n", (161876, 161898), False, 'import pulumi\n'), ((161950, 162004), 'pulumi.set', 'pulumi.set', (['__self__', '"""nat_ip_address"""', 'nat_ip_address'], {}), "(__self__, 'nat_ip_address', nat_ip_address)\n", (161960, 162004), False, 'import pulumi\n'), ((162056, 162110), 'pulumi.set', 'pulumi.set', (['__self__', '"""nat_ip_version"""', 'nat_ip_version'], {}), "(__self__, 'nat_ip_version', nat_ip_version)\n", (162066, 162110), False, 'import pulumi\n'), ((162157, 162201), 'pulumi.set', 'pulumi.set', (['__self__', '"""subnet_id"""', 'subnet_id'], {}), "(__self__, 'subnet_id', subnet_id)\n", (162167, 162201), False, 'import pulumi\n'), ((170825, 170873), 'pulumi.set', 'pulumi.set', (['__self__', '"""description"""', 'description'], {}), "(__self__, 'description', description)\n", (170835, 170873), False, 'import pulumi\n'), ((170919, 170961), 'pulumi.set', 'pulumi.set', (['__self__', '"""hostname"""', 'hostname'], {}), "(__self__, 'hostname', hostname)\n", (170929, 170961), False, 'import pulumi\n'), ((171005, 171043), 'pulumi.set', 'pulumi.set', (['__self__', '"""labels"""', 'labels'], {}), "(__self__, 'labels', labels)\n", (171015, 171043), False, 'import pulumi\n'), ((171089, 171131), 'pulumi.set', 'pulumi.set', (['__self__', '"""metadata"""', 'metadata'], {}), "(__self__, 'metadata', metadata)\n", (171099, 171131), False, 'import pulumi\n'), ((171173, 171207), 'pulumi.set', 'pulumi.set', (['__self__', '"""name"""', 'name'], {}), "(__self__, 'name', name)\n", (171183, 171207), False, 'import pulumi\n'), ((171261, 171319), 'pulumi.set', 'pulumi.set', (['__self__', '"""network_settings"""', 'network_settings'], {}), "(__self__, 'network_settings', network_settings)\n", (171271, 171319), False, 'import pulumi\n'), ((171373, 171431), 'pulumi.set', 'pulumi.set', (['__self__', '"""placement_policy"""', 'placement_policy'], {}), "(__self__, 'placement_policy', placement_policy)\n", (171383, 171431), False, 'import pulumi\n'), ((171480, 171528), 'pulumi.set', 'pulumi.set', (['__self__', '"""platform_id"""', 'platform_id'], {}), "(__self__, 'platform_id', platform_id)\n", (171490, 171528), False, 'import pulumi\n'), ((171583, 171643), 'pulumi.set', 'pulumi.set', (['__self__', '"""scheduling_policy"""', 'scheduling_policy'], {}), "(__self__, 'scheduling_policy', scheduling_policy)\n", (171593, 171643), False, 'import pulumi\n'), ((171696, 171752), 'pulumi.set', 'pulumi.set', (['__self__', '"""secondary_disks"""', 'secondary_disks'], {}), "(__self__, 'secondary_disks', secondary_disks)\n", (171706, 171752), False, 'import pulumi\n'), ((171808, 171870), 'pulumi.set', 'pulumi.set', (['__self__', '"""service_account_id"""', 'service_account_id'], {}), "(__self__, 'service_account_id', service_account_id)\n", (171818, 171870), False, 'import pulumi\n'), ((180452, 180500), 'pulumi.set', 'pulumi.set', (['__self__', '"""device_name"""', 'device_name'], {}), "(__self__, 'device_name', device_name)\n", (180462, 180500), False, 'import pulumi\n'), ((180545, 180585), 'pulumi.set', 'pulumi.set', (['__self__', '"""disk_id"""', 'disk_id'], {}), "(__self__, 'disk_id', disk_id)\n", (180555, 180585), False, 'import pulumi\n'), ((180640, 180700), 'pulumi.set', 'pulumi.set', (['__self__', '"""initialize_params"""', 'initialize_params'], {}), "(__self__, 'initialize_params', initialize_params)\n", (180650, 180700), False, 'import pulumi\n'), ((180742, 180776), 'pulumi.set', 'pulumi.set', (['__self__', '"""mode"""', 'mode'], {}), "(__self__, 'mode', mode)\n", (180752, 180776), False, 'import pulumi\n'), ((183421, 183469), 'pulumi.set', 'pulumi.set', (['__self__', '"""description"""', 'description'], {}), "(__self__, 'description', description)\n", (183431, 183469), False, 'import pulumi\n'), ((183515, 183557), 'pulumi.set', 'pulumi.set', (['__self__', '"""image_id"""', 'image_id'], {}), "(__self__, 'image_id', image_id)\n", (183525, 183557), False, 'import pulumi\n'), ((183599, 183633), 'pulumi.set', 'pulumi.set', (['__self__', '"""size"""', 'size'], {}), "(__self__, 'size', size)\n", (183609, 183633), False, 'import pulumi\n'), ((183682, 183730), 'pulumi.set', 'pulumi.set', (['__self__', '"""snapshot_id"""', 'snapshot_id'], {}), "(__self__, 'snapshot_id', snapshot_id)\n", (183692, 183730), False, 'import pulumi\n'), ((183772, 183806), 'pulumi.set', 'pulumi.set', (['__self__', '"""type"""', 'type'], {}), "(__self__, 'type', type)\n", (183782, 183806), False, 'import pulumi\n'), ((188178, 188226), 'pulumi.set', 'pulumi.set', (['__self__', '"""dns_records"""', 'dns_records'], {}), "(__self__, 'dns_records', dns_records)\n", (188188, 188226), False, 'import pulumi\n'), ((188274, 188320), 'pulumi.set', 'pulumi.set', (['__self__', '"""ip_address"""', 'ip_address'], {}), "(__self__, 'ip_address', ip_address)\n", (188284, 188320), False, 'import pulumi\n'), ((188362, 188396), 'pulumi.set', 'pulumi.set', (['__self__', '"""ipv4"""', 'ipv4'], {}), "(__self__, 'ipv4', ipv4)\n", (188372, 188396), False, 'import pulumi\n'), ((188438, 188472), 'pulumi.set', 'pulumi.set', (['__self__', '"""ipv6"""', 'ipv6'], {}), "(__self__, 'ipv6', ipv6)\n", (188448, 188472), False, 'import pulumi\n'), ((188522, 188572), 'pulumi.set', 'pulumi.set', (['__self__', '"""ipv6_address"""', 'ipv6_address'], {}), "(__self__, 'ipv6_address', ipv6_address)\n", (188532, 188572), False, 'import pulumi\n'), ((188626, 188684), 'pulumi.set', 'pulumi.set', (['__self__', '"""ipv6_dns_records"""', 'ipv6_dns_records'], {}), "(__self__, 'ipv6_dns_records', ipv6_dns_records)\n", (188636, 188684), False, 'import pulumi\n'), ((188725, 188757), 'pulumi.set', 'pulumi.set', (['__self__', '"""nat"""', 'nat'], {}), "(__self__, 'nat', nat)\n", (188735, 188757), False, 'import pulumi\n'), ((188810, 188866), 'pulumi.set', 'pulumi.set', (['__self__', '"""nat_dns_records"""', 'nat_dns_records'], {}), "(__self__, 'nat_dns_records', nat_dns_records)\n", (188820, 188866), False, 'import pulumi\n'), ((188918, 188972), 'pulumi.set', 'pulumi.set', (['__self__', '"""nat_ip_address"""', 'nat_ip_address'], {}), "(__self__, 'nat_ip_address', nat_ip_address)\n", (188928, 188972), False, 'import pulumi\n'), ((189020, 189066), 'pulumi.set', 'pulumi.set', (['__self__', '"""network_id"""', 'network_id'], {}), "(__self__, 'network_id', network_id)\n", (189030, 189066), False, 'import pulumi\n'), ((189122, 189184), 'pulumi.set', 'pulumi.set', (['__self__', '"""security_group_ids"""', 'security_group_ids'], {}), "(__self__, 'security_group_ids', security_group_ids)\n", (189132, 189184), False, 'import pulumi\n'), ((189232, 189278), 'pulumi.set', 'pulumi.set', (['__self__', '"""subnet_ids"""', 'subnet_ids'], {}), "(__self__, 'subnet_ids', subnet_ids)\n", (189242, 189278), False, 'import pulumi\n'), ((195027, 195075), 'pulumi.set', 'pulumi.set', (['__self__', '"""dns_zone_id"""', 'dns_zone_id'], {}), "(__self__, 'dns_zone_id', dns_zone_id)\n", (195037, 195075), False, 'import pulumi\n'), ((195116, 195148), 'pulumi.set', 'pulumi.set', (['__self__', '"""ptr"""', 'ptr'], {}), "(__self__, 'ptr', ptr)\n", (195126, 195148), False, 'import pulumi\n'), ((195189, 195221), 'pulumi.set', 'pulumi.set', (['__self__', '"""ttl"""', 'ttl'], {}), "(__self__, 'ttl', ttl)\n", (195199, 195221), False, 'import pulumi\n'), ((197264, 197312), 'pulumi.set', 'pulumi.set', (['__self__', '"""dns_zone_id"""', 'dns_zone_id'], {}), "(__self__, 'dns_zone_id', dns_zone_id)\n", (197274, 197312), False, 'import pulumi\n'), ((197353, 197385), 'pulumi.set', 'pulumi.set', (['__self__', '"""ptr"""', 'ptr'], {}), "(__self__, 'ptr', ptr)\n", (197363, 197385), False, 'import pulumi\n'), ((197426, 197458), 'pulumi.set', 'pulumi.set', (['__self__', '"""ttl"""', 'ttl'], {}), "(__self__, 'ttl', ttl)\n", (197436, 197458), False, 'import pulumi\n'), ((199500, 199548), 'pulumi.set', 'pulumi.set', (['__self__', '"""dns_zone_id"""', 'dns_zone_id'], {}), "(__self__, 'dns_zone_id', dns_zone_id)\n", (199510, 199548), False, 'import pulumi\n'), ((199589, 199621), 'pulumi.set', 'pulumi.set', (['__self__', '"""ptr"""', 'ptr'], {}), "(__self__, 'ptr', ptr)\n", (199599, 199621), False, 'import pulumi\n'), ((199662, 199694), 'pulumi.set', 'pulumi.set', (['__self__', '"""ttl"""', 'ttl'], {}), "(__self__, 'ttl', ttl)\n", (199672, 199694), False, 'import pulumi\n'), ((201303, 201337), 'pulumi.set', 'pulumi.set', (['__self__', '"""type"""', 'type'], {}), "(__self__, 'type', type)\n", (201313, 201337), False, 'import pulumi\n'), ((203231, 203283), 'pulumi.set', 'pulumi.set', (['__self__', '"""core_fraction"""', 'core_fraction'], {}), "(__self__, 'core_fraction', core_fraction)\n", (203241, 203283), False, 'import pulumi\n'), ((203325, 203359), 'pulumi.set', 'pulumi.set', (['__self__', '"""gpus"""', 'gpus'], {}), "(__self__, 'gpus', gpus)\n", (203335, 203359), False, 'import pulumi\n'), ((204948, 204996), 'pulumi.set', 'pulumi.set', (['__self__', '"""preemptible"""', 'preemptible'], {}), "(__self__, 'preemptible', preemptible)\n", (204958, 204996), False, 'import pulumi\n'), ((206425, 206473), 'pulumi.set', 'pulumi.set', (['__self__', '"""device_name"""', 'device_name'], {}), "(__self__, 'device_name', device_name)\n", (206435, 206473), False, 'import pulumi\n'), ((206518, 206558), 'pulumi.set', 'pulumi.set', (['__self__', '"""disk_id"""', 'disk_id'], {}), "(__self__, 'disk_id', disk_id)\n", (206528, 206558), False, 'import pulumi\n'), ((206613, 206673), 'pulumi.set', 'pulumi.set', (['__self__', '"""initialize_params"""', 'initialize_params'], {}), "(__self__, 'initialize_params', initialize_params)\n", (206623, 206673), False, 'import pulumi\n'), ((206715, 206749), 'pulumi.set', 'pulumi.set', (['__self__', '"""mode"""', 'mode'], {}), "(__self__, 'mode', mode)\n", (206725, 206749), False, 'import pulumi\n'), ((209409, 209457), 'pulumi.set', 'pulumi.set', (['__self__', '"""description"""', 'description'], {}), "(__self__, 'description', description)\n", (209419, 209457), False, 'import pulumi\n'), ((209503, 209545), 'pulumi.set', 'pulumi.set', (['__self__', '"""image_id"""', 'image_id'], {}), "(__self__, 'image_id', image_id)\n", (209513, 209545), False, 'import pulumi\n'), ((209587, 209621), 'pulumi.set', 'pulumi.set', (['__self__', '"""size"""', 'size'], {}), "(__self__, 'size', size)\n", (209597, 209621), False, 'import pulumi\n'), ((209670, 209718), 'pulumi.set', 'pulumi.set', (['__self__', '"""snapshot_id"""', 'snapshot_id'], {}), "(__self__, 'snapshot_id', snapshot_id)\n", (209680, 209718), False, 'import pulumi\n'), ((209760, 209794), 'pulumi.set', 'pulumi.set', (['__self__', '"""type"""', 'type'], {}), "(__self__, 'type', type)\n", (209770, 209794), False, 'import pulumi\n'), ((212786, 212872), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_opening_traffic_duration"""', 'max_opening_traffic_duration'], {}), "(__self__, 'max_opening_traffic_duration',\n max_opening_traffic_duration)\n", (212796, 212872), False, 'import pulumi\n'), ((212920, 212974), 'pulumi.set', 'pulumi.set', (['__self__', '"""status_message"""', 'status_message'], {}), "(__self__, 'status_message', status_message)\n", (212930, 212974), False, 'import pulumi\n'), ((213036, 213110), 'pulumi.set', 'pulumi.set', (['__self__', '"""target_group_description"""', 'target_group_description'], {}), "(__self__, 'target_group_description', target_group_description)\n", (213046, 213110), False, 'import pulumi\n'), ((213163, 213219), 'pulumi.set', 'pulumi.set', (['__self__', '"""target_group_id"""', 'target_group_id'], {}), "(__self__, 'target_group_id', target_group_id)\n", (213173, 213219), False, 'import pulumi\n'), ((213276, 213340), 'pulumi.set', 'pulumi.set', (['__self__', '"""target_group_labels"""', 'target_group_labels'], {}), "(__self__, 'target_group_labels', target_group_labels)\n", (213286, 213340), False, 'import pulumi\n'), ((213395, 213455), 'pulumi.set', 'pulumi.set', (['__self__', '"""target_group_name"""', 'target_group_name'], {}), "(__self__, 'target_group_name', target_group_name)\n", (213405, 213455), False, 'import pulumi\n'), ((217107, 217153), 'pulumi.set', 'pulumi.set', (['__self__', '"""auto_scale"""', 'auto_scale'], {}), "(__self__, 'auto_scale', auto_scale)\n", (217117, 217153), False, 'import pulumi\n'), ((217202, 217250), 'pulumi.set', 'pulumi.set', (['__self__', '"""fixed_scale"""', 'fixed_scale'], {}), "(__self__, 'fixed_scale', fixed_scale)\n", (217212, 217250), False, 'import pulumi\n'), ((217303, 217359), 'pulumi.set', 'pulumi.set', (['__self__', '"""test_auto_scale"""', 'test_auto_scale'], {}), "(__self__, 'test_auto_scale', test_auto_scale)\n", (217313, 217359), False, 'import pulumi\n'), ((221312, 221382), 'pulumi.set', 'pulumi.set', (['__self__', '"""cpu_utilization_target"""', 'cpu_utilization_target'], {}), "(__self__, 'cpu_utilization_target', cpu_utilization_target)\n", (221322, 221382), False, 'import pulumi\n'), ((221432, 221482), 'pulumi.set', 'pulumi.set', (['__self__', '"""custom_rules"""', 'custom_rules'], {}), "(__self__, 'custom_rules', custom_rules)\n", (221442, 221482), False, 'import pulumi\n'), ((221528, 221570), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_size"""', 'max_size'], {}), "(__self__, 'max_size', max_size)\n", (221538, 221570), False, 'import pulumi\n'), ((221621, 221673), 'pulumi.set', 'pulumi.set', (['__self__', '"""min_zone_size"""', 'min_zone_size'], {}), "(__self__, 'min_zone_size', min_zone_size)\n", (221631, 221673), False, 'import pulumi\n'), ((221733, 221803), 'pulumi.set', 'pulumi.set', (['__self__', '"""stabilization_duration"""', 'stabilization_duration'], {}), "(__self__, 'stabilization_duration', stabilization_duration)\n", (221743, 221803), False, 'import pulumi\n'), ((221856, 221912), 'pulumi.set', 'pulumi.set', (['__self__', '"""warmup_duration"""', 'warmup_duration'], {}), "(__self__, 'warmup_duration', warmup_duration)\n", (221866, 221912), False, 'import pulumi\n'), ((227688, 227732), 'pulumi.set', 'pulumi.set', (['__self__', '"""folder_id"""', 'folder_id'], {}), "(__self__, 'folder_id', folder_id)\n", (227698, 227732), False, 'import pulumi\n'), ((227776, 227814), 'pulumi.set', 'pulumi.set', (['__self__', '"""labels"""', 'labels'], {}), "(__self__, 'labels', labels)\n", (227786, 227814), False, 'import pulumi\n'), ((227859, 227899), 'pulumi.set', 'pulumi.set', (['__self__', '"""service"""', 'service'], {}), "(__self__, 'service', service)\n", (227869, 227899), False, 'import pulumi\n'), ((233739, 233809), 'pulumi.set', 'pulumi.set', (['__self__', '"""cpu_utilization_target"""', 'cpu_utilization_target'], {}), "(__self__, 'cpu_utilization_target', cpu_utilization_target)\n", (233749, 233809), False, 'import pulumi\n'), ((233859, 233909), 'pulumi.set', 'pulumi.set', (['__self__', '"""custom_rules"""', 'custom_rules'], {}), "(__self__, 'custom_rules', custom_rules)\n", (233869, 233909), False, 'import pulumi\n'), ((233955, 233997), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_size"""', 'max_size'], {}), "(__self__, 'max_size', max_size)\n", (233965, 233997), False, 'import pulumi\n'), ((234048, 234100), 'pulumi.set', 'pulumi.set', (['__self__', '"""min_zone_size"""', 'min_zone_size'], {}), "(__self__, 'min_zone_size', min_zone_size)\n", (234058, 234100), False, 'import pulumi\n'), ((234160, 234230), 'pulumi.set', 'pulumi.set', (['__self__', '"""stabilization_duration"""', 'stabilization_duration'], {}), "(__self__, 'stabilization_duration', stabilization_duration)\n", (234170, 234230), False, 'import pulumi\n'), ((234283, 234339), 'pulumi.set', 'pulumi.set', (['__self__', '"""warmup_duration"""', 'warmup_duration'], {}), "(__self__, 'warmup_duration', warmup_duration)\n", (234293, 234339), False, 'import pulumi\n'), ((240127, 240171), 'pulumi.set', 'pulumi.set', (['__self__', '"""folder_id"""', 'folder_id'], {}), "(__self__, 'folder_id', folder_id)\n", (240137, 240171), False, 'import pulumi\n'), ((240215, 240253), 'pulumi.set', 'pulumi.set', (['__self__', '"""labels"""', 'labels'], {}), "(__self__, 'labels', labels)\n", (240225, 240253), False, 'import pulumi\n'), ((240298, 240338), 'pulumi.set', 'pulumi.set', (['__self__', '"""service"""', 'service'], {}), "(__self__, 'service', service)\n", (240308, 240338), False, 'import pulumi\n'), ((246324, 246372), 'pulumi.set', 'pulumi.set', (['__self__', '"""dns_records"""', 'dns_records'], {}), "(__self__, 'dns_records', dns_records)\n", (246334, 246372), False, 'import pulumi\n'), ((246415, 246451), 'pulumi.set', 'pulumi.set', (['__self__', '"""index"""', 'index'], {}), "(__self__, 'index', index)\n", (246425, 246451), False, 'import pulumi\n'), ((246499, 246545), 'pulumi.set', 'pulumi.set', (['__self__', '"""ip_address"""', 'ip_address'], {}), "(__self__, 'ip_address', ip_address)\n", (246509, 246545), False, 'import pulumi\n'), ((246587, 246621), 'pulumi.set', 'pulumi.set', (['__self__', '"""ipv4"""', 'ipv4'], {}), "(__self__, 'ipv4', ipv4)\n", (246597, 246621), False, 'import pulumi\n'), ((246663, 246697), 'pulumi.set', 'pulumi.set', (['__self__', '"""ipv6"""', 'ipv6'], {}), "(__self__, 'ipv6', ipv6)\n", (246673, 246697), False, 'import pulumi\n'), ((246747, 246797), 'pulumi.set', 'pulumi.set', (['__self__', '"""ipv6_address"""', 'ipv6_address'], {}), "(__self__, 'ipv6_address', ipv6_address)\n", (246757, 246797), False, 'import pulumi\n'), ((246851, 246909), 'pulumi.set', 'pulumi.set', (['__self__', '"""ipv6_dns_records"""', 'ipv6_dns_records'], {}), "(__self__, 'ipv6_dns_records', ipv6_dns_records)\n", (246861, 246909), False, 'import pulumi\n'), ((246958, 247006), 'pulumi.set', 'pulumi.set', (['__self__', '"""mac_address"""', 'mac_address'], {}), "(__self__, 'mac_address', mac_address)\n", (246968, 247006), False, 'import pulumi\n'), ((247047, 247079), 'pulumi.set', 'pulumi.set', (['__self__', '"""nat"""', 'nat'], {}), "(__self__, 'nat', nat)\n", (247057, 247079), False, 'import pulumi\n'), ((247132, 247188), 'pulumi.set', 'pulumi.set', (['__self__', '"""nat_dns_records"""', 'nat_dns_records'], {}), "(__self__, 'nat_dns_records', nat_dns_records)\n", (247142, 247188), False, 'import pulumi\n'), ((247240, 247294), 'pulumi.set', 'pulumi.set', (['__self__', '"""nat_ip_address"""', 'nat_ip_address'], {}), "(__self__, 'nat_ip_address', nat_ip_address)\n", (247250, 247294), False, 'import pulumi\n'), ((247346, 247400), 'pulumi.set', 'pulumi.set', (['__self__', '"""nat_ip_version"""', 'nat_ip_version'], {}), "(__self__, 'nat_ip_version', nat_ip_version)\n", (247356, 247400), False, 'import pulumi\n'), ((247456, 247518), 'pulumi.set', 'pulumi.set', (['__self__', '"""security_group_ids"""', 'security_group_ids'], {}), "(__self__, 'security_group_ids', security_group_ids)\n", (247466, 247518), False, 'import pulumi\n'), ((254131, 254179), 'pulumi.set', 'pulumi.set', (['__self__', '"""dns_zone_id"""', 'dns_zone_id'], {}), "(__self__, 'dns_zone_id', dns_zone_id)\n", (254141, 254179), False, 'import pulumi\n'), ((254220, 254252), 'pulumi.set', 'pulumi.set', (['__self__', '"""ptr"""', 'ptr'], {}), "(__self__, 'ptr', ptr)\n", (254230, 254252), False, 'import pulumi\n'), ((254293, 254325), 'pulumi.set', 'pulumi.set', (['__self__', '"""ttl"""', 'ttl'], {}), "(__self__, 'ttl', ttl)\n", (254303, 254325), False, 'import pulumi\n'), ((256377, 256425), 'pulumi.set', 'pulumi.set', (['__self__', '"""dns_zone_id"""', 'dns_zone_id'], {}), "(__self__, 'dns_zone_id', dns_zone_id)\n", (256387, 256425), False, 'import pulumi\n'), ((256466, 256498), 'pulumi.set', 'pulumi.set', (['__self__', '"""ptr"""', 'ptr'], {}), "(__self__, 'ptr', ptr)\n", (256476, 256498), False, 'import pulumi\n'), ((256539, 256571), 'pulumi.set', 'pulumi.set', (['__self__', '"""ttl"""', 'ttl'], {}), "(__self__, 'ttl', ttl)\n", (256549, 256571), False, 'import pulumi\n'), ((258622, 258670), 'pulumi.set', 'pulumi.set', (['__self__', '"""dns_zone_id"""', 'dns_zone_id'], {}), "(__self__, 'dns_zone_id', dns_zone_id)\n", (258632, 258670), False, 'import pulumi\n'), ((258711, 258743), 'pulumi.set', 'pulumi.set', (['__self__', '"""ptr"""', 'ptr'], {}), "(__self__, 'ptr', ptr)\n", (258721, 258743), False, 'import pulumi\n'), ((258784, 258816), 'pulumi.set', 'pulumi.set', (['__self__', '"""ttl"""', 'ttl'], {}), "(__self__, 'ttl', ttl)\n", (258794, 258816), False, 'import pulumi\n'), ((261587, 261639), 'pulumi.set', 'pulumi.set', (['__self__', '"""core_fraction"""', 'core_fraction'], {}), "(__self__, 'core_fraction', core_fraction)\n", (261597, 261639), False, 'import pulumi\n'), ((261681, 261715), 'pulumi.set', 'pulumi.set', (['__self__', '"""gpus"""', 'gpus'], {}), "(__self__, 'gpus', gpus)\n", (261691, 261715), False, 'import pulumi\n'), ((263271, 263319), 'pulumi.set', 'pulumi.set', (['__self__', '"""preemptible"""', 'preemptible'], {}), "(__self__, 'preemptible', preemptible)\n", (263281, 263319), False, 'import pulumi\n'), ((264648, 264696), 'pulumi.set', 'pulumi.set', (['__self__', '"""auto_delete"""', 'auto_delete'], {}), "(__self__, 'auto_delete', auto_delete)\n", (264658, 264696), False, 'import pulumi\n'), ((264745, 264793), 'pulumi.set', 'pulumi.set', (['__self__', '"""device_name"""', 'device_name'], {}), "(__self__, 'device_name', device_name)\n", (264755, 264793), False, 'import pulumi\n'), ((264835, 264869), 'pulumi.set', 'pulumi.set', (['__self__', '"""mode"""', 'mode'], {}), "(__self__, 'mode', mode)\n", (264845, 264869), False, 'import pulumi\n'), ((267319, 267357), 'pulumi.set', 'pulumi.set', (['__self__', '"""hadoop"""', 'hadoop'], {}), "(__self__, 'hadoop', hadoop)\n", (267329, 267357), False, 'import pulumi\n'), ((267405, 267451), 'pulumi.set', 'pulumi.set', (['__self__', '"""version_id"""', 'version_id'], {}), "(__self__, 'version_id', version_id)\n", (267415, 267451), False, 'import pulumi\n'), ((269745, 269791), 'pulumi.set', 'pulumi.set', (['__self__', '"""properties"""', 'properties'], {}), "(__self__, 'properties', properties)\n", (269755, 269791), False, 'import pulumi\n'), ((269837, 269879), 'pulumi.set', 'pulumi.set', (['__self__', '"""services"""', 'services'], {}), "(__self__, 'services', services)\n", (269847, 269879), False, 'import pulumi\n'), ((269932, 269988), 'pulumi.set', 'pulumi.set', (['__self__', '"""ssh_public_keys"""', 'ssh_public_keys'], {}), "(__self__, 'ssh_public_keys', ssh_public_keys)\n", (269942, 269988), False, 'import pulumi\n'), ((273175, 273237), 'pulumi.set', 'pulumi.set', (['__self__', '"""autoscaling_config"""', 'autoscaling_config'], {}), "(__self__, 'autoscaling_config', autoscaling_config)\n", (273185, 273237), False, 'import pulumi\n'), ((273277, 273307), 'pulumi.set', 'pulumi.set', (['__self__', '"""id"""', 'id'], {}), "(__self__, 'id', id)\n", (273287, 273307), False, 'import pulumi\n'), ((278134, 278204), 'pulumi.set', 'pulumi.set', (['__self__', '"""cpu_utilization_target"""', 'cpu_utilization_target'], {}), "(__self__, 'cpu_utilization_target', cpu_utilization_target)\n", (278144, 278204), False, 'import pulumi\n'), ((278262, 278328), 'pulumi.set', 'pulumi.set', (['__self__', '"""decommission_timeout"""', 'decommission_timeout'], {}), "(__self__, 'decommission_timeout', decommission_timeout)\n", (278272, 278328), False, 'import pulumi\n'), ((278386, 278452), 'pulumi.set', 'pulumi.set', (['__self__', '"""measurement_duration"""', 'measurement_duration'], {}), "(__self__, 'measurement_duration', measurement_duration)\n", (278396, 278452), False, 'import pulumi\n'), ((278501, 278549), 'pulumi.set', 'pulumi.set', (['__self__', '"""preemptible"""', 'preemptible'], {}), "(__self__, 'preemptible', preemptible)\n", (278511, 278549), False, 'import pulumi\n'), ((278609, 278679), 'pulumi.set', 'pulumi.set', (['__self__', '"""stabilization_duration"""', 'stabilization_duration'], {}), "(__self__, 'stabilization_duration', stabilization_duration)\n", (278619, 278679), False, 'import pulumi\n'), ((278732, 278788), 'pulumi.set', 'pulumi.set', (['__self__', '"""warmup_duration"""', 'warmup_duration'], {}), "(__self__, 'warmup_duration', warmup_duration)\n", (278742, 278788), False, 'import pulumi\n'), ((283273, 283323), 'pulumi.set', 'pulumi.set', (['__self__', '"""disk_type_id"""', 'disk_type_id'], {}), "(__self__, 'disk_type_id', disk_type_id)\n", (283283, 283323), False, 'import pulumi\n'), ((285484, 285522), 'pulumi.set', 'pulumi.set', (['__self__', '"""sha256"""', 'sha256'], {}), "(__self__, 'sha256', sha256)\n", (285494, 285522), False, 'import pulumi\n'), ((286697, 286763), 'pulumi.set', 'pulumi.set', (['__self__', '"""zone_instances_limit"""', 'zone_instances_limit'], {}), "(__self__, 'zone_instances_limit', zone_instances_limit)\n", (286707, 286763), False, 'import pulumi\n'), ((286820, 286884), 'pulumi.set', 'pulumi.set', (['__self__', '"""zone_requests_limit"""', 'zone_requests_limit'], {}), "(__self__, 'zone_requests_limit', zone_requests_limit)\n", (286830, 286884), False, 'import pulumi\n'), ((289146, 289200), 'pulumi.set', 'pulumi.set', (['__self__', '"""retry_attempts"""', 'retry_attempts'], {}), "(__self__, 'retry_attempts', retry_attempts)\n", (289156, 289200), False, 'import pulumi\n'), ((289252, 289306), 'pulumi.set', 'pulumi.set', (['__self__', '"""retry_interval"""', 'retry_interval'], {}), "(__self__, 'retry_interval', retry_interval)\n", (289262, 289306), False, 'import pulumi\n'), ((289362, 289424), 'pulumi.set', 'pulumi.set', (['__self__', '"""service_account_id"""', 'service_account_id'], {}), "(__self__, 'service_account_id', service_account_id)\n", (289372, 289424), False, 'import pulumi\n'), ((289465, 289497), 'pulumi.set', 'pulumi.set', (['__self__', '"""tag"""', 'tag'], {}), "(__self__, 'tag', tag)\n", (289475, 289497), False, 'import pulumi\n'), ((291255, 291299), 'pulumi.set', 'pulumi.set', (['__self__', '"""device_id"""', 'device_id'], {}), "(__self__, 'device_id', device_id)\n", (291265, 291299), False, 'import pulumi\n'), ((291342, 291378), 'pulumi.set', 'pulumi.set', (['__self__', '"""topic"""', 'topic'], {}), "(__self__, 'topic', topic)\n", (291352, 291378), False, 'import pulumi\n'), ((292621, 292667), 'pulumi.set', 'pulumi.set', (['__self__', '"""batch_size"""', 'batch_size'], {}), "(__self__, 'batch_size', batch_size)\n", (292631, 292667), False, 'import pulumi\n'), ((294357, 294403), 'pulumi.set', 'pulumi.set', (['__self__', '"""batch_size"""', 'batch_size'], {}), "(__self__, 'batch_size', batch_size)\n", (294367, 294403), False, 'import pulumi\n'), ((296777, 296823), 'pulumi.set', 'pulumi.set', (['__self__', '"""batch_size"""', 'batch_size'], {}), "(__self__, 'batch_size', batch_size)\n", (296787, 296823), False, 'import pulumi\n'), ((296879, 296941), 'pulumi.set', 'pulumi.set', (['__self__', '"""visibility_timeout"""', 'visibility_timeout'], {}), "(__self__, 'visibility_timeout', visibility_timeout)\n", (296889, 296941), False, 'import pulumi\n'), ((298955, 298993), 'pulumi.set', 'pulumi.set', (['__self__', '"""create"""', 'create'], {}), "(__self__, 'create', create)\n", (298965, 298993), False, 'import pulumi\n'), ((299037, 299075), 'pulumi.set', 'pulumi.set', (['__self__', '"""delete"""', 'delete'], {}), "(__self__, 'delete', delete)\n", (299047, 299075), False, 'import pulumi\n'), ((299119, 299157), 'pulumi.set', 'pulumi.set', (['__self__', '"""prefix"""', 'prefix'], {}), "(__self__, 'prefix', prefix)\n", (299129, 299157), False, 'import pulumi\n'), ((299201, 299239), 'pulumi.set', 'pulumi.set', (['__self__', '"""suffix"""', 'suffix'], {}), "(__self__, 'suffix', suffix)\n", (299211, 299239), False, 'import pulumi\n'), ((299283, 299321), 'pulumi.set', 'pulumi.set', (['__self__', '"""update"""', 'update'], {}), "(__self__, 'update', update)\n", (299293, 299321), False, 'import pulumi\n'), ((301606, 301644), 'pulumi.set', 'pulumi.set', (['__self__', '"""key_id"""', 'key_id'], {}), "(__self__, 'key_id', key_id)\n", (301616, 301644), False, 'import pulumi\n'), ((305323, 305393), 'pulumi.set', 'pulumi.set', (['__self__', '"""cluster_ca_certificate"""', 'cluster_ca_certificate'], {}), "(__self__, 'cluster_ca_certificate', cluster_ca_certificate)\n", (305333, 305393), False, 'import pulumi\n'), ((305450, 305514), 'pulumi.set', 'pulumi.set', (['__self__', '"""external_v4_address"""', 'external_v4_address'], {}), "(__self__, 'external_v4_address', external_v4_address)\n", (305460, 305514), False, 'import pulumi\n'), ((305572, 305638), 'pulumi.set', 'pulumi.set', (['__self__', '"""external_v4_endpoint"""', 'external_v4_endpoint'], {}), "(__self__, 'external_v4_endpoint', external_v4_endpoint)\n", (305582, 305638), False, 'import pulumi\n'), ((305695, 305759), 'pulumi.set', 'pulumi.set', (['__self__', '"""internal_v4_address"""', 'internal_v4_address'], {}), "(__self__, 'internal_v4_address', internal_v4_address)\n", (305705, 305759), False, 'import pulumi\n'), ((305817, 305883), 'pulumi.set', 'pulumi.set', (['__self__', '"""internal_v4_endpoint"""', 'internal_v4_endpoint'], {}), "(__self__, 'internal_v4_endpoint', internal_v4_endpoint)\n", (305827, 305883), False, 'import pulumi\n'), ((305939, 306001), 'pulumi.set', 'pulumi.set', (['__self__', '"""maintenance_policy"""', 'maintenance_policy'], {}), "(__self__, 'maintenance_policy', maintenance_policy)\n", (305949, 306001), False, 'import pulumi\n'), ((306048, 306092), 'pulumi.set', 'pulumi.set', (['__self__', '"""public_ip"""', 'public_ip'], {}), "(__self__, 'public_ip', public_ip)\n", (306058, 306092), False, 'import pulumi\n'), ((306138, 306180), 'pulumi.set', 'pulumi.set', (['__self__', '"""regional"""', 'regional'], {}), "(__self__, 'regional', regional)\n", (306148, 306180), False, 'import pulumi\n'), ((306236, 306298), 'pulumi.set', 'pulumi.set', (['__self__', '"""security_group_ids"""', 'security_group_ids'], {}), "(__self__, 'security_group_ids', security_group_ids)\n", (306246, 306298), False, 'import pulumi\n'), ((306343, 306383), 'pulumi.set', 'pulumi.set', (['__self__', '"""version"""', 'version'], {}), "(__self__, 'version', version)\n", (306353, 306383), False, 'import pulumi\n'), ((306433, 306483), 'pulumi.set', 'pulumi.set', (['__self__', '"""version_info"""', 'version_info'], {}), "(__self__, 'version_info', version_info)\n", (306443, 306483), False, 'import pulumi\n'), ((306526, 306562), 'pulumi.set', 'pulumi.set', (['__self__', '"""zonal"""', 'zonal'], {}), "(__self__, 'zonal', zonal)\n", (306536, 306562), False, 'import pulumi\n'), ((313541, 313605), 'pulumi.set', 'pulumi.set', (['__self__', '"""maintenance_windows"""', 'maintenance_windows'], {}), "(__self__, 'maintenance_windows', maintenance_windows)\n", (313551, 313605), False, 'import pulumi\n'), ((315339, 315371), 'pulumi.set', 'pulumi.set', (['__self__', '"""day"""', 'day'], {}), "(__self__, 'day', day)\n", (315349, 315371), False, 'import pulumi\n'), ((316849, 316893), 'pulumi.set', 'pulumi.set', (['__self__', '"""locations"""', 'locations'], {}), "(__self__, 'locations', locations)\n", (316859, 316893), False, 'import pulumi\n'), ((318231, 318275), 'pulumi.set', 'pulumi.set', (['__self__', '"""subnet_id"""', 'subnet_id'], {}), "(__self__, 'subnet_id', subnet_id)\n", (318241, 318275), False, 'import pulumi\n'), ((318317, 318351), 'pulumi.set', 'pulumi.set', (['__self__', '"""zone"""', 'zone'], {}), "(__self__, 'zone', zone)\n", (318327, 318351), False, 'import pulumi\n'), ((320291, 320347), 'pulumi.set', 'pulumi.set', (['__self__', '"""current_version"""', 'current_version'], {}), "(__self__, 'current_version', current_version)\n", (320301, 320347), False, 'import pulumi\n'), ((320407, 320477), 'pulumi.set', 'pulumi.set', (['__self__', '"""new_revision_available"""', 'new_revision_available'], {}), "(__self__, 'new_revision_available', new_revision_available)\n", (320417, 320477), False, 'import pulumi\n'), ((320535, 320601), 'pulumi.set', 'pulumi.set', (['__self__', '"""new_revision_summary"""', 'new_revision_summary'], {}), "(__self__, 'new_revision_summary', new_revision_summary)\n", (320545, 320601), False, 'import pulumi\n'), ((320657, 320719), 'pulumi.set', 'pulumi.set', (['__self__', '"""version_deprecated"""', 'version_deprecated'], {}), "(__self__, 'version_deprecated', version_deprecated)\n", (320667, 320719), False, 'import pulumi\n'), ((323208, 323252), 'pulumi.set', 'pulumi.set', (['__self__', '"""subnet_id"""', 'subnet_id'], {}), "(__self__, 'subnet_id', subnet_id)\n", (323218, 323252), False, 'import pulumi\n'), ((323294, 323328), 'pulumi.set', 'pulumi.set', (['__self__', '"""zone"""', 'zone'], {}), "(__self__, 'zone', zone)\n", (323304, 323328), False, 'import pulumi\n'), ((324421, 324459), 'pulumi.set', 'pulumi.set', (['__self__', '"""cilium"""', 'cilium'], {}), "(__self__, 'cilium', cilium)\n", (324431, 324459), False, 'import pulumi\n'), ((325576, 325620), 'pulumi.set', 'pulumi.set', (['__self__', '"""locations"""', 'locations'], {}), "(__self__, 'locations', locations)\n", (325586, 325620), False, 'import pulumi\n'), ((326721, 326867), 'warnings.warn', 'warnings.warn', (['"""The \'subnet_id\' field has been deprecated. Please use \'subnet_ids under network_interface\' instead."""', 'DeprecationWarning'], {}), '(\n "The \'subnet_id\' field has been deprecated. Please use \'subnet_ids under network_interface\' instead."\n , DeprecationWarning)\n', (326734, 326867), False, 'import warnings\n'), ((326874, 327027), 'pulumi.log.warn', 'pulumi.log.warn', (['"""subnet_id is deprecated: The \'subnet_id\' field has been deprecated. Please use \'subnet_ids under network_interface\' instead."""'], {}), '(\n "subnet_id is deprecated: The \'subnet_id\' field has been deprecated. Please use \'subnet_ids under network_interface\' instead."\n )\n', (326889, 327027), False, 'import pulumi\n'), ((327068, 327112), 'pulumi.set', 'pulumi.set', (['__self__', '"""subnet_id"""', 'subnet_id'], {}), "(__self__, 'subnet_id', subnet_id)\n", (327078, 327112), False, 'import pulumi\n'), ((327154, 327188), 'pulumi.set', 'pulumi.set', (['__self__', '"""zone"""', 'zone'], {}), "(__self__, 'zone', zone)\n", (327164, 327188), False, 'import pulumi\n'), ((332212, 332256), 'pulumi.set', 'pulumi.set', (['__self__', '"""boot_disk"""', 'boot_disk'], {}), "(__self__, 'boot_disk', boot_disk)\n", (332222, 332256), False, 'import pulumi\n'), ((332302, 332344), 'pulumi.set', 'pulumi.set', (['__self__', '"""metadata"""', 'metadata'], {}), "(__self__, 'metadata', metadata)\n", (332312, 332344), False, 'import pulumi\n'), ((332385, 332518), 'warnings.warn', 'warnings.warn', (['"""The \'nat\' field has been deprecated. Please use \'nat under network_interface\' instead."""', 'DeprecationWarning'], {}), '(\n "The \'nat\' field has been deprecated. Please use \'nat under network_interface\' instead."\n , DeprecationWarning)\n', (332398, 332518), False, 'import warnings\n'), ((332525, 332659), 'pulumi.log.warn', 'pulumi.log.warn', (['"""nat is deprecated: The \'nat\' field has been deprecated. Please use \'nat under network_interface\' instead."""'], {}), '(\n "nat is deprecated: The \'nat\' field has been deprecated. Please use \'nat under network_interface\' instead."\n )\n', (332540, 332659), False, 'import pulumi\n'), ((332694, 332726), 'pulumi.set', 'pulumi.set', (['__self__', '"""nat"""', 'nat'], {}), "(__self__, 'nat', nat)\n", (332704, 332726), False, 'import pulumi\n'), ((332789, 332865), 'pulumi.set', 'pulumi.set', (['__self__', '"""network_acceleration_type"""', 'network_acceleration_type'], {}), "(__self__, 'network_acceleration_type', network_acceleration_type)\n", (332799, 332865), False, 'import pulumi\n'), ((332921, 332983), 'pulumi.set', 'pulumi.set', (['__self__', '"""network_interfaces"""', 'network_interfaces'], {}), "(__self__, 'network_interfaces', network_interfaces)\n", (332931, 332983), False, 'import pulumi\n'), ((333037, 333095), 'pulumi.set', 'pulumi.set', (['__self__', '"""placement_policy"""', 'placement_policy'], {}), "(__self__, 'placement_policy', placement_policy)\n", (333047, 333095), False, 'import pulumi\n'), ((333144, 333192), 'pulumi.set', 'pulumi.set', (['__self__', '"""platform_id"""', 'platform_id'], {}), "(__self__, 'platform_id', platform_id)\n", (333154, 333192), False, 'import pulumi\n'), ((333239, 333283), 'pulumi.set', 'pulumi.set', (['__self__', '"""resources"""', 'resources'], {}), "(__self__, 'resources', resources)\n", (333249, 333283), False, 'import pulumi\n'), ((333338, 333398), 'pulumi.set', 'pulumi.set', (['__self__', '"""scheduling_policy"""', 'scheduling_policy'], {}), "(__self__, 'scheduling_policy', scheduling_policy)\n", (333348, 333398), False, 'import pulumi\n'), ((338432, 338466), 'pulumi.set', 'pulumi.set', (['__self__', '"""size"""', 'size'], {}), "(__self__, 'size', size)\n", (338442, 338466), False, 'import pulumi\n'), ((338508, 338542), 'pulumi.set', 'pulumi.set', (['__self__', '"""type"""', 'type'], {}), "(__self__, 'type', type)\n", (338518, 338542), False, 'import pulumi\n'), ((340320, 340354), 'pulumi.set', 'pulumi.set', (['__self__', '"""ipv4"""', 'ipv4'], {}), "(__self__, 'ipv4', ipv4)\n", (340330, 340354), False, 'import pulumi\n'), ((340396, 340430), 'pulumi.set', 'pulumi.set', (['__self__', '"""ipv6"""', 'ipv6'], {}), "(__self__, 'ipv6', ipv6)\n", (340406, 340430), False, 'import pulumi\n'), ((340471, 340503), 'pulumi.set', 'pulumi.set', (['__self__', '"""nat"""', 'nat'], {}), "(__self__, 'nat', nat)\n", (340481, 340503), False, 'import pulumi\n'), ((340559, 340621), 'pulumi.set', 'pulumi.set', (['__self__', '"""security_group_ids"""', 'security_group_ids'], {}), "(__self__, 'security_group_ids', security_group_ids)\n", (340569, 340621), False, 'import pulumi\n'), ((343734, 343786), 'pulumi.set', 'pulumi.set', (['__self__', '"""core_fraction"""', 'core_fraction'], {}), "(__self__, 'core_fraction', core_fraction)\n", (343744, 343786), False, 'import pulumi\n'), ((343829, 343865), 'pulumi.set', 'pulumi.set', (['__self__', '"""cores"""', 'cores'], {}), "(__self__, 'cores', cores)\n", (343839, 343865), False, 'import pulumi\n'), ((343907, 343941), 'pulumi.set', 'pulumi.set', (['__self__', '"""gpus"""', 'gpus'], {}), "(__self__, 'gpus', gpus)\n", (343917, 343941), False, 'import pulumi\n'), ((343985, 344023), 'pulumi.set', 'pulumi.set', (['__self__', '"""memory"""', 'memory'], {}), "(__self__, 'memory', memory)\n", (343995, 344023), False, 'import pulumi\n'), ((345446, 345494), 'pulumi.set', 'pulumi.set', (['__self__', '"""preemptible"""', 'preemptible'], {}), "(__self__, 'preemptible', preemptible)\n", (345456, 345494), False, 'import pulumi\n'), ((346997, 347061), 'pulumi.set', 'pulumi.set', (['__self__', '"""maintenance_windows"""', 'maintenance_windows'], {}), "(__self__, 'maintenance_windows', maintenance_windows)\n", (347007, 347061), False, 'import pulumi\n'), ((348971, 349003), 'pulumi.set', 'pulumi.set', (['__self__', '"""day"""', 'day'], {}), "(__self__, 'day', day)\n", (348981, 349003), False, 'import pulumi\n'), ((350457, 350503), 'pulumi.set', 'pulumi.set', (['__self__', '"""auto_scale"""', 'auto_scale'], {}), "(__self__, 'auto_scale', auto_scale)\n", (350467, 350503), False, 'import pulumi\n'), ((350552, 350600), 'pulumi.set', 'pulumi.set', (['__self__', '"""fixed_scale"""', 'fixed_scale'], {}), "(__self__, 'fixed_scale', fixed_scale)\n", (350562, 350600), False, 'import pulumi\n'), ((353424, 353458), 'pulumi.set', 'pulumi.set', (['__self__', '"""size"""', 'size'], {}), "(__self__, 'size', size)\n", (353434, 353458), False, 'import pulumi\n'), ((355061, 355117), 'pulumi.set', 'pulumi.set', (['__self__', '"""current_version"""', 'current_version'], {}), "(__self__, 'current_version', current_version)\n", (355071, 355117), False, 'import pulumi\n'), ((355177, 355247), 'pulumi.set', 'pulumi.set', (['__self__', '"""new_revision_available"""', 'new_revision_available'], {}), "(__self__, 'new_revision_available', new_revision_available)\n", (355187, 355247), False, 'import pulumi\n'), ((355305, 355371), 'pulumi.set', 'pulumi.set', (['__self__', '"""new_revision_summary"""', 'new_revision_summary'], {}), "(__self__, 'new_revision_summary', new_revision_summary)\n", (355315, 355371), False, 'import pulumi\n'), ((355427, 355489), 'pulumi.set', 'pulumi.set', (['__self__', '"""version_deprecated"""', 'version_deprecated'], {}), "(__self__, 'version_deprecated', version_deprecated)\n", (355437, 355489), False, 'import pulumi\n'), ((360921, 360981), 'pulumi.set', 'pulumi.set', (['__self__', '"""healthy_threshold"""', 'healthy_threshold'], {}), "(__self__, 'healthy_threshold', healthy_threshold)\n", (360931, 360981), False, 'import pulumi\n'), ((361031, 361081), 'pulumi.set', 'pulumi.set', (['__self__', '"""http_options"""', 'http_options'], {}), "(__self__, 'http_options', http_options)\n", (361041, 361081), False, 'import pulumi\n'), ((361127, 361169), 'pulumi.set', 'pulumi.set', (['__self__', '"""interval"""', 'interval'], {}), "(__self__, 'interval', interval)\n", (361137, 361169), False, 'import pulumi\n'), ((361218, 361266), 'pulumi.set', 'pulumi.set', (['__self__', '"""tcp_options"""', 'tcp_options'], {}), "(__self__, 'tcp_options', tcp_options)\n", (361228, 361266), False, 'import pulumi\n'), ((361311, 361351), 'pulumi.set', 'pulumi.set', (['__self__', '"""timeout"""', 'timeout'], {}), "(__self__, 'timeout', timeout)\n", (361321, 361351), False, 'import pulumi\n'), ((361408, 361472), 'pulumi.set', 'pulumi.set', (['__self__', '"""unhealthy_threshold"""', 'unhealthy_threshold'], {}), "(__self__, 'unhealthy_threshold', unhealthy_threshold)\n", (361418, 361472), False, 'import pulumi\n'), ((365080, 365114), 'pulumi.set', 'pulumi.set', (['__self__', '"""path"""', 'path'], {}), "(__self__, 'path', path)\n", (365090, 365114), False, 'import pulumi\n'), ((367866, 367934), 'pulumi.set', 'pulumi.set', (['__self__', '"""external_address_spec"""', 'external_address_spec'], {}), "(__self__, 'external_address_spec', external_address_spec)\n", (367876, 367934), False, 'import pulumi\n'), ((367993, 368061), 'pulumi.set', 'pulumi.set', (['__self__', '"""internal_address_spec"""', 'internal_address_spec'], {}), "(__self__, 'internal_address_spec', internal_address_spec)\n", (368003, 368061), False, 'import pulumi\n'), ((368107, 368149), 'pulumi.set', 'pulumi.set', (['__self__', '"""protocol"""', 'protocol'], {}), "(__self__, 'protocol', protocol)\n", (368117, 368149), False, 'import pulumi\n'), ((368198, 368246), 'pulumi.set', 'pulumi.set', (['__self__', '"""target_port"""', 'target_port'], {}), "(__self__, 'target_port', target_port)\n", (368208, 368246), False, 'import pulumi\n'), ((371386, 371426), 'pulumi.set', 'pulumi.set', (['__self__', '"""address"""', 'address'], {}), "(__self__, 'address', address)\n", (371396, 371426), False, 'import pulumi\n'), ((371474, 371520), 'pulumi.set', 'pulumi.set', (['__self__', '"""ip_version"""', 'ip_version'], {}), "(__self__, 'ip_version', ip_version)\n", (371484, 371520), False, 'import pulumi\n'), ((373261, 373301), 'pulumi.set', 'pulumi.set', (['__self__', '"""address"""', 'address'], {}), "(__self__, 'address', address)\n", (373271, 373301), False, 'import pulumi\n'), ((373349, 373395), 'pulumi.set', 'pulumi.set', (['__self__', '"""ip_version"""', 'ip_version'], {}), "(__self__, 'ip_version', ip_version)\n", (373359, 373395), False, 'import pulumi\n'), ((376790, 376834), 'pulumi.set', 'pulumi.set', (['__self__', '"""data_lens"""', 'data_lens'], {}), "(__self__, 'data_lens', data_lens)\n", (376800, 376834), False, 'import pulumi\n'), ((376879, 376919), 'pulumi.set', 'pulumi.set', (['__self__', '"""metrika"""', 'metrika'], {}), "(__self__, 'metrika', metrika)\n", (376889, 376919), False, 'import pulumi\n'), ((376967, 377013), 'pulumi.set', 'pulumi.set', (['__self__', '"""serverless"""', 'serverless'], {}), "(__self__, 'serverless', serverless)\n", (376977, 377013), False, 'import pulumi\n'), ((377058, 377098), 'pulumi.set', 'pulumi.set', (['__self__', '"""web_sql"""', 'web_sql'], {}), "(__self__, 'web_sql', web_sql)\n", (377068, 377098), False, 'import pulumi\n'), ((379012, 379048), 'pulumi.set', 'pulumi.set', (['__self__', '"""hours"""', 'hours'], {}), "(__self__, 'hours', hours)\n", (379022, 379048), False, 'import pulumi\n'), ((379093, 379133), 'pulumi.set', 'pulumi.set', (['__self__', '"""minutes"""', 'minutes'], {}), "(__self__, 'minutes', minutes)\n", (379103, 379133), False, 'import pulumi\n'), ((380483, 380521), 'pulumi.set', 'pulumi.set', (['__self__', '"""config"""', 'config'], {}), "(__self__, 'config', config)\n", (380493, 380521), False, 'import pulumi\n'), ((385562, 385628), 'pulumi.set', 'pulumi.set', (['__self__', '"""background_pool_size"""', 'background_pool_size'], {}), "(__self__, 'background_pool_size', background_pool_size)\n", (385572, 385628), False, 'import pulumi\n'), ((385695, 385783), 'pulumi.set', 'pulumi.set', (['__self__', '"""background_schedule_pool_size"""', 'background_schedule_pool_size'], {}), "(__self__, 'background_schedule_pool_size',\n background_schedule_pool_size)\n", (385705, 385783), False, 'import pulumi\n'), ((385829, 385879), 'pulumi.set', 'pulumi.set', (['__self__', '"""compressions"""', 'compressions'], {}), "(__self__, 'compressions', compressions)\n", (385839, 385879), False, 'import pulumi\n'), ((385928, 385976), 'pulumi.set', 'pulumi.set', (['__self__', '"""geobase_uri"""', 'geobase_uri'], {}), "(__self__, 'geobase_uri', geobase_uri)\n", (385938, 385976), False, 'import pulumi\n'), ((386030, 386088), 'pulumi.set', 'pulumi.set', (['__self__', '"""graphite_rollups"""', 'graphite_rollups'], {}), "(__self__, 'graphite_rollups', graphite_rollups)\n", (386040, 386088), False, 'import pulumi\n'), ((386131, 386167), 'pulumi.set', 'pulumi.set', (['__self__', '"""kafka"""', 'kafka'], {}), "(__self__, 'kafka', kafka)\n", (386141, 386167), False, 'import pulumi\n'), ((386217, 386267), 'pulumi.set', 'pulumi.set', (['__self__', '"""kafka_topics"""', 'kafka_topics'], {}), "(__self__, 'kafka_topics', kafka_topics)\n", (386227, 386267), False, 'import pulumi\n'), ((386323, 386385), 'pulumi.set', 'pulumi.set', (['__self__', '"""keep_alive_timeout"""', 'keep_alive_timeout'], {}), "(__self__, 'keep_alive_timeout', keep_alive_timeout)\n", (386333, 386385), False, 'import pulumi\n'), ((386432, 386476), 'pulumi.set', 'pulumi.set', (['__self__', '"""log_level"""', 'log_level'], {}), "(__self__, 'log_level', log_level)\n", (386442, 386476), False, 'import pulumi\n'), ((386529, 386585), 'pulumi.set', 'pulumi.set', (['__self__', '"""mark_cache_size"""', 'mark_cache_size'], {}), "(__self__, 'mark_cache_size', mark_cache_size)\n", (386539, 386585), False, 'import pulumi\n'), ((386645, 386715), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_concurrent_queries"""', 'max_concurrent_queries'], {}), "(__self__, 'max_concurrent_queries', max_concurrent_queries)\n", (386655, 386715), False, 'import pulumi\n'), ((386768, 386824), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_connections"""', 'max_connections'], {}), "(__self__, 'max_connections', max_connections)\n", (386778, 386824), False, 'import pulumi\n'), ((386888, 386966), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_partition_size_to_drop"""', 'max_partition_size_to_drop'], {}), "(__self__, 'max_partition_size_to_drop', max_partition_size_to_drop)\n", (386898, 386966), False, 'import pulumi\n'), ((387026, 387096), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_table_size_to_drop"""', 'max_table_size_to_drop'], {}), "(__self__, 'max_table_size_to_drop', max_table_size_to_drop)\n", (387036, 387096), False, 'import pulumi\n'), ((387144, 387190), 'pulumi.set', 'pulumi.set', (['__self__', '"""merge_tree"""', 'merge_tree'], {}), "(__self__, 'merge_tree', merge_tree)\n", (387154, 387190), False, 'import pulumi\n'), ((387246, 387308), 'pulumi.set', 'pulumi.set', (['__self__', '"""metric_log_enabled"""', 'metric_log_enabled'], {}), "(__self__, 'metric_log_enabled', metric_log_enabled)\n", (387256, 387308), False, 'import pulumi\n'), ((387371, 387447), 'pulumi.set', 'pulumi.set', (['__self__', '"""metric_log_retention_size"""', 'metric_log_retention_size'], {}), "(__self__, 'metric_log_retention_size', metric_log_retention_size)\n", (387381, 387447), False, 'import pulumi\n'), ((387510, 387586), 'pulumi.set', 'pulumi.set', (['__self__', '"""metric_log_retention_time"""', 'metric_log_retention_time'], {}), "(__self__, 'metric_log_retention_time', metric_log_retention_time)\n", (387520, 387586), False, 'import pulumi\n'), ((387647, 387719), 'pulumi.set', 'pulumi.set', (['__self__', '"""part_log_retention_size"""', 'part_log_retention_size'], {}), "(__self__, 'part_log_retention_size', part_log_retention_size)\n", (387657, 387719), False, 'import pulumi\n'), ((387780, 387852), 'pulumi.set', 'pulumi.set', (['__self__', '"""part_log_retention_time"""', 'part_log_retention_time'], {}), "(__self__, 'part_log_retention_time', part_log_retention_time)\n", (387790, 387852), False, 'import pulumi\n'), ((387914, 387988), 'pulumi.set', 'pulumi.set', (['__self__', '"""query_log_retention_size"""', 'query_log_retention_size'], {}), "(__self__, 'query_log_retention_size', query_log_retention_size)\n", (387924, 387988), False, 'import pulumi\n'), ((388050, 388124), 'pulumi.set', 'pulumi.set', (['__self__', '"""query_log_retention_time"""', 'query_log_retention_time'], {}), "(__self__, 'query_log_retention_time', query_log_retention_time)\n", (388060, 388124), False, 'import pulumi\n'), ((388186, 388260), 'pulumi.set', 'pulumi.set', (['__self__', '"""query_thread_log_enabled"""', 'query_thread_log_enabled'], {}), "(__self__, 'query_thread_log_enabled', query_thread_log_enabled)\n", (388196, 388260), False, 'import pulumi\n'), ((388329, 388421), 'pulumi.set', 'pulumi.set', (['__self__', '"""query_thread_log_retention_size"""', 'query_thread_log_retention_size'], {}), "(__self__, 'query_thread_log_retention_size',\n query_thread_log_retention_size)\n", (388339, 388421), False, 'import pulumi\n'), ((388486, 388578), 'pulumi.set', 'pulumi.set', (['__self__', '"""query_thread_log_retention_time"""', 'query_thread_log_retention_time'], {}), "(__self__, 'query_thread_log_retention_time',\n query_thread_log_retention_time)\n", (388496, 388578), False, 'import pulumi\n'), ((388620, 388662), 'pulumi.set', 'pulumi.set', (['__self__', '"""rabbitmq"""', 'rabbitmq'], {}), "(__self__, 'rabbitmq', rabbitmq)\n", (388630, 388662), False, 'import pulumi\n'), ((388716, 388774), 'pulumi.set', 'pulumi.set', (['__self__', '"""text_log_enabled"""', 'text_log_enabled'], {}), "(__self__, 'text_log_enabled', text_log_enabled)\n", (388726, 388774), False, 'import pulumi\n'), ((388826, 388880), 'pulumi.set', 'pulumi.set', (['__self__', '"""text_log_level"""', 'text_log_level'], {}), "(__self__, 'text_log_level', text_log_level)\n", (388836, 388880), False, 'import pulumi\n'), ((388941, 389013), 'pulumi.set', 'pulumi.set', (['__self__', '"""text_log_retention_size"""', 'text_log_retention_size'], {}), "(__self__, 'text_log_retention_size', text_log_retention_size)\n", (388951, 389013), False, 'import pulumi\n'), ((389074, 389146), 'pulumi.set', 'pulumi.set', (['__self__', '"""text_log_retention_time"""', 'text_log_retention_time'], {}), "(__self__, 'text_log_retention_time', text_log_retention_time)\n", (389084, 389146), False, 'import pulumi\n'), ((389192, 389234), 'pulumi.set', 'pulumi.set', (['__self__', '"""timezone"""', 'timezone'], {}), "(__self__, 'timezone', timezone)\n", (389202, 389234), False, 'import pulumi\n'), ((389289, 389349), 'pulumi.set', 'pulumi.set', (['__self__', '"""trace_log_enabled"""', 'trace_log_enabled'], {}), "(__self__, 'trace_log_enabled', trace_log_enabled)\n", (389299, 389349), False, 'import pulumi\n'), ((389411, 389485), 'pulumi.set', 'pulumi.set', (['__self__', '"""trace_log_retention_size"""', 'trace_log_retention_size'], {}), "(__self__, 'trace_log_retention_size', trace_log_retention_size)\n", (389421, 389485), False, 'import pulumi\n'), ((389547, 389621), 'pulumi.set', 'pulumi.set', (['__self__', '"""trace_log_retention_time"""', 'trace_log_retention_time'], {}), "(__self__, 'trace_log_retention_time', trace_log_retention_time)\n", (389557, 389621), False, 'import pulumi\n'), ((389682, 389754), 'pulumi.set', 'pulumi.set', (['__self__', '"""uncompressed_cache_size"""', 'uncompressed_cache_size'], {}), "(__self__, 'uncompressed_cache_size', uncompressed_cache_size)\n", (389692, 389754), False, 'import pulumi\n'), ((406009, 406051), 'pulumi.set', 'pulumi.set', (['__self__', '"""patterns"""', 'patterns'], {}), "(__self__, 'patterns', patterns)\n", (406019, 406051), False, 'import pulumi\n'), ((407645, 407683), 'pulumi.set', 'pulumi.set', (['__self__', '"""regexp"""', 'regexp'], {}), "(__self__, 'regexp', regexp)\n", (407655, 407683), False, 'import pulumi\n'), ((407731, 407777), 'pulumi.set', 'pulumi.set', (['__self__', '"""retentions"""', 'retentions'], {}), "(__self__, 'retentions', retentions)\n", (407741, 407777), False, 'import pulumi\n'), ((410829, 410883), 'pulumi.set', 'pulumi.set', (['__self__', '"""sasl_mechanism"""', 'sasl_mechanism'], {}), "(__self__, 'sasl_mechanism', sasl_mechanism)\n", (410839, 410883), False, 'import pulumi\n'), ((410934, 410986), 'pulumi.set', 'pulumi.set', (['__self__', '"""sasl_password"""', 'sasl_password'], {}), "(__self__, 'sasl_password', sasl_password)\n", (410944, 410986), False, 'import pulumi\n'), ((411037, 411089), 'pulumi.set', 'pulumi.set', (['__self__', '"""sasl_username"""', 'sasl_username'], {}), "(__self__, 'sasl_username', sasl_username)\n", (411047, 411089), False, 'import pulumi\n'), ((411144, 411204), 'pulumi.set', 'pulumi.set', (['__self__', '"""security_protocol"""', 'security_protocol'], {}), "(__self__, 'security_protocol', security_protocol)\n", (411154, 411204), False, 'import pulumi\n'), ((413344, 413386), 'pulumi.set', 'pulumi.set', (['__self__', '"""settings"""', 'settings'], {}), "(__self__, 'settings', settings)\n", (413354, 413386), False, 'import pulumi\n'), ((414945, 414999), 'pulumi.set', 'pulumi.set', (['__self__', '"""sasl_mechanism"""', 'sasl_mechanism'], {}), "(__self__, 'sasl_mechanism', sasl_mechanism)\n", (414955, 414999), False, 'import pulumi\n'), ((415050, 415102), 'pulumi.set', 'pulumi.set', (['__self__', '"""sasl_password"""', 'sasl_password'], {}), "(__self__, 'sasl_password', sasl_password)\n", (415060, 415102), False, 'import pulumi\n'), ((415153, 415205), 'pulumi.set', 'pulumi.set', (['__self__', '"""sasl_username"""', 'sasl_username'], {}), "(__self__, 'sasl_username', sasl_username)\n", (415163, 415205), False, 'import pulumi\n'), ((415260, 415320), 'pulumi.set', 'pulumi.set', (['__self__', '"""security_protocol"""', 'security_protocol'], {}), "(__self__, 'security_protocol', security_protocol)\n", (415270, 415320), False, 'import pulumi\n'), ((419309, 419417), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_bytes_to_merge_at_min_space_in_pool"""', 'max_bytes_to_merge_at_min_space_in_pool'], {}), "(__self__, 'max_bytes_to_merge_at_min_space_in_pool',\n max_bytes_to_merge_at_min_space_in_pool)\n", (419319, 419417), False, 'import pulumi\n'), ((419481, 419571), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_replicated_merges_in_queue"""', 'max_replicated_merges_in_queue'], {}), "(__self__, 'max_replicated_merges_in_queue',\n max_replicated_merges_in_queue)\n", (419491, 419571), False, 'import pulumi\n'), ((419662, 419810), 'pulumi.set', 'pulumi.set', (['__self__', '"""number_of_free_entries_in_pool_to_lower_max_size_of_merge"""', 'number_of_free_entries_in_pool_to_lower_max_size_of_merge'], {}), "(__self__,\n 'number_of_free_entries_in_pool_to_lower_max_size_of_merge',\n number_of_free_entries_in_pool_to_lower_max_size_of_merge)\n", (419672, 419810), False, 'import pulumi\n'), ((419861, 419929), 'pulumi.set', 'pulumi.set', (['__self__', '"""parts_to_delay_insert"""', 'parts_to_delay_insert'], {}), "(__self__, 'parts_to_delay_insert', parts_to_delay_insert)\n", (419871, 419929), False, 'import pulumi\n'), ((419988, 420056), 'pulumi.set', 'pulumi.set', (['__self__', '"""parts_to_throw_insert"""', 'parts_to_throw_insert'], {}), "(__self__, 'parts_to_throw_insert', parts_to_throw_insert)\n", (419998, 420056), False, 'import pulumi\n'), ((420125, 420217), 'pulumi.set', 'pulumi.set', (['__self__', '"""replicated_deduplication_window"""', 'replicated_deduplication_window'], {}), "(__self__, 'replicated_deduplication_window',\n replicated_deduplication_window)\n", (420135, 420217), False, 'import pulumi\n'), ((420290, 420398), 'pulumi.set', 'pulumi.set', (['__self__', '"""replicated_deduplication_window_seconds"""', 'replicated_deduplication_window_seconds'], {}), "(__self__, 'replicated_deduplication_window_seconds',\n replicated_deduplication_window_seconds)\n", (420300, 420398), False, 'import pulumi\n'), ((425171, 425213), 'pulumi.set', 'pulumi.set', (['__self__', '"""password"""', 'password'], {}), "(__self__, 'password', password)\n", (425181, 425213), False, 'import pulumi\n'), ((425259, 425301), 'pulumi.set', 'pulumi.set', (['__self__', '"""username"""', 'username'], {}), "(__self__, 'username', username)\n", (425269, 425301), False, 'import pulumi\n'), ((432373, 432431), 'pulumi.set', 'pulumi.set', (['__self__', '"""assign_public_ip"""', 'assign_public_ip'], {}), "(__self__, 'assign_public_ip', assign_public_ip)\n", (432383, 432431), False, 'import pulumi\n'), ((432473, 432507), 'pulumi.set', 'pulumi.set', (['__self__', '"""fqdn"""', 'fqdn'], {}), "(__self__, 'fqdn', fqdn)\n", (432483, 432507), False, 'import pulumi\n'), ((432555, 432601), 'pulumi.set', 'pulumi.set', (['__self__', '"""shard_name"""', 'shard_name'], {}), "(__self__, 'shard_name', shard_name)\n", (432565, 432601), False, 'import pulumi\n'), ((432648, 432692), 'pulumi.set', 'pulumi.set', (['__self__', '"""subnet_id"""', 'subnet_id'], {}), "(__self__, 'subnet_id', subnet_id)\n", (432658, 432692), False, 'import pulumi\n'), ((435911, 435943), 'pulumi.set', 'pulumi.set', (['__self__', '"""day"""', 'day'], {}), "(__self__, 'day', day)\n", (435921, 435943), False, 'import pulumi\n'), ((435985, 436019), 'pulumi.set', 'pulumi.set', (['__self__', '"""hour"""', 'hour'], {}), "(__self__, 'hour', hour)\n", (435995, 436019), False, 'import pulumi\n'), ((439585, 439633), 'pulumi.set', 'pulumi.set', (['__self__', '"""description"""', 'description'], {}), "(__self__, 'description', description)\n", (439595, 439633), False, 'import pulumi\n'), ((441940, 441988), 'pulumi.set', 'pulumi.set', (['__self__', '"""permissions"""', 'permissions'], {}), "(__self__, 'permissions', permissions)\n", (441950, 441988), False, 'import pulumi\n'), ((442032, 442070), 'pulumi.set', 'pulumi.set', (['__self__', '"""quotas"""', 'quotas'], {}), "(__self__, 'quotas', quotas)\n", (442042, 442070), False, 'import pulumi\n'), ((442116, 442158), 'pulumi.set', 'pulumi.set', (['__self__', '"""settings"""', 'settings'], {}), "(__self__, 'settings', settings)\n", (442126, 442158), False, 'import pulumi\n'), ((446030, 446068), 'pulumi.set', 'pulumi.set', (['__self__', '"""errors"""', 'errors'], {}), "(__self__, 'errors', errors)\n", (446040, 446068), False, 'import pulumi\n'), ((446120, 446174), 'pulumi.set', 'pulumi.set', (['__self__', '"""execution_time"""', 'execution_time'], {}), "(__self__, 'execution_time', execution_time)\n", (446130, 446174), False, 'import pulumi\n'), ((446219, 446259), 'pulumi.set', 'pulumi.set', (['__self__', '"""queries"""', 'queries'], {}), "(__self__, 'queries', queries)\n", (446229, 446259), False, 'import pulumi\n'), ((446306, 446350), 'pulumi.set', 'pulumi.set', (['__self__', '"""read_rows"""', 'read_rows'], {}), "(__self__, 'read_rows', read_rows)\n", (446316, 446350), False, 'import pulumi\n'), ((446399, 446447), 'pulumi.set', 'pulumi.set', (['__self__', '"""result_rows"""', 'result_rows'], {}), "(__self__, 'result_rows', result_rows)\n", (446409, 446447), False, 'import pulumi\n'), ((468784, 468850), 'pulumi.set', 'pulumi.set', (['__self__', '"""add_http_cors_header"""', 'add_http_cors_header'], {}), "(__self__, 'add_http_cors_header', add_http_cors_header)\n", (468794, 468850), False, 'import pulumi\n'), ((468897, 468941), 'pulumi.set', 'pulumi.set', (['__self__', '"""allow_ddl"""', 'allow_ddl'], {}), "(__self__, 'allow_ddl', allow_ddl)\n", (468907, 468941), False, 'import pulumi\n'), ((468986, 469026), 'pulumi.set', 'pulumi.set', (['__self__', '"""compile"""', 'compile'], {}), "(__self__, 'compile', compile)\n", (468996, 469026), False, 'import pulumi\n'), ((469083, 469147), 'pulumi.set', 'pulumi.set', (['__self__', '"""compile_expressions"""', 'compile_expressions'], {}), "(__self__, 'compile_expressions', compile_expressions)\n", (469093, 469147), False, 'import pulumi\n'), ((469200, 469256), 'pulumi.set', 'pulumi.set', (['__self__', '"""connect_timeout"""', 'connect_timeout'], {}), "(__self__, 'connect_timeout', connect_timeout)\n", (469210, 469256), False, 'import pulumi\n'), ((469323, 469411), 'pulumi.set', 'pulumi.set', (['__self__', '"""count_distinct_implementation"""', 'count_distinct_implementation'], {}), "(__self__, 'count_distinct_implementation',\n count_distinct_implementation)\n", (469333, 469411), False, 'import pulumi\n'), ((469467, 469537), 'pulumi.set', 'pulumi.set', (['__self__', '"""distinct_overflow_mode"""', 'distinct_overflow_mode'], {}), "(__self__, 'distinct_overflow_mode', distinct_overflow_mode)\n", (469477, 469537), False, 'import pulumi\n'), ((469615, 469725), 'pulumi.set', 'pulumi.set', (['__self__', '"""distributed_aggregation_memory_efficient"""', 'distributed_aggregation_memory_efficient'], {}), "(__self__, 'distributed_aggregation_memory_efficient',\n distributed_aggregation_memory_efficient)\n", (469625, 469725), False, 'import pulumi\n'), ((469787, 469873), 'pulumi.set', 'pulumi.set', (['__self__', '"""distributed_ddl_task_timeout"""', 'distributed_ddl_task_timeout'], {}), "(__self__, 'distributed_ddl_task_timeout',\n distributed_ddl_task_timeout)\n", (469797, 469873), False, 'import pulumi\n'), ((469931, 470005), 'pulumi.set', 'pulumi.set', (['__self__', '"""distributed_product_mode"""', 'distributed_product_mode'], {}), "(__self__, 'distributed_product_mode', distributed_product_mode)\n", (469941, 470005), False, 'import pulumi\n'), ((470084, 470196), 'pulumi.set', 'pulumi.set', (['__self__', '"""empty_result_for_aggregation_by_empty_set"""', 'empty_result_for_aggregation_by_empty_set'], {}), "(__self__, 'empty_result_for_aggregation_by_empty_set',\n empty_result_for_aggregation_by_empty_set)\n", (470094, 470196), False, 'import pulumi\n'), ((470253, 470325), 'pulumi.set', 'pulumi.set', (['__self__', '"""enable_http_compression"""', 'enable_http_compression'], {}), "(__self__, 'enable_http_compression', enable_http_compression)\n", (470263, 470325), False, 'import pulumi\n'), ((470413, 470543), 'pulumi.set', 'pulumi.set', (['__self__', '"""fallback_to_stale_replicas_for_distributed_queries"""', 'fallback_to_stale_replicas_for_distributed_queries'], {}), "(__self__, 'fallback_to_stale_replicas_for_distributed_queries',\n fallback_to_stale_replicas_for_distributed_queries)\n", (470423, 470543), False, 'import pulumi\n'), ((470596, 470660), 'pulumi.set', 'pulumi.set', (['__self__', '"""force_index_by_date"""', 'force_index_by_date'], {}), "(__self__, 'force_index_by_date', force_index_by_date)\n", (470606, 470660), False, 'import pulumi\n'), ((470715, 470775), 'pulumi.set', 'pulumi.set', (['__self__', '"""force_primary_key"""', 'force_primary_key'], {}), "(__self__, 'force_primary_key', force_primary_key)\n", (470725, 470775), False, 'import pulumi\n'), ((470835, 470905), 'pulumi.set', 'pulumi.set', (['__self__', '"""group_by_overflow_mode"""', 'group_by_overflow_mode'], {}), "(__self__, 'group_by_overflow_mode', group_by_overflow_mode)\n", (470845, 470905), False, 'import pulumi\n'), ((470971, 471057), 'pulumi.set', 'pulumi.set', (['__self__', '"""group_by_two_level_threshold"""', 'group_by_two_level_threshold'], {}), "(__self__, 'group_by_two_level_threshold',\n group_by_two_level_threshold)\n", (470981, 471057), False, 'import pulumi\n'), ((471125, 471223), 'pulumi.set', 'pulumi.set', (['__self__', '"""group_by_two_level_threshold_bytes"""', 'group_by_two_level_threshold_bytes'], {}), "(__self__, 'group_by_two_level_threshold_bytes',\n group_by_two_level_threshold_bytes)\n", (471135, 471223), False, 'import pulumi\n'), ((471280, 471352), 'pulumi.set', 'pulumi.set', (['__self__', '"""http_connection_timeout"""', 'http_connection_timeout'], {}), "(__self__, 'http_connection_timeout', http_connection_timeout)\n", (471290, 471352), False, 'import pulumi\n'), ((471420, 471510), 'pulumi.set', 'pulumi.set', (['__self__', '"""http_headers_progress_interval"""', 'http_headers_progress_interval'], {}), "(__self__, 'http_headers_progress_interval',\n http_headers_progress_interval)\n", (471430, 471510), False, 'import pulumi\n'), ((471564, 471630), 'pulumi.set', 'pulumi.set', (['__self__', '"""http_receive_timeout"""', 'http_receive_timeout'], {}), "(__self__, 'http_receive_timeout', http_receive_timeout)\n", (471574, 471630), False, 'import pulumi\n'), ((471685, 471745), 'pulumi.set', 'pulumi.set', (['__self__', '"""http_send_timeout"""', 'http_send_timeout'], {}), "(__self__, 'http_send_timeout', http_send_timeout)\n", (471695, 471745), False, 'import pulumi\n'), ((471823, 471933), 'pulumi.set', 'pulumi.set', (['__self__', '"""input_format_defaults_for_omitted_fields"""', 'input_format_defaults_for_omitted_fields'], {}), "(__self__, 'input_format_defaults_for_omitted_fields',\n input_format_defaults_for_omitted_fields)\n", (471833, 471933), False, 'import pulumi\n'), ((472008, 472120), 'pulumi.set', 'pulumi.set', (['__self__', '"""input_format_values_interpret_expressions"""', 'input_format_values_interpret_expressions'], {}), "(__self__, 'input_format_values_interpret_expressions',\n input_format_values_interpret_expressions)\n", (472018, 472120), False, 'import pulumi\n'), ((472167, 472219), 'pulumi.set', 'pulumi.set', (['__self__', '"""insert_quorum"""', 'insert_quorum'], {}), "(__self__, 'insert_quorum', insert_quorum)\n", (472177, 472219), False, 'import pulumi\n'), ((472278, 472346), 'pulumi.set', 'pulumi.set', (['__self__', '"""insert_quorum_timeout"""', 'insert_quorum_timeout'], {}), "(__self__, 'insert_quorum_timeout', insert_quorum_timeout)\n", (472288, 472346), False, 'import pulumi\n'), ((472402, 472464), 'pulumi.set', 'pulumi.set', (['__self__', '"""join_overflow_mode"""', 'join_overflow_mode'], {}), "(__self__, 'join_overflow_mode', join_overflow_mode)\n", (472412, 472464), False, 'import pulumi\n'), ((472516, 472570), 'pulumi.set', 'pulumi.set', (['__self__', '"""join_use_nulls"""', 'join_use_nulls'], {}), "(__self__, 'join_use_nulls', join_use_nulls)\n", (472526, 472570), False, 'import pulumi\n'), ((472638, 472728), 'pulumi.set', 'pulumi.set', (['__self__', '"""joined_subquery_requires_alias"""', 'joined_subquery_requires_alias'], {}), "(__self__, 'joined_subquery_requires_alias',\n joined_subquery_requires_alias)\n", (472648, 472728), False, 'import pulumi\n'), ((472800, 472906), 'pulumi.set', 'pulumi.set', (['__self__', '"""low_cardinality_allow_in_native_format"""', 'low_cardinality_allow_in_native_format'], {}), "(__self__, 'low_cardinality_allow_in_native_format',\n low_cardinality_allow_in_native_format)\n", (472810, 472906), False, 'import pulumi\n'), ((472953, 473005), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_ast_depth"""', 'max_ast_depth'], {}), "(__self__, 'max_ast_depth', max_ast_depth)\n", (472963, 473005), False, 'import pulumi\n'), ((473059, 473117), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_ast_elements"""', 'max_ast_elements'], {}), "(__self__, 'max_ast_elements', max_ast_elements)\n", (473069, 473117), False, 'import pulumi\n'), ((473169, 473223), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_block_size"""', 'max_block_size'], {}), "(__self__, 'max_block_size', max_block_size)\n", (473179, 473223), False, 'import pulumi\n'), ((473295, 473393), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_bytes_before_external_group_by"""', 'max_bytes_before_external_group_by'], {}), "(__self__, 'max_bytes_before_external_group_by',\n max_bytes_before_external_group_by)\n", (473305, 473393), False, 'import pulumi\n'), ((473457, 473547), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_bytes_before_external_sort"""', 'max_bytes_before_external_sort'], {}), "(__self__, 'max_bytes_before_external_sort',\n max_bytes_before_external_sort)\n", (473467, 473547), False, 'import pulumi\n'), ((473602, 473670), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_bytes_in_distinct"""', 'max_bytes_in_distinct'], {}), "(__self__, 'max_bytes_in_distinct', max_bytes_in_distinct)\n", (473612, 473670), False, 'import pulumi\n'), ((473725, 473785), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_bytes_in_join"""', 'max_bytes_in_join'], {}), "(__self__, 'max_bytes_in_join', max_bytes_in_join)\n", (473735, 473785), False, 'import pulumi\n'), ((473839, 473897), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_bytes_in_set"""', 'max_bytes_in_set'], {}), "(__self__, 'max_bytes_in_set', max_bytes_in_set)\n", (473849, 473897), False, 'import pulumi\n'), ((473952, 474012), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_bytes_to_read"""', 'max_bytes_to_read'], {}), "(__self__, 'max_bytes_to_read', max_bytes_to_read)\n", (473962, 474012), False, 'import pulumi\n'), ((474067, 474127), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_bytes_to_sort"""', 'max_bytes_to_sort'], {}), "(__self__, 'max_bytes_to_sort', max_bytes_to_sort)\n", (474077, 474127), False, 'import pulumi\n'), ((474186, 474254), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_bytes_to_transfer"""', 'max_bytes_to_transfer'], {}), "(__self__, 'max_bytes_to_transfer', max_bytes_to_transfer)\n", (474196, 474254), False, 'import pulumi\n'), ((474311, 474375), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_columns_to_read"""', 'max_columns_to_read'], {}), "(__self__, 'max_columns_to_read', max_columns_to_read)\n", (474321, 474375), False, 'import pulumi\n'), ((474431, 474493), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_execution_time"""', 'max_execution_time'], {}), "(__self__, 'max_execution_time', max_execution_time)\n", (474441, 474493), False, 'import pulumi\n'), ((474556, 474632), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_expanded_ast_elements"""', 'max_expanded_ast_elements'], {}), "(__self__, 'max_expanded_ast_elements', max_expanded_ast_elements)\n", (474566, 474632), False, 'import pulumi\n'), ((474691, 474759), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_insert_block_size"""', 'max_insert_block_size'], {}), "(__self__, 'max_insert_block_size', max_insert_block_size)\n", (474701, 474759), False, 'import pulumi\n'), ((474813, 474871), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_memory_usage"""', 'max_memory_usage'], {}), "(__self__, 'max_memory_usage', max_memory_usage)\n", (474823, 474871), False, 'import pulumi\n'), ((474934, 475010), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_memory_usage_for_user"""', 'max_memory_usage_for_user'], {}), "(__self__, 'max_memory_usage_for_user', max_memory_usage_for_user)\n", (474944, 475010), False, 'import pulumi\n'), ((475069, 475137), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_network_bandwidth"""', 'max_network_bandwidth'], {}), "(__self__, 'max_network_bandwidth', max_network_bandwidth)\n", (475079, 475137), False, 'import pulumi\n'), ((475205, 475295), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_network_bandwidth_for_user"""', 'max_network_bandwidth_for_user'], {}), "(__self__, 'max_network_bandwidth_for_user',\n max_network_bandwidth_for_user)\n", (475215, 475295), False, 'import pulumi\n'), ((475343, 475397), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_query_size"""', 'max_query_size'], {}), "(__self__, 'max_query_size', max_query_size)\n", (475353, 475397), False, 'import pulumi\n'), ((475476, 475588), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_replica_delay_for_distributed_queries"""', 'max_replica_delay_for_distributed_queries'], {}), "(__self__, 'max_replica_delay_for_distributed_queries',\n max_replica_delay_for_distributed_queries)\n", (475486, 475588), False, 'import pulumi\n'), ((475638, 475696), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_result_bytes"""', 'max_result_bytes'], {}), "(__self__, 'max_result_bytes', max_result_bytes)\n", (475648, 475696), False, 'import pulumi\n'), ((475749, 475805), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_result_rows"""', 'max_result_rows'], {}), "(__self__, 'max_result_rows', max_result_rows)\n", (475759, 475805), False, 'import pulumi\n'), ((475863, 475929), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_rows_in_distinct"""', 'max_rows_in_distinct'], {}), "(__self__, 'max_rows_in_distinct', max_rows_in_distinct)\n", (475873, 475929), False, 'import pulumi\n'), ((475983, 476041), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_rows_in_join"""', 'max_rows_in_join'], {}), "(__self__, 'max_rows_in_join', max_rows_in_join)\n", (475993, 476041), False, 'import pulumi\n'), ((476094, 476150), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_rows_in_set"""', 'max_rows_in_set'], {}), "(__self__, 'max_rows_in_set', max_rows_in_set)\n", (476104, 476150), False, 'import pulumi\n'), ((476208, 476274), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_rows_to_group_by"""', 'max_rows_to_group_by'], {}), "(__self__, 'max_rows_to_group_by', max_rows_to_group_by)\n", (476218, 476274), False, 'import pulumi\n'), ((476328, 476386), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_rows_to_read"""', 'max_rows_to_read'], {}), "(__self__, 'max_rows_to_read', max_rows_to_read)\n", (476338, 476386), False, 'import pulumi\n'), ((476440, 476498), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_rows_to_sort"""', 'max_rows_to_sort'], {}), "(__self__, 'max_rows_to_sort', max_rows_to_sort)\n", (476450, 476498), False, 'import pulumi\n'), ((476556, 476622), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_rows_to_transfer"""', 'max_rows_to_transfer'], {}), "(__self__, 'max_rows_to_transfer', max_rows_to_transfer)\n", (476566, 476622), False, 'import pulumi\n'), ((476681, 476749), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_temporary_columns"""', 'max_temporary_columns'], {}), "(__self__, 'max_temporary_columns', max_temporary_columns)\n", (476691, 476749), False, 'import pulumi\n'), ((476818, 476910), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_temporary_non_const_columns"""', 'max_temporary_non_const_columns'], {}), "(__self__, 'max_temporary_non_const_columns',\n max_temporary_non_const_columns)\n", (476828, 476910), False, 'import pulumi\n'), ((476955, 477003), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_threads"""', 'max_threads'], {}), "(__self__, 'max_threads', max_threads)\n", (476965, 477003), False, 'import pulumi\n'), ((477074, 477170), 'pulumi.set', 'pulumi.set', (['__self__', '"""merge_tree_max_bytes_to_use_cache"""', 'merge_tree_max_bytes_to_use_cache'], {}), "(__self__, 'merge_tree_max_bytes_to_use_cache',\n merge_tree_max_bytes_to_use_cache)\n", (477084, 477170), False, 'import pulumi\n'), ((477236, 477330), 'pulumi.set', 'pulumi.set', (['__self__', '"""merge_tree_max_rows_to_use_cache"""', 'merge_tree_max_rows_to_use_cache'], {}), "(__self__, 'merge_tree_max_rows_to_use_cache',\n merge_tree_max_rows_to_use_cache)\n", (477246, 477330), False, 'import pulumi\n'), ((477404, 477514), 'pulumi.set', 'pulumi.set', (['__self__', '"""merge_tree_min_bytes_for_concurrent_read"""', 'merge_tree_min_bytes_for_concurrent_read'], {}), "(__self__, 'merge_tree_min_bytes_for_concurrent_read',\n merge_tree_min_bytes_for_concurrent_read)\n", (477414, 477514), False, 'import pulumi\n'), ((477587, 477695), 'pulumi.set', 'pulumi.set', (['__self__', '"""merge_tree_min_rows_for_concurrent_read"""', 'merge_tree_min_rows_for_concurrent_read'], {}), "(__self__, 'merge_tree_min_rows_for_concurrent_read',\n merge_tree_min_rows_for_concurrent_read)\n", (477597, 477695), False, 'import pulumi\n'), ((477755, 477833), 'pulumi.set', 'pulumi.set', (['__self__', '"""min_bytes_to_use_direct_io"""', 'min_bytes_to_use_direct_io'], {}), "(__self__, 'min_bytes_to_use_direct_io', min_bytes_to_use_direct_io)\n", (477765, 477833), False, 'import pulumi\n'), ((477891, 477957), 'pulumi.set', 'pulumi.set', (['__self__', '"""min_count_to_compile"""', 'min_count_to_compile'], {}), "(__self__, 'min_count_to_compile', min_count_to_compile)\n", (477901, 477957), False, 'import pulumi\n'), ((478026, 478118), 'pulumi.set', 'pulumi.set', (['__self__', '"""min_count_to_compile_expression"""', 'min_count_to_compile_expression'], {}), "(__self__, 'min_count_to_compile_expression',\n min_count_to_compile_expression)\n", (478036, 478118), False, 'import pulumi\n'), ((478171, 478235), 'pulumi.set', 'pulumi.set', (['__self__', '"""min_execution_speed"""', 'min_execution_speed'], {}), "(__self__, 'min_execution_speed', min_execution_speed)\n", (478181, 478235), False, 'import pulumi\n'), ((478298, 478374), 'pulumi.set', 'pulumi.set', (['__self__', '"""min_execution_speed_bytes"""', 'min_execution_speed_bytes'], {}), "(__self__, 'min_execution_speed_bytes', min_execution_speed_bytes)\n", (478308, 478374), False, 'import pulumi\n'), ((478439, 478524), 'pulumi.set', 'pulumi.set', (['__self__', '"""min_insert_block_size_bytes"""', 'min_insert_block_size_bytes'], {}), "(__self__, 'min_insert_block_size_bytes', min_insert_block_size_bytes\n )\n", (478449, 478524), False, 'import pulumi\n'), ((478583, 478661), 'pulumi.set', 'pulumi.set', (['__self__', '"""min_insert_block_size_rows"""', 'min_insert_block_size_rows'], {}), "(__self__, 'min_insert_block_size_rows', min_insert_block_size_rows)\n", (478593, 478661), False, 'import pulumi\n'), ((478737, 478843), 'pulumi.set', 'pulumi.set', (['__self__', '"""output_format_json_quote64bit_integers"""', 'output_format_json_quote64bit_integers'], {}), "(__self__, 'output_format_json_quote64bit_integers',\n output_format_json_quote64bit_integers)\n", (478747, 478843), False, 'import pulumi\n'), ((478911, 479009), 'pulumi.set', 'pulumi.set', (['__self__', '"""output_format_json_quote_denormals"""', 'output_format_json_quote_denormals'], {}), "(__self__, 'output_format_json_quote_denormals',\n output_format_json_quote_denormals)\n", (478921, 479009), False, 'import pulumi\n'), ((479051, 479093), 'pulumi.set', 'pulumi.set', (['__self__', '"""priority"""', 'priority'], {}), "(__self__, 'priority', priority)\n", (479061, 479093), False, 'import pulumi\n'), ((479141, 479187), 'pulumi.set', 'pulumi.set', (['__self__', '"""quota_mode"""', 'quota_mode'], {}), "(__self__, 'quota_mode', quota_mode)\n", (479151, 479187), False, 'import pulumi\n'), ((479243, 479305), 'pulumi.set', 'pulumi.set', (['__self__', '"""read_overflow_mode"""', 'read_overflow_mode'], {}), "(__self__, 'read_overflow_mode', read_overflow_mode)\n", (479253, 479305), False, 'import pulumi\n'), ((479351, 479393), 'pulumi.set', 'pulumi.set', (['__self__', '"""readonly"""', 'readonly'], {}), "(__self__, 'readonly', readonly)\n", (479361, 479393), False, 'import pulumi\n'), ((479446, 479502), 'pulumi.set', 'pulumi.set', (['__self__', '"""receive_timeout"""', 'receive_timeout'], {}), "(__self__, 'receive_timeout', receive_timeout)\n", (479456, 479502), False, 'import pulumi\n'), ((479573, 479669), 'pulumi.set', 'pulumi.set', (['__self__', '"""replication_alter_partitions_sync"""', 'replication_alter_partitions_sync'], {}), "(__self__, 'replication_alter_partitions_sync',\n replication_alter_partitions_sync)\n", (479583, 479669), False, 'import pulumi\n'), ((479723, 479789), 'pulumi.set', 'pulumi.set', (['__self__', '"""result_overflow_mode"""', 'result_overflow_mode'], {}), "(__self__, 'result_overflow_mode', result_overflow_mode)\n", (479733, 479789), False, 'import pulumi\n'), ((479856, 479944), 'pulumi.set', 'pulumi.set', (['__self__', '"""select_sequential_consistency"""', 'select_sequential_consistency'], {}), "(__self__, 'select_sequential_consistency',\n select_sequential_consistency)\n", (479866, 479944), False, 'import pulumi\n'), ((480007, 480095), 'pulumi.set', 'pulumi.set', (['__self__', '"""send_progress_in_http_headers"""', 'send_progress_in_http_headers'], {}), "(__self__, 'send_progress_in_http_headers',\n send_progress_in_http_headers)\n", (480017, 480095), False, 'import pulumi\n'), ((480141, 480191), 'pulumi.set', 'pulumi.set', (['__self__', '"""send_timeout"""', 'send_timeout'], {}), "(__self__, 'send_timeout', send_timeout)\n", (480151, 480191), False, 'import pulumi\n'), ((480246, 480306), 'pulumi.set', 'pulumi.set', (['__self__', '"""set_overflow_mode"""', 'set_overflow_mode'], {}), "(__self__, 'set_overflow_mode', set_overflow_mode)\n", (480256, 480306), False, 'import pulumi\n'), ((480367, 480439), 'pulumi.set', 'pulumi.set', (['__self__', '"""skip_unavailable_shards"""', 'skip_unavailable_shards'], {}), "(__self__, 'skip_unavailable_shards', skip_unavailable_shards)\n", (480377, 480439), False, 'import pulumi\n'), ((480495, 480557), 'pulumi.set', 'pulumi.set', (['__self__', '"""sort_overflow_mode"""', 'sort_overflow_mode'], {}), "(__self__, 'sort_overflow_mode', sort_overflow_mode)\n", (480505, 480557), False, 'import pulumi\n'), ((480616, 480684), 'pulumi.set', 'pulumi.set', (['__self__', '"""timeout_overflow_mode"""', 'timeout_overflow_mode'], {}), "(__self__, 'timeout_overflow_mode', timeout_overflow_mode)\n", (480626, 480684), False, 'import pulumi\n'), ((480744, 480814), 'pulumi.set', 'pulumi.set', (['__self__', '"""transfer_overflow_mode"""', 'transfer_overflow_mode'], {}), "(__self__, 'transfer_overflow_mode', transfer_overflow_mode)\n", (480754, 480814), False, 'import pulumi\n'), ((480869, 480929), 'pulumi.set', 'pulumi.set', (['__self__', '"""transform_null_in"""', 'transform_null_in'], {}), "(__self__, 'transform_null_in', transform_null_in)\n", (480879, 480929), False, 'import pulumi\n'), ((480989, 481059), 'pulumi.set', 'pulumi.set', (['__self__', '"""use_uncompressed_cache"""', 'use_uncompressed_cache'], {}), "(__self__, 'use_uncompressed_cache', use_uncompressed_cache)\n", (480999, 481059), False, 'import pulumi\n'), ((525456, 525500), 'pulumi.set', 'pulumi.set', (['__self__', '"""resources"""', 'resources'], {}), "(__self__, 'resources', resources)\n", (525466, 525500), False, 'import pulumi\n'), ((526683, 526727), 'pulumi.set', 'pulumi.set', (['__self__', '"""disk_size"""', 'disk_size'], {}), "(__self__, 'disk_size', disk_size)\n", (526693, 526727), False, 'import pulumi\n'), ((526777, 526827), 'pulumi.set', 'pulumi.set', (['__self__', '"""disk_type_id"""', 'disk_type_id'], {}), "(__self__, 'disk_type_id', disk_type_id)\n", (526787, 526827), False, 'import pulumi\n'), ((526883, 526945), 'pulumi.set', 'pulumi.set', (['__self__', '"""resource_preset_id"""', 'resource_preset_id'], {}), "(__self__, 'resource_preset_id', resource_preset_id)\n", (526893, 526945), False, 'import pulumi\n'), ((529720, 529760), 'pulumi.set', 'pulumi.set', (['__self__', '"""edition"""', 'edition'], {}), "(__self__, 'edition', edition)\n", (529730, 529760), False, 'import pulumi\n'), ((529809, 529857), 'pulumi.set', 'pulumi.set', (['__self__', '"""master_node"""', 'master_node'], {}), "(__self__, 'master_node', master_node)\n", (529819, 529857), False, 'import pulumi\n'), ((529902, 529942), 'pulumi.set', 'pulumi.set', (['__self__', '"""plugins"""', 'plugins'], {}), "(__self__, 'plugins', plugins)\n", (529912, 529942), False, 'import pulumi\n'), ((529987, 530027), 'pulumi.set', 'pulumi.set', (['__self__', '"""version"""', 'version'], {}), "(__self__, 'version', version)\n", (529997, 530027), False, 'import pulumi\n'), ((539239, 539297), 'pulumi.set', 'pulumi.set', (['__self__', '"""assign_public_ip"""', 'assign_public_ip'], {}), "(__self__, 'assign_public_ip', assign_public_ip)\n", (539249, 539297), False, 'import pulumi\n'), ((539339, 539373), 'pulumi.set', 'pulumi.set', (['__self__', '"""fqdn"""', 'fqdn'], {}), "(__self__, 'fqdn', fqdn)\n", (539349, 539373), False, 'import pulumi\n'), ((539420, 539464), 'pulumi.set', 'pulumi.set', (['__self__', '"""subnet_id"""', 'subnet_id'], {}), "(__self__, 'subnet_id', subnet_id)\n", (539430, 539464), False, 'import pulumi\n'), ((542236, 542280), 'pulumi.set', 'pulumi.set', (['__self__', '"""data_lens"""', 'data_lens'], {}), "(__self__, 'data_lens', data_lens)\n", (542246, 542280), False, 'import pulumi\n'), ((542325, 542365), 'pulumi.set', 'pulumi.set', (['__self__', '"""web_sql"""', 'web_sql'], {}), "(__self__, 'web_sql', web_sql)\n", (542335, 542365), False, 'import pulumi\n'), ((543588, 543624), 'pulumi.set', 'pulumi.set', (['__self__', '"""hours"""', 'hours'], {}), "(__self__, 'hours', hours)\n", (543598, 543624), False, 'import pulumi\n'), ((543669, 543709), 'pulumi.set', 'pulumi.set', (['__self__', '"""minutes"""', 'minutes'], {}), "(__self__, 'minutes', minutes)\n", (543679, 543709), False, 'import pulumi\n'), ((544969, 545027), 'pulumi.set', 'pulumi.set', (['__self__', '"""assign_public_ip"""', 'assign_public_ip'], {}), "(__self__, 'assign_public_ip', assign_public_ip)\n", (544979, 545027), False, 'import pulumi\n'), ((545069, 545103), 'pulumi.set', 'pulumi.set', (['__self__', '"""fqdn"""', 'fqdn'], {}), "(__self__, 'fqdn', fqdn)\n", (545079, 545103), False, 'import pulumi\n'), ((548505, 548539), 'pulumi.set', 'pulumi.set', (['__self__', '"""fqdn"""', 'fqdn'], {}), "(__self__, 'fqdn', fqdn)\n", (548515, 548539), False, 'import pulumi\n'), ((552957, 553015), 'pulumi.set', 'pulumi.set', (['__self__', '"""assign_public_ip"""', 'assign_public_ip'], {}), "(__self__, 'assign_public_ip', assign_public_ip)\n", (552967, 553015), False, 'import pulumi\n'), ((553066, 553118), 'pulumi.set', 'pulumi.set', (['__self__', '"""brokers_count"""', 'brokers_count'], {}), "(__self__, 'brokers_count', brokers_count)\n", (553076, 553118), False, 'import pulumi\n'), ((553171, 553227), 'pulumi.set', 'pulumi.set', (['__self__', '"""schema_registry"""', 'schema_registry'], {}), "(__self__, 'schema_registry', schema_registry)\n", (553181, 553227), False, 'import pulumi\n'), ((553281, 553339), 'pulumi.set', 'pulumi.set', (['__self__', '"""unmanaged_topics"""', 'unmanaged_topics'], {}), "(__self__, 'unmanaged_topics', unmanaged_topics)\n", (553291, 553339), False, 'import pulumi\n'), ((553386, 553430), 'pulumi.set', 'pulumi.set', (['__self__', '"""zookeeper"""', 'zookeeper'], {}), "(__self__, 'zookeeper', zookeeper)\n", (553396, 553430), False, 'import pulumi\n'), ((557338, 557388), 'pulumi.set', 'pulumi.set', (['__self__', '"""kafka_config"""', 'kafka_config'], {}), "(__self__, 'kafka_config', kafka_config)\n", (557348, 557388), False, 'import pulumi\n'), ((559633, 559709), 'pulumi.set', 'pulumi.set', (['__self__', '"""auto_create_topics_enable"""', 'auto_create_topics_enable'], {}), "(__self__, 'auto_create_topics_enable', auto_create_topics_enable)\n", (559643, 559709), False, 'import pulumi\n'), ((559763, 559821), 'pulumi.set', 'pulumi.set', (['__self__', '"""compression_type"""', 'compression_type'], {}), "(__self__, 'compression_type', compression_type)\n", (559773, 559821), False, 'import pulumi\n'), ((559885, 559963), 'pulumi.set', 'pulumi.set', (['__self__', '"""default_replication_factor"""', 'default_replication_factor'], {}), "(__self__, 'default_replication_factor', default_replication_factor)\n", (559895, 559963), False, 'import pulumi\n'), ((560028, 560113), 'pulumi.set', 'pulumi.set', (['__self__', '"""log_flush_interval_messages"""', 'log_flush_interval_messages'], {}), "(__self__, 'log_flush_interval_messages', log_flush_interval_messages\n )\n", (560038, 560113), False, 'import pulumi\n'), ((560167, 560235), 'pulumi.set', 'pulumi.set', (['__self__', '"""log_flush_interval_ms"""', 'log_flush_interval_ms'], {}), "(__self__, 'log_flush_interval_ms', log_flush_interval_ms)\n", (560177, 560235), False, 'import pulumi\n'), ((560304, 560396), 'pulumi.set', 'pulumi.set', (['__self__', '"""log_flush_scheduler_interval_ms"""', 'log_flush_scheduler_interval_ms'], {}), "(__self__, 'log_flush_scheduler_interval_ms',\n log_flush_scheduler_interval_ms)\n", (560314, 560396), False, 'import pulumi\n'), ((560445, 560501), 'pulumi.set', 'pulumi.set', (['__self__', '"""log_preallocate"""', 'log_preallocate'], {}), "(__self__, 'log_preallocate', log_preallocate)\n", (560455, 560501), False, 'import pulumi\n'), ((560558, 560622), 'pulumi.set', 'pulumi.set', (['__self__', '"""log_retention_bytes"""', 'log_retention_bytes'], {}), "(__self__, 'log_retention_bytes', log_retention_bytes)\n", (560568, 560622), False, 'import pulumi\n'), ((560679, 560743), 'pulumi.set', 'pulumi.set', (['__self__', '"""log_retention_hours"""', 'log_retention_hours'], {}), "(__self__, 'log_retention_hours', log_retention_hours)\n", (560689, 560743), False, 'import pulumi\n'), ((560802, 560870), 'pulumi.set', 'pulumi.set', (['__self__', '"""log_retention_minutes"""', 'log_retention_minutes'], {}), "(__self__, 'log_retention_minutes', log_retention_minutes)\n", (560812, 560870), False, 'import pulumi\n'), ((560924, 560982), 'pulumi.set', 'pulumi.set', (['__self__', '"""log_retention_ms"""', 'log_retention_ms'], {}), "(__self__, 'log_retention_ms', log_retention_ms)\n", (560934, 560982), False, 'import pulumi\n'), ((561037, 561097), 'pulumi.set', 'pulumi.set', (['__self__', '"""log_segment_bytes"""', 'log_segment_bytes'], {}), "(__self__, 'log_segment_bytes', log_segment_bytes)\n", (561047, 561097), False, 'import pulumi\n'), ((561149, 561203), 'pulumi.set', 'pulumi.set', (['__self__', '"""num_partitions"""', 'num_partitions'], {}), "(__self__, 'num_partitions', num_partitions)\n", (561159, 561203), False, 'import pulumi\n'), ((561268, 561353), 'pulumi.set', 'pulumi.set', (['__self__', '"""socket_receive_buffer_bytes"""', 'socket_receive_buffer_bytes'], {}), "(__self__, 'socket_receive_buffer_bytes', socket_receive_buffer_bytes\n )\n", (561278, 561353), False, 'import pulumi\n'), ((561410, 561484), 'pulumi.set', 'pulumi.set', (['__self__', '"""socket_send_buffer_bytes"""', 'socket_send_buffer_bytes'], {}), "(__self__, 'socket_send_buffer_bytes', socket_send_buffer_bytes)\n", (561420, 561484), False, 'import pulumi\n'), ((569170, 569214), 'pulumi.set', 'pulumi.set', (['__self__', '"""resources"""', 'resources'], {}), "(__self__, 'resources', resources)\n", (569180, 569214), False, 'import pulumi\n'), ((570395, 570439), 'pulumi.set', 'pulumi.set', (['__self__', '"""disk_size"""', 'disk_size'], {}), "(__self__, 'disk_size', disk_size)\n", (570405, 570439), False, 'import pulumi\n'), ((570489, 570539), 'pulumi.set', 'pulumi.set', (['__self__', '"""disk_type_id"""', 'disk_type_id'], {}), "(__self__, 'disk_type_id', disk_type_id)\n", (570499, 570539), False, 'import pulumi\n'), ((570595, 570657), 'pulumi.set', 'pulumi.set', (['__self__', '"""resource_preset_id"""', 'resource_preset_id'], {}), "(__self__, 'resource_preset_id', resource_preset_id)\n", (570605, 570657), False, 'import pulumi\n'), ((572939, 572997), 'pulumi.set', 'pulumi.set', (['__self__', '"""assign_public_ip"""', 'assign_public_ip'], {}), "(__self__, 'assign_public_ip', assign_public_ip)\n", (572949, 572997), False, 'import pulumi\n'), ((573041, 573079), 'pulumi.set', 'pulumi.set', (['__self__', '"""health"""', 'health'], {}), "(__self__, 'health', health)\n", (573051, 573079), False, 'import pulumi\n'), ((573121, 573155), 'pulumi.set', 'pulumi.set', (['__self__', '"""name"""', 'name'], {}), "(__self__, 'name', name)\n", (573131, 573155), False, 'import pulumi\n'), ((573197, 573231), 'pulumi.set', 'pulumi.set', (['__self__', '"""role"""', 'role'], {}), "(__self__, 'role', role)\n", (573207, 573231), False, 'import pulumi\n'), ((573278, 573322), 'pulumi.set', 'pulumi.set', (['__self__', '"""subnet_id"""', 'subnet_id'], {}), "(__self__, 'subnet_id', subnet_id)\n", (573288, 573322), False, 'import pulumi\n'), ((573367, 573407), 'pulumi.set', 'pulumi.set', (['__self__', '"""zone_id"""', 'zone_id'], {}), "(__self__, 'zone_id', zone_id)\n", (573377, 573407), False, 'import pulumi\n'), ((576466, 576516), 'pulumi.set', 'pulumi.set', (['__self__', '"""topic_config"""', 'topic_config'], {}), "(__self__, 'topic_config', topic_config)\n", (576476, 576516), False, 'import pulumi\n'), ((579052, 579106), 'pulumi.set', 'pulumi.set', (['__self__', '"""cleanup_policy"""', 'cleanup_policy'], {}), "(__self__, 'cleanup_policy', cleanup_policy)\n", (579062, 579106), False, 'import pulumi\n'), ((579160, 579218), 'pulumi.set', 'pulumi.set', (['__self__', '"""compression_type"""', 'compression_type'], {}), "(__self__, 'compression_type', compression_type)\n", (579170, 579218), False, 'import pulumi\n'), ((579275, 579339), 'pulumi.set', 'pulumi.set', (['__self__', '"""delete_retention_ms"""', 'delete_retention_ms'], {}), "(__self__, 'delete_retention_ms', delete_retention_ms)\n", (579285, 579339), False, 'import pulumi\n'), ((579397, 579463), 'pulumi.set', 'pulumi.set', (['__self__', '"""file_delete_delay_ms"""', 'file_delete_delay_ms'], {}), "(__self__, 'file_delete_delay_ms', file_delete_delay_ms)\n", (579407, 579463), False, 'import pulumi\n'), ((579515, 579569), 'pulumi.set', 'pulumi.set', (['__self__', '"""flush_messages"""', 'flush_messages'], {}), "(__self__, 'flush_messages', flush_messages)\n", (579525, 579569), False, 'import pulumi\n'), ((579615, 579657), 'pulumi.set', 'pulumi.set', (['__self__', '"""flush_ms"""', 'flush_ms'], {}), "(__self__, 'flush_ms', flush_ms)\n", (579625, 579657), False, 'import pulumi\n'), ((579712, 579772), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_message_bytes"""', 'max_message_bytes'], {}), "(__self__, 'max_message_bytes', max_message_bytes)\n", (579722, 579772), False, 'import pulumi\n'), ((579831, 579899), 'pulumi.set', 'pulumi.set', (['__self__', '"""min_compaction_lag_ms"""', 'min_compaction_lag_ms'], {}), "(__self__, 'min_compaction_lag_ms', min_compaction_lag_ms)\n", (579841, 579899), False, 'import pulumi\n'), ((579956, 580020), 'pulumi.set', 'pulumi.set', (['__self__', '"""min_insync_replicas"""', 'min_insync_replicas'], {}), "(__self__, 'min_insync_replicas', min_insync_replicas)\n", (579966, 580020), False, 'import pulumi\n'), ((580069, 580117), 'pulumi.set', 'pulumi.set', (['__self__', '"""preallocate"""', 'preallocate'], {}), "(__self__, 'preallocate', preallocate)\n", (580079, 580117), False, 'import pulumi\n'), ((580170, 580226), 'pulumi.set', 'pulumi.set', (['__self__', '"""retention_bytes"""', 'retention_bytes'], {}), "(__self__, 'retention_bytes', retention_bytes)\n", (580180, 580226), False, 'import pulumi\n'), ((580276, 580326), 'pulumi.set', 'pulumi.set', (['__self__', '"""retention_ms"""', 'retention_ms'], {}), "(__self__, 'retention_ms', retention_ms)\n", (580286, 580326), False, 'import pulumi\n'), ((580377, 580429), 'pulumi.set', 'pulumi.set', (['__self__', '"""segment_bytes"""', 'segment_bytes'], {}), "(__self__, 'segment_bytes', segment_bytes)\n", (580387, 580429), False, 'import pulumi\n'), ((585323, 585371), 'pulumi.set', 'pulumi.set', (['__self__', '"""permissions"""', 'permissions'], {}), "(__self__, 'permissions', permissions)\n", (585333, 585371), False, 'import pulumi\n'), ((588623, 588677), 'pulumi.set', 'pulumi.set', (['__self__', '"""cleanup_policy"""', 'cleanup_policy'], {}), "(__self__, 'cleanup_policy', cleanup_policy)\n", (588633, 588677), False, 'import pulumi\n'), ((588731, 588789), 'pulumi.set', 'pulumi.set', (['__self__', '"""compression_type"""', 'compression_type'], {}), "(__self__, 'compression_type', compression_type)\n", (588741, 588789), False, 'import pulumi\n'), ((588846, 588910), 'pulumi.set', 'pulumi.set', (['__self__', '"""delete_retention_ms"""', 'delete_retention_ms'], {}), "(__self__, 'delete_retention_ms', delete_retention_ms)\n", (588856, 588910), False, 'import pulumi\n'), ((588968, 589034), 'pulumi.set', 'pulumi.set', (['__self__', '"""file_delete_delay_ms"""', 'file_delete_delay_ms'], {}), "(__self__, 'file_delete_delay_ms', file_delete_delay_ms)\n", (588978, 589034), False, 'import pulumi\n'), ((589086, 589140), 'pulumi.set', 'pulumi.set', (['__self__', '"""flush_messages"""', 'flush_messages'], {}), "(__self__, 'flush_messages', flush_messages)\n", (589096, 589140), False, 'import pulumi\n'), ((589186, 589228), 'pulumi.set', 'pulumi.set', (['__self__', '"""flush_ms"""', 'flush_ms'], {}), "(__self__, 'flush_ms', flush_ms)\n", (589196, 589228), False, 'import pulumi\n'), ((589283, 589343), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_message_bytes"""', 'max_message_bytes'], {}), "(__self__, 'max_message_bytes', max_message_bytes)\n", (589293, 589343), False, 'import pulumi\n'), ((589402, 589470), 'pulumi.set', 'pulumi.set', (['__self__', '"""min_compaction_lag_ms"""', 'min_compaction_lag_ms'], {}), "(__self__, 'min_compaction_lag_ms', min_compaction_lag_ms)\n", (589412, 589470), False, 'import pulumi\n'), ((589527, 589591), 'pulumi.set', 'pulumi.set', (['__self__', '"""min_insync_replicas"""', 'min_insync_replicas'], {}), "(__self__, 'min_insync_replicas', min_insync_replicas)\n", (589537, 589591), False, 'import pulumi\n'), ((589640, 589688), 'pulumi.set', 'pulumi.set', (['__self__', '"""preallocate"""', 'preallocate'], {}), "(__self__, 'preallocate', preallocate)\n", (589650, 589688), False, 'import pulumi\n'), ((589741, 589797), 'pulumi.set', 'pulumi.set', (['__self__', '"""retention_bytes"""', 'retention_bytes'], {}), "(__self__, 'retention_bytes', retention_bytes)\n", (589751, 589797), False, 'import pulumi\n'), ((589847, 589897), 'pulumi.set', 'pulumi.set', (['__self__', '"""retention_ms"""', 'retention_ms'], {}), "(__self__, 'retention_ms', retention_ms)\n", (589857, 589897), False, 'import pulumi\n'), ((589948, 590000), 'pulumi.set', 'pulumi.set', (['__self__', '"""segment_bytes"""', 'segment_bytes'], {}), "(__self__, 'segment_bytes', segment_bytes)\n", (589958, 590000), False, 'import pulumi\n'), ((595316, 595354), 'pulumi.set', 'pulumi.set', (['__self__', '"""access"""', 'access'], {}), "(__self__, 'access', access)\n", (595326, 595354), False, 'import pulumi\n'), ((595411, 595475), 'pulumi.set', 'pulumi.set', (['__self__', '"""backup_window_start"""', 'backup_window_start'], {}), "(__self__, 'backup_window_start', backup_window_start)\n", (595421, 595475), False, 'import pulumi\n'), ((595542, 595630), 'pulumi.set', 'pulumi.set', (['__self__', '"""feature_compatibility_version"""', 'feature_compatibility_version'], {}), "(__self__, 'feature_compatibility_version',\n feature_compatibility_version)\n", (595552, 595630), False, 'import pulumi\n'), ((597804, 597848), 'pulumi.set', 'pulumi.set', (['__self__', '"""data_lens"""', 'data_lens'], {}), "(__self__, 'data_lens', data_lens)\n", (597814, 597848), False, 'import pulumi\n'), ((598658, 598694), 'pulumi.set', 'pulumi.set', (['__self__', '"""hours"""', 'hours'], {}), "(__self__, 'hours', hours)\n", (598668, 598694), False, 'import pulumi\n'), ((598739, 598779), 'pulumi.set', 'pulumi.set', (['__self__', '"""minutes"""', 'minutes'], {}), "(__self__, 'minutes', minutes)\n", (598749, 598779), False, 'import pulumi\n'), ((601880, 601938), 'pulumi.set', 'pulumi.set', (['__self__', '"""assign_public_ip"""', 'assign_public_ip'], {}), "(__self__, 'assign_public_ip', assign_public_ip)\n", (601890, 601938), False, 'import pulumi\n'), ((601982, 602020), 'pulumi.set', 'pulumi.set', (['__self__', '"""health"""', 'health'], {}), "(__self__, 'health', health)\n", (601992, 602020), False, 'import pulumi\n'), ((602062, 602096), 'pulumi.set', 'pulumi.set', (['__self__', '"""name"""', 'name'], {}), "(__self__, 'name', name)\n", (602072, 602096), False, 'import pulumi\n'), ((602138, 602172), 'pulumi.set', 'pulumi.set', (['__self__', '"""role"""', 'role'], {}), "(__self__, 'role', role)\n", (602148, 602172), False, 'import pulumi\n'), ((602220, 602266), 'pulumi.set', 'pulumi.set', (['__self__', '"""shard_name"""', 'shard_name'], {}), "(__self__, 'shard_name', shard_name)\n", (602230, 602266), False, 'import pulumi\n'), ((602308, 602342), 'pulumi.set', 'pulumi.set', (['__self__', '"""type"""', 'type'], {}), "(__self__, 'type', type)\n", (602318, 602342), False, 'import pulumi\n'), ((606247, 606279), 'pulumi.set', 'pulumi.set', (['__self__', '"""day"""', 'day'], {}), "(__self__, 'day', day)\n", (606257, 606279), False, 'import pulumi\n'), ((606321, 606355), 'pulumi.set', 'pulumi.set', (['__self__', '"""hour"""', 'hour'], {}), "(__self__, 'hour', hour)\n", (606331, 606355), False, 'import pulumi\n'), ((610224, 610272), 'pulumi.set', 'pulumi.set', (['__self__', '"""permissions"""', 'permissions'], {}), "(__self__, 'permissions', permissions)\n", (610234, 610272), False, 'import pulumi\n'), ((611856, 611892), 'pulumi.set', 'pulumi.set', (['__self__', '"""roles"""', 'roles'], {}), "(__self__, 'roles', roles)\n", (611866, 611892), False, 'import pulumi\n'), ((613145, 613189), 'pulumi.set', 'pulumi.set', (['__self__', '"""data_lens"""', 'data_lens'], {}), "(__self__, 'data_lens', data_lens)\n", (613155, 613189), False, 'import pulumi\n'), ((613234, 613274), 'pulumi.set', 'pulumi.set', (['__self__', '"""web_sql"""', 'web_sql'], {}), "(__self__, 'web_sql', web_sql)\n", (613244, 613274), False, 'import pulumi\n'), ((614554, 614590), 'pulumi.set', 'pulumi.set', (['__self__', '"""hours"""', 'hours'], {}), "(__self__, 'hours', hours)\n", (614564, 614590), False, 'import pulumi\n'), ((614635, 614675), 'pulumi.set', 'pulumi.set', (['__self__', '"""minutes"""', 'minutes'], {}), "(__self__, 'minutes', minutes)\n", (614645, 614675), False, 'import pulumi\n'), ((618024, 618082), 'pulumi.set', 'pulumi.set', (['__self__', '"""assign_public_ip"""', 'assign_public_ip'], {}), "(__self__, 'assign_public_ip', assign_public_ip)\n", (618034, 618082), False, 'import pulumi\n'), ((618124, 618158), 'pulumi.set', 'pulumi.set', (['__self__', '"""fqdn"""', 'fqdn'], {}), "(__self__, 'fqdn', fqdn)\n", (618134, 618158), False, 'import pulumi\n'), ((618200, 618234), 'pulumi.set', 'pulumi.set', (['__self__', '"""name"""', 'name'], {}), "(__self__, 'name', name)\n", (618210, 618234), False, 'import pulumi\n'), ((618290, 618352), 'pulumi.set', 'pulumi.set', (['__self__', '"""replication_source"""', 'replication_source'], {}), "(__self__, 'replication_source', replication_source)\n", (618300, 618352), False, 'import pulumi\n'), ((618413, 618485), 'pulumi.set', 'pulumi.set', (['__self__', '"""replication_source_name"""', 'replication_source_name'], {}), "(__self__, 'replication_source_name', replication_source_name)\n", (618423, 618485), False, 'import pulumi\n'), ((618532, 618576), 'pulumi.set', 'pulumi.set', (['__self__', '"""subnet_id"""', 'subnet_id'], {}), "(__self__, 'subnet_id', subnet_id)\n", (618542, 618576), False, 'import pulumi\n'), ((622390, 622422), 'pulumi.set', 'pulumi.set', (['__self__', '"""day"""', 'day'], {}), "(__self__, 'day', day)\n", (622400, 622422), False, 'import pulumi\n'), ((622464, 622498), 'pulumi.set', 'pulumi.set', (['__self__', '"""hour"""', 'hour'], {}), "(__self__, 'hour', hour)\n", (622474, 622498), False, 'import pulumi\n'), ((625962, 625996), 'pulumi.set', 'pulumi.set', (['__self__', '"""time"""', 'time'], {}), "(__self__, 'time', time)\n", (625972, 625996), False, 'import pulumi\n'), ((628873, 628941), 'pulumi.set', 'pulumi.set', (['__self__', '"""authentication_plugin"""', 'authentication_plugin'], {}), "(__self__, 'authentication_plugin', authentication_plugin)\n", (628883, 628941), False, 'import pulumi\n'), ((628996, 629056), 'pulumi.set', 'pulumi.set', (['__self__', '"""connection_limits"""', 'connection_limits'], {}), "(__self__, 'connection_limits', connection_limits)\n", (629006, 629056), False, 'import pulumi\n'), ((629112, 629174), 'pulumi.set', 'pulumi.set', (['__self__', '"""global_permissions"""', 'global_permissions'], {}), "(__self__, 'global_permissions', global_permissions)\n", (629122, 629174), False, 'import pulumi\n'), ((629223, 629271), 'pulumi.set', 'pulumi.set', (['__self__', '"""permissions"""', 'permissions'], {}), "(__self__, 'permissions', permissions)\n", (629233, 629271), False, 'import pulumi\n'), ((633093, 633167), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_connections_per_hour"""', 'max_connections_per_hour'], {}), "(__self__, 'max_connections_per_hour', max_connections_per_hour)\n", (633103, 633167), False, 'import pulumi\n'), ((633227, 633297), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_questions_per_hour"""', 'max_questions_per_hour'], {}), "(__self__, 'max_questions_per_hour', max_questions_per_hour)\n", (633237, 633297), False, 'import pulumi\n'), ((633355, 633421), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_updates_per_hour"""', 'max_updates_per_hour'], {}), "(__self__, 'max_updates_per_hour', max_updates_per_hour)\n", (633365, 633421), False, 'import pulumi\n'), ((633479, 633545), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_user_connections"""', 'max_user_connections'], {}), "(__self__, 'max_user_connections', max_user_connections)\n", (633489, 633545), False, 'import pulumi\n'), ((635998, 636034), 'pulumi.set', 'pulumi.set', (['__self__', '"""roles"""', 'roles'], {}), "(__self__, 'roles', roles)\n", (636008, 636034), False, 'import pulumi\n'), ((638733, 638777), 'pulumi.set', 'pulumi.set', (['__self__', '"""databases"""', 'databases'], {}), "(__self__, 'databases', databases)\n", (638743, 638777), False, 'import pulumi\n'), ((638831, 638889), 'pulumi.set', 'pulumi.set', (['__self__', '"""maxmemory_policy"""', 'maxmemory_policy'], {}), "(__self__, 'maxmemory_policy', maxmemory_policy)\n", (638841, 638889), False, 'import pulumi\n'), ((638949, 639019), 'pulumi.set', 'pulumi.set', (['__self__', '"""notify_keyspace_events"""', 'notify_keyspace_events'], {}), "(__self__, 'notify_keyspace_events', notify_keyspace_events)\n", (638959, 639019), False, 'import pulumi\n'), ((639080, 639152), 'pulumi.set', 'pulumi.set', (['__self__', '"""slowlog_log_slower_than"""', 'slowlog_log_slower_than'], {}), "(__self__, 'slowlog_log_slower_than', slowlog_log_slower_than)\n", (639090, 639152), False, 'import pulumi\n'), ((639205, 639261), 'pulumi.set', 'pulumi.set', (['__self__', '"""slowlog_max_len"""', 'slowlog_max_len'], {}), "(__self__, 'slowlog_max_len', slowlog_max_len)\n", (639215, 639261), False, 'import pulumi\n'), ((639306, 639346), 'pulumi.set', 'pulumi.set', (['__self__', '"""timeout"""', 'timeout'], {}), "(__self__, 'timeout', timeout)\n", (639316, 639346), False, 'import pulumi\n'), ((643523, 643557), 'pulumi.set', 'pulumi.set', (['__self__', '"""fqdn"""', 'fqdn'], {}), "(__self__, 'fqdn', fqdn)\n", (643533, 643557), False, 'import pulumi\n'), ((643605, 643651), 'pulumi.set', 'pulumi.set', (['__self__', '"""shard_name"""', 'shard_name'], {}), "(__self__, 'shard_name', shard_name)\n", (643615, 643651), False, 'import pulumi\n'), ((643698, 643742), 'pulumi.set', 'pulumi.set', (['__self__', '"""subnet_id"""', 'subnet_id'], {}), "(__self__, 'subnet_id', subnet_id)\n", (643708, 643742), False, 'import pulumi\n'), ((646122, 646154), 'pulumi.set', 'pulumi.set', (['__self__', '"""day"""', 'day'], {}), "(__self__, 'day', day)\n", (646132, 646154), False, 'import pulumi\n'), ((646196, 646230), 'pulumi.set', 'pulumi.set', (['__self__', '"""hour"""', 'hour'], {}), "(__self__, 'hour', hour)\n", (646206, 646230), False, 'import pulumi\n'), ((648042, 648092), 'pulumi.set', 'pulumi.set', (['__self__', '"""disk_type_id"""', 'disk_type_id'], {}), "(__self__, 'disk_type_id', disk_type_id)\n", (648052, 648092), False, 'import pulumi\n'), ((649620, 649656), 'pulumi.set', 'pulumi.set', (['__self__', '"""hours"""', 'hours'], {}), "(__self__, 'hours', hours)\n", (649630, 649656), False, 'import pulumi\n'), ((649701, 649741), 'pulumi.set', 'pulumi.set', (['__self__', '"""minutes"""', 'minutes'], {}), "(__self__, 'minutes', minutes)\n", (649711, 649741), False, 'import pulumi\n'), ((651921, 651979), 'pulumi.set', 'pulumi.set', (['__self__', '"""assign_public_ip"""', 'assign_public_ip'], {}), "(__self__, 'assign_public_ip', assign_public_ip)\n", (651931, 651979), False, 'import pulumi\n'), ((652021, 652055), 'pulumi.set', 'pulumi.set', (['__self__', '"""fqdn"""', 'fqdn'], {}), "(__self__, 'fqdn', fqdn)\n", (652031, 652055), False, 'import pulumi\n'), ((652102, 652146), 'pulumi.set', 'pulumi.set', (['__self__', '"""subnet_id"""', 'subnet_id'], {}), "(__self__, 'subnet_id', subnet_id)\n", (652112, 652146), False, 'import pulumi\n'), ((656151, 656199), 'pulumi.set', 'pulumi.set', (['__self__', '"""permissions"""', 'permissions'], {}), "(__self__, 'permissions', permissions)\n", (656161, 656199), False, 'import pulumi\n'), ((658009, 658045), 'pulumi.set', 'pulumi.set', (['__self__', '"""roles"""', 'roles'], {}), "(__self__, 'roles', roles)\n", (658019, 658045), False, 'import pulumi\n'), ((660278, 660334), 'pulumi.set', 'pulumi.set', (['__self__', '"""allowed_headers"""', 'allowed_headers'], {}), "(__self__, 'allowed_headers', allowed_headers)\n", (660288, 660334), False, 'import pulumi\n'), ((660386, 660440), 'pulumi.set', 'pulumi.set', (['__self__', '"""expose_headers"""', 'expose_headers'], {}), "(__self__, 'expose_headers', expose_headers)\n", (660396, 660440), False, 'import pulumi\n'), ((660493, 660549), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_age_seconds"""', 'max_age_seconds'], {}), "(__self__, 'max_age_seconds', max_age_seconds)\n", (660503, 660549), False, 'import pulumi\n'), ((663328, 663358), 'pulumi.set', 'pulumi.set', (['__self__', '"""id"""', 'id'], {}), "(__self__, 'id', id)\n", (663338, 663358), False, 'import pulumi\n'), ((663399, 663431), 'pulumi.set', 'pulumi.set', (['__self__', '"""uri"""', 'uri'], {}), "(__self__, 'uri', uri)\n", (663409, 663431), False, 'import pulumi\n'), ((666801, 666907), 'pulumi.set', 'pulumi.set', (['__self__', '"""abort_incomplete_multipart_upload_days"""', 'abort_incomplete_multipart_upload_days'], {}), "(__self__, 'abort_incomplete_multipart_upload_days',\n abort_incomplete_multipart_upload_days)\n", (666811, 666907), False, 'import pulumi\n'), ((666951, 666997), 'pulumi.set', 'pulumi.set', (['__self__', '"""expiration"""', 'expiration'], {}), "(__self__, 'expiration', expiration)\n", (666961, 666997), False, 'import pulumi\n'), ((667037, 667067), 'pulumi.set', 'pulumi.set', (['__self__', '"""id"""', 'id'], {}), "(__self__, 'id', id)\n", (667047, 667067), False, 'import pulumi\n'), ((667134, 667222), 'pulumi.set', 'pulumi.set', (['__self__', '"""noncurrent_version_expiration"""', 'noncurrent_version_expiration'], {}), "(__self__, 'noncurrent_version_expiration',\n noncurrent_version_expiration)\n", (667144, 667222), False, 'import pulumi\n'), ((667286, 667376), 'pulumi.set', 'pulumi.set', (['__self__', '"""noncurrent_version_transitions"""', 'noncurrent_version_transitions'], {}), "(__self__, 'noncurrent_version_transitions',\n noncurrent_version_transitions)\n", (667296, 667376), False, 'import pulumi\n'), ((667416, 667454), 'pulumi.set', 'pulumi.set', (['__self__', '"""prefix"""', 'prefix'], {}), "(__self__, 'prefix', prefix)\n", (667426, 667454), False, 'import pulumi\n'), ((667503, 667551), 'pulumi.set', 'pulumi.set', (['__self__', '"""transitions"""', 'transitions'], {}), "(__self__, 'transitions', transitions)\n", (667513, 667551), False, 'import pulumi\n'), ((672261, 672295), 'pulumi.set', 'pulumi.set', (['__self__', '"""date"""', 'date'], {}), "(__self__, 'date', date)\n", (672271, 672295), False, 'import pulumi\n'), ((672337, 672371), 'pulumi.set', 'pulumi.set', (['__self__', '"""days"""', 'days'], {}), "(__self__, 'days', days)\n", (672347, 672371), False, 'import pulumi\n'), ((672437, 672523), 'pulumi.set', 'pulumi.set', (['__self__', '"""expired_object_delete_marker"""', 'expired_object_delete_marker'], {}), "(__self__, 'expired_object_delete_marker',\n expired_object_delete_marker)\n", (672447, 672523), False, 'import pulumi\n'), ((674206, 674240), 'pulumi.set', 'pulumi.set', (['__self__', '"""days"""', 'days'], {}), "(__self__, 'days', days)\n", (674216, 674240), False, 'import pulumi\n'), ((675213, 675247), 'pulumi.set', 'pulumi.set', (['__self__', '"""days"""', 'days'], {}), "(__self__, 'days', days)\n", (675223, 675247), False, 'import pulumi\n'), ((676839, 676873), 'pulumi.set', 'pulumi.set', (['__self__', '"""date"""', 'date'], {}), "(__self__, 'date', date)\n", (676849, 676873), False, 'import pulumi\n'), ((676915, 676949), 'pulumi.set', 'pulumi.set', (['__self__', '"""days"""', 'days'], {}), "(__self__, 'days', days)\n", (676925, 676949), False, 'import pulumi\n'), ((678627, 678679), 'pulumi.set', 'pulumi.set', (['__self__', '"""target_prefix"""', 'target_prefix'], {}), "(__self__, 'target_prefix', target_prefix)\n", (678637, 678679), False, 'import pulumi\n'), ((683553, 683593), 'pulumi.set', 'pulumi.set', (['__self__', '"""enabled"""', 'enabled'], {}), "(__self__, 'enabled', enabled)\n", (683563, 683593), False, 'import pulumi\n'), ((685276, 685330), 'pulumi.set', 'pulumi.set', (['__self__', '"""error_document"""', 'error_document'], {}), "(__self__, 'error_document', error_document)\n", (685286, 685330), False, 'import pulumi\n'), ((685382, 685436), 'pulumi.set', 'pulumi.set', (['__self__', '"""index_document"""', 'index_document'], {}), "(__self__, 'index_document', index_document)\n", (685392, 685436), False, 'import pulumi\n'), ((685498, 685572), 'pulumi.set', 'pulumi.set', (['__self__', '"""redirect_all_requests_to"""', 'redirect_all_requests_to'], {}), "(__self__, 'redirect_all_requests_to', redirect_all_requests_to)\n", (685508, 685572), False, 'import pulumi\n'), ((685623, 685675), 'pulumi.set', 'pulumi.set', (['__self__', '"""routing_rules"""', 'routing_rules'], {}), "(__self__, 'routing_rules', routing_rules)\n", (685633, 685675), False, 'import pulumi\n'), ((688471, 688511), 'pulumi.set', 'pulumi.set', (['__self__', '"""address"""', 'address'], {}), "(__self__, 'address', address)\n", (688481, 688511), False, 'import pulumi\n'), ((688573, 688647), 'pulumi.set', 'pulumi.set', (['__self__', '"""ddos_protection_provider"""', 'ddos_protection_provider'], {}), "(__self__, 'ddos_protection_provider', ddos_protection_provider)\n", (688583, 688647), False, 'import pulumi\n'), ((688709, 688783), 'pulumi.set', 'pulumi.set', (['__self__', '"""outgoing_smtp_capability"""', 'outgoing_smtp_capability'], {}), "(__self__, 'outgoing_smtp_capability', outgoing_smtp_capability)\n", (688719, 688783), False, 'import pulumi\n'), ((688828, 688868), 'pulumi.set', 'pulumi.set', (['__self__', '"""zone_id"""', 'zone_id'], {}), "(__self__, 'zone_id', zone_id)\n", (688838, 688868), False, 'import pulumi\n'), ((691635, 691683), 'pulumi.set', 'pulumi.set', (['__self__', '"""description"""', 'description'], {}), "(__self__, 'description', description)\n", (691645, 691683), False, 'import pulumi\n'), ((691730, 691774), 'pulumi.set', 'pulumi.set', (['__self__', '"""from_port"""', 'from_port'], {}), "(__self__, 'from_port', from_port)\n", (691740, 691774), False, 'import pulumi\n'), ((691814, 691844), 'pulumi.set', 'pulumi.set', (['__self__', '"""id"""', 'id'], {}), "(__self__, 'id', id)\n", (691824, 691844), False, 'import pulumi\n'), ((691888, 691926), 'pulumi.set', 'pulumi.set', (['__self__', '"""labels"""', 'labels'], {}), "(__self__, 'labels', labels)\n", (691898, 691926), False, 'import pulumi\n'), ((691968, 692002), 'pulumi.set', 'pulumi.set', (['__self__', '"""port"""', 'port'], {}), "(__self__, 'port', port)\n", (691978, 692002), False, 'import pulumi\n'), ((692057, 692117), 'pulumi.set', 'pulumi.set', (['__self__', '"""predefined_target"""', 'predefined_target'], {}), "(__self__, 'predefined_target', predefined_target)\n", (692067, 692117), False, 'import pulumi\n'), ((692172, 692232), 'pulumi.set', 'pulumi.set', (['__self__', '"""security_group_id"""', 'security_group_id'], {}), "(__self__, 'security_group_id', security_group_id)\n", (692182, 692232), False, 'import pulumi\n'), ((692277, 692317), 'pulumi.set', 'pulumi.set', (['__self__', '"""to_port"""', 'to_port'], {}), "(__self__, 'to_port', to_port)\n", (692287, 692317), False, 'import pulumi\n'), ((692369, 692423), 'pulumi.set', 'pulumi.set', (['__self__', '"""v4_cidr_blocks"""', 'v4_cidr_blocks'], {}), "(__self__, 'v4_cidr_blocks', v4_cidr_blocks)\n", (692379, 692423), False, 'import pulumi\n'), ((692475, 692529), 'pulumi.set', 'pulumi.set', (['__self__', '"""v6_cidr_blocks"""', 'v6_cidr_blocks'], {}), "(__self__, 'v6_cidr_blocks', v6_cidr_blocks)\n", (692485, 692529), False, 'import pulumi\n'), ((697181, 697229), 'pulumi.set', 'pulumi.set', (['__self__', '"""description"""', 'description'], {}), "(__self__, 'description', description)\n", (697191, 697229), False, 'import pulumi\n'), ((697276, 697320), 'pulumi.set', 'pulumi.set', (['__self__', '"""from_port"""', 'from_port'], {}), "(__self__, 'from_port', from_port)\n", (697286, 697320), False, 'import pulumi\n'), ((697360, 697390), 'pulumi.set', 'pulumi.set', (['__self__', '"""id"""', 'id'], {}), "(__self__, 'id', id)\n", (697370, 697390), False, 'import pulumi\n'), ((697434, 697472), 'pulumi.set', 'pulumi.set', (['__self__', '"""labels"""', 'labels'], {}), "(__self__, 'labels', labels)\n", (697444, 697472), False, 'import pulumi\n'), ((697514, 697548), 'pulumi.set', 'pulumi.set', (['__self__', '"""port"""', 'port'], {}), "(__self__, 'port', port)\n", (697524, 697548), False, 'import pulumi\n'), ((697603, 697663), 'pulumi.set', 'pulumi.set', (['__self__', '"""predefined_target"""', 'predefined_target'], {}), "(__self__, 'predefined_target', predefined_target)\n", (697613, 697663), False, 'import pulumi\n'), ((697718, 697778), 'pulumi.set', 'pulumi.set', (['__self__', '"""security_group_id"""', 'security_group_id'], {}), "(__self__, 'security_group_id', security_group_id)\n", (697728, 697778), False, 'import pulumi\n'), ((697823, 697863), 'pulumi.set', 'pulumi.set', (['__self__', '"""to_port"""', 'to_port'], {}), "(__self__, 'to_port', to_port)\n", (697833, 697863), False, 'import pulumi\n'), ((697915, 697969), 'pulumi.set', 'pulumi.set', (['__self__', '"""v4_cidr_blocks"""', 'v4_cidr_blocks'], {}), "(__self__, 'v4_cidr_blocks', v4_cidr_blocks)\n", (697925, 697969), False, 'import pulumi\n'), ((698021, 698075), 'pulumi.set', 'pulumi.set', (['__self__', '"""v6_cidr_blocks"""', 'v6_cidr_blocks'], {}), "(__self__, 'v6_cidr_blocks', v6_cidr_blocks)\n", (698031, 698075), False, 'import pulumi\n'), ((701954, 702016), 'pulumi.set', 'pulumi.set', (['__self__', '"""destination_prefix"""', 'destination_prefix'], {}), "(__self__, 'destination_prefix', destination_prefix)\n", (701964, 702016), False, 'import pulumi\n'), ((702070, 702128), 'pulumi.set', 'pulumi.set', (['__self__', '"""next_hop_address"""', 'next_hop_address'], {}), "(__self__, 'next_hop_address', next_hop_address)\n", (702080, 702128), False, 'import pulumi\n'), ((704132, 704180), 'pulumi.set', 'pulumi.set', (['__self__', '"""description"""', 'description'], {}), "(__self__, 'description', description)\n", (704142, 704180), False, 'import pulumi\n'), ((704227, 704271), 'pulumi.set', 'pulumi.set', (['__self__', '"""from_port"""', 'from_port'], {}), "(__self__, 'from_port', from_port)\n", (704237, 704271), False, 'import pulumi\n'), ((704311, 704341), 'pulumi.set', 'pulumi.set', (['__self__', '"""id"""', 'id'], {}), "(__self__, 'id', id)\n", (704321, 704341), False, 'import pulumi\n'), ((704385, 704423), 'pulumi.set', 'pulumi.set', (['__self__', '"""labels"""', 'labels'], {}), "(__self__, 'labels', labels)\n", (704395, 704423), False, 'import pulumi\n'), ((704465, 704499), 'pulumi.set', 'pulumi.set', (['__self__', '"""port"""', 'port'], {}), "(__self__, 'port', port)\n", (704475, 704499), False, 'import pulumi\n'), ((704554, 704614), 'pulumi.set', 'pulumi.set', (['__self__', '"""predefined_target"""', 'predefined_target'], {}), "(__self__, 'predefined_target', predefined_target)\n", (704564, 704614), False, 'import pulumi\n'), ((704669, 704729), 'pulumi.set', 'pulumi.set', (['__self__', '"""security_group_id"""', 'security_group_id'], {}), "(__self__, 'security_group_id', security_group_id)\n", (704679, 704729), False, 'import pulumi\n'), ((704774, 704814), 'pulumi.set', 'pulumi.set', (['__self__', '"""to_port"""', 'to_port'], {}), "(__self__, 'to_port', to_port)\n", (704784, 704814), False, 'import pulumi\n'), ((704866, 704920), 'pulumi.set', 'pulumi.set', (['__self__', '"""v4_cidr_blocks"""', 'v4_cidr_blocks'], {}), "(__self__, 'v4_cidr_blocks', v4_cidr_blocks)\n", (704876, 704920), False, 'import pulumi\n'), ((704972, 705026), 'pulumi.set', 'pulumi.set', (['__self__', '"""v6_cidr_blocks"""', 'v6_cidr_blocks'], {}), "(__self__, 'v6_cidr_blocks', v6_cidr_blocks)\n", (704982, 705026), False, 'import pulumi\n'), ((709651, 709699), 'pulumi.set', 'pulumi.set', (['__self__', '"""description"""', 'description'], {}), "(__self__, 'description', description)\n", (709661, 709699), False, 'import pulumi\n'), ((709746, 709790), 'pulumi.set', 'pulumi.set', (['__self__', '"""from_port"""', 'from_port'], {}), "(__self__, 'from_port', from_port)\n", (709756, 709790), False, 'import pulumi\n'), ((709830, 709860), 'pulumi.set', 'pulumi.set', (['__self__', '"""id"""', 'id'], {}), "(__self__, 'id', id)\n", (709840, 709860), False, 'import pulumi\n'), ((709904, 709942), 'pulumi.set', 'pulumi.set', (['__self__', '"""labels"""', 'labels'], {}), "(__self__, 'labels', labels)\n", (709914, 709942), False, 'import pulumi\n'), ((709984, 710018), 'pulumi.set', 'pulumi.set', (['__self__', '"""port"""', 'port'], {}), "(__self__, 'port', port)\n", (709994, 710018), False, 'import pulumi\n'), ((710073, 710133), 'pulumi.set', 'pulumi.set', (['__self__', '"""predefined_target"""', 'predefined_target'], {}), "(__self__, 'predefined_target', predefined_target)\n", (710083, 710133), False, 'import pulumi\n'), ((710188, 710248), 'pulumi.set', 'pulumi.set', (['__self__', '"""security_group_id"""', 'security_group_id'], {}), "(__self__, 'security_group_id', security_group_id)\n", (710198, 710248), False, 'import pulumi\n'), ((710293, 710333), 'pulumi.set', 'pulumi.set', (['__self__', '"""to_port"""', 'to_port'], {}), "(__self__, 'to_port', to_port)\n", (710303, 710333), False, 'import pulumi\n'), ((710385, 710439), 'pulumi.set', 'pulumi.set', (['__self__', '"""v4_cidr_blocks"""', 'v4_cidr_blocks'], {}), "(__self__, 'v4_cidr_blocks', v4_cidr_blocks)\n", (710395, 710439), False, 'import pulumi\n'), ((710491, 710545), 'pulumi.set', 'pulumi.set', (['__self__', '"""v6_cidr_blocks"""', 'v6_cidr_blocks'], {}), "(__self__, 'v6_cidr_blocks', v6_cidr_blocks)\n", (710501, 710545), False, 'import pulumi\n'), ((714618, 714666), 'pulumi.set', 'pulumi.set', (['__self__', '"""domain_name"""', 'domain_name'], {}), "(__self__, 'domain_name', domain_name)\n", (714628, 714666), False, 'import pulumi\n'), ((714723, 714787), 'pulumi.set', 'pulumi.set', (['__self__', '"""domain_name_servers"""', 'domain_name_servers'], {}), "(__self__, 'domain_name_servers', domain_name_servers)\n", (714733, 714787), False, 'import pulumi\n'), ((714836, 714884), 'pulumi.set', 'pulumi.set', (['__self__', '"""ntp_servers"""', 'ntp_servers'], {}), "(__self__, 'ntp_servers', ntp_servers)\n", (714846, 714884), False, 'import pulumi\n'), ((716499, 716537), 'pulumi.set', 'pulumi.set', (['__self__', '"""region"""', 'region'], {}), "(__self__, 'region', region)\n", (716509, 716537), False, 'import pulumi\n'), ((760595, 760643), 'pulumi.set', 'pulumi.set', (['__self__', '"""preemptible"""', 'preemptible'], {}), "(__self__, 'preemptible', preemptible)\n", (760605, 760643), False, 'import pulumi\n'), ((761336, 761402), 'pulumi.set', 'pulumi.set', (['__self__', '"""zone_instances_limit"""', 'zone_instances_limit'], {}), "(__self__, 'zone_instances_limit', zone_instances_limit)\n", (761346, 761402), False, 'import pulumi\n'), ((761459, 761523), 'pulumi.set', 'pulumi.set', (['__self__', '"""zone_requests_limit"""', 'zone_requests_limit'], {}), "(__self__, 'zone_requests_limit', zone_requests_limit)\n", (761469, 761523), False, 'import pulumi\n'), ((766431, 766489), 'pulumi.set', 'pulumi.set', (['__self__', '"""assign_public_ip"""', 'assign_public_ip'], {}), "(__self__, 'assign_public_ip', assign_public_ip)\n", (766441, 766489), False, 'import pulumi\n'), ((766540, 766592), 'pulumi.set', 'pulumi.set', (['__self__', '"""brokers_count"""', 'brokers_count'], {}), "(__self__, 'brokers_count', brokers_count)\n", (766550, 766592), False, 'import pulumi\n'), ((766645, 766701), 'pulumi.set', 'pulumi.set', (['__self__', '"""schema_registry"""', 'schema_registry'], {}), "(__self__, 'schema_registry', schema_registry)\n", (766655, 766701), False, 'import pulumi\n'), ((766755, 766813), 'pulumi.set', 'pulumi.set', (['__self__', '"""unmanaged_topics"""', 'unmanaged_topics'], {}), "(__self__, 'unmanaged_topics', unmanaged_topics)\n", (766765, 766813), False, 'import pulumi\n'), ((770499, 770549), 'pulumi.set', 'pulumi.set', (['__self__', '"""kafka_config"""', 'kafka_config'], {}), "(__self__, 'kafka_config', kafka_config)\n", (770509, 770549), False, 'import pulumi\n'), ((772565, 772641), 'pulumi.set', 'pulumi.set', (['__self__', '"""auto_create_topics_enable"""', 'auto_create_topics_enable'], {}), "(__self__, 'auto_create_topics_enable', auto_create_topics_enable)\n", (772575, 772641), False, 'import pulumi\n'), ((772695, 772753), 'pulumi.set', 'pulumi.set', (['__self__', '"""compression_type"""', 'compression_type'], {}), "(__self__, 'compression_type', compression_type)\n", (772705, 772753), False, 'import pulumi\n'), ((772817, 772895), 'pulumi.set', 'pulumi.set', (['__self__', '"""default_replication_factor"""', 'default_replication_factor'], {}), "(__self__, 'default_replication_factor', default_replication_factor)\n", (772827, 772895), False, 'import pulumi\n'), ((772960, 773045), 'pulumi.set', 'pulumi.set', (['__self__', '"""log_flush_interval_messages"""', 'log_flush_interval_messages'], {}), "(__self__, 'log_flush_interval_messages', log_flush_interval_messages\n )\n", (772970, 773045), False, 'import pulumi\n'), ((773099, 773167), 'pulumi.set', 'pulumi.set', (['__self__', '"""log_flush_interval_ms"""', 'log_flush_interval_ms'], {}), "(__self__, 'log_flush_interval_ms', log_flush_interval_ms)\n", (773109, 773167), False, 'import pulumi\n'), ((773236, 773328), 'pulumi.set', 'pulumi.set', (['__self__', '"""log_flush_scheduler_interval_ms"""', 'log_flush_scheduler_interval_ms'], {}), "(__self__, 'log_flush_scheduler_interval_ms',\n log_flush_scheduler_interval_ms)\n", (773246, 773328), False, 'import pulumi\n'), ((773377, 773433), 'pulumi.set', 'pulumi.set', (['__self__', '"""log_preallocate"""', 'log_preallocate'], {}), "(__self__, 'log_preallocate', log_preallocate)\n", (773387, 773433), False, 'import pulumi\n'), ((773490, 773554), 'pulumi.set', 'pulumi.set', (['__self__', '"""log_retention_bytes"""', 'log_retention_bytes'], {}), "(__self__, 'log_retention_bytes', log_retention_bytes)\n", (773500, 773554), False, 'import pulumi\n'), ((773611, 773675), 'pulumi.set', 'pulumi.set', (['__self__', '"""log_retention_hours"""', 'log_retention_hours'], {}), "(__self__, 'log_retention_hours', log_retention_hours)\n", (773621, 773675), False, 'import pulumi\n'), ((773734, 773802), 'pulumi.set', 'pulumi.set', (['__self__', '"""log_retention_minutes"""', 'log_retention_minutes'], {}), "(__self__, 'log_retention_minutes', log_retention_minutes)\n", (773744, 773802), False, 'import pulumi\n'), ((773856, 773914), 'pulumi.set', 'pulumi.set', (['__self__', '"""log_retention_ms"""', 'log_retention_ms'], {}), "(__self__, 'log_retention_ms', log_retention_ms)\n", (773866, 773914), False, 'import pulumi\n'), ((773969, 774029), 'pulumi.set', 'pulumi.set', (['__self__', '"""log_segment_bytes"""', 'log_segment_bytes'], {}), "(__self__, 'log_segment_bytes', log_segment_bytes)\n", (773979, 774029), False, 'import pulumi\n'), ((774081, 774135), 'pulumi.set', 'pulumi.set', (['__self__', '"""num_partitions"""', 'num_partitions'], {}), "(__self__, 'num_partitions', num_partitions)\n", (774091, 774135), False, 'import pulumi\n'), ((774200, 774285), 'pulumi.set', 'pulumi.set', (['__self__', '"""socket_receive_buffer_bytes"""', 'socket_receive_buffer_bytes'], {}), "(__self__, 'socket_receive_buffer_bytes', socket_receive_buffer_bytes\n )\n", (774210, 774285), False, 'import pulumi\n'), ((774342, 774416), 'pulumi.set', 'pulumi.set', (['__self__', '"""socket_send_buffer_bytes"""', 'socket_send_buffer_bytes'], {}), "(__self__, 'socket_send_buffer_bytes', socket_send_buffer_bytes)\n", (774352, 774416), False, 'import pulumi\n'), ((784908, 784958), 'pulumi.set', 'pulumi.set', (['__self__', '"""topic_config"""', 'topic_config'], {}), "(__self__, 'topic_config', topic_config)\n", (784918, 784958), False, 'import pulumi\n'), ((787550, 787604), 'pulumi.set', 'pulumi.set', (['__self__', '"""cleanup_policy"""', 'cleanup_policy'], {}), "(__self__, 'cleanup_policy', cleanup_policy)\n", (787560, 787604), False, 'import pulumi\n'), ((787658, 787716), 'pulumi.set', 'pulumi.set', (['__self__', '"""compression_type"""', 'compression_type'], {}), "(__self__, 'compression_type', compression_type)\n", (787668, 787716), False, 'import pulumi\n'), ((787773, 787837), 'pulumi.set', 'pulumi.set', (['__self__', '"""delete_retention_ms"""', 'delete_retention_ms'], {}), "(__self__, 'delete_retention_ms', delete_retention_ms)\n", (787783, 787837), False, 'import pulumi\n'), ((787895, 787961), 'pulumi.set', 'pulumi.set', (['__self__', '"""file_delete_delay_ms"""', 'file_delete_delay_ms'], {}), "(__self__, 'file_delete_delay_ms', file_delete_delay_ms)\n", (787905, 787961), False, 'import pulumi\n'), ((788013, 788067), 'pulumi.set', 'pulumi.set', (['__self__', '"""flush_messages"""', 'flush_messages'], {}), "(__self__, 'flush_messages', flush_messages)\n", (788023, 788067), False, 'import pulumi\n'), ((788113, 788155), 'pulumi.set', 'pulumi.set', (['__self__', '"""flush_ms"""', 'flush_ms'], {}), "(__self__, 'flush_ms', flush_ms)\n", (788123, 788155), False, 'import pulumi\n'), ((788210, 788270), 'pulumi.set', 'pulumi.set', (['__self__', '"""max_message_bytes"""', 'max_message_bytes'], {}), "(__self__, 'max_message_bytes', max_message_bytes)\n", (788220, 788270), False, 'import pulumi\n'), ((788329, 788397), 'pulumi.set', 'pulumi.set', (['__self__', '"""min_compaction_lag_ms"""', 'min_compaction_lag_ms'], {}), "(__self__, 'min_compaction_lag_ms', min_compaction_lag_ms)\n", (788339, 788397), False, 'import pulumi\n'), ((788454, 788518), 'pulumi.set', 'pulumi.set', (['__self__', '"""min_insync_replicas"""', 'min_insync_replicas'], {}), "(__self__, 'min_insync_replicas', min_insync_replicas)\n", (788464, 788518), False, 'import pulumi\n'), ((788567, 788615), 'pulumi.set', 'pulumi.set', (['__self__', '"""preallocate"""', 'preallocate'], {}), "(__self__, 'preallocate', preallocate)\n", (788577, 788615), False, 'import pulumi\n'), ((788668, 788724), 'pulumi.set', 'pulumi.set', (['__self__', '"""retention_bytes"""', 'retention_bytes'], {}), "(__self__, 'retention_bytes', retention_bytes)\n", (788678, 788724), False, 'import pulumi\n'), ((788774, 788824), 'pulumi.set', 'pulumi.set', (['__self__', '"""retention_ms"""', 'retention_ms'], {}), "(__self__, 'retention_ms', retention_ms)\n", (788784, 788824), False, 'import pulumi\n'), ((788875, 788927), 'pulumi.set', 'pulumi.set', (['__self__', '"""segment_bytes"""', 'segment_bytes'], {}), "(__self__, 'segment_bytes', segment_bytes)\n", (788885, 788927), False, 'import pulumi\n'), ((793379, 793427), 'pulumi.set', 'pulumi.set', (['__self__', '"""permissions"""', 'permissions'], {}), "(__self__, 'permissions', permissions)\n", (793389, 793427), False, 'import pulumi\n')]
""" Project resources Many configuration and scripting resources are extracted here. """ from shared.tools.snapshot.utils import encode, hashmapToDict def extract_project_props(client_context): global_props = client_context.getGlobalProps() configuration = { 'permissions': hashmapToDict(global_props.getPermissionEnabledMap()), 'roles': { 'client': dict((category, [role.strip() for role in role_string.split(',') if role ]) for category, role_string in hashmapToDict( global_props.getRequiredClientRolesMap() ).items()), 'delete' : [role.strip() for role in global_props.getRequiredDeleteRoles()], 'publish' : [role.strip() for role in global_props.getRequiredPublishRoles()], 'resource': [role.strip() for role in global_props.getRequiredResourceRoles()], 'required': [role.strip() for role in global_props.getRequiredRoles()], 'save' : [role.strip() for role in global_props.getRequiredSaveRoles()], 'view' : [role.strip() for role in global_props.getRequiredViewRoles()], }, 'auditing': global_props.isAuditingEnabled(), 'legacy': global_props.isLegacyProject(), 'commitMessageMode': global_props.getCommitMessageMode().toString(), # enum 'defaultSQLTagsProviderRate': global_props.getSqltagsClientPollRate(), } defaultable_attributes = set([ 'auditProfileName', 'authProfileName', 'defaultDatasourceName', 'defaultSQLTagsProviderName', 'publishMode', ]) for attribute in defaultable_attributes: try: # to get the Java getter first # it's slightly more reliable than the Jython auto-attribute, in general getter_name = 'get' + attribute[0].upper() + attribute[1:] value = getattr(global_props, getter_name)() except AttributeError: try: # the Jython attribute value = getattr(global_props, attribute) except AttributeError: value = None if value is None: continue configuration[attribute] = value return dict([ encode(configuration), ]) def extract_gatewayevents(resource_objects): assert len(resource_objects) == 1, 'Resource is expected to be contained in one root object' client_script_config = resource_objects[0] scripts = {} script = client_script_config.getStartupScript() if script: scripts['startup.py'] = script script = client_script_config.getShutdownScript() if script: scripts['shutdown.py'] = script timer_scripts = client_script_config.getTimerScripts() for timer_script in timer_scripts: suffix, serialized = encode({ 'enabled': timer_script.isEnabled(), 'timing': 'delay' if timer_script.isFixedDelay() else 'rate', 'period': timer_script.getDelay(), 'threading': 'shared' if timer_script.isSharedThread() else 'dedicated', }) scripts['timer/%s%s' % (timer_script.getName(), suffix)] = serialized scripts['timer/%s.py' % timer_script.getName()] = timer_scripts[timer_script] for tag_script in client_script_config.getTagChangeScripts(): suffix, serialized = encode({ 'name': tag_script.getName(), 'tags': [tag_path for tag_path in tag_script.getPaths()], 'triggers': [t.toString() for t in tag_script.getChangeTypes()], 'enabled': tag_script.isEnabled(), }) scripts['tag-change/%s%s' % (tag_script.getName(), suffix)] = serialized scripts['tag-change/%s.py' % tag_script.getName()] = tag_script.getScript() message_scripts = client_script_config.getMessageHandlerScripts() for message_script in message_scripts: suffix, serialized = encode({ 'name': message_script.getName(), 'threading': str(message_script.getThreadType()), 'enabled': message_script.isEnabled(), }) scripts['message/%s%s' % (message_script.getName(),suffix)] = serialized scripts['message/%s.py' % message_script.getName()] = message_scripts[message_script] return scripts def extract_clientevents(resource_objects): assert len(resource_objects) == 1, 'Resource is expected to be contained in one root object' client_script_config = resource_objects[0] scripts = {} script = client_script_config.getStartupScript() if script: scripts['startup.py'] = script script = client_script_config.getShutdownScript() if script: scripts['shutdown.py'] = script script = client_script_config.getShutdownAllowedScript() if script: scripts['shutdown-intercept.py'] = script key_schema_pattern = re.compile("(\[(?P<modifiers>.*)\] )?(?P<key>.*) \((?P<action>.*)\)") key_modifier_pattern = re.compile("(Button \d|\w+)") key_scripts = client_script_config.getKeyScripts() for kix, key_script in enumerate(key_scripts): key_config = key_schema_pattern.match(key_script.getDisplay()).groupdict() suffix, serialized = encode({ 'action': key_config['action'], 'key': key_config['key'].replace("'", ''), 'modifiers': key_modifier_pattern.findall(key_config['modifiers']) if key_config['modifiers'] else [] }) scripts['key/%s%s' % (key_script.getDisplay(), suffix)] = serialized scripts['key/%s.py' % key_script.getDisplay()] = key_scripts[key_script] timer_scripts = client_script_config.getTimerScripts() for timer_script in timer_scripts: suffix, serialized = encode({ 'enabled': timer_script.isEnabled(), 'timing': 'delay' if timer_script.isFixedDelay() else 'rate', 'period': timer_script.getDelay(), 'threading': 'shared' if timer_script.isSharedThread() else 'dedicated', }) scripts['timer/%s%s' % (timer_script.getName(), suffix)] = serialized scripts['timer/%s.py' % timer_script.getName()] = timer_scripts[timer_script] for tag_script in client_script_config.getTagChangeScripts(): suffix, serialized = encode({ 'name': tag_script.getName(), 'tags': [tag_path for tag_path in tag_script.getPaths()], 'triggers': [t.toString() for t in tag_script.getChangeTypes()], 'enabled': tag_script.isEnabled(), }) scripts['tag-change/%s%s' % (tag_script.getName(), suffix)] = serialized scripts['tag-change/%s.py' % tag_script.getName()] = tag_script.getScript() def traverse_menu(parent_path, menu_node, mutable_dict): for mix, child in enumerate(menu_node.getChildren() or []): suffix, serialized = encode({ 'name': child.getName(), 'icon': child.getIconPath(), 'mnemonic': child.getMnemonic(), 'description': child.getDescription(), 'accelerator': child.getAccelerator(), }) mutable_dict['%s/entry-%02d%s' % ('/'.join(parent_path), mix, suffix)] = serialized mutable_dict['%s/entry-%02d.py' % ('/'.join(parent_path), mix)] = child.getScript() traverse_menu(parent_path + [child.getName() or ('Submenu-%02d' % mix)], child, mutable_dict) menu_root = client_script_config.getMenuRoot() traverse_menu(['menu'], menu_root, scripts) message_scripts = client_script_config.getMessageHandlerScripts() for message_script in message_scripts: suffix, serialized = encode({ 'name': message_script.getName(), 'threading': str(message_script.getThreadType()), 'enabled': message_script.isEnabled(), }) scripts['message/%s%s' % (message_script.getName(), suffix)] = serialized scripts['message/%s.py' % message_script.getName()] = message_scripts[message_script] return scripts def extract_namedquery(resource_objects): assert len(resource_objects) == 1, 'Resource is expected to be contained in one root object' named_query = resource_objects[0] info = { 'query': named_query.getQuery(), 'database': named_query.getDatabase() or '-default-', 'parameters': dict( (param.getIdentifier(), { 'sql_type' : str(param.getSqlType()), 'type' : str(param.getType()), 'identifier': str(param.getIdentifier()), }) for param in named_query.getParameters() ), 'type': named_query.getType(), } return dict([ ('.sql', format_sql(info['query'])), encode(info), ]) def extract_project_script(resource_objects): assert len(resource_objects) == 1, 'Resource is expected to be contained in one root object' script = resource_objects[0] return { '.py': script, } # Ready for the dispatcher EXTRACTORS = { 'sr.script.project': extract_project_script, 'named-query': extract_namedquery, 'client.event.scripts': extract_clientevents, 'event.scripts': extract_gatewayevents, 'project/properties': extract_project_props, }
[ "shared.tools.snapshot.utils.encode" ]
[((2021, 2042), 'shared.tools.snapshot.utils.encode', 'encode', (['configuration'], {}), '(configuration)\n', (2027, 2042), False, 'from shared.tools.snapshot.utils import encode, hashmapToDict\n'), ((7840, 7852), 'shared.tools.snapshot.utils.encode', 'encode', (['info'], {}), '(info)\n', (7846, 7852), False, 'from shared.tools.snapshot.utils import encode, hashmapToDict\n')]
# -*- coding: utf-8 -*- """Polynomial techniques for fitting baselines to experimental data. Created on Feb. 27, 2021 @author: <NAME> The function penalized_poly was adapted from MATLAB code from https://www.mathworks.com/matlabcentral/fileexchange/27429-background-correction (accessed March 18, 2021), which was licensed under the BSD-2-clause below. License: 2-clause BSD Copyright (c) 2012, <NAME> All rights reserved. Redistribution and use in source and binary forms, with or without modification, are permitted provided that the following conditions are met: * Redistributions of source code must retain the above copyright notice, this list of conditions and the following disclaimer. * Redistributions in binary form must reproduce the above copyright notice, this list of conditions and the following disclaimer in the documentation and/or other materials provided with the distribution THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. The function loess was adapted from code from https://gist.github.com/agramfort/850437 (accessed March 25, 2021), which was licensed under the BSD-3-clause below. # Authors: <NAME> <<EMAIL>> # # License: BSD (3-clause) Copyright (c) 2015, <NAME> All rights reserved. Redistribution and use in source and binary forms, with or without modification, are permitted provided that the following conditions are met: 1. Redistributions of source code must retain the above copyright notice, this list of conditions and the following disclaimer. 2. Redistributions in binary form must reproduce the above copyright notice, this list of conditions and the following disclaimer in the documentation and/or other materials provided with the distribution. 3. Neither the name of the copyright holder nor the names of its contributors may be used to endorse or promote products derived from this software without specific prior written permission. THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. """ from math import ceil import warnings import numpy as np from . import _weighting from ._algorithm_setup import _get_vander, _setup_polynomial from ._compat import jit, prange from .utils import ( _MIN_FLOAT, ParameterWarning, _convert_coef, _interp_inplace, _inverted_sort, relative_difference ) def poly(data, x_data=None, poly_order=2, weights=None, return_coef=False): """ Computes a polynomial that fits the baseline of the data. Parameters ---------- data : array-like, shape (N,) The y-values of the measured data, with N data points. x_data : array-like, shape (N,), optional The x-values of the measured data. Default is None, which will create an array from -1 to 1 with N points. poly_order : int, optional The polynomial order for fitting the baseline. Default is 2. weights : array-like, shape (N,), optional The weighting array. If None (default), then will be an array with size equal to N and all values set to 1. return_coef : bool, optional If True, will convert the polynomial coefficients for the fit baseline to a form that fits the input x_data and return them in the params dictionary. Default is False, since the conversion takes time. Returns ------- baseline : numpy.ndarray, shape (N,) The calculated baseline. params : dict A dictionary with the following items: * 'weights': numpy.ndarray, shape (N,) The weight array used for fitting the data. * 'coef': numpy.ndarray, shape (poly_order,) Only if `return_coef` is True. The array of polynomial parameters for the baseline, in increasing order. Can be used to create a polynomial using numpy.polynomial.polynomial.Polynomial(). Notes ----- To only fit regions without peaks, supply a weight array with zero values at the indices where peaks are located. """ y, x, weight_array, original_domain = _setup_polynomial(data, x_data, weights) fit_polynomial = np.polynomial.Polynomial.fit(x, y, poly_order, w=np.sqrt(weight_array)) baseline = fit_polynomial(x) params = {'weights': weight_array} if return_coef: params['coef'] = fit_polynomial.convert(window=original_domain).coef return baseline, params def modpoly(data, x_data=None, poly_order=2, tol=1e-3, max_iter=250, weights=None, use_original=False, mask_initial_peaks=False, return_coef=False): """ The modified polynomial (ModPoly) baseline algorithm. Parameters ---------- data : array-like, shape (N,) The y-values of the measured data, with N data points. x_data : array-like, shape (N,), optional The x-values of the measured data. Default is None, which will create an array from -1 to 1 with N points. poly_order : int, optional The polynomial order for fitting the baseline. Default is 2. tol : float, optional The exit criteria. Default is 1e-3. max_iter : int, optional The maximum number of iterations. Default is 250. weights : array-like, shape (N,), optional The weighting array. If None (default), then will be an array with size equal to N and all values set to 1. use_original : bool, optional If False (default), will compare the baseline of each iteration with the y-values of that iteration [1]_ when choosing minimum values. If True, will compare the baseline with the original y-values given by `data` [2]_. mask_initial_peaks : bool, optional If True, will mask any data where the initial baseline fit + the standard deviation of the residual is less than measured data [3]_. Default is False. return_coef : bool, optional If True, will convert the polynomial coefficients for the fit baseline to a form that fits the input x_data and return them in the params dictionary. Default is False, since the conversion takes time. Returns ------- baseline : numpy.ndarray, shape (N,) The calculated baseline. params : dict A dictionary with the following items: * 'weights': numpy.ndarray, shape (N,) The weight array used for fitting the data. * 'tol_history': numpy.ndarray An array containing the calculated tolerance values for each iteration. The length of the array is the number of iterations completed. If the last value in the array is greater than the input `tol` value, then the function did not converge. * 'coef': numpy.ndarray, shape (poly_order + 1,) Only if `return_coef` is True. The array of polynomial parameters for the baseline, in increasing order. Can be used to create a polynomial using numpy.polynomial.polynomial.Polynomial(). Notes ----- Algorithm originally developed in [2]_ and then slightly modified in [1]_. References ---------- .. [1] <NAME>., et al. Baseline correction by improved iterative polynomial fitting with automatic threshold. Chemometrics and Intelligent Laboratory Systems, 2006, 82, 59-65. .. [2] <NAME>., et al. Automated method for subtraction of fluorescence from biological raman spectra. Applied Spectroscopy, 2003, 57(11), 1363-1367. .. [3] <NAME>., et al. Automated Autofluorescence Background Subtraction Algorithm for Biomedical Raman Spectroscopy, Applied Spectroscopy, 2007, 61(11), 1225-1232. """ y, x, weight_array, original_domain, vander, pseudo_inverse = _setup_polynomial( data, x_data, weights, poly_order, True, True, True ) sqrt_w = np.sqrt(weight_array) if use_original: y0 = y coef = np.dot(pseudo_inverse, sqrt_w * y) baseline = np.dot(vander, coef) if mask_initial_peaks: # use baseline + deviation since without deviation, half of y should be above baseline weight_array[baseline + np.std(y - baseline) < y] = 0 sqrt_w = np.sqrt(weight_array) vander, pseudo_inverse = _get_vander(x, poly_order, sqrt_w) tol_history = np.empty(max_iter) for i in range(max_iter): baseline_old = baseline y = np.minimum(y0 if use_original else y, baseline) coef = np.dot(pseudo_inverse, sqrt_w * y) baseline = np.dot(vander, coef) calc_difference = relative_difference(baseline_old, baseline) tol_history[i] = calc_difference if calc_difference < tol: break params = {'weights': weight_array, 'tol_history': tol_history[:i + 1]} if return_coef: params['coef'] = _convert_coef(coef, original_domain) return baseline, params def imodpoly(data, x_data=None, poly_order=2, tol=1e-3, max_iter=250, weights=None, use_original=False, mask_initial_peaks=True, return_coef=False, num_std=1): """ The improved modofied polynomial (IModPoly) baseline algorithm. Parameters ---------- data : array-like, shape (N,) The y-values of the measured data, with N data points. x_data : array-like, shape (N,), optional The x-values of the measured data. Default is None, which will create an array from -1 to 1 with N points. poly_order : int, optional The polynomial order for fitting the baseline. Default is 2. tol : float, optional The exit criteria. Default is 1e-3. max_iter : int, optional The maximum number of iterations. Default is 250. weights : array-like, shape (N,), optional The weighting array. If None (default), then will be an array with size equal to N and all values set to 1. use_original : bool, optional If False (default), will compare the baseline of each iteration with the y-values of that iteration [4]_ when choosing minimum values. If True, will compare the baseline with the original y-values given by `data` [5]_. mask_initial_peaks : bool, optional If True (default), will mask any data where the initial baseline fit + the standard deviation of the residual is less than measured data [6]_. return_coef : bool, optional If True, will convert the polynomial coefficients for the fit baseline to a form that fits the input x_data and return them in the params dictionary. Default is False, since the conversion takes time. num_std : float, optional The number of standard deviations to include when thresholding. Default is 1. Returns ------- baseline : numpy.ndarray, shape (N,) The calculated baseline. params : dict A dictionary with the following items: * 'weights': numpy.ndarray, shape (N,) The weight array used for fitting the data. * 'tol_history': numpy.ndarray An array containing the calculated tolerance values for each iteration. The length of the array is the number of iterations completed. If the last value in the array is greater than the input `tol` value, then the function did not converge. * 'coef': numpy.ndarray, shape (poly_order + 1,) Only if `return_coef` is True. The array of polynomial parameters for the baseline, in increasing order. Can be used to create a polynomial using numpy.polynomial.polynomial.Polynomial(). Notes ----- Algorithm originally developed in [6]_. References ---------- .. [4] <NAME>., et al. Baseline correction by improved iterative polynomial fitting with automatic threshold. Chemometrics and Intelligent Laboratory Systems, 2006, 82, 59-65. .. [5] <NAME>., et al. Automated method for subtraction of fluorescence from biological raman spectra. Applied Spectroscopy, 2003, 57(11), 1363-1367. .. [6] <NAME>., et al. Automated Autofluorescence Background Subtraction Algorithm for Biomedical Raman Spectroscopy, Applied Spectroscopy, 2007, 61(11), 1225-1232. """ y, x, weight_array, original_domain, vander, pseudo_inverse = _setup_polynomial( data, x_data, weights, poly_order, True, True, True ) sqrt_w = np.sqrt(weight_array) if use_original: y0 = y coef = np.dot(pseudo_inverse, sqrt_w * y) baseline = np.dot(vander, coef) deviation = np.std(y - baseline) if mask_initial_peaks: weight_array[baseline + deviation < y] = 0 sqrt_w = np.sqrt(weight_array) vander, pseudo_inverse = _get_vander(x, poly_order, sqrt_w) tol_history = np.empty(max_iter) for i in range(max_iter): y = np.minimum(y0 if use_original else y, baseline + num_std * deviation) coef = np.dot(pseudo_inverse, sqrt_w * y) baseline = np.dot(vander, coef) new_deviation = np.std(y - baseline) # use new_deviation as dividing term in relative difference calc_difference = relative_difference(new_deviation, deviation) tol_history[i] = calc_difference if calc_difference < tol: break deviation = new_deviation params = {'weights': weight_array, 'tol_history': tol_history[:i + 1]} if return_coef: params['coef'] = _convert_coef(coef, original_domain) return baseline, params # adapted from (https://www.mathworks.com/matlabcentral/fileexchange/27429-background-correction); # see license above def _huber_loss(residual, threshold=1.0, alpha_factor=0.99, symmetric=True): """ The Huber non-quadratic cost function. Parameters ---------- residual : numpy.ndarray, shape (N,) The residual array. threshold : float, optional Any residual values below the threshold are given quadratic loss. Default is 1.0. alpha_factor : float, optional The scale between 0 and 1 to multiply the cost function's alpha_max value (see Notes below). Default is 0.99. symmetric : bool, optional If True (default), the cost function is symmetric and applies the same weighting for positive and negative values. If False, will apply weights asymmetrically so that only positive weights are given the non-quadratic weigting and negative weights have normal, quadratic weighting. Returns ------- weights : numpy.ndarray, shape (N,) The weight array. Notes ----- The returned result is -residual + alpha_factor * alpha_max * phi'(residual) where phi'(x) is the derivative of the huber loss function, phi(x). References ---------- <NAME>., et al. Background removal from spectra by designing and minimising a non-quadratic cost function. Chemometrics and Intelligent Laboratory Systems, 2005, 76(2), 121–133. """ alpha = alpha_factor * 0.5 # alpha_max for huber is 0.5 if symmetric: mask = (np.abs(residual) < threshold) weights = ( mask * residual * (2 * alpha - 1) + (~mask) * 2 * alpha * threshold * np.sign(residual) ) else: mask = (residual < threshold) weights = ( mask * residual * (2 * alpha - 1) + (~mask) * (2 * alpha * threshold - residual) ) return weights # adapted from (https://www.mathworks.com/matlabcentral/fileexchange/27429-background-correction); # see license above def _truncated_quadratic_loss(residual, threshold=1.0, alpha_factor=0.99, symmetric=True): """ The Truncated-Quadratic non-quadratic cost function. Parameters ---------- residual : numpy.ndarray, shape (N,) The residual array. threshold : float, optional Any residual values below the threshold are given quadratic loss. Default is 1.0. alpha_factor : float, optional The scale between 0 and 1 to multiply the cost function's alpha_max value (see Notes below). Default is 0.99. symmetric : bool, optional If True (default), the cost function is symmetric and applies the same weighting for positive and negative values. If False, will apply weights asymmetrically so that only positive weights are given the non-quadratic weigting and negative weights have normal, quadratic weighting. Returns ------- weights : numpy.ndarray, shape (N,) The weight array. Notes ----- The returned result is -residual + alpha_factor * alpha_max * phi'(residual) where phi'(x) is the derivative of the truncated quadratic function, phi(x). References ---------- Mazet, V., et al. Background removal from spectra by designing and minimising a non-quadratic cost function. Chemometrics and Intelligent Laboratory Systems, 2005, 76(2), 121–133. """ alpha = alpha_factor * 0.5 # alpha_max for truncated quadratic is 0.5 if symmetric: mask = (np.abs(residual) < threshold) else: mask = (residual < threshold) return mask * residual * (2 * alpha - 1) - (~mask) * residual def _indec_loss(residual, threshold=1.0, alpha_factor=0.99, symmetric=True): """ The Indec non-quadratic cost function. Parameters ---------- residual : numpy.ndarray, shape (N,) The residual array. threshold : float, optional Any residual values below the threshold are given quadratic loss. Default is 1.0. alpha_factor : float, optional The scale between 0 and 1 to multiply the cost function's alpha_max value (see Notes below). Default is 0.99. symmetric : bool, optional If True (default), the cost function is symmetric and applies the same weighting for positive and negative values. If False, will apply weights asymmetrically so that only positive weights are given the non-quadratic weigting and negative weights have normal, quadratic weighting. Returns ------- weights : numpy.ndarray, shape (N,) The weight array. Notes ----- The returned result is -residual + alpha_factor * alpha_max * phi'(residual) where phi'(x) is the derivative of the Indec function, phi(x). References ---------- <NAME>., et al. Goldindec: A Novel Algorithm for Raman Spectrum Baseline Correction. Applied Spectroscopy, 2015, 69(7), 834-842. <NAME>., et al. Background removal from spectra by designing and minimising a non-quadratic cost function. Chemometrics and Intelligent Laboratory Systems, 2005, 76(2), 121–133. """ alpha = alpha_factor * 0.5 # alpha_max for indec is 0.5 if symmetric: mask = (np.abs(residual) < threshold) multiple = np.sign(residual) else: mask = (residual < threshold) # multiple=1 is same as sign(residual) since residual is always > 0 # for asymmetric case, but this allows not doing the sign calculation multiple = 1 weights = ( mask * residual * (2 * alpha - 1) - (~mask) * ( residual + alpha * multiple * threshold**3 / np.maximum(2 * residual**2, _MIN_FLOAT) ) ) return weights def _identify_loss_method(loss_method): """ Identifies the symmetry for the given loss method. Parameters ---------- loss_method : str The loss method to use. Should have the symmetry identifier as the prefix. Returns ------- symmetric : bool True if `loss_method` had 's_' or 'symmetric_' as the prefix, else False. str The input `loss_method` value without the first section that indicated the symmetry. Raises ------ ValueError Raised if the loss method does not have the correct form. """ prefix, *split_method = loss_method.lower().split('_') if prefix not in ('a', 's', 'asymmetric', 'symmetric') or not split_method: raise ValueError('must specify loss function symmetry by appending "a_" or "s_"') if prefix in ('a', 'asymmetric'): symmetric = False else: symmetric = True return symmetric, '_'.join(split_method) # adapted from (https://www.mathworks.com/matlabcentral/fileexchange/27429-background-correction); # see license above def penalized_poly(data, x_data=None, poly_order=2, tol=1e-3, max_iter=250, weights=None, cost_function='asymmetric_truncated_quadratic', threshold=None, alpha_factor=0.99, return_coef=False): """ Fits a polynomial baseline using a non-quadratic cost function. The non-quadratic cost functions penalize residuals with larger values, giving a more robust fit compared to normal least-squares. Parameters ---------- data : array-like, shape (N,) The y-values of the measured data, with N data points. x_data : array-like, shape (N,), optional The x-values of the measured data. Default is None, which will create an array from -1 to 1 with N points. poly_order : int, optional The polynomial order for fitting the baseline. Default is 2. tol : float, optional The exit criteria. Default is 1e-3. max_iter : int, optional The maximum number of iterations. Default is 250. weights : array-like, shape (N,), optional The weighting array. If None (default), then will be an array with size equal to N and all values set to 1. cost_function : str, optional The non-quadratic cost function to minimize. Must indicate symmetry of the method by appending 'a' or 'asymmetric' for asymmetric loss, and 's' or 'symmetric' for symmetric loss. Default is 'asymmetric_truncated_quadratic'. Available methods, and their associated reference, are: * 'asymmetric_truncated_quadratic'[7]_ * 'symmetric_truncated_quadratic'[7]_ * 'asymmetric_huber'[7]_ * 'symmetric_huber'[7]_ * 'asymmetric_indec'[8]_ * 'symmetric_indec'[8]_ threshold : float, optional The threshold value for the loss method, where the function goes from quadratic loss (such as used for least squares) to non-quadratic. For symmetric loss methods, residual values with absolute value less than threshold will have quadratic loss. For asymmetric loss methods, residual values less than the threshold will have quadratic loss. Default is None, which sets `threshold` to one-tenth of the standard deviation of the input data. alpha_factor : float, optional A value between 0 and 1 that controls the value of the penalty. Default is 0.99. Typically should not need to change this value. return_coef : bool, optional If True, will convert the polynomial coefficients for the fit baseline to a form that fits the input x_data and return them in the params dictionary. Default is False, since the conversion takes time. Returns ------- baseline : numpy.ndarray, shape (N,) The calculated baseline. params : dict A dictionary with the following items: * 'weights': numpy.ndarray, shape (N,) The weight array used for fitting the data. * 'tol_history': numpy.ndarray An array containing the calculated tolerance values for each iteration. The length of the array is the number of iterations completed. If the last value in the array is greater than the input `tol` value, then the function did not converge. * 'coef': numpy.ndarray, shape (poly_order + 1,) Only if `return_coef` is True. The array of polynomial parameters for the baseline, in increasing order. Can be used to create a polynomial using numpy.polynomial.polynomial.Polynomial(). Raises ------ ValueError Raised if `alpha_factor` is not between 0 and 1. Notes ----- In baseline literature, this procedure is sometimes called "backcor". References ---------- .. [7] <NAME>., et al. Background removal from spectra by designing and minimising a non-quadratic cost function. Chemometrics and Intelligent Laboratory Systems, 2005, 76(2), 121–133. .. [8] <NAME>., et al. Goldindec: A Novel Algorithm for Raman Spectrum Baseline Correction. Applied Spectroscopy, 2015, 69(7), 834-842. """ if not 0 < alpha_factor <= 1: raise ValueError('alpha_factor must be between 0 and 1') symmetric_loss, method = _identify_loss_method(cost_function) loss_function = { 'huber': _huber_loss, 'truncated_quadratic': _truncated_quadratic_loss, 'indec': _indec_loss }[method] y, x, weight_array, original_domain, vander, pseudo_inverse = _setup_polynomial( data, x_data, weights, poly_order, return_vander=True, return_pinv=True ) if threshold is None: threshold = np.std(y) / 10 loss_kwargs = { 'threshold': threshold, 'alpha_factor': alpha_factor, 'symmetric': symmetric_loss } sqrt_w = np.sqrt(weight_array) y = sqrt_w * y coef = np.dot(pseudo_inverse, y) baseline = np.dot(vander, coef) tol_history = np.empty(max_iter) for i in range(max_iter): baseline_old = baseline coef = np.dot(pseudo_inverse, y + loss_function(y - sqrt_w * baseline, **loss_kwargs)) baseline = np.dot(vander, coef) calc_difference = relative_difference(baseline_old, baseline) tol_history[i] = calc_difference if calc_difference < tol: break params = {'weights': weight_array, 'tol_history': tol_history[:i + 1]} if return_coef: params['coef'] = _convert_coef(coef, original_domain) return baseline, params def _tukey_square(residual, scale=3, symmetric=False): """ The square root of Tukey's bisquare function. Parameters ---------- residual : numpy.ndarray, shape (N,) The residual array of the fit. scale : float, optional A scale factor applied to the weighted residuals to control the robustness of the fit. Default is 3.0. symmetric : bool, optional If False (default), will apply weighting asymmetrically, with residuals < 0 having full weight. If True, will apply weighting the same for both positive and negative residuals, which is regular LOESS. Returns ------- weights : numpy.ndarray, shape (N,) The weighting array. Notes ----- The function is technically sqrt(Tukey's bisquare) since the outer power of 2 is not performed. This is intentional, so that the square root for weighting in least squares does not need to be done, speeding up the calculation. References ---------- <NAME>., et al., Baseline subtraction using robust local regression estimation. J. Quantitative Spectroscopy and Radiative Transfer, 2001, 68, 179-193. """ if symmetric: inner = residual / scale weights = np.maximum(0, 1 - inner * inner) else: weights = np.ones_like(residual) mask = residual > 0 inner = residual[mask] / scale weights[mask] = np.maximum(0, 1 - inner * inner) return weights def _median_absolute_value(values): """ Computes the median absolute value (MAV) of an array. Parameters ---------- values : array-like The array of values to use for the calculation. Returns ------- float The scaled median absolute value for the input array. Notes ----- The 1/0.6744897501960817 scale factor is to make the result comparable to the standard deviation of a Gaussian distribution. The divisor is obtained by calculating the value at which the cumulative distribution function of a Gaussian distribution is 0.75 (see https://en.wikipedia.org/wiki/Median_absolute_deviation), which can be obtained by:: from scipy.special import ndtri ndtri(0.75) # equals 0.6744897501960817 To calculate the median absolute difference (MAD) using this function, simply do:: _median_absolute_value(values - np.median(values)) References ---------- <NAME>., et al., Baseline subtraction using robust local regression estimation. J. Quantitative Spectroscopy and Radiative Transfer, 2001, 68, 179-193. https://en.wikipedia.org/wiki/Median_absolute_deviation. """ return np.median(np.abs(values)) / 0.6744897501960817 @jit(nopython=True, cache=True) def _loess_solver(AT, b): """ Solves the equation `A x = b` given `A.T` and `b`. Parameters ---------- AT : numpy.ndarray, shape (M, N) The transposed `A` matrix. b : numpy.ndarray, shape (N,) The `b` array. Returns ------- numpy.ndarray, shape (N,) The solution to the normal equation. Notes ----- Uses np.linalg.solve (which uses LU decomposition) rather than np.linalg.lstsq (which uses SVD) since solve is ~30-60% faster. np.linalg.solve requires ``A.T * A``, which squares the condition number of ``A``, but on tested datasets the relative difference when using solve vs lstsq (using np.allclose) is ~1e-10 to 1e-13 for poly_orders of 1 or 2, which seems fine; the relative differences increase to ~1e-6 to 1e-9 for a poly_order of 3, and ~1e-4 to 1e-6 for a poly_order of 4, but loess should use a poly_order <= 2, so that should not be a problem. """ return np.linalg.solve(AT.dot(AT.T), AT.dot(b)) @jit(nopython=True, cache=True, parallel=True) def _fill_skips(x, baseline, skips): """ Fills in the skipped baseline points using linear interpolation. Parameters ---------- x : numpy.ndarray The array of x-values. baseline : numpy.ndarray The array of baseline values with all fit points allocated. All skipped points will be filled in using interpolation. skips : numpy.ndarray, shape (G, 2) The array of left and right indices that define the windows for interpolation, with length G being the number of interpolation segments. Indices are set such that `baseline[skips[i][0]:skips[i][1]]` will have fitted values at the first and last indices and all other values (the slice [1:-1]) will be calculated by interpolation. Notes ----- All changes to `baseline` are done inplace. """ for i in prange(skips.shape[0]): window = skips[i] left = window[0] right = window[1] _interp_inplace(x[left:right], baseline[left:right], baseline[left], baseline[right - 1]) # adapted from (https://gist.github.com/agramfort/850437); see license above @jit(nopython=True, cache=True, parallel=True) def _loess_low_memory(x, y, weights, coefs, vander, num_x, windows, fits): """ A version of loess that uses near constant memory. The distance-weighted kernel for each x-value is computed each loop, rather than cached, so memory usage is low but the calculation is slightly slower. Parameters ---------- x : numpy.ndarray, shape (N,) The x-values of the measured data, with N data points. y : numpy.ndarray, shape (N,) The y-values of the measured data, with N points. weights : numpy.ndarray, shape (N,) The array of weights. coefs : numpy.ndarray, shape (N, poly_order + 1) The array of polynomial coefficients (with polynomial order poly_order), for each value in `x`. vander : numpy.ndarray, shape (N, poly_order + 1) The Vandermonde matrix for the `x` array. num_x : int The number of data points in `x`, also known as N. windows : numpy.ndarray, shape (F, 2) An array of left and right indices that define the fitting window for each fit x-value. The length is F, which is the total number of fit points. If `fit_dx` is <= 0, F is equal to N, the total number of x-values. fits : numpy.ndarray, shape (F,) The array of indices indicating which x-values to fit. Notes ----- The coefficient array, `coefs`, is modified inplace. """ baseline = np.empty(num_x) y_fit = y * weights vander_fit = vander.T * weights for idx in prange(fits.shape[0]): i = fits[idx] window = windows[idx] left = window[0] right = window[1] difference = np.abs(x[left:right] - x[i]) difference = difference / max(difference[0], difference[-1]) difference = difference * difference * difference difference = 1 - difference kernel = np.sqrt(difference * difference * difference) coef = _loess_solver( kernel * vander_fit[:, left:right], kernel * y_fit[left:right] ) baseline[i] = vander[i].dot(coef) coefs[i] = coef return baseline # adapted from (https://gist.github.com/agramfort/850437); see license above @jit(nopython=True, cache=True, parallel=True) def _loess_first_loop(x, y, weights, coefs, vander, total_points, num_x, windows, fits): """ The initial fit for loess that also caches the window values for each x-value. Parameters ---------- x : numpy.ndarray, shape (N,) The x-values of the measured data, with N data points. y : numpy.ndarray, shape (N,) The y-values of the measured data, with N points. weights : numpy.ndarray, shape (N,) The array of weights. coefs : numpy.ndarray, shape (N, poly_order + 1) The array of polynomial coefficients (with polynomial order poly_order), for each value in `x`. vander : numpy.ndarray, shape (N, poly_order + 1) The Vandermonde matrix for the `x` array. total_points : int The number of points to include when fitting each x-value. num_x : int The number of data points in `x`, also known as N. windows : numpy.ndarray, shape (F, 2) An array of left and right indices that define the fitting window for each fit x-value. The length is F, which is the total number of fit points. If `fit_dx` is <= 0, F is equal to N, the total number of x-values. fits : numpy.ndarray, shape (F,) The array of indices indicating which x-values to fit. Returns ------- kernels : numpy.ndarray, shape (num_x, total_points) The array containing the distance-weighted kernel for each x-value. Notes ----- The coefficient array, `coefs`, is modified inplace. """ kernels = np.empty((num_x, total_points)) baseline = np.empty(num_x) y_fit = y * weights vander_fit = vander.T * weights for idx in prange(fits.shape[0]): i = fits[idx] window = windows[idx] left = window[0] right = window[1] difference = np.abs(x[left:right] - x[i]) difference = difference / max(difference[0], difference[-1]) difference = difference * difference * difference difference = 1 - difference kernel = np.sqrt(difference * difference * difference) kernels[i] = kernel coef = _loess_solver( kernel * vander_fit[:, left:right], kernel * y_fit[left:right] ) baseline[i] = vander[i].dot(coef) coefs[i] = coef return kernels, baseline @jit(nopython=True, cache=True, parallel=True) def _loess_nonfirst_loops(y, weights, coefs, vander, kernels, windows, num_x, fits): """ The loess fit to use after the first loop that uses the cached window values. Parameters ---------- y : numpy.ndarray, shape (N,) The y-values of the measured data, with N points. weights : numpy.ndarray, shape (N,) The array of weights. coefs : numpy.ndarray, shape (N, poly_order + 1) The array of polynomial coefficients (with polynomial order poly_order), for each value in `x`. vander : numpy.ndarray, shape (N, poly_order + 1) The Vandermonde matrix for the `x` array. kernels : numpy.ndarray, shape (N, total_points) The array containing the distance-weighted kernel for each x-value. Each kernel has a length of total_points. windows : numpy.ndarray, shape (F, 2) An array of left and right indices that define the fitting window for each fit x-value. The length is F, which is the total number of fit points. If `fit_dx` is <= 0, F is equal to N, the total number of x-values. num_x : int The total number of values, N. fits : numpy.ndarray, shape (F,) The array of indices indicating which x-values to fit. Notes ----- The coefficient array, `coefs`, is modified inplace. """ baseline = np.empty(num_x) y_fit = y * weights vander_fit = vander.T * weights for idx in prange(fits.shape[0]): i = fits[idx] window = windows[idx] left = window[0] right = window[1] kernel = kernels[i] coef = _loess_solver( kernel * vander_fit[:, left:right], kernel * y_fit[left:right] ) baseline[i] = vander[i].dot(coef) coefs[i] = coef return baseline @jit(nopython=True, cache=True) def _determine_fits(x, num_x, total_points, delta): """ Determines the x-values to fit and the left and right indices for each fit x-value. The windows are set before fitting so that fitting can be done in parallel when numba is installed, since the left and right indices would otherwise need to be determined in order. Similarly, determining which x-values to fit would not be able to be done in parallel since it requires knowledge of the last x-value fit. Parameters ---------- x : numpy.ndarray, shape (N,) The array of x-values. num_x : int The total number of x-values, N. total_points : int The number of values to include in each fitting window. delta : float If `delta` is > 0, will skip all but the last x-value in the range x_last + `delta`, where x_last is the last x-value to be fit. Fits all x-values if `delta` is <= 0. Returns ------- windows : numpy.ndarray, shape (F, 2) An array of left and right indices that define the fitting window for each fit x-value. The length is F, which is the total number of fit points. If `fit_dx` is <= 0, F is equal to N, the total number of x-values. Indices are set such that the number of values in `x[windows[i][0]:windows[i][1]] is equal to `total_points`. fits : numpy.ndarray, shape (F,) The array of indices indicating which x-values to fit. skips : numpy.ndarray, shape (G, 2) The array of left and right indices that define the windows for interpolation, with length G being the number of interpolation segments. G is 0 if `fit_dx` is <= 0. Indices are set such that `baseline[skips[i][0]:skips[i][1]]` will have fitted values at the first and last indices and all other values (the slice [1:-1]) will be calculated by interpolation. Notes ----- The dtype `np.intp` is used for `fits`, `skips`, and `windows` to be consistent with numpy since numpy internally uses that type when referring to indices. """ # faster to allocate array and return only filled in sections # rather than constanly appending to a list if delta > 0: check_fits = True fits = np.empty(num_x, dtype=np.intp) fits[0] = 0 # always fit first item skips = np.empty((num_x, 2), dtype=np.intp) else: # TODO maybe use another function when fitting all points in order # to skip the if check_fits check for every x-value; does it affect # calculation time that much? check_fits = False # TODO once numba minimum version is >= 0.47, can use dtype kwarg in np.arange fits = np.arange(num_x).astype(np.intp) # numba cannot compile in nopython mode when directly creating # np.array([], dtype=np.intp), so work-around by creating np.array([[0, 0]]) # and then index with [:total_skips], which becomes np.array([]) # since total_skips is 0 when delta is <= 0. skips = np.array([[0, 0]], dtype=np.intp) windows = np.empty((num_x, 2), dtype=np.intp) windows[0] = (0, total_points) total_fits = 1 total_skips = 0 skip_start = 0 skip_range = x[0] + delta left = 0 right = total_points for i in range(1, num_x - 1): x_val = x[i] if check_fits: # use x[i+1] rather than x[i] since it ensures that the last value within # the range x_last_fit + delta is used; x[i+1] is also guranteed to be >= x[i] if x[i + 1] < skip_range: if not skip_start: skip_start = i continue else: skip_range = x_val + delta fits[total_fits] = i if skip_start: skips[total_skips] = (skip_start - 1, i + 1) total_skips += 1 skip_start = 0 while right < num_x and x_val - x[left] > x[right] - x_val: left += 1 right += 1 window = windows[total_fits] window[0] = left window[1] = right total_fits += 1 if skip_start: # fit second to last x-value fits[total_fits] = num_x - 2 if x[-1] - x[-2] < x[-2] - x[num_x - total_points]: windows[total_fits] = (num_x - total_points, num_x) else: windows[total_fits] = (num_x - total_points - 1, num_x - 1) total_fits += 1 skips[total_skips] = (skip_start - 1, num_x - 1) total_skips += 1 # always fit last item fits[total_fits] = num_x - 1 windows[total_fits] = (num_x - total_points, num_x) total_fits += 1 return windows[:total_fits], fits[:total_fits], skips[:total_skips] def loess(data, x_data=None, fraction=0.2, total_points=None, poly_order=1, scale=3.0, tol=1e-3, max_iter=10, symmetric_weights=False, use_threshold=False, num_std=1, use_original=False, weights=None, return_coef=False, conserve_memory=True, delta=0.0): """ Locally estimated scatterplot smoothing (LOESS). Performs polynomial regression at each data point using the nearest points. Parameters ---------- data : array-like, shape (N,) The y-values of the measured data, with N data points. x_data : array-like, shape (N,), optional The x-values of the measured data. Default is None, which will create an array from -1 to 1 with N points. fraction : float, optional The fraction of N data points to include for the fitting on each point. Default is 0.2. Not used if `total_points` is not None. total_points : int, optional The total number of points to include for the fitting on each point. Default is None, which will use `fraction` * N to determine the number of points. scale : float, optional A scale factor applied to the weighted residuals to control the robustness of the fit. Default is 3.0, as used in [9]_. Note that the original loess procedure in [10]_ used a `scale` of ~4.05. poly_order : int, optional The polynomial order for fitting the baseline. Default is 1. tol : float, optional The exit criteria. Default is 1e-3. max_iter : int, optional The maximum number of iterations. Default is 10. symmetric_weights : bool, optional If False (default), will apply weighting asymmetrically, with residuals < 0 having a weight of 1, according to [9]_. If True, will apply weighting the same for both positive and negative residuals, which is regular LOESS. If `use_threshold` is True, this parameter is ignored. use_threshold : bool, optional If False (default), will compute weights each iteration to perform the robust fitting, which is regular LOESS. If True, will apply a threshold on the data being fit each iteration, based on the maximum values of the data and the fit baseline, as proposed by [11]_, similar to the modpoly and imodpoly techniques. num_std : float, optional The number of standard deviations to include when thresholding. Default is 1, which is the value used for the imodpoly technique. Only used if `use_threshold` is True. use_original : bool, optional If False (default), will compare the baseline of each iteration with the y-values of that iteration [12]_ when choosing minimum values for thresholding. If True, will compare the baseline with the original y-values given by `data` [13]_. Only used if `use_threshold` is True. weights : array-like, shape (N,), optional The weighting array. If None (default), then will be an array with size equal to N and all values set to 1. return_coef : bool, optional If True, will convert the polynomial coefficients for the fit baseline to a form that fits the input x_data and return them in the params dictionary. Default is False, since the conversion takes time. conserve_memory : bool, optional If False, will cache the distance-weighted kernels for each value in `x_data` on the first iteration and reuse them on subsequent iterations to save time. The shape of the array of kernels is (len(`x_data`), `total_points`). If True (default), will recalculate the kernels each iteration, which uses very little memory, but is slower. Can usually set to False unless `x_data` and`total_points` are quite large and the function causes memory issues when cacheing the kernels. If numba is installed, there is no significant time difference since the calculations are sped up. delta : float, optional If `delta` is > 0, will skip all but the last x-value in the range x_last + `delta`, where x_last is the last x-value to be fit using weighted least squares, and instead use linear interpolation to calculate the fit for those x-values (same behavior as in statsmodels [14]_ and Cleveland's original Fortran lowess implementation [15]_). Fits all x-values if `delta` is <= 0. Default is 0.0. Note that `x_data` is scaled to fit in the range [-1, 1], so `delta` should likewise be scaled. For example, if the desired `delta` value was ``0.01 * (max(x_data) - min(x_data))``, then the correctly scaled `delta` would be 0.02 (ie. ``0.01 * (1 - (-1))``). Returns ------- baseline : numpy.ndarray, shape (N,) The calculated baseline. params : dict A dictionary with the following items: * 'weights': numpy.ndarray, shape (N,) The weight array used for fitting the data. Does NOT contain the individual distance-weighted kernels for each x-value. * 'tol_history': numpy.ndarray An array containing the calculated tolerance values for each iteration. The length of the array is the number of iterations completed. If the last value in the array is greater than the input `tol` value, then the function did not converge. * 'coef': numpy.ndarray, shape (N, poly_order + 1) Only if `return_coef` is True. The array of polynomial parameters for the baseline, in increasing order. Can be used to create a polynomial using numpy.polynomial.polynomial.Polynomial(). If `delta` is > 0, the coefficients for any skipped x-value will all be 0. Raises ------ ValueError Raised if the number of points per window for the fitting is less than `poly_order` + 1 or greater than the total number of points. Notes ----- The iterative, robust, aspect of the fitting can be achieved either through reweighting based on the residuals (the typical usage), or thresholding the fit data based on the residuals, as proposed by [11]_, similar to the modpoly and imodpoly techniques. In baseline literature, this procedure is sometimes called "rbe", meaning "robust baseline estimate". References ---------- .. [9] <NAME>., et al. Baseline subtraction using robust local regression estimation. J. Quantitative Spectroscopy and Radiative Transfer, 2001, 68, 179-193. .. [10] <NAME>. Robust locally weighted regression and smoothing scatterplots. Journal of the American Statistical Association, 1979, 74(368), 829-836. .. [11] <NAME>. Comparison of Several Methods of Chromatographic Baseline Removal with a New Approach Based on Quantile Regression. Chromatographia, 2011, 73, 721-731. .. [12] <NAME>., et al. Baseline correction by improved iterative polynomial fitting with automatic threshold. Chemometrics and Intelligent Laboratory Systems, 2006, 82, 59-65. .. [13] <NAME>., et al. Automated method for subtraction of fluorescence from biological raman spectra. Applied Spectroscopy, 2003, 57(11), 1363-1367. .. [14] https://github.com/statsmodels/statsmodels. .. [15] https://www.netlib.org/go (lowess.f is the file). """ y, x, weight_array, original_domain = _setup_polynomial(data, x_data, weights, poly_order) num_x = x.shape[0] if total_points is None: total_points = ceil(fraction * num_x) if total_points < poly_order + 1: raise ValueError('total points must be greater than polynomial order + 1') elif total_points > num_x: raise ValueError(( 'points per window is higher than total number of points; lower either ' '"fraction" or "total_points"' )) elif poly_order > 2: warnings.warn( ('polynomial orders greater than 2 can have numerical issues;' ' consider using a polynomial order of 1 or 2 instead'), ParameterWarning ) sort_x = x_data is not None if sort_x: sort_order = np.argsort(x, kind='mergesort') # to ensure x is increasing x = x[sort_order] y = y[sort_order] weight_array = weight_array[sort_order] if use_original: y0 = y # find the indices for fitting beforehand so that the fitting can be done # in parallel; cast delta as float so numba does not have to compile for # both int and float windows, fits, skips = _determine_fits(x, num_x, total_points, float(delta)) # np.polynomial.polynomial.polyvander returns a Fortran-ordered array, which # when matrix multiplied with the C-ordered coefficient array gives a warning # when using numba, so convert Vandermonde matrix to C-ordering. vander = np.ascontiguousarray(_get_vander(x, poly_order, calc_pinv=False)) baseline = y coefs = np.zeros((num_x, poly_order + 1)) tol_history = np.empty(max_iter + 1) sqrt_w = np.sqrt(weight_array) # do max_iter + 1 since a max_iter of 0 would return y as baseline otherwise for i in range(max_iter + 1): baseline_old = baseline if conserve_memory: baseline = _loess_low_memory( x, y, sqrt_w, coefs, vander, num_x, windows, fits ) elif i == 0: kernels, baseline = _loess_first_loop( x, y, sqrt_w, coefs, vander, total_points, num_x, windows, fits ) else: baseline = _loess_nonfirst_loops( y, sqrt_w, coefs, vander, kernels, windows, num_x, fits ) _fill_skips(x, baseline, skips) calc_difference = relative_difference(baseline_old, baseline) tol_history[i] = calc_difference if calc_difference < tol: break if use_threshold: y = np.minimum( y0 if use_original else y, baseline + num_std * np.std(y - baseline) ) else: residual = y - baseline # TODO median_absolute_value can be 0 if more than half of residuals are # 0 (perfect fit); can that ever really happen? if so, should prevent dividing by 0 sqrt_w = _tukey_square( residual / _median_absolute_value(residual), scale, symmetric_weights ) params = {'weights': sqrt_w**2, 'tol_history': tol_history[:i + 1]} if return_coef: # TODO maybe leave out the coefficients from the rest of the calculations # since they are otherwise unused, and just fit x vs baseline here; would # save a little memory; is providing coefficients for loess even useful? params['coef'] = np.array([_convert_coef(coef, original_domain) for coef in coefs]) if sort_x: inverted_order = _inverted_sort(sort_order) baseline = baseline[inverted_order] params['weights'] = params['weights'][inverted_order] if return_coef: params['coef'] = params['coef'][inverted_order] return baseline, params def quant_reg(data, x_data=None, poly_order=2, quantile=0.05, tol=1e-6, max_iter=250, weights=None, eps=None, return_coef=False): """ Approximates the baseline of the data using quantile regression. Parameters ---------- data : array-like, shape (N,) The y-values of the measured data, with N data points. x_data : array-like, shape (N,), optional The x-values of the measured data. Default is None, which will create an array from -1 to 1 with N points. poly_order : int, optional The polynomial order for fitting the baseline. Default is 2. quantile : float, optional The quantile at which to fit the baseline. Default is 0.05. tol : float, optional The exit criteria. Default is 1e-6. For extreme quantiles (`quantile` < 0.01 or `quantile` > 0.99), may need to use a lower value to get a good fit. max_iter : int, optional The maximum number of iterations. Default is 250. For extreme quantiles (`quantile` < 0.01 or `quantile` > 0.99), may need to use a higher value to ensure convergence. weights : array-like, shape (N,), optional The weighting array. If None (default), then will be an array with size equal to N and all values set to 1. eps : float, optional A small value added to the square of the residual to prevent dividing by 0. Default is None, which uses the square of the maximum-absolute-value of the fit each iteration multiplied by 1e-6. return_coef : bool, optional If True, will convert the polynomial coefficients for the fit baseline to a form that fits the input `x_data` and return them in the params dictionary. Default is False, since the conversion takes time. Returns ------- baseline : numpy.ndarray, shape (N,) The calculated baseline. params : dict A dictionary with the following items: * 'weights': numpy.ndarray, shape (N,) The weight array used for fitting the data. * 'tol_history': numpy.ndarray An array containing the calculated tolerance values for each iteration. The length of the array is the number of iterations completed. If the last value in the array is greater than the input `tol` value, then the function did not converge. * 'coef': numpy.ndarray, shape (poly_order + 1,) Only if `return_coef` is True. The array of polynomial parameters for the baseline, in increasing order. Can be used to create a polynomial using numpy.polynomial.polynomial.Polynomial(). Raises ------ ValueError Raised if `quantile` is not between 0 and 1. Notes ----- Application of quantile regression for baseline fitting as described in [16]_. Performs quantile regression using iteratively reweighted least squares (IRLS) as described in [17]_. References ---------- .. [16] Komsta, Ł. Comparison of Several Methods of Chromatographic Baseline Removal with a New Approach Based on Quantile Regression. Chromatographia, 2011, 73, 721-731. .. [17] <NAME>., et al. Simultaneous estimation of quantile curves using quantile sheets. AStA Advances in Statistical Analysis, 2013, 97, 77-87. """ # TODO provide a way to estimate best poly_order based on AIC like in Komsta? could be # useful for all polynomial methods; maybe could be an optimizer function if not 0 < quantile < 1: raise ValueError('quantile must be between 0 and 1.') y, x, weight_array, original_domain, vander = _setup_polynomial( data, x_data, weights, poly_order, return_vander=True ) # estimate first iteration using least squares coef = np.linalg.lstsq(vander * weight_array[:, None], y * weight_array, None)[0] baseline = vander @ coef tol_history = np.empty(max_iter) for i in range(max_iter): baseline_old = baseline weight_array = np.sqrt(_weighting._quantile(y, baseline, quantile, eps)) coef = np.linalg.lstsq(vander * weight_array[:, None], y * weight_array, None)[0] baseline = vander @ coef # relative_difference(baseline_old, baseline, 1) gives nearly same result and # the l2 norm is faster to calculate, so use that instead of l1 norm calc_difference = relative_difference(baseline_old, baseline) tol_history[i] = calc_difference if calc_difference < tol: break params = {'weights': weight_array**2, 'tol_history': tol_history[:i + 1]} if return_coef: params['coef'] = _convert_coef(coef, original_domain) return baseline, params def goldindec(data, x_data=None, poly_order=2, tol=1e-3, max_iter=250, weights=None, cost_function='asymmetric_indec', peak_ratio=0.5, alpha_factor=0.99, tol_2=1e-3, tol_3=1e-6, max_iter_2=100, return_coef=False): """ Fits a polynomial baseline using a non-quadratic cost function. The non-quadratic cost functions penalize residuals with larger values, giving a more robust fit compared to normal least-squares. Parameters ---------- data : array-like, shape (N,) The y-values of the measured data, with N data points. x_data : array-like, shape (N,), optional The x-values of the measured data. Default is None, which will create an array from -1 to 1 with N points. poly_order : int, optional The polynomial order for fitting the baseline. Default is 2. tol : float, optional The exit criteria for the fitting with a given threshold value. Default is 1e-3. max_iter : int, optional The maximum number of iterations for fitting a threshold value. Default is 250. weights : array-like, shape (N,), optional The weighting array. If None (default), then will be an array with size equal to N and all values set to 1. cost_function : str, optional The non-quadratic cost function to minimize. Unlike :func:`.penalized_poly`, this function only works with asymmetric cost functions, so the symmetry prefix ('a' or 'asymmetric') is optional (eg. 'indec' and 'a_indec' are the same). Default is 'asymmetric_indec'. Available methods, and their associated reference, are: * 'asymmetric_indec'[18]_ * 'asymmetric_truncated_quadratic'[19]_ * 'asymmetric_huber'[19]_ peak_ratio : float, optional A value between 0 and 1 that designates how many points in the data belong to peaks. Values are valid within ~10% of the actual peak ratio. Default is 0.5. alpha_factor : float, optional A value between 0 and 1 that controls the value of the penalty. Default is 0.99. Typically should not need to change this value. tol_2 : float, optional The exit criteria for the difference between the optimal up-down ratio (number of points above 0 in the residual compared to number of points below 0) and the up-down ratio for a given threshold value. Default is 1e-3. tol_3 : float, optional The exit criteria for the relative change in the threshold value. Default is 1e-6. max_iter_2 : float, optional The number of iterations for iterating between different threshold values. Default is 100. return_coef : bool, optional If True, will convert the polynomial coefficients for the fit baseline to a form that fits the input x_data and return them in the params dictionary. Default is False, since the conversion takes time. Returns ------- baseline : numpy.ndarray, shape (N,) The calculated baseline. params : dict A dictionary with the following items: * 'weights': numpy.ndarray, shape (N,) The weight array used for fitting the data. * 'tol_history': numpy.ndarray, shape (J, K) An array containing the calculated tolerance values for each iteration of both threshold values and fit values. Index 0 are the tolerence values for the difference in up-down ratios, index 1 are the tolerance values for the relative change in the threshold, and indices >= 2 are the tolerance values for each fit. All values that were not used in fitting have values of 0. Shape J is 2 plus the number of iterations for the threshold to converge (related to `max_iter_2`, `tol_2`, `tol_3`), and shape K is the maximum of the number of iterations for the threshold and the maximum number of iterations for all of the fits of the various threshold values (related to `max_iter` and `tol`). * 'threshold' : float The optimal threshold value. Could be used in :func:`.penalized_poly` for fitting other similar data. * 'coef': numpy.ndarray, shape (poly_order + 1,) Only if `return_coef` is True. The array of polynomial parameters for the baseline, in increasing order. Can be used to create a polynomial using numpy.polynomial.polynomial.Polynomial(). Raises ------ ValueError Raised if `alpha_factor` or `peak_ratio` are not between 0 and 1, or if the specified cost function is symmetric. References ---------- .. [18] <NAME>., et al. Goldindec: A Novel Algorithm for Raman Spectrum Baseline Correction. Applied Spectroscopy, 2015, 69(7), 834-842. .. [19] <NAME>., et al. Background removal from spectra by designing and minimising a non-quadratic cost function. Chemometrics and Intelligent Laboratory Systems, 2005, 76(2), 121–133. """ if not 0 < alpha_factor <= 1: raise ValueError('alpha_factor must be between 0 and 1') elif not 0 < peak_ratio < 1: raise ValueError('peak_ratio must be between 0 and 1') try: symmetric_loss, method = _identify_loss_method(cost_function) except ValueError: # do not require a prefix since cost must be asymmetric symmetric_loss, method = _identify_loss_method('a_' + cost_function) if symmetric_loss: # symmetric cost functions don't work due to how the up-down ratio vs # peak_ratio function was created in the reference; in theory, could simulate # spectra with both positive and negative peaks following the reference # and build another empirical function, but would likely need to also # add other parameters detailing the percent of positive vs negative peaks, # etc., so it's not worth the effort raise ValueError('goldindec only works for asymmetric cost functions') loss_function = { 'huber': _huber_loss, 'truncated_quadratic': _truncated_quadratic_loss, 'indec': _indec_loss }[method] y, x, weight_array, original_domain, vander, pseudo_inverse = _setup_polynomial( data, x_data, weights, poly_order, return_vander=True, return_pinv=True ) num_y = y.shape[0] up_down_ratio_goal = ( 0.7679 + 11.2358 * peak_ratio - 39.7064 * peak_ratio**2 + 92.3583 * peak_ratio**3 ) # TODO reference states threshold must be <= 2 for half-quadratic minimization to # be valid for indec cost function, and normalized y so that threshold is always <= 2; # however, it seems to work fine without normalization; just be aware in case errors # occur, may have to normalize y in both this function and penalized_poly sqrt_w = np.sqrt(weight_array) y_fit = sqrt_w * y coef = np.dot(pseudo_inverse, y_fit) initial_baseline = np.dot(vander, coef) a = 0 # reference used b=1, but normalized y before fitting; instead, set b as max of # initial residual b = abs((y - initial_baseline).max()) threshold = a + 0.618 * (b - a) loss_kwargs = { 'threshold': threshold, 'alpha_factor': alpha_factor, 'symmetric': symmetric_loss } # have to use zeros rather than empty for tol_history since each inner fit may # have a different number of iterations tol_history = np.zeros((max_iter_2 + 2, max(max_iter, max_iter_2))) j_max = 0 for i in range(max_iter_2): baseline = initial_baseline for j in range(max_iter): baseline_old = baseline coef = np.dot( pseudo_inverse, y_fit + loss_function(y_fit - sqrt_w * baseline, **loss_kwargs) ) baseline = np.dot(vander, coef) calc_difference = relative_difference(baseline_old, baseline) tol_history[i + 2, j] = calc_difference if calc_difference < tol: break if j > j_max: j_max = j up_count = (y > baseline).sum() up_down_ratio = up_count / max(1, num_y - up_count) calc_difference = up_down_ratio - up_down_ratio_goal tol_history[0, i] = calc_difference if calc_difference > tol_2: a = threshold elif calc_difference < -tol_2: b = threshold else: break threshold = a + 0.618 * (b - a) # this exit criteria was not stated in the reference, but the change in threshold # becomes zero fairly quickly, so need to also exit rather than needlessly # continuing to calculate with the same threshold value calc_difference = relative_difference(loss_kwargs['threshold'], threshold) tol_history[1, i] = calc_difference if calc_difference < tol_3: break loss_kwargs['threshold'] = threshold params = { 'weights': weight_array, 'tol_history': tol_history[:i + 3, :max(i, j_max) + 1], 'threshold': loss_kwargs['threshold'] } if return_coef: params['coef'] = _convert_coef(coef, original_domain) return baseline, params
[ "numpy.ones_like", "numpy.abs", "numpy.sqrt", "numpy.minimum", "math.ceil", "numpy.argsort", "numpy.array", "numpy.dot", "numpy.zeros", "numpy.empty", "numpy.sign", "numpy.linalg.lstsq", "numpy.std", "warnings.warn", "numpy.maximum", "numpy.arange" ]
[((9204, 9225), 'numpy.sqrt', 'np.sqrt', (['weight_array'], {}), '(weight_array)\n', (9211, 9225), True, 'import numpy as np\n'), ((9274, 9308), 'numpy.dot', 'np.dot', (['pseudo_inverse', '(sqrt_w * y)'], {}), '(pseudo_inverse, sqrt_w * y)\n', (9280, 9308), True, 'import numpy as np\n'), ((9324, 9344), 'numpy.dot', 'np.dot', (['vander', 'coef'], {}), '(vander, coef)\n', (9330, 9344), True, 'import numpy as np\n'), ((9655, 9673), 'numpy.empty', 'np.empty', (['max_iter'], {}), '(max_iter)\n', (9663, 9673), True, 'import numpy as np\n'), ((13785, 13806), 'numpy.sqrt', 'np.sqrt', (['weight_array'], {}), '(weight_array)\n', (13792, 13806), True, 'import numpy as np\n'), ((13855, 13889), 'numpy.dot', 'np.dot', (['pseudo_inverse', '(sqrt_w * y)'], {}), '(pseudo_inverse, sqrt_w * y)\n', (13861, 13889), True, 'import numpy as np\n'), ((13905, 13925), 'numpy.dot', 'np.dot', (['vander', 'coef'], {}), '(vander, coef)\n', (13911, 13925), True, 'import numpy as np\n'), ((13942, 13962), 'numpy.std', 'np.std', (['(y - baseline)'], {}), '(y - baseline)\n', (13948, 13962), True, 'import numpy as np\n'), ((14167, 14185), 'numpy.empty', 'np.empty', (['max_iter'], {}), '(max_iter)\n', (14175, 14185), True, 'import numpy as np\n'), ((26700, 26721), 'numpy.sqrt', 'np.sqrt', (['weight_array'], {}), '(weight_array)\n', (26707, 26721), True, 'import numpy as np\n'), ((26753, 26778), 'numpy.dot', 'np.dot', (['pseudo_inverse', 'y'], {}), '(pseudo_inverse, y)\n', (26759, 26778), True, 'import numpy as np\n'), ((26794, 26814), 'numpy.dot', 'np.dot', (['vander', 'coef'], {}), '(vander, coef)\n', (26800, 26814), True, 'import numpy as np\n'), ((26833, 26851), 'numpy.empty', 'np.empty', (['max_iter'], {}), '(max_iter)\n', (26841, 26851), True, 'import numpy as np\n'), ((33845, 33860), 'numpy.empty', 'np.empty', (['num_x'], {}), '(num_x)\n', (33853, 33860), True, 'import numpy as np\n'), ((36211, 36242), 'numpy.empty', 'np.empty', (['(num_x, total_points)'], {}), '((num_x, total_points))\n', (36219, 36242), True, 'import numpy as np\n'), ((36258, 36273), 'numpy.empty', 'np.empty', (['num_x'], {}), '(num_x)\n', (36266, 36273), True, 'import numpy as np\n'), ((38394, 38409), 'numpy.empty', 'np.empty', (['num_x'], {}), '(num_x)\n', (38402, 38409), True, 'import numpy as np\n'), ((41977, 42012), 'numpy.empty', 'np.empty', (['(num_x, 2)'], {'dtype': 'np.intp'}), '((num_x, 2), dtype=np.intp)\n', (41985, 42012), True, 'import numpy as np\n'), ((52749, 52782), 'numpy.zeros', 'np.zeros', (['(num_x, poly_order + 1)'], {}), '((num_x, poly_order + 1))\n', (52757, 52782), True, 'import numpy as np\n'), ((52801, 52823), 'numpy.empty', 'np.empty', (['(max_iter + 1)'], {}), '(max_iter + 1)\n', (52809, 52823), True, 'import numpy as np\n'), ((52837, 52858), 'numpy.sqrt', 'np.sqrt', (['weight_array'], {}), '(weight_array)\n', (52844, 52858), True, 'import numpy as np\n'), ((58879, 58897), 'numpy.empty', 'np.empty', (['max_iter'], {}), '(max_iter)\n', (58887, 58897), True, 'import numpy as np\n'), ((66570, 66591), 'numpy.sqrt', 'np.sqrt', (['weight_array'], {}), '(weight_array)\n', (66577, 66591), True, 'import numpy as np\n'), ((66627, 66656), 'numpy.dot', 'np.dot', (['pseudo_inverse', 'y_fit'], {}), '(pseudo_inverse, y_fit)\n', (66633, 66656), True, 'import numpy as np\n'), ((66680, 66700), 'numpy.dot', 'np.dot', (['vander', 'coef'], {}), '(vander, coef)\n', (66686, 66700), True, 'import numpy as np\n'), ((9546, 9567), 'numpy.sqrt', 'np.sqrt', (['weight_array'], {}), '(weight_array)\n', (9553, 9567), True, 'import numpy as np\n'), ((9748, 9795), 'numpy.minimum', 'np.minimum', (['(y0 if use_original else y)', 'baseline'], {}), '(y0 if use_original else y, baseline)\n', (9758, 9795), True, 'import numpy as np\n'), ((9811, 9845), 'numpy.dot', 'np.dot', (['pseudo_inverse', '(sqrt_w * y)'], {}), '(pseudo_inverse, sqrt_w * y)\n', (9817, 9845), True, 'import numpy as np\n'), ((9865, 9885), 'numpy.dot', 'np.dot', (['vander', 'coef'], {}), '(vander, coef)\n', (9871, 9885), True, 'import numpy as np\n'), ((14058, 14079), 'numpy.sqrt', 'np.sqrt', (['weight_array'], {}), '(weight_array)\n', (14065, 14079), True, 'import numpy as np\n'), ((14228, 14297), 'numpy.minimum', 'np.minimum', (['(y0 if use_original else y)', '(baseline + num_std * deviation)'], {}), '(y0 if use_original else y, baseline + num_std * deviation)\n', (14238, 14297), True, 'import numpy as np\n'), ((14313, 14347), 'numpy.dot', 'np.dot', (['pseudo_inverse', '(sqrt_w * y)'], {}), '(pseudo_inverse, sqrt_w * y)\n', (14319, 14347), True, 'import numpy as np\n'), ((14367, 14387), 'numpy.dot', 'np.dot', (['vander', 'coef'], {}), '(vander, coef)\n', (14373, 14387), True, 'import numpy as np\n'), ((14412, 14432), 'numpy.std', 'np.std', (['(y - baseline)'], {}), '(y - baseline)\n', (14418, 14432), True, 'import numpy as np\n'), ((20273, 20290), 'numpy.sign', 'np.sign', (['residual'], {}), '(residual)\n', (20280, 20290), True, 'import numpy as np\n'), ((27028, 27048), 'numpy.dot', 'np.dot', (['vander', 'coef'], {}), '(vander, coef)\n', (27034, 27048), True, 'import numpy as np\n'), ((28659, 28691), 'numpy.maximum', 'np.maximum', (['(0)', '(1 - inner * inner)'], {}), '(0, 1 - inner * inner)\n', (28669, 28691), True, 'import numpy as np\n'), ((28720, 28742), 'numpy.ones_like', 'np.ones_like', (['residual'], {}), '(residual)\n', (28732, 28742), True, 'import numpy as np\n'), ((28834, 28866), 'numpy.maximum', 'np.maximum', (['(0)', '(1 - inner * inner)'], {}), '(0, 1 - inner * inner)\n', (28844, 28866), True, 'import numpy as np\n'), ((34084, 34112), 'numpy.abs', 'np.abs', (['(x[left:right] - x[i])'], {}), '(x[left:right] - x[i])\n', (34090, 34112), True, 'import numpy as np\n'), ((34293, 34338), 'numpy.sqrt', 'np.sqrt', (['(difference * difference * difference)'], {}), '(difference * difference * difference)\n', (34300, 34338), True, 'import numpy as np\n'), ((36497, 36525), 'numpy.abs', 'np.abs', (['(x[left:right] - x[i])'], {}), '(x[left:right] - x[i])\n', (36503, 36525), True, 'import numpy as np\n'), ((36706, 36751), 'numpy.sqrt', 'np.sqrt', (['(difference * difference * difference)'], {}), '(difference * difference * difference)\n', (36713, 36751), True, 'import numpy as np\n'), ((41141, 41171), 'numpy.empty', 'np.empty', (['num_x'], {'dtype': 'np.intp'}), '(num_x, dtype=np.intp)\n', (41149, 41171), True, 'import numpy as np\n'), ((41233, 41268), 'numpy.empty', 'np.empty', (['(num_x, 2)'], {'dtype': 'np.intp'}), '((num_x, 2), dtype=np.intp)\n', (41241, 41268), True, 'import numpy as np\n'), ((41928, 41961), 'numpy.array', 'np.array', (['[[0, 0]]'], {'dtype': 'np.intp'}), '([[0, 0]], dtype=np.intp)\n', (41936, 41961), True, 'import numpy as np\n'), ((51319, 51341), 'math.ceil', 'ceil', (['(fraction * num_x)'], {}), '(fraction * num_x)\n', (51323, 51341), False, 'from math import ceil\n'), ((51948, 51979), 'numpy.argsort', 'np.argsort', (['x'], {'kind': '"""mergesort"""'}), "(x, kind='mergesort')\n", (51958, 51979), True, 'import numpy as np\n'), ((58757, 58828), 'numpy.linalg.lstsq', 'np.linalg.lstsq', (['(vander * weight_array[:, None])', '(y * weight_array)', 'None'], {}), '(vander * weight_array[:, None], y * weight_array, None)\n', (58772, 58828), True, 'import numpy as np\n'), ((5537, 5558), 'numpy.sqrt', 'np.sqrt', (['weight_array'], {}), '(weight_array)\n', (5544, 5558), True, 'import numpy as np\n'), ((16471, 16487), 'numpy.abs', 'np.abs', (['residual'], {}), '(residual)\n', (16477, 16487), True, 'import numpy as np\n'), ((18482, 18498), 'numpy.abs', 'np.abs', (['residual'], {}), '(residual)\n', (18488, 18498), True, 'import numpy as np\n'), ((20224, 20240), 'numpy.abs', 'np.abs', (['residual'], {}), '(residual)\n', (20230, 20240), True, 'import numpy as np\n'), ((26555, 26564), 'numpy.std', 'np.std', (['y'], {}), '(y)\n', (26561, 26564), True, 'import numpy as np\n'), ((30109, 30123), 'numpy.abs', 'np.abs', (['values'], {}), '(values)\n', (30115, 30123), True, 'import numpy as np\n'), ((59056, 59127), 'numpy.linalg.lstsq', 'np.linalg.lstsq', (['(vander * weight_array[:, None])', '(y * weight_array)', 'None'], {}), '(vander * weight_array[:, None], y * weight_array, None)\n', (59071, 59127), True, 'import numpy as np\n'), ((67532, 67552), 'numpy.dot', 'np.dot', (['vander', 'coef'], {}), '(vander, coef)\n', (67538, 67552), True, 'import numpy as np\n'), ((16615, 16632), 'numpy.sign', 'np.sign', (['residual'], {}), '(residual)\n', (16622, 16632), True, 'import numpy as np\n'), ((41597, 41613), 'numpy.arange', 'np.arange', (['num_x'], {}), '(num_x)\n', (41606, 41613), True, 'import numpy as np\n'), ((51693, 51849), 'warnings.warn', 'warnings.warn', (['"""polynomial orders greater than 2 can have numerical issues; consider using a polynomial order of 1 or 2 instead"""', 'ParameterWarning'], {}), "(\n 'polynomial orders greater than 2 can have numerical issues; consider using a polynomial order of 1 or 2 instead'\n , ParameterWarning)\n", (51706, 51849), False, 'import warnings\n'), ((9499, 9519), 'numpy.std', 'np.std', (['(y - baseline)'], {}), '(y - baseline)\n', (9505, 9519), True, 'import numpy as np\n'), ((20651, 20692), 'numpy.maximum', 'np.maximum', (['(2 * residual ** 2)', '_MIN_FLOAT'], {}), '(2 * residual ** 2, _MIN_FLOAT)\n', (20661, 20692), True, 'import numpy as np\n'), ((53792, 53812), 'numpy.std', 'np.std', (['(y - baseline)'], {}), '(y - baseline)\n', (53798, 53812), True, 'import numpy as np\n')]
from django.urls import path from . import views urlpatterns = [ path('<int:sproperty_id>/',views.contact,name="contact"), path('',views.contact_submit,name="contact-submit"), ]
[ "django.urls.path" ]
[((75, 133), 'django.urls.path', 'path', (['"""<int:sproperty_id>/"""', 'views.contact'], {'name': '"""contact"""'}), "('<int:sproperty_id>/', views.contact, name='contact')\n", (79, 133), False, 'from django.urls import path\n'), ((137, 190), 'django.urls.path', 'path', (['""""""', 'views.contact_submit'], {'name': '"""contact-submit"""'}), "('', views.contact_submit, name='contact-submit')\n", (141, 190), False, 'from django.urls import path\n')]
#!/usr/bin/env python # # Copyright 2010 Google Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # from google.appengine.ext import db from google.appengine.ext import webapp from google.appengine.ext.webapp import util class IndexHost(db.Model): record_type = db.StringProperty() record_name = db.StringProperty() listed = db.BooleanProperty() class NameServer(db.Model): ip = db.StringProperty() hostname = db.StringProperty() name = db.StringProperty() listed = db.BooleanProperty() city = db.StringProperty() region = db.StringProperty() country = db.StringProperty() country_code = db.StringProperty() coordinates = db.GeoPtProperty() is_global = db.BooleanProperty() is_regional = db.BooleanProperty() is_custom = db.BooleanProperty() url = db.LinkProperty() timestamp = db.DateTimeProperty(auto_now_add=True) class Submission(db.Model): client_id = db.IntegerProperty() submit_id = db.IntegerProperty() class_c = db.StringProperty() timestamp = db.DateTimeProperty(auto_now_add=True) listed = db.BooleanProperty() hidden = db.BooleanProperty() city = db.StringProperty() region = db.StringProperty() country = db.StringProperty() country_code = db.StringProperty() coordinates = db.GeoPtProperty() # de-normalized data, also duplicated in RunResults (though much slower) best_nameserver = db.ReferenceProperty(NameServer, collection_name='best_submissions') best_improvement = db.FloatProperty() primary_nameserver = db.ReferenceProperty(NameServer, collection_name="primary_submissions") class SubmissionConfig(db.Model): submission = db.ReferenceProperty(Submission, collection_name='config') input_source = db.StringProperty() benchmark_thread_count = db.IntegerProperty() health_thread_count = db.IntegerProperty() health_timeout = db.FloatProperty() timeout = db.FloatProperty() query_count = db.IntegerProperty() run_count = db.IntegerProperty() platform = db.StringProperty() version = db.StringProperty() class SubmissionNameServer(db.Model): nameserver = db.ReferenceProperty(NameServer, collection_name='submissions') submission = db.ReferenceProperty(Submission, collection_name='nameservers') is_error_prone = db.BooleanProperty() is_disabled = db.BooleanProperty() is_reference = db.BooleanProperty() overall_average = db.FloatProperty() check_average = db.FloatProperty() averages = db.ListProperty(float) duration_min = db.FloatProperty() duration_max = db.FloatProperty() error_count = db.IntegerProperty() timeout_count = db.IntegerProperty() nx_count = db.IntegerProperty() position = db.IntegerProperty() sys_position = db.IntegerProperty() version = db.StringProperty() node_ids = db.ListProperty(str) # TODO(tstromberg): Remove obsoleted improvement variable improvement = db.FloatProperty() diff = db.FloatProperty() notes = db.ListProperty(str) port_behavior = db.StringProperty() # Store one row per run for run_results, since we do not need to do much with them. class RunResult(db.Model): submission_nameserver = db.ReferenceProperty(SubmissionNameServer, collection_name='results') run_number = db.IntegerProperty() durations = db.ListProperty(float) answer_counts = db.ListProperty(int) # We may want to compare index results, so we will store one row per record class IndexResult(db.Model): submission_nameserver = db.ReferenceProperty(SubmissionNameServer, collection_name='index_results') index_host = db.ReferenceProperty(IndexHost, collection_name='results') duration = db.FloatProperty() answer_count = db.IntegerProperty() ttl = db.IntegerProperty() response = db.StringProperty()
[ "google.appengine.ext.db.FloatProperty", "google.appengine.ext.db.BooleanProperty", "google.appengine.ext.db.GeoPtProperty", "google.appengine.ext.db.IntegerProperty", "google.appengine.ext.db.ReferenceProperty", "google.appengine.ext.db.ListProperty", "google.appengine.ext.db.DateTimeProperty", "goog...
[((766, 785), 'google.appengine.ext.db.StringProperty', 'db.StringProperty', ([], {}), '()\n', (783, 785), False, 'from google.appengine.ext import db\n'), ((802, 821), 'google.appengine.ext.db.StringProperty', 'db.StringProperty', ([], {}), '()\n', (819, 821), False, 'from google.appengine.ext import db\n'), ((833, 853), 'google.appengine.ext.db.BooleanProperty', 'db.BooleanProperty', ([], {}), '()\n', (851, 853), False, 'from google.appengine.ext import db\n'), ((890, 909), 'google.appengine.ext.db.StringProperty', 'db.StringProperty', ([], {}), '()\n', (907, 909), False, 'from google.appengine.ext import db\n'), ((923, 942), 'google.appengine.ext.db.StringProperty', 'db.StringProperty', ([], {}), '()\n', (940, 942), False, 'from google.appengine.ext import db\n'), ((952, 971), 'google.appengine.ext.db.StringProperty', 'db.StringProperty', ([], {}), '()\n', (969, 971), False, 'from google.appengine.ext import db\n'), ((983, 1003), 'google.appengine.ext.db.BooleanProperty', 'db.BooleanProperty', ([], {}), '()\n', (1001, 1003), False, 'from google.appengine.ext import db\n'), ((1013, 1032), 'google.appengine.ext.db.StringProperty', 'db.StringProperty', ([], {}), '()\n', (1030, 1032), False, 'from google.appengine.ext import db\n'), ((1044, 1063), 'google.appengine.ext.db.StringProperty', 'db.StringProperty', ([], {}), '()\n', (1061, 1063), False, 'from google.appengine.ext import db\n'), ((1076, 1095), 'google.appengine.ext.db.StringProperty', 'db.StringProperty', ([], {}), '()\n', (1093, 1095), False, 'from google.appengine.ext import db\n'), ((1113, 1132), 'google.appengine.ext.db.StringProperty', 'db.StringProperty', ([], {}), '()\n', (1130, 1132), False, 'from google.appengine.ext import db\n'), ((1149, 1167), 'google.appengine.ext.db.GeoPtProperty', 'db.GeoPtProperty', ([], {}), '()\n', (1165, 1167), False, 'from google.appengine.ext import db\n'), ((1182, 1202), 'google.appengine.ext.db.BooleanProperty', 'db.BooleanProperty', ([], {}), '()\n', (1200, 1202), False, 'from google.appengine.ext import db\n'), ((1219, 1239), 'google.appengine.ext.db.BooleanProperty', 'db.BooleanProperty', ([], {}), '()\n', (1237, 1239), False, 'from google.appengine.ext import db\n'), ((1254, 1274), 'google.appengine.ext.db.BooleanProperty', 'db.BooleanProperty', ([], {}), '()\n', (1272, 1274), False, 'from google.appengine.ext import db\n'), ((1283, 1300), 'google.appengine.ext.db.LinkProperty', 'db.LinkProperty', ([], {}), '()\n', (1298, 1300), False, 'from google.appengine.ext import db\n'), ((1315, 1353), 'google.appengine.ext.db.DateTimeProperty', 'db.DateTimeProperty', ([], {'auto_now_add': '(True)'}), '(auto_now_add=True)\n', (1334, 1353), False, 'from google.appengine.ext import db\n'), ((1400, 1420), 'google.appengine.ext.db.IntegerProperty', 'db.IntegerProperty', ([], {}), '()\n', (1418, 1420), False, 'from google.appengine.ext import db\n'), ((1435, 1455), 'google.appengine.ext.db.IntegerProperty', 'db.IntegerProperty', ([], {}), '()\n', (1453, 1455), False, 'from google.appengine.ext import db\n'), ((1468, 1487), 'google.appengine.ext.db.StringProperty', 'db.StringProperty', ([], {}), '()\n', (1485, 1487), False, 'from google.appengine.ext import db\n'), ((1502, 1540), 'google.appengine.ext.db.DateTimeProperty', 'db.DateTimeProperty', ([], {'auto_now_add': '(True)'}), '(auto_now_add=True)\n', (1521, 1540), False, 'from google.appengine.ext import db\n'), ((1552, 1572), 'google.appengine.ext.db.BooleanProperty', 'db.BooleanProperty', ([], {}), '()\n', (1570, 1572), False, 'from google.appengine.ext import db\n'), ((1584, 1604), 'google.appengine.ext.db.BooleanProperty', 'db.BooleanProperty', ([], {}), '()\n', (1602, 1604), False, 'from google.appengine.ext import db\n'), ((1614, 1633), 'google.appengine.ext.db.StringProperty', 'db.StringProperty', ([], {}), '()\n', (1631, 1633), False, 'from google.appengine.ext import db\n'), ((1645, 1664), 'google.appengine.ext.db.StringProperty', 'db.StringProperty', ([], {}), '()\n', (1662, 1664), False, 'from google.appengine.ext import db\n'), ((1677, 1696), 'google.appengine.ext.db.StringProperty', 'db.StringProperty', ([], {}), '()\n', (1694, 1696), False, 'from google.appengine.ext import db\n'), ((1714, 1733), 'google.appengine.ext.db.StringProperty', 'db.StringProperty', ([], {}), '()\n', (1731, 1733), False, 'from google.appengine.ext import db\n'), ((1750, 1768), 'google.appengine.ext.db.GeoPtProperty', 'db.GeoPtProperty', ([], {}), '()\n', (1766, 1768), False, 'from google.appengine.ext import db\n'), ((1867, 1935), 'google.appengine.ext.db.ReferenceProperty', 'db.ReferenceProperty', (['NameServer'], {'collection_name': '"""best_submissions"""'}), "(NameServer, collection_name='best_submissions')\n", (1887, 1935), False, 'from google.appengine.ext import db\n'), ((1957, 1975), 'google.appengine.ext.db.FloatProperty', 'db.FloatProperty', ([], {}), '()\n', (1973, 1975), False, 'from google.appengine.ext import db\n'), ((1999, 2070), 'google.appengine.ext.db.ReferenceProperty', 'db.ReferenceProperty', (['NameServer'], {'collection_name': '"""primary_submissions"""'}), "(NameServer, collection_name='primary_submissions')\n", (2019, 2070), False, 'from google.appengine.ext import db\n'), ((2121, 2179), 'google.appengine.ext.db.ReferenceProperty', 'db.ReferenceProperty', (['Submission'], {'collection_name': '"""config"""'}), "(Submission, collection_name='config')\n", (2141, 2179), False, 'from google.appengine.ext import db\n'), ((2199, 2218), 'google.appengine.ext.db.StringProperty', 'db.StringProperty', ([], {}), '()\n', (2216, 2218), False, 'from google.appengine.ext import db\n'), ((2246, 2266), 'google.appengine.ext.db.IntegerProperty', 'db.IntegerProperty', ([], {}), '()\n', (2264, 2266), False, 'from google.appengine.ext import db\n'), ((2291, 2311), 'google.appengine.ext.db.IntegerProperty', 'db.IntegerProperty', ([], {}), '()\n', (2309, 2311), False, 'from google.appengine.ext import db\n'), ((2331, 2349), 'google.appengine.ext.db.FloatProperty', 'db.FloatProperty', ([], {}), '()\n', (2347, 2349), False, 'from google.appengine.ext import db\n'), ((2362, 2380), 'google.appengine.ext.db.FloatProperty', 'db.FloatProperty', ([], {}), '()\n', (2378, 2380), False, 'from google.appengine.ext import db\n'), ((2397, 2417), 'google.appengine.ext.db.IntegerProperty', 'db.IntegerProperty', ([], {}), '()\n', (2415, 2417), False, 'from google.appengine.ext import db\n'), ((2432, 2452), 'google.appengine.ext.db.IntegerProperty', 'db.IntegerProperty', ([], {}), '()\n', (2450, 2452), False, 'from google.appengine.ext import db\n'), ((2466, 2485), 'google.appengine.ext.db.StringProperty', 'db.StringProperty', ([], {}), '()\n', (2483, 2485), False, 'from google.appengine.ext import db\n'), ((2498, 2517), 'google.appengine.ext.db.StringProperty', 'db.StringProperty', ([], {}), '()\n', (2515, 2517), False, 'from google.appengine.ext import db\n'), ((2574, 2637), 'google.appengine.ext.db.ReferenceProperty', 'db.ReferenceProperty', (['NameServer'], {'collection_name': '"""submissions"""'}), "(NameServer, collection_name='submissions')\n", (2594, 2637), False, 'from google.appengine.ext import db\n'), ((2653, 2716), 'google.appengine.ext.db.ReferenceProperty', 'db.ReferenceProperty', (['Submission'], {'collection_name': '"""nameservers"""'}), "(Submission, collection_name='nameservers')\n", (2673, 2716), False, 'from google.appengine.ext import db\n'), ((2738, 2758), 'google.appengine.ext.db.BooleanProperty', 'db.BooleanProperty', ([], {}), '()\n', (2756, 2758), False, 'from google.appengine.ext import db\n'), ((2775, 2795), 'google.appengine.ext.db.BooleanProperty', 'db.BooleanProperty', ([], {}), '()\n', (2793, 2795), False, 'from google.appengine.ext import db\n'), ((2813, 2833), 'google.appengine.ext.db.BooleanProperty', 'db.BooleanProperty', ([], {}), '()\n', (2831, 2833), False, 'from google.appengine.ext import db\n'), ((2854, 2872), 'google.appengine.ext.db.FloatProperty', 'db.FloatProperty', ([], {}), '()\n', (2870, 2872), False, 'from google.appengine.ext import db\n'), ((2891, 2909), 'google.appengine.ext.db.FloatProperty', 'db.FloatProperty', ([], {}), '()\n', (2907, 2909), False, 'from google.appengine.ext import db\n'), ((2923, 2945), 'google.appengine.ext.db.ListProperty', 'db.ListProperty', (['float'], {}), '(float)\n', (2938, 2945), False, 'from google.appengine.ext import db\n'), ((2963, 2981), 'google.appengine.ext.db.FloatProperty', 'db.FloatProperty', ([], {}), '()\n', (2979, 2981), False, 'from google.appengine.ext import db\n'), ((2999, 3017), 'google.appengine.ext.db.FloatProperty', 'db.FloatProperty', ([], {}), '()\n', (3015, 3017), False, 'from google.appengine.ext import db\n'), ((3034, 3054), 'google.appengine.ext.db.IntegerProperty', 'db.IntegerProperty', ([], {}), '()\n', (3052, 3054), False, 'from google.appengine.ext import db\n'), ((3073, 3093), 'google.appengine.ext.db.IntegerProperty', 'db.IntegerProperty', ([], {}), '()\n', (3091, 3093), False, 'from google.appengine.ext import db\n'), ((3107, 3127), 'google.appengine.ext.db.IntegerProperty', 'db.IntegerProperty', ([], {}), '()\n', (3125, 3127), False, 'from google.appengine.ext import db\n'), ((3141, 3161), 'google.appengine.ext.db.IntegerProperty', 'db.IntegerProperty', ([], {}), '()\n', (3159, 3161), False, 'from google.appengine.ext import db\n'), ((3179, 3199), 'google.appengine.ext.db.IntegerProperty', 'db.IntegerProperty', ([], {}), '()\n', (3197, 3199), False, 'from google.appengine.ext import db\n'), ((3212, 3231), 'google.appengine.ext.db.StringProperty', 'db.StringProperty', ([], {}), '()\n', (3229, 3231), False, 'from google.appengine.ext import db\n'), ((3245, 3265), 'google.appengine.ext.db.ListProperty', 'db.ListProperty', (['str'], {}), '(str)\n', (3260, 3265), False, 'from google.appengine.ext import db\n'), ((3342, 3360), 'google.appengine.ext.db.FloatProperty', 'db.FloatProperty', ([], {}), '()\n', (3358, 3360), False, 'from google.appengine.ext import db\n'), ((3370, 3388), 'google.appengine.ext.db.FloatProperty', 'db.FloatProperty', ([], {}), '()\n', (3386, 3388), False, 'from google.appengine.ext import db\n'), ((3399, 3419), 'google.appengine.ext.db.ListProperty', 'db.ListProperty', (['str'], {}), '(str)\n', (3414, 3419), False, 'from google.appengine.ext import db\n'), ((3438, 3457), 'google.appengine.ext.db.StringProperty', 'db.StringProperty', ([], {}), '()\n', (3455, 3457), False, 'from google.appengine.ext import db\n'), ((3596, 3665), 'google.appengine.ext.db.ReferenceProperty', 'db.ReferenceProperty', (['SubmissionNameServer'], {'collection_name': '"""results"""'}), "(SubmissionNameServer, collection_name='results')\n", (3616, 3665), False, 'from google.appengine.ext import db\n'), ((3681, 3701), 'google.appengine.ext.db.IntegerProperty', 'db.IntegerProperty', ([], {}), '()\n', (3699, 3701), False, 'from google.appengine.ext import db\n'), ((3716, 3738), 'google.appengine.ext.db.ListProperty', 'db.ListProperty', (['float'], {}), '(float)\n', (3731, 3738), False, 'from google.appengine.ext import db\n'), ((3757, 3777), 'google.appengine.ext.db.ListProperty', 'db.ListProperty', (['int'], {}), '(int)\n', (3772, 3777), False, 'from google.appengine.ext import db\n'), ((3910, 3985), 'google.appengine.ext.db.ReferenceProperty', 'db.ReferenceProperty', (['SubmissionNameServer'], {'collection_name': '"""index_results"""'}), "(SubmissionNameServer, collection_name='index_results')\n", (3930, 3985), False, 'from google.appengine.ext import db\n'), ((4001, 4059), 'google.appengine.ext.db.ReferenceProperty', 'db.ReferenceProperty', (['IndexHost'], {'collection_name': '"""results"""'}), "(IndexHost, collection_name='results')\n", (4021, 4059), False, 'from google.appengine.ext import db\n'), ((4073, 4091), 'google.appengine.ext.db.FloatProperty', 'db.FloatProperty', ([], {}), '()\n', (4089, 4091), False, 'from google.appengine.ext import db\n'), ((4109, 4129), 'google.appengine.ext.db.IntegerProperty', 'db.IntegerProperty', ([], {}), '()\n', (4127, 4129), False, 'from google.appengine.ext import db\n'), ((4138, 4158), 'google.appengine.ext.db.IntegerProperty', 'db.IntegerProperty', ([], {}), '()\n', (4156, 4158), False, 'from google.appengine.ext import db\n'), ((4172, 4191), 'google.appengine.ext.db.StringProperty', 'db.StringProperty', ([], {}), '()\n', (4189, 4191), False, 'from google.appengine.ext import db\n')]
""" Running operational space control with the PyGame display, using an exponential additive signal when to push away from joints. The target location can be moved by clicking on the background. """ import numpy as np from abr_control.arms import threejoint as arm # from abr_control.arms import twojoint as arm from abr_control.interfaces import PyGame from abr_control.controllers import OSC, AvoidJointLimits, Damping print('\nClick to move the target.\n') # initialize our robot config robot_config = arm.Config(use_cython=True) # create our arm simulation arm_sim = arm.ArmSim(robot_config) avoid = AvoidJointLimits( robot_config, min_joint_angles=[np.pi/5.0]*robot_config.N_JOINTS, max_joint_angles=[np.pi/2.0]*robot_config.N_JOINTS, max_torque=[100.0]*robot_config.N_JOINTS) # damp the movements of the arm damping = Damping(robot_config, kv=10) # create an operational space controller ctrlr = OSC(robot_config, kp=100, null_controllers=[avoid, damping], # control (x, y) out of [x, y, z, alpha, beta, gamma] ctrlr_dof=[True, True, False, False, False, False]) def on_click(self, mouse_x, mouse_y): self.target[0] = self.mouse_x self.target[1] = self.mouse_y # create our interface interface = PyGame(robot_config, arm_sim, dt=.001, on_click=on_click, q_init=[np.pi/4, np.pi/2, np.pi/2]) interface.connect() # create a target [x, y, z]] target_xyz = [0, 2, 0] # create a target orientation [alpha, beta, gamma] target_angles = [0, 0, 0] interface.set_target(target_xyz) try: print('\nSimulation starting...\n') count = 0 while 1: # get arm feedback feedback = interface.get_feedback() hand_xyz = robot_config.Tx('EE', feedback['q']) target = np.hstack([target_xyz, target_angles]) # generate an operational space control signal u = ctrlr.generate( q=feedback['q'], dq=feedback['dq'], target=target, ) new_target_xy = interface.get_mousexy() if new_target_xy is not None: target_xyz[:2] = new_target_xy interface.set_target(target_xyz) # apply the control signal, step the sim forward interface.send_forces( u, update_display=True if count % 20 == 0 else False) count += 1 finally: # stop and reset the simulation interface.disconnect() print('Simulation terminated...')
[ "numpy.hstack", "abr_control.controllers.OSC", "abr_control.interfaces.PyGame", "abr_control.controllers.AvoidJointLimits", "abr_control.arms.threejoint.Config", "abr_control.controllers.Damping", "abr_control.arms.threejoint.ArmSim" ]
[((509, 536), 'abr_control.arms.threejoint.Config', 'arm.Config', ([], {'use_cython': '(True)'}), '(use_cython=True)\n', (519, 536), True, 'from abr_control.arms import threejoint as arm\n'), ((575, 599), 'abr_control.arms.threejoint.ArmSim', 'arm.ArmSim', (['robot_config'], {}), '(robot_config)\n', (585, 599), True, 'from abr_control.arms import threejoint as arm\n'), ((609, 804), 'abr_control.controllers.AvoidJointLimits', 'AvoidJointLimits', (['robot_config'], {'min_joint_angles': '([np.pi / 5.0] * robot_config.N_JOINTS)', 'max_joint_angles': '([np.pi / 2.0] * robot_config.N_JOINTS)', 'max_torque': '([100.0] * robot_config.N_JOINTS)'}), '(robot_config, min_joint_angles=[np.pi / 5.0] *\n robot_config.N_JOINTS, max_joint_angles=[np.pi / 2.0] * robot_config.\n N_JOINTS, max_torque=[100.0] * robot_config.N_JOINTS)\n', (625, 804), False, 'from abr_control.controllers import OSC, AvoidJointLimits, Damping\n'), ((845, 873), 'abr_control.controllers.Damping', 'Damping', (['robot_config'], {'kv': '(10)'}), '(robot_config, kv=10)\n', (852, 873), False, 'from abr_control.controllers import OSC, AvoidJointLimits, Damping\n'), ((923, 1040), 'abr_control.controllers.OSC', 'OSC', (['robot_config'], {'kp': '(100)', 'null_controllers': '[avoid, damping]', 'ctrlr_dof': '[True, True, False, False, False, False]'}), '(robot_config, kp=100, null_controllers=[avoid, damping], ctrlr_dof=[\n True, True, False, False, False, False])\n', (926, 1040), False, 'from abr_control.controllers import OSC, AvoidJointLimits, Damping\n'), ((1258, 1363), 'abr_control.interfaces.PyGame', 'PyGame', (['robot_config', 'arm_sim'], {'dt': '(0.001)', 'on_click': 'on_click', 'q_init': '[np.pi / 4, np.pi / 2, np.pi / 2]'}), '(robot_config, arm_sim, dt=0.001, on_click=on_click, q_init=[np.pi / \n 4, np.pi / 2, np.pi / 2])\n', (1264, 1363), False, 'from abr_control.interfaces import PyGame\n'), ((1793, 1831), 'numpy.hstack', 'np.hstack', (['[target_xyz, target_angles]'], {}), '([target_xyz, target_angles])\n', (1802, 1831), True, 'import numpy as np\n')]
import itertools from unittest import TestCase import numpy as np from utils.data import ArrayInfo, image_array_to_rgb from utils.data.mappers import * class ImageUtilsTestCase(TestCase): def test_image_array_to_rgb(self): np.random.seed(1234) def f(batch_size, n_channels, channel_last, the_channel_last, use_info, bit_depth, dequantize, scale_to): shape = [31, 32] if n_channels is not None: if the_channel_last: shape = shape + [n_channels] else: shape = [n_channels] + shape the_info = ArrayInfo( shape=shape, min_val=0, max_val=255, is_discrete=True, n_discrete_vals=256, bit_depth=8) x = np.random.randint(0, 256, size=batch_size + shape) mappers = [] ans_mappers = None if bit_depth not in (None, 8): mappers.append(ReduceToBitDepth(bit_depth)) ans_mappers = ReduceToBitDepth(bit_depth) if dequantize: mappers.append(Dequantize(epsilon=1e-5)) if scale_to: mappers.append(ScaleToRange(*scale_to)) if mappers: m = ArrayMapperList(mappers) y_the_info = m.fit(the_info) y = m.transform(x) else: y_the_info = the_info y = x ans = x if ans_mappers is not None: ans_mappers.fit(the_info) ans = ans_mappers.transform(ans) if n_channels is None: ans = np.reshape(ans, ans.shape + (1,)) elif not the_channel_last: ans = np.transpose( ans, list(range(len(ans.shape) - 3)) + [-2, -1, -3] ) info = y_the_info if use_info else None out = image_array_to_rgb(y, info, channel_last) self.assertEqual(out.dtype, np.uint8) np.testing.assert_equal(out, ans) for (batch_size, n_channels, channel_last, the_channel_last, use_info, bit_depth, dequantize, scale_to) in itertools.product( ([], [7], [3, 4]), (None, 1, 3), (None, True, False), (True, False), (True, False), (8, 5), (True, False), (None, (0, 1), (-1, 1),), ): # skip inconsistent parameter combination if (n_channels is None and channel_last is not None) or \ (channel_last is not None and channel_last != the_channel_last): continue if n_channels is None and batch_size: continue # use_info = False is not supported along with dequantize or bit-depth if not use_info and (dequantize or bit_depth != 8): continue f(batch_size, n_channels, channel_last, the_channel_last, use_info, bit_depth, dequantize, scale_to)
[ "numpy.reshape", "numpy.testing.assert_equal", "itertools.product", "utils.data.ArrayInfo", "utils.data.image_array_to_rgb", "numpy.random.randint", "numpy.random.seed" ]
[((240, 260), 'numpy.random.seed', 'np.random.seed', (['(1234)'], {}), '(1234)\n', (254, 260), True, 'import numpy as np\n'), ((2216, 2375), 'itertools.product', 'itertools.product', (['([], [7], [3, 4])', '(None, 1, 3)', '(None, True, False)', '(True, False)', '(True, False)', '(8, 5)', '(True, False)', '(None, (0, 1), (-1, 1))'], {}), '(([], [7], [3, 4]), (None, 1, 3), (None, True, False), (\n True, False), (True, False), (8, 5), (True, False), (None, (0, 1), (-1, 1))\n )\n', (2233, 2375), False, 'import itertools\n'), ((639, 741), 'utils.data.ArrayInfo', 'ArrayInfo', ([], {'shape': 'shape', 'min_val': '(0)', 'max_val': '(255)', 'is_discrete': '(True)', 'n_discrete_vals': '(256)', 'bit_depth': '(8)'}), '(shape=shape, min_val=0, max_val=255, is_discrete=True,\n n_discrete_vals=256, bit_depth=8)\n', (648, 741), False, 'from utils.data import ArrayInfo, image_array_to_rgb\n'), ((787, 837), 'numpy.random.randint', 'np.random.randint', (['(0)', '(256)'], {'size': '(batch_size + shape)'}), '(0, 256, size=batch_size + shape)\n', (804, 837), True, 'import numpy as np\n'), ((1949, 1990), 'utils.data.image_array_to_rgb', 'image_array_to_rgb', (['y', 'info', 'channel_last'], {}), '(y, info, channel_last)\n', (1967, 1990), False, 'from utils.data import ArrayInfo, image_array_to_rgb\n'), ((2053, 2086), 'numpy.testing.assert_equal', 'np.testing.assert_equal', (['out', 'ans'], {}), '(out, ans)\n', (2076, 2086), True, 'import numpy as np\n'), ((1659, 1692), 'numpy.reshape', 'np.reshape', (['ans', '(ans.shape + (1,))'], {}), '(ans, ans.shape + (1,))\n', (1669, 1692), True, 'import numpy as np\n')]
# -*- coding: utf-8 -*- # closeness-server (c) <NAME> from closeness_server import create_app app = create_app()
[ "closeness_server.create_app" ]
[((101, 113), 'closeness_server.create_app', 'create_app', ([], {}), '()\n', (111, 113), False, 'from closeness_server import create_app\n')]
from sqlalchemy import Table, Column, Integer, String from sqlalchemy.orm import mapper from .database import metadata, db_session class User(object): query = db_session.query_property() def __init__(self, name=None, email=None): self.name = name self.email = email def __repr__(self): return '<User %r>' % (self.name) users = Table( 'users', metadata, Column('id', Integer, primary_key=True), Column('name', String(50), unique=True), Column('email', String(120), unique=True) ) mapper(User, users)
[ "sqlalchemy.orm.mapper", "sqlalchemy.String", "sqlalchemy.Column" ]
[((538, 557), 'sqlalchemy.orm.mapper', 'mapper', (['User', 'users'], {}), '(User, users)\n', (544, 557), False, 'from sqlalchemy.orm import mapper\n'), ((404, 443), 'sqlalchemy.Column', 'Column', (['"""id"""', 'Integer'], {'primary_key': '(True)'}), "('id', Integer, primary_key=True)\n", (410, 443), False, 'from sqlalchemy import Table, Column, Integer, String\n'), ((464, 474), 'sqlalchemy.String', 'String', (['(50)'], {}), '(50)\n', (470, 474), False, 'from sqlalchemy import Table, Column, Integer, String\n'), ((510, 521), 'sqlalchemy.String', 'String', (['(120)'], {}), '(120)\n', (516, 521), False, 'from sqlalchemy import Table, Column, Integer, String\n')]
import dash_bootstrap_components as dbc import dash_html_components as html """ Creation of jumbotrons for a better Homepage display """ left_jumbotron = dbc.Col( html.Div( [ html.H2("Your Favorite Finance Hub", className="display-3"), html.Hr(className="my-2"), html.P( "Get the latest news from the news agencies " "And track the live Stock & Crypto data. Free." ), ], className="h-100 p-5 text-white bg-dark rounded-1", ), md=12, ) right_jumbotron = dbc.Col( html.Div( [ html.Hr(className="my-2"), ], className="h-100 p-5 bg-light border rounded-3", ), md=16, ) jumbotron = dbc.Row( [left_jumbotron], className="align-items-md-stretch", ) second_jumbotron = dbc.Row(children=[right_jumbotron], className="align-items-md-stretch" )
[ "dash_bootstrap_components.Row", "dash_html_components.P", "dash_html_components.H2", "dash_html_components.Hr" ]
[((744, 805), 'dash_bootstrap_components.Row', 'dbc.Row', (['[left_jumbotron]'], {'className': '"""align-items-md-stretch"""'}), "([left_jumbotron], className='align-items-md-stretch')\n", (751, 805), True, 'import dash_bootstrap_components as dbc\n'), ((836, 907), 'dash_bootstrap_components.Row', 'dbc.Row', ([], {'children': '[right_jumbotron]', 'className': '"""align-items-md-stretch"""'}), "(children=[right_jumbotron], className='align-items-md-stretch')\n", (843, 907), True, 'import dash_bootstrap_components as dbc\n'), ((200, 259), 'dash_html_components.H2', 'html.H2', (['"""Your Favorite Finance Hub"""'], {'className': '"""display-3"""'}), "('Your Favorite Finance Hub', className='display-3')\n", (207, 259), True, 'import dash_html_components as html\n'), ((273, 298), 'dash_html_components.Hr', 'html.Hr', ([], {'className': '"""my-2"""'}), "(className='my-2')\n", (280, 298), True, 'import dash_html_components as html\n'), ((312, 420), 'dash_html_components.P', 'html.P', (['"""Get the latest news from the news agencies And track the live Stock & Crypto data. Free."""'], {}), "(\n 'Get the latest news from the news agencies And track the live Stock & Crypto data. Free.'\n )\n", (318, 420), True, 'import dash_html_components as html\n'), ((616, 641), 'dash_html_components.Hr', 'html.Hr', ([], {'className': '"""my-2"""'}), "(className='my-2')\n", (623, 641), True, 'import dash_html_components as html\n')]
#!/usr/bin/python # -*- coding: UTF-8 -*- from numpy import * import heapq class solutions: def childs_with_sugers(self,childs,sugers): """ 分糖给小朋友,一个小朋友只能拿一块糖,糖不能分割 :param childs: :param sugers: :return: """ childs=sorted(childs) sugers=sorted(sugers) res=[] # if len(childs) > len(sugers) : j=0 for i,child in enumerate(childs): while j < len(sugers): if sugers[j] >= child: res.append([child,sugers[j]]) j+=1 break j=j+1 else: # 正常结束 while 循环 则运行下面代码 print('suger is not enough') # 糖已经分配完了,有的小朋友没有糖吃 break # 跳出for 循环 return res def regions_overlap(self,regions,L): """ 区间覆盖问题 给定一个长度为 m的区间,再给出 n条线段的起点和终点(注意这里是闭区间), 求最少使用多少条线段可以将整个区间完全覆盖。 ref: https://www.cnblogs.com/acgoto/p/9824723.html :param regions: [ [2,6],[1,4],[3,6],[3,7],[6,8],[2,4],[3,5] ] :param L: 8 :return: [ [1,4] ,[3,7],[6,8] ] """ regions=sorted(regions, key=lambda d: d[0]) # 按照区间的 左端点进行排序 # print(regions) right_most=1 res=[] while right_most<L: left_small=list(filter(lambda x: x[0] <= right_most, regions)) # 过滤出左端点 小于 right_most 的区间 right_max=max(left_small,key=lambda x:x[1]) # 选这些区间 中 右端点最大的一个 res.append(right_max) right_most=right_max[1] #更新 已覆盖线段的 右端点 return res def max_regions_not_intersect(self, regions, L): """ 最多 不相交区间(活动选择问题) 假设我们有 n 个区间,区间的起始端点和结束端点分别是[l1, r1],[l2, r2],[l3, r3],……,[ln, rn]。我们从这 n 个区间中选出一部分区间, 这部分区间满足两两不相交(端点相交的情况不算相交),最多能选出多少个区间呢? ref: https://time.geekbang.org/column/article/73188 :param regions: [[6,8],[2,4],[3,5],[1,5],[5,9],[8,10]] :param L: 10 :return: [[2,4],[6,8],[8,10]] """ regions = sorted(regions, key=lambda d: d[0]) # 按照区间的 左端点进行排序 # print(regions) right_most=0 res=[] while right_most<L: left_small=list(filter(lambda x: x[0] >= right_most, regions)) # 过滤出左端点 大于 right_most 的区间 ,这样能避免重合 right_max=min(left_small,key=lambda x:x[1]) # 选这些区间 中 右端点最小的一个,这样 能留出更多的剩余空间 res.append(right_max) right_most=right_max[1] #更新 已覆盖线段的 右端点 return res def activity_selection(self, regions): """ 活动选择问题 假设我们有 n 个活动,活动的 开始端点和结束时间 分别是[l1, r1],[l2, r2],[l3, r3],……,[ln, rn] 我们从这 n 个 活动中选出一部分活动, 这部分活动的时间不冲突,即 满足 两两不相交(端点相交的情况不算相交),最多能选出多少个活动呢? 贪心策略: 结束时间 早的 活动 优先 :param regions: [[6,8],[2,4],[3,5],[1,5],[5,9],[8,10]] :return: [[2,4],[6,8],[8,10]] """ regions = sorted(regions, key=lambda d: d[1]) # 按照 结束时间 对活动 进行排序 selected=[] #被选中的活动 next_earliest_start=0 # 保证与上一个活动不冲突的 情况下,下一个活动的最早开始时间 for region in regions: start=region[0] end=region[1] if start >= next_earliest_start: # 活动的开始时间 满足 要求 selected.append(region) next_earliest_start=end return selected def minimum_delay_scheduling(self, duration, deadline): """ 最小延迟 调度 问题 n 项任务,每一项任务 消耗的时间为 duration,每一项任务的 截止时间为 deadline 任务超过截止时间才完成会产生延迟 求 所有任务中 发生最大延迟的任务 所产生的延迟 达到最小的 调度策略,在这一策略下的 最大延迟 贪心策略: 截止时间 早的任务优先 :param duration: [5,8,4,10,3] :param deadline: [10,12,15,11,20] :return: """ N=len(duration) # 任务个数 deadline=array(deadline) jobs_deadline= [ (i,deadline[i]) for i in range(N)] # (任务标号, 任务截止时间) jobs_deadline = sorted(jobs_deadline, key=lambda ele: ele[1]) jobs_delay = zeros(N, dtype=int) # 记录 各个任务的 延迟时间 current_time=0 # 记录 当前的时间点 for job in jobs_deadline: job_NO=job[0] job_deadline=job[1] current_time+=duration[job_NO] if current_time > job_deadline: # 时间超过了 deadline jobs_delay[job_NO]= current_time-job_deadline print('jobs_delay:',jobs_delay) max_delay=max(jobs_delay) return max_delay def bulk_bag_problem(self, weights,values, capacity): """ 散装背包问题 (背包问题变形) 与普通背包问题 的关键区别:每件物品均可分解 一个可以容纳 100kg 物品的背包, 有 5 种豆子,每种豆子的 总量 和 总价值都各不相同。 为了让背包中所装物品的总价值最大,我们如何选择在背包中装哪些豆子?每种豆子又该装多少呢 采用贪心策略:尽可能多放 单位重量价值 最大的 物品 :param weights: [100,30,60,20,50] :param values: [100,90,120,80,75] :param capacity: 100 :return: """ N=len(weights) # 物品种类 weights=array(weights) values=array(values) unit_weight_value= values/weights unit_weight_value= [ (i,unit_weight_value[i]) for i in range(N)] # (标号,物品的 单位重量的价值) unit_weight_value= sorted(unit_weight_value,key=lambda ele:ele[1],reverse=True) #按照 物品单位重量的价值 逆序排序 [(3, 4.0), (1, 3.0), (2, 2.0), (4, 1.5), (0, 1.0)] bag_weight=capacity bag_value=0 bag_items= zeros(N, dtype=float) # 背包中 每一样物品的 重量 for item in unit_weight_value: if bag_weight <=0: break item_NO=item[0] if bag_weight >= weights[item_NO]: # 背包 容量 足以装下全部的 item_NO 物品 bag_weight -= weights[item_NO] # bag_value += values[item_NO] bag_items[item_NO]=weights[item_NO] else: # 背包 容量 不足 bag_items[item_NO]=bag_weight # 剩下的背包容量 全部装 item_NO 物品 bag_weight = 0 # bag_value += bag_weight*item[1] return bag_value,bag_items def optimal_loading(self, weights, capacity): """ 最优装载问题 (背包问题变形) 即 物品的价值都为 1 的 01 背包问题 n 个集装箱 重量为 weights, 船的载重能力限制为 capacity, 每一个集装箱 都小于 capacity, 如何 选择 使得 可以装载更多 的集装箱 采用贪心策略:轻的物品优先装入 :param weights: [100,30,60,20,50] :param capacity: 100 :return: """ N=len(weights) # 物品种类 weights=array(weights) items_weight= [ (i,weights[i]) for i in range(N)] # (标号,物品的重量) items_weight = sorted(items_weight, key=lambda ele: ele[1]) bag_weight=capacity bag_items= [] # 记录 放入背包的物品 for item in items_weight: if bag_weight <=0: break item_NO=item[0] if bag_weight >= item[1] : # 背包 容量 能装下 item_NO 物品 bag_weight -= weights[item_NO] bag_items.append(item_NO) return bag_items class ComapreHeap(object): def __init__(self, initial=None, key=lambda x: x): self.key = key if initial: self._data = [(key(item), item) for item in initial] heapq.heapify(self._data) else: self._data = [] def push(self, item): heapq.heappush(self._data, (self.key(item), item)) def pop(self): return heapq.heappop(self._data)[1] class TreeNode(object): def __init__(self,key=None,value=None): self.key=key self.value=value self.left=None self.right=None class huffman_tree: """ 霍夫曼 前缀编码 使用 {0,1} 按照字符出现的 频率,并根据贪心策略 生成二叉编码树对字符集进行前缀编码 ref: (1)《算法导论》 (2)https://time.geekbang.org/column/article/73188 """ def __init__(self, char_list): self.huffman_encode_tree=self.__encode(char_list) self.__decode_all() def encode(self, string): """ 将字符串 编码为 以 {0,1} 表示的 字节流 :param string: 'acbf' :return: '01001011100' """ bytes=[] for char in string: bytes.append(self.char_dict[char]) return ''.join(bytes) def decode(self,bytes): """ 解码一段 以 {0,1} 表示的 字节流 :param bytes: '01001011100' :return: 'acbf' """ root = self.huffman_encode_tree p=root res_string=[] for byte in bytes: if byte=='0': # 走左子树 p=p.left else: # 走右子树 p=p.right #TODO : 解码失败的处理 if p.left==None and p.right==None: # 走到叶子节点了 res_string.append(p.key) p=root return ''.join(res_string) def __decode_all(self): """ 返回 霍夫曼 编码树 上 char_list 中所有字符 和其 对应的 编码 :return: char_dict= {'a': '0', 'c': '100', 'b': '101', 'f': '1100', 'e': '1101', 'd': '111'} """ root=self.huffman_encode_tree self.char_dict={} self.__tree_pre_order(root,[]) for char,char_bytes in self.char_dict.items(): self.char_dict[char]=''.join(str(e) for e in char_bytes) return self.char_dict def __tree_pre_order(self,root,pre_list): if root.left== None and root.right == None: # 说明到达叶子节点 self.char_dict[root.key]=pre_list else: if root.left !=None: self.__tree_pre_order(root.left,pre_list+[0]) if root.right!=None: self.__tree_pre_order(root.right, pre_list + [1]) def __encode(self,char_list): """ 将 char_list 中的字符,生成一颗 huffman 编码树 :param char_list: [('a',45),('b',13),('c',12),('d',16),('e',9),('f',5)] :return: """ leaf_nodes= [TreeNode(ele[0] ,ele[1]) for ele in char_list ] heap_nodes=ComapreHeap(leaf_nodes,key=lambda x:x.value) N=len(char_list) root_node=None for i in range(N-1): # N 为叶节点个数,要执行N-1次的 叶节点的合并操作 root_node=TreeNode() left_node=heap_nodes.pop() right_node=heap_nodes.pop() root_node.key='s'+str(i) # 非叶子节点的 Key root_node.value=left_node.value+ right_node.value root_node.left=left_node root_node.right=right_node # print('root:',root_node.value) # print('root.left:', root_node.left.value) # print('root.right:', root_node.right.value) heap_nodes.push(root_node) return root_node if __name__ == '__main__': sol = solutions() childs=[3,4,5,6,7,8] # 小孩0-小孩5 想要的糖果的 重量 sugers=[1,2,3,4,5] # 现有的各个糖果的 重量 # print(sol.childs_with_sugers(childs,sugers)) regions=[ [2,6],[1,4],[3,6],[3,7],[6,8],[2,4],[3,5] ] L=8 # print(sol.regions_overlap(regions,L)) regions=[[6,8],[2,4],[3,5],[1,5],[5,9],[8,10]] L=10 # print(sol.max_regions_not_intersect(regions, L)) # print(sol.activity_selection(regions)) duration= [5, 8, 4, 10, 3] deadline= [10, 12, 15, 11, 20] # print(sol.minimum_delay_scheduling(duration,deadline)) weights= [100, 30, 60, 20, 50] values= [100, 90, 120, 80, 75] capacity= 100 # print(sol.bulk_bag_problem(weights,values,capacity)) # print(sol.optimal_loading(weights,capacity)) char_list=[('a',45),('b',13),('c',12),('d',16),('e',9),('f',5)] huffman_tree=huffman_tree(char_list) print('char_dict:',huffman_tree.char_dict) bytes= '01001011100' print(huffman_tree.decode(bytes)) print(huffman_tree.encode('acbf'))
[ "heapq.heappop", "heapq.heapify" ]
[((7105, 7130), 'heapq.heapify', 'heapq.heapify', (['self._data'], {}), '(self._data)\n', (7118, 7130), False, 'import heapq\n'), ((7294, 7319), 'heapq.heappop', 'heapq.heappop', (['self._data'], {}), '(self._data)\n', (7307, 7319), False, 'import heapq\n')]
# -*- coding: utf-8 -*- # flake8: noqa # noreorder """ Pytube: a very serious Python library for downloading YouTube Videos. """ __title__ = 'my_pytube' __version__ = '9.5.2' __author__ = '<NAME>' __license__ = 'MIT License' __copyright__ = 'Copyright 2019 <NAME>' #import logging #import query #import streams #import captions #import contrib #import __main__ from my_pytube.logging import create_logger from my_pytube.query import CaptionQuery from my_pytube.query import StreamQuery from my_pytube.streams import Stream from my_pytube.captions import Caption from my_pytube.contrib.playlist import Playlist from my_pytube.__main__ import YouTube logger = create_logger() logger.info('%s v%s', __title__, __version__)
[ "my_pytube.logging.create_logger" ]
[((663, 678), 'my_pytube.logging.create_logger', 'create_logger', ([], {}), '()\n', (676, 678), False, 'from my_pytube.logging import create_logger\n')]
## @ingroup Methods-Aerodynamics-Airfoil_Panel_Method # panel_geometry.py # Created: Mar 2021, <NAME> # --------------------------------------- #------------------------------- # Imports # ---------------------------------------------------------------------- import SUAVE from SUAVE.Core import Units import numpy as np # ---------------------------------------------------------------------- # panel_geometry.py # ---------------------------------------------------------------------- ## @ingroup Methods-Aerodynamics-Airfoil_Panel_Method def panel_geometry(x,y,npanel,nalpha,nRe): """Computes airfoil surface panelization parameters for later use in the computation of the matrix of influence coefficients. Assumptions: None Source: None Inputs: x - Vector of x coordinates of the surface nodes [unitless] y - Vector of y coordinates of the surface nodes [unitless] npanel - Number of panels on the airfoil [unitless] Outputs: l - Panel lengths [unitless] st - np.sin(theta) for each panel [radians] ct - np.cos(theta) for each panel [radians] xbar - x-coordinate of the midpoint of each panel [unitless] ybar - y-coordinate of the midpoint of each panel [unitless] Properties Used: N/A """ # compute various geometrical quantities l = np.sqrt((x[1:] -x[:-1])**2 +(y[1:] -y[:-1])**2) st = (y[1:] -y[:-1])/l ct = (x[1:] -x[:-1])/l xbar = (x[1:] +x[:-1])/2 ybar = (y[1:] +y[:-1])/2 norm = np.zeros((npanel,2,nalpha,nRe)) norm[:,0,:,:] = -st norm[:,1,:,:] = ct return l,st,ct,xbar,ybar,norm
[ "numpy.zeros", "numpy.sqrt" ]
[((1778, 1832), 'numpy.sqrt', 'np.sqrt', (['((x[1:] - x[:-1]) ** 2 + (y[1:] - y[:-1]) ** 2)'], {}), '((x[1:] - x[:-1]) ** 2 + (y[1:] - y[:-1]) ** 2)\n', (1785, 1832), True, 'import numpy as np\n'), ((1962, 1996), 'numpy.zeros', 'np.zeros', (['(npanel, 2, nalpha, nRe)'], {}), '((npanel, 2, nalpha, nRe))\n', (1970, 1996), True, 'import numpy as np\n')]
#!/usr/bin/python """ (dummy-)Mutates multiple residues on a PDB-formatted structure. HADDOCK will then reconstruct the residue according to its topology. Usage: python pdb_multimutate.py pdbFL <mutation list file> The format of mutation list: chain resi resn_wt resn_mut Example: python pdb_multimutate.py 3mrb.pdb list_mutations for mutating residue 7 Serine of chain A to Alanine: A 7 SER ALA Author: {0} Email: {1} """ from __future__ import print_function import os import sys import re __author__ = "<NAME>; <NAME>; <NAME>" __email__ = "<EMAIL>; <EMAIL>; <EMAIL>" USAGE = __doc__.format(__author__, __email__) def check_input(args): """Checks whether to read from stdin/file and validates user input/options.""" if len(args) == 2: if not os.path.isfile(args[0]): sys.stderr.write('File not found: ' + args[0] + '\n') sys.stderr.write(USAGE) sys.exit(1) else: sys.stderr.write(USAGE) sys.exit(1) def mutate(structure_fhandle, chain, resi, resn_wt, resn_mut): mutated_structure = [] flag = 0 # 0: residue-to-be-mutated NOT found in structure; 1 found. atom_set = set(["CA", "C", "O", "N", "CB"]) # keep main chain atoms and CB atom of side chain for line in structure_fhandle: if line[0:4] == 'ATOM' or line[0:6] == 'HETATM' or line[0:6] == 'ANISOU': s_chain = line[21].strip() s_resi = line[22:26].strip() s_resn = line[17:20].strip() s_atom = line[12:16].strip() if s_chain == chain and s_resi == resi and s_resn == resn_wt: flag = 1 if s_atom in atom_set: line = line[0:17]+resn_mut+line[20:] else: continue mutated_structure.append(line) if flag ==0: sys.stderr.write('WARNING: ' + chain + ":" + resi +":" + resn_wt + ' does NOT exist in the strcuture\n') return mutated_structure def _print_mutants(pdbFL, mutationFL): # read pdb file into memory f_pdb = open(pdbFL, 'r') structure = [l for l in f_pdb] f_pdb.close() new_pdbFL = os.path.splitext(pdbFL)[0] + '_mutated.pdb' # mutate pdb file print("Generated mutant files:") f_mut = open(mutationFL, 'r') for line in f_mut: line = re.sub('[\n\r]','', line) i = line.split() if len(i) == 4: chain, resi, resn_wt, resn_mut = i print (i) else: sys.stderr.write('WARNING: Unrecognized mutation format in line "{0}"\n'.format(" ".join(i))) continue structure = mutate(structure, chain, resi, resn_wt, resn_mut) if structure: m_file = open(new_pdbFL, 'w') print(new_pdbFL + ' generated') m_file.write(''.join(structure)) m_file.close() if __name__ == "__main__": check_input(sys.argv[1:]) pdbFL = sys.argv[1] mutationFL = sys.argv[2] _print_mutants(pdbFL, mutationFL)
[ "os.path.splitext", "os.path.isfile", "sys.stderr.write", "sys.exit", "re.sub" ]
[((970, 993), 'sys.stderr.write', 'sys.stderr.write', (['USAGE'], {}), '(USAGE)\n', (986, 993), False, 'import sys\n'), ((1002, 1013), 'sys.exit', 'sys.exit', (['(1)'], {}), '(1)\n', (1010, 1013), False, 'import sys\n'), ((1869, 1981), 'sys.stderr.write', 'sys.stderr.write', (["('WARNING: ' + chain + ':' + resi + ':' + resn_wt +\n ' does NOT exist in the strcuture\\n')"], {}), '(\'WARNING: \' + chain + \':\' + resi + \':\' + resn_wt +\n """ does NOT exist in the strcuture\n""")\n', (1885, 1981), False, 'import sys\n'), ((2352, 2378), 're.sub', 're.sub', (["'[\\n\\r]'", '""""""', 'line'], {}), "('[\\n\\r]', '', line)\n", (2358, 2378), False, 'import re\n'), ((801, 824), 'os.path.isfile', 'os.path.isfile', (['args[0]'], {}), '(args[0])\n', (815, 824), False, 'import os\n'), ((838, 891), 'sys.stderr.write', 'sys.stderr.write', (["('File not found: ' + args[0] + '\\n')"], {}), "('File not found: ' + args[0] + '\\n')\n", (854, 891), False, 'import sys\n'), ((904, 927), 'sys.stderr.write', 'sys.stderr.write', (['USAGE'], {}), '(USAGE)\n', (920, 927), False, 'import sys\n'), ((940, 951), 'sys.exit', 'sys.exit', (['(1)'], {}), '(1)\n', (948, 951), False, 'import sys\n'), ((2176, 2199), 'os.path.splitext', 'os.path.splitext', (['pdbFL'], {}), '(pdbFL)\n', (2192, 2199), False, 'import os\n')]
import marshal from multiprocessing import Condition, Process, Queue, Pipe import os from threading import Timer from types import FunctionType import pickle from celery import Celery from flask import Flask, url_for from grams.grams import Histogram from grams.markov import MC import time def make_app(): def make_model(): def _make_model(corpus, n_sentences=10): # global cv def _generate(): markovchain = MC(corpus) return markovchain.generate generate = _generate() child_conn.send(generate(n_sentences)) while True: if parent_conn.poll(): ## previously sent message got consumed # send another child_conn.send(generate(n_sentences)) parent_conn, child_conn = Pipe(duplex=True) with open("res/the_adventures_of_sherlock_holmes.txt", "r") as f: f_out = f.read() make_process = Process(target=_make_model, args=(f_out,)) make_process.start() return parent_conn, make_process # init app flask_app = Flask(__name__) parent_conn, make_process = make_model() @flask_app.route("/") def home(): if parent_conn.poll(): return parent_conn.recv() return "loading..." return flask_app flask_app = make_app() if __name__ == "__main__": flask_app.run(debug=True, port=8080)
[ "multiprocessing.Process", "multiprocessing.Pipe", "grams.markov.MC", "flask.Flask" ]
[((1155, 1170), 'flask.Flask', 'Flask', (['__name__'], {}), '(__name__)\n', (1160, 1170), False, 'from flask import Flask, url_for\n'), ((864, 881), 'multiprocessing.Pipe', 'Pipe', ([], {'duplex': '(True)'}), '(duplex=True)\n', (868, 881), False, 'from multiprocessing import Condition, Process, Queue, Pipe\n'), ((1009, 1051), 'multiprocessing.Process', 'Process', ([], {'target': '_make_model', 'args': '(f_out,)'}), '(target=_make_model, args=(f_out,))\n', (1016, 1051), False, 'from multiprocessing import Condition, Process, Queue, Pipe\n'), ((468, 478), 'grams.markov.MC', 'MC', (['corpus'], {}), '(corpus)\n', (470, 478), False, 'from grams.markov import MC\n')]
import abc import numbers from typing import Union import numpy as np from river import base, optim, utils VectorLike = Union[utils.VectorDict, np.ndarray] __all__ = ["Initializer", "Scheduler", "Optimizer", "Loss"] class Initializer(base.Base, abc.ABC): """An initializer is used to set initial weights in a model.""" @abc.abstractmethod def __call__(self, shape=1): """Returns a fresh set of weights. Parameters ---------- shape Indicates how many weights to return. If `1`, then a single scalar value will be returned. """ class Scheduler(base.Base, abc.ABC): """Can be used to program the learning rate schedule of an `optim.base.Optimizer`.""" @abc.abstractmethod def get(self, t: int) -> float: """Returns the learning rate at a given iteration. Parameters ---------- t The iteration number. """ def __repr__(self): return f"{self.__class__.__name__}({vars(self)})" class Optimizer(base.Base): """Optimizer interface. Every optimizer inherits from this base interface. Parameters ---------- lr Attributes ---------- learning_rate : float Returns the current learning rate value. """ def __init__(self, lr: Union[Scheduler, float]): if isinstance(lr, numbers.Number): lr = optim.schedulers.Constant(lr) self.lr = lr self.n_iterations = 0 @property def learning_rate(self) -> float: return self.lr.get(self.n_iterations) def look_ahead(self, w: dict) -> dict: """Updates a weight vector before a prediction is made. Parameters: w (dict): A dictionary of weight parameters. The weights are modified in-place. Returns: The updated weights. """ return w def _step_with_dict(self, w: dict, g: dict) -> dict: raise NotImplementedError def _step_with_vector(self, w: VectorLike, g: VectorLike) -> VectorLike: raise NotImplementedError def step( self, w: Union[dict, VectorLike], g: Union[dict, VectorLike] ) -> Union[dict, VectorLike]: """Updates a weight vector given a gradient. Parameters ---------- w A vector-like object containing weights. The weights are modified in-place. g A vector-like object of gradients. Returns ------- The updated weights. """ if isinstance(w, VectorLike.__args__) and isinstance(g, VectorLike.__args__): try: w = self._step_with_vector(w, g) self.n_iterations += 1 return w except NotImplementedError: pass w = self._step_with_dict(w, g) self.n_iterations += 1 return w def __repr__(self): return f"{self.__class__.__name__}({vars(self)})" class Loss(base.Base, abc.ABC): """Base class for all loss functions.""" def __repr__(self): return f"{self.__class__.__name__}({vars(self)})" @abc.abstractmethod def __call__(self, y_true, y_pred): """Returns the loss. Parameters ---------- y_true Ground truth(s). y_pred Prediction(s). Returns ------- The loss(es). """ @abc.abstractmethod def gradient(self, y_true, y_pred): """Return the gradient with respect to y_pred. Parameters ---------- y_true Ground truth(s). y_pred Prediction(s). Returns ------- The gradient(s). """ @abc.abstractmethod def mean_func(self, y_pred): """Mean function. This is the inverse of the link function. Typically, a loss function takes as input the raw output of a model. In the case of classification, the raw output would be logits. The mean function can be used to convert the raw output into a value that makes sense to the user, such as a probability. Parameters ---------- y_pred Raw prediction(s). Returns ------- The adjusted prediction(s). References ---------- [^1]: [Wikipedia section on link and mean function](https://www.wikiwand.com/en/Generalized_linear_model#/Link_function) """
[ "river.optim.schedulers.Constant" ]
[((1421, 1450), 'river.optim.schedulers.Constant', 'optim.schedulers.Constant', (['lr'], {}), '(lr)\n', (1446, 1450), False, 'from river import base, optim, utils\n')]
#!/usr/bin/env python3 """Quantize and convert a keras model to tensorflowjs model.""" import argparse import pathlib import tempfile import numpy as np import tensorflow as tf import tensorflowjs as tfjs def main() -> None: """Execute the main routine.""" ## # Parse command-line arguments ## parser = argparse.ArgumentParser(description=__doc__) parser.add_argument( "--keras_path", help="Path to the keras model", required=True) parser.add_argument( "--tfjs_dir", help="Path to the tensorflowjs model", required=True) args = parser.parse_args() keras_path = pathlib.Path(args.keras_path) tfjs_dir = pathlib.Path(args.tfjs_dir) ## # Check command-line arguments ## if not keras_path.exists(): raise FileNotFoundError("Keras model does not exist: {}".format( keras_path)) if not keras_path.is_file(): raise RuntimeError("Expected keras model to be a file: {}".format( keras_path)) if tfjs_dir.exists() and not tfjs_dir.is_dir(): raise NotADirectoryError( "Expected tfjs model path to be a directory: {}".format( tfjs_dir)) ## # Prepare ## tfjs_dir.mkdir(exist_ok=True, parents=True) ## # Quantize and convert # # (paraphrased from tensorflowjs/converters/converter.py ## model = tf.keras.models.load_model(str(keras_path)) with tempfile.TemporaryDirectory( prefix='savedmodel', dir=str(tfjs_dir)) as temp_savedmodel_dir: tf.keras.experimental.export_saved_model( model, str(temp_savedmodel_dir), serving_only=True) quantization_dtype = np.uint8 # quantize to 1-byte skip_op_check = False # default strip_debug_ops = False # default tfjs.converters.tf_saved_model_conversion_v2.convert_tf_saved_model( str(temp_savedmodel_dir), str(tfjs_dir), signature_def='serving_default', saved_model_tags='serve', quantization_dtype=quantization_dtype, skip_op_check=skip_op_check, strip_debug_ops=strip_debug_ops) if __name__ == "__main__": main()
[ "argparse.ArgumentParser", "pathlib.Path" ]
[((329, 373), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {'description': '__doc__'}), '(description=__doc__)\n', (352, 373), False, 'import argparse\n'), ((654, 683), 'pathlib.Path', 'pathlib.Path', (['args.keras_path'], {}), '(args.keras_path)\n', (666, 683), False, 'import pathlib\n'), ((699, 726), 'pathlib.Path', 'pathlib.Path', (['args.tfjs_dir'], {}), '(args.tfjs_dir)\n', (711, 726), False, 'import pathlib\n')]
from flask import render_template_string from werkzeug.test import EnvironBuilder, run_wsgi_app class TestTemplateGlobals: def test_static_mode(self, single_use_app): @single_use_app.route("/test-globals") def test_globals(): return render_template_string("""{{ static_mode }}""") client = single_use_app.test_client() single_use_app.config["STATIC_MODE"] = False assert client.get("/test-globals").get_data(as_text=True) == "False" single_use_app.config["STATIC_MODE"] = True assert client.get("/test-globals").get_data(as_text=True) == "True" assert client.get("/test-globals?static_mode=no").get_data(as_text=True) == "False" assert client.get("/test-globals?static_mode=yes").get_data(as_text=True) == "True" class TestHeaders: def test_http_headers(self, single_use_app): builder = EnvironBuilder(path="/", method="GET") env = builder.get_environ() (app_iter, status, headers) = run_wsgi_app(single_use_app, env) assert headers.get("X-Frame-Options") == "deny" assert headers.get("X-Content-Type-Options") == "nosniff" assert headers.get("X-XSS-Protection") == "1; mode=block" assert headers.get("X-Permitted-Cross-Domain-Policies") == "none" assert headers.get("Content-Security-Policy")
[ "flask.render_template_string", "werkzeug.test.EnvironBuilder", "werkzeug.test.run_wsgi_app" ]
[((895, 933), 'werkzeug.test.EnvironBuilder', 'EnvironBuilder', ([], {'path': '"""/"""', 'method': '"""GET"""'}), "(path='/', method='GET')\n", (909, 933), False, 'from werkzeug.test import EnvironBuilder, run_wsgi_app\n'), ((1009, 1042), 'werkzeug.test.run_wsgi_app', 'run_wsgi_app', (['single_use_app', 'env'], {}), '(single_use_app, env)\n', (1021, 1042), False, 'from werkzeug.test import EnvironBuilder, run_wsgi_app\n'), ((267, 310), 'flask.render_template_string', 'render_template_string', (['"""{{ static_mode }}"""'], {}), "('{{ static_mode }}')\n", (289, 310), False, 'from flask import render_template_string\n')]
#!/usr/bin/env python ######################################################################## # RSA2ELK, by <NAME> # Converts Netwitness log parser configuration to Logstash configuration # see https://github.com/blookot/rsa2elk ######################################################################## import config import re import csv import sys # carriage return for logstash conf, just \n for linux CR = "\n" # insert n times a tab def t(n): t="" for i in range(0,n): t=t+"\t" return t # Replace all RSA date syntax with LS date filter one def convertDate(s): dateList = set() pattern = re.compile(",'([^']+)'") for dStr in pattern.finditer(s): c = dStr.group(1) # first, replace all non coding (not having a %) chars, like T, cf https://www.elastic.co/guide/en/logstash/current/plugins-filters-date.html#plugins-filters-date-match c = re.sub("([^%])([a-zA-Z]+)", r"\1'\2'", c) # replace the specific chars by their logstash date filter equivalent c = c.replace("%C", "M/d/yy H:m:s") c = c.replace("%R", "MMMM") c = c.replace("%B", "MMM") c = c.replace("%M", "MM") c = c.replace("%G", "M") c = c.replace("%D", "dd") c = c.replace("%F", "d") c = c.replace("%H", "HH") c = c.replace("%I", "HH") c = c.replace("%N", "H") c = c.replace("%T", "mm") c = c.replace("%U", "m") c = c.replace("%J", "D") c = c.replace("%P", "a") c = c.replace("%S", "ss") c = c.replace("%O", "s") c = c.replace("%Y", "yy") c = c.replace("%W", "yyyy") c = c.replace("%Z", "H:m:s") c = c.replace("%A", "D") c = c.replace("%Q", "a") # AM/PM c = c.replace("%K", "") # undocumented... seen in cef.xml c = c.replace("%L", "") # undocumented... seen in cef.xml c = c.replace("%E", "") # undocumented... seen in v20_trendmicromsg.xml c = c.replace("%X", "UNIX") if '%' in c: print("Missing a condition in date conversion: " + c) dateList.add(c) return "\"" + "\", \"".join(dateList) + "\"" # converting STRCAT def convertStrcat(s): c = "" regex = re.compile("^[a-zA-Z]+.*") # just saying the string starts with a letter # grab first ( iFirstPar = s.find("(") iEndPar = s.find(")", iFirstPar) if iFirstPar > 0 and iEndPar > 0: idParts = str.split(s[iFirstPar+1:iEndPar], ",") # check for static strings vs fields for idPart in idParts: if "'" in idPart: # catenate a string c = c + str.strip(idPart.replace("'", "")) elif "\"" in idPart: # catenate a string c = c + str.strip(idPart.r("\"", "")) elif regex.match(idPart.strip()): # first character is [a-z] ie a field c = c + "%{" + idPart.strip() + "}" elif idPart == "\t": # just a tab c = c + "\\t" else: # any other char, just append c = c + idPart.strip() return c else: if config.DEBUG: print("Couldn't parse STRCAT string") return "" # escaping " in grok content, and adding anchors if passed as param def escapeGrok(s): if s == "": return s if config.NO_GROK_ANCHORS: return "\"" + str.strip(s.replace("\"","\\\"")) + "\"" else: return "\"^" + str.strip(s.replace("\"","\\\"")) + "$\"" # escaping " in dissect content def escapeDissect(s): if s == "": return s return "\"" + str.strip(s.replace("\"","\\\"")) + "\"" # escapes a string in logstash (only escaping ") def escapeString(s): return str.strip(s.replace("\"","\\\"")) # escape special characters in grok : \ . ^ $ * + - ? ( ) [ ] { } | def escapeRegex(s): # s = str(s,'utf-8') s = s.replace("\\", "\\\\"); s = s.replace(".", "\\."); s = s.replace("^", "\\^"); s = s.replace("$", "\\$") s = s.replace("*", "\\*"); s = s.replace("+", "\\+"); s = s.replace("-", "\\-"); s = s.replace("?", "\\?") s = s.replace("(", "\\("); s = s.replace(")", "\\)"); s = s.replace("[", "\\["); s = s.replace("]", "\\]") s = s.replace("{", "\\{"); s = s.replace("}", "\\}"); s = s.replace("|", "\\|") s = s.replace(chr(9), "\\t"); s = s.replace(chr(10), "\\n"); s = s.replace(chr(13), "\\r"); s = s.replace(" ", "\\s") return s # remove dots in field names def removeDots(fldName): if "." in fldName: return "[" + fldName.replace(".","][") + "]" else: return fldName # transform a nested field into a nested es mapping def generateFieldMapping(fldName, fldType): # either key.subkey as RSA format if "." in fldName: # split on . flds = fldName.split(".") if len(flds) == 2: config.esMap["mappings"]["properties"][flds[0]]["properties"][flds[1]] = {"type": fldType} elif len(flds) == 3: config.esMap["mappings"]["properties"][flds[0]]["properties"][flds[1]]["properties"][flds[2]] = {"type": fldType} elif len(flds) == 4: config.esMap["mappings"]["properties"][flds[0]]["properties"][flds[1]]["properties"][flds[2]]["properties"][flds[3]] = {"type": fldType} elif len(flds) == 5: config.esMap["mappings"]["properties"][flds[0]]["properties"][flds[1]]["properties"][flds[2]]["properties"][flds[3]]["properties"][flds[4]] = {"type": fldType} else: print("Error: are you really using more than 5 levels of nested field?") sys.exit(-1) # or [key][subkey] after mutate elif "][" in fldName: # split on ][ flds = fldName[1:-1].split("][") if len(flds) == 2: config.esMap["mappings"]["properties"][flds[0]]["properties"][flds[1]] = {"type": fldType} elif len(flds) == 3: config.esMap["mappings"]["properties"][flds[0]]["properties"][flds[1]]["properties"][flds[2]] = {"type": fldType} elif len(flds) == 4: config.esMap["mappings"]["properties"][flds[0]]["properties"][flds[1]]["properties"][flds[2]]["properties"][flds[3]] = {"type": fldType} elif len(flds) == 5: config.esMap["mappings"]["properties"][flds[0]]["properties"][flds[1]]["properties"][flds[2]]["properties"][flds[3]]["properties"][flds[4]] = {"type": fldType} else: print("Error: are you really using more than 5 levels of nested field?") sys.exit(-1) # plain key else: config.esMap["mappings"]["properties"][fldName] = {"type": fldType} # read table-map and populate dicts def initMapping(): noComma = False; noSemiColon = False # trying to open table map file with , delimiter first (french way!) with open(config.MAPPING_FILE,'r') as csvFile: try: reader = csv.DictReader(csvFile, delimiter=',') for row in reader: # if field is flagged as "Transient", we don't take it into consideration if row['ecsName'] != "" and row['flags'] != "Transient": config.ecsField[row['envisionName']] = row['ecsName'] config.ecsType[row['envisionName']] = row['ecsType'] except: noComma = True if noComma: with open(config.MAPPING_FILE,'r') as csvFile: try: reader = csv.DictReader(csvFile, delimiter=';') for row in reader: # TODO: not sure what I should do with the fields marked "Transient" if row['ecsName'] != "": # and row['flags'] != "Transient": config.ecsField[row['envisionName']] = row['ecsName'] config.ecsType[row['envisionName']] = row['ecsType'] except: noSemiColon = True if noComma and noSemiColon: print("CSV mapping file expecting ; or , delimiter. Exiting...") sys.exit(-1) # get a possible valuemap func and store the fields def getValueMap(fld,vmFunc): # form is *getEventCategoryActivity(action) pattern = re.compile("\*([^\(]+)\(([^\)]+)\)") m = pattern.match(vmFunc) if m is not None: k,v = m.group(1),m.group(2) # let's look for the func in the valuemap funcs we've stored if k in config.valueMap: # record the func parameter as the key config.valueMap[k]["fld"] = v # record the destination field as well config.valueMap[k]["newFld"] = fld config.allFields.add(fld)
[ "csv.DictReader", "re.compile", "config.allFields.add", "sys.exit", "re.sub" ]
[((625, 649), 're.compile', 're.compile', (['""",\'([^\']+)\'"""'], {}), '(",\'([^\']+)\'")\n', (635, 649), False, 'import re\n'), ((2228, 2254), 're.compile', 're.compile', (['"""^[a-zA-Z]+.*"""'], {}), "('^[a-zA-Z]+.*')\n", (2238, 2254), False, 'import re\n'), ((8064, 8105), 're.compile', 're.compile', (['"""\\\\*([^\\\\(]+)\\\\(([^\\\\)]+)\\\\)"""'], {}), "('\\\\*([^\\\\(]+)\\\\(([^\\\\)]+)\\\\)')\n", (8074, 8105), False, 'import re\n'), ((902, 944), 're.sub', 're.sub', (['"""([^%])([a-zA-Z]+)"""', '"""\\\\1\'\\\\2\'"""', 'c'], {}), '(\'([^%])([a-zA-Z]+)\', "\\\\1\'\\\\2\'", c)\n', (908, 944), False, 'import re\n'), ((7907, 7919), 'sys.exit', 'sys.exit', (['(-1)'], {}), '(-1)\n', (7915, 7919), False, 'import sys\n'), ((6805, 6843), 'csv.DictReader', 'csv.DictReader', (['csvFile'], {'delimiter': '""","""'}), "(csvFile, delimiter=',')\n", (6819, 6843), False, 'import csv\n'), ((8494, 8519), 'config.allFields.add', 'config.allFields.add', (['fld'], {}), '(fld)\n', (8514, 8519), False, 'import config\n'), ((7341, 7379), 'csv.DictReader', 'csv.DictReader', (['csvFile'], {'delimiter': '""";"""'}), "(csvFile, delimiter=';')\n", (7355, 7379), False, 'import csv\n'), ((5570, 5582), 'sys.exit', 'sys.exit', (['(-1)'], {}), '(-1)\n', (5578, 5582), False, 'import sys\n'), ((6435, 6447), 'sys.exit', 'sys.exit', (['(-1)'], {}), '(-1)\n', (6443, 6447), False, 'import sys\n')]
# -*- coding: utf-8 -*- import os import shutil import io import zipfile from commons.Utils import path_utils class ZipHelper(object): @classmethod def _get_arcname(cls, old_arcname, new_arcname): if old_arcname is not None: return os.path.join(old_arcname, new_arcname) else: return new_arcname @classmethod def zip(cls, src, dest, arcname=None): if os.path.isdir(src): for root, filepath_list, filename_list in os.walk(src): for filepath in filepath_list: ZipHelper.zip( src=os.path.join(root, filepath), dest=dest, arcname=ZipHelper._get_arcname(arcname, filepath) ) for filename in filename_list: with zipfile.ZipFile(dest, 'a', zipfile.ZIP_DEFLATED) as zf: zf.write(os.path.join(root, filename), arcname=ZipHelper._get_arcname(arcname, filename)) break elif os.path.isfile(src): filepath, filename = os.path.split(src) with zipfile.ZipFile(dest, 'a', zipfile.ZIP_DEFLATED) as zf: zf.write(src, arcname=ZipHelper._get_arcname(arcname, filename)) @classmethod def zip_file(cls, file, zipname=None, exclude_parent=False): if not os.path.isfile(file): return False folder, filename = os.path.split(file) if exclude_parent: base_path, pathname = path_utils.get_path_split(folder) if zipname is None: zipname = os.path.join(base_path, "{0}.zip".format(pathname)) ZipHelper.zip(file, zipname, arcname=None) else: filename_base, filename_ext = os.path.splitext(filename) if zipname is None: zipname = os.path.join(folder, "{0}.zip".format(filename_base)) ZipHelper.zip(file, zipname, arcname=None) return True @classmethod def zip_folder(cls, folder, zipname=None, exclude_parent=False): if not os.path.isdir(folder): return False base_path, pathname = path_utils.get_path_split(folder) if zipname is None: zipname = os.path.join(base_path, "{0}.zip".format(pathname)) if exclude_parent: ZipHelper.zip(folder, zipname, arcname=pathname) else: ZipHelper.zip(folder, zipname, arcname=None) return True @classmethod def unzip(cls, zipname, dest=None): """服务器端解压文件生成r_list.dat文件""" zf = zipfile.ZipFile(zipname, 'r') dest = dest.replace('\\', '/') if dest.endswith('/'): dest = dest[:-1] for filename in zf.namelist(): fname = filename new_file = '%s/%s' % (dest, fname) if new_file.endswith('/') or new_file.endswith('\\'): if not os.path.exists(new_file): os.makedirs(new_file) else: filepath = os.path.split(new_file)[0] if not os.path.exists(filepath): os.makedirs(filepath) try: f = open(new_file, 'wb') except UnicodeEncodeError: f = open(new_file.encode('utf8'), 'wb') f.write(zf.read(filename)) f.close() class InMemoryZip(object): """用法: imz = InMemoryZip() imz.append("test.txt", "Another test").append("test2.txt", "Still another") imz.writetofile("test.zip") """ def __init__(self): # Create the in-memory file-like object self.in_memory_zip = io.StringIO() def append(self, filename_in_zip, file_contents): """Appends a file with name filename_in_zip and contents of file_contents to the in-memory zip.""" # Get a handle to the in-memory zip in append mode zf = zipfile.ZipFile(self.in_memory_zip, "a", zipfile.ZIP_DEFLATED, False) # Write the file to the in-memory zip zf.writestr(filename_in_zip, file_contents) # Mark the files as having been created on Windows so that # Unix permissions are not inferred as 0000 for zfile in zf.filelist: zfile.create_system = 0 return self def read(self): '''Returns a string with the contents of the in-memory zip.''' self.in_memory_zip.seek(0) return self.in_memory_zip.read() def writetofile(self, filename): '''Writes the in-memory zip to a file.''' with open(filename, "wb") as f: f.write(self.read())
[ "os.path.exists", "zipfile.ZipFile", "os.makedirs", "os.path.join", "os.path.splitext", "os.path.split", "os.path.isfile", "os.path.isdir", "commons.Utils.path_utils.get_path_split", "io.StringIO", "os.walk" ]
[((419, 437), 'os.path.isdir', 'os.path.isdir', (['src'], {}), '(src)\n', (432, 437), False, 'import os\n'), ((1407, 1426), 'os.path.split', 'os.path.split', (['file'], {}), '(file)\n', (1420, 1426), False, 'import os\n'), ((2139, 2172), 'commons.Utils.path_utils.get_path_split', 'path_utils.get_path_split', (['folder'], {}), '(folder)\n', (2164, 2172), False, 'from commons.Utils import path_utils\n'), ((2564, 2593), 'zipfile.ZipFile', 'zipfile.ZipFile', (['zipname', '"""r"""'], {}), "(zipname, 'r')\n", (2579, 2593), False, 'import zipfile\n'), ((3652, 3665), 'io.StringIO', 'io.StringIO', ([], {}), '()\n', (3663, 3665), False, 'import io\n'), ((3908, 3977), 'zipfile.ZipFile', 'zipfile.ZipFile', (['self.in_memory_zip', '"""a"""', 'zipfile.ZIP_DEFLATED', '(False)'], {}), "(self.in_memory_zip, 'a', zipfile.ZIP_DEFLATED, False)\n", (3923, 3977), False, 'import zipfile\n'), ((263, 301), 'os.path.join', 'os.path.join', (['old_arcname', 'new_arcname'], {}), '(old_arcname, new_arcname)\n', (275, 301), False, 'import os\n'), ((493, 505), 'os.walk', 'os.walk', (['src'], {}), '(src)\n', (500, 505), False, 'import os\n'), ((1007, 1026), 'os.path.isfile', 'os.path.isfile', (['src'], {}), '(src)\n', (1021, 1026), False, 'import os\n'), ((1332, 1352), 'os.path.isfile', 'os.path.isfile', (['file'], {}), '(file)\n', (1346, 1352), False, 'import os\n'), ((1489, 1522), 'commons.Utils.path_utils.get_path_split', 'path_utils.get_path_split', (['folder'], {}), '(folder)\n', (1514, 1522), False, 'from commons.Utils import path_utils\n'), ((1744, 1770), 'os.path.splitext', 'os.path.splitext', (['filename'], {}), '(filename)\n', (1760, 1770), False, 'import os\n'), ((2060, 2081), 'os.path.isdir', 'os.path.isdir', (['folder'], {}), '(folder)\n', (2073, 2081), False, 'import os\n'), ((1061, 1079), 'os.path.split', 'os.path.split', (['src'], {}), '(src)\n', (1074, 1079), False, 'import os\n'), ((1097, 1145), 'zipfile.ZipFile', 'zipfile.ZipFile', (['dest', '"""a"""', 'zipfile.ZIP_DEFLATED'], {}), "(dest, 'a', zipfile.ZIP_DEFLATED)\n", (1112, 1145), False, 'import zipfile\n'), ((2897, 2921), 'os.path.exists', 'os.path.exists', (['new_file'], {}), '(new_file)\n', (2911, 2921), False, 'import os\n'), ((2943, 2964), 'os.makedirs', 'os.makedirs', (['new_file'], {}), '(new_file)\n', (2954, 2964), False, 'import os\n'), ((3010, 3033), 'os.path.split', 'os.path.split', (['new_file'], {}), '(new_file)\n', (3023, 3033), False, 'import os\n'), ((3060, 3084), 'os.path.exists', 'os.path.exists', (['filepath'], {}), '(filepath)\n', (3074, 3084), False, 'import os\n'), ((3106, 3127), 'os.makedirs', 'os.makedirs', (['filepath'], {}), '(filepath)\n', (3117, 3127), False, 'import os\n'), ((802, 850), 'zipfile.ZipFile', 'zipfile.ZipFile', (['dest', '"""a"""', 'zipfile.ZIP_DEFLATED'], {}), "(dest, 'a', zipfile.ZIP_DEFLATED)\n", (817, 850), False, 'import zipfile\n'), ((617, 645), 'os.path.join', 'os.path.join', (['root', 'filepath'], {}), '(root, filepath)\n', (629, 645), False, 'import os\n'), ((891, 919), 'os.path.join', 'os.path.join', (['root', 'filename'], {}), '(root, filename)\n', (903, 919), False, 'import os\n')]